kopyl commited on
Commit
096c926
1 Parent(s): 8fae579

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +183 -0
  2. .gitignore +0 -11
  3. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/HP2-人声vocals+非人声instrumentals.pth +3 -0
  4. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/Retrieval_based_Voice_Conversion_WebUI.ipynb +381 -0
  5. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/hubert_base.pt +3 -0
  6. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/.DS_Store +0 -0
  7. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/kanye/D_2333333.pth +3 -0
  8. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/kanye/G_2333333.pth +3 -0
  9. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/kanye/added_IVF875_Flat_nprobe_7.index +3 -0
  10. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/kanye/total_fea.npy +3 -0
  11. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/.DS_Store +0 -0
  12. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/0_gt_wavs/mute32k.wav +0 -0
  13. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/0_gt_wavs/mute40k.wav +0 -0
  14. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/0_gt_wavs/mute48k.wav +0 -0
  15. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/1_16k_wavs/mute.wav +0 -0
  16. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/2a_f0/mute.wav.npy +3 -0
  17. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/2b-f0nsf/mute.wav.npy +3 -0
  18. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/3_feature256/mute.npy +3 -0
  19. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/.DS_Store +0 -0
  20. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/.gitignore +2 -0
  21. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/D32k.pth +3 -0
  22. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/D40k.pth +3 -0
  23. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/D48k.pth +3 -0
  24. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/G32k.pth +3 -0
  25. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/G40k.pth +3 -0
  26. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/G48k.pth +3 -0
  27. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/f0D32k.pth +3 -0
  28. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/f0D40k.pth +3 -0
  29. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/f0G32k.pth +3 -0
  30. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/f0G40k.pth +3 -0
  31. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/f0G48k.pth +3 -0
  32. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/test-voice.mp3 +3 -0
  33. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/weights/.gitignore +2 -0
  34. docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/weights/kanye.pth +3 -0
  35. docker-no-splitter/models/.DS_Store +0 -0
  36. docker-no-splitter/models/ariana-grande.pth +3 -0
  37. docker-no-splitter/models/drake.pth +3 -0
  38. docker-no-splitter/models/iu.pth +3 -0
  39. docker-no-splitter/models/jungkook.pth +3 -0
  40. docker-no-splitter/models/kanye.pth +3 -0
  41. docker-no-splitter/models/rose-bp.pth +3 -0
  42. docker-no-splitter/pretrained_models/.DS_Store +0 -0
  43. docker-no-splitter/pretrained_models/2stems/.probe +1 -0
  44. docker-no-splitter/pretrained_models/2stems/checkpoint +2 -0
  45. docker-no-splitter/pretrained_models/2stems/model.data-00000-of-00001 +3 -0
  46. docker-no-splitter/pretrained_models/2stems/model.index +0 -0
  47. docker-no-splitter/pretrained_models/2stems/model.meta +0 -0
  48. docker/Retrieval-based-Voice-Conversion-WebUI/HP2-人声vocals+非人声instrumentals.pth +3 -0
  49. docker/Retrieval-based-Voice-Conversion-WebUI/Retrieval_based_Voice_Conversion_WebUI.ipynb +381 -0
  50. docker/Retrieval-based-Voice-Conversion-WebUI/hubert_base.pt +3 -0
.gitattributes CHANGED
@@ -33,3 +33,186 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ docker/Retrieval-based-Voice-Conversion-WebUI/logs/kanye/added_IVF875_Flat_nprobe_7.index filter=lfs diff=lfs merge=lfs -text
37
+ docker/Retrieval-based-Voice-Conversion-WebUI/test-voice.mp3 filter=lfs diff=lfs merge=lfs -text
38
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/_soundfile_data/libsndfile_x86_64.so filter=lfs diff=lfs merge=lfs -text
39
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/clang/native/libclang.so filter=lfs diff=lfs merge=lfs -text
40
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/google/protobuf/pyext/_message.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
41
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/grpc/_cython/cygrpc.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
42
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/h5py/h5t.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
43
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/h5py.libs/libhdf5-45610c27.so.200.2.0 filter=lfs diff=lfs merge=lfs -text
44
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/llvmlite/binding/libllvmlite.so filter=lfs diff=lfs merge=lfs -text
45
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/numpy/core/_multiarray_umath.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
46
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/numpy/core/_simd.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
47
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/numpy/random/_generator.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
48
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/numpy.libs/libgfortran-040039e1.so.5.0.0 filter=lfs diff=lfs merge=lfs -text
49
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/numpy.libs/libopenblas64_p-r0-2f7c42d4.3.18.so filter=lfs diff=lfs merge=lfs -text
50
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/pandas/_libs/algos.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
51
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/pandas/_libs/groupby.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
52
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/pandas/_libs/hashtable.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
53
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/pandas/_libs/interval.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
54
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/pandas/_libs/join.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
55
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/scipy/fft/_pocketfft/pypocketfft.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
56
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/scipy/linalg/_flapack.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
57
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/scipy/misc/face.dat filter=lfs diff=lfs merge=lfs -text
58
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/scipy/optimize/_highs/_highs_wrapper.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
59
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/scipy/sparse/_sparsetools.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
60
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/scipy/spatial/_qhull.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
61
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/scipy/special/_ufuncs.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
62
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/scipy/special/cython_special.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
63
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/scipy/stats/_unuran/unuran_wrapper.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
64
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/scipy.libs/libgfortran-040039e1.so.5.0.0 filter=lfs diff=lfs merge=lfs -text
65
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/scipy.libs/libopenblasp-r0-41284840.3.18.so filter=lfs diff=lfs merge=lfs -text
66
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/sklearn/_loss/_loss.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
67
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorboard_data_server/bin/server filter=lfs diff=lfs merge=lfs -text
68
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/compiler/tf2tensorrt/_pywrap_py_utils.so filter=lfs diff=lfs merge=lfs -text
69
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/compiler/tf2xla/ops/_xla_ops.so filter=lfs diff=lfs merge=lfs -text
70
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/core/kernels/libtfkernel_sobol_op.so filter=lfs diff=lfs merge=lfs -text
71
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/libtensorflow_cc.so.2 filter=lfs diff=lfs merge=lfs -text
72
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/libtensorflow_framework.so.2 filter=lfs diff=lfs merge=lfs -text
73
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/lite/experimental/microfrontend/python/ops/_audio_microfrontend_op.so filter=lfs diff=lfs merge=lfs -text
74
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/lite/python/analyzer_wrapper/_pywrap_analyzer_wrapper.so filter=lfs diff=lfs merge=lfs -text
75
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/lite/python/interpreter_wrapper/_pywrap_tensorflow_interpreter_wrapper.so filter=lfs diff=lfs merge=lfs -text
76
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/lite/python/metrics/_pywrap_tensorflow_lite_metrics_wrapper.so filter=lfs diff=lfs merge=lfs -text
77
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/lite/python/optimize/_pywrap_tensorflow_lite_calibration_wrapper.so filter=lfs diff=lfs merge=lfs -text
78
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/_pywrap_mlir.so filter=lfs diff=lfs merge=lfs -text
79
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/_pywrap_parallel_device.so filter=lfs diff=lfs merge=lfs -text
80
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/_pywrap_py_exception_registry.so filter=lfs diff=lfs merge=lfs -text
81
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/_pywrap_quantize_training.so filter=lfs diff=lfs merge=lfs -text
82
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/_pywrap_tensorflow_internal.so filter=lfs diff=lfs merge=lfs -text
83
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/_pywrap_tfcompile.so filter=lfs diff=lfs merge=lfs -text
84
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/_pywrap_tfe.so filter=lfs diff=lfs merge=lfs -text
85
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/autograph/impl/testing/pybind_for_testing.so filter=lfs diff=lfs merge=lfs -text
86
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/client/_pywrap_debug_events_writer.so filter=lfs diff=lfs merge=lfs -text
87
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/client/_pywrap_device_lib.so filter=lfs diff=lfs merge=lfs -text
88
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/client/_pywrap_events_writer.so filter=lfs diff=lfs merge=lfs -text
89
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/client/_pywrap_tf_session.so filter=lfs diff=lfs merge=lfs -text
90
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/data/experimental/service/_pywrap_server_lib.so filter=lfs diff=lfs merge=lfs -text
91
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/data/experimental/service/_pywrap_utils.so filter=lfs diff=lfs merge=lfs -text
92
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/flags_pybind.so filter=lfs diff=lfs merge=lfs -text
93
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/framework/_errors_test_helper.so filter=lfs diff=lfs merge=lfs -text
94
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/framework/_op_def_library_pybind.so filter=lfs diff=lfs merge=lfs -text
95
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/framework/_op_def_registry.so filter=lfs diff=lfs merge=lfs -text
96
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/framework/_op_def_util.so filter=lfs diff=lfs merge=lfs -text
97
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/framework/_proto_comparators.so filter=lfs diff=lfs merge=lfs -text
98
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/framework/_pywrap_python_api_dispatcher.so filter=lfs diff=lfs merge=lfs -text
99
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/framework/_pywrap_python_api_info.so filter=lfs diff=lfs merge=lfs -text
100
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/framework/_pywrap_python_api_parameter_converter.so filter=lfs diff=lfs merge=lfs -text
101
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/framework/_test_metrics_util.so filter=lfs diff=lfs merge=lfs -text
102
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/grappler/_pywrap_tf_cluster.so filter=lfs diff=lfs merge=lfs -text
103
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/grappler/_pywrap_tf_item.so filter=lfs diff=lfs merge=lfs -text
104
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/grappler/_pywrap_tf_optimizer.so filter=lfs diff=lfs merge=lfs -text
105
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/lib/io/_pywrap_file_io.so filter=lfs diff=lfs merge=lfs -text
106
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/lib/io/_pywrap_record_io.so filter=lfs diff=lfs merge=lfs -text
107
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/platform/_pywrap_tf2.so filter=lfs diff=lfs merge=lfs -text
108
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/profiler/internal/_pywrap_profiler.so filter=lfs diff=lfs merge=lfs -text
109
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/saved_model/pywrap_saved_model.so filter=lfs diff=lfs merge=lfs -text
110
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/util/_pywrap_checkpoint_reader.so filter=lfs diff=lfs merge=lfs -text
111
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/util/_pywrap_kernel_registry.so filter=lfs diff=lfs merge=lfs -text
112
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/util/_pywrap_stat_summarizer.so filter=lfs diff=lfs merge=lfs -text
113
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/util/_pywrap_tfprof.so filter=lfs diff=lfs merge=lfs -text
114
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/util/_pywrap_transform_graph.so filter=lfs diff=lfs merge=lfs -text
115
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/util/_pywrap_util_port.so filter=lfs diff=lfs merge=lfs -text
116
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/util/_pywrap_utils.so filter=lfs diff=lfs merge=lfs -text
117
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow/python/util/_tf_stack.so filter=lfs diff=lfs merge=lfs -text
118
+ docker/deezer-spleeter-env/lib/python3.10/site-packages/tensorflow_io_gcs_filesystem/core/python/ops/libtensorflow_io_gcs_filesystem.so filter=lfs diff=lfs merge=lfs -text
119
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/_soundfile_data/libsndfile_x86_64.so filter=lfs diff=lfs merge=lfs -text
120
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/clang/native/libclang.so filter=lfs diff=lfs merge=lfs -text
121
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/google/protobuf/pyext/_message.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
122
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/grpc/_cython/cygrpc.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
123
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/h5py/h5t.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
124
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/h5py.libs/libhdf5-45610c27.so.200.2.0 filter=lfs diff=lfs merge=lfs -text
125
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/llvmlite/binding/libllvmlite.so filter=lfs diff=lfs merge=lfs -text
126
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/numpy/core/_multiarray_umath.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
127
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/numpy/core/_simd.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
128
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/numpy/random/_generator.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
129
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/numpy.libs/libgfortran-040039e1.so.5.0.0 filter=lfs diff=lfs merge=lfs -text
130
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/numpy.libs/libopenblas64_p-r0-2f7c42d4.3.18.so filter=lfs diff=lfs merge=lfs -text
131
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/pandas/_libs/algos.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
132
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/pandas/_libs/groupby.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
133
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/pandas/_libs/hashtable.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
134
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/pandas/_libs/interval.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
135
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/pandas/_libs/join.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
136
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/scipy/fft/_pocketfft/pypocketfft.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
137
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/scipy/linalg/_flapack.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
138
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/scipy/misc/face.dat filter=lfs diff=lfs merge=lfs -text
139
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/scipy/optimize/_highs/_highs_wrapper.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
140
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/scipy/sparse/_sparsetools.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
141
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/scipy/spatial/_qhull.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
142
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/scipy/special/_ufuncs.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
143
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/scipy/special/cython_special.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
144
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/scipy/stats/_unuran/unuran_wrapper.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
145
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/scipy.libs/libgfortran-040039e1.so.5.0.0 filter=lfs diff=lfs merge=lfs -text
146
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/scipy.libs/libopenblasp-r0-41284840.3.18.so filter=lfs diff=lfs merge=lfs -text
147
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/sklearn/_loss/_loss.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text
148
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorboard_data_server/bin/server filter=lfs diff=lfs merge=lfs -text
149
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/compiler/tf2tensorrt/_pywrap_py_utils.so filter=lfs diff=lfs merge=lfs -text
150
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/compiler/tf2xla/ops/_xla_ops.so filter=lfs diff=lfs merge=lfs -text
151
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/core/kernels/libtfkernel_sobol_op.so filter=lfs diff=lfs merge=lfs -text
152
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/libtensorflow_cc.so.2 filter=lfs diff=lfs merge=lfs -text
153
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/libtensorflow_framework.so.2 filter=lfs diff=lfs merge=lfs -text
154
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/lite/experimental/microfrontend/python/ops/_audio_microfrontend_op.so filter=lfs diff=lfs merge=lfs -text
155
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/lite/python/analyzer_wrapper/_pywrap_analyzer_wrapper.so filter=lfs diff=lfs merge=lfs -text
156
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/lite/python/interpreter_wrapper/_pywrap_tensorflow_interpreter_wrapper.so filter=lfs diff=lfs merge=lfs -text
157
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/lite/python/metrics/_pywrap_tensorflow_lite_metrics_wrapper.so filter=lfs diff=lfs merge=lfs -text
158
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/lite/python/optimize/_pywrap_tensorflow_lite_calibration_wrapper.so filter=lfs diff=lfs merge=lfs -text
159
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/_pywrap_mlir.so filter=lfs diff=lfs merge=lfs -text
160
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/_pywrap_parallel_device.so filter=lfs diff=lfs merge=lfs -text
161
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/_pywrap_py_exception_registry.so filter=lfs diff=lfs merge=lfs -text
162
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/_pywrap_quantize_training.so filter=lfs diff=lfs merge=lfs -text
163
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/_pywrap_tensorflow_internal.so filter=lfs diff=lfs merge=lfs -text
164
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/_pywrap_tfcompile.so filter=lfs diff=lfs merge=lfs -text
165
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/_pywrap_tfe.so filter=lfs diff=lfs merge=lfs -text
166
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/autograph/impl/testing/pybind_for_testing.so filter=lfs diff=lfs merge=lfs -text
167
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/client/_pywrap_debug_events_writer.so filter=lfs diff=lfs merge=lfs -text
168
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/client/_pywrap_device_lib.so filter=lfs diff=lfs merge=lfs -text
169
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/client/_pywrap_events_writer.so filter=lfs diff=lfs merge=lfs -text
170
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/client/_pywrap_tf_session.so filter=lfs diff=lfs merge=lfs -text
171
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/data/experimental/service/_pywrap_server_lib.so filter=lfs diff=lfs merge=lfs -text
172
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/data/experimental/service/_pywrap_utils.so filter=lfs diff=lfs merge=lfs -text
173
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/flags_pybind.so filter=lfs diff=lfs merge=lfs -text
174
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/framework/_errors_test_helper.so filter=lfs diff=lfs merge=lfs -text
175
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/framework/_op_def_library_pybind.so filter=lfs diff=lfs merge=lfs -text
176
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/framework/_op_def_registry.so filter=lfs diff=lfs merge=lfs -text
177
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/framework/_op_def_util.so filter=lfs diff=lfs merge=lfs -text
178
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/framework/_proto_comparators.so filter=lfs diff=lfs merge=lfs -text
179
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/framework/_pywrap_python_api_dispatcher.so filter=lfs diff=lfs merge=lfs -text
180
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/framework/_pywrap_python_api_info.so filter=lfs diff=lfs merge=lfs -text
181
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/framework/_pywrap_python_api_parameter_converter.so filter=lfs diff=lfs merge=lfs -text
182
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/framework/_test_metrics_util.so filter=lfs diff=lfs merge=lfs -text
183
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/grappler/_pywrap_tf_cluster.so filter=lfs diff=lfs merge=lfs -text
184
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/grappler/_pywrap_tf_item.so filter=lfs diff=lfs merge=lfs -text
185
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/grappler/_pywrap_tf_optimizer.so filter=lfs diff=lfs merge=lfs -text
186
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/lib/io/_pywrap_file_io.so filter=lfs diff=lfs merge=lfs -text
187
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/lib/io/_pywrap_record_io.so filter=lfs diff=lfs merge=lfs -text
188
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/platform/_pywrap_tf2.so filter=lfs diff=lfs merge=lfs -text
189
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/profiler/internal/_pywrap_profiler.so filter=lfs diff=lfs merge=lfs -text
190
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/saved_model/pywrap_saved_model.so filter=lfs diff=lfs merge=lfs -text
191
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/util/_pywrap_checkpoint_reader.so filter=lfs diff=lfs merge=lfs -text
192
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/util/_pywrap_kernel_registry.so filter=lfs diff=lfs merge=lfs -text
193
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/util/_pywrap_stat_summarizer.so filter=lfs diff=lfs merge=lfs -text
194
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/util/_pywrap_tfprof.so filter=lfs diff=lfs merge=lfs -text
195
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/util/_pywrap_transform_graph.so filter=lfs diff=lfs merge=lfs -text
196
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/util/_pywrap_util_port.so filter=lfs diff=lfs merge=lfs -text
197
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/util/_pywrap_utils.so filter=lfs diff=lfs merge=lfs -text
198
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow/python/util/_tf_stack.so filter=lfs diff=lfs merge=lfs -text
199
+ docker/deezer-spleeter-env/lib64/python3.10/site-packages/tensorflow_io_gcs_filesystem/core/python/ops/libtensorflow_io_gcs_filesystem.so filter=lfs diff=lfs merge=lfs -text
200
+ docker/pretrained_models/2stems/model.data-00000-of-00001 filter=lfs diff=lfs merge=lfs -text
201
+ docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/kanye/added_IVF875_Flat_nprobe_7.index filter=lfs diff=lfs merge=lfs -text
202
+ docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/test-voice.mp3 filter=lfs diff=lfs merge=lfs -text
203
+ docker-no-splitter/pretrained_models/2stems/model.data-00000-of-00001 filter=lfs diff=lfs merge=lfs -text
204
+ music/Game[[:space:]]feat.Tupac[[:space:]]and[[:space:]]Easy[[:space:]]E[[:space:]]-[[:space:]]T/Game[[:space:]]feat.Tupac[[:space:]]and[[:space:]]Easy[[:space:]]E[[:space:]]-[[:space:]]T.mp3 filter=lfs diff=lfs merge=lfs -text
205
+ music/Game[[:space:]]feat.Tupac[[:space:]]and[[:space:]]Easy[[:space:]]E[[:space:]]-[[:space:]]T/vocal+instrumental/Game[[:space:]]feat.Tupac[[:space:]]and[[:space:]]Easy[[:space:]]E[[:space:]]-[[:space:]]T[[:space:]](Filtered[[:space:]]Instrumental).mp3 filter=lfs diff=lfs merge=lfs -text
206
+ music/Game[[:space:]]feat.Tupac[[:space:]]and[[:space:]]Easy[[:space:]]E[[:space:]]-[[:space:]]T/vocal+instrumental/Game[[:space:]]feat.Tupac[[:space:]]and[[:space:]]Easy[[:space:]]E[[:space:]]-[[:space:]]T[[:space:]](Filtered[[:space:]]Vocals).mp3 filter=lfs diff=lfs merge=lfs -text
207
+ music/cupid/cupid_instrumental_trimmed.mp3 filter=lfs diff=lfs merge=lfs -text
208
+ music/cupid/cupid_voice_trimmed.mp3 filter=lfs diff=lfs merge=lfs -text
209
+ music/cupid/instrumental.mp3 filter=lfs diff=lfs merge=lfs -text
210
+ music/cupid/voice.mp3 filter=lfs diff=lfs merge=lfs -text
211
+ music/vanessa/combined.mp3 filter=lfs diff=lfs merge=lfs -text
212
+ music/vanessa/fc0ced7e-21f0-4548-90d0-f4d210f8dc88.wav filter=lfs diff=lfs merge=lfs -text
213
+ music/vanessa/instrumental.mp3 filter=lfs diff=lfs merge=lfs -text
214
+ music/vanessa/instrumental_trimmed.mp3 filter=lfs diff=lfs merge=lfs -text
215
+ music/vanessa/voice.mp3 filter=lfs diff=lfs merge=lfs -text
216
+ music/vanessa/voice_trimmed.mp3 filter=lfs diff=lfs merge=lfs -text
217
+ music/vanessa-cropped-song/vanessa-cropped-song.mp3 filter=lfs diff=lfs merge=lfs -text
218
+ music/vanessa-cropped-song/vanessa-cropped-song.wav filter=lfs diff=lfs merge=lfs -text
.gitignore CHANGED
@@ -1,11 +0,0 @@
1
- models/
2
- logs/
3
- pretrained/
4
- weights/
5
- music/
6
- *.ipynb
7
- HP2-人声vocals+非人声instrumentals.pth
8
- test-voice.mp3
9
- hubert_base.pt
10
- docker/deezer-spleeter-env
11
- pretrained_models/
 
 
 
 
 
 
 
 
 
 
 
 
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/HP2-人声vocals+非人声instrumentals.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39796caa5db18d7f9382d8ac997ac967bfd85f7761014bb807d2543cc844ef05
3
+ size 63454827
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/Retrieval_based_Voice_Conversion_WebUI.ipynb ADDED
@@ -0,0 +1,381 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "nbformat": 4,
3
+ "nbformat_minor": 0,
4
+ "metadata": {
5
+ "colab": {
6
+ "private_outputs": true,
7
+ "provenance": []
8
+ },
9
+ "kernelspec": {
10
+ "name": "python3",
11
+ "display_name": "Python 3"
12
+ },
13
+ "language_info": {
14
+ "name": "python"
15
+ },
16
+ "accelerator": "GPU",
17
+ "gpuClass": "standard"
18
+ },
19
+ "cells": [
20
+ {
21
+ "cell_type": "markdown",
22
+ "source": [
23
+ "[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/liujing04/Retrieval-based-Voice-Conversion-WebUI/blob/main/Retrieval_based_Voice_Conversion_WebUI.ipynb)"
24
+ ],
25
+ "metadata": {
26
+ "id": "ZFFCx5J80SGa"
27
+ }
28
+ },
29
+ {
30
+ "cell_type": "code",
31
+ "execution_count": null,
32
+ "metadata": {
33
+ "id": "GmFP6bN9dvOq"
34
+ },
35
+ "outputs": [],
36
+ "source": [
37
+ "#@title 查看显卡\n",
38
+ "!nvidia-smi"
39
+ ]
40
+ },
41
+ {
42
+ "cell_type": "code",
43
+ "source": [
44
+ "#@title 安装依赖\n",
45
+ "!apt-get -y install build-essential python3-dev ffmpeg\n",
46
+ "!pip3 install --upgrade setuptools wheel\n",
47
+ "!pip3 install --upgrade pip\n",
48
+ "!pip3 install faiss-gpu fairseq gradio ffmpeg ffmpeg-python praat-parselmouth pyworld numpy==1.23.5 numba==0.56.4 librosa==0.9.2"
49
+ ],
50
+ "metadata": {
51
+ "id": "wjddIFr1oS3W"
52
+ },
53
+ "execution_count": null,
54
+ "outputs": []
55
+ },
56
+ {
57
+ "cell_type": "code",
58
+ "source": [
59
+ "#@title 克隆仓库\n",
60
+ "\n",
61
+ "!git clone --depth=1 -b stable https://github.com/fumiama/Retrieval-based-Voice-Conversion-WebUI\n",
62
+ "%cd /content/Retrieval-based-Voice-Conversion-WebUI\n",
63
+ "!mkdir -p pretrained uvr5_weights"
64
+ ],
65
+ "metadata": {
66
+ "id": "ge_97mfpgqTm"
67
+ },
68
+ "execution_count": null,
69
+ "outputs": []
70
+ },
71
+ {
72
+ "cell_type": "code",
73
+ "source": [
74
+ "#@title 更新仓库(一般无需执行)\n",
75
+ "!git pull"
76
+ ],
77
+ "metadata": {
78
+ "id": "BLDEZADkvlw1"
79
+ },
80
+ "execution_count": null,
81
+ "outputs": []
82
+ },
83
+ {
84
+ "cell_type": "code",
85
+ "source": [
86
+ "#@title 安装aria2\n",
87
+ "!apt -y install -qq aria2"
88
+ ],
89
+ "metadata": {
90
+ "id": "pqE0PrnuRqI2"
91
+ },
92
+ "execution_count": null,
93
+ "outputs": []
94
+ },
95
+ {
96
+ "cell_type": "code",
97
+ "source": [
98
+ "#@title 下载底模\n",
99
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/D32k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o D32k.pth\n",
100
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/D40k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o D40k.pth\n",
101
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/D48k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o D48k.pth\n",
102
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/G32k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o G32k.pth\n",
103
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/G40k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o G40k.pth\n",
104
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/G48k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o G48k.pth\n",
105
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/f0D32k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o f0D32k.pth\n",
106
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/f0D40k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o f0D40k.pth\n",
107
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/f0D48k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o f0D48k.pth\n",
108
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/f0G32k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o f0G32k.pth\n",
109
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/f0G40k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o f0G40k.pth\n",
110
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/f0G48k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o f0G48k.pth"
111
+ ],
112
+ "metadata": {
113
+ "id": "UG3XpUwEomUz"
114
+ },
115
+ "execution_count": null,
116
+ "outputs": []
117
+ },
118
+ {
119
+ "cell_type": "code",
120
+ "source": [
121
+ "#@title 下载人声分离模型\n",
122
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/uvr5_weights/HP2-人声vocals+非人声instrumentals.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/uvr5_weights -o HP2-人声vocals+非人声instrumentals.pth\n",
123
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/uvr5_weights/HP5-主旋律人声vocals+其他instrumentals.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/uvr5_weights -o HP5-主旋律人声vocals+其他instrumentals.pth"
124
+ ],
125
+ "metadata": {
126
+ "id": "HugjmZqZRuiF"
127
+ },
128
+ "execution_count": null,
129
+ "outputs": []
130
+ },
131
+ {
132
+ "cell_type": "code",
133
+ "source": [
134
+ "#@title 下载hubert_base\n",
135
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/hubert_base.pt -d /content/Retrieval-based-Voice-Conversion-WebUI -o hubert_base.pt"
136
+ ],
137
+ "metadata": {
138
+ "id": "2RCaT9FTR0ej"
139
+ },
140
+ "execution_count": null,
141
+ "outputs": []
142
+ },
143
+ {
144
+ "cell_type": "code",
145
+ "source": [
146
+ "#@title 挂载谷歌云盘\n",
147
+ "\n",
148
+ "from google.colab import drive\n",
149
+ "drive.mount('/content/drive')"
150
+ ],
151
+ "metadata": {
152
+ "id": "jwu07JgqoFON"
153
+ },
154
+ "execution_count": null,
155
+ "outputs": []
156
+ },
157
+ {
158
+ "cell_type": "code",
159
+ "source": [
160
+ "#@title 从谷歌云盘加载打包好的数据集到/content/dataset\n",
161
+ "\n",
162
+ "#@markdown 数据集位置\n",
163
+ "DATASET = \"/content/drive/MyDrive/dataset/lulu20230327_32k.zip\" #@param {type:\"string\"}\n",
164
+ "\n",
165
+ "!mkdir -p /content/dataset\n",
166
+ "!unzip -d /content/dataset -B {DATASET}"
167
+ ],
168
+ "metadata": {
169
+ "id": "Mwk7Q0Loqzjx"
170
+ },
171
+ "execution_count": null,
172
+ "outputs": []
173
+ },
174
+ {
175
+ "cell_type": "code",
176
+ "source": [
177
+ "#@title 重命名数据集中的重名文件\n",
178
+ "!ls -a /content/dataset/\n",
179
+ "!rename 's/(\\w+)\\.(\\w+)~(\\d*)/$1_$3.$2/' /content/dataset/*.*~*"
180
+ ],
181
+ "metadata": {
182
+ "id": "PDlFxWHWEynD"
183
+ },
184
+ "execution_count": null,
185
+ "outputs": []
186
+ },
187
+ {
188
+ "cell_type": "code",
189
+ "source": [
190
+ "#@title 启动web\n",
191
+ "%cd /content/Retrieval-based-Voice-Conversion-WebUI\n",
192
+ "# %load_ext tensorboard\n",
193
+ "# %tensorboard --logdir /content/Retrieval-based-Voice-Conversion-WebUI/logs\n",
194
+ "!python3 infer-web.py --colab --pycmd python3"
195
+ ],
196
+ "metadata": {
197
+ "id": "7vh6vphDwO0b"
198
+ },
199
+ "execution_count": null,
200
+ "outputs": []
201
+ },
202
+ {
203
+ "cell_type": "code",
204
+ "source": [
205
+ "#@title 手动将训练后的模型文件备份到谷歌云盘\n",
206
+ "#@markdown 需要自己查看logs文件夹下模型的文件名,手动修改下方命令末尾的文件名\n",
207
+ "\n",
208
+ "#@markdown 模型名\n",
209
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
210
+ "#@markdown 模型epoch\n",
211
+ "MODELEPOCH = 9600 #@param {type:\"integer\"}\n",
212
+ "\n",
213
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/G_{MODELEPOCH}.pth /content/drive/MyDrive/{MODELNAME}_D_{MODELEPOCH}.pth\n",
214
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/D_{MODELEPOCH}.pth /content/drive/MyDrive/{MODELNAME}_G_{MODELEPOCH}.pth\n",
215
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/added_*.index /content/drive/MyDrive/\n",
216
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/total_*.npy /content/drive/MyDrive/\n",
217
+ "\n",
218
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/weights/{MODELNAME}.pth /content/drive/MyDrive/{MODELNAME}{MODELEPOCH}.pth"
219
+ ],
220
+ "metadata": {
221
+ "id": "FgJuNeAwx5Y_"
222
+ },
223
+ "execution_count": null,
224
+ "outputs": []
225
+ },
226
+ {
227
+ "cell_type": "code",
228
+ "source": [
229
+ "#@title 从谷歌云盘恢复pth\n",
230
+ "#@markdown 需要自己查看logs文件夹下模型的文件名,手动修改下方命令末尾的文件名\n",
231
+ "\n",
232
+ "#@markdown 模型名\n",
233
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
234
+ "#@markdown 模型epoch\n",
235
+ "MODELEPOCH = 7500 #@param {type:\"integer\"}\n",
236
+ "\n",
237
+ "!mkdir -p /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}\n",
238
+ "\n",
239
+ "!cp /content/drive/MyDrive/{MODELNAME}_D_{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/G_{MODELEPOCH}.pth\n",
240
+ "!cp /content/drive/MyDrive/{MODELNAME}_G_{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/D_{MODELEPOCH}.pth\n",
241
+ "!cp /content/drive/MyDrive/*.index /content/\n",
242
+ "!cp /content/drive/MyDrive/*.npy /content/\n",
243
+ "!cp /content/drive/MyDrive/{MODELNAME}{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/weights/{MODELNAME}.pth"
244
+ ],
245
+ "metadata": {
246
+ "id": "OVQoLQJXS7WX"
247
+ },
248
+ "execution_count": null,
249
+ "outputs": []
250
+ },
251
+ {
252
+ "cell_type": "code",
253
+ "source": [
254
+ "#@title 手动预处理(不推荐)\n",
255
+ "#@markdown 模型名\n",
256
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
257
+ "#@markdown 采样率\n",
258
+ "BITRATE = 48000 #@param {type:\"integer\"}\n",
259
+ "#@markdown 使用的进程数\n",
260
+ "THREADCOUNT = 8 #@param {type:\"integer\"}\n",
261
+ "\n",
262
+ "!python3 trainset_preprocess_pipeline_print.py /content/dataset {BITRATE} {THREADCOUNT} logs/{MODELNAME} True\n"
263
+ ],
264
+ "metadata": {
265
+ "id": "ZKAyuKb9J6dz"
266
+ },
267
+ "execution_count": null,
268
+ "outputs": []
269
+ },
270
+ {
271
+ "cell_type": "code",
272
+ "source": [
273
+ "#@title 手动提取特征(不推荐)\n",
274
+ "#@markdown 模型名\n",
275
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
276
+ "#@markdown 使用的进程数\n",
277
+ "THREADCOUNT = 8 #@param {type:\"integer\"}\n",
278
+ "#@markdown 音高提取算法\n",
279
+ "ALGO = \"harvest\" #@param {type:\"string\"}\n",
280
+ "\n",
281
+ "!python3 extract_f0_print.py logs/{MODELNAME} {THREADCOUNT} {ALGO}\n",
282
+ "\n",
283
+ "!python3 extract_feature_print.py cpu 1 0 0 logs/{MODELNAME}\n"
284
+ ],
285
+ "metadata": {
286
+ "id": "CrxJqzAUKmPJ"
287
+ },
288
+ "execution_count": null,
289
+ "outputs": []
290
+ },
291
+ {
292
+ "cell_type": "code",
293
+ "source": [
294
+ "#@title 手动训练(不推荐)\n",
295
+ "#@markdown 模型名\n",
296
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
297
+ "#@markdown 使用的GPU\n",
298
+ "USEGPU = \"0\" #@param {type:\"string\"}\n",
299
+ "#@markdown 批大小\n",
300
+ "BATCHSIZE = 32 #@param {type:\"integer\"}\n",
301
+ "#@markdown 停止的epoch\n",
302
+ "MODELEPOCH = 3200 #@param {type:\"integer\"}\n",
303
+ "#@markdown 保存epoch间隔\n",
304
+ "EPOCHSAVE = 100 #@param {type:\"integer\"}\n",
305
+ "#@markdown 采样率\n",
306
+ "MODELSAMPLE = \"48k\" #@param {type:\"string\"}\n",
307
+ "#@markdown 是否缓存训练集\n",
308
+ "CACHEDATA = 1 #@param {type:\"integer\"}\n",
309
+ "#@markdown 是否仅保存最新的ckpt文件\n",
310
+ "ONLYLATEST = 0 #@param {type:\"integer\"}\n",
311
+ "\n",
312
+ "!python3 train_nsf_sim_cache_sid_load_pretrain.py -e lulu -sr {MODELSAMPLE} -f0 1 -bs {BATCHSIZE} -g {USEGPU} -te {MODELEPOCH} -se {EPOCHSAVE} -pg pretrained/f0G{MODELSAMPLE}.pth -pd pretrained/f0D{MODELSAMPLE}.pth -l {ONLYLATEST} -c {CACHEDATA}\n"
313
+ ],
314
+ "metadata": {
315
+ "id": "IMLPLKOaKj58"
316
+ },
317
+ "execution_count": null,
318
+ "outputs": []
319
+ },
320
+ {
321
+ "cell_type": "code",
322
+ "source": [
323
+ "#@title 删除其它pth,只留选中的(慎点,仔细看代码)\n",
324
+ "#@markdown 模型名\n",
325
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
326
+ "#@markdown 选中模型epoch\n",
327
+ "MODELEPOCH = 9600 #@param {type:\"integer\"}\n",
328
+ "\n",
329
+ "!echo \"备份选中的模型。。。\"\n",
330
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/G_{MODELEPOCH}.pth /content/{MODELNAME}_D_{MODELEPOCH}.pth\n",
331
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/D_{MODELEPOCH}.pth /content/{MODELNAME}_G_{MODELEPOCH}.pth\n",
332
+ "\n",
333
+ "!echo \"正在删除。。。\"\n",
334
+ "!ls /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}\n",
335
+ "!rm /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/*.pth\n",
336
+ "\n",
337
+ "!echo \"恢复选中的模型。。。\"\n",
338
+ "!mv /content/{MODELNAME}_D_{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/G_{MODELEPOCH}.pth \n",
339
+ "!mv /content/{MODELNAME}_G_{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/D_{MODELEPOCH}.pth\n",
340
+ "\n",
341
+ "!echo \"删除完成\"\n",
342
+ "!ls /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}"
343
+ ],
344
+ "metadata": {
345
+ "id": "haYA81hySuDl"
346
+ },
347
+ "execution_count": null,
348
+ "outputs": []
349
+ },
350
+ {
351
+ "cell_type": "code",
352
+ "source": [
353
+ "#@title 清除项目���所有文件,只留选中的模型(慎点,仔细看代码)\n",
354
+ "#@markdown 模型名\n",
355
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
356
+ "#@markdown 选中模型epoch\n",
357
+ "MODELEPOCH = 9600 #@param {type:\"integer\"}\n",
358
+ "\n",
359
+ "!echo \"备份选中的模型。。。\"\n",
360
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/G_{MODELEPOCH}.pth /content/{MODELNAME}_D_{MODELEPOCH}.pth\n",
361
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/D_{MODELEPOCH}.pth /content/{MODELNAME}_G_{MODELEPOCH}.pth\n",
362
+ "\n",
363
+ "!echo \"正在删除。。。\"\n",
364
+ "!ls /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}\n",
365
+ "!rm -rf /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/*\n",
366
+ "\n",
367
+ "!echo \"恢复选中的模型。。。\"\n",
368
+ "!mv /content/{MODELNAME}_D_{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/G_{MODELEPOCH}.pth \n",
369
+ "!mv /content/{MODELNAME}_G_{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/D_{MODELEPOCH}.pth\n",
370
+ "\n",
371
+ "!echo \"删除完成\"\n",
372
+ "!ls /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}"
373
+ ],
374
+ "metadata": {
375
+ "id": "QhSiPTVPoIRh"
376
+ },
377
+ "execution_count": null,
378
+ "outputs": []
379
+ }
380
+ ]
381
+ }
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/hubert_base.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f54b40fd2802423a5643779c4861af1e9ee9c1564dc9d32f54f20b5ffba7db96
3
+ size 189507909
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/.DS_Store ADDED
Binary file (6.15 kB). View file
 
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/kanye/D_2333333.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74584537cac9d53532cba4310c6f31b91a085c3991c3b43ce04ad4bc504b11b2
3
+ size 659775237
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/kanye/G_2333333.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:825f08efbfbd24c50576822cbfbf036d31e3544546666d1562b7b49af4355308
3
+ size 437003421
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/kanye/added_IVF875_Flat_nprobe_7.index ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fd5f21995b96231bbb1534814df3d9593975439fc580b664114dc6fe0ac2d5a
3
+ size 36141811
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/kanye/total_fea.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc59e38154b3a65b5cfabe5da337d40a506e150b62ac1f6380b211e7372b3bb0
3
+ size 34965632
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/.DS_Store ADDED
Binary file (8.2 kB). View file
 
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/0_gt_wavs/mute32k.wav ADDED
Binary file (192 kB). View file
 
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/0_gt_wavs/mute40k.wav ADDED
Binary file (240 kB). View file
 
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/0_gt_wavs/mute48k.wav ADDED
Binary file (288 kB). View file
 
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/1_16k_wavs/mute.wav ADDED
Binary file (96.1 kB). View file
 
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/2a_f0/mute.wav.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b9acf9ab7facdb032e1d687fe35182670b0b94566c4b209ae48c239d19956a6
3
+ size 1332
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/2b-f0nsf/mute.wav.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30792849c8e72d67e6691754077f2888b101cb741e9c7f193c91dd9692870c87
3
+ size 2536
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/logs/mute/3_feature256/mute.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64d5abbac078e19a3f649c0d78a02cb33a71407ded3ddf2db78e6b803d0c0126
3
+ size 152704
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/.DS_Store ADDED
Binary file (6.15 kB). View file
 
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/.gitignore ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ *
2
+ !.gitignore
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/D32k.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ab20645829460fdad0d3c44254f1ab53c32cae50c22a66c926ae5aa30abda6f
3
+ size 109978943
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/D40k.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:547f66dbbcd9023b9051ed244d12ab043ba8a4e854b154cc28761ac7c002909b
3
+ size 109978943
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/D48k.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cc013fa60ed9c3f902f5bd99f48c7e3b9352d763d4d3cd6bc241c37b0bfd9ad
3
+ size 109978943
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/G32k.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90e0458e39efd50beef48ab398c9f88cec2b405d69565e2ad958a25882936aa1
3
+ size 72653893
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/G40k.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1e7e413d5b1009777c39cd232ac4d91a81382161d19350fe15bf3f137b01425
3
+ size 72763063
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/G48k.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d4a19433aa20617d0712dce2d379620b2f9b80e21b58d6081ba442f203e26d0
3
+ size 72850501
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/f0D32k.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:294db3087236e2c75260d6179056791c9231245daf5d0485545d9e54c4057c77
3
+ size 109978943
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/f0D40k.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d4f5a441594b470d67579958b2fd4c6b992852ded28ff9e72eda67abcebe423
3
+ size 109978943
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/f0G32k.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68c08e16e47d2796427b518c9b702edeae5c9d11f8d62608215f10c776d6170b
3
+ size 72795627
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/f0G40k.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a562c00b615367ed2bd97857346a26ce05b5c231629264f5ca2e19ce5f8d59ea
3
+ size 72909665
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/pretrained/f0G48k.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f36a4b1e1c3009ee72e5e6c572e0418d8585094e94b00ca9bd5c2b6c8a6c404
3
+ size 73008619
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/test-voice.mp3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:400171cb980e181d7962a04b56c7251b2348004d22722e7a621c1a90900ee757
3
+ size 1362706
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/weights/.gitignore ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ *
2
+ !.gitignore
docker-no-splitter/Retrieval-based-Voice-Conversion-WebUI/weights/kanye.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7229ed4c322d7c38758b3ea788c4ab6ad7c674d444005bbbbeef5e469c3a20d
3
+ size 54996633
docker-no-splitter/models/.DS_Store ADDED
Binary file (8.2 kB). View file
 
docker-no-splitter/models/ariana-grande.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b60a9217ce171667039f3bab5056aa81105509c3d21b8520bf22520a70c6633b
3
+ size 55127639
docker-no-splitter/models/drake.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7158c3bd17a324b0911ebfbadc6fd20c47fa475b3a8635effb80174db152e03d
3
+ size 55026095
docker-no-splitter/models/iu.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fb08c8e2822bc508e1f3111b40be71ccc09bf298857b76e238705fa24872b32
3
+ size 54995128
docker-no-splitter/models/jungkook.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c846648c9447cd91690f1d17c3ac2f868f951ed2eb04af94ac98fc26580a88bd
3
+ size 55027130
docker-no-splitter/models/kanye.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7229ed4c322d7c38758b3ea788c4ab6ad7c674d444005bbbbeef5e469c3a20d
3
+ size 54996633
docker-no-splitter/models/rose-bp.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a488df159ea8ba96e31140bb0f040e8131dcede9362e25d4128293bdc0a1d3d
3
+ size 55026095
docker-no-splitter/pretrained_models/.DS_Store ADDED
Binary file (8.2 kB). View file
 
docker-no-splitter/pretrained_models/2stems/.probe ADDED
@@ -0,0 +1 @@
 
 
1
+ OK
docker-no-splitter/pretrained_models/2stems/checkpoint ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ model_checkpoint_path: "model"
2
+ all_model_checkpoint_paths: "model"
docker-no-splitter/pretrained_models/2stems/model.data-00000-of-00001 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7747f9fd2c782306dbec1504360fbb645a097a48446f23486c3ff9c89bc11788
3
+ size 78614080
docker-no-splitter/pretrained_models/2stems/model.index ADDED
Binary file (5.24 kB). View file
 
docker-no-splitter/pretrained_models/2stems/model.meta ADDED
Binary file (806 kB). View file
 
docker/Retrieval-based-Voice-Conversion-WebUI/HP2-人声vocals+非人声instrumentals.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39796caa5db18d7f9382d8ac997ac967bfd85f7761014bb807d2543cc844ef05
3
+ size 63454827
docker/Retrieval-based-Voice-Conversion-WebUI/Retrieval_based_Voice_Conversion_WebUI.ipynb ADDED
@@ -0,0 +1,381 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "nbformat": 4,
3
+ "nbformat_minor": 0,
4
+ "metadata": {
5
+ "colab": {
6
+ "private_outputs": true,
7
+ "provenance": []
8
+ },
9
+ "kernelspec": {
10
+ "name": "python3",
11
+ "display_name": "Python 3"
12
+ },
13
+ "language_info": {
14
+ "name": "python"
15
+ },
16
+ "accelerator": "GPU",
17
+ "gpuClass": "standard"
18
+ },
19
+ "cells": [
20
+ {
21
+ "cell_type": "markdown",
22
+ "source": [
23
+ "[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/liujing04/Retrieval-based-Voice-Conversion-WebUI/blob/main/Retrieval_based_Voice_Conversion_WebUI.ipynb)"
24
+ ],
25
+ "metadata": {
26
+ "id": "ZFFCx5J80SGa"
27
+ }
28
+ },
29
+ {
30
+ "cell_type": "code",
31
+ "execution_count": null,
32
+ "metadata": {
33
+ "id": "GmFP6bN9dvOq"
34
+ },
35
+ "outputs": [],
36
+ "source": [
37
+ "#@title 查看显卡\n",
38
+ "!nvidia-smi"
39
+ ]
40
+ },
41
+ {
42
+ "cell_type": "code",
43
+ "source": [
44
+ "#@title 安装依赖\n",
45
+ "!apt-get -y install build-essential python3-dev ffmpeg\n",
46
+ "!pip3 install --upgrade setuptools wheel\n",
47
+ "!pip3 install --upgrade pip\n",
48
+ "!pip3 install faiss-gpu fairseq gradio ffmpeg ffmpeg-python praat-parselmouth pyworld numpy==1.23.5 numba==0.56.4 librosa==0.9.2"
49
+ ],
50
+ "metadata": {
51
+ "id": "wjddIFr1oS3W"
52
+ },
53
+ "execution_count": null,
54
+ "outputs": []
55
+ },
56
+ {
57
+ "cell_type": "code",
58
+ "source": [
59
+ "#@title 克隆仓库\n",
60
+ "\n",
61
+ "!git clone --depth=1 -b stable https://github.com/fumiama/Retrieval-based-Voice-Conversion-WebUI\n",
62
+ "%cd /content/Retrieval-based-Voice-Conversion-WebUI\n",
63
+ "!mkdir -p pretrained uvr5_weights"
64
+ ],
65
+ "metadata": {
66
+ "id": "ge_97mfpgqTm"
67
+ },
68
+ "execution_count": null,
69
+ "outputs": []
70
+ },
71
+ {
72
+ "cell_type": "code",
73
+ "source": [
74
+ "#@title 更新仓库(一般无需执行)\n",
75
+ "!git pull"
76
+ ],
77
+ "metadata": {
78
+ "id": "BLDEZADkvlw1"
79
+ },
80
+ "execution_count": null,
81
+ "outputs": []
82
+ },
83
+ {
84
+ "cell_type": "code",
85
+ "source": [
86
+ "#@title 安装aria2\n",
87
+ "!apt -y install -qq aria2"
88
+ ],
89
+ "metadata": {
90
+ "id": "pqE0PrnuRqI2"
91
+ },
92
+ "execution_count": null,
93
+ "outputs": []
94
+ },
95
+ {
96
+ "cell_type": "code",
97
+ "source": [
98
+ "#@title 下载底模\n",
99
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/D32k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o D32k.pth\n",
100
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/D40k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o D40k.pth\n",
101
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/D48k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o D48k.pth\n",
102
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/G32k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o G32k.pth\n",
103
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/G40k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o G40k.pth\n",
104
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/G48k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o G48k.pth\n",
105
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/f0D32k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o f0D32k.pth\n",
106
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/f0D40k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o f0D40k.pth\n",
107
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/f0D48k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o f0D48k.pth\n",
108
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/f0G32k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o f0G32k.pth\n",
109
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/f0G40k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o f0G40k.pth\n",
110
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/pretrained/f0G48k.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/pretrained -o f0G48k.pth"
111
+ ],
112
+ "metadata": {
113
+ "id": "UG3XpUwEomUz"
114
+ },
115
+ "execution_count": null,
116
+ "outputs": []
117
+ },
118
+ {
119
+ "cell_type": "code",
120
+ "source": [
121
+ "#@title 下载人声分离模型\n",
122
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/uvr5_weights/HP2-人声vocals+非人声instrumentals.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/uvr5_weights -o HP2-人声vocals+非人声instrumentals.pth\n",
123
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/uvr5_weights/HP5-主旋律人声vocals+其他instrumentals.pth -d /content/Retrieval-based-Voice-Conversion-WebUI/uvr5_weights -o HP5-主旋律人声vocals+其他instrumentals.pth"
124
+ ],
125
+ "metadata": {
126
+ "id": "HugjmZqZRuiF"
127
+ },
128
+ "execution_count": null,
129
+ "outputs": []
130
+ },
131
+ {
132
+ "cell_type": "code",
133
+ "source": [
134
+ "#@title 下载hubert_base\n",
135
+ "!aria2c --console-log-level=error -c -x 16 -s 16 -k 1M https://huggingface.co/lj1995/VoiceConversionWebUI/resolve/main/hubert_base.pt -d /content/Retrieval-based-Voice-Conversion-WebUI -o hubert_base.pt"
136
+ ],
137
+ "metadata": {
138
+ "id": "2RCaT9FTR0ej"
139
+ },
140
+ "execution_count": null,
141
+ "outputs": []
142
+ },
143
+ {
144
+ "cell_type": "code",
145
+ "source": [
146
+ "#@title 挂载谷歌云盘\n",
147
+ "\n",
148
+ "from google.colab import drive\n",
149
+ "drive.mount('/content/drive')"
150
+ ],
151
+ "metadata": {
152
+ "id": "jwu07JgqoFON"
153
+ },
154
+ "execution_count": null,
155
+ "outputs": []
156
+ },
157
+ {
158
+ "cell_type": "code",
159
+ "source": [
160
+ "#@title 从谷歌云盘加载打包好的数据集到/content/dataset\n",
161
+ "\n",
162
+ "#@markdown 数据集位置\n",
163
+ "DATASET = \"/content/drive/MyDrive/dataset/lulu20230327_32k.zip\" #@param {type:\"string\"}\n",
164
+ "\n",
165
+ "!mkdir -p /content/dataset\n",
166
+ "!unzip -d /content/dataset -B {DATASET}"
167
+ ],
168
+ "metadata": {
169
+ "id": "Mwk7Q0Loqzjx"
170
+ },
171
+ "execution_count": null,
172
+ "outputs": []
173
+ },
174
+ {
175
+ "cell_type": "code",
176
+ "source": [
177
+ "#@title 重命名数据集中的重名文件\n",
178
+ "!ls -a /content/dataset/\n",
179
+ "!rename 's/(\\w+)\\.(\\w+)~(\\d*)/$1_$3.$2/' /content/dataset/*.*~*"
180
+ ],
181
+ "metadata": {
182
+ "id": "PDlFxWHWEynD"
183
+ },
184
+ "execution_count": null,
185
+ "outputs": []
186
+ },
187
+ {
188
+ "cell_type": "code",
189
+ "source": [
190
+ "#@title 启动web\n",
191
+ "%cd /content/Retrieval-based-Voice-Conversion-WebUI\n",
192
+ "# %load_ext tensorboard\n",
193
+ "# %tensorboard --logdir /content/Retrieval-based-Voice-Conversion-WebUI/logs\n",
194
+ "!python3 infer-web.py --colab --pycmd python3"
195
+ ],
196
+ "metadata": {
197
+ "id": "7vh6vphDwO0b"
198
+ },
199
+ "execution_count": null,
200
+ "outputs": []
201
+ },
202
+ {
203
+ "cell_type": "code",
204
+ "source": [
205
+ "#@title 手动将训练后的模型文件备份到谷歌云盘\n",
206
+ "#@markdown 需要自己查看logs文件夹下模型的文件名,手动修改下方命令末尾的文件名\n",
207
+ "\n",
208
+ "#@markdown 模型名\n",
209
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
210
+ "#@markdown 模型epoch\n",
211
+ "MODELEPOCH = 9600 #@param {type:\"integer\"}\n",
212
+ "\n",
213
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/G_{MODELEPOCH}.pth /content/drive/MyDrive/{MODELNAME}_D_{MODELEPOCH}.pth\n",
214
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/D_{MODELEPOCH}.pth /content/drive/MyDrive/{MODELNAME}_G_{MODELEPOCH}.pth\n",
215
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/added_*.index /content/drive/MyDrive/\n",
216
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/total_*.npy /content/drive/MyDrive/\n",
217
+ "\n",
218
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/weights/{MODELNAME}.pth /content/drive/MyDrive/{MODELNAME}{MODELEPOCH}.pth"
219
+ ],
220
+ "metadata": {
221
+ "id": "FgJuNeAwx5Y_"
222
+ },
223
+ "execution_count": null,
224
+ "outputs": []
225
+ },
226
+ {
227
+ "cell_type": "code",
228
+ "source": [
229
+ "#@title 从谷歌云盘恢复pth\n",
230
+ "#@markdown 需要自己查看logs文件夹下模型的文件名,手动修改下方命令末尾的文件名\n",
231
+ "\n",
232
+ "#@markdown 模型名\n",
233
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
234
+ "#@markdown 模型epoch\n",
235
+ "MODELEPOCH = 7500 #@param {type:\"integer\"}\n",
236
+ "\n",
237
+ "!mkdir -p /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}\n",
238
+ "\n",
239
+ "!cp /content/drive/MyDrive/{MODELNAME}_D_{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/G_{MODELEPOCH}.pth\n",
240
+ "!cp /content/drive/MyDrive/{MODELNAME}_G_{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/D_{MODELEPOCH}.pth\n",
241
+ "!cp /content/drive/MyDrive/*.index /content/\n",
242
+ "!cp /content/drive/MyDrive/*.npy /content/\n",
243
+ "!cp /content/drive/MyDrive/{MODELNAME}{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/weights/{MODELNAME}.pth"
244
+ ],
245
+ "metadata": {
246
+ "id": "OVQoLQJXS7WX"
247
+ },
248
+ "execution_count": null,
249
+ "outputs": []
250
+ },
251
+ {
252
+ "cell_type": "code",
253
+ "source": [
254
+ "#@title 手动预处理(不推荐)\n",
255
+ "#@markdown 模型名\n",
256
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
257
+ "#@markdown 采样率\n",
258
+ "BITRATE = 48000 #@param {type:\"integer\"}\n",
259
+ "#@markdown 使用的进程数\n",
260
+ "THREADCOUNT = 8 #@param {type:\"integer\"}\n",
261
+ "\n",
262
+ "!python3 trainset_preprocess_pipeline_print.py /content/dataset {BITRATE} {THREADCOUNT} logs/{MODELNAME} True\n"
263
+ ],
264
+ "metadata": {
265
+ "id": "ZKAyuKb9J6dz"
266
+ },
267
+ "execution_count": null,
268
+ "outputs": []
269
+ },
270
+ {
271
+ "cell_type": "code",
272
+ "source": [
273
+ "#@title 手动提取特征(不推荐)\n",
274
+ "#@markdown 模型名\n",
275
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
276
+ "#@markdown 使用的进程数\n",
277
+ "THREADCOUNT = 8 #@param {type:\"integer\"}\n",
278
+ "#@markdown 音高提取算法\n",
279
+ "ALGO = \"harvest\" #@param {type:\"string\"}\n",
280
+ "\n",
281
+ "!python3 extract_f0_print.py logs/{MODELNAME} {THREADCOUNT} {ALGO}\n",
282
+ "\n",
283
+ "!python3 extract_feature_print.py cpu 1 0 0 logs/{MODELNAME}\n"
284
+ ],
285
+ "metadata": {
286
+ "id": "CrxJqzAUKmPJ"
287
+ },
288
+ "execution_count": null,
289
+ "outputs": []
290
+ },
291
+ {
292
+ "cell_type": "code",
293
+ "source": [
294
+ "#@title 手动训练(不推荐)\n",
295
+ "#@markdown 模型名\n",
296
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
297
+ "#@markdown 使用的GPU\n",
298
+ "USEGPU = \"0\" #@param {type:\"string\"}\n",
299
+ "#@markdown 批大小\n",
300
+ "BATCHSIZE = 32 #@param {type:\"integer\"}\n",
301
+ "#@markdown 停止的epoch\n",
302
+ "MODELEPOCH = 3200 #@param {type:\"integer\"}\n",
303
+ "#@markdown 保存epoch间隔\n",
304
+ "EPOCHSAVE = 100 #@param {type:\"integer\"}\n",
305
+ "#@markdown 采样率\n",
306
+ "MODELSAMPLE = \"48k\" #@param {type:\"string\"}\n",
307
+ "#@markdown 是否缓存训练集\n",
308
+ "CACHEDATA = 1 #@param {type:\"integer\"}\n",
309
+ "#@markdown 是否仅保存最新的ckpt文件\n",
310
+ "ONLYLATEST = 0 #@param {type:\"integer\"}\n",
311
+ "\n",
312
+ "!python3 train_nsf_sim_cache_sid_load_pretrain.py -e lulu -sr {MODELSAMPLE} -f0 1 -bs {BATCHSIZE} -g {USEGPU} -te {MODELEPOCH} -se {EPOCHSAVE} -pg pretrained/f0G{MODELSAMPLE}.pth -pd pretrained/f0D{MODELSAMPLE}.pth -l {ONLYLATEST} -c {CACHEDATA}\n"
313
+ ],
314
+ "metadata": {
315
+ "id": "IMLPLKOaKj58"
316
+ },
317
+ "execution_count": null,
318
+ "outputs": []
319
+ },
320
+ {
321
+ "cell_type": "code",
322
+ "source": [
323
+ "#@title 删除其它pth,只留选中的(慎点,仔细看代码)\n",
324
+ "#@markdown 模型名\n",
325
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
326
+ "#@markdown 选中模型epoch\n",
327
+ "MODELEPOCH = 9600 #@param {type:\"integer\"}\n",
328
+ "\n",
329
+ "!echo \"备份选中的模型。。。\"\n",
330
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/G_{MODELEPOCH}.pth /content/{MODELNAME}_D_{MODELEPOCH}.pth\n",
331
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/D_{MODELEPOCH}.pth /content/{MODELNAME}_G_{MODELEPOCH}.pth\n",
332
+ "\n",
333
+ "!echo \"正在删除。。。\"\n",
334
+ "!ls /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}\n",
335
+ "!rm /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/*.pth\n",
336
+ "\n",
337
+ "!echo \"恢复选中的模型。。。\"\n",
338
+ "!mv /content/{MODELNAME}_D_{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/G_{MODELEPOCH}.pth \n",
339
+ "!mv /content/{MODELNAME}_G_{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/D_{MODELEPOCH}.pth\n",
340
+ "\n",
341
+ "!echo \"删除完成\"\n",
342
+ "!ls /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}"
343
+ ],
344
+ "metadata": {
345
+ "id": "haYA81hySuDl"
346
+ },
347
+ "execution_count": null,
348
+ "outputs": []
349
+ },
350
+ {
351
+ "cell_type": "code",
352
+ "source": [
353
+ "#@title 清除项目���所有文件,只留选中的模型(慎点,仔细看代码)\n",
354
+ "#@markdown 模型名\n",
355
+ "MODELNAME = \"lulu\" #@param {type:\"string\"}\n",
356
+ "#@markdown 选中模型epoch\n",
357
+ "MODELEPOCH = 9600 #@param {type:\"integer\"}\n",
358
+ "\n",
359
+ "!echo \"备份选中的模型。。。\"\n",
360
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/G_{MODELEPOCH}.pth /content/{MODELNAME}_D_{MODELEPOCH}.pth\n",
361
+ "!cp /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/D_{MODELEPOCH}.pth /content/{MODELNAME}_G_{MODELEPOCH}.pth\n",
362
+ "\n",
363
+ "!echo \"正在删除。。。\"\n",
364
+ "!ls /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}\n",
365
+ "!rm -rf /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/*\n",
366
+ "\n",
367
+ "!echo \"恢复选中的模型。。。\"\n",
368
+ "!mv /content/{MODELNAME}_D_{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/G_{MODELEPOCH}.pth \n",
369
+ "!mv /content/{MODELNAME}_G_{MODELEPOCH}.pth /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}/D_{MODELEPOCH}.pth\n",
370
+ "\n",
371
+ "!echo \"删除完成\"\n",
372
+ "!ls /content/Retrieval-based-Voice-Conversion-WebUI/logs/{MODELNAME}"
373
+ ],
374
+ "metadata": {
375
+ "id": "QhSiPTVPoIRh"
376
+ },
377
+ "execution_count": null,
378
+ "outputs": []
379
+ }
380
+ ]
381
+ }
docker/Retrieval-based-Voice-Conversion-WebUI/hubert_base.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f54b40fd2802423a5643779c4861af1e9ee9c1564dc9d32f54f20b5ffba7db96
3
+ size 189507909