:zap: [Enhance] Align with openai api /models format
Browse files
components/llm_models_loader.js
CHANGED
@@ -24,7 +24,7 @@ export async function setup_available_models_on_select(default_option = null) {
|
|
24 |
}
|
25 |
|
26 |
select.val(default_model);
|
27 |
-
console.log(`
|
28 |
}
|
29 |
|
30 |
export async function setup_temperature_on_select(default_option = null) {
|
|
|
24 |
}
|
25 |
|
26 |
select.val(default_model);
|
27 |
+
console.log(`default_ model: ${select.val()}`);
|
28 |
}
|
29 |
|
30 |
export async function setup_temperature_on_select(default_option = null) {
|
networks/llm_requester.js
CHANGED
@@ -136,14 +136,15 @@ export class AvailableModelsRequester {
|
|
136 |
return fetch(this.backend_request_endpoint, this.backend_request_params)
|
137 |
.then((response) => response.json())
|
138 |
.then((response_json) => {
|
139 |
-
response_json.
|
|
|
140 |
if (!(item.id in available_models)) {
|
141 |
available_models.push(item.id);
|
142 |
}
|
143 |
});
|
144 |
available_models.sort();
|
145 |
available_models = [...new Set(available_models)];
|
146 |
-
console.log(available_models);
|
147 |
})
|
148 |
.catch((error) => {
|
149 |
console.error("Error:", error);
|
|
|
136 |
return fetch(this.backend_request_endpoint, this.backend_request_params)
|
137 |
.then((response) => response.json())
|
138 |
.then((response_json) => {
|
139 |
+
let data = response_json.data;
|
140 |
+
data.forEach((item) => {
|
141 |
if (!(item.id in available_models)) {
|
142 |
available_models.push(item.id);
|
143 |
}
|
144 |
});
|
145 |
available_models.sort();
|
146 |
available_models = [...new Set(available_models)];
|
147 |
+
console.log("available_models:", available_models);
|
148 |
})
|
149 |
.catch((error) => {
|
150 |
console.error("Error:", error);
|