diff --git a/self_hosting_machinery/inference/stream_results.py b/self_hosting_machinery/inference/stream_results.py index 2f9815c2..475b86c4 100644 --- a/self_hosting_machinery/inference/stream_results.py +++ b/self_hosting_machinery/inference/stream_results.py @@ -53,7 +53,7 @@ def url_complain_doesnt_work(): def model_guid_allowed_characters(name): - return re.sub(r"[^a-zA-Z0-9_]", "_", name) + return re.sub(r"[^a-zA-Z0-9_\.]", "_", name) def validate_description_dict( @@ -242,7 +242,7 @@ def upload_result( progress[original_batch[b]["id"]] = tmp upload_dict["progress"] = progress upload_dict["check_cancelled"] = [call["id"] for call in original_batch] - upload_dict["model_name"] = description_dict["model"] + upload_dict["model_name"] = description_dict["model"].replace("/vllm", "") self.upload_q.put(copy.deepcopy(upload_dict)) if DEBUG_UPLOAD_NOT_SEPARATE_PROCESS: _upload_results_loop(self.upload_q, self.cancelled_q) diff --git a/self_hosting_machinery/scripts/first_run.py b/self_hosting_machinery/scripts/first_run.py index f2716008..dfae7b85 100644 --- a/self_hosting_machinery/scripts/first_run.py +++ b/self_hosting_machinery/scripts/first_run.py @@ -25,3 +25,4 @@ def convert_old_configs(): convert_old_configs() model_assigner = ModelAssigner() assign_gpus_if_first_run_detected(model_assigner) + model_assigner.models_to_watchdog_configs() # removes deprecated models