@@ -48,6 +48,9 @@ def createFinetuneRequest(
4848 lora_trainable_modules : str | None = "all-linear" ,
4949 suffix : str | None = None ,
5050 wandb_api_key : str | None = None ,
51+ wandb_base_url : str | None = None ,
52+ wandb_project_name : str | None = None ,
53+ wandb_name : str | None = None ,
5154 train_on_inputs : bool | Literal ["auto" ] = "auto" ,
5255) -> FinetuneRequest :
5356 if batch_size == "max" :
@@ -118,6 +121,9 @@ def createFinetuneRequest(
118121 training_type = training_type ,
119122 suffix = suffix ,
120123 wandb_key = wandb_api_key ,
124+ wandb_base_url = wandb_base_url ,
125+ wandb_project_name = wandb_project_name ,
126+ wandb_name = wandb_name ,
121127 train_on_inputs = train_on_inputs ,
122128 )
123129
@@ -150,6 +156,9 @@ def create(
150156 lora_trainable_modules : str | None = "all-linear" ,
151157 suffix : str | None = None ,
152158 wandb_api_key : str | None = None ,
159+ wandb_base_url : str | None = None ,
160+ wandb_project_name : str | None = None ,
161+ wandb_name : str | None = None ,
153162 verbose : bool = False ,
154163 model_limits : FinetuneTrainingLimits | None = None ,
155164 train_on_inputs : bool | Literal ["auto" ] = "auto" ,
@@ -182,6 +191,12 @@ def create(
182191 Defaults to None.
183192 wandb_api_key (str, optional): API key for Weights & Biases integration.
184193 Defaults to None.
194+ wandb_base_url (str, optional): Base URL for Weights & Biases integration.
195+ Defaults to None.
196+ wandb_project_name (str, optional): Project name for Weights & Biases integration.
197+ Defaults to None.
198+ wandb_name (str, optional): Run name for Weights & Biases integration.
199+ Defaults to None.
185200 verbose (bool, optional): whether to print the job parameters before submitting a request.
186201 Defaults to False.
187202 model_limits (FinetuneTrainingLimits, optional): Limits for the hyperparameters the model in Fine-tuning.
@@ -225,6 +240,9 @@ def create(
225240 lora_trainable_modules = lora_trainable_modules ,
226241 suffix = suffix ,
227242 wandb_api_key = wandb_api_key ,
243+ wandb_base_url = wandb_base_url ,
244+ wandb_project_name = wandb_project_name ,
245+ wandb_name = wandb_name ,
228246 train_on_inputs = train_on_inputs ,
229247 )
230248
@@ -479,6 +497,9 @@ async def create(
479497 lora_trainable_modules : str | None = "all-linear" ,
480498 suffix : str | None = None ,
481499 wandb_api_key : str | None = None ,
500+ wandb_base_url : str | None = None ,
501+ wandb_project_name : str | None = None ,
502+ wandb_name : str | None = None ,
482503 verbose : bool = False ,
483504 model_limits : FinetuneTrainingLimits | None = None ,
484505 train_on_inputs : bool | Literal ["auto" ] = "auto" ,
@@ -511,6 +532,12 @@ async def create(
511532 Defaults to None.
512533 wandb_api_key (str, optional): API key for Weights & Biases integration.
513534 Defaults to None.
535+ wandb_base_url (str, optional): Base URL for Weights & Biases integration.
536+ Defaults to None.
537+ wandb_project_name (str, optional): Project name for Weights & Biases integration.
538+ Defaults to None.
539+ wandb_name (str, optional): Run name for Weights & Biases integration.
540+ Defaults to None.
514541 verbose (bool, optional): whether to print the job parameters before submitting a request.
515542 Defaults to False.
516543 model_limits (FinetuneTrainingLimits, optional): Limits for the hyperparameters the model in Fine-tuning.
@@ -554,6 +581,9 @@ async def create(
554581 lora_trainable_modules = lora_trainable_modules ,
555582 suffix = suffix ,
556583 wandb_api_key = wandb_api_key ,
584+ wandb_base_url = wandb_base_url ,
585+ wandb_project_name = wandb_project_name ,
586+ wandb_name = wandb_name ,
557587 train_on_inputs = train_on_inputs ,
558588 )
559589
0 commit comments