@@ -298,8 +298,8 @@ def create(
298298 lr_scheduler_type (Literal["linear", "cosine"]): Learning rate scheduler type. Defaults to "linear".
299299 min_lr_ratio (float, optional): Min learning rate ratio of the initial learning rate for
300300 the learning rate scheduler. Defaults to 0.0.
301- num_cycles (float, optional): Number of cycles for cosine learning rate scheduler. Defaults to 0.5.
302- warmup_ratio (float, optional): Warmup ratio for learning rate scheduler.
301+ num_cycles (float, optional): Number of cycles for the cosine learning rate scheduler. Defaults to 0.5.
302+ warmup_ratio (float, optional): Warmup ratio for the learning rate scheduler.
303303 max_grad_norm (float, optional): Max gradient norm. Defaults to 1.0, set to 0 to disable.
304304 weight_decay (float, optional): Weight decay. Defaults to 0.0.
305305 lora (bool, optional): Whether to use LoRA adapters. Defaults to True.
@@ -677,8 +677,8 @@ async def create(
677677 lr_scheduler_type (Literal["linear", "cosine"]): Learning rate scheduler type. Defaults to "linear".
678678 min_lr_ratio (float, optional): Min learning rate ratio of the initial learning rate for
679679 the learning rate scheduler. Defaults to 0.0.
680- num_cycles (float, optional): Number of cycles for cosine learning rate scheduler. Defaults to 0.5.
681- warmup_ratio (float, optional): Warmup ratio for learning rate scheduler.
680+ num_cycles (float, optional): Number of cycles for the cosine learning rate scheduler. Defaults to 0.5.
681+ warmup_ratio (float, optional): Warmup ratio for the learning rate scheduler.
682682 max_grad_norm (float, optional): Max gradient norm. Defaults to 1.0, set to 0 to disable.
683683 weight_decay (float, optional): Weight decay. Defaults to 0.0.
684684 lora (bool, optional): Whether to use LoRA adapters. Defaults to True.
0 commit comments