You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: lightning_examples/finetuning-scheduler/finetuning-scheduler.py
+6-9Lines changed: 6 additions & 9 deletions
Original file line number
Diff line number
Diff line change
@@ -554,9 +554,7 @@ def train() -> None:
554
554
# the implicit schedule will limit fine-tuning to just the last 4 parameters of the model, which is only a small fraction
555
555
# of the parameters you'd want to tune for maximum performance. Since the implicit schedule is quite computationally
556
556
# intensive and most useful for exploring model behavior, leaving [max_depth](https://finetuning-scheduler.readthedocs.io/en/stable/api/finetuning_scheduler.fts.html?highlight=max_depth#finetuning_scheduler.fts.FinetuningScheduler.params.max_depth) 1 allows us to demo implicit mode
557
-
# behavior while keeping the computational cost and runtime of this notebook reasonable. To review how a full implicit
558
-
# mode run compares to the ``nofts_baseline`` and ``fts_explicit`` scenarios, please see the the following
# behavior while keeping the computational cost and runtime of this notebook reasonable.
560
558
561
559
562
560
# %%
@@ -579,16 +577,15 @@ def train() -> None:
579
577
# %% [markdown]
580
578
# ### Reviewing the Training Results
581
579
#
582
-
# See the [tensorboard experiment summaries](https://tensorboard.dev/experiment/n7U8XhrzRbmvVzC4SQSpWw/) to get a sense
583
-
# of the relative computational and performance tradeoffs associated with these [FinetuningScheduler](https://finetuning-scheduler.readthedocs.io/en/stable/api/finetuning_scheduler.fts.html#finetuning_scheduler.fts.FinetuningScheduler) configurations.
584
-
# The summary compares a full ``fts_implicit`` execution to ``fts_explicit`` and ``nofts_baseline`` scenarios using DDP
580
+
# It's worth considering the relative computational and performance tradeoffs associated with different [FinetuningScheduler](https://finetuning-scheduler.readthedocs.io/en/stable/api/finetuning_scheduler.fts.html#finetuning_scheduler.fts.FinetuningScheduler) configurations.
581
+
# The example below compares ``fts_implicit`` execution to ``fts_explicit`` and ``nofts_baseline`` scenarios using DDP
585
582
# training with 2 GPUs. The full logs/schedules for all three scenarios are available
586
583
# [here](https://drive.google.com/file/d/1LrUcisRLHeJgh_BDOOD_GUBPp5iHAkoR/view?usp=sharing) and the checkpoints
587
584
# produced in the scenarios [here](https://drive.google.com/file/d/1t7myBgcqcZ9ax_IT9QVk-vFH_l_o5UXB/view?usp=sharing)
# Note that given execution context differences, there could be a modest variation in performance from the tensorboard summaries generated by this notebook.
594
591
#
@@ -597,7 +594,7 @@ def train() -> None:
597
594
# greater fine-tuning flexibility for model exploration in research. For example, glancing at DeBERTa-v3's implicit training
598
595
# run, a critical tuning transition point is immediately apparent:
0 commit comments