diff --git a/.github/workflows/checks.yml b/.github/workflows/checks.yml index 9a802831e..727c807d8 100644 --- a/.github/workflows/checks.yml +++ b/.github/workflows/checks.yml @@ -240,6 +240,7 @@ jobs: # - "No_Position_Experiment" - "Othello_GPT" - "Patchscopes_Generation_Demo" + - "Santa_Coder" # - "T5" steps: - uses: actions/checkout@v3 diff --git a/demos/Santa_Coder.ipynb b/demos/Santa_Coder.ipynb index 0c95abd1d..e607da867 100644 --- a/demos/Santa_Coder.ipynb +++ b/demos/Santa_Coder.ipynb @@ -54,7 +54,7 @@ }, { "cell_type": "code", - "execution_count": 2, + "execution_count": null, "id": "da9f5a40", "metadata": {}, "outputs": [ @@ -97,11 +97,7 @@ "\n", "import transformer_lens\n", "import transformer_lens.utils as utils\n", - "from transformer_lens.hook_points import (\n", - " HookedRootModule,\n", - " HookPoint,\n", - ") # Hooking utilities\n", - "from transformer_lens import HookedTransformer, HookedTransformerConfig, FactoredMatrix, ActivationCache\n", + "from transformer_lens.model_bridge import TransformerBridge\n", "\n", "torch.set_grad_enabled(False)\n", "\n", @@ -132,7 +128,7 @@ }, { "cell_type": "code", - "execution_count": 4, + "execution_count": null, "id": "1f7ac1e1", "metadata": {}, "outputs": [ @@ -154,7 +150,8 @@ "source": [ "# Disable folding norms and folding norms and biases so that intermediate value\n", "# in between transformer blocks can be compared\n", - "bloom = HookedTransformer.from_pretrained(\"bloom-560m\",fold_ln=False, fold_value_biases=False, center_writing_weights=False)" + "bloom = TransformerBridge.boot_transformers(\"bloom-560m\",fold_ln=False, fold_value_biases=False, center_writing_weights=False)\n", + "bloom.enable_compatibility_mode()" ] }, {