Skip to content

Added functionality of storing layers activations output.#145

Merged
copybara-service[bot] merged 3 commits intogoogle:devfrom
atorero:dev
Apr 12, 2024
Merged

Added functionality of storing layers activations output.#145
copybara-service[bot] merged 3 commits intogoogle:devfrom
atorero:dev

Conversation

@atorero
Copy link
Contributor

@atorero atorero commented Apr 12, 2024

Adding a way to save activations output after applying each layer. This can be useful for two purposes:

  • Debugging the model and comparing the output to another working implementation
  • Quantising weights or activations and getting insights of the accumulated error.

There will be a similar functionality in the Python code and a diffing tool that will compare two outputs and calculating some metrics for them. See the output of such a tool when comparing outputs of (3 tokens)x(27 layers) between the Python and the C++ (8-bit) Recurrent Gemma implementations:

image

Copy link
Member

@jan-wassenberg jan-wassenberg left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nice, thanks for updating the PR!

@jan-wassenberg jan-wassenberg added the copybara-import Trigger Copybara for merging pull requests label Apr 12, 2024
@copybara-service copybara-service bot merged commit 05e7e2b into google:dev Apr 12, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

copybara-import Trigger Copybara for merging pull requests

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants