You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This commit was created on GitHub.com and signed with GitHub’s verified signature.
The key has expired.
[3.1.33]
Fixed
Two small fixes to SampleK. Before the device was not set correctly leading to issues when running sampling on GPUs. Furthermore, SampleK did not return the top-k values correctly.
[3.1.32]
Added
Sockeye now supports blocking cross-attention between decoder and encoded prepended tokens.
If the source contains prepended text and a tag indicating the end of prepended text,
Sockeye supports blocking the cross-attention between decoder and encoded prepended tokens (including the tag).
To enable this operation, specify --end-of-prepending-tag for training or data preparation,
and --transformer-block-prepended-cross-attention for training.
Changed
Sockeye uses a new dictionary-based prepared data format that supports storing length of prepended source tokens
(version 7). The previous format (version 6) is still supported.