Skip to content

Tags: ismstat/sockeye

Tags

2.3.14

Toggle 2.3.14's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Update python-publish.yml to Python 3.7 (awslabs#949)

2.3.10

Toggle 2.3.10's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Avoid circular import: move cleanup method to training.py (awslabs#932)

2.3.8

Toggle 2.3.8's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
fix warmup of learning rate so that it works properly w/ continued tr… (

awslabs#926)

Co-authored-by: Steven Bradtke sjbradt <[email protected]>

2.3.2

Toggle 2.3.2's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Add some Sockeye publications (awslabs#912)

2.2.8

Toggle 2.2.8's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Make source/target data required for scoring. (awslabs#904)

2.2.0

Toggle 2.2.0's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Interleaved Multi-head Attention Operators (awslabs#884)

Replaced batched dot product in multi-head attention with interleaved_matmul attention operators to improve performance. Also changes the batch-major data to time-major format while in the model to comply with the new operator requirements.

2.1.21

Toggle 2.1.21's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Add ability to cache encoder outputs of model (awslabs#858)

2.1.17

Toggle 2.1.17's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Add SSRU layer and decoder (awslabs#851)

- Adds layers.SSRU, which implements a Simpler Simple Recurrent Unit as described by Kim et al, 2019.
- Adds ssru_transformer option to --decoder, which enables the usage of SSRUs as a replacement for the decoder-side self-attention layers.
- Reduces the number of arguments for MultiHeadSelfAttention.hybrid_forward().
previous_keys and previous_values should now be input together as previous_states, a list containing two symbols.

2.1.16

Toggle 2.1.16's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Fix batch sizing errors introduced in c00da52 (awslabs#849)

2.1.13

Toggle 2.1.13's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature. The key has expired.
Option to suppress console output for secondary workers (awslabs#841)