-
Notifications
You must be signed in to change notification settings - Fork 37
Detokenization parallelization public repo #69
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Detokenization parallelization public repo #69
Conversation
|
|
CNN 1k data set is too small now, result is not reliable. Please use full valid set. |
|
|
For |
This is because pytorch dataloader API only supports number of samples to be multiple of batch size. This is why last batch is dropped. |
|
@JiushengChen For Mbart and T5 , do we have larger dataset? I couldn't find it in benchmark scripts. |
The root cause may be here. Try to change it to be |
Yes, I have larger data in my local. Please leave out these two, I will update them today. |
I already did. |
Could you please explain more? I saw your code here use |
benchmarks/models/hf_distibart.sh
Outdated
| grep -E "transformers_v3.0.2\+fastseq_v.* hf.sshleifer.distilbart-cnn-12-6.tar.gz cnn_dm.1k/raw val 64 " perf | awk '{s+=$13}END{print s/NR}' | bash range.sh 13 100 | ||
| grep -E "transformers_v3.0.2\+fastseq_v.* hf.sshleifer.distilbart-cnn-12-6.tar.gz cnn_dm.1k/raw val 64 " perf | awk '{s+=$13}END{print s/NR}' | bash range.sh 15.2 100 | ||
| # todo: bigger bs doesn't increase speed | ||
| grep -E "transformers_v3.0.2\+fastseq_v.* hf.sshleifer.distilbart-cnn-12-6.tar.gz cnn_dm.1k/raw val 128 " perf | awk '{s+=$13}END{print s/NR}' | bash range.sh 13.5 100 | ||
| grep -E "transformers_v3.0.2\+fastseq_v.* hf.sshleifer.distilbart-cnn-12-6.tar.gz cnn_dm.1k/raw val 128 " perf | awk '{s+=$13}END{print s/NR}' | bash range.sh 15.9 100 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
same here
|
Benchmarks on Larger dataset:
|
|
Before/After
|
Synced offline. Included last batch. |
Moving #37 here.