Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel
Alexei Baevski 67af40c9cc
allow specifying max_tokens for generation
6 years ago
..
56f9ec3c38
Use ATen built-in conv_tbc method (#66)
6 years ago
d3795d6cd1
Merge internal changes (#136)
6 years ago
12a47a64ce
fix flag copy paste (decoder-normalize-before)
5 years ago
d6be0c7e00
Use FP32 for multi-head attention softmax
5 years ago
7c7634f638
Support --warmup-updates with fixed LR schedule
5 years ago
6641520612
fairseq-py goes distributed (#106)
6 years ago
ec0031df7b
Merge internal changes (#163)
6 years ago
8afb77612c
Fix tests
5 years ago
a04c4cf46f
Fix --remove-bpe to strip trailing BPE symbols
5 years ago
8fcdb9b726
Fix Flake8
5 years ago
8fcdb9b726
Fix Flake8
5 years ago
29153e279f
Update dataset code for use by https://github.com/pytorch/translate/pull/62 (#161)
6 years ago
0daba38ecb
Save and restore wall time in checkpoints
5 years ago
6641520612
fairseq-py goes distributed (#106)
6 years ago
67af40c9cc
allow specifying max_tokens for generation
5 years ago
0e8414f9ae
Remove sweep_log prefix from json progress bar
5 years ago
9f1b37ddcd
fix alignment when using uneven batches and left pad
5 years ago
c53b2ee06a
Remove padding from --score-reference
5 years ago
8fcdb9b726
Fix Flake8
5 years ago
8fcdb9b726
Fix Flake8
5 years ago
7ee1d28458
Add FP16 support
5 years ago

Comments

Loading...