Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
F
fairseq
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Container Registry
Model registry
Operate
Environments
Monitor
Incidents
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
Simon Will
fairseq
Repository graph
Repository graph
You can move around the graph by using the arrow keys.
master
Select Git revision
Branches
4
master
default
protected
remove-decoder-and-encoder-ff
remove-decoder-ff
remove-encoder-ff
Tags
6
v0.6.2
v0.6.1
v0.6.0
v0.5.0
v0.4.0
v0.3.0
10 results
Begin with the selected commit
Created with Raphaël 2.2.0
6
May
5
3
2
1
30
Apr
29
27
26
25
24
22
21
17
16
15
12
10
9
7
5
4
3
2
1
28
Mar
26
19
15
14
12
11
4
2
1
28
Feb
25
23
22
19
15
12
8
6
5
1
30
Jan
29
25
24
16
15
14
9
7
5
4
28
Dec
27
26
24
18
10
8
7
6
5
4
30
Nov
28
27
26
25
19
18
17
16
14
13
9
8
7
6
1
31
Oct
29
26
25
22
20
19
17
5
3
2
30
Sep
25
24
17
7
3
16
Aug
1
31
Jul
26
25
19
11
10
8
7
2
28
Jun
26
25
24
21
20
15
24
May
22
21
9
1
2
Apr
28
Mar
29
28
26
25
7
5
1
27
Feb
12
8
31
Jan
29
26
22
12
5
6
Dec
1
24
Nov
13
12
11
8
1
24
Oct
19
17
14
13
12
11
2
28
Sep
26
24
19
18
15
14
Add script for preparing WMT18-Multi30k with BPE
master
master
Remove two-layer ff network from transformer encoder
remove-decoder-…
remove-decoder-and-encoder-ff
Remove two-layer ff network from transformer decoder
remove-decoder-ff
remove-decoder-ff
Remove two-layer ff network from transformer encoder
remove-encoder-ff
remove-encoder-ff
Bugfix in size of multi-corpus dataset
an option to raise exception if oom happens during fairseq.trainer.train_step (#2)
added bert large architecture (#698)
Make learned positional embedding optional
Move distributed_init into DistributedFairseqModel (#687)
Validate on all sets based on --save-interval-updates
Fix inconsistent gradient check
Make CTC work with more encoder-only models
Make MultiCorpusSampledDataset and IndexedCachedDataset Picklable
add ConcatDataset support for XLM
Support dataset upsampling / relative ratio in PytorchTranslateTask (#494)
Better OOM recovery
Add default noising argument in WordNoiser initialization (#664)
addding polynomial lr scheduler (#683)
Merge internal changes
Add rm_pt.py helper script for removing checkpoint files
Merge internal changes (#654)
Add more details in error message when sentence length > max tokens (#672)
Fix upgrade_state_dict for XLM Transformer sentence encoder (#680)
Update README.md (#679)
Update comments and citations
Add args and sys.argv to tensorboard (#673)
Add small comments for MonolingualDataset and TokenBlockDataset
Passing kwargs in setup_task in fairseq_task (#670)
Fix fairseq unittest timeouts (#667)
XLM for NMT: option to only load encoder or decoder (#666)
Load a XLM model into transformer encoder / decoder for MT training (#629)
Add gelu and gelu_fast as possible activation functions (#653)
Added link to blog post (#662)
added link to sample stories
Don't reload best validation loss when using --reset-optimizer
Fix generation with --no-early-stop (#627)
reduce memory footprint for average_checkpoints (#647)
Open BlockPairDataset for MaskedLMData to work (#641)
Enable custom sampling strategy in MultiCorpusSampledDataset (#639)
Black formatting for multi_corpus_sampled_dataset.py (#638)
Loading