You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Jan 5, 2023. It is now read-only.
- You can now use `$HOME` and `$USER` in your configuration files.
60
+
- Fixed an overflow error that would cause NMT with more than 255 tokens to fail.
61
+
- METEOR worker process is now correctly killed after validations.
62
+
- Many runs of an experiment are now suffixed with a unique random string instead of incremental integers to avoid race conditions in cluster setups.
63
+
- Replaced `utils.nn.get_network_topology()` with a new `Topology`[class](nmtpytorch/utils/topology.py) that will parse the `direction` string of the model in a more smart way.
64
+
- If `CUDA_VISIBLE_DEVICES` is set, the `GPUManager` will always honor it.
65
+
- Dropped creation of temporary/advisory lock files under `/tmp` for GPU reservation.
66
+
- Time measurements during training are now structered into batch overhead, training and evaluation timings.
67
+
-**Datasets**
68
+
- Added `TextDataset` for standalone text file reading.
69
+
- Added `OneHotDataset`, a variant of `TextDataset` where the sequences are not prefixed/suffixed with `<bos>` and `<eos>` respectively.
70
+
- Added experimental `MultiParallelDataset` that merges an arbitrary number of parallel datasets together.
71
+
-**nmtpy translate**
72
+
-`.nodbl` and `.nounk` suffixes are now added to output files for `--avoid-double` and `--avoid-unk` arguments respectively.
73
+
- A model-agnostic enough `beam_search()` is now separated out into its own file `nmtpytorch/search.py`.
74
+
-`max_len` default is increased to 200.
75
+
57
76
### v1.1 (25/01/2018)
58
77
59
78
- New experimental `Multi30kDataset` and `ImageFolderDataset` classes
@@ -233,17 +252,12 @@ steps are to:
233
252
as an attribute of the model (i.e. `self.encoder = ...`) in order for
234
253
PyTorch to work correctly.
235
254
- Create and store relevant dataset objects in `load_data()`
236
-
- Define a `get_iterator()` method to obtain a dataset-specific iterator
237
-
depending on train/dev and loss/beam-search variants.
238
255
- Define `compute_loss()` which takes a data iterator and
239
256
computes the loss over it. This method is used for dev set perplexities.
240
-
- Define `aux_loss()` where you may want to define an additional loss term.
241
-
Return just `0` if you do not want to use it.
257
+
- Set `aux_loss` attribute for an additional loss term.
242
258
- Define `forward()` which takes a dictionary with keys as data sources and
243
259
returns the batch training loss. This is the method called from the `mainloop`
244
260
during training.
245
-
- Define `beam_search()` which takes a data iterator and generates hypotheses
246
-
using beam search. The default implementation in `NMT` is a batched GPU version.
247
261
248
262
Feel free to copy the methods from `NMT` if you do not need to modify
0 commit comments