Skip to content
This repository was archived by the owner on Jan 5, 2023. It is now read-only.

Commit 4f56c04

Browse files
authored
Update README for v1.2.0
1 parent b243cae commit 4f56c04

File tree

1 file changed

+20
-6
lines changed

1 file changed

+20
-6
lines changed

README.md

Lines changed: 20 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -54,6 +54,25 @@ nmtpy train -C <config file> train.<opt>:<val> model.<opt>:<val> ...
5454

5555
## Release Notes
5656

57+
### v1.2 (20/02/2018)
58+
59+
- You can now use `$HOME` and `$USER` in your configuration files.
60+
- Fixed an overflow error that would cause NMT with more than 255 tokens to fail.
61+
- METEOR worker process is now correctly killed after validations.
62+
- Many runs of an experiment are now suffixed with a unique random string instead of incremental integers to avoid race conditions in cluster setups.
63+
- Replaced `utils.nn.get_network_topology()` with a new `Topology` [class](nmtpytorch/utils/topology.py) that will parse the `direction` string of the model in a more smart way.
64+
- If `CUDA_VISIBLE_DEVICES` is set, the `GPUManager` will always honor it.
65+
- Dropped creation of temporary/advisory lock files under `/tmp` for GPU reservation.
66+
- Time measurements during training are now structered into batch overhead, training and evaluation timings.
67+
- **Datasets**
68+
- Added `TextDataset` for standalone text file reading.
69+
- Added `OneHotDataset`, a variant of `TextDataset` where the sequences are not prefixed/suffixed with `<bos>` and `<eos>` respectively.
70+
- Added experimental `MultiParallelDataset` that merges an arbitrary number of parallel datasets together.
71+
- **nmtpy translate**
72+
- `.nodbl` and `.nounk` suffixes are now added to output files for `--avoid-double` and `--avoid-unk` arguments respectively.
73+
- A model-agnostic enough `beam_search()` is now separated out into its own file `nmtpytorch/search.py`.
74+
- `max_len` default is increased to 200.
75+
5776
### v1.1 (25/01/2018)
5877

5978
- New experimental `Multi30kDataset` and `ImageFolderDataset` classes
@@ -233,17 +252,12 @@ steps are to:
233252
as an attribute of the model (i.e. `self.encoder = ...`) in order for
234253
PyTorch to work correctly.
235254
- Create and store relevant dataset objects in `load_data()`
236-
- Define a `get_iterator()` method to obtain a dataset-specific iterator
237-
depending on train/dev and loss/beam-search variants.
238255
- Define `compute_loss()` which takes a data iterator and
239256
computes the loss over it. This method is used for dev set perplexities.
240-
- Define `aux_loss()` where you may want to define an additional loss term.
241-
Return just `0` if you do not want to use it.
257+
- Set `aux_loss` attribute for an additional loss term.
242258
- Define `forward()` which takes a dictionary with keys as data sources and
243259
returns the batch training loss. This is the method called from the `mainloop`
244260
during training.
245-
- Define `beam_search()` which takes a data iterator and generates hypotheses
246-
using beam search. The default implementation in `NMT` is a batched GPU version.
247261

248262
Feel free to copy the methods from `NMT` if you do not need to modify
249263
some of them.

0 commit comments

Comments
 (0)