Skip to content
This repository has been archived by the owner on Dec 11, 2023. It is now read-only.

Added num_encoder_layers/num_decoder_layers to WMT16 standard hparams. #269

Closed
wants to merge 39 commits into from

Conversation

tiberiu92
Copy link

This is based on the fix #265

@googlebot
Copy link

Thanks for your pull request. It looks like this may be your first contribution to a Google open source project (if not, look below for help). Before we can look at your pull request, you'll need to sign a Contributor License Agreement (CLA).

📝 Please visit https://cla.developers.google.com/ to sign.

Once you've signed (or fixed any issues), please reply here (e.g. I signed it!) and we'll verify it.


What to do if you already signed the CLA

Individual signers
Corporate signers
  • Your company has a Point of Contact who decides which employees are authorized to participate. Ask your POC to be added to the group of authorized contributors. If you don't know who your Point of Contact is, direct the project maintainer to go/cla#troubleshoot.
  • The email used to register you as an authorized contributor must be the email used for the Git commit. Check your existing CLA data and verify that your email is set on your git commits.
  • The email used to register you as an authorized contributor must also be attached to your GitHub account.

@tiberiu92
Copy link
Author

I signed it !

@googlebot
Copy link

We found a Contributor License Agreement for you (the sender of this pull request), but were unable to find agreements for the commit author(s). If you authored these, maybe you used a different email address in the git commits than was used to sign the CLA (login here to double check)? If these were authored by someone else, then they will need to sign a CLA as well, and confirm that they're okay with these being contributed to Google.
In order to pass this check, please resolve this problem and have the pull request author add another comment and the bot will run again. If the bot doesn't comment, it means it doesn't think anything has changed.

lmthang and others added 26 commits August 25, 2018 11:27
to allow for flexibility in extending models.
Clean up and factor train.py

PiperOrigin-RevId: 180703151
  (a) During inference, given --ckpt, we can try to load hparams in the same dir
  (b) When loading models and override_loaded_hparams=False, we still overwrite
      ["beam_width", "length_penalty_weight", "sampling_temperature",
       "num_translations_per_input"]
  (c) Introduce _add_argument to smartly add argument to hparams, so
      extend_hparams can be called when loading hparams.
      This is useful for old checkpoints.
  (d) Handle old checkpoints before the separation of num_layers into
      num_encoder_layers and num_decoder_layers.
Minor clean-ups of misc_utils.py.

PiperOrigin-RevId: 180989949
PiperOrigin-RevId: 181096467
Update attention_model.py so that we can specify GNMT encoder without attention.

PiperOrigin-RevId: 181117151
…ferCheckpoint();

Rename ckpt into ckpt_path in inference.py and model_helper.py

PiperOrigin-RevId: 181260899
op.device actually returns what the user requested not the actual device. This can be misleading as it can return "GPU0" even if no GPU is available.

For context see: tensorflow/tensorflow#1344

PiperOrigin-RevId: 181261953
Rename _get_best_results to get_best_results.
Update avg_grad_norm computation to divide by the number of examples instead.

PiperOrigin-RevId: 181346178
PiperOrigin-RevId: 181399302
Add an option include_embeddings to allow for
  appending embedding layer in front of encoder state list.
Properly handle the case when time_major=True.

PiperOrigin-RevId: 183117301
Useful when vocab size is very large.

PiperOrigin-RevId: 183184262
PiperOrigin-RevId: 183778701
- Allow the construction of encoders from sequences different from the default source sequence.
- Cleanups.

PiperOrigin-RevId: 184301964
PiperOrigin-RevId: 184795279
Minor updates to nmt.py to print logging info on embedding files

PiperOrigin-RevId: 185313574
…e entry

that doesn't have the correct size.

Handle attention_architecture == "" same as attention_architecture == "standard".

Use separate embedding partitioner for encoder and decoder.

PiperOrigin-RevId: 185489121
PiperOrigin-RevId: 186098897
lmthang and others added 12 commits August 25, 2018 11:30
Added a implicit flag extract_encoder_layers to get
  intermediate layers from GNMT models and skip decoder.

PiperOrigin-RevId: 191678516
PiperOrigin-RevId: 191720585
PiperOrigin-RevId: 191804041
PiperOrigin-RevId: 196283435
PiperOrigin-RevId: 203278814
PiperOrigin-RevId: 207180389
PiperOrigin-RevId: 207608855
PiperOrigin-RevId: 208349749
@tiberiu92 tiberiu92 closed this Oct 17, 2018
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

6 participants