GitXplorerGitXplorer
e

autotransformer

public
9 stars
0 forks
1 issues

Commits

List of commits on branch master.
Unverified
7990379ba39077b8f57c54d0a5bdc66e1b4a075c

added references

eeyaler committed 6 years ago
Unverified
4de6677c961a2d1b25866ea970a64a35c0f8cfcf

add references to readme

eeyaler committed 6 years ago
Unverified
ce9f27335b442cabc75a222a1b6061cb74a41d72

first code push

eeyaler committed 6 years ago
Unverified
91518a59c426a8996dbff88459b471e4f6331d9a

added transformer_tpu.txt

eeyaler committed 6 years ago
Verified
f5c43a24d8ad32a1ec9b862cc3a0f6da483b4f11

Update README.md

eeyaler committed 6 years ago
Verified
8d0a08ef03fed7b6a35528dfe76e39c231155977

Merge pull request #1 from eyaler/add-license-1

eeyaler committed 6 years ago

README

The README file for this repository.

Autotransformer

A transformer variant for semantic continuous sentence representation and generation

We modify the transformer to give a fixed-length and distributed sentence representation. This allows sampling from the latent space to generate sentences, semantic interpolations, sentence analogies etc. The code is based on the Tensorflow Tensor2Tensor library and trains on TPU out of the box.

based on:

Vaswani et al., Attention Is All You Need, arxiv.org/abs/1706.03762

Cer et al., Universal Sentence Encoder, arxiv.org/abs/1803.11175

Cifka et al., Eval all, trust a few, do wrong to none: Comparing sentence generation models, arxiv.org/abs/1804.07972

Cifka and Bojar, Are BLEU and Meaning Representation in Opposition? arxiv.org/abs/1805.06536

motivation:

Bowman et al., Generating Sentences from a Continuous Space, arxiv.org/abs/1511.06349

Gan et al., Learning Generic Sentence Representations Using Convolutional Neural Networks arxiv.org/abs/1611.07897

Semeniuta et al., A Hybrid Convolutional Variational Autoencoder for Text Generation, arxiv.org/abs/1702.02390

Zhao et al, Adversarially Regularized Autoencoders, arxiv.org/abs/1706.04223

https://blogs.helsinki.fi/language-technology/files/2017/09/FINMT2017-Bojar.pdf

A rough guide to training transformer on TPU can be found here:

https://github.com/eyaler/autotransformer/blob/master/transformer_tpu.txt

Acknowledgements: This work was supported by Deep Learning Camp Jeju 2018 which was organized by TensorFlow Korea User Group