Difference between revisions of "NLP Status Report 2016-11-14"

From cslt Wiki
Jump to: navigation, search
Line 35: Line 35:
* try rnng on GPU
* try rnng on GPU
* read the code of Feng
* read the code of Feng
* modify model  
* modify model [http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2f/RNNG%2Bmm%E5%AE%9E%E9%AA%8C%E6%8A%A5%E5%91%8A.pdf report]
[http://cslt.riit.tsinghua.edu.cn/mediawiki/images/2/2f/RNNG%2Bmm%E5%AE%9E%E9%AA%8C%E6%8A%A5%E5%91%8A.pdf report]
* try MKL
* try MKL

Revision as of 05:22, 14 November 2016

Date People Last Week This Week
2016/10/31 Yang Feng
  • added new features to rnng+mn, including automatically detecting wrong sentences, swapping memories more frequently and filtering memory units to speed up.
  • ran experiments for rnng+mn [report]
  • read the code of sequence-to-sequence with tensorflow
  • recruited interns
  • Huilan work summary
  • optimize rnng+MN;
  • discuss the code with Jiyuan;
  • work with Andy at NMT;
  • Intern interviews
  • Huilan work.
Jiyuan Zhang
  • checked previous code about encoder-memory
  • completed code about decoder-memory,running
  • continue to modify memory model
  • read some related papers
Andi Zhang
  • ran NMT (cs-en) on gpu, but bleu is low, could be resulting from a small corpus
  • ran NMT on paraphrase data set
  • wrote MemN2N ducument
  • run NMT (fr-en) to get a bleu as that in the paper
  • run paraphrase for validation
Shiyue Zhang
  • try rnng on GPU
  • read the code of Feng
  • modify model report
  • try MKL
  • modify model