Skip to content

Conversation

@jinmang2
Copy link
Owner

@jinmang2 jinmang2 commented Apr 12, 2022

목적

Efficient Transformer family를 fairseq + huggingface style로 개발한다.

참고할 Libraries

TODO

  • past key value
  • efficient transformers
  • dynamic padding
  • gradient checkpoint
  • various positional embedding
  • reversible residual connection
  • chunk feed forward network
  • 3d, 4d multi-head scaled dot product attention
  • various heads
  • generation mixin
  • parrallelism mixin
  • pushtohub mixin
  • how to make tokenization script?
  • trainer

후보

  • memorized transformer
  • compressive transformer
  • transformer-xl
  • retrieval enhanced transformer
  • LoRA
  • p-tuning
  • Bi-encoder, Cross-encoder, Poly-encoder, ColBERT

@jinmang2 jinmang2 self-assigned this Apr 12, 2022
@review-notebook-app
Copy link

Check out this pull request on  ReviewNB

See visual diffs & provide feedback on Jupyter Notebooks.


Powered by ReviewNB

@jinmang2 jinmang2 changed the title Project: Efficient Transformers Project: Advanced Transformers Apr 12, 2022
@jinmang2 jinmang2 force-pushed the project-efficient-transformer branch from 0bd8ef7 to 61911b2 Compare April 12, 2022 16:30
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants