230 Forks
3024 Stars
3024 Watchers

flash-linear-attention

🚀 Efficient implementations of state-of-the-art linear attention models

How to download and setup flash-linear-attention

Open terminal and run command
git clone https://github.com/fla-org/flash-linear-attention.git
git clone is used to create a copy or clone of flash-linear-attention repositories. You pass git clone a repository URL.
it supports a few different network protocols and corresponding URL formats.

Also you may download zip file with flash-linear-attention https://github.com/fla-org/flash-linear-attention/archive/master.zip

Or simply clone flash-linear-attention with SSH
[email protected]:fla-org/flash-linear-attention.git

If you have some problems with flash-linear-attention

You may open issue on flash-linear-attention support forum (system) here: https://github.com/fla-org/flash-linear-attention/issues

Similar to flash-linear-attention repositories

Here you may see flash-linear-attention alternatives and analogs

 natural-language-processing    lectures    spaCy    HanLP    gensim    MatchZoo    tensorflow-nlp    Awesome-pytorch-list    spacy-models    Repo-2017    stanford-tensorflow-tutorials    awesome-nlp    nlp_tasks    nltk    pattern    TextBlob    CoreNLP    allennlp    mycroft-core    practical-pytorch    textract    languagetool    MITIE    machine_learning_examples    prose    arXivTimes    ltp    libpostal    sling    DeepNLP-models-Pytorch