Topic: long-context Goto Github
Some thing interesting about long-context
Some thing interesting about long-context
long-context,RAN: Recurrent Attention Networks for Long-text Modeling | Findings of ACL23
Organization: 4ai
Home Page: https://rannet.readthedocs.io/en/latest/index.html
long-context,[DEPRECIATED] Very fast, large music transformer with 8k sequence length, efficient heptabit MIDI notes encoding, true full MIDI instruments range, chords counters and outro tokens
User: asigalov61
Home Page: https://soundcloud.com/aleksandr-sigalov-61/sets/heptabit-music-transformer
long-context,LooGLE: Long Context Evaluation for Long-Context Language Models
Organization: bigai-nlco
long-context,Papers of Long Context Language Model
User: davendw49
long-context,PyTorch implementation of Infini-Transformer from "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" (https://arxiv.org/abs/2404.07143)
User: dingo-actual
long-context,Code and documents of LongLoRA and LongAlpaca (ICLR 2024 Oral)
Organization: dvlab-research
Home Page: http://arxiv.org/abs/2309.12307
long-context,Implementation of paper "LM-Infinite: Simple On-the-Fly Length Generalization for Large Language Models"
User: glaciohound
Home Page: https://arxiv.org/abs/2308.16137
long-context,TriForce: Lossless Acceleration of Long Sequence Generation with Hierarchical Speculative Decoding
Organization: infini-ai-lab
Home Page: https://infini-ai-lab.github.io/TriForce/
long-context,Official release of InternLM2 7B and 20B base and chat models. 200K context support
Organization: internlm
Home Page: https://internlm.intern-ai.org.cn/
long-context,The official repo for "LLoCo: Learning Long Contexts Offline"
User: jeffreysijuntan
Home Page: https://arxiv.org/pdf/2404.07979.pdf
long-context,Transformers with Arbitrarily Large Context
User: lhao499
long-context,Streamlined variant of Long-Range Arena with pinned dependencies, automated data downloads, and deterministic shuffling.
User: lucaslingle
long-context,My own attempt at a long context genomics model, leveraging recent advances in long context attention modeling (Flash Attention + other hierarchical methods)
User: lucidrains
long-context,Implementation of MEGABYTE, Predicting Million-byte Sequences with Multiscale Transformers, in Pytorch
User: lucidrains
long-context,Implementation of Perceiver AR, Deepmind's new long-context attention network based on Perceiver architecture, in Pytorch
User: lucidrains
long-context,Implementation of Recurrent Memory Transformer, Neurips 2022 paper, in Pytorch
User: lucidrains
long-context,Implementation of π Ring Attention, from Liu et al. at Berkeley AI, in Pytorch
User: lucidrains
long-context,needle in a haystack for LLMs
User: melvinebenezer
long-context,awesome llm plaza: daily tracking all sorts of awesome topics of llm, e.g. llm for coding, robotics, reasoning, multimod etc.
Organization: metame-ai
long-context,Counting-Stars (β )
User: nick7nlp
Home Page: https://arxiv.org/pdf/2403.11802.pdf
long-context,open-source code for paper: Retrieval Head Mechanistically Explains Long-Context Factuality
User: nightdessert
Home Page: https://arxiv.org/abs/2404.15574
long-context,Finetuning and evaluating LLMs to extract GHG emissions from PDF reports using RAG and grammar-based decoding.
User: nopperl
Home Page: https://huggingface.co/spaces/nopperl/emission-extractor
long-context,The official implementation of "Ada-LEval: Evaluating long-context LLMs with length-adaptable benchmarks"
Organization: open-compass
long-context,Codes for the paper "βBench: Extending Long Context Evaluation Beyond 100K Tokens": https://arxiv.org/abs/2402.13718
Organization: openbmb
long-context,The code of our paper "InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memory"
Organization: thunlp
long-context,"Found in the Middle: How Language Models Use Long Contexts Better via Plug-and-Play Positional Encoding" Zhenyu Zhang, Runjin Chen, Shiwei Liu, Zhewei Yao, Olatunji Ruwase, Beidi Chen, Xiaoxia Wu, Zhangyang Wang.
Organization: vita-group
long-context,LongQLoRA: Extent Context Length of LLMs Efficiently
User: yangjianxin1
A declarative, efficient, and flexible JavaScript library for building user interfaces.
π Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. πππ
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google β€οΈ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.