Comments (6)
This may be a template and not a model - perhaps llm -t auto
Not sure what the YAML would look like.
A model might be better through, since then you could combine a template with the -m auto
option.
from llm.
I think it's a special model called -m auto
How should it handle some users not having GPT-4 32k access?
I think it should try anyway and error if they don't have the model - it would have errored anyway since they were over 32k tokens.
from llm.
Also need to consider the 3.5 4k vs 16k, guessing this is going to be a pattern that continues as well; models that are "the same" but differ in context length (and pricing).
I think there needs to be some concept of "flavors" of models, and in llm
you should be able to select the base "flavor" you want and have the model be selected based on a number of other factors (including context length).
from llm.
Worth noting this is a problem that other tools are facing right now as well. I'm not aware of any consensus on how to handle it as of yet, but it's probably worth looking into.
from llm.
this is relevant to the new -c
flag as well, as a conversation that fits in the context of one model may outgrow it, and ideally you can continue the conversation without interuption.
from llm.
Dropped from the 0.5 milestone, it's not critical for that.
I'm actually thinking this might make more sense as a llm-auto
plugin. It could be expanded to cover all kinds of other heuristics, not just the length of the context.
from llm.
Related Issues (20)
- Mechanism for recording a different model ID from the one requested HOT 1
- How to handle fake messages that were not part of real coversations? HOT 3
- llm-llamafile is missing from plugin directory
- How to cut off the LLM in chat mode
- IndexError on Windows for llm chat HOT 1
- llm-groq does not support llama 3 HOT 2
- some plugins fail to install with "Connection refused" error HOT 1
- UI around chat history HOT 1
- [plugin] add IBM watsonx
- A rapidly convert Files to Prompts Using Rust
- Enhancement idea: implement a self help
- Asynchronous API support HOT 1
- Add API documentation on how to import and use this tool as a Python library HOT 1
- Support for GPT-4o HOT 3
- Fix for latest mypy
- Rename the gpt-4-turbo aliases HOT 1
- All I ever get is "insufficient_quota" HOT 5
- llm 0.14: Can't run <<llm chat>> on Windows 11 HOT 1
- llm keys set openai
- Would you be up for a PR that shows help on no options?
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from llm.