Comments (2)
To address the discrepancy in answers between Langchain-Chatchat and direct model inference, consider the following steps:
-
Model Configuration: Ensure both Langchain-Chatchat and your direct inference setup use identical model configurations, including parameters and version.
-
Input Processing: Verify that input data preprocessing, such as tokenization and normalization, is consistent across both setups. Langchain-Chatchat has specific preprocessing steps, including handling message history and constructing prompts, which should be mirrored in your direct inference process.
-
Inference Engine Settings: Align the settings of the underlying inference engine (e.g., PyTorch, TensorFlow) between both setups to ensure consistency.
-
Model Versions: Update both Langchain-Chatchat and your direct inference setup to the latest model versions if discrepancies exist.
-
External Dependencies: Ensure that all external dependencies, such as language model libraries, are the same version in both environments.
-
Knowledge Base: If Langchain-Chatchat uses a local knowledge base, make sure it's also used in your direct inference setup.
-
Debugging and Logs: Review logs from both setups for any errors or warnings that could explain the output differences.
-
Documentation and Community: Consult the official documentation and community forums for advice on achieving consistent inference results.
For detailed guidance, refer to Langchain-Chatchat's documentation and source code, especially sections related to chat processing, configuration examples, and knowledge base integration. This approach will help you identify and resolve the differences in answers between the two setups.
Details
Sources
- [BUG] 使用微调模型进行对话异常,其他工具调用正常
- knowledge_base/samples/content/test_files/langchain-ChatGLM_open.jsonl
- knowledge_base/samples/content/test_files/langchain-ChatGLM_closed.jsonl
- server/chat/chat.py
- configs/kb_config.py.example
- server/chat/search_engine_chat.py
- README_en.md
- README_ja.md
- README.md
- Dockerfile
- .gitmodules
- requirements_lite.txt
- server/api.py
- server/chat/knowledge_base_chat.py
- server/agent/tools/search_knowledgebase_once.py
- configs/basic_config.py.example
- startup.py
- server/agent/tools/search_knowledgebase_complex.py
About Dosu
This response is meant to be useful and save you time. It is not meant to be a precise solution, but rather a starting point for your own research.
Help me be more useful! Please leave a 👍 if this is helpful and 👎 if it is irrelevant.
To continue the conversation, mention @dosu.
from langchain-chatchat.
我也发现这个问题了
from langchain-chatchat.
Related Issues (20)
- cannot import name 'legacy' from 'llama_index'
- [BUG]httpcore.RemoteProtocolError: peer closed connection without sending complete message body (incomplete chunked read) HOT 3
- 配置好vllm配置之后启动startup.py时候遇到一个问题,似乎是缺少一个参数配置,具体报错如下
- 运行后卡柱不动
- 在初始化过程中,显存爆了 HOT 1
- 知识库匹配需要原文回答,而不是润色的内容
- 这个项目中的fastapi 怎么在vscode 调试时 重载代码? uvicorn --reload时报错
- 知识库问答调用add_message_to_db 无法保存response
- UnstructuredMarkdownLoader会丢失Markdown的结构标识
- 怎么使用splite
- [FEATURE] 搜索引擎支持百度api
- 请问怎么看最后传给大模型的历史对话和用户询问,我目前只跟踪到以下部分。
- http://127.0.0.1:7861/knowledge_base/kb_summary_api/summary_file_to_vector_store 进行文件总结 响应7分钟后出现一下报错,是超时了吗,这个逻辑是在哪啊,
- 上传文件不支持json格式的吗,似乎上传后解析不了
- 问答的第二个问题就会卡住中断
- cannot import name 'PDFResourceManag er' from 'pdfminer.converter'这是什么错误[BUG] 简洁阐述问题 / Concise description of the issue HOT 1
- 如何同时接入多个封装好的openai api格式大模型 HOT 2
- 请问怎么计算输出的token数量?这个框架有现成支持的方法吗
- 调用讯飞星火api,没有输出回答 HOT 2
- [BUG] 知识库问答匹配不到/ Concise description of the issue HOT 4
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from langchain-chatchat.