Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ollama 加载 glm-4-9b-chat 胡言乱语 #521

Open
2 tasks
siegrainwong opened this issue Aug 30, 2024 · 2 comments
Open
2 tasks

ollama 加载 glm-4-9b-chat 胡言乱语 #521

siegrainwong opened this issue Aug 30, 2024 · 2 comments
Assignees

Comments

@siegrainwong
Copy link

siegrainwong commented Aug 30, 2024

System Info / 系統信息

cuda: 12.6
transformer: 4.44.0
OS: win10
python: 3.11.4
ollama: 0.3.8 & 0.2.3
配置: RTX3090 12700kf

Who can help? / 谁可以帮助到您?

No response

Information / 问题信息

  • The official example scripts / 官方的示例脚本
  • My own modified scripts / 我自己修改的脚本和任务

Reproduction / 复现过程

  1. download gguf model from https://www.modelscope.cn/models/llm-research/glm-4-9b-chat-gguf/files
  2. ollama create xxx
  3. ollama serve & open open-webui

只要我不点停就会一直写下去,没在别的model上发现过这种情况(gemma2-7b\ yi-9b),根据以往记录下了0.2.3的ollama但响应差不多
image

Expected behavior / 期待表现

跑原模型时挺正常
image

@zhipuch
Copy link
Collaborator

zhipuch commented Aug 30, 2024

#323
#333

@zhipuch zhipuch self-assigned this Aug 30, 2024
@siegrainwong
Copy link
Author

开过flash attention,不起作用
image
image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants