-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Why is it not working? #2
Comments
现在它可以正常运行,我不知道是为什么。 |
另外一个问题是,stable-diffusion每次都是迅速响应,flux每次要加载很久,这是因为在加载模型吗?每次post请求,都需要重新加载flux模型吗? |
时间都花费在了加载模型上 |
If no new log messages show on the screen, please check if the port you're using is in use or not.
The model is loaded ONLY at the stage of context initialization, and ONLY loaded once. Could you please share with us the following environment info?
Thanks a lot! |
感谢您的回复! |
Thanks for the feedback. We'll check the issue ASAP. Thanks! |
Thanks a lot. |
After checking the design of wasmedge-stablediffusion, the context should remain after loading. |
The models of the stable-diffusion series will not be dropped, while those of the flux series will be dropped. |
@hydai According to the investigation, |
Thank you! I'm looking forward to it very very much! |
@MMMazart Thanks for your quick feedback!
You have to do directory mappings as the running environment is an wasm sandbox. That's why you can see wasmedge --dir .:/Users/sam/workspace/demo/sd/dev sd-api-server.wasm \
--model-name flux1-dev \
--diffusion-model flux1-dev-Q4_0.gguf \
--vae ae.safetensors \
--clip-l clip_l.safetensors \
--t5xxl t5xxl-Q8_0.gguf
Yeah, the major target of
Could you please provide more details about the issue, such as request you used, CPU/GPU, memory/vram, and etc. That would help us reproduce the issue. In addition, our
I have no idea about |
My environment information is the same as mentioned before and has not been changed. This problem occurs every time. |
@MMMazart Do you mind sharing with us the prompt you're using? BTW, the issue is triggered while using |
prompt:"a lovely cat holding a sign says 'flux.cpp'". Yes, both flux.1-dev and flux.1-schnell will trigger this problem. |
"a cat" will trigger it, too. This seems to have nothing to do with the prompt. |
@MMMazart Could you share with us the request? For example, steps. |
headers = { data = { } This is my request, which is the same as the example |
|
After first inference is completed, it can be seen that the memory is released. So, the second request directly results in an error. |
@MMMazart Which version of CUDA are you using? |
@MMMazart We don't have A100, so we tried to reproduce the issue in the environment of |
I see that in your snapshot, it shows that only one request has been sent? It will crash on the second request. Can you send multiple requests? In my environment, after the first request, the context will be deleted. Thanks! |
我把cuda版本换成了12.2,在第一次request后,context还是会被清除。I changed the CUDA version to 12.2. After the first request, the context will still be cleared. My Ubuntu version is 22.04, but it seems that the biggest difference is on the GPU. |
I don't think it has anything to do with machines, GPU etc. I am getting the very same behaviour on a Macbook M3 Pro 48GB of shared RAM. At the second request the server crashes:
I followed the steps for the FLUX example. Server runs with:
The client request:
The second time I execute this request the server crashes. |
Thanks for reporting, @fabiopolimeni and @MMMazart . Will release a new version to solve this problem. See the upstream issue. WasmEdge/WasmEdge#3803 |
Hi @fabiopolimeni and @MMMazart |
|
My CUDA version is 11.5, but it seems to be unsupported. I switched to version 12.2, which works. |
It's weird. This error shows the address it tried to bind is in use. And it's not related to the cuda version. Could you check if you run the cuda-11 version when there are no other applications using the same address/port? |
This is indeed strange, but I was using the same port before and after. It worked after changing the CUDA version. |
./wasmedge --dir .:. sd-api-server.wasm --model-name sd-v1.4 --model /mnt/data/zhangmingyang/t2i/models/stable-diffusion-v-1-4-GGUF/stable-diffusion-v1-4-Q8_0.gguf I executed this command, and the result is as follows, but when I send the request 'curl -X POST 'http://localhost:8080/v1/images/generations' --header 'Content-Type: application/json' --data '{"model": "sd-v1.4", "prompt": "A cute baby sea otter"}'', there is no response. What is going on?
The text was updated successfully, but these errors were encountered: