could not load model (no success): Unexpected err=ModuleNotFoundError(\"No module named 'optimum'\"), type(err)=<class 'ModuleNotFoundError' #3191
Unanswered
noblerboy2004
asked this question in
Q&A
Replies: 2 comments 1 reply
-
works here with the following container image:
and:
|
Beta Was this translation helpful? Give feedback.
0 replies
-
ah - got it sorry. You are trying to use the |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
LocalAI version:
Docker 3.6 Running on Docker-Container
Environment, CPU architecture, OS, and Version:
Linux 162fd9400319 5.15.153.1-microsoft-standard-WSL2 #1 SMP Fri Mar 29 23:14:13 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
Describe the bug
I'm trying to use https://huggingface.co/intfloat/multilingual-e5-base for embedding. I install model from localhost:8080. When running command $ curl http://localhost:8080/v1/embeddings -H "Content-Type: application/json" -d '{"input": "Test","model": "openvino-all-MiniLM-L6-v2"}'
Output will be: Administrator@mainserver MINGW64 /d/Flowise
$ curl http://localhost:8080/v1/embeddings -H "Content-Type: application/json" -d '{"input": "Test","model": "openvino-all-MiniLM-L6-v2"}'
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 250 100 196 100 54 31 8 0:00:06 0:00:06 --:--:-- 49{"error":{"code":500,"message":"could not load model (no success): Unexpected err=ModuleNotFoundError("No module named 'optimum'"), type(err)=\u003cclass 'ModuleNotFoundError'\u003e","type":""}}
I tried to install pip install optimum, but the result is the same.
To Reproduce
Expected behavior
Logs
08-07 12:36:10 5:36AM DBG GRPC(all-MiniLM-L6-v2-127.0.0.1:38999): stderr Server started. Listening on: 127.0.0.1:38999
2024-08-07 12:36:11 5:36AM DBG GRPC Service Ready
2024-08-07 12:36:11 5:36AM DBG GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:} sizeCache:0 unknownFields:[] Model:all-MiniLM-L6-v2 ContextSize:8192 Seed:940775446 NBatch:512 F16Memory:false MLock:false MMap:true VocabOnly:false LowVRAM:false Embeddings:true NUMA:false NGPULayers:99999999 MainGPU: TensorSplit: Threads:32 LibrarySearchPath: RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0 ModelFile:/models/all-MiniLM-L6-v2 Device: UseTriton:false ModelBaseName: UseFastTokenizer:false PipelineType: SchedulerType: CUDA:false CFGScale:0 IMG2IMG:false CLIPModel: CLIPSubfolder: CLIPSkip:0 ControlNet: Tokenizer: LoraBase: LoraAdapter: LoraScale:0 NoMulMatQ:false DraftModel: AudioPath: Quantization: GPUMemoryUtilization:0 TrustRemoteCode:false EnforceEager:false SwapSpace:0 MaxModelLen:0 TensorParallelSize:0 MMProj: RopeScaling: YarnExtFactor:0 YarnAttnFactor:0 YarnBetaFast:0 YarnBetaSlow:0 Type:OVModelForFeatureExtraction FlashAttention:false NoKVOffload:false}
2024-08-07 12:36:11 5:36AM DBG GRPC(all-MiniLM-L6-v2-127.0.0.1:38999): stderr Error: No module named 'optimum'
2024-08-07 12:36:11 5:36AM ERR Server error error="could not load model (no success): Unexpected err=ModuleNotFoundError("No module named 'optimum'"), type(err)=<class 'ModuleNotFoundError'>" ip=172.18.0.1 latency=6.15637602s method=POST status=500 url=/v1/embeddings
2024-08-07 12:36:58 5:36AM INF Success ip=127.0.0.1 latency="86.705µs"
Additional context
Beta Was this translation helpful? Give feedback.
All reactions