Skip to content

Out of date huggingface-embeddings backend #1783

@TwinFinz

Description

@TwinFinz

LocalAI version:
LocalAI version: c8be839 (c8be839)

Environment, CPU architecture, OS, and Version:
Linux LocalAi 6.5.11-8-pve #1 SMP PREEMPT_DYNAMIC PMX 6.5.11-8 (2024-01-30T12:27Z) x86_64 x86_64 x86_64 GNU/Linux

Describe the bug
Out of date huggingface-embeddings backend cause model to not load/function.

To Reproduce
Use the following yaml to attempt to use "nomic-embed" model.

yaml

name: nomic-embed
backend: huggingface-embeddings
embeddings: true
parameters:
  model: nomic-ai/nomic-embed-text-v1

Expected behavior
Generate and return of embeddings.

Logs
7:52AM DBG Loading Model nomic-ai/nomic-embed-text-v1 with gRPC (file: /build/models/nomic-ai/nomic-embed-text-v1) (backend: huggingface-embeddings): {backendString:huggingface-embeddings model:nomic-ai/nomic-embed-text-v1 threads:4 assetDir:/tmp/localai/backend_data context:{emptyCtx:{}} gRPCOptions:0xc000338000 externalBackends:map[autogptq:/build/backend/python/autogptq/run.sh bark:/build/backend/python/bark/run.sh coqui:/build/backend/python/coqui/run.sh diffusers:/build/backend/python/diffusers/run.sh exllama:/build/backend/python/exllama/run.sh exllama2:/build/backend/python/exllama2/run.sh huggingface-embeddings:/build/backend/python/sentencetransformers/run.sh mamba:/build/backend/python/mamba/run.sh petals:/build/backend/python/petals/run.sh sentencetransformers:/build/backend/python/sentencetransformers/run.sh transformers:/build/backend/python/transformers/run.sh transformers-musicgen:/build/backend/python/transformers-musicgen/run.sh vall-e-x:/build/backend/python/vall-e-x/run.sh vllm:/build/backend/python/vllm/run.sh] grpcAttempts:20 grpcAttemptsDelay:2 singleActiveBackend:true parallelRequests:true}
7:52AM DBG Loading external backend: /build/backend/python/sentencetransformers/run.sh
7:52AM DBG Loading GRPC Process: /build/backend/python/sentencetransformers/run.sh
7:52AM DBG GRPC Service for nomic-ai/nomic-embed-text-v1 will be running at: '127.0.0.1:44505'
7:52AM DBG GRPC Service state dir: /tmp/go-processmanager1812659671
7:52AM DBG GRPC Service Started
7:52AM DBG GRPC(nomic-ai/nomic-embed-text-v1-127.0.0.1:44505): stderr Server started. Listening on: 127.0.0.1:44505
7:52AM DBG GRPC Service Ready
7:52AM DBG GRPC: Loading model with options: {state:{NoUnkeyedLiterals:{} DoNotCompare:[] DoNotCopy:[] atomicMessageInfo:} sizeCache:0 unknownFields:[] Model:nomic-ai/nomic-embed-text-v1 ContextSize:0 Seed:0 NBatch:512 F16Memory:false MLock:false MMap:false VocabOnly:false LowVRAM:false Embeddings:true NUMA:false NGPULayers:0 MainGPU: TensorSplit: Threads:4 LibrarySearchPath: RopeFreqBase:0 RopeFreqScale:0 RMSNormEps:0 NGQA:0 ModelFile:/build/models/nomic-ai/nomic-embed-text-v1 Device: UseTriton:false ModelBaseName: UseFastTokenizer:false PipelineType: SchedulerType: CUDA:false CFGScale:0 IMG2IMG:false CLIPModel: CLIPSubfolder: CLIPSkip:0 ControlNet: Tokenizer: LoraBase: LoraAdapter: LoraScale:0 NoMulMatQ:false DraftModel: AudioPath: Quantization: MMProj: RopeScaling: YarnExtFactor:0 YarnAttnFactor:0 YarnBetaFast:0 YarnBetaSlow:0 Type:}
7:52AM DBG [WatchDog] Watchdog checks for idle connections
7:52AM DBG GRPC(nomic-ai/nomic-embed-text-v1-127.0.0.1:44505): stderr You try to use a model that was created with version 2.4.0.dev0, however, your version is 2.2.2. This might cause unexpected behavior or errors. In that case, try to update to the latest version.
7:52AM DBG GRPC(nomic-ai/nomic-embed-text-v1-127.0.0.1:44505): stderr
7:52AM DBG GRPC(nomic-ai/nomic-embed-text-v1-127.0.0.1:44505): stderr
7:52AM DBG GRPC(nomic-ai/nomic-embed-text-v1-127.0.0.1:44505): stderr
[192.168.0.106]:7512 500 - POST /embeddings

Additional context
Looks like something just needs to be updated. Should be pretty simple.

Metadata

Metadata

Assignees

No one assigned

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions