1. add id2label, label2id to config.json.
  2. remove new. prefix from the key of the weight.
2024-11-17T10:45:42.615968Z  INFO text_embeddings_router: router/src/main.rs:175: Args { model_id: "./gte-************-********-*ase", revision: None, tokenization_workers: None, dtype: Some(Float16), pooling: Some(Cls), max_concurrent_requests: 512, max_batch_tokens: 16384, max_batch_requests: None, max_client_batch_size: 32, auto_truncate: false, default_prompt_name: None, default_prompt: None, hf_api_token: None, hostname: "d58481309b13", port: 12345, uds_path: "/tmp/text-embeddings-inference-server", huggingface_hub_cache: None, payload_limit: 2000000, api_key: None, json_output: false, otlp_endpoint: None, otlp_service_name: "text-embeddings-inference.server", cors_allow_origin: None }
2024-11-17T10:45:42.616155Z  WARN text_embeddings_router: router/src/lib.rs:377: `--pooling` arg is set but model is a classifier. Ignoring `--pooling` arg.
2024-11-17T10:45:43.253591Z  WARN text_embeddings_router: router/src/lib.rs:202: Could not find a Sentence Transformers config
2024-11-17T10:45:43.253612Z  INFO text_embeddings_router: router/src/lib.rs:206: Maximum number of tokens per request: 8192
2024-11-17T10:45:43.253626Z  INFO text_embeddings_core::tokenization: core/src/tokenization.rs:28: Starting 4 tokenization workers
2024-11-17T10:45:44.723069Z  INFO text_embeddings_router: router/src/lib.rs:248: Starting model backend
2024-11-17T10:45:44.930135Z  INFO text_embeddings_backend_candle: backends/candle/src/lib.rs:354: Starting FlashGTE model on Cuda(CudaDevice(DeviceId(1)))
2024-11-17T10:45:54.536773Z  INFO text_embeddings_router: router/src/lib.rs:264: Warming up model
2024-11-17T10:45:55.200022Z  WARN text_embeddings_router: router/src/lib.rs:326: Invalid hostname, defaulting to 0.0.0.0
2024-11-17T10:45:55.202256Z  INFO text_embeddings_router::http::server: router/src/http/server.rs:1812: Starting HTTP server: 0.0.0.0:12345
2024-11-17T10:45:55.202273Z  INFO text_embeddings_router::http::server: router/src/http/server.rs:1813: Ready
kozistr changed pull request status to open
thenlper changed pull request status to merged

Sign up or log in to comment