Interface EmbeddingModelConfig


public interface EmbeddingModelConfig
  • Field Details

  • Method Details

    • inferenceEndpointUrl

      @WithDefault("https://api-inference.huggingface.co/pipeline/feature-extraction/sentence-transformers/all-MiniLM-L6-v2") URL inferenceEndpointUrl()
      The URL of the inference endpoint for the embedding.

      When using Hugging Face with the inference API, the URL is https://api-inference.huggingface.co/pipeline/feature-extraction/<model-id>, for example https://api-inference.huggingface.co/pipeline/feature-extraction/sentence-transformers/all-mpnet-base-v2.

      When using a deployed inference endpoint, the URL is the URL of the endpoint. When using a local hugging face model, the URL is the URL of the local model.

    • waitForModel

      @WithDefault("true") Boolean waitForModel()
      If the model is not ready, wait for it instead of receiving 503. It limits the number of requests required to get your inference done. It is advised to only set this flag to true after receiving a 503 error as it will limit hanging in your application to known places