-
Notifications
You must be signed in to change notification settings - Fork 8.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ELSER / interface endpoint installation and management #192461
Comments
Pinging @elastic/appex-ai-infra (Team:AI Infra) |
From #188554:
Apparently, today the kibana system user does not have the Note that we don't strictly need to be able to manage any inference endpoint, but we will need the Kibana user to at least be able to create (and then manage) the ELSER-based inference endpoints we're planning to use for the KB indices. |
Here's some prior art from when we tried to enable it a few months back: Gist was we would let the search team drive the need and work with Kibana Security on enabling. I think with the requirements we have now we can drive the need now too 🙂 And tangentially, since I haven't seen this mentioned, I just wanted to note that some of the Search UI's like within Search Connectors already support some flows where they install ELSER via an Inference Endpoint. This has some interesting interactions with how the assistants currently deploy ELSER via the edit: Just ran into this again today setting up the Slack Connector If you click Which results in a second ELSER deployment showing up in the ML Trained Models UI: |
FYI, we're re-attempting to add the |
We (observability AI assistant) have received requests about supporting other languages. Will this work provide support for E5 as well or is it ELSER only? |
elastic/elasticsearch#114051 allowed the Kibana user to manage any kind of inference endpoint, so installing an E5 endpoint can be done too. Note that this issue may be superseded by the default elastic endpoint feature that ES is currently working on, and we might be able to leverage that feature without any additional work on our side. |
This is a pre-requisite for the unified knowledge base.
Being able to use semantic search within a cluster has some prerequisites: it requires an inference endpoint deployed on the cluster, and deploying an inference endpoint requires having at least one ML node present on the cluster. Deploying an inference endpoint, even with ELSER, isn’t necessarily trivial in some scenarios (e.g air gap mode).
For internal usages of semantic search (such as retrieval from the knowledge bases), our solution teams are using our internal ELSER model. However, for this, they’re forced to set up the inference endpoint, and to manage its lifecycle.
The text was updated successfully, but these errors were encountered: