-
Notifications
You must be signed in to change notification settings - Fork 25.2k
Semantic Text Chunking Indexing Pressure #125517
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Semantic Text Chunking Indexing Pressure #125517
Conversation
…emitted to XContent
x-pack/plugin/inference/src/main/java/org/elasticsearch/xpack/inference/InferenceException.java
Outdated
Show resolved
Hide resolved
@elasticmachine update branch |
@elasticmachine update branch |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM, thanks for iterating
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Thanks @Mikep86
Hi @Mikep86, I've created a changelog YAML for you. |
Pinging @elastic/es-distributed-indexing (Team:Distributed Indexing) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks, @Mikep86.
Let’s update the PR title and summary to reflect the use of indexing pressure.
It would be great if someone from @elastic/es-distributed-indexing could review this section.
💔 Backport failed
You can use sqren/backport to manually backport by running |
💚 All backports created successfully
Questions ?Please refer to the Backport tool documentation |
We have observed many OOMs due to the memory required to inject chunked inference results for semantic_text fields. This PR uses coordinating indexing pressure to account for this memory usage. When indexing pressure memory usage exceeds the threshold set by indexing_pressure.memory.limit, chunked inference result injection will be suspended to prevent OOMs. (cherry picked from commit 85713f7) # Conflicts: # server/src/main/java/org/elasticsearch/node/NodeConstruction.java # server/src/main/java/org/elasticsearch/node/PluginServiceInstances.java # x-pack/plugin/inference/src/main/java/org/elasticsearch/xpack/inference/InferencePlugin.java # x-pack/plugin/inference/src/test/java/org/elasticsearch/xpack/inference/action/filter/ShardBulkInferenceActionFilterTests.java
* Semantic Text Chunking Indexing Pressure (#125517) We have observed many OOMs due to the memory required to inject chunked inference results for semantic_text fields. This PR uses coordinating indexing pressure to account for this memory usage. When indexing pressure memory usage exceeds the threshold set by indexing_pressure.memory.limit, chunked inference result injection will be suspended to prevent OOMs. (cherry picked from commit 85713f7) # Conflicts: # server/src/main/java/org/elasticsearch/node/NodeConstruction.java # server/src/main/java/org/elasticsearch/node/PluginServiceInstances.java # x-pack/plugin/inference/src/main/java/org/elasticsearch/xpack/inference/InferencePlugin.java # x-pack/plugin/inference/src/test/java/org/elasticsearch/xpack/inference/action/filter/ShardBulkInferenceActionFilterTests.java * [CI] Auto commit changes from spotless --------- Co-authored-by: elasticsearchmachine <[email protected]>
We have observed many OOMs due to the memory required to inject chunked inference results for
semantic_text
fields. This PR uses coordinating indexing pressure to account for this memory usage. When indexing pressure memory usage exceeds the threshold set byindexing_pressure.memory.limit
, chunked inference result injection will be suspended to prevent OOMs.