Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FEATURE] Improve filtering condition push down in Flint data source #148

Open
dai-chen opened this issue Nov 10, 2023 · 1 comment
Open
Labels

Comments

@dai-chen
Copy link
Collaborator

Is your feature request related to a problem?

Currently there are only limited support for push down optimization in Flint data source. For example, value set column in skipping index is actually array and should use ARRAY_CONTAINS in filtering condition. Because of no push down support, index query has to use = instead.

What solution would you like?

  1. Support push down for more operators in filtering condition
  2. [TBD] Support array field by field metadata
@dai-chen
Copy link
Collaborator Author

This requirement is part of integrating OpenSearch as a data source in Spark, which includes:

  1. OpenSearch data types
  2. OpenSearch scalar and aggregate functions
  3. OpenSearch DSL pushdown capabilities

One approach is to leverage the OpenSearch SQL plugin by integrating OpenSearch as a JDBC connector via the OpenSearch SQL JDBC driver. If this proves to be a viable long-term solution, future enhancements, such as high performance communication via Apache Arrow, can be encapsulated and managed within the JDBC driver, ensuring a centralized and efficient implementation.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

1 participant