Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FEATURE] Avoid expensive S3 listing by using file list in skipping index #218

Open
dai-chen opened this issue Jan 10, 2024 · 0 comments
Open
Labels
DataSource:File enhancement New feature or request

Comments

@dai-chen
Copy link
Collaborator

Is your feature request related to a problem?

Currently, any batch or streaming job on Spark table incurs expensive S3 listing to generate the input file list. Hive table has its own partition information in catalog but requires MSCK refresh this information manually and on a regular basis.

What solution would you like?

Actually only the S3 listing in skipping index is inevitable. Any direct query or streaming refreshing in covering index and materialized view doesn't need to do this again (unless strong consistency with latest source file is required).

In this case, the source file list seen so far can be found in file path column in Flint skipping index. The challenge is just to figure out if we can reconstruct FileStatus for Spark.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
DataSource:File enhancement New feature or request
Development

No branches or pull requests

1 participant