Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Spark RSR docs for FilOz review #6

Draft
wants to merge 4 commits into
base: biglep/spark-rsr-docs-filoz-feedback-attempt-2
Choose a base branch
from

Conversation

BigLep
Copy link
Member

@BigLep BigLep commented Nov 9, 2024

Putting Spark RSR docs in a PR so easier to collect questions/suggestions from FilOz during their 2024-11-10 review before Space Meridian meeting.

FilOz: please add suggestions, ask contextual comments, etc.

BigLep and others added 3 commits November 5, 2024 19:45
* Added initial spark retrieval success rate definiiton.
This from a export from https://www.notion.so/protocollabs/Spark-Request-Based-Non-Committee-Global-Retrieval-Success-Rate-4c5e8c47c45f467f80392d00cac2aae4#1a5fc6a081904c88b6d9ac7752e652a8
I haven't fixed the anchor links yet.

* Fixed the relative links.
This was done with a script that was added as well.

* Removed mangled markdown tables with a a request for a todo to give clean HTML

* Added a versions and support/questions/feedback section

* start porting result code table

* add result status codes

* fix rendering

* fill 3rd column

* more result code texts

* finish table

* empty lines in tables break formatting

* ` -> <code />

* more formatting fixes

* Finish Per Request (non-committee) Score vs. Committee Scoring

* Update service-level-indicators/spark-retrieval-success-rate.md

Co-authored-by: Julian Gruber <[email protected]>

* Update service-level-indicators/spark-retrieval-success-rate.md

Co-authored-by: Julian Gruber <[email protected]>

* Update service-level-indicators/spark-retrieval-success-rate.md

Co-authored-by: Julian Gruber <[email protected]>

* Update service-level-indicators/spark-retrieval-success-rate.md

Co-authored-by: Julian Gruber <[email protected]>

* Update service-level-indicators/spark-retrieval-success-rate.md

Co-authored-by: Julian Gruber <[email protected]>

* Update service-level-indicators/spark-retrieval-success-rate.md

Co-authored-by: Julian Gruber <[email protected]>

* Update service-level-indicators/spark-retrieval-success-rate.md

Co-authored-by: Julian Gruber <[email protected]>

* Update service-level-indicators/spark-retrieval-success-rate.md

Co-authored-by: Julian Gruber <[email protected]>

* Update service-level-indicators/spark-retrieval-success-rate.md

Co-authored-by: Julian Gruber <[email protected]>

* Small table formating adjustments

* Removed bolded headings

* Apply suggestions from code review

Co-authored-by: Julian Gruber <[email protected]>

---------

Co-authored-by: Julian Gruber <[email protected]>
Co-authored-by: patrickwoodhead <[email protected]>
* Added initial readme.
This includes:
* Project background
* Service class definition and listing
* Example service classes started for "warm" and "cold"
* SLO definition
* SLI definition and listing (referencing docs created in other PRs)
* Project tenets
* project conventions
* Start of an FAQ

* Typo / small cleanups

* badge cleanup

* more badge cleanup
* Added an initial sector health rate SLI definition

* Moved to use propertly formula formatting

* Typo fixes

* More fixiups
- Created by X on a daily basis as part of Y
- Consumed by anyone wanting to analyze or display SP retrievability performance.

## Callouts/Concerns with this SLI
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

2024-11-10 sync conversation notes:

  • Currently we are checking weighted on data stored. This means small-storing SPs won't get checked.
    • @Stebalien had the idea of blending a weighting of how much data an SP has and the time since have last checked an SP.
  • Idea of making it a two round thing
  • Idea of having some known failures cases where you expect nodes to fail as a way to catch "dishonest actors"

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Let's discuss the ideas in space-meridian/roadmap#190

Copy link
Member Author

@BigLep BigLep left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

2024-11-10 conversation between FilOz and Space Meridian in Bangkok:

  • PayloadCID
    • Label field should have the CID of the CAR that is stored.
    • Spark wants to treat "payload" as the data that a user wants to store.
    • Note: need to make it really clear on what an SP should do. Document the conventions.
    • Really want to either
      • Make deals with CARv2
      • Deal should have the index of the data stored
  • IPNI needs to do some sort of sticky cache to keep results for at least 20 minutes
  • Is it needed in IPNI to get pieceCids to CIDs
  • Spark needs to do more investigation on how to ensure spark checks are indiscerable from real requests. Right now a smart SP can see what are the Spark checks and then just be on their best behavior for the expected spark checks.
  • Masih's video on CommPact: https://www.youtube.com/watch?v=ndLKjXQ1awI&list=PL_0VrY55uV19y0b03a-CXNuvlgJR3G0EG&index=15
  • Steven idea of doing IPLD checks and pieceCid/range checks
  • 🎬 Schedule recurring Spark RSR working group

@juliangruber
Copy link
Member

Right now a smart SP can see what are the Spark checks and then just be on their best behavior for the expected spark checks.

How could a SP do that?

@BigLep
Copy link
Member Author

BigLep commented Nov 18, 2024

Right now a smart SP can see what are the Spark checks and then just be on their best behavior for the expected spark checks.

How could a SP do that?

@juliangruber : my understanding is that an SP can look at the Round Retrieval Task List, see which rows they are the listed "minerId" for, and then just make sure to serve the corresponding payloadCids in that 20 minute period.

BigLep added a commit that referenced this pull request Nov 18, 2024
Incorporated some feedback from #6
These are mostly spellcheck / formatting updates.
The only substantive change is the "callout/concern with this SLI" about how "checker traffic is pretty easy to differentiate from real traffic for a Storage Provider"
* Minor update to Spark RSR v1 docs
Incorporated some feedback from #6
These are mostly spellcheck / formatting updates.
The only substantive change is the "callout/concern with this SLI" about how "checker traffic is pretty easy to differentiate from real traffic for a Storage Provider"

* Update service-level-indicators/spark-retrieval-success-rate.md

Co-authored-by: Julian Gruber <[email protected]>

---------

Co-authored-by: Julian Gruber <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants