-
Notifications
You must be signed in to change notification settings - Fork 13
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Out of storage error when logging custom model to model registry #106
Comments
Hi Mugl3, Thank you for reporting. The error message is actually reporting the storage in the Snowflake Virtual Warehouse is running out disk space when loading the model. You could switch to use Snowpark Optimized Warehouse which has a larger disk to load your model. We are also working on this to make it possible to load a larger model in Standard Warehouse in the next few releases. |
Hi Wzhao, Are you able to confirm the max disk space by warehouse type and size? Thanks |
Hi Mugl3, Typically standard warehouse has several hundreds MBs temporary storage and Snowpark Optimized Warehouse has multiple GBs temporary storage. The disk space issue is happening when loading the model to the temporary storage and then loading into the RAM of the warehouse. As I mentioned above, we have noticed this issue and are currently working to loading it directly without using the temporary storage. For your report on "I have tried an XL Snowpark WH with the same error as before. It's only a small model that easily fits in RAM.", could you please share the query ID, so that we could further investigate. Thank you! |
When I went to get the query I found out the sample code replaced the Snowpark warehouse with a std one. Have updated the code and it no longer fails with the same error. I am trying to deploy a base sentence-transformers embedding model. I have included my code to ease troubleshooting:
Now define model class
Create custom model context
Test the model loads & outputs
Infer signature
Log the model
When logging the model I receive this error, Query ID Is it an issue with how I define conda dependencies in last code block? I also tried it like this |
Hi Mugl3, Sorry for the late reply. I believe the issue is because the version is not aligned in your local environment vs Snowflake environment. You could specify the version you want when specifying conda dependencies like Also our library recently adds out-of-box support for sentence transformer models recently, thus you could try log the model directly without using the custom model. |
Closing this given lack of activity since the suggestion to pin transformer or use native transformer model support. Please feel free to re-open if you have any other questions. |
I am following the guide to load a finetuned embeddings model into the registry using Sentence-Transformers.
I have successfully vectorized the code such that the model can be registered if needed.
When I get to the stage of actually uploading the local files to the Snowflake account I receive an error:
The device I am operating on has over 800gb available. The model is only 265 MBs.
The below is the SQL command showing in Snowflake UI that failed
CREATE MODEL MRCM_HOL_DB.MRCM_HOL_SCHEMA.MY_STCUSTOM_MODEL WITH VERSION VERSION_1 FROM @MRCM_HOL_DB.MRCM_HOL_SCHEMA.SNOWPARK_TEMP_STAGE_IZQPCZQXTY/model
Query ID: 01b56116-3202-ab08-0000-0001d77282d9
This was running on venv py3.8 with Snowflake trial account for testing. Please let me know if you need more logs.
The text was updated successfully, but these errors were encountered: