chore(deps): update dependency mlflow to v2.15.1 - autoclosed #60
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR contains the following updates:
==2.1.1
->==2.15.1
Release Notes
mlflow/mlflow (mlflow)
v2.15.1
Compare Source
MLflow 2.15.1 is a patch release that addresses several bug fixes.
Bug fixes:
mlflow.evaluate
crash on binary classification with data subset only contains single class (#12825, @serena-ruan)Documentation updates:
Small bug fixes and documentation updates:
#12823, #12860, #12844, #12843, @B-Step62; #12863, #12828, @harupy; #12845, @djliden; #12820, @annzhang-db; #12831, @chenmoneygithub
v2.15.0
Compare Source
We are excited to announce the release candidate for MLflow 2.15.0. This release includes many major features and improvements!
Major features:
LlamaIndex Flavor🦙 - MLflow now offers a native integration with LlamaIndex, one of the most popular libraries for building GenAI apps centered around custom data. This integration allows you to log LlamaIndex indices within MLflow, allowing for the loading and deployment of your indexed data for inference tasks with different engine types. MLflow also provides comprehensive tracing support for LlamaIndex operations, offering unprecedented transparency into complex queries. Check out the MLflow LlamaIndex documentation to get started! (#12633, @michael-berk, @B-Step62)
OpenAI Tracing🔍 - We've enhanced our OpenAI integration with a new tracing feature that works seamlessly with MLflow OpenAI autologging. You can now enable tracing of their OpenAI API usage with a single
mlflow.openai.autolog()
call, thereby MLflow will automatically log valuable metadata such as token usage and a history of your interactions, providing deeper insights into your OpenAI-powered applications. To start exploring this new capability, please check out the tracing documentation! (#12267, @gabrielfu)Enhanced Model Deployment with New Validation Feature✅ - To improve the reliability of model deployments, MLflow has added a new method to validate your model before deploying it to an inference endpoint. This feature helps to eliminate typical errors in input and output handling, streamlining the process of model deployment and increasing confidence in your deployed models. By catching potential issues early, you can ensure a smoother transition from development to production. (#12710, @serena-ruan)
Custom Metrics Definition Recording for Evaluations📊 - We've strengthened the flexibility of defining custom metrics for model evaluation by automatically logging and versioning metrics definitions, including models used as judges and prompt templates. With this new capability, you can ensure reproducibility of evaluations across different runs and easily reuse evaluation setups for consistency, facilitating more meaningful comparisons between different models or versions. (#12487, #12509, @xq-yin)
Databricks SDK Integration🔐 - MLflow's interaction with Databricks endpoints has been fully migrated to use the Databricks SDK. This change brings more robust and reliable connections between MLflow and Databricks, and access to the latest Databricks features and capabilities. We mark the legacy databricks-cli support as deprecated and will remove in the future release. (#12313, @WeichenXu123)
Spark VectorUDT Support💥 - MLflow's Model Signature framework now supports Spark Vector UDT (User Defined Type), enabling logging and deployment of models using Spark VectorUDT with robust type validation. (#12758, @WeichenXu123)
Other Notable Changes
Features:
parent_id
as a parameter to thestart_run
fluent API for alternative control flows (#12721, @Flametaa)mlflow gc
(#12451, @M4nouel)ChatModel
interface for GenAI flavors (#12612, @WeichenXu123)iloc
for accessing rows (#12410, @julcsii)Bug fixes:
.batch
call due to thread unsafety (#12701, @B-Step62)log_model
issue in MLflow >= 2.13 that causes databricks DLT py4j service crashing (#12514, @WeichenXu123)predict_stream
forAgentExecutor
and other non-Runnable chains (#12518, @B-Step62)Documentation updates:
fork
vsspawn
method when using multiprocessing for parallel runs (#12337, @B-Step62)extract_fields
formlflow.search_traces
(#12319, @xq-yin)Small bug fixes and documentation updates:
#12727, #12709, #12685, #12667, #12673, #12602, #12601, #12655, #12641, #12635, #12634, #12584, #12428, #12388, #12352, #12298, #12750, #12727, #12757, @daniellok-db; #12726, #12733, #12691, #12622, #12579, #12581, #12285, #12311, #12357, #12339, #12338, #12705, #12797, #12787, #12784, #12771, #12737, @B-Step62; #12715, @hubertzub-db; #12722, #12804, @annzhang-db; #12676, #12680, #12665, #12664, #12671, #12651, #12649, #12647, #12637, #12632, #12603, #12343, #12328, #12286, #12793, #12770, @serena-ruan; #12670, #12613, #12473, #12506, #12485, #12477, #12468, #12464, #12443, #12807, #12800, #10874, #12761, @WeichenXu123; #12690, #12678, #12686, #12545, #12621, #12598, #12583, #12582, #12510, #12580, #12570, #12571, #12559, #12538, #12537, #12519, #12515, #12507, #12508, #12502, #12499, #12497, #12447, #12467, #12426, #12448, #12430, #12420, #12385, #12371, #12359, #12284, #12345, #12316, #12287, #12303, #12291, #12795, #12786, #12796, #12792, #12791, #12778, #12777, #12755, #12751, #12753, #12749, @harupy; #12742, #12702, #12742 @edwardfeng-db; #12605, @alxhslm; #12662, @freemso; #12577, @rafyzg; #12512, @Jaishree2310; #12491, #1274, @BenWilson2; #12549, @besarthoxhaj; #12476, @jessechancy; #12541, @amanjam; #12479, #12472, #12433, #12289, @xq-yin; #12486, #12474, #11406, @jgiannuzzi; #12463, @jsuchome; #12460, @Venki1402; #12449, @yukimori; #12318, @RistoAle97; #12440, @victolee0; #12416, @Dev-98; #11771, @lababidi; #12417, @dannikay; #12663, @cgilviadee; #12410, @julcsii; #12600, @ZTZK; #12803, @hcmturner; #12747, @michael-berk; #12342, @kriscon-db; #12766, @artjen;
v2.14.3
Compare Source
MLflow 2.14.3 is a patch release that addresses bug fixes and additional documentation for released features
Features:
Bug fixes:
Documentation updates:
Small bug fixes and documentation updates:
#12556, #12628, @B-Step62; #12582, #12560, @harupy; #12553, @nojaf
v2.14.2
Compare Source
MLflow 2.14.2 is a patch release that includes several important bug fixes and documentation enhancements.
Bug fixes:
llm/v1/xxx
task definitions. (#12551, @B-Step62)log_model
introduced in MLflow 2.13.0 that causes Databricks DLT service to crash in some situations (#12514, @WeichenXu123)predict_stream
implementation for LangChain AgentExecutor and other non-Runnable chains (#12518, @B-Step62)predict_proba
inference method in thesklearn
flavor when loading an sklearn pipeline object aspyfunc
(#12554, @WeichenXu123)Documentation updates:
Small bug fixes and documentation updates:
#12311, #12285, #12535, #12543, #12320, #12444, @B-Step62; #12310, #12340, @serena-ruan; #12409, #12432, #12471, #12497, #12499, @harupy; #12555, @nojaf; #12472, #12431, @xq-yin; #12530, #12529, #12528, #12527, #12526, #12524, #12531, #12523, #12525, #12522, @dbczumar; #12483, @jsuchome; #12465, #12441, @BenWilson2; #12450, @StarryZhang-whu
v2.14.1
Compare Source
MLflow 2.14.1 is a patch release that contains several bug fixes and documentation improvements
Bug fixes:
install_mlflow=False
(#12388, @daniellok-db)Documentation updates:
Small bug fixes and documentation updates:
#12415, #12396, #12394, @harupy; #12403, #12382, @BenWilson2; #12397, @B-Step62
v2.14.0
Compare Source
MLflow 2.14.0 includes several major features and improvements that we're very excited to announce!
Major features:
Other Notable Features:
Bug fixes:
Documentation updates:
Small bug fixes and documentation updates:
#12359, #12308, #12350, #12284, #12345, #12316, #12287, #12303, #12291, #12288, #12265, #12170, #12248, #12263, #12249, #12251, #12239, #12241, #12240, #12235, #12242, #12172, #12215, #12228, #12216, #12164, #12225, #12203, #12181, #12198, #12195, #12192, #12146, #12171, #12163, #12166, #12124, #12106, #12113, #12112, #12074, #12077, #12058, @harupy; #12355, #12326, #12114, #12343, #12328, #12327, #12340, #12286, #12310, #12200, #12209, #12189, #12194, #12201, #12196, #12174, #12107, @serena-ruan; #12364, #12352, #12354, #12353, #12351, #12298, #12297, #12220, #12155, @daniellok-db; #12311, #12357, #12346, #12312, #12339, #12281, #12283, #12282, #12268, #12236, #12247, #12199, #12232, #12233, #12221, #12229, #12207, #12212, #12193, #12167, #12137, #12147, #12148, #12138, #12127, #12065, @B-Step62; #12289, #12253, #12330 @xq-yin; #11771, @lababidi; #12280, #12275, @BenWilson2; #12246, #12244, #12211, #12066, #12061, @WeichenXu123; #12278, @sunishsheth2009; #12136, @kriscon-db; #11911, @jessechancy; #12169, @hubertzub-db
v2.13.2
Compare Source
MLflow 2.13.2 is a patch release that includes several bug fixes and integration improvements to existing
features.
Features:
urllib
's connection number and max size (#12227, @chenmoneygithub)Bug fixes:
mlflow[gateway]
as dependency when usingmlflow.deployment
module (#12264, @B-Step62)/
before logging as params (#12190, @sunishsheth2009)Small bug fixes and documentation updates:
#12268, #12210, @B-Step62; #12214, @harupy; #12223, #12226, @annzhang-db; #12260, #12237, @prithvikannan; #12261, @BenWilson2; #12231, @serena-ruan; #12238, @sunishsheth2009
v2.13.1
Compare Source
MLflow 2.13.1 is a patch release that includes several bug fixes and integration improvements to existing features. New features that are introduced in this patch release are intended to provide a foundation to further major features that will be released in the next release.
Features:
mlflow[langchain]
extra that installs recommended versions of langchain with MLflow (#12182, @sunishsheth2009)Bug fixes:
getUserLocalTempDir
andgetUserNFSTempDir
to replacegetReplLocalTempDir
andgetReplNFSTempDir
in databricks runtime (#12105, @WeichenXu123)load_context
when inferring signature in pyfunc (#12099, @sunishsheth2009)Small bug fixes and documentation updates:
#12180, #12152, #12128, #12126, #12100, #12086, #12084, #12079, #12071, #12067, #12062, @serena-ruan; #12175, #12167, #12137, #12134, #12127, #12123, #12111, #12109, #12078, #12080, #12064, @B-Step62; #12142, @2maz; #12171, #12168, #12159, #12153, #12144, #12104, #12095, #12083, @harupy; #12160, @aravind-segu; #11990, @kriscon-db; #12178, #12176, #12090, #12036, @sunishsheth2009; #12162, #12110, #12088, #11937, #12075, @daniellok-db; #12133, #12131, @prithvikannan; #12132, #12035, @annzhang-db; #12121, #12120, @liangz1; #12122, #12094, @dbczumar; #12098, #12055, @mparkhe
v2.13.0
Compare Source
MLflow 2.13.0 includes several major features and improvements
With this release, we're happy to introduce several features that enhance the usability of MLflow broadly across a range of use cases.
Major Features and Improvements:
Streamable Python Models: The newly introduced
predict_stream
API for Python Models allows for custom model implementations that support the return of a generator object, permitting full customization for GenAI applications.Enhanced Code Dependency Inference: A new feature for automatically inferrring code dependencies based on detected dependencies within a model's implementation. As a supplement to the
code_paths
parameter, the introducedinfer_model_code_paths
option when logging a model will determine which additional code modules are needed in order to ensure that your models can be loaded in isolation, deployed, and reliably stored.Standardization of MLflow Deployment Server: Outputs from the Deployment Server's endpoints now conform to OpenAI's interfaces to provide a simpler integration with commonly used services.
Features:
Togetherai
as a supported provider for the MLflow Deployments Server (#11557, @FotiosBistas)predict_stream
API support for Python Models (#11791, @WeichenXu123)Configuration
📅 Schedule: Branch creation - "on the 2nd and 4th day instance on sunday" (UTC), Automerge - At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR is behind base branch, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
This PR was generated by Mend Renovate. View the repository job log.