-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Question: best way to use sqlx with connections to two different databases? #121
Comments
We don't currently have a solution to this for the query macros beyond recommending to use multiple crates so the compilation units would be different. I know @abonander has some ideas how to make this nicer so I'm marking this as an enhancement. |
A workaround is to use cargo workspace: create one crate for MySql, one for Postgres, and third that imports them. So you can have different |
Copying over my thoughts from #224 as I realized we didn't put these up here. We were thinking a decent approach would be to generate a set of macros with a higher level macro.. e.g., mod one { sqlx::database!(env!("ONE_DATABASE_URL")); }
mod two { sqlx::database!(env!("TWO_DATABASE_URL")); }
one::query!("...") // mysql://...
two::query!("...") // postgres://... |
@mehcode To be honest, I don't like the idea of using macro here. Because it seems to increase both compile time and build size. Besides, I think a flexible and extensible approach is using SQL comment. e.g. sqlx::query!(
r#"
-- DATABASE_URL=postgres://...
select * from foo
"#
) will specifies the sqlx::query!(
r#"
-- DATABASE=bar
select * from foo
"#
) will tells sqlx to use |
I ran into this issue as well as I wanted separate postgres databases for a microserverish approach (I've worked around it for now using schemas). Using a comment, apart from being pretty verbose, feels too magical to me. Being able to customise or generate the |
My 0.02. I'd rather just have something like sqlx::query!([sql string], [optional db url or env var], [params]). If you leave it off then it uses the current behavior of DATABASE_URL env var, but if you supply the optional env var it will use that instead or you could just put a const string there and no env vars would be used. so perhaps:
Not sure if that would cause the macro to be too complicated, but something like that would be ideal for me. I can easily see that the query is being directed from a place different then the default fairly easily and hopefully doesn't add too much computational size or overhead. |
The idea with the Have a look back at @mehcode's example usage; the idea is that by invoking the macro in a submodule you can namespace the query macros so you know at a glance which DB it's talking to: mod one { sqlx::database!(env!("ONE_DATABASE_URL")); }
mod two { sqlx::database!(env!("TWO_DATABASE_URL")); }
one::query!("...") // mysql://...
two::query!("...") // postgres://... It doesn't make sense to me to hardcode the database URL at the use site; you'd have to remember to include the URL or the env var every time you write another In fact, I'd like to refactor the proc macro itself so that we don't have 4 separate proc macro definitions which all do about the same thing. Instead, I'm thinking the facade will pass in key-value pairs that dictate its behavior. |
Another thought is the |
Like setting It should probably be called |
I suggest using sql comment simply because it's extendable. It's is definitely better If we can have a config!(
DATABASE_URL = "...",
// ... more
) |
The comment would be significantly more complex to implement because we would have to parse the SQL in a comment-aware manner, which probably means rolling our own parser. We currently treat the string as opaque and I'd like to keep it that way unless we have a really good reason. |
If we did have a compelling reason to implement per-invocation config, I think I'd prefer it as attributes inside the invocation, which we can parse with sqlx:;query!(
#[db(env = "OTHER_ENV_VAR")]
"select ..."
); However, I think the |
I can see the benefits of a config macro in a module and then using that module to indicate which config you were trying to use. I just want it to be obvious and hopefully hard to screw up which database you are using for the query. I am waaayyy to good and making those mistakes :p |
I'm feeling like |
I think to use multiple crates is better. |
After thinking on this more, I'm not saying I don't want to do something with a meta-macro but multiple crates definitely feels like the cleanest solution to me. With #267, separate |
Suppose we have an application to transfer data from PostgreSQL to MySQL, using multiple crates isn't helpful, especial when I use query! to compile or run. I prefer mod one { sqlx::database!(env!("ONE_DATABASE_URL")); }
mod two { sqlx::database!(env!("TWO_DATABASE_URL")); }
one::query!("...") // mysql://...
two::query!("...") // postgres://... |
The idea with multiple crates is you'd have the query methods inside each crate. So if you want to query both at the same time that'd be a minimum of three crates. Yeah, I can see how that's not ideal but it is clean. The macros are going through an overhaul with the offline stuff but afterwords I'm sure we can revisit this. I think the only issue here is what to call |
I'm taking this feature into account with the offline refactor, this will probably be the next thing I tackle. |
Ideas for names (I don't mind a longer, more descriptive name as this should be used at most a handful of times in a given project):
|
I'm running into the similar issues here. My application is essentially a bridge between not 2 but many more databases. Has any progress been made on this since may? I understand the multiple crates solution is available, but that is a little hard to scale and maintain for a use case with many different database connections. I understand that multiple connections can be supported if you don't use compile-time checked queries, but unfortunately the sqlx-cli tool (which I really love) also depends on the hardcoded DATABASE_URL. |
I made a patch to allow specifying which env var is used in use like so: diff --git a/sqlx-macros/src/query/input.rs b/sqlx-macros/src/query/input.rs
index ddcd6d4a..138f3e75 100644
--- a/sqlx-macros/src/query/input.rs
+++ b/sqlx-macros/src/query/input.rs
@@ -8,6 +8,8 @@ use syn::{ExprArray, Type};
/// Macro input shared by `query!()` and `query_file!()`
pub struct QueryMacroInput {
+ pub(super) env_var: Option<String>,
+
pub(super) src: String,
#[cfg_attr(not(feature = "offline"), allow(dead_code))]
@@ -36,6 +38,7 @@ impl Parse for QueryMacroInput {
let mut args: Option<Vec<Expr>> = None;
let mut record_type = RecordType::Generated;
let mut checked = true;
+ let mut env_var = None;
let mut expect_comma = false;
@@ -48,7 +51,9 @@ impl Parse for QueryMacroInput {
let _ = input.parse::<syn::token::Eq>()?;
- if key == "source" {
+ if key == "env_var" {
+ env_var = Some(input.parse::<LitStr>()?.value());
+ } else if key == "source" {
let span = input.span();
let query_str = Punctuated::<LitStr, Token![+]>::parse_separated_nonempty(input)?
.iter()
@@ -80,6 +85,7 @@ impl Parse for QueryMacroInput {
let arg_exprs = args.unwrap_or_default();
Ok(QueryMacroInput {
+ env_var,
src: src.resolve(src_span)?,
src_span,
record_type,
diff --git a/sqlx-macros/src/query/mod.rs b/sqlx-macros/src/query/mod.rs
index 3ccab313..1ddfd24a 100644
--- a/sqlx-macros/src/query/mod.rs
+++ b/sqlx-macros/src/query/mod.rs
@@ -34,27 +34,33 @@ pub fn expand_input(input: QueryMacroInput) -> crate::Result<TokenStream> {
.map_err(|e| format!("failed to load environment from {:?}, {}", env_path, e))?
}
+ let db_var = input
+ .env_var
+ .as_ref()
+ .map_or_else(|| "DATABASE_URL".to_string(), |s| s.to_string());
+
// if `dotenv` wasn't initialized by the above we make sure to do it here
match (
dotenv::var("SQLX_OFFLINE")
.map(|s| s.to_lowercase() == "true")
.unwrap_or(false),
- dotenv::var("DATABASE_URL"),
+ dotenv::var(&db_var),
) {
(false, Ok(db_url)) => expand_from_db(input, &db_url),
#[cfg(feature = "offline")]
- _ => {
+ (_, db_url) => {
let data_file_path = std::path::Path::new(&manifest_dir).join("sqlx-data.json");
if data_file_path.exists() {
expand_from_file(input, data_file_path)
} else {
- Err(
- "`DATABASE_URL` must be set, or `cargo sqlx prepare` must have been run \
- and sqlx-data.json must exist, to use query macros"
- .into(),
+ Err(format!(
+ "`{}` must be set, or `cargo sqlx prepare` must have been run \
+ and sqlx-data.json must exist, to use query macros",
+ &db_var
)
+ .into())
}
}
@@ -64,7 +70,7 @@ pub fn expand_input(input: QueryMacroInput) -> crate::Result<TokenStream> {
}
#[cfg(not(feature = "offline"))]
- (false, Err(_)) => Err("`DATABASE_URL` must be set to use query macros".into()),
+ (false, Err(_)) => Err(format!("`{}` must be set to use query macros", &db_var,).into()),
}
}
diff --git a/src/macros.rs b/src/macros.rs
index 876e8fcb..3262c641 100644
--- a/src/macros.rs
+++ b/src/macros.rs
@@ -512,6 +512,12 @@ macro_rules! query_file_unchecked (
#[macro_export]
#[cfg_attr(docsrs, doc(cfg(feature = "macros")))]
macro_rules! query_as (
+ ($var:expr, $out_struct:path, $query:expr) => ( {
+ $crate::sqlx_macros::expand_query!(env_var = $var, record = $out_struct, source = $query)
+ });
+ ($var:expr, $out_struct:path, $query:expr, $($args:tt)*) => ( {
+ $crate::sqlx_macros::expand_query!(env_var = $var, record = $out_struct, source = $query, args = [$($args)*])
+ });
($out_struct:path, $query:expr) => ( {
$crate::sqlx_macros::expand_query!(record = $out_struct, source = $query)
}); |
In the case of the workaround by splitting into independent workspace crates, each with its .env; How do you handle the case for offline building, for example in CI/CD?. I've so far tried to generate the sqlx-data.json in each workspace independently, to no avail. It picks up the DATABASE_URL env var from one of the crates and checks some different one with it, failing while checking (mix of pgsql and mysql in my case). The "--merge" option does not change any of this, the problem being the environment variables handling. Suggestions? Opened #1223 |
What's the best solution to this in 2023? For example, I'd like to query |
Just chiming in that I'd also like to see better multi-db support. Our project uses many SQLite databases, all created at runtime, so one-crate-per-DB is not even possible. We only have 7 or so distinct schemas, so we could conceivably have separate crates for each of those, with an empty db for each just to have the right schema for compiletime verification, but even so that feels heavy handed, and we also can't make use of migrations against the actual databases (I haven't explored sqlx migrations enough to know if it could even support migrations for runtime-created dbs though). |
Is it possible set EDIT: yes, this fn main() {
// Create a new temporary file. We need to `.keep()` it so that it isn't
// deleted when it goes out of scope at the end of this function.
let (_, tempfile) = tempfile::NamedTempFile::new().unwrap().keep().unwrap();
let database_path = tempfile.as_path().to_str().unwrap();
let schema_file = "../schema.sql";
// Ensure that the build script is rerun if schema.sql changes
println!("cargo:rerun-if-changed={schema_file}");
// Execute SQL commands from schema.sql on the temporary database
rusqlite::Connection::open(database_path)
.unwrap()
.execute_batch(&std::fs::read_to_string(schema_file).unwrap())
.unwrap();
// Prepare the DATABASE_URL in the format that sqlx expects for SQLite
println!("cargo:rustc-env=DATABASE_URL=sqlite:{database_path}");
} |
Obviously, life changes so this didn't happen yet. However, this would be really helpful for me (as long as #[sqlx::test] works with it). |
Is it not possible to just make it so you could run Or perhaps you allow specifying multiple database-urls and then it tries each database in succession for each query until one succeeds, then uses that to save the query data. That solution wouldn't need any modification of the current macros, I believe. |
@Victor-N-Suadicani I think the main problem with that is: when / how are stale files in |
Wouldn't specifying multiple database-urls in a single |
commenting on this because it's an issue for me as well |
To add a new idea that I haven't seen in this thread before: Maybe we could use That way it would be easy to support multiple types of DBs at once (like this issue originally targeted) without introducing any change to the lib consumers and it would be very consistent to use. Only downside is, that this wouldn't support multiple DBs of the same type (e.g. two postgres DBs). The two DB types is a very common issue for me, since I'd like to provide some of my software with support for multiple DBs and also use compile time checked queries and right now that's not easy to do. |
As you might have seen, I started an implementation in #3278. Sadly I'm kind of stuck there, because the If the query family of macros had an additional parameter to determine the type of query (could be optional for cases where multiple DB types are used), it would be easy to resolve this, but that would mean changing the API (although it wouldn't really be breaking, if it's optional). I implemented this by passing in the |
@Snapstromegon Oh neat! That's awesome. I've got a question or two, but I'll leave it on your PR :) As for macros, when I was starting my macro journey, this article was tremendous for me. I'm not sure where you are on your journey but the testing parts specifically helped me. |
Thanks @esmevane, I'll give it a read. Right now I'm working through Jon Gjengset's Crust of Rust streams around macros and that's ~10h of videos to watch. |
For my usecase I want to have multiple crates using different (sqlite) databases in a single workspace. Using multiple I had the idea to set the // my-workspace/some-crate/build.rs
fn main() {
println!(
"cargo::rustc-env=DATABASE_URL=sqlite:{}/schema.db",
std::env::var("CARGO_MANIFEST_DIR").unwrap(),
);
} This way I can be sure every crate always uses the right database file. Edit: Improved build script that runs build.rsuse std::{
path::PathBuf,
process::Command,
};
fn main() {
let crate_dir = PathBuf::from(std::env::var("CARGO_MANIFEST_DIR").unwrap());
let out_dir = PathBuf::from(std::env::var("OUT_DIR").unwrap());
let migrations_dir = crate_dir.join("migrations");
let db_file = out_dir.join("schema.db");
let db_url = format!("sqlite:{}", db_file.display());
println!("cargo::rerun-if-changed={}", migrations_dir.display());
if db_file.exists() {
std::fs::remove_file(&db_file).unwrap();
}
std::fs::write(&db_file, b"").unwrap();
let exit_status = Command::new("sqlx")
.arg("migrate")
.arg("run")
.arg("--source")
.arg(&migrations_dir)
.arg("--database-url")
.arg(&db_url)
.spawn()
.unwrap()
.wait()
.unwrap();
if !exit_status.success() {
panic!("sqlx failed: {exit_status}");
}
println!("cargo::rustc-env=DATABASE_URL={}", db_url,);
} |
So excited for that PR @Aderinom !! We were just about to invest in writing that ourselves. Much appreciated! |
Our intended solution for this (and many different use-cases besides) is to allow multiple crates in the same workspace to rename their I'm working on it as fast as I can, and hope to have it ready for a short alpha cycle for |
It's so good to see it's happening! |
Would love to see this asap also! Is there any chance of an ETA @abonander? |
I have a program with two databases. One MySQL and one Postgres where I am syncing data from one to the other. Is that possible with sqlx? What is the best way to do that since the query macro has a hardcoded DATABASE_URL env var to determine the database information?
The text was updated successfully, but these errors were encountered: