From 7de68c01a61e8b9ddd79cd8352ee26c9682e0ffa Mon Sep 17 00:00:00 2001 From: Julian Date: Sat, 1 Feb 2025 15:14:00 +0100 Subject: [PATCH 01/16] use db settings --- crates/pg_configuration/src/database.rs | 14 ++---- crates/pg_workspace/src/settings.rs | 13 ++--- crates/pg_workspace/src/workspace/server.rs | 54 ++++++++++++++++----- 3 files changed, 51 insertions(+), 30 deletions(-) diff --git a/crates/pg_configuration/src/database.rs b/crates/pg_configuration/src/database.rs index 9b35ae61..e9cd1649 100644 --- a/crates/pg_configuration/src/database.rs +++ b/crates/pg_configuration/src/database.rs @@ -26,6 +26,10 @@ pub struct DatabaseConfiguration { /// The name of the database. #[partial(bpaf(long("database")))] pub database: String, + + /// The connection timeout in seconds. + #[partial(bpaf(long("conn_timeout")))] + pub conn_timeout: Option, } impl Default for DatabaseConfiguration { @@ -36,15 +40,7 @@ impl Default for DatabaseConfiguration { username: "postgres".to_string(), password: "postgres".to_string(), database: "postgres".to_string(), + conn_timeout: Some(10), } } } - -impl DatabaseConfiguration { - pub fn to_connection_string(&self) -> String { - format!( - "postgres://{}:{}@{}:{}/{}", - self.username, self.password, self.host, self.port, self.database - ) - } -} diff --git a/crates/pg_workspace/src/settings.rs b/crates/pg_workspace/src/settings.rs index 9f6fa661..6023d771 100644 --- a/crates/pg_workspace/src/settings.rs +++ b/crates/pg_workspace/src/settings.rs @@ -5,6 +5,7 @@ use std::{ num::NonZeroU64, path::{Path, PathBuf}, sync::{RwLock, RwLockReadGuard, RwLockWriteGuard}, + time::Duration, }; use ignore::gitignore::{Gitignore, GitignoreBuilder}; @@ -266,6 +267,7 @@ pub struct DatabaseSettings { pub username: String, pub password: String, pub database: String, + pub conn_timeout: Duration, } impl Default for DatabaseSettings { @@ -276,19 +278,11 @@ impl Default for DatabaseSettings { username: "postgres".to_string(), password: "postgres".to_string(), database: "postgres".to_string(), + conn_timeout: Duration::from_secs(10), } } } -impl DatabaseSettings { - pub fn to_connection_string(&self) -> String { - format!( - "postgres://{}:{}@{}:{}/{}", - self.username, self.password, self.host, self.port, self.database - ) - } -} - impl From for DatabaseSettings { fn from(value: PartialDatabaseConfiguration) -> Self { let d = DatabaseSettings::default(); @@ -298,6 +292,7 @@ impl From for DatabaseSettings { username: value.username.unwrap_or(d.username), password: value.password.unwrap_or(d.password), database: value.database.unwrap_or(d.database), + conn_timeout: value.conn_timeout.unwrap_or(d.conn_timeout), } } } diff --git a/crates/pg_workspace/src/workspace/server.rs b/crates/pg_workspace/src/workspace/server.rs index bf90c2c3..76401373 100644 --- a/crates/pg_workspace/src/workspace/server.rs +++ b/crates/pg_workspace/src/workspace/server.rs @@ -1,4 +1,7 @@ -use std::{fs, future::Future, panic::RefUnwindSafe, path::Path, sync::RwLock}; +use std::{ + fs, future::Future, panic::RefUnwindSafe, path::Path, str::FromStr, sync::RwLock, + time::Duration, +}; use analyser::AnalyserVisitorBuilder; use change::StatementChange; @@ -12,7 +15,11 @@ use pg_fs::{ConfigName, PgLspPath}; use pg_query::PgQueryStore; use pg_schema_cache::SchemaCache; use pg_typecheck::TypecheckParams; -use sqlx::PgPool; +use sqlx::{ + pool::PoolOptions, + postgres::{PgConnectOptions, PgPoolOptions}, + PgPool, +}; use std::sync::LazyLock; use tokio::runtime::Runtime; use tracing::info; @@ -20,7 +27,7 @@ use tree_sitter::TreeSitterStore; use crate::{ configuration::to_analyser_rules, - settings::{Settings, SettingsHandle, SettingsHandleMut}, + settings::{DatabaseSettings, Settings, SettingsHandle, SettingsHandleMut}, workspace::PullDiagnosticsResult, WorkspaceError, }; @@ -41,7 +48,6 @@ mod tree_sitter; #[derive(Default)] struct DbConnection { pool: Option, - connection_string: Option, } // Global Tokio Runtime @@ -53,16 +59,41 @@ impl DbConnection { self.pool.clone() } - pub(crate) fn set_connection(&mut self, connection_string: &str) -> Result<(), WorkspaceError> { - if self.connection_string.is_none() - || self.connection_string.as_ref().unwrap() != connection_string - { - self.connection_string = Some(connection_string.to_string()); - self.pool = Some(PgPool::connect_lazy(connection_string)?); + pub(crate) fn set_connection( + &mut self, + settings: &DatabaseSettings, + ) -> Result<(), WorkspaceError> { + if self.is_already_connected(settings) { + return Ok(()); } + let config = PgConnectOptions::new() + .host(&settings.host) + .port(settings.port) + .username(&settings.username) + .password(&settings.password) + .database(&settings.database); + + self.pool = Some( + PoolOptions::new() + .acquire_timeout(settings.conn_timeout.clone()) + .connect_lazy_with(config), + ); + Ok(()) } + + fn is_already_connected(&self, settings: &DatabaseSettings) -> bool { + self.pool + .as_ref() + .map(|p| p.connect_options()) + .is_some_and(|opts| { + opts.get_host() == settings.host + && opts.get_port() == settings.port + && opts.get_username() == settings.username + && opts.get_database() == Some(&settings.database) + }) + } } pub(super) struct WorkspaceServer { @@ -122,11 +153,10 @@ impl WorkspaceServer { fn refresh_db_connection(&self) -> Result<(), WorkspaceError> { let s = self.settings(); - let connection_string = s.as_ref().db.to_connection_string(); self.connection .write() .unwrap() - .set_connection(&connection_string)?; + .set_connection(&s.as_ref().db)?; self.reload_schema_cache()?; From e2a975f11a166a8d18e1b594014d9d5e4d55ed3e Mon Sep 17 00:00:00 2001 From: Julian Date: Sat, 1 Feb 2025 15:33:59 +0100 Subject: [PATCH 02/16] eager init --- crates/pg_configuration/src/database.rs | 4 +- crates/pg_workspace/src/workspace/server.rs | 45 +++++++++------------ 2 files changed, 20 insertions(+), 29 deletions(-) diff --git a/crates/pg_configuration/src/database.rs b/crates/pg_configuration/src/database.rs index e9cd1649..57752bc9 100644 --- a/crates/pg_configuration/src/database.rs +++ b/crates/pg_configuration/src/database.rs @@ -29,7 +29,7 @@ pub struct DatabaseConfiguration { /// The connection timeout in seconds. #[partial(bpaf(long("conn_timeout")))] - pub conn_timeout: Option, + pub conn_timeout_secs: Option, } impl Default for DatabaseConfiguration { @@ -40,7 +40,7 @@ impl Default for DatabaseConfiguration { username: "postgres".to_string(), password: "postgres".to_string(), database: "postgres".to_string(), - conn_timeout: Some(10), + conn_timeout_secs: Some(10), } } } diff --git a/crates/pg_workspace/src/workspace/server.rs b/crates/pg_workspace/src/workspace/server.rs index 76401373..5ffa0083 100644 --- a/crates/pg_workspace/src/workspace/server.rs +++ b/crates/pg_workspace/src/workspace/server.rs @@ -1,7 +1,4 @@ -use std::{ - fs, future::Future, panic::RefUnwindSafe, path::Path, str::FromStr, sync::RwLock, - time::Duration, -}; +use std::{fs, future::Future, panic::RefUnwindSafe, path::Path, sync::RwLock}; use analyser::AnalyserVisitorBuilder; use change::StatementChange; @@ -15,11 +12,7 @@ use pg_fs::{ConfigName, PgLspPath}; use pg_query::PgQueryStore; use pg_schema_cache::SchemaCache; use pg_typecheck::TypecheckParams; -use sqlx::{ - pool::PoolOptions, - postgres::{PgConnectOptions, PgPoolOptions}, - PgPool, -}; +use sqlx::{pool::PoolOptions, postgres::PgConnectOptions, PgPool, Postgres}; use std::sync::LazyLock; use tokio::runtime::Runtime; use tracing::info; @@ -59,11 +52,8 @@ impl DbConnection { self.pool.clone() } - pub(crate) fn set_connection( - &mut self, - settings: &DatabaseSettings, - ) -> Result<(), WorkspaceError> { - if self.is_already_connected(settings) { + pub(crate) fn connect(&mut self, settings: &DatabaseSettings) -> Result<(), WorkspaceError> { + if self.matches_current_connection(settings) { return Ok(()); } @@ -74,16 +64,21 @@ impl DbConnection { .password(&settings.password) .database(&settings.database); - self.pool = Some( - PoolOptions::new() - .acquire_timeout(settings.conn_timeout.clone()) - .connect_lazy_with(config), - ); + let timeout = settings.conn_timeout.clone(); + + let maybe_pool = run_async(async move { + PoolOptions::::new() + .acquire_timeout(timeout) + .connect_with(config) + .await + })?; + + self.pool = Some(maybe_pool?); Ok(()) } - fn is_already_connected(&self, settings: &DatabaseSettings) -> bool { + fn matches_current_connection(&self, settings: &DatabaseSettings) -> bool { self.pool .as_ref() .map(|p| p.connect_options()) @@ -153,11 +148,7 @@ impl WorkspaceServer { fn refresh_db_connection(&self) -> Result<(), WorkspaceError> { let s = self.settings(); - self.connection - .write() - .unwrap() - .set_connection(&s.as_ref().db)?; - + self.connection.write().unwrap().connect(&s.as_ref().db)?; self.reload_schema_cache()?; Ok(()) @@ -165,14 +156,14 @@ impl WorkspaceServer { fn reload_schema_cache(&self) -> Result<(), WorkspaceError> { tracing::info!("Reloading schema cache"); - // TODO return error if db connection is not available + if let Some(c) = self.connection.read().unwrap().get_pool() { let maybe_schema_cache = run_async(async move { SchemaCache::load(&c).await })?; let schema_cache = maybe_schema_cache?; - let mut cache = self.schema_cache.write().unwrap(); *cache = schema_cache; } else { + // if we can't get a connection, we'l reset the schema cache let mut cache = self.schema_cache.write().unwrap(); *cache = SchemaCache::default(); } From afc93e4b88ba73c28d3aa66d0fa20bbc989bffe3 Mon Sep 17 00:00:00 2001 From: Julian Date: Sat, 1 Feb 2025 15:39:18 +0100 Subject: [PATCH 03/16] . --- crates/pg_workspace/src/workspace/server.rs | 25 +++++++++++++++------ 1 file changed, 18 insertions(+), 7 deletions(-) diff --git a/crates/pg_workspace/src/workspace/server.rs b/crates/pg_workspace/src/workspace/server.rs index 5ffa0083..ddea112c 100644 --- a/crates/pg_workspace/src/workspace/server.rs +++ b/crates/pg_workspace/src/workspace/server.rs @@ -53,7 +53,7 @@ impl DbConnection { } pub(crate) fn connect(&mut self, settings: &DatabaseSettings) -> Result<(), WorkspaceError> { - if self.matches_current_connection(settings) { + if self.is_already_connected(settings) { return Ok(()); } @@ -78,7 +78,7 @@ impl DbConnection { Ok(()) } - fn matches_current_connection(&self, settings: &DatabaseSettings) -> bool { + fn is_already_connected(&self, settings: &DatabaseSettings) -> bool { self.pool .as_ref() .map(|p| p.connect_options()) @@ -148,10 +148,22 @@ impl WorkspaceServer { fn refresh_db_connection(&self) -> Result<(), WorkspaceError> { let s = self.settings(); - self.connection.write().unwrap().connect(&s.as_ref().db)?; - self.reload_schema_cache()?; + match self.connection.write().unwrap().connect(&s.as_ref().db) { + Ok(_) => { + self.reload_schema_cache()?; + Ok(()) + } + Err(e) => { + tracing::error!("Failed to connect to database: {:?}", e); + self.reset_schema_cache(); + Err(e) + } + } + } - Ok(()) + fn reset_schema_cache(&self) { + let mut cache = self.schema_cache.write().unwrap(); + *cache = SchemaCache::default(); } fn reload_schema_cache(&self) -> Result<(), WorkspaceError> { @@ -164,8 +176,7 @@ impl WorkspaceServer { *cache = schema_cache; } else { // if we can't get a connection, we'l reset the schema cache - let mut cache = self.schema_cache.write().unwrap(); - *cache = SchemaCache::default(); + self.reset_schema_cache(); } tracing::info!("Schema cache reloaded"); From 120625a9f9922f6b603cd0ae4c23218cdb41c5cd Mon Sep 17 00:00:00 2001 From: Julian Date: Mon, 3 Feb 2025 08:56:04 +0100 Subject: [PATCH 04/16] remove ::new() --- crates/pg_completions/src/context.rs | 14 +++++++------- crates/pg_configuration/src/lib.rs | 1 + crates/pg_schema_cache/src/schema_cache.rs | 4 ---- crates/pg_workspace/src/settings.rs | 9 ++++++--- crates/pg_workspace/src/workspace/server.rs | 2 +- 5 files changed, 15 insertions(+), 15 deletions(-) diff --git a/crates/pg_completions/src/context.rs b/crates/pg_completions/src/context.rs index a5fb0c6b..d3ff1052 100644 --- a/crates/pg_completions/src/context.rs +++ b/crates/pg_completions/src/context.rs @@ -266,7 +266,7 @@ mod tests { position: (position as u32).into(), text, tree: Some(&tree), - schema: &pg_schema_cache::SchemaCache::new(), + schema: &pg_schema_cache::SchemaCache::default(), }; let ctx = CompletionContext::new(¶ms); @@ -298,7 +298,7 @@ mod tests { position: (position as u32).into(), text, tree: Some(&tree), - schema: &pg_schema_cache::SchemaCache::new(), + schema: &pg_schema_cache::SchemaCache::default(), }; let ctx = CompletionContext::new(¶ms); @@ -332,7 +332,7 @@ mod tests { position: (position as u32).into(), text, tree: Some(&tree), - schema: &pg_schema_cache::SchemaCache::new(), + schema: &pg_schema_cache::SchemaCache::default(), }; let ctx = CompletionContext::new(¶ms); @@ -357,7 +357,7 @@ mod tests { position: (position as u32).into(), text, tree: Some(&tree), - schema: &pg_schema_cache::SchemaCache::new(), + schema: &pg_schema_cache::SchemaCache::default(), }; let ctx = CompletionContext::new(¶ms); @@ -385,7 +385,7 @@ mod tests { position: (position as u32).into(), text, tree: Some(&tree), - schema: &pg_schema_cache::SchemaCache::new(), + schema: &pg_schema_cache::SchemaCache::default(), }; let ctx = CompletionContext::new(¶ms); @@ -411,7 +411,7 @@ mod tests { position: (position as u32).into(), text, tree: Some(&tree), - schema: &pg_schema_cache::SchemaCache::new(), + schema: &pg_schema_cache::SchemaCache::default(), }; let ctx = CompletionContext::new(¶ms); @@ -436,7 +436,7 @@ mod tests { position: (position as u32).into(), text, tree: Some(&tree), - schema: &pg_schema_cache::SchemaCache::new(), + schema: &pg_schema_cache::SchemaCache::default(), }; let ctx = CompletionContext::new(¶ms); diff --git a/crates/pg_configuration/src/lib.rs b/crates/pg_configuration/src/lib.rs index 9dd64f89..7190271d 100644 --- a/crates/pg_configuration/src/lib.rs +++ b/crates/pg_configuration/src/lib.rs @@ -104,6 +104,7 @@ impl PartialConfiguration { username: Some("postgres".to_string()), password: Some("postgres".to_string()), database: Some("postgres".to_string()), + conn_timeout_secs: Some(10), }), } } diff --git a/crates/pg_schema_cache/src/schema_cache.rs b/crates/pg_schema_cache/src/schema_cache.rs index 8d73e631..77a0526a 100644 --- a/crates/pg_schema_cache/src/schema_cache.rs +++ b/crates/pg_schema_cache/src/schema_cache.rs @@ -18,10 +18,6 @@ pub struct SchemaCache { } impl SchemaCache { - pub fn new() -> SchemaCache { - SchemaCache::default() - } - pub async fn load(pool: &PgPool) -> Result { let (schemas, tables, functions, types, versions, columns) = futures_util::try_join!( Schema::load(pool), diff --git a/crates/pg_workspace/src/settings.rs b/crates/pg_workspace/src/settings.rs index 6023d771..5d8f9acd 100644 --- a/crates/pg_workspace/src/settings.rs +++ b/crates/pg_workspace/src/settings.rs @@ -267,7 +267,7 @@ pub struct DatabaseSettings { pub username: String, pub password: String, pub database: String, - pub conn_timeout: Duration, + pub conn_timeout_secs: Duration, } impl Default for DatabaseSettings { @@ -278,7 +278,7 @@ impl Default for DatabaseSettings { username: "postgres".to_string(), password: "postgres".to_string(), database: "postgres".to_string(), - conn_timeout: Duration::from_secs(10), + conn_timeout_secs: Duration::from_secs(10), } } } @@ -292,7 +292,10 @@ impl From for DatabaseSettings { username: value.username.unwrap_or(d.username), password: value.password.unwrap_or(d.password), database: value.database.unwrap_or(d.database), - conn_timeout: value.conn_timeout.unwrap_or(d.conn_timeout), + conn_timeout_secs: value + .conn_timeout_secs + .map(|s| Duration::from_secs(s.into())) + .unwrap_or(d.conn_timeout_secs), } } } diff --git a/crates/pg_workspace/src/workspace/server.rs b/crates/pg_workspace/src/workspace/server.rs index ddea112c..9ce25332 100644 --- a/crates/pg_workspace/src/workspace/server.rs +++ b/crates/pg_workspace/src/workspace/server.rs @@ -64,7 +64,7 @@ impl DbConnection { .password(&settings.password) .database(&settings.database); - let timeout = settings.conn_timeout.clone(); + let timeout = settings.conn_timeout_secs.clone(); let maybe_pool = run_async(async move { PoolOptions::::new() From 34eef56fbe42594797a3ae0826b256fc63372701 Mon Sep 17 00:00:00 2001 From: Julian Date: Mon, 3 Feb 2025 10:20:06 +0100 Subject: [PATCH 05/16] fix(workspace)?: dont hang on db connection errors --- crates/pg_lsp/src/handlers/completions.rs | 16 +- crates/pg_schema_cache/src/schema_cache.rs | 19 ++ crates/pg_workspace/src/workspace.rs | 3 - crates/pg_workspace/src/workspace/client.rs | 4 - crates/pg_workspace/src/workspace/server.rs | 232 +++++------------- .../src/workspace/server/async_helper.rs | 21 ++ .../src/workspace/server/db_connection.rs | 34 +++ .../workspace/server/schema_cache_manager.rs | 54 ++++ pglsp.toml | 2 +- 9 files changed, 206 insertions(+), 179 deletions(-) create mode 100644 crates/pg_workspace/src/workspace/server/async_helper.rs create mode 100644 crates/pg_workspace/src/workspace/server/db_connection.rs create mode 100644 crates/pg_workspace/src/workspace/server/schema_cache_manager.rs diff --git a/crates/pg_lsp/src/handlers/completions.rs b/crates/pg_lsp/src/handlers/completions.rs index f13526cd..83ade9f4 100644 --- a/crates/pg_lsp/src/handlers/completions.rs +++ b/crates/pg_lsp/src/handlers/completions.rs @@ -1,6 +1,6 @@ use crate::session::Session; use anyhow::Result; -use pg_workspace::workspace; +use pg_workspace::{workspace, WorkspaceError}; use tower_lsp::lsp_types::{self, CompletionItem, CompletionItemLabelDetails}; #[tracing::instrument(level = "trace", skip_all)] @@ -26,12 +26,22 @@ pub fn get_completions( pg_lsp_converters::negotiated_encoding(client_capabilities), )?; - let completion_result = session + let completion_result = match session .workspace .get_completions(workspace::CompletionParams { path, position: offset, - })?; + }) { + Ok(result) => result, + Err(e) => match e { + WorkspaceError::DatabaseConnectionError(_) => { + return Ok(lsp_types::CompletionResponse::Array(vec![])); + } + _ => { + return Err(e.into()); + } + }, + }; let items: Vec = completion_result .into_iter() diff --git a/crates/pg_schema_cache/src/schema_cache.rs b/crates/pg_schema_cache/src/schema_cache.rs index 77a0526a..0fdb7f39 100644 --- a/crates/pg_schema_cache/src/schema_cache.rs +++ b/crates/pg_schema_cache/src/schema_cache.rs @@ -9,6 +9,8 @@ use crate::versions::Version; #[derive(Debug, Clone, Default)] pub struct SchemaCache { + cached_conn_str: String, + pub schemas: Vec, pub tables: Vec, pub functions: Vec, @@ -29,6 +31,7 @@ impl SchemaCache { )?; Ok(SchemaCache { + cached_conn_str: SchemaCache::pool_to_conn_str(pool), schemas, tables, functions, @@ -38,6 +41,22 @@ impl SchemaCache { }) } + fn pool_to_conn_str(pool: &PgPool) -> String { + let conn = pool.connect_options(); + + format!( + "postgres://{}:@{}:{}/{}", + conn.get_username(), + conn.get_host(), + conn.get_port(), + conn.get_database().unwrap_or("default") + ) + } + + pub fn has_already_cached_connection(&self, pool: &PgPool) -> bool { + self.cached_conn_str == SchemaCache::pool_to_conn_str(pool) + } + /// Applies an AST node to the repository /// /// For example, alter table add column will add the column to the table if it does not exist diff --git a/crates/pg_workspace/src/workspace.rs b/crates/pg_workspace/src/workspace.rs index cbfd3756..0dfb9f7c 100644 --- a/crates/pg_workspace/src/workspace.rs +++ b/crates/pg_workspace/src/workspace.rs @@ -119,9 +119,6 @@ pub trait Workspace: Send + Sync + RefUnwindSafe { params: CompletionParams, ) -> Result; - /// Refresh the schema cache for this workspace - fn refresh_schema_cache(&self) -> Result<(), WorkspaceError>; - /// Update the global settings for this workspace fn update_settings(&self, params: UpdateSettingsParams) -> Result<(), WorkspaceError>; diff --git a/crates/pg_workspace/src/workspace/client.rs b/crates/pg_workspace/src/workspace/client.rs index c5059be9..36e67363 100644 --- a/crates/pg_workspace/src/workspace/client.rs +++ b/crates/pg_workspace/src/workspace/client.rs @@ -117,10 +117,6 @@ where self.request("pglsp/get_file_content", params) } - fn refresh_schema_cache(&self) -> Result<(), WorkspaceError> { - self.request("pglsp/refresh_schema_cache", ()) - } - fn pull_diagnostics( &self, params: super::PullDiagnosticsParams, diff --git a/crates/pg_workspace/src/workspace/server.rs b/crates/pg_workspace/src/workspace/server.rs index 9ce25332..a94b29e9 100644 --- a/crates/pg_workspace/src/workspace/server.rs +++ b/crates/pg_workspace/src/workspace/server.rs @@ -1,8 +1,10 @@ -use std::{fs, future::Future, panic::RefUnwindSafe, path::Path, sync::RwLock}; +use std::{fs, panic::RefUnwindSafe, path::Path, sync::RwLock}; use analyser::AnalyserVisitorBuilder; +use async_helper::run_async; use change::StatementChange; use dashmap::{DashMap, DashSet}; +use db_connection::DbConnection; use document::{Document, Statement}; use futures::{stream, StreamExt}; use pg_analyse::{AnalyserOptions, AnalysisFilter}; @@ -10,17 +12,14 @@ use pg_analyser::{Analyser, AnalyserConfig, AnalyserContext}; use pg_diagnostics::{serde::Diagnostic as SDiagnostic, Diagnostic, DiagnosticExt, Severity}; use pg_fs::{ConfigName, PgLspPath}; use pg_query::PgQueryStore; -use pg_schema_cache::SchemaCache; use pg_typecheck::TypecheckParams; -use sqlx::{pool::PoolOptions, postgres::PgConnectOptions, PgPool, Postgres}; -use std::sync::LazyLock; -use tokio::runtime::Runtime; +use schema_cache_manager::SchemaCacheManager; use tracing::info; use tree_sitter::TreeSitterStore; use crate::{ configuration::to_analyser_rules, - settings::{DatabaseSettings, Settings, SettingsHandle, SettingsHandleMut}, + settings::{Settings, SettingsHandle, SettingsHandleMut}, workspace::PullDiagnosticsResult, WorkspaceError, }; @@ -31,72 +30,21 @@ use super::{ }; mod analyser; +mod async_helper; mod change; +mod db_connection; mod document; mod migration; mod pg_query; +mod schema_cache_manager; mod tree_sitter; -/// Simple helper to manage the db connection and the associated connection string -#[derive(Default)] -struct DbConnection { - pool: Option, -} - -// Global Tokio Runtime -static RUNTIME: LazyLock = - LazyLock::new(|| Runtime::new().expect("Failed to create Tokio runtime")); - -impl DbConnection { - pub(crate) fn get_pool(&self) -> Option { - self.pool.clone() - } - - pub(crate) fn connect(&mut self, settings: &DatabaseSettings) -> Result<(), WorkspaceError> { - if self.is_already_connected(settings) { - return Ok(()); - } - - let config = PgConnectOptions::new() - .host(&settings.host) - .port(settings.port) - .username(&settings.username) - .password(&settings.password) - .database(&settings.database); - - let timeout = settings.conn_timeout_secs.clone(); - - let maybe_pool = run_async(async move { - PoolOptions::::new() - .acquire_timeout(timeout) - .connect_with(config) - .await - })?; - - self.pool = Some(maybe_pool?); - - Ok(()) - } - - fn is_already_connected(&self, settings: &DatabaseSettings) -> bool { - self.pool - .as_ref() - .map(|p| p.connect_options()) - .is_some_and(|opts| { - opts.get_host() == settings.host - && opts.get_port() == settings.port - && opts.get_username() == settings.username - && opts.get_database() == Some(&settings.database) - }) - } -} - pub(super) struct WorkspaceServer { /// global settings object for this workspace settings: RwLock, /// Stores the schema cache for this workspace - schema_cache: RwLock, + schema_cache: SchemaCacheManager, /// Stores the document (text content + version number) associated with a URL documents: DashMap, @@ -131,7 +79,7 @@ impl WorkspaceServer { tree_sitter: TreeSitterStore::new(), pg_query: PgQueryStore::new(), changed_stmts: DashSet::default(), - schema_cache: RwLock::default(), + schema_cache: SchemaCacheManager::default(), connection: RwLock::default(), } } @@ -145,44 +93,6 @@ impl WorkspaceServer { SettingsHandleMut::new(&self.settings) } - fn refresh_db_connection(&self) -> Result<(), WorkspaceError> { - let s = self.settings(); - - match self.connection.write().unwrap().connect(&s.as_ref().db) { - Ok(_) => { - self.reload_schema_cache()?; - Ok(()) - } - Err(e) => { - tracing::error!("Failed to connect to database: {:?}", e); - self.reset_schema_cache(); - Err(e) - } - } - } - - fn reset_schema_cache(&self) { - let mut cache = self.schema_cache.write().unwrap(); - *cache = SchemaCache::default(); - } - - fn reload_schema_cache(&self) -> Result<(), WorkspaceError> { - tracing::info!("Reloading schema cache"); - - if let Some(c) = self.connection.read().unwrap().get_pool() { - let maybe_schema_cache = run_async(async move { SchemaCache::load(&c).await })?; - let schema_cache = maybe_schema_cache?; - let mut cache = self.schema_cache.write().unwrap(); - *cache = schema_cache; - } else { - // if we can't get a connection, we'l reset the schema cache - self.reset_schema_cache(); - } - tracing::info!("Schema cache reloaded"); - - Ok(()) - } - fn is_ignored_by_migration_config(&self, path: &Path) -> bool { let set = self.settings(); set.as_ref() @@ -233,11 +143,6 @@ impl WorkspaceServer { } impl Workspace for WorkspaceServer { - #[tracing::instrument(level = "trace", skip(self))] - fn refresh_schema_cache(&self) -> Result<(), WorkspaceError> { - self.reload_schema_cache() - } - /// Update the global settings for this workspace /// /// ## Panics @@ -254,10 +159,15 @@ impl Workspace for WorkspaceServer { params.gitignore_matches.as_slice(), )?; - self.refresh_db_connection()?; - tracing::info!("Updated settings in workspace"); + self.connection + .write() + .unwrap() + .set_conn_settings(&self.settings().as_ref().db); + + tracing::info!("Updated Db connection settings"); + Ok(()) } @@ -390,51 +300,51 @@ impl Workspace for WorkspaceServer { // run diagnostics for each statement in parallel if its mostly i/o work if let Ok(connection) = self.connection.read() { - if let Some(pool) = connection.get_pool() { - let typecheck_params: Vec<_> = doc - .iter_statements_with_text_and_range() - .map(|(stmt, range, text)| { - let ast = self.pg_query.get_ast(&stmt); - let tree = self.tree_sitter.get_parse_tree(&stmt); - (text.to_string(), ast, tree, *range) - }) - .collect(); - - let pool_clone = pool.clone(); - let path_clone = params.path.clone(); - let async_results = run_async(async move { - stream::iter(typecheck_params) - .map(|(text, ast, tree, range)| { - let pool = pool_clone.clone(); - let path = path_clone.clone(); - async move { - if let Some(ast) = ast { - pg_typecheck::check_sql(TypecheckParams { - conn: &pool, - sql: &text, - ast: &ast, - tree: tree.as_deref(), - }) - .await - .map(|d| { - let r = d.location().span.map(|span| span + range.start()); - - d.with_file_path(path.as_path().display().to_string()) - .with_file_span(r.unwrap_or(range)) - }) - } else { - None - } + let pool = connection.get_pool(); + + let typecheck_params: Vec<_> = doc + .iter_statements_with_text_and_range() + .map(|(stmt, range, text)| { + let ast = self.pg_query.get_ast(&stmt); + let tree = self.tree_sitter.get_parse_tree(&stmt); + (text.to_string(), ast, tree, *range) + }) + .collect(); + + let pool_clone = pool.clone(); + let path_clone = params.path.clone(); + let async_results = run_async(async move { + stream::iter(typecheck_params) + .map(|(text, ast, tree, range)| { + let pool = pool_clone.clone(); + let path = path_clone.clone(); + async move { + if let Some(ast) = ast { + pg_typecheck::check_sql(TypecheckParams { + conn: &pool, + sql: &text, + ast: &ast, + tree: tree.as_deref(), + }) + .await + .map(|d| { + let r = d.location().span.map(|span| span + range.start()); + + d.with_file_path(path.as_path().display().to_string()) + .with_file_span(r.unwrap_or(range)) + }) + } else { + None } - }) - .buffer_unordered(10) - .collect::>() - .await - })?; - - for result in async_results.into_iter().flatten() { - diagnostics.push(SDiagnostic::new(result)); - } + } + }) + .buffer_unordered(10) + .collect::>() + .await + })?; + + for result in async_results.into_iter().flatten() { + diagnostics.push(SDiagnostic::new(result)); } } @@ -529,14 +439,12 @@ impl Workspace for WorkspaceServer { tracing::debug!("Found the statement. We're looking for position {:?}. Statement Range {:?} to {:?}. Statement: {}", position, stmt_range.start(), stmt_range.end(), text); - let schema_cache = self - .schema_cache - .read() - .map_err(|_| WorkspaceError::runtime("Unable to load SchemaCache"))?; + let pool = self.connection.read().unwrap().get_pool(); + let schema_cache = self.schema_cache.load(pool)?; let result = pg_completions::complete(pg_completions::CompletionParams { position, - schema: &schema_cache, + schema: schema_cache.as_ref(), tree: tree.as_deref(), text: text.to_string(), }); @@ -550,15 +458,3 @@ impl Workspace for WorkspaceServer { fn is_dir(path: &Path) -> bool { path.is_dir() || (path.is_symlink() && fs::read_link(path).is_ok_and(|path| path.is_dir())) } - -/// Use this function to run async functions in the workspace, which is a sync trait called from an -/// async context. -/// -/// Checkout https://greptime.com/blogs/2023-03-09-bridging-async-and-sync-rust for details. -fn run_async(future: F) -> Result -where - F: Future + Send + 'static, - R: Send + 'static, -{ - futures::executor::block_on(async { RUNTIME.spawn(future).await.map_err(|e| e.into()) }) -} diff --git a/crates/pg_workspace/src/workspace/server/async_helper.rs b/crates/pg_workspace/src/workspace/server/async_helper.rs new file mode 100644 index 00000000..896a63a4 --- /dev/null +++ b/crates/pg_workspace/src/workspace/server/async_helper.rs @@ -0,0 +1,21 @@ +use std::{future::Future, sync::LazyLock}; + +use tokio::runtime::Runtime; + +use crate::WorkspaceError; + +// Global Tokio Runtime +static RUNTIME: LazyLock = + LazyLock::new(|| Runtime::new().expect("Failed to create Tokio runtime")); + +/// Use this function to run async functions in the workspace, which is a sync trait called from an +/// async context. +/// +/// Checkout https://greptime.com/blogs/2023-03-09-bridging-async-and-sync-rust for details. +pub fn run_async(future: F) -> Result +where + F: Future + Send + 'static, + R: Send + 'static, +{ + futures::executor::block_on(async { RUNTIME.spawn(future).await.map_err(|e| e.into()) }) +} diff --git a/crates/pg_workspace/src/workspace/server/db_connection.rs b/crates/pg_workspace/src/workspace/server/db_connection.rs new file mode 100644 index 00000000..1dc2e127 --- /dev/null +++ b/crates/pg_workspace/src/workspace/server/db_connection.rs @@ -0,0 +1,34 @@ +use sqlx::{pool::PoolOptions, postgres::PgConnectOptions, PgPool, Postgres}; + +use crate::settings::DatabaseSettings; + +#[derive(Default)] +pub struct DbConnection { + pool: Option, +} + +impl DbConnection { + /// Requires that you call `set_conn_settings` at least once before getting a pool. + pub(crate) fn get_pool(&self) -> PgPool { + self.pool + .clone() + .expect("The database has never been properly initialized.") + } + + pub(crate) fn set_conn_settings(&mut self, settings: &DatabaseSettings) { + let config = PgConnectOptions::new() + .host(&settings.host) + .port(settings.port) + .username(&settings.username) + .password(&settings.password) + .database(&settings.database); + + let timeout = settings.conn_timeout_secs.clone(); + + let pool = PoolOptions::::new() + .acquire_timeout(timeout) + .connect_lazy_with(config); + + self.pool = Some(pool); + } +} diff --git a/crates/pg_workspace/src/workspace/server/schema_cache_manager.rs b/crates/pg_workspace/src/workspace/server/schema_cache_manager.rs new file mode 100644 index 00000000..36f026f5 --- /dev/null +++ b/crates/pg_workspace/src/workspace/server/schema_cache_manager.rs @@ -0,0 +1,54 @@ +use std::sync::{RwLock, RwLockReadGuard}; + +use pg_schema_cache::SchemaCache; +use sqlx::PgPool; + +use crate::WorkspaceError; + +use super::async_helper::run_async; + +#[derive(Debug)] +pub(crate) struct SchemaCacheHandle<'a> { + inner: RwLockReadGuard<'a, SchemaCache>, +} + +impl<'a> SchemaCacheHandle<'a> { + pub(crate) fn new(cache: &'a RwLock) -> Self { + Self { + inner: cache.read().unwrap(), + } + } + + pub(crate) fn wrap(inner: RwLockReadGuard<'a, SchemaCache>) -> Self { + Self { inner } + } +} + +impl AsRef for SchemaCacheHandle<'_> { + fn as_ref(&self) -> &SchemaCache { + &self.inner + } +} + +#[derive(Default)] +pub struct SchemaCacheManager { + cache: RwLock, +} + +impl SchemaCacheManager { + pub fn load(&self, pool: PgPool) -> Result { + let cache_lock = self.cache.read().unwrap(); + + if cache_lock.has_already_cached_connection(&pool) { + Ok(SchemaCacheHandle::wrap(cache_lock)) + } else { + let maybe_refreshed = run_async(async move { SchemaCache::load(&pool).await })?; + let refreshed = maybe_refreshed?; + + let mut cache = self.cache.write().unwrap(); + *cache = refreshed; + + Ok(SchemaCacheHandle::new(&self.cache)) + } + } +} diff --git a/pglsp.toml b/pglsp.toml index ea394361..f616a30d 100644 --- a/pglsp.toml +++ b/pglsp.toml @@ -7,7 +7,7 @@ use_ignore_file = false ignore = [] [db] -host = "127.0.0.1" +host = "127.1.1.1" port = 5432 username = "postgres" password = "postgres" From 8a24148e4169ea88657280bfb9108c70f98d2a75 Mon Sep 17 00:00:00 2001 From: Julian Date: Mon, 3 Feb 2025 10:27:02 +0100 Subject: [PATCH 06/16] change toml --- pglsp.toml | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/pglsp.toml b/pglsp.toml index f616a30d..4f7d325d 100644 --- a/pglsp.toml +++ b/pglsp.toml @@ -7,11 +7,12 @@ use_ignore_file = false ignore = [] [db] -host = "127.1.1.1" +host = "127.0.0.1" port = 5432 username = "postgres" password = "postgres" database = "postgres" +conn_timeout_secs = 10 # [migrations] # migrations_dir = "migrations" From 4ffabc2d66519c3ce2c4638dd510579de4c65146 Mon Sep 17 00:00:00 2001 From: Julian Date: Mon, 3 Feb 2025 10:40:16 +0100 Subject: [PATCH 07/16] a little safer --- crates/pg_schema_cache/src/schema_cache.rs | 22 +++++++++++++++------- 1 file changed, 15 insertions(+), 7 deletions(-) diff --git a/crates/pg_schema_cache/src/schema_cache.rs b/crates/pg_schema_cache/src/schema_cache.rs index 0fdb7f39..154cafcd 100644 --- a/crates/pg_schema_cache/src/schema_cache.rs +++ b/crates/pg_schema_cache/src/schema_cache.rs @@ -44,13 +44,21 @@ impl SchemaCache { fn pool_to_conn_str(pool: &PgPool) -> String { let conn = pool.connect_options(); - format!( - "postgres://{}:@{}:{}/{}", - conn.get_username(), - conn.get_host(), - conn.get_port(), - conn.get_database().unwrap_or("default") - ) + match conn.get_database() { + None => format!( + "postgres://{}:@{}:{}", + conn.get_username(), + conn.get_host(), + conn.get_port() + ), + Some(db) => format!( + "postgres://{}:@{}:{}/{}", + conn.get_username(), + conn.get_host(), + conn.get_port(), + db + ), + } } pub fn has_already_cached_connection(&self, pool: &PgPool) -> bool { From 8bd878b98d7491b70406ffc36d264698052c7d5f Mon Sep 17 00:00:00 2001 From: Julian Date: Tue, 4 Feb 2025 08:06:23 +0100 Subject: [PATCH 08/16] lift cache key --- crates/pg_schema_cache/src/schema_cache.rs | 27 ---------- .../workspace/server/schema_cache_manager.rs | 53 ++++++++++++++----- 2 files changed, 41 insertions(+), 39 deletions(-) diff --git a/crates/pg_schema_cache/src/schema_cache.rs b/crates/pg_schema_cache/src/schema_cache.rs index 154cafcd..77a0526a 100644 --- a/crates/pg_schema_cache/src/schema_cache.rs +++ b/crates/pg_schema_cache/src/schema_cache.rs @@ -9,8 +9,6 @@ use crate::versions::Version; #[derive(Debug, Clone, Default)] pub struct SchemaCache { - cached_conn_str: String, - pub schemas: Vec, pub tables: Vec
, pub functions: Vec, @@ -31,7 +29,6 @@ impl SchemaCache { )?; Ok(SchemaCache { - cached_conn_str: SchemaCache::pool_to_conn_str(pool), schemas, tables, functions, @@ -41,30 +38,6 @@ impl SchemaCache { }) } - fn pool_to_conn_str(pool: &PgPool) -> String { - let conn = pool.connect_options(); - - match conn.get_database() { - None => format!( - "postgres://{}:@{}:{}", - conn.get_username(), - conn.get_host(), - conn.get_port() - ), - Some(db) => format!( - "postgres://{}:@{}:{}/{}", - conn.get_username(), - conn.get_host(), - conn.get_port(), - db - ), - } - } - - pub fn has_already_cached_connection(&self, pool: &PgPool) -> bool { - self.cached_conn_str == SchemaCache::pool_to_conn_str(pool) - } - /// Applies an AST node to the repository /// /// For example, alter table add column will add the column to the table if it does not exist diff --git a/crates/pg_workspace/src/workspace/server/schema_cache_manager.rs b/crates/pg_workspace/src/workspace/server/schema_cache_manager.rs index 36f026f5..9df910e3 100644 --- a/crates/pg_workspace/src/workspace/server/schema_cache_manager.rs +++ b/crates/pg_workspace/src/workspace/server/schema_cache_manager.rs @@ -7,48 +7,77 @@ use crate::WorkspaceError; use super::async_helper::run_async; -#[derive(Debug)] pub(crate) struct SchemaCacheHandle<'a> { - inner: RwLockReadGuard<'a, SchemaCache>, + inner: RwLockReadGuard<'a, SchemaCacheManagerInner>, } impl<'a> SchemaCacheHandle<'a> { - pub(crate) fn new(cache: &'a RwLock) -> Self { + pub(crate) fn new(cache: &'a RwLock) -> Self { Self { inner: cache.read().unwrap(), } } - pub(crate) fn wrap(inner: RwLockReadGuard<'a, SchemaCache>) -> Self { + pub(crate) fn wrap(inner: RwLockReadGuard<'a, SchemaCacheManagerInner>) -> Self { Self { inner } } } impl AsRef for SchemaCacheHandle<'_> { fn as_ref(&self) -> &SchemaCache { - &self.inner + &self.inner.cache } } +#[derive(Default)] +pub(crate) struct SchemaCacheManagerInner { + cache: SchemaCache, + conn_str: String, +} + #[derive(Default)] pub struct SchemaCacheManager { - cache: RwLock, + inner: RwLock, } impl SchemaCacheManager { pub fn load(&self, pool: PgPool) -> Result { - let cache_lock = self.cache.read().unwrap(); + let inner = self.inner.read().unwrap(); - if cache_lock.has_already_cached_connection(&pool) { - Ok(SchemaCacheHandle::wrap(cache_lock)) + if pool_to_conn_str(&pool) == inner.conn_str { + Ok(SchemaCacheHandle::wrap(inner)) } else { + let new_conn_str = pool_to_conn_str(&pool); + let maybe_refreshed = run_async(async move { SchemaCache::load(&pool).await })?; let refreshed = maybe_refreshed?; - let mut cache = self.cache.write().unwrap(); - *cache = refreshed; + let mut inner = self.inner.write().unwrap(); + + inner.cache = refreshed; + inner.conn_str = new_conn_str; - Ok(SchemaCacheHandle::new(&self.cache)) + Ok(SchemaCacheHandle::new(&self.inner)) } } } + +fn pool_to_conn_str(pool: &PgPool) -> String { + let conn = pool.connect_options(); + + match conn.get_database() { + None => format!( + "postgres://{}:@{}:{}", + conn.get_username(), + conn.get_host(), + conn.get_port() + ), + Some(db) => format!( + "postgres://{}:@{}:{}/{}", + conn.get_username(), + conn.get_host(), + conn.get_port(), + db + ), + } +} From 221e9cc9ad7a45df78bc16943baa1a5e01d191b3 Mon Sep 17 00:00:00 2001 From: Julian Date: Tue, 4 Feb 2025 09:15:45 +0100 Subject: [PATCH 09/16] skip db checks via flag --- crates/pg_cli/src/cli_options.rs | 4 +++ crates/pg_cli/src/commands/mod.rs | 3 ++ .../pg_cli/src/execute/process_file/check.rs | 1 + crates/pg_lsp/src/session.rs | 1 + crates/pg_workspace/src/workspace.rs | 1 + crates/pg_workspace/src/workspace/server.rs | 30 ++++++++++++------- .../src/workspace/server/db_connection.rs | 11 +++---- 7 files changed, 35 insertions(+), 16 deletions(-) diff --git a/crates/pg_cli/src/cli_options.rs b/crates/pg_cli/src/cli_options.rs index 24d7a3c3..575c25fe 100644 --- a/crates/pg_cli/src/cli_options.rs +++ b/crates/pg_cli/src/cli_options.rs @@ -18,6 +18,10 @@ pub struct CliOptions { #[bpaf(long("use-server"), switch, fallback(false))] pub use_server: bool, + /// Skip over files containing syntax errors instead of emitting an error diagnostic. + #[bpaf(long("skip-db"), switch, fallback(false))] + pub skip_db: bool, + /// Print additional diagnostics, and some diagnostics show more information. Also, print out what files were processed and which ones were modified. #[bpaf(long("verbose"), switch, fallback(false))] pub verbose: bool, diff --git a/crates/pg_cli/src/commands/mod.rs b/crates/pg_cli/src/commands/mod.rs index 6b47ecc6..6fe3472d 100644 --- a/crates/pg_cli/src/commands/mod.rs +++ b/crates/pg_cli/src/commands/mod.rs @@ -35,8 +35,10 @@ pub enum PgLspCommand { Check { #[bpaf(external(partial_configuration), hide_usage, optional)] configuration: Option, + #[bpaf(external, hide_usage)] cli_options: CliOptions, + /// Use this option when you want to format code piped from `stdin`, and print the output to `stdout`. /// /// The file doesn't need to exist on disk, what matters is the extension of the file. Based on the extension, we know how to check the code. @@ -286,6 +288,7 @@ pub(crate) trait CommandRunner: Sized { configuration, vcs_base_path, gitignore_matches, + skip_db: cli_options.skip_db, })?; let execution = self.get_execution(cli_options, console, workspace)?; diff --git a/crates/pg_cli/src/execute/process_file/check.rs b/crates/pg_cli/src/execute/process_file/check.rs index fa5b522b..f364587e 100644 --- a/crates/pg_cli/src/execute/process_file/check.rs +++ b/crates/pg_cli/src/execute/process_file/check.rs @@ -35,6 +35,7 @@ pub(crate) fn check_with_guard<'ctx>( max_diagnostics, only, skip, + true, ) .with_file_path_and_code( workspace_file.path.display().to_string(), diff --git a/crates/pg_lsp/src/session.rs b/crates/pg_lsp/src/session.rs index 9af4c83d..d8379617 100644 --- a/crates/pg_lsp/src/session.rs +++ b/crates/pg_lsp/src/session.rs @@ -464,6 +464,7 @@ impl Session { configuration: fs_configuration, vcs_base_path, gitignore_matches, + skip_db: false, }); if let Err(error) = result { diff --git a/crates/pg_workspace/src/workspace.rs b/crates/pg_workspace/src/workspace.rs index 0dfb9f7c..16e6d135 100644 --- a/crates/pg_workspace/src/workspace.rs +++ b/crates/pg_workspace/src/workspace.rs @@ -90,6 +90,7 @@ pub struct UpdateSettingsParams { pub vcs_base_path: Option, pub gitignore_matches: Vec, pub workspace_directory: Option, + pub skip_db: bool, } #[derive(Debug, serde::Serialize, serde::Deserialize)] diff --git a/crates/pg_workspace/src/workspace/server.rs b/crates/pg_workspace/src/workspace/server.rs index a94b29e9..a8734da2 100644 --- a/crates/pg_workspace/src/workspace/server.rs +++ b/crates/pg_workspace/src/workspace/server.rs @@ -161,10 +161,12 @@ impl Workspace for WorkspaceServer { tracing::info!("Updated settings in workspace"); - self.connection - .write() - .unwrap() - .set_conn_settings(&self.settings().as_ref().db); + if !params.skip_db { + self.connection + .write() + .unwrap() + .set_conn_settings(&self.settings().as_ref().db); + } tracing::info!("Updated Db connection settings"); @@ -298,10 +300,12 @@ impl Workspace for WorkspaceServer { let mut diagnostics: Vec = vec![]; - // run diagnostics for each statement in parallel if its mostly i/o work - if let Ok(connection) = self.connection.read() { - let pool = connection.get_pool(); - + if let Some(pool) = self + .connection + .read() + .expect("DbConnection RwLock panicked") + .get_pool() + { let typecheck_params: Vec<_> = doc .iter_statements_with_text_and_range() .map(|(stmt, range, text)| { @@ -311,12 +315,12 @@ impl Workspace for WorkspaceServer { }) .collect(); - let pool_clone = pool.clone(); + // run diagnostics for each statement in parallel if its mostly i/o work let path_clone = params.path.clone(); let async_results = run_async(async move { stream::iter(typecheck_params) .map(|(text, ast, tree, range)| { - let pool = pool_clone.clone(); + let pool = pool.clone(); let path = path_clone.clone(); async move { if let Some(ast) = ast { @@ -412,6 +416,11 @@ impl Workspace for WorkspaceServer { ¶ms.position ); + let pool = match self.connection.read().unwrap().get_pool() { + Some(pool) => pool, + None => return Ok(pg_completions::CompletionResult::default()), + }; + let doc = self .documents .get(¶ms.path) @@ -439,7 +448,6 @@ impl Workspace for WorkspaceServer { tracing::debug!("Found the statement. We're looking for position {:?}. Statement Range {:?} to {:?}. Statement: {}", position, stmt_range.start(), stmt_range.end(), text); - let pool = self.connection.read().unwrap().get_pool(); let schema_cache = self.schema_cache.load(pool)?; let result = pg_completions::complete(pg_completions::CompletionParams { diff --git a/crates/pg_workspace/src/workspace/server/db_connection.rs b/crates/pg_workspace/src/workspace/server/db_connection.rs index 1dc2e127..3a747342 100644 --- a/crates/pg_workspace/src/workspace/server/db_connection.rs +++ b/crates/pg_workspace/src/workspace/server/db_connection.rs @@ -1,3 +1,5 @@ +use std::time::Duration; + use sqlx::{pool::PoolOptions, postgres::PgConnectOptions, PgPool, Postgres}; use crate::settings::DatabaseSettings; @@ -8,11 +10,9 @@ pub struct DbConnection { } impl DbConnection { - /// Requires that you call `set_conn_settings` at least once before getting a pool. - pub(crate) fn get_pool(&self) -> PgPool { - self.pool - .clone() - .expect("The database has never been properly initialized.") + /// There might be no pool available if the user decides to skip db checks. + pub(crate) fn get_pool(&self) -> Option { + self.pool.clone() } pub(crate) fn set_conn_settings(&mut self, settings: &DatabaseSettings) { @@ -27,6 +27,7 @@ impl DbConnection { let pool = PoolOptions::::new() .acquire_timeout(timeout) + .acquire_slow_threshold(Duration::from_secs(2)) .connect_lazy_with(config); self.pool = Some(pool); From 96901d9d1ead590621ab85340b17a12e6fdd8b78 Mon Sep 17 00:00:00 2001 From: Julian Date: Tue, 4 Feb 2025 09:16:05 +0100 Subject: [PATCH 10/16] hm --- crates/pg_cli/src/execute/process_file/check.rs | 1 - 1 file changed, 1 deletion(-) diff --git a/crates/pg_cli/src/execute/process_file/check.rs b/crates/pg_cli/src/execute/process_file/check.rs index f364587e..fa5b522b 100644 --- a/crates/pg_cli/src/execute/process_file/check.rs +++ b/crates/pg_cli/src/execute/process_file/check.rs @@ -35,7 +35,6 @@ pub(crate) fn check_with_guard<'ctx>( max_diagnostics, only, skip, - true, ) .with_file_path_and_code( workspace_file.path.display().to_string(), From acce4e38a56601c5a46bef425fa24f4196cb6615 Mon Sep 17 00:00:00 2001 From: Julian Date: Sat, 8 Feb 2025 16:29:18 +0100 Subject: [PATCH 11/16] like this maybe? --- crates/pg_cli/src/execute/mod.rs | 6 +++ crates/pg_cli/src/execute/process_file.rs | 1 + .../pg_cli/src/execute/process_file/check.rs | 5 ++ crates/pg_cli/src/execute/traverse.rs | 46 ++++++++++++++++--- crates/pg_cli/src/reporter/github.rs | 15 ++++++ crates/pg_cli/src/reporter/gitlab.rs | 19 +++++++- crates/pg_cli/src/reporter/junit.rs | 17 +++++++ crates/pg_cli/src/reporter/mod.rs | 11 +++++ crates/pg_cli/src/reporter/terminal.rs | 16 +++++++ crates/pg_workspace/src/workspace.rs | 1 + crates/pg_workspace/src/workspace/server.rs | 9 ++-- pglsp.toml | 4 +- 12 files changed, 137 insertions(+), 13 deletions(-) diff --git a/crates/pg_cli/src/execute/mod.rs b/crates/pg_cli/src/execute/mod.rs index c18f8ec2..36d6dfbf 100644 --- a/crates/pg_cli/src/execute/mod.rs +++ b/crates/pg_cli/src/execute/mod.rs @@ -9,6 +9,7 @@ use crate::reporter::github::{GithubReporter, GithubReporterVisitor}; use crate::reporter::gitlab::{GitLabReporter, GitLabReporterVisitor}; use crate::reporter::junit::{JunitReporter, JunitReporterVisitor}; use crate::reporter::terminal::{ConsoleReporter, ConsoleReporterVisitor}; +use crate::reporter::UserHintsPayload; use crate::{CliDiagnostic, CliSession, DiagnosticsPayload, Reporter}; use pg_diagnostics::{category, Category}; use pg_fs::PgLspPath; @@ -242,6 +243,7 @@ pub fn execute_mode( summary, evaluated_paths, diagnostics, + user_hints, } = traverse(&execution, &mut session, cli_options, paths)?; let console = session.app.console; let errors = summary.errors; @@ -260,6 +262,7 @@ pub fn execute_mode( }, execution: execution.clone(), evaluated_paths, + user_hints_payload: UserHintsPayload { hints: user_hints }, }; reporter.write(&mut ConsoleReporterVisitor(console))?; } @@ -271,6 +274,7 @@ pub fn execute_mode( diagnostics, }, execution: execution.clone(), + user_hints: UserHintsPayload { hints: user_hints }, }; reporter.write(&mut GithubReporterVisitor(console))?; } @@ -282,6 +286,7 @@ pub fn execute_mode( diagnostics, }, execution: execution.clone(), + user_hints: UserHintsPayload { hints: user_hints }, }; reporter.write(&mut GitLabReporterVisitor::new( console, @@ -297,6 +302,7 @@ pub fn execute_mode( diagnostics, }, execution: execution.clone(), + user_hints: UserHintsPayload { hints: user_hints }, }; reporter.write(&mut JunitReporterVisitor::new(console))?; } diff --git a/crates/pg_cli/src/execute/process_file.rs b/crates/pg_cli/src/execute/process_file.rs index 6135c012..4df1f51a 100644 --- a/crates/pg_cli/src/execute/process_file.rs +++ b/crates/pg_cli/src/execute/process_file.rs @@ -46,6 +46,7 @@ pub(crate) enum Message { diagnostics: Vec, skipped_diagnostics: u32, }, + Hint(String), } impl Message { diff --git a/crates/pg_cli/src/execute/process_file/check.rs b/crates/pg_cli/src/execute/process_file/check.rs index fa5b522b..9795d847 100644 --- a/crates/pg_cli/src/execute/process_file/check.rs +++ b/crates/pg_cli/src/execute/process_file/check.rs @@ -28,6 +28,7 @@ pub(crate) fn check_with_guard<'ctx>( let (only, skip) = (Vec::new(), Vec::new()); let max_diagnostics = ctx.remaining_diagnostics.load(Ordering::Relaxed); + let pull_diagnostics_result = workspace_file .guard() .pull_diagnostics( @@ -41,6 +42,10 @@ pub(crate) fn check_with_guard<'ctx>( category!("check"), )?; + if pull_diagnostics_result.skipped_db_checks { + ctx.set_skipped_db_conn(true); + } + let no_diagnostics = pull_diagnostics_result.diagnostics.is_empty() && pull_diagnostics_result.skipped_diagnostics == 0; diff --git a/crates/pg_cli/src/execute/traverse.rs b/crates/pg_cli/src/execute/traverse.rs index 873b8905..40a36c78 100644 --- a/crates/pg_cli/src/execute/traverse.rs +++ b/crates/pg_cli/src/execute/traverse.rs @@ -14,7 +14,7 @@ use pg_workspace::workspace::IsPathIgnoredParams; use pg_workspace::{Workspace, WorkspaceError}; use rustc_hash::FxHashSet; use std::collections::BTreeSet; -use std::sync::atomic::AtomicU32; +use std::sync::atomic::{AtomicBool, AtomicU32}; use std::sync::RwLock; use std::{ env::current_dir, @@ -33,6 +33,7 @@ pub(crate) struct TraverseResult { pub(crate) summary: TraversalSummary, pub(crate) evaluated_paths: BTreeSet, pub(crate) diagnostics: Vec, + pub(crate) user_hints: Vec, } pub(crate) fn traverse( @@ -72,6 +73,7 @@ pub(crate) fn traverse( let unchanged = AtomicUsize::new(0); let matches = AtomicUsize::new(0); let skipped = AtomicUsize::new(0); + let skipped_db_conn = AtomicBool::new(false); let fs = &*session.app.fs; let workspace = &*session.app.workspace; @@ -84,7 +86,7 @@ pub(crate) fn traverse( .with_diagnostic_level(cli_options.diagnostic_level) .with_max_diagnostics(max_diagnostics); - let (duration, evaluated_paths, diagnostics) = thread::scope(|s| { + let (duration, evaluated_paths, diagnostics, mut user_hints) = thread::scope(|s| { let handler = thread::Builder::new() .name(String::from("pglsp::console")) .spawn_scoped(s, || printer.run(receiver, recv_files)) @@ -104,15 +106,16 @@ pub(crate) fn traverse( changed: &changed, unchanged: &unchanged, skipped: &skipped, + skipped_db_conn: &skipped_db_conn, messages: sender, remaining_diagnostics: &remaining_diagnostics, evaluated_paths: RwLock::default(), }, ); // wait for the main thread to finish - let diagnostics = handler.join().unwrap(); + let (diagnostics, user_hints) = handler.join().unwrap(); - (elapsed, evaluated_paths, diagnostics) + (elapsed, evaluated_paths, diagnostics, user_hints) }); let errors = printer.errors(); @@ -123,6 +126,20 @@ pub(crate) fn traverse( let skipped = skipped.load(Ordering::Relaxed); let suggested_fixes_skipped = printer.skipped_fixes(); let diagnostics_not_printed = printer.not_printed_diagnostics(); + + if duration.as_secs() >= 2 { + user_hints.push(format!( + "The traversal took longer than expected ({}s). Consider using the `--skip-db` option if your Postgres connection is slow.", + duration.as_secs() + )); + } + + if skipped_db_conn.load(Ordering::Relaxed) { + user_hints.push(format!( + "Skipped all checks requiring database connections.", + )); + } + Ok(TraverseResult { summary: TraversalSummary { changed, @@ -137,6 +154,7 @@ pub(crate) fn traverse( }, evaluated_paths, diagnostics, + user_hints, }) } @@ -288,10 +306,15 @@ impl<'ctx> DiagnosticsPrinter<'ctx> { should_print } - fn run(&self, receiver: Receiver, interner: Receiver) -> Vec { + fn run( + &self, + receiver: Receiver, + interner: Receiver, + ) -> (Vec, Vec) { let mut paths: FxHashSet = FxHashSet::default(); let mut diagnostics_to_print = vec![]; + let mut hints_to_print = vec![]; while let Ok(msg) = receiver.recv() { match msg { @@ -306,6 +329,10 @@ impl<'ctx> DiagnosticsPrinter<'ctx> { self.errors.fetch_add(1, Ordering::Relaxed); } + Message::Hint(hint) => { + hints_to_print.push(hint); + } + Message::Error(mut err) => { let location = err.location(); if self.should_skip_diagnostic(err.severity(), err.tags()) { @@ -381,7 +408,8 @@ impl<'ctx> DiagnosticsPrinter<'ctx> { } } } - diagnostics_to_print + + (diagnostics_to_print, hints_to_print) } } @@ -403,6 +431,8 @@ pub(crate) struct TraversalOptions<'ctx, 'app> { matches: &'ctx AtomicUsize, /// Shared atomic counter storing the number of skipped files skipped: &'ctx AtomicUsize, + /// Shared atomic bool tracking whether we used a DB connection + skipped_db_conn: &'ctx AtomicBool, /// Channel sending messages to the display thread pub(crate) messages: Sender, /// The approximate number of diagnostics the console will print before @@ -434,6 +464,10 @@ impl TraversalOptions<'_, '_> { self.messages.send(msg.into()).ok(); } + pub(crate) fn set_skipped_db_conn(&self, has_skipped: bool) { + self.skipped_db_conn.store(has_skipped, Ordering::Relaxed); + } + pub(crate) fn protected_file(&self, pglsp_path: &PgLspPath) { self.push_diagnostic( WorkspaceError::protected_file(pglsp_path.display().to_string()).into(), diff --git a/crates/pg_cli/src/reporter/github.rs b/crates/pg_cli/src/reporter/github.rs index 6b1588b1..ecb70b03 100644 --- a/crates/pg_cli/src/reporter/github.rs +++ b/crates/pg_cli/src/reporter/github.rs @@ -3,14 +3,18 @@ use pg_console::{markup, Console, ConsoleExt}; use pg_diagnostics::PrintGitHubDiagnostic; use std::io; +use super::UserHintsPayload; + pub(crate) struct GithubReporter { pub(crate) diagnostics_payload: DiagnosticsPayload, pub(crate) execution: Execution, + pub(crate) user_hints: UserHintsPayload, } impl Reporter for GithubReporter { fn write(self, visitor: &mut dyn ReporterVisitor) -> io::Result<()> { visitor.report_diagnostics(&self.execution, self.diagnostics_payload)?; + visitor.report_user_hints(&self.execution, self.user_hints)?; Ok(()) } } @@ -42,4 +46,15 @@ impl ReporterVisitor for GithubReporterVisitor<'_> { Ok(()) } + + fn report_user_hints( + &mut self, + _execution: &Execution, + payload: super::UserHintsPayload, + ) -> io::Result<()> { + for hint in payload.hints { + self.0.log(markup! {{hint}}); + } + Ok(()) + } } diff --git a/crates/pg_cli/src/reporter/gitlab.rs b/crates/pg_cli/src/reporter/gitlab.rs index ea3fd285..6fe580b3 100644 --- a/crates/pg_cli/src/reporter/gitlab.rs +++ b/crates/pg_cli/src/reporter/gitlab.rs @@ -12,14 +12,18 @@ use std::{ path::{Path, PathBuf}, }; +use super::UserHintsPayload; + pub struct GitLabReporter { - pub execution: Execution, - pub diagnostics: DiagnosticsPayload, + pub(crate) execution: Execution, + pub(crate) diagnostics: DiagnosticsPayload, + pub(crate) user_hints: UserHintsPayload, } impl Reporter for GitLabReporter { fn write(self, visitor: &mut dyn ReporterVisitor) -> std::io::Result<()> { visitor.report_diagnostics(&self.execution, self.diagnostics)?; + visitor.report_user_hints(&self.execution, self.user_hints)?; Ok(()) } } @@ -72,6 +76,17 @@ impl ReporterVisitor for GitLabReporterVisitor<'_> { self.console.log(markup!({ diagnostics })); Ok(()) } + + fn report_user_hints( + &mut self, + _execution: &Execution, + payload: super::UserHintsPayload, + ) -> std::io::Result<()> { + for hint in payload.hints { + self.console.log(markup! {{hint}}); + } + Ok(()) + } } struct GitLabDiagnostics<'a>( diff --git a/crates/pg_cli/src/reporter/junit.rs b/crates/pg_cli/src/reporter/junit.rs index c10059fc..2fe0fb30 100644 --- a/crates/pg_cli/src/reporter/junit.rs +++ b/crates/pg_cli/src/reporter/junit.rs @@ -6,16 +6,20 @@ use quick_junit::{NonSuccessKind, Report, TestCase, TestCaseStatus, TestSuite}; use std::fmt::{Display, Formatter}; use std::io; +use super::UserHintsPayload; + pub(crate) struct JunitReporter { pub(crate) diagnostics_payload: DiagnosticsPayload, pub(crate) execution: Execution, pub(crate) summary: TraversalSummary, + pub(crate) user_hints: UserHintsPayload, } impl Reporter for JunitReporter { fn write(self, visitor: &mut dyn ReporterVisitor) -> io::Result<()> { visitor.report_summary(&self.execution, self.summary)?; visitor.report_diagnostics(&self.execution, self.diagnostics_payload)?; + visitor.report_user_hints(&self.execution, self.user_hints)?; Ok(()) } } @@ -118,4 +122,17 @@ impl ReporterVisitor for JunitReporterVisitor<'_> { Ok(()) } + + fn report_user_hints( + &mut self, + _execution: &Execution, + payload: super::UserHintsPayload, + ) -> io::Result<()> { + for hint in payload.hints { + self.1.log(markup! { + {hint} + }); + } + Ok(()) + } } diff --git a/crates/pg_cli/src/reporter/mod.rs b/crates/pg_cli/src/reporter/mod.rs index adc7c023..25593056 100644 --- a/crates/pg_cli/src/reporter/mod.rs +++ b/crates/pg_cli/src/reporter/mod.rs @@ -17,6 +17,10 @@ pub struct DiagnosticsPayload { pub diagnostic_level: Severity, } +pub struct UserHintsPayload { + pub hints: Vec, +} + /// A type that holds the result of the traversal #[derive(Debug, Default, Serialize, Copy, Clone)] pub struct TraversalSummary { @@ -60,4 +64,11 @@ pub trait ReporterVisitor { execution: &Execution, payload: DiagnosticsPayload, ) -> io::Result<()>; + + /// Writes a diagnostics + fn report_user_hints( + &mut self, + execution: &Execution, + payload: UserHintsPayload, + ) -> io::Result<()>; } diff --git a/crates/pg_cli/src/reporter/terminal.rs b/crates/pg_cli/src/reporter/terminal.rs index 5fd493a2..6cfbb90a 100644 --- a/crates/pg_cli/src/reporter/terminal.rs +++ b/crates/pg_cli/src/reporter/terminal.rs @@ -10,11 +10,14 @@ use std::collections::BTreeSet; use std::io; use std::time::Duration; +use super::UserHintsPayload; + pub(crate) struct ConsoleReporter { pub(crate) summary: TraversalSummary, pub(crate) diagnostics_payload: DiagnosticsPayload, pub(crate) execution: Execution, pub(crate) evaluated_paths: BTreeSet, + pub(crate) user_hints_payload: UserHintsPayload, } impl Reporter for ConsoleReporter { @@ -22,6 +25,7 @@ impl Reporter for ConsoleReporter { let verbose = self.diagnostics_payload.verbose; visitor.report_diagnostics(&self.execution, self.diagnostics_payload)?; visitor.report_summary(&self.execution, self.summary)?; + visitor.report_user_hints(&self.execution, self.user_hints_payload)?; if verbose { visitor.report_handled_paths(self.evaluated_paths)?; } @@ -115,6 +119,18 @@ impl ReporterVisitor for ConsoleReporterVisitor<'_> { Ok(()) } + + fn report_user_hints( + &mut self, + _execution: &Execution, + payload: UserHintsPayload, + ) -> io::Result<()> { + for hint in payload.hints { + self.0.log(markup! {{hint}}); + } + self.0.log(markup! {{"\n"}}); + Ok(()) + } } struct Files(usize); diff --git a/crates/pg_workspace/src/workspace.rs b/crates/pg_workspace/src/workspace.rs index 16e6d135..ba04d961 100644 --- a/crates/pg_workspace/src/workspace.rs +++ b/crates/pg_workspace/src/workspace.rs @@ -53,6 +53,7 @@ pub struct PullDiagnosticsResult { pub diagnostics: Vec, pub errors: usize, pub skipped_diagnostics: u64, + pub skipped_db_checks: bool, } #[derive(Debug, Clone, Copy, serde::Serialize, serde::Deserialize, PartialEq)] diff --git a/crates/pg_workspace/src/workspace/server.rs b/crates/pg_workspace/src/workspace/server.rs index a8734da2..5d479a85 100644 --- a/crates/pg_workspace/src/workspace/server.rs +++ b/crates/pg_workspace/src/workspace/server.rs @@ -300,12 +300,14 @@ impl Workspace for WorkspaceServer { let mut diagnostics: Vec = vec![]; - if let Some(pool) = self + let maybe_pool = self .connection .read() .expect("DbConnection RwLock panicked") - .get_pool() - { + .get_pool(); + + let skipped_db_checks = maybe_pool.is_none(); + if let Some(pool) = maybe_pool { let typecheck_params: Vec<_> = doc .iter_statements_with_text_and_range() .map(|(stmt, range, text)| { @@ -402,6 +404,7 @@ impl Workspace for WorkspaceServer { diagnostics, errors, skipped_diagnostics: 0, + skipped_db_checks, }) } diff --git a/pglsp.toml b/pglsp.toml index 4f7d325d..043ec1dc 100644 --- a/pglsp.toml +++ b/pglsp.toml @@ -7,12 +7,12 @@ use_ignore_file = false ignore = [] [db] -host = "127.0.0.1" +host = "127.1.1.1" port = 5432 username = "postgres" password = "postgres" database = "postgres" -conn_timeout_secs = 10 +conn_timeout_secs = 3 # [migrations] # migrations_dir = "migrations" From e51c14985ac527d3b71144ad0383401af99378f8 Mon Sep 17 00:00:00 2001 From: Julian Date: Sat, 8 Feb 2025 16:29:36 +0100 Subject: [PATCH 12/16] whops --- pglsp.toml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/pglsp.toml b/pglsp.toml index 043ec1dc..4f7d325d 100644 --- a/pglsp.toml +++ b/pglsp.toml @@ -7,12 +7,12 @@ use_ignore_file = false ignore = [] [db] -host = "127.1.1.1" +host = "127.0.0.1" port = 5432 username = "postgres" password = "postgres" database = "postgres" -conn_timeout_secs = 3 +conn_timeout_secs = 10 # [migrations] # migrations_dir = "migrations" From 5e174c7fcd167e2595c2cb199621e84bc8fb0309 Mon Sep 17 00:00:00 2001 From: Julian Date: Sun, 9 Feb 2025 09:00:34 +0100 Subject: [PATCH 13/16] adjust --- crates/pg_cli/src/cli_options.rs | 2 +- crates/pg_cli/src/execute/mod.rs | 6 ---- crates/pg_cli/src/execute/process_file.rs | 1 - crates/pg_cli/src/execute/traverse.rs | 44 ++++------------------- crates/pg_cli/src/reporter/github.rs | 15 -------- crates/pg_cli/src/reporter/gitlab.rs | 15 -------- crates/pg_cli/src/reporter/junit.rs | 17 --------- crates/pg_cli/src/reporter/mod.rs | 11 ------ crates/pg_cli/src/reporter/terminal.rs | 16 --------- 9 files changed, 7 insertions(+), 120 deletions(-) diff --git a/crates/pg_cli/src/cli_options.rs b/crates/pg_cli/src/cli_options.rs index 575c25fe..20e18c8c 100644 --- a/crates/pg_cli/src/cli_options.rs +++ b/crates/pg_cli/src/cli_options.rs @@ -18,7 +18,7 @@ pub struct CliOptions { #[bpaf(long("use-server"), switch, fallback(false))] pub use_server: bool, - /// Skip over files containing syntax errors instead of emitting an error diagnostic. + /// Skip connecting to the database and only run checks that don't require a database connection. #[bpaf(long("skip-db"), switch, fallback(false))] pub skip_db: bool, diff --git a/crates/pg_cli/src/execute/mod.rs b/crates/pg_cli/src/execute/mod.rs index 36d6dfbf..c18f8ec2 100644 --- a/crates/pg_cli/src/execute/mod.rs +++ b/crates/pg_cli/src/execute/mod.rs @@ -9,7 +9,6 @@ use crate::reporter::github::{GithubReporter, GithubReporterVisitor}; use crate::reporter::gitlab::{GitLabReporter, GitLabReporterVisitor}; use crate::reporter::junit::{JunitReporter, JunitReporterVisitor}; use crate::reporter::terminal::{ConsoleReporter, ConsoleReporterVisitor}; -use crate::reporter::UserHintsPayload; use crate::{CliDiagnostic, CliSession, DiagnosticsPayload, Reporter}; use pg_diagnostics::{category, Category}; use pg_fs::PgLspPath; @@ -243,7 +242,6 @@ pub fn execute_mode( summary, evaluated_paths, diagnostics, - user_hints, } = traverse(&execution, &mut session, cli_options, paths)?; let console = session.app.console; let errors = summary.errors; @@ -262,7 +260,6 @@ pub fn execute_mode( }, execution: execution.clone(), evaluated_paths, - user_hints_payload: UserHintsPayload { hints: user_hints }, }; reporter.write(&mut ConsoleReporterVisitor(console))?; } @@ -274,7 +271,6 @@ pub fn execute_mode( diagnostics, }, execution: execution.clone(), - user_hints: UserHintsPayload { hints: user_hints }, }; reporter.write(&mut GithubReporterVisitor(console))?; } @@ -286,7 +282,6 @@ pub fn execute_mode( diagnostics, }, execution: execution.clone(), - user_hints: UserHintsPayload { hints: user_hints }, }; reporter.write(&mut GitLabReporterVisitor::new( console, @@ -302,7 +297,6 @@ pub fn execute_mode( diagnostics, }, execution: execution.clone(), - user_hints: UserHintsPayload { hints: user_hints }, }; reporter.write(&mut JunitReporterVisitor::new(console))?; } diff --git a/crates/pg_cli/src/execute/process_file.rs b/crates/pg_cli/src/execute/process_file.rs index 4df1f51a..6135c012 100644 --- a/crates/pg_cli/src/execute/process_file.rs +++ b/crates/pg_cli/src/execute/process_file.rs @@ -46,7 +46,6 @@ pub(crate) enum Message { diagnostics: Vec, skipped_diagnostics: u32, }, - Hint(String), } impl Message { diff --git a/crates/pg_cli/src/execute/traverse.rs b/crates/pg_cli/src/execute/traverse.rs index 40a36c78..ee51125c 100644 --- a/crates/pg_cli/src/execute/traverse.rs +++ b/crates/pg_cli/src/execute/traverse.rs @@ -14,7 +14,7 @@ use pg_workspace::workspace::IsPathIgnoredParams; use pg_workspace::{Workspace, WorkspaceError}; use rustc_hash::FxHashSet; use std::collections::BTreeSet; -use std::sync::atomic::{AtomicBool, AtomicU32}; +use std::sync::atomic::AtomicU32; use std::sync::RwLock; use std::{ env::current_dir, @@ -33,7 +33,6 @@ pub(crate) struct TraverseResult { pub(crate) summary: TraversalSummary, pub(crate) evaluated_paths: BTreeSet, pub(crate) diagnostics: Vec, - pub(crate) user_hints: Vec, } pub(crate) fn traverse( @@ -73,7 +72,6 @@ pub(crate) fn traverse( let unchanged = AtomicUsize::new(0); let matches = AtomicUsize::new(0); let skipped = AtomicUsize::new(0); - let skipped_db_conn = AtomicBool::new(false); let fs = &*session.app.fs; let workspace = &*session.app.workspace; @@ -86,7 +84,7 @@ pub(crate) fn traverse( .with_diagnostic_level(cli_options.diagnostic_level) .with_max_diagnostics(max_diagnostics); - let (duration, evaluated_paths, diagnostics, mut user_hints) = thread::scope(|s| { + let (duration, evaluated_paths, diagnostics) = thread::scope(|s| { let handler = thread::Builder::new() .name(String::from("pglsp::console")) .spawn_scoped(s, || printer.run(receiver, recv_files)) @@ -106,16 +104,15 @@ pub(crate) fn traverse( changed: &changed, unchanged: &unchanged, skipped: &skipped, - skipped_db_conn: &skipped_db_conn, messages: sender, remaining_diagnostics: &remaining_diagnostics, evaluated_paths: RwLock::default(), }, ); // wait for the main thread to finish - let (diagnostics, user_hints) = handler.join().unwrap(); + let diagnostics = handler.join().unwrap(); - (elapsed, evaluated_paths, diagnostics, user_hints) + (elapsed, evaluated_paths, diagnostics) }); let errors = printer.errors(); @@ -127,19 +124,6 @@ pub(crate) fn traverse( let suggested_fixes_skipped = printer.skipped_fixes(); let diagnostics_not_printed = printer.not_printed_diagnostics(); - if duration.as_secs() >= 2 { - user_hints.push(format!( - "The traversal took longer than expected ({}s). Consider using the `--skip-db` option if your Postgres connection is slow.", - duration.as_secs() - )); - } - - if skipped_db_conn.load(Ordering::Relaxed) { - user_hints.push(format!( - "Skipped all checks requiring database connections.", - )); - } - Ok(TraverseResult { summary: TraversalSummary { changed, @@ -154,7 +138,6 @@ pub(crate) fn traverse( }, evaluated_paths, diagnostics, - user_hints, }) } @@ -306,15 +289,10 @@ impl<'ctx> DiagnosticsPrinter<'ctx> { should_print } - fn run( - &self, - receiver: Receiver, - interner: Receiver, - ) -> (Vec, Vec) { + fn run(&self, receiver: Receiver, interner: Receiver) -> Vec { let mut paths: FxHashSet = FxHashSet::default(); let mut diagnostics_to_print = vec![]; - let mut hints_to_print = vec![]; while let Ok(msg) = receiver.recv() { match msg { @@ -329,10 +307,6 @@ impl<'ctx> DiagnosticsPrinter<'ctx> { self.errors.fetch_add(1, Ordering::Relaxed); } - Message::Hint(hint) => { - hints_to_print.push(hint); - } - Message::Error(mut err) => { let location = err.location(); if self.should_skip_diagnostic(err.severity(), err.tags()) { @@ -409,7 +383,7 @@ impl<'ctx> DiagnosticsPrinter<'ctx> { } } - (diagnostics_to_print, hints_to_print) + diagnostics_to_print } } @@ -431,8 +405,6 @@ pub(crate) struct TraversalOptions<'ctx, 'app> { matches: &'ctx AtomicUsize, /// Shared atomic counter storing the number of skipped files skipped: &'ctx AtomicUsize, - /// Shared atomic bool tracking whether we used a DB connection - skipped_db_conn: &'ctx AtomicBool, /// Channel sending messages to the display thread pub(crate) messages: Sender, /// The approximate number of diagnostics the console will print before @@ -464,10 +436,6 @@ impl TraversalOptions<'_, '_> { self.messages.send(msg.into()).ok(); } - pub(crate) fn set_skipped_db_conn(&self, has_skipped: bool) { - self.skipped_db_conn.store(has_skipped, Ordering::Relaxed); - } - pub(crate) fn protected_file(&self, pglsp_path: &PgLspPath) { self.push_diagnostic( WorkspaceError::protected_file(pglsp_path.display().to_string()).into(), diff --git a/crates/pg_cli/src/reporter/github.rs b/crates/pg_cli/src/reporter/github.rs index ecb70b03..6b1588b1 100644 --- a/crates/pg_cli/src/reporter/github.rs +++ b/crates/pg_cli/src/reporter/github.rs @@ -3,18 +3,14 @@ use pg_console::{markup, Console, ConsoleExt}; use pg_diagnostics::PrintGitHubDiagnostic; use std::io; -use super::UserHintsPayload; - pub(crate) struct GithubReporter { pub(crate) diagnostics_payload: DiagnosticsPayload, pub(crate) execution: Execution, - pub(crate) user_hints: UserHintsPayload, } impl Reporter for GithubReporter { fn write(self, visitor: &mut dyn ReporterVisitor) -> io::Result<()> { visitor.report_diagnostics(&self.execution, self.diagnostics_payload)?; - visitor.report_user_hints(&self.execution, self.user_hints)?; Ok(()) } } @@ -46,15 +42,4 @@ impl ReporterVisitor for GithubReporterVisitor<'_> { Ok(()) } - - fn report_user_hints( - &mut self, - _execution: &Execution, - payload: super::UserHintsPayload, - ) -> io::Result<()> { - for hint in payload.hints { - self.0.log(markup! {{hint}}); - } - Ok(()) - } } diff --git a/crates/pg_cli/src/reporter/gitlab.rs b/crates/pg_cli/src/reporter/gitlab.rs index 6fe580b3..473f4f63 100644 --- a/crates/pg_cli/src/reporter/gitlab.rs +++ b/crates/pg_cli/src/reporter/gitlab.rs @@ -12,18 +12,14 @@ use std::{ path::{Path, PathBuf}, }; -use super::UserHintsPayload; - pub struct GitLabReporter { pub(crate) execution: Execution, pub(crate) diagnostics: DiagnosticsPayload, - pub(crate) user_hints: UserHintsPayload, } impl Reporter for GitLabReporter { fn write(self, visitor: &mut dyn ReporterVisitor) -> std::io::Result<()> { visitor.report_diagnostics(&self.execution, self.diagnostics)?; - visitor.report_user_hints(&self.execution, self.user_hints)?; Ok(()) } } @@ -76,17 +72,6 @@ impl ReporterVisitor for GitLabReporterVisitor<'_> { self.console.log(markup!({ diagnostics })); Ok(()) } - - fn report_user_hints( - &mut self, - _execution: &Execution, - payload: super::UserHintsPayload, - ) -> std::io::Result<()> { - for hint in payload.hints { - self.console.log(markup! {{hint}}); - } - Ok(()) - } } struct GitLabDiagnostics<'a>( diff --git a/crates/pg_cli/src/reporter/junit.rs b/crates/pg_cli/src/reporter/junit.rs index 2fe0fb30..c10059fc 100644 --- a/crates/pg_cli/src/reporter/junit.rs +++ b/crates/pg_cli/src/reporter/junit.rs @@ -6,20 +6,16 @@ use quick_junit::{NonSuccessKind, Report, TestCase, TestCaseStatus, TestSuite}; use std::fmt::{Display, Formatter}; use std::io; -use super::UserHintsPayload; - pub(crate) struct JunitReporter { pub(crate) diagnostics_payload: DiagnosticsPayload, pub(crate) execution: Execution, pub(crate) summary: TraversalSummary, - pub(crate) user_hints: UserHintsPayload, } impl Reporter for JunitReporter { fn write(self, visitor: &mut dyn ReporterVisitor) -> io::Result<()> { visitor.report_summary(&self.execution, self.summary)?; visitor.report_diagnostics(&self.execution, self.diagnostics_payload)?; - visitor.report_user_hints(&self.execution, self.user_hints)?; Ok(()) } } @@ -122,17 +118,4 @@ impl ReporterVisitor for JunitReporterVisitor<'_> { Ok(()) } - - fn report_user_hints( - &mut self, - _execution: &Execution, - payload: super::UserHintsPayload, - ) -> io::Result<()> { - for hint in payload.hints { - self.1.log(markup! { - {hint} - }); - } - Ok(()) - } } diff --git a/crates/pg_cli/src/reporter/mod.rs b/crates/pg_cli/src/reporter/mod.rs index 25593056..adc7c023 100644 --- a/crates/pg_cli/src/reporter/mod.rs +++ b/crates/pg_cli/src/reporter/mod.rs @@ -17,10 +17,6 @@ pub struct DiagnosticsPayload { pub diagnostic_level: Severity, } -pub struct UserHintsPayload { - pub hints: Vec, -} - /// A type that holds the result of the traversal #[derive(Debug, Default, Serialize, Copy, Clone)] pub struct TraversalSummary { @@ -64,11 +60,4 @@ pub trait ReporterVisitor { execution: &Execution, payload: DiagnosticsPayload, ) -> io::Result<()>; - - /// Writes a diagnostics - fn report_user_hints( - &mut self, - execution: &Execution, - payload: UserHintsPayload, - ) -> io::Result<()>; } diff --git a/crates/pg_cli/src/reporter/terminal.rs b/crates/pg_cli/src/reporter/terminal.rs index 6cfbb90a..5fd493a2 100644 --- a/crates/pg_cli/src/reporter/terminal.rs +++ b/crates/pg_cli/src/reporter/terminal.rs @@ -10,14 +10,11 @@ use std::collections::BTreeSet; use std::io; use std::time::Duration; -use super::UserHintsPayload; - pub(crate) struct ConsoleReporter { pub(crate) summary: TraversalSummary, pub(crate) diagnostics_payload: DiagnosticsPayload, pub(crate) execution: Execution, pub(crate) evaluated_paths: BTreeSet, - pub(crate) user_hints_payload: UserHintsPayload, } impl Reporter for ConsoleReporter { @@ -25,7 +22,6 @@ impl Reporter for ConsoleReporter { let verbose = self.diagnostics_payload.verbose; visitor.report_diagnostics(&self.execution, self.diagnostics_payload)?; visitor.report_summary(&self.execution, self.summary)?; - visitor.report_user_hints(&self.execution, self.user_hints_payload)?; if verbose { visitor.report_handled_paths(self.evaluated_paths)?; } @@ -119,18 +115,6 @@ impl ReporterVisitor for ConsoleReporterVisitor<'_> { Ok(()) } - - fn report_user_hints( - &mut self, - _execution: &Execution, - payload: UserHintsPayload, - ) -> io::Result<()> { - for hint in payload.hints { - self.0.log(markup! {{hint}}); - } - self.0.log(markup! {{"\n"}}); - Ok(()) - } } struct Files(usize); From 3d8f1b5467d6a126fd899a38582245a9432354df Mon Sep 17 00:00:00 2001 From: Julian Date: Sun, 9 Feb 2025 09:02:02 +0100 Subject: [PATCH 14/16] that too --- crates/pg_workspace/src/workspace.rs | 1 - crates/pg_workspace/src/workspace/server.rs | 9 +++------ 2 files changed, 3 insertions(+), 7 deletions(-) diff --git a/crates/pg_workspace/src/workspace.rs b/crates/pg_workspace/src/workspace.rs index ba04d961..16e6d135 100644 --- a/crates/pg_workspace/src/workspace.rs +++ b/crates/pg_workspace/src/workspace.rs @@ -53,7 +53,6 @@ pub struct PullDiagnosticsResult { pub diagnostics: Vec, pub errors: usize, pub skipped_diagnostics: u64, - pub skipped_db_checks: bool, } #[derive(Debug, Clone, Copy, serde::Serialize, serde::Deserialize, PartialEq)] diff --git a/crates/pg_workspace/src/workspace/server.rs b/crates/pg_workspace/src/workspace/server.rs index 5d479a85..a8734da2 100644 --- a/crates/pg_workspace/src/workspace/server.rs +++ b/crates/pg_workspace/src/workspace/server.rs @@ -300,14 +300,12 @@ impl Workspace for WorkspaceServer { let mut diagnostics: Vec = vec![]; - let maybe_pool = self + if let Some(pool) = self .connection .read() .expect("DbConnection RwLock panicked") - .get_pool(); - - let skipped_db_checks = maybe_pool.is_none(); - if let Some(pool) = maybe_pool { + .get_pool() + { let typecheck_params: Vec<_> = doc .iter_statements_with_text_and_range() .map(|(stmt, range, text)| { @@ -404,7 +402,6 @@ impl Workspace for WorkspaceServer { diagnostics, errors, skipped_diagnostics: 0, - skipped_db_checks, }) } From eb6feaf39adbb2f42f1a5a012a999af9dddd48a0 Mon Sep 17 00:00:00 2001 From: Julian Date: Sun, 9 Feb 2025 09:02:33 +0100 Subject: [PATCH 15/16] cool --- crates/pg_cli/src/execute/process_file/check.rs | 4 ---- 1 file changed, 4 deletions(-) diff --git a/crates/pg_cli/src/execute/process_file/check.rs b/crates/pg_cli/src/execute/process_file/check.rs index 9795d847..866fa740 100644 --- a/crates/pg_cli/src/execute/process_file/check.rs +++ b/crates/pg_cli/src/execute/process_file/check.rs @@ -42,10 +42,6 @@ pub(crate) fn check_with_guard<'ctx>( category!("check"), )?; - if pull_diagnostics_result.skipped_db_checks { - ctx.set_skipped_db_conn(true); - } - let no_diagnostics = pull_diagnostics_result.diagnostics.is_empty() && pull_diagnostics_result.skipped_diagnostics == 0; From fbaab3bc8266763464131c548635aee9baa540de Mon Sep 17 00:00:00 2001 From: Julian Date: Sun, 9 Feb 2025 18:28:46 +0100 Subject: [PATCH 16/16] debug representation --- crates/pg_configuration/src/database.rs | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/crates/pg_configuration/src/database.rs b/crates/pg_configuration/src/database.rs index 57752bc9..2feb0330 100644 --- a/crates/pg_configuration/src/database.rs +++ b/crates/pg_configuration/src/database.rs @@ -28,8 +28,8 @@ pub struct DatabaseConfiguration { pub database: String, /// The connection timeout in seconds. - #[partial(bpaf(long("conn_timeout")))] - pub conn_timeout_secs: Option, + #[partial(bpaf(long("conn_timeout_secs"), fallback(Some(10)), debug_fallback))] + pub conn_timeout_secs: u16, } impl Default for DatabaseConfiguration { @@ -40,7 +40,7 @@ impl Default for DatabaseConfiguration { username: "postgres".to_string(), password: "postgres".to_string(), database: "postgres".to_string(), - conn_timeout_secs: Some(10), + conn_timeout_secs: 10, } } }