mirror of
https://github.com/Bubberstation/Bubberstation.git
synced 2025-12-10 01:34:01 +00:00
* ONLY SKYRAT CHANGES * ACTUALLY SKYRAT CHANGES * yolo, revert later * Update alternate_byond_versions.txt Co-authored-by: AnturK <AnturK@users.noreply.github.com>
576 lines
19 KiB
Plaintext
576 lines
19 KiB
Plaintext
SUBSYSTEM_DEF(dbcore)
|
|
name = "Database"
|
|
flags = SS_TICKER
|
|
wait = 10 // Not seconds because we're running on SS_TICKER
|
|
runlevels = RUNLEVEL_INIT|RUNLEVEL_LOBBY|RUNLEVELS_DEFAULT
|
|
init_order = INIT_ORDER_DBCORE
|
|
priority = FIRE_PRIORITY_DATABASE
|
|
|
|
var/failed_connection_timeout = 0
|
|
|
|
var/schema_mismatch = 0
|
|
var/db_minor = 0
|
|
var/db_major = 0
|
|
var/failed_connections = 0
|
|
|
|
var/last_error
|
|
|
|
var/max_concurrent_queries = 25
|
|
|
|
/// Number of all queries, reset to 0 when logged in SStime_track. Used by SStime_track
|
|
var/all_queries_num = 0
|
|
/// Number of active queries, reset to 0 when logged in SStime_track. Used by SStime_track
|
|
var/queries_active_num = 0
|
|
/// Number of standby queries, reset to 0 when logged in SStime_track. Used by SStime_track
|
|
var/queries_standby_num = 0
|
|
|
|
/// All the current queries that exist.
|
|
var/list/all_queries = list()
|
|
/// Queries being checked for timeouts.
|
|
var/list/processing_queries
|
|
|
|
/// Queries currently being handled by database driver
|
|
var/list/datum/db_query/queries_active = list()
|
|
/// Queries pending execution that will be handled this controller firing
|
|
var/list/datum/db_query/queries_new
|
|
/// Queries pending execution, mapped to complete arguments
|
|
var/list/datum/db_query/queries_standby = list()
|
|
/// Queries left to handle during controller firing
|
|
var/list/datum/db_query/queries_current
|
|
|
|
var/connection // Arbitrary handle returned from rust_g.
|
|
|
|
/datum/controller/subsystem/dbcore/Initialize()
|
|
//We send warnings to the admins during subsystem init, as the clients will be New'd and messages
|
|
//will queue properly with goonchat
|
|
switch(schema_mismatch)
|
|
if(1)
|
|
message_admins("Database schema ([db_major].[db_minor]) doesn't match the latest schema version ([DB_MAJOR_VERSION].[DB_MINOR_VERSION]), this may lead to undefined behaviour or errors")
|
|
if(2)
|
|
message_admins("Could not get schema version from database")
|
|
|
|
return SS_INIT_SUCCESS
|
|
|
|
/datum/controller/subsystem/dbcore/stat_entry(msg)
|
|
msg = "P:[length(all_queries)]|Active:[length(queries_active)]|Standby:[length(queries_standby)]"
|
|
return ..()
|
|
|
|
/// Resets the tracking numbers on the subsystem. Used by SStime_track.
|
|
/datum/controller/subsystem/dbcore/proc/reset_tracking()
|
|
all_queries_num = 0
|
|
queries_active_num = 0
|
|
queries_standby_num = 0
|
|
|
|
/datum/controller/subsystem/dbcore/fire(resumed = FALSE)
|
|
if(!IsConnected())
|
|
return
|
|
|
|
if(!resumed)
|
|
queries_new = null
|
|
if(!length(queries_active) && !length(queries_standby) && !length(all_queries))
|
|
processing_queries = null
|
|
queries_current = null
|
|
return
|
|
queries_current = queries_active.Copy()
|
|
processing_queries = all_queries.Copy()
|
|
|
|
while(length(processing_queries))
|
|
var/datum/db_query/query = popleft(processing_queries)
|
|
if(world.time - query.last_activity_time > (5 MINUTES))
|
|
stack_trace("Found undeleted query, check the sql.log for the undeleted query and add a delete call to the query datum.")
|
|
log_sql("Undeleted query: \"[query.sql]\" LA: [query.last_activity] LAT: [query.last_activity_time]")
|
|
qdel(query)
|
|
if(MC_TICK_CHECK)
|
|
return
|
|
|
|
// First handle the already running queries
|
|
while(length(queries_current))
|
|
var/datum/db_query/query = popleft(queries_current)
|
|
if(!process_query(query))
|
|
queries_active -= query
|
|
if(MC_TICK_CHECK)
|
|
return
|
|
|
|
// Then strap on extra new queries as possible
|
|
if(isnull(queries_new))
|
|
if(!length(queries_standby))
|
|
return
|
|
queries_new = queries_standby.Copy(1, min(length(queries_standby), max_concurrent_queries) + 1)
|
|
|
|
while(length(queries_new) && length(queries_active) < max_concurrent_queries)
|
|
var/datum/db_query/query = popleft(queries_new)
|
|
queries_standby.Remove(query)
|
|
create_active_query(query)
|
|
if(MC_TICK_CHECK)
|
|
return
|
|
|
|
/// Helper proc for handling queued new queries
|
|
/datum/controller/subsystem/dbcore/proc/create_active_query(datum/db_query/query)
|
|
PRIVATE_PROC(TRUE)
|
|
SHOULD_NOT_SLEEP(TRUE)
|
|
if(IsAdminAdvancedProcCall())
|
|
return FALSE
|
|
run_query(query)
|
|
queries_active_num++
|
|
queries_active += query
|
|
return query
|
|
|
|
/datum/controller/subsystem/dbcore/proc/process_query(datum/db_query/query)
|
|
PRIVATE_PROC(TRUE)
|
|
SHOULD_NOT_SLEEP(TRUE)
|
|
if(IsAdminAdvancedProcCall())
|
|
return FALSE
|
|
if(QDELETED(query))
|
|
return FALSE
|
|
if(query.process(wait))
|
|
queries_active -= query
|
|
return FALSE
|
|
return TRUE
|
|
|
|
/datum/controller/subsystem/dbcore/proc/run_query_sync(datum/db_query/query)
|
|
if(IsAdminAdvancedProcCall())
|
|
return
|
|
run_query(query)
|
|
UNTIL(query.process())
|
|
return query
|
|
|
|
/datum/controller/subsystem/dbcore/proc/run_query(datum/db_query/query)
|
|
if(IsAdminAdvancedProcCall())
|
|
return
|
|
query.job_id = rustg_sql_query_async(connection, query.sql, json_encode(query.arguments))
|
|
|
|
/datum/controller/subsystem/dbcore/proc/queue_query(datum/db_query/query)
|
|
if(IsAdminAdvancedProcCall())
|
|
return
|
|
queries_standby_num++
|
|
queries_standby |= query
|
|
|
|
/datum/controller/subsystem/dbcore/Recover()
|
|
connection = SSdbcore.connection
|
|
|
|
/datum/controller/subsystem/dbcore/Shutdown()
|
|
//This is as close as we can get to the true round end before Disconnect() without changing where it's called, defeating the reason this is a subsystem
|
|
if(SSdbcore.Connect())
|
|
for(var/datum/db_query/query in queries_current)
|
|
run_query(query)
|
|
|
|
// SKYRAT EDIT START - SQL-based logging
|
|
for(var/table in queued_log_entries_by_table)
|
|
MassInsert(table, rows = queued_log_entries_by_table[table], duplicate_key = FALSE, ignore_errors = FALSE, delayed = FALSE, warn = FALSE, async = TRUE, special_columns = null)
|
|
// SKYRAT EDIT END
|
|
|
|
var/datum/db_query/query_round_shutdown = SSdbcore.NewQuery(
|
|
"UPDATE [format_table_name("round")] SET shutdown_datetime = Now(), end_state = :end_state WHERE id = :round_id",
|
|
list("end_state" = SSticker.end_state, "round_id" = GLOB.round_id)
|
|
)
|
|
query_round_shutdown.Execute()
|
|
qdel(query_round_shutdown)
|
|
if(IsConnected())
|
|
Disconnect()
|
|
|
|
//nu
|
|
/datum/controller/subsystem/dbcore/can_vv_get(var_name)
|
|
if(var_name == NAMEOF(src, connection))
|
|
return FALSE
|
|
if(var_name == NAMEOF(src, all_queries))
|
|
return FALSE
|
|
if(var_name == NAMEOF(src, queries_active))
|
|
return FALSE
|
|
if(var_name == NAMEOF(src, queries_new))
|
|
return FALSE
|
|
if(var_name == NAMEOF(src, queries_standby))
|
|
return FALSE
|
|
if(var_name == NAMEOF(src, queries_active))
|
|
return FALSE
|
|
|
|
return ..()
|
|
|
|
/datum/controller/subsystem/dbcore/vv_edit_var(var_name, var_value)
|
|
if(var_name == NAMEOF(src, connection))
|
|
return FALSE
|
|
if(var_name == NAMEOF(src, all_queries))
|
|
return FALSE
|
|
if(var_name == NAMEOF(src, queries_active))
|
|
return FALSE
|
|
if(var_name == NAMEOF(src, queries_new))
|
|
return FALSE
|
|
if(var_name == NAMEOF(src, queries_standby))
|
|
return FALSE
|
|
if(var_name == NAMEOF(src, queries_active))
|
|
return FALSE
|
|
return ..()
|
|
|
|
/datum/controller/subsystem/dbcore/proc/Connect()
|
|
if(IsConnected())
|
|
return TRUE
|
|
|
|
if(failed_connection_timeout <= world.time) //it's been more than 5 seconds since we failed to connect, reset the counter
|
|
failed_connections = 0
|
|
|
|
if(failed_connections > 5) //If it failed to establish a connection more than 5 times in a row, don't bother attempting to connect for 5 seconds.
|
|
failed_connection_timeout = world.time + 50
|
|
return FALSE
|
|
|
|
if(!CONFIG_GET(flag/sql_enabled))
|
|
return FALSE
|
|
|
|
var/user = CONFIG_GET(string/feedback_login)
|
|
var/pass = CONFIG_GET(string/feedback_password)
|
|
var/db = CONFIG_GET(string/feedback_database)
|
|
var/address = CONFIG_GET(string/address)
|
|
var/port = CONFIG_GET(number/port)
|
|
var/timeout = max(CONFIG_GET(number/async_query_timeout), CONFIG_GET(number/blocking_query_timeout))
|
|
var/thread_limit = CONFIG_GET(number/bsql_thread_limit)
|
|
|
|
max_concurrent_queries = CONFIG_GET(number/max_concurrent_queries)
|
|
|
|
var/result = json_decode(rustg_sql_connect_pool(json_encode(list(
|
|
"host" = address,
|
|
"port" = port,
|
|
"user" = user,
|
|
"pass" = pass,
|
|
"db_name" = db,
|
|
"read_timeout" = timeout,
|
|
"write_timeout" = timeout,
|
|
"max_threads" = thread_limit,
|
|
))))
|
|
. = (result["status"] == "ok")
|
|
if (.)
|
|
connection = result["handle"]
|
|
else
|
|
connection = null
|
|
last_error = result["data"]
|
|
log_sql("Connect() failed | [last_error]")
|
|
++failed_connections
|
|
|
|
/datum/controller/subsystem/dbcore/proc/CheckSchemaVersion()
|
|
if(CONFIG_GET(flag/sql_enabled))
|
|
if(Connect())
|
|
log_world("Database connection established.")
|
|
var/datum/db_query/query_db_version = NewQuery("SELECT major, minor FROM [format_table_name("schema_revision")] ORDER BY date DESC LIMIT 1")
|
|
query_db_version.Execute()
|
|
if(query_db_version.NextRow())
|
|
db_major = text2num(query_db_version.item[1])
|
|
db_minor = text2num(query_db_version.item[2])
|
|
if(db_major != DB_MAJOR_VERSION || db_minor != DB_MINOR_VERSION)
|
|
schema_mismatch = 1 // flag admin message about mismatch
|
|
log_sql("Database schema ([db_major].[db_minor]) doesn't match the latest schema version ([DB_MAJOR_VERSION].[DB_MINOR_VERSION]), this may lead to undefined behaviour or errors")
|
|
else
|
|
schema_mismatch = 2 //flag admin message about no schema version
|
|
log_sql("Could not get schema version from database")
|
|
qdel(query_db_version)
|
|
else
|
|
log_sql("Your server failed to establish a connection with the database.")
|
|
else
|
|
log_sql("Database is not enabled in configuration.")
|
|
|
|
/datum/controller/subsystem/dbcore/proc/SetRoundID()
|
|
if(!Connect())
|
|
return
|
|
var/datum/db_query/query_round_initialize = SSdbcore.NewQuery(/* SKYRAT EDIT CHANGE - MULTISERVER */
|
|
"INSERT INTO [format_table_name("round")] (initialize_datetime, server_name, server_ip, server_port) VALUES (Now(), :server_name, INET_ATON(:internet_address), :port)",
|
|
list("server_name" = CONFIG_GET(string/serversqlname), "internet_address" = world.internet_address || "0", "port" = "[world.port]") // SKYRAT EDIT CHANGE - MULTISERVER
|
|
)
|
|
query_round_initialize.Execute(async = FALSE)
|
|
GLOB.round_id = "[query_round_initialize.last_insert_id]"
|
|
qdel(query_round_initialize)
|
|
|
|
/datum/controller/subsystem/dbcore/proc/SetRoundStart()
|
|
if(!Connect())
|
|
return
|
|
var/datum/db_query/query_round_start = SSdbcore.NewQuery(
|
|
"UPDATE [format_table_name("round")] SET start_datetime = Now() WHERE id = :round_id",
|
|
list("round_id" = GLOB.round_id)
|
|
)
|
|
query_round_start.Execute()
|
|
qdel(query_round_start)
|
|
|
|
/datum/controller/subsystem/dbcore/proc/SetRoundEnd()
|
|
if(!Connect())
|
|
return
|
|
var/datum/db_query/query_round_end = SSdbcore.NewQuery(
|
|
"UPDATE [format_table_name("round")] SET end_datetime = Now(), game_mode_result = :game_mode_result, station_name = :station_name WHERE id = :round_id",
|
|
list("game_mode_result" = SSticker.mode_result, "station_name" = station_name(), "round_id" = GLOB.round_id)
|
|
)
|
|
query_round_end.Execute()
|
|
qdel(query_round_end)
|
|
|
|
/datum/controller/subsystem/dbcore/proc/Disconnect()
|
|
failed_connections = 0
|
|
if (connection)
|
|
rustg_sql_disconnect_pool(connection)
|
|
connection = null
|
|
|
|
/datum/controller/subsystem/dbcore/proc/IsConnected()
|
|
if (!CONFIG_GET(flag/sql_enabled))
|
|
return FALSE
|
|
if (!connection)
|
|
return FALSE
|
|
return json_decode(rustg_sql_connected(connection))["status"] == "online"
|
|
|
|
/datum/controller/subsystem/dbcore/proc/ErrorMsg()
|
|
if(!CONFIG_GET(flag/sql_enabled))
|
|
return "Database disabled by configuration"
|
|
return last_error
|
|
|
|
/datum/controller/subsystem/dbcore/proc/ReportError(error)
|
|
last_error = error
|
|
|
|
/datum/controller/subsystem/dbcore/proc/NewQuery(sql_query, arguments)
|
|
if(IsAdminAdvancedProcCall())
|
|
log_admin_private("ERROR: Advanced admin proc call led to sql query: [sql_query]. Query has been blocked")
|
|
message_admins("ERROR: Advanced admin proc call led to sql query. Query has been blocked")
|
|
return FALSE
|
|
return new /datum/db_query(connection, sql_query, arguments)
|
|
|
|
/** QuerySelect
|
|
Run a list of query datums in parallel, blocking until they all complete.
|
|
* queries - List of queries or single query datum to run.
|
|
* warn - Controls rather warn_execute() or Execute() is called.
|
|
* qdel - If you don't care about the result or checking for errors, you can have the queries be deleted afterwards.
|
|
This can be combined with invoke_async as a way of running queries async without having to care about waiting for them to finish so they can be deleted.
|
|
*/
|
|
/datum/controller/subsystem/dbcore/proc/QuerySelect(list/queries, warn = FALSE, qdel = FALSE)
|
|
if (!islist(queries))
|
|
if (!istype(queries, /datum/db_query))
|
|
CRASH("Invalid query passed to QuerySelect: [queries]")
|
|
queries = list(queries)
|
|
else
|
|
queries = queries.Copy() //we don't want to hide bugs in the parent caller by removing invalid values from this list.
|
|
|
|
for (var/datum/db_query/query as anything in queries)
|
|
if (!istype(query))
|
|
queries -= query
|
|
stack_trace("Invalid query passed to QuerySelect: `[query]` [REF(query)]")
|
|
continue
|
|
|
|
if (warn)
|
|
INVOKE_ASYNC(query, TYPE_PROC_REF(/datum/db_query, warn_execute))
|
|
else
|
|
INVOKE_ASYNC(query, TYPE_PROC_REF(/datum/db_query, Execute))
|
|
|
|
for (var/datum/db_query/query as anything in queries)
|
|
query.sync()
|
|
if (qdel)
|
|
qdel(query)
|
|
|
|
|
|
|
|
/*
|
|
Takes a list of rows (each row being an associated list of column => value) and inserts them via a single mass query.
|
|
Rows missing columns present in other rows will resolve to SQL NULL
|
|
You are expected to do your own escaping of the data, and expected to provide your own quotes for strings.
|
|
The duplicate_key arg can be true to automatically generate this part of the query
|
|
or set to a string that is appended to the end of the query
|
|
Ignore_errors instructes mysql to continue inserting rows if some of them have errors.
|
|
the erroneous row(s) aren't inserted and there isn't really any way to know why or why errored
|
|
Delayed insert mode was removed in mysql 7 and only works with MyISAM type tables,
|
|
It was included because it is still supported in mariadb.
|
|
It does not work with duplicate_key and the mysql server ignores it in those cases
|
|
*/
|
|
/datum/controller/subsystem/dbcore/proc/MassInsert(table, list/rows, duplicate_key = FALSE, ignore_errors = FALSE, delayed = FALSE, warn = FALSE, async = TRUE, special_columns = null)
|
|
if (!table || !rows || !istype(rows))
|
|
return
|
|
|
|
// Prepare column list
|
|
var/list/columns = list()
|
|
var/list/has_question_mark = list()
|
|
for (var/list/row in rows)
|
|
for (var/column in row)
|
|
columns[column] = "?"
|
|
has_question_mark[column] = TRUE
|
|
for (var/column in special_columns)
|
|
columns[column] = special_columns[column]
|
|
has_question_mark[column] = findtext(special_columns[column], "?")
|
|
|
|
// Prepare SQL query full of placeholders
|
|
var/list/query_parts = list("INSERT")
|
|
if (delayed)
|
|
query_parts += " DELAYED"
|
|
if (ignore_errors)
|
|
query_parts += " IGNORE"
|
|
query_parts += " INTO "
|
|
query_parts += table
|
|
query_parts += "\n([columns.Join(", ")])\nVALUES"
|
|
|
|
var/list/arguments = list()
|
|
var/has_row = FALSE
|
|
for (var/list/row in rows)
|
|
if (has_row)
|
|
query_parts += ","
|
|
query_parts += "\n ("
|
|
var/has_col = FALSE
|
|
for (var/column in columns)
|
|
if (has_col)
|
|
query_parts += ", "
|
|
if (has_question_mark[column])
|
|
var/name = "p[arguments.len]"
|
|
query_parts += replacetext(columns[column], "?", ":[name]")
|
|
arguments[name] = row[column]
|
|
else
|
|
query_parts += columns[column]
|
|
has_col = TRUE
|
|
query_parts += ")"
|
|
has_row = TRUE
|
|
|
|
if (duplicate_key == TRUE)
|
|
var/list/column_list = list()
|
|
for (var/column in columns)
|
|
column_list += "[column] = VALUES([column])"
|
|
query_parts += "\nON DUPLICATE KEY UPDATE [column_list.Join(", ")]"
|
|
else if (duplicate_key != FALSE)
|
|
query_parts += duplicate_key
|
|
|
|
var/datum/db_query/Query = NewQuery(query_parts.Join(), arguments)
|
|
if (warn)
|
|
. = Query.warn_execute(async)
|
|
else
|
|
. = Query.Execute(async)
|
|
qdel(Query)
|
|
|
|
/datum/db_query
|
|
// Inputs
|
|
var/connection
|
|
var/sql
|
|
var/arguments
|
|
|
|
var/datum/callback/success_callback
|
|
var/datum/callback/fail_callback
|
|
|
|
// Status information
|
|
/// Current status of the query.
|
|
var/status
|
|
/// Job ID of the query passed by rustg.
|
|
var/job_id
|
|
var/last_error
|
|
var/last_activity
|
|
var/last_activity_time
|
|
|
|
// Output
|
|
var/list/list/rows
|
|
var/next_row_to_take = 1
|
|
var/affected
|
|
var/last_insert_id
|
|
|
|
var/list/item //list of data values populated by NextRow()
|
|
|
|
/datum/db_query/New(connection, sql, arguments)
|
|
SSdbcore.all_queries += src
|
|
SSdbcore.all_queries_num++
|
|
Activity("Created")
|
|
item = list()
|
|
|
|
src.connection = connection
|
|
src.sql = sql
|
|
src.arguments = arguments
|
|
|
|
/datum/db_query/Destroy()
|
|
Close()
|
|
SSdbcore.all_queries -= src
|
|
SSdbcore.queries_standby -= src
|
|
SSdbcore.queries_active -= src
|
|
return ..()
|
|
|
|
/datum/db_query/CanProcCall(proc_name)
|
|
//fuck off kevinz
|
|
return FALSE
|
|
|
|
/datum/db_query/proc/Activity(activity)
|
|
last_activity = activity
|
|
last_activity_time = world.time
|
|
|
|
/datum/db_query/proc/warn_execute(async = TRUE)
|
|
. = Execute(async)
|
|
if(!.)
|
|
to_chat(usr, span_danger("A SQL error occurred during this operation, check the server logs."))
|
|
|
|
/datum/db_query/proc/Execute(async = TRUE, log_error = TRUE)
|
|
Activity("Execute")
|
|
if(status == DB_QUERY_STARTED)
|
|
CRASH("Attempted to start a new query while waiting on the old one")
|
|
|
|
if(!SSdbcore.IsConnected())
|
|
last_error = "No connection!"
|
|
return FALSE
|
|
|
|
var/start_time
|
|
if(!async)
|
|
start_time = REALTIMEOFDAY
|
|
Close()
|
|
status = DB_QUERY_STARTED
|
|
if(async)
|
|
if(!MC_RUNNING(SSdbcore.init_stage))
|
|
SSdbcore.run_query_sync(src)
|
|
else
|
|
SSdbcore.queue_query(src)
|
|
sync()
|
|
else
|
|
var/job_result_str = rustg_sql_query_blocking(connection, sql, json_encode(arguments))
|
|
store_data(json_decode(job_result_str))
|
|
|
|
. = (status != DB_QUERY_BROKEN)
|
|
var/timed_out = !. && findtext(last_error, "Operation timed out")
|
|
if(!. && log_error)
|
|
log_sql("[last_error] | Query used: [sql] | Arguments: [json_encode(arguments)]")
|
|
if(!async && timed_out)
|
|
log_query_debug("Query execution started at [start_time]")
|
|
log_query_debug("Query execution ended at [REALTIMEOFDAY]")
|
|
log_query_debug("Slow query timeout detected.")
|
|
log_query_debug("Query used: [sql]")
|
|
slow_query_check()
|
|
|
|
/// Sleeps until execution of the query has finished.
|
|
/datum/db_query/proc/sync()
|
|
while(status < DB_QUERY_FINISHED)
|
|
stoplag()
|
|
|
|
/datum/db_query/process(delta_time)
|
|
if(status >= DB_QUERY_FINISHED)
|
|
return
|
|
|
|
status = DB_QUERY_STARTED
|
|
var/job_result = rustg_sql_check_query(job_id)
|
|
if(job_result == RUSTG_JOB_NO_RESULTS_YET)
|
|
return
|
|
|
|
store_data(json_decode(job_result))
|
|
return TRUE
|
|
|
|
/datum/db_query/proc/store_data(result)
|
|
switch(result["status"])
|
|
if("ok")
|
|
rows = result["rows"]
|
|
affected = result["affected"]
|
|
last_insert_id = result["last_insert_id"]
|
|
status = DB_QUERY_FINISHED
|
|
return
|
|
if("err")
|
|
last_error = result["data"]
|
|
status = DB_QUERY_BROKEN
|
|
return
|
|
if("offline")
|
|
last_error = "CONNECTION OFFLINE"
|
|
status = DB_QUERY_BROKEN
|
|
return
|
|
|
|
|
|
/datum/db_query/proc/slow_query_check()
|
|
message_admins("HEY! A database query timed out. Did the server just hang? <a href='?_src_=holder;[HrefToken()];slowquery=yes'>\[YES\]</a>|<a href='?_src_=holder;[HrefToken()];slowquery=no'>\[NO\]</a>")
|
|
|
|
/datum/db_query/proc/NextRow(async = TRUE)
|
|
Activity("NextRow")
|
|
|
|
if (rows && next_row_to_take <= rows.len)
|
|
item = rows[next_row_to_take]
|
|
next_row_to_take++
|
|
return !!item
|
|
else
|
|
return FALSE
|
|
|
|
/datum/db_query/proc/ErrorMsg()
|
|
return last_error
|
|
|
|
/datum/db_query/proc/Close()
|
|
rows = null
|
|
item = null
|