mirror of
https://github.com/ParadiseSS13/Paradise.git
synced 2025-12-19 14:51:27 +00:00
530 lines
17 KiB
Plaintext
530 lines
17 KiB
Plaintext
SUBSYSTEM_DEF(garbage)
|
|
name = "Garbage"
|
|
priority = FIRE_PRIORITY_GARBAGE
|
|
wait = 2 SECONDS
|
|
flags = SS_POST_FIRE_TIMING|SS_BACKGROUND|SS_NO_INIT
|
|
runlevels = RUNLEVELS_DEFAULT | RUNLEVEL_LOBBY
|
|
init_order = INIT_ORDER_GARBAGE // AA 2020: Why does this have an init order if it has SS_NO_INIT? | AA 2022: Its used for shutdown
|
|
offline_implications = "Garbage collection is no longer functional, and objects will not be qdel'd. Immediate server restart recommended."
|
|
cpu_display = SS_CPUDISPLAY_HIGH
|
|
|
|
var/list/collection_timeout = list(GC_FILTER_QUEUE, GC_CHECK_QUEUE, GC_DEL_QUEUE) // deciseconds to wait before moving something up in the queue to the next level
|
|
|
|
//Stat tracking
|
|
var/delslasttick = 0 // number of del()'s we've done this tick
|
|
var/gcedlasttick = 0 // number of things that gc'ed last tick
|
|
var/totaldels = 0
|
|
var/totalgcs = 0
|
|
|
|
var/highest_del_time = 0
|
|
var/highest_del_tickusage = 0
|
|
|
|
var/list/pass_counts
|
|
var/list/fail_counts
|
|
|
|
var/list/items = list() // Holds our qdel_item statistics datums
|
|
|
|
//Queue
|
|
var/list/queues
|
|
|
|
#ifdef REFERENCE_TRACKING
|
|
var/list/reference_find_on_fail = list()
|
|
var/ref_search_stop = FALSE
|
|
#endif
|
|
|
|
|
|
/datum/controller/subsystem/garbage/PreInit()
|
|
InitQueues()
|
|
|
|
/datum/controller/subsystem/garbage/get_stat_details()
|
|
var/list/msg = list()
|
|
var/list/counts = list()
|
|
for(var/list/L in queues)
|
|
counts += length(L)
|
|
msg += "Q:[counts.Join(", ")] | D:[delslasttick] | S:[gcedlasttick] |"
|
|
msg += "GCR:"
|
|
if(!(delslasttick + gcedlasttick))
|
|
msg += "n/a|"
|
|
else
|
|
msg += "[round((gcedlasttick / (delslasttick + gcedlasttick)) * 100, 0.01)]% |"
|
|
|
|
msg += "TD:[totaldels] | TS:[totalgcs] |"
|
|
if(!(totaldels + totalgcs))
|
|
msg += "n/a|"
|
|
else
|
|
msg += "TGCR:[round((totalgcs / (totaldels + totalgcs)) * 100, 0.01)]% |"
|
|
msg += " P:[pass_counts.Join(",")]"
|
|
msg += " | F:[fail_counts.Join(",")]"
|
|
return msg.Join("")
|
|
|
|
/datum/controller/subsystem/garbage/get_metrics()
|
|
. = ..()
|
|
var/list/cust = list()
|
|
if((delslasttick + gcedlasttick) == 0) // Account for DIV0
|
|
cust["gcr"] = 0
|
|
else
|
|
cust["gcr"] = (gcedlasttick / (delslasttick + gcedlasttick))
|
|
cust["total_harddels"] = totaldels
|
|
cust["total_softdels"] = totalgcs
|
|
var/i = 0
|
|
for(var/list/L in queues)
|
|
i++
|
|
cust["queue_[i]"] = length(L)
|
|
|
|
.["custom"] = cust
|
|
|
|
/datum/controller/subsystem/garbage/Shutdown()
|
|
//Adds the del() log to the qdel log file
|
|
var/list/dellog = list()
|
|
|
|
//sort by how long it's wasted hard deleting
|
|
sortTim(items, GLOBAL_PROC_REF(cmp_qdel_item_time), TRUE)
|
|
for(var/path in items)
|
|
var/datum/qdel_item/I = items[path]
|
|
dellog += "Path: [path]"
|
|
if(I.failures)
|
|
dellog += "\tFailures: [I.failures]"
|
|
dellog += "\tqdel() Count: [I.qdels]"
|
|
dellog += "\tDestroy() Cost: [I.destroy_time]ms"
|
|
if(I.hard_deletes)
|
|
dellog += "\tTotal Hard Deletes [I.hard_deletes]"
|
|
dellog += "\tTime Spent Hard Deleting: [I.hard_delete_time]ms"
|
|
if(I.slept_destroy)
|
|
dellog += "\tSleeps: [I.slept_destroy]"
|
|
if(I.no_respect_force)
|
|
dellog += "\tIgnored force: [I.no_respect_force] times"
|
|
if(I.no_hint)
|
|
dellog += "\tNo hint: [I.no_hint] times"
|
|
log_qdel(dellog.Join("\n"))
|
|
|
|
/datum/controller/subsystem/garbage/fire()
|
|
//the fact that this resets its processing each fire (rather then resume where it left off) is intentional.
|
|
var/queue = GC_QUEUE_FILTER
|
|
|
|
while(state == SS_RUNNING)
|
|
switch(queue)
|
|
if(GC_QUEUE_FILTER)
|
|
HandleQueue(GC_QUEUE_FILTER)
|
|
queue = GC_QUEUE_FILTER + 1
|
|
if(GC_QUEUE_CHECK)
|
|
HandleQueue(GC_QUEUE_CHECK)
|
|
queue = GC_QUEUE_CHECK + 1
|
|
if(GC_QUEUE_HARDDELETE)
|
|
HandleQueue(GC_QUEUE_HARDDELETE)
|
|
if(state == SS_PAUSED) //make us wait again before the next run.
|
|
state = SS_RUNNING
|
|
break
|
|
|
|
/datum/controller/subsystem/garbage/proc/InitQueues()
|
|
if(isnull(queues)) // Only init the queues if they don't already exist, prevents overriding of recovered lists
|
|
queues = new(GC_QUEUE_COUNT)
|
|
pass_counts = new(GC_QUEUE_COUNT)
|
|
fail_counts = new(GC_QUEUE_COUNT)
|
|
for(var/i in 1 to GC_QUEUE_COUNT)
|
|
queues[i] = list()
|
|
pass_counts[i] = 0
|
|
fail_counts[i] = 0
|
|
|
|
#define IS_DELETED(datum, gcd_at_time) (isnull(##datum) || ##datum.gc_destroyed != gcd_at_time)
|
|
|
|
/datum/controller/subsystem/garbage/proc/HandleQueue(level = GC_QUEUE_FILTER)
|
|
if(level == GC_QUEUE_FILTER)
|
|
delslasttick = 0
|
|
gcedlasttick = 0
|
|
var/cut_off_time = world.time - collection_timeout[level] //ignore entries newer then this
|
|
var/list/queue = queues[level]
|
|
var/static/lastlevel
|
|
var/static/count = 0
|
|
if(count) //runtime last run before we could do this.
|
|
var/c = count
|
|
count = 0 //so if we runtime on the Cut, we don't try again.
|
|
var/list/lastqueue = queues[lastlevel]
|
|
lastqueue.Cut(1, c + 1)
|
|
|
|
lastlevel = level
|
|
|
|
//We do this rather then for(var/refID in queue) because that sort of for loop copies the whole list.
|
|
//Normally this isn't expensive, but the gc queue can grow to 40k items, and that gets costly/causes overrun.
|
|
for(var/i in 1 to length(queue))
|
|
var/list/L = queue[i]
|
|
if(length(L) < GC_QUEUE_ITEM_INDEX_COUNT)
|
|
count++
|
|
if(MC_TICK_CHECK)
|
|
return
|
|
continue
|
|
|
|
var/queued_at_time = L[GC_QUEUE_ITEM_QUEUE_TIME]
|
|
if(queued_at_time > cut_off_time)
|
|
break // Everything else is newer, skip them
|
|
count++
|
|
|
|
var/GCd_at_time = L[GC_QUEUE_ITEM_GCD_DESTROYED]
|
|
|
|
var/refID = L[GC_QUEUE_ITEM_REF]
|
|
var/datum/D
|
|
D = locate(refID)
|
|
|
|
if(IS_DELETED(D, GCd_at_time)) // So if something else coincidently gets the same ref, it's not deleted by mistake
|
|
++gcedlasttick
|
|
++totalgcs
|
|
pass_counts[level]++
|
|
#ifdef REFERENCE_TRACKING
|
|
reference_find_on_fail -= text_ref(D) //It's deleted we don't care anymore.
|
|
#endif
|
|
if(MC_TICK_CHECK)
|
|
return
|
|
continue
|
|
|
|
// Something's still referring to the qdel'd object.
|
|
fail_counts[level]++
|
|
|
|
#ifdef REFERENCE_TRACKING
|
|
var/ref_searching = FALSE
|
|
#endif
|
|
|
|
switch(level)
|
|
if(GC_QUEUE_CHECK)
|
|
#ifdef REFERENCE_TRACKING
|
|
if(reference_find_on_fail[text_ref(D)] && !ref_search_stop)
|
|
INVOKE_ASYNC(D, TYPE_PROC_REF(/datum, find_references))
|
|
ref_searching = TRUE
|
|
#ifdef GC_FAILURE_HARD_LOOKUP
|
|
else if(!ref_search_stop)
|
|
INVOKE_ASYNC(D, TYPE_PROC_REF(/datum, find_references))
|
|
ref_searching = TRUE
|
|
#endif
|
|
reference_find_on_fail -= text_ref(D)
|
|
#endif
|
|
var/type = D.type
|
|
var/datum/qdel_item/I = items[type]
|
|
#ifdef REFERENCE_TRACKING
|
|
log_gc("GC: -- [text_ref(D)] | [type] was unable to be GC'd --")
|
|
#endif
|
|
I.failures++
|
|
if(GC_QUEUE_HARDDELETE)
|
|
HardDelete(D)
|
|
if(MC_TICK_CHECK)
|
|
return
|
|
continue
|
|
|
|
Queue(D, level + 1)
|
|
|
|
#ifdef REFERENCE_TRACKING
|
|
if(ref_searching)
|
|
return
|
|
#endif
|
|
|
|
if(MC_TICK_CHECK)
|
|
return
|
|
if(count)
|
|
queue.Cut(1, count + 1)
|
|
count = 0
|
|
|
|
#undef IS_DELETED
|
|
|
|
/datum/controller/subsystem/garbage/proc/Queue(datum/D, level = GC_QUEUE_FILTER)
|
|
if(isnull(D))
|
|
return
|
|
if(level > GC_QUEUE_COUNT)
|
|
HardDelete(D)
|
|
return
|
|
var/queue_time = world.time
|
|
|
|
var/refid = text_ref(D)
|
|
var/static/uid = 0
|
|
if(D.gc_destroyed <= 0)
|
|
uid = WRAP(uid + 1, 1, SHORT_REAL_LIMIT - 1)
|
|
D.gc_destroyed = uid
|
|
|
|
var/list/queue = queues[level]
|
|
|
|
queue[++queue.len] = list(queue_time, refid, D.gc_destroyed) // not += for byond reasons
|
|
|
|
//this is mainly to separate things profile wise.
|
|
/datum/controller/subsystem/garbage/proc/HardDelete(datum/D)
|
|
var/time = world.timeofday
|
|
var/tick = TICK_USAGE
|
|
var/ticktime = world.time
|
|
++delslasttick
|
|
++totaldels
|
|
var/type = D.type
|
|
var/refID = text_ref(D)
|
|
|
|
del(D)
|
|
|
|
tick = (TICK_USAGE - tick + ((world.time - ticktime) / world.tick_lag * 100))
|
|
|
|
var/datum/qdel_item/I = items[type]
|
|
|
|
I.hard_deletes++
|
|
I.hard_delete_time += TICK_DELTA_TO_MS(tick)
|
|
|
|
|
|
if(tick > highest_del_tickusage)
|
|
highest_del_tickusage = tick
|
|
time = world.timeofday - time
|
|
if(!time && TICK_DELTA_TO_MS(tick) > 1)
|
|
time = TICK_DELTA_TO_MS(tick) / 100
|
|
if(time > highest_del_time)
|
|
highest_del_time = time
|
|
if(time > 10)
|
|
log_game("Error: [type]([refID]) took longer than 1 second to delete (took [time / 10] seconds to delete)")
|
|
message_admins("Error: [type]([refID]) took longer than 1 second to delete (took [time / 10] seconds to delete).")
|
|
postpone(time)
|
|
|
|
/datum/controller/subsystem/garbage/Recover()
|
|
InitQueues() //We first need to create the queues before recovering data
|
|
if(istype(SSgarbage.queues))
|
|
for(var/i in 1 to SSgarbage.queues.len)
|
|
queues[i] |= SSgarbage.queues[i]
|
|
|
|
|
|
/datum/qdel_item
|
|
var/name = ""
|
|
var/qdels = 0 //Total number of times it's passed thru qdel.
|
|
var/destroy_time = 0 //Total amount of milliseconds spent processing this type's Destroy()
|
|
var/failures = 0 //Times it was queued for soft deletion but failed to soft delete.
|
|
var/hard_deletes = 0 //Different from failures because it also includes QDEL_HINT_HARDDEL deletions
|
|
var/hard_delete_time = 0//Total amount of milliseconds spent hard deleting this type.
|
|
var/no_respect_force = 0//Number of times it's not respected force=TRUE
|
|
var/no_hint = 0 //Number of times it's not even bother to give a qdel hint
|
|
var/slept_destroy = 0 //Number of times it's slept in its destroy
|
|
|
|
/datum/qdel_item/New(mytype)
|
|
name = "[mytype]"
|
|
|
|
#ifdef REFERENCE_TRACKING
|
|
/proc/qdel_and_find_ref_if_fail(datum/thing_to_del, force = FALSE)
|
|
thing_to_del.qdel_and_find_ref_if_fail(force)
|
|
|
|
/datum/proc/qdel_and_find_ref_if_fail(force = FALSE)
|
|
SSgarbage.reference_find_on_fail[text_ref(src)] = TRUE
|
|
qdel(src, force)
|
|
|
|
#endif
|
|
|
|
// Should be treated as a replacement for the 'del' keyword.
|
|
// Datums passed to this will be given a chance to clean up references to allow the GC to collect them.
|
|
/proc/qdel(datum/D, force = FALSE, ...)
|
|
if(!istype(D))
|
|
del(D)
|
|
return
|
|
|
|
var/datum/qdel_item/I = SSgarbage.items[D.type]
|
|
if(!I)
|
|
I = SSgarbage.items[D.type] = new /datum/qdel_item(D.type)
|
|
I.qdels++
|
|
|
|
if(isnull(D.gc_destroyed))
|
|
if(SEND_SIGNAL(D, COMSIG_PARENT_PREQDELETED, force)) // Give the components a chance to prevent their parent from being deleted
|
|
return
|
|
D.gc_destroyed = GC_CURRENTLY_BEING_QDELETED
|
|
var/start_time = world.time
|
|
var/start_tick = world.tick_usage
|
|
SEND_SIGNAL(D, COMSIG_PARENT_QDELETING, force) // Let the (remaining) components know about the result of Destroy
|
|
var/hint = D.Destroy(arglist(args.Copy(2))) // Let our friend know they're about to get fucked up.
|
|
if(world.time != start_time)
|
|
I.slept_destroy++
|
|
else
|
|
I.destroy_time += TICK_USAGE_TO_MS(start_tick)
|
|
if(!D)
|
|
return
|
|
switch(hint)
|
|
if(QDEL_HINT_QUEUE) //qdel should queue the object for deletion.
|
|
SSgarbage.Queue(D)
|
|
if(QDEL_HINT_IWILLGC)
|
|
D.gc_destroyed = world.time
|
|
return
|
|
if(QDEL_HINT_LETMELIVE) //qdel should let the object live after calling destory.
|
|
if(!force)
|
|
D.gc_destroyed = null //clear the gc variable (important!)
|
|
return
|
|
// Returning LETMELIVE after being told to force destroy
|
|
// indicates the objects Destroy() does not respect force
|
|
#ifdef TESTING
|
|
if(!I.no_respect_force)
|
|
testing("WARNING: [D.type] has been force deleted, but is \
|
|
returning an immortal QDEL_HINT, indicating it does \
|
|
not respect the force flag for qdel(). It has been \
|
|
placed in the queue, further instances of this type \
|
|
will also be queued.")
|
|
#endif
|
|
I.no_respect_force++
|
|
|
|
SSgarbage.Queue(D)
|
|
if(QDEL_HINT_HARDDEL) //qdel should assume this object won't gc, and queue a hard delete
|
|
SSgarbage.Queue(D, GC_QUEUE_HARDDELETE)
|
|
if(QDEL_HINT_HARDDEL_NOW) //qdel should assume this object won't gc, and hard del it post haste.
|
|
SSgarbage.HardDelete(D)
|
|
if(QDEL_HINT_FINDREFERENCE)//qdel will, if TESTING is enabled, display all references to this object, then queue the object for deletion.
|
|
SSgarbage.Queue(D)
|
|
#ifdef REFERENCE_TRACKING
|
|
D.find_references()
|
|
#endif
|
|
if(QDEL_HINT_IFFAIL_FINDREFERENCE)
|
|
SSgarbage.Queue(D)
|
|
#ifdef REFERENCE_TRACKING
|
|
SSgarbage.reference_find_on_fail[text_ref(D)] = TRUE
|
|
#endif
|
|
else
|
|
#ifdef REFERENCE_TRACKING
|
|
if(!I.no_hint)
|
|
log_gc("WARNING: [D.type] is not returning a qdel hint. It is being placed in the queue. Further instances of this type will also be queued.")
|
|
#endif
|
|
I.no_hint++
|
|
SSgarbage.Queue(D)
|
|
else if(D.gc_destroyed == GC_CURRENTLY_BEING_QDELETED)
|
|
CRASH("[D.type] destroy proc was called multiple times, likely due to a qdel loop in the Destroy logic")
|
|
|
|
#ifdef REFERENCE_TRACKING
|
|
|
|
/datum/proc/find_refs()
|
|
set category = "Debug"
|
|
set name = "Find References"
|
|
|
|
if(!check_rights(R_DEBUG))
|
|
return
|
|
find_references(FALSE)
|
|
|
|
/datum/proc/find_references(skip_alert)
|
|
running_find_references = type
|
|
if(usr && usr.client)
|
|
if(usr.client.running_find_references)
|
|
log_gc("CANCELLED search for references to a [usr.client.running_find_references].")
|
|
usr.client.running_find_references = null
|
|
running_find_references = null
|
|
//restart the garbage collector
|
|
SSgarbage.can_fire = 1
|
|
SSgarbage.next_fire = world.time + world.tick_lag
|
|
return
|
|
|
|
if(!skip_alert)
|
|
if(alert("Running this will lock everything up for about 5 minutes. Would you like to begin the search?", "Find References", "Yes", "No") == "No")
|
|
running_find_references = null
|
|
return
|
|
|
|
//this keeps the garbage collector from failing to collect objects being searched for in here
|
|
SSgarbage.can_fire = 0
|
|
|
|
if(usr && usr.client)
|
|
usr.client.running_find_references = type
|
|
|
|
log_gc("Beginning search for references to a [type].")
|
|
var/starting_time = world.time
|
|
|
|
DoSearchVar(GLOB, "GLOB") //globals
|
|
log_gc("Finished searching globals")
|
|
for(var/datum/thing in world) //atoms (don't beleive it's lies)
|
|
DoSearchVar(thing, "World -> [thing.type]", search_time = starting_time)
|
|
|
|
log_gc("Finished searching atoms")
|
|
|
|
for(var/datum/thing) //datums
|
|
DoSearchVar(thing, "World -> [thing.type]", search_time = starting_time)
|
|
|
|
log_gc("Finished searching datums")
|
|
|
|
for(var/client/thing) //clients
|
|
DoSearchVar(thing, "World -> [thing.type]", search_time = starting_time)
|
|
|
|
log_gc("Finished searching clients")
|
|
|
|
log_gc("Completed search for references to a [type].")
|
|
#ifdef FIND_REF_NOTIFY_ON_COMPLETE
|
|
rustg_create_toast("ParadiseSS13", "GC search complete for [type]")
|
|
#endif
|
|
if(usr && usr.client)
|
|
usr.client.running_find_references = null
|
|
running_find_references = null
|
|
|
|
//restart the garbage collector
|
|
SSgarbage.can_fire = 1
|
|
SSgarbage.next_fire = world.time + world.tick_lag
|
|
|
|
/datum/proc/qdel_then_find_references()
|
|
set category = "Debug"
|
|
set name = "qdel() then Find References"
|
|
if(!check_rights(R_DEBUG))
|
|
return
|
|
|
|
qdel(src, TRUE) //force a qdel
|
|
if(!running_find_references)
|
|
find_references(TRUE)
|
|
|
|
/datum/proc/qdel_then_if_fail_find_references()
|
|
set category = "Debug"
|
|
set name = "qdel() then Find References if GC failure"
|
|
if(!check_rights(R_DEBUG))
|
|
return
|
|
|
|
qdel_and_find_ref_if_fail(src, TRUE)
|
|
|
|
/datum/proc/DoSearchVar(potential_container, container_name, recursive_limit = 64, search_time = world.time)
|
|
if((usr?.client && !usr.client.running_find_references) || SSgarbage.ref_search_stop)
|
|
return
|
|
|
|
if(!recursive_limit)
|
|
log_gc("Recursion limit reached. [container_name]")
|
|
return
|
|
|
|
//Check each time you go down a layer. This makes it a bit slow, but it won't effect the rest of the game at all
|
|
#ifndef FIND_REF_NO_CHECK_TICK
|
|
CHECK_TICK
|
|
#endif
|
|
|
|
if(istype(potential_container, /datum))
|
|
var/datum/datum_container = potential_container
|
|
if(datum_container.last_find_references == search_time)
|
|
return
|
|
|
|
datum_container.last_find_references = search_time
|
|
var/list/vars_list = datum_container.vars
|
|
|
|
for(var/varname in vars_list)
|
|
#ifndef FIND_REF_NO_CHECK_TICK
|
|
CHECK_TICK
|
|
#endif
|
|
if(varname == "vars" || varname == "vis_locs") //Fun fact, vis_locs don't count for references
|
|
continue
|
|
var/variable = vars_list[varname]
|
|
|
|
if(variable == src)
|
|
log_gc("Found [type] [text_ref(src)] in [datum_container.type]'s [text_ref(datum_container)] [varname] var. [container_name]")
|
|
continue
|
|
|
|
if(islist(variable))
|
|
DoSearchVar(variable, "[container_name] [text_ref(datum_container)] -> [varname] (list)", recursive_limit - 1, search_time)
|
|
|
|
else if(islist(potential_container))
|
|
var/normal = IS_NORMAL_LIST(potential_container)
|
|
var/list/potential_cache = potential_container
|
|
for(var/element_in_list in potential_cache)
|
|
#ifndef FIND_REF_NO_CHECK_TICK
|
|
CHECK_TICK
|
|
#endif
|
|
//Check normal entrys
|
|
if(element_in_list == src)
|
|
log_gc("Found [type] [text_ref(src)] in list [container_name].")
|
|
continue
|
|
|
|
var/assoc_val = null
|
|
if(!isnum(element_in_list) && normal)
|
|
assoc_val = potential_cache[element_in_list]
|
|
//Check assoc entrys
|
|
if(assoc_val == src)
|
|
log_gc("Found [type] [text_ref(src)] in list [container_name]\[[element_in_list]\]")
|
|
continue
|
|
//We need to run both of these checks, since our object could be hiding in either of them
|
|
//Check normal sublists
|
|
if(islist(element_in_list))
|
|
DoSearchVar(element_in_list, "[container_name] -> [element_in_list] (list)", recursive_limit - 1, search_time)
|
|
//Check assoc sublists
|
|
if(islist(assoc_val))
|
|
DoSearchVar(potential_container[element_in_list], "[container_name]\[[element_in_list]\] -> [assoc_val] (list)", recursive_limit - 1, search_time)
|
|
|
|
#ifndef FIND_REF_NO_CHECK_TICK
|
|
CHECK_TICK
|
|
#endif
|
|
|
|
#endif
|