thumb|400px|... After that, the working memory contents is discarded in favor of the compressed copy, and the role of ''working'' and ''free'' memory are exchanged (depicted).
In computer science
, garbage collection (GC) is a form of automatic memory management
. The ''garbage collector'', or just ''collector'', attempts to reclaim ''garbage
'', or memory occupied by objects
that are no longer in use by the program
. Garbage collection was invented by American computer scientist John McCarthy
around 1959 to simplify manual memory management in Lisp
Garbage collection relieves the programmer from performing manual memory management
where the programmer specifies what objects to deallocate and return to the memory system and when to do so. Other similar techniques include stack allocation
, region inference
, memory ownership, and combinations of multiple techniques. Garbage collection may take a significant proportion of total processing time in a program and, as a result, can have significant influence on performance
Resources other than memory, such as network socket
s, database handle
s, user interaction window
and device descriptors, are not typically handled by garbage collection. Methods
for managing such resources, particularly destructors
, may suffice to manage memory as well, leaving no need for GC. Some GC systems allow such other resources to be associated with a region of memory that, when collected, causes the work of reclaiming these resources.
The basic principles of garbage collection are to find data objects in a program that cannot be accessed in the future, and to reclaim the resources used by those objects.
Many programming language
s require garbage collection, either as part of the language specification
(for example, Java
and most scripting language
s) or effectively for practical implementation (for example, formal languages like lambda calculus
); these are said to be ''garbage collected languages''. Other languages were designed for use with manual memory management, but have garbage-collected implementations available (for example, C
). Some languages, like Ada
, and C++/CLI
, allow both garbage collection and manual memory management
to co-exist in the same application by using separate heaps
for collected and manually managed objects; others, like D
, are garbage-collected but allow the user to manually delete objects and also entirely disable garbage collection when speed is required.
While integrating garbage collection into the language's compiler
and runtime system
enables a much wider choice of methods, ''post-hoc'' GC systems exist, such as Automatic Reference Counting
(ARC), including some that do not require recompilation. (''Post-hoc'' GC is sometimes distinguished as ''litter collection''.) The garbage collector will almost always be closely integrated with the memory allocator
Garbage collection frees the programmer from manually dealing with memory deallocation. As a result, certain categories of bug
s are eliminated or substantially reduced:
* ''Dangling pointer
bugs'', which occur when a piece of memory is freed while there are still pointer
s to it, and one of those pointers is dereferenced
. By then the memory may have been reassigned to another use, with unpredictable results.
* ''Double free bugs'', which occur when the program tries to free a region of memory that has already been freed, and perhaps already been allocated again.
* Certain kinds of ''memory leak
s'', in which a program fails to free memory occupied by objects that have become unreachable
, which can lead to memory exhaustion. (Garbage collection typically does not deal with the unbounded accumulation of data that is reachable, but that will actually not be used by the program.)
* Efficient implementations of persistent data structure
Some of the bugs addressed by garbage collection have security implications.
Typically, garbage collection has certain disadvantages, including consuming additional resources, performance impacts, possible stalls in program execution, and incompatibility with manual resource management.
Garbage collection consumes computing resources in deciding which memory to free, even though the programmer may have already known this information. The penalty for the convenience of not annotating object lifetime manually in the source code is overhead
, which can lead to decreased or uneven performance.
A peer-reviewed paper from 2005 came to the conclusion that GC needs five times the memory to compensate for this overhead and to perform as fast as explicit memory management. Interaction with memory hierarchy effects can make this overhead intolerable in circumstances that are hard to predict or to detect in routine testing. The impact on performance was also given by Apple as a reason for not adopting garbage collection in iOS
despite being the most desired feature.
The moment when the garbage is actually collected can be unpredictable, resulting in stalls (pauses to shift/free memory) scattered throughout a session. Unpredictable stalls can be unacceptable in real-time environments
, in transaction processing
, or in interactive programs. Incremental, concurrent, and real-time garbage collectors address these problems, with varying trade-offs.
The modern GC implementations try to minimize blocking "stop-the-world
" stalls by doing as much work as possible in the background (i.e. on a separate thread), for example marking unreachable garbage instances while the application process continues to run. In spite of these advancements, for example in the .NET CLR paradigm
it is still very difficult to maintain large heaps (millions of objects) with resident objects that get promoted to older generations without incurring noticeable delays (sometimes a few seconds).
The need for explicit manual resource management (release/close) for non-GCed resources in an object oriented language becomes transitive to composition. That is: in a non-deterministic GC system, if a resource or a resource-like object requires manual resource management (release/close), and this object is used as "part of" another object, then the composed object will also become a resource-like object that itself requires manual resource management (release/close).
Tracing garbage collection
is the most common type of garbage collection, so much so that "garbage collection" often refers to tracing garbage collection, rather than other methods such as reference counting
. The overall strategy consists of determining which objects should be garbage collected by tracing which objects are ''reachable'' by a chain of references from certain root objects, and considering the rest as garbage and collecting them. However, there are a large number of algorithms used in implementation, with widely varying complexity and performance characteristics.
Reference counting garbage collection is where each object has a count of the number of references to it. Garbage is identified by having a reference count of zero. An object's reference count is incremented when a reference to it is created, and decremented when a reference is destroyed. When the count reaches zero, the object's memory is reclaimed.
As with manual memory management, and unlike tracing garbage collection, reference counting guarantees that objects are destroyed as soon as their last reference is destroyed, and usually only accesses memory which is either in CPU cache
s, in objects to be freed, or directly pointed by those, and thus tends to not have significant negative side effects on CPU cache and virtual memory
There are a number of disadvantages to reference counting; this can generally be solved or mitigated by more sophisticated algorithms:
; Cycles: If two or more objects refer to each other, they can create a cycle whereby neither will be collected as their mutual references never let their reference counts become zero. Some garbage collection systems using reference counting (like the one in CPython
) use specific cycle-detecting algorithms to deal with this issue. Another strategy is to use weak reference
s for the "backpointers" which create cycles. Under reference counting, a weak reference is similar to a weak reference under a tracing garbage collector. It is a special reference object whose existence does not increment the reference count of the referent object. Furthermore, a weak reference is safe in that when the referent object becomes garbage, any weak reference to it ''lapses'', rather than being permitted to remain dangling, meaning that it turns into a predictable value, such as a null reference.
; Space overhead (reference count): Reference counting requires space to be allocated for each object to store its reference count. The count may be stored adjacent to the object's memory or in a side table somewhere else, but in either case, every single reference-counted object requires additional storage for its reference count. Memory space with the size of an unsigned pointer is commonly used for this task, meaning that 32 or 64 bits of reference count storage must be allocated for each object. On some systems, it may be possible to mitigate this overhead by using a tagged pointer
to store the reference count in unused areas of the object's memory. Often, an architecture does not actually allow programs to access the full range of memory addresses that could be stored in its native pointer size; certain number of high bits in the address is either ignored or required to be zero. If an object reliably has a pointer at a certain location, the reference count can be stored in the unused bits of the pointer. For example, each object in Objective-C
has a pointer to its class
at the beginning of its memory; on the ARM64
architecture using iOS 7
, 19 unused bits of this class pointer are used to store the object's reference count.
; Speed overhead (increment/decrement):In naive implementations, each assignment of a reference and each reference falling out of scope often require modifications of one or more reference counters. However, in the common case, when a reference is copied from an outer scope variable into an inner scope variable, such that the lifetime of the inner variable is bounded by the lifetime of the outer one, the reference incrementing can be eliminated. The outer variable "owns" the reference. In the programming language C++, this technique is readily implemented and demonstrated with the use of
references. Reference counting in C++ is usually implemented using "smart pointer
s" whose constructors, destructors and assignment operators manage the references. A smart pointer can be passed by reference to a function, which avoids the need to copy-construct a new smart pointer (which would increase the reference count on entry into the function and decrease it on exit). Instead the function receives a reference to the smart pointer which is produced inexpensively. The Deutsch-Bobrow method of reference counting capitalizes on the fact that most reference count updates are in fact generated by references stored in local variables. It ignores these references, only counting references in the heap, but before an object with reference count zero can be deleted, the system must verify with a scan of the stack and registers that no other reference to it still exists. A further substantial decrease in the overhead on counter updates can be obtained by update coalescing introduced by Levanoni and Petrank
Consider a pointer that in a given interval of the execution is updated several times. It first points to an object
, then to an object
, and so forth until at the end of the interval it points to some object
. A reference counting algorithm would typically execute
. But most of these updates are redundant. In order to have the reference count properly evaluated at the end of the interval it is enough to perform
. Levanoni and Petrank measured an elimination of more than 99% of the counter updates in typical Java benchmarks.
; Requires atomicity:When used in a multithreaded
environment, these modifications (increment and decrement) may need to be atomic operation
s such as compare-and-swap
, at least for any objects which are shared, or potentially shared among multiple threads. Atomic operations are expensive on a multiprocessor, and even more expensive if they have to be emulated with software algorithms. It is possible to avoid this issue by adding per-thread or per-CPU reference counts and only accessing the global reference count when the local reference counts become or are no longer zero (or, alternatively, using a binary tree of reference counts, or even giving up deterministic destruction in exchange for not having a global reference count at all), but this adds significant memory overhead and thus tends to be only useful in special cases (it is used, for example, in the reference counting of Linux kernel modules).
Update coalescing by Levanoni and Petrank
can be used to eliminate all atomic operations from the write-barrier. Counters are never updated by the program threads in the course of program execution. They are only modified by the collector which executes as a single additional thread with no synchronization. This method can be used as a stop-the-world mechanism for parallel programs, and also with a concurrent reference counting collector.
; Not real-time: Naive implementations of reference counting do not in general provide real-time behavior, because any pointer assignment can potentially cause a number of objects bounded only by total allocated memory size to be recursively freed while the thread is unable to perform other work. It is possible to avoid this issue by delegating the freeing of objects whose reference count dropped to zero to other threads, at the cost of extra overhead.
is a compile-time technique that can convert heap allocation
s to stack allocation
s, thereby reducing the amount of garbage collection to be done. This analysis determines whether an object allocated inside a function is accessible outside of it. If a function-local allocation is found to be accessible to another function or thread, the allocation is said to “escape” and cannot be done on the stack. Otherwise, the object may be allocated directly on the stack and released when the function returns, bypassing the heap and associated memory management costs.
Heartbeat and timestamp
Heartbeat-based garbage collection is a technique to free heterogenous resources such as file handlers and network pointers, rather than unused memory like traditional garbage collection algorithms. An example is a network algorithm that sends a "heartbeat" signal to a monitor. Once the client fails to send a heartbeat to the monitor serve, the network connection is closed and the resources are freed. Timestamp methods can work as garbage collectors for collection of unused memory but have serious drawbacks for this task and are used when other methods are not practical (i.e. network tasks).
Generally speaking, higher-level programming languages
are more likely to have garbage collection as a standard feature. In some languages that do not have built in garbage collection, it can be added through a library, as with the Boehm garbage collector
for C and C++.
Most functional programming language
s, such as ML
, and APL
, have garbage collection built in. Lisp
is especially notable as both the first functional programming language
and the first language to introduce garbage collection.
Other dynamic languages, such as Ruby
(but not Perl
5 or PHP
also tend to use GC. Object-oriented programming
languages such as Smalltalk
usually provide integrated garbage collection. Notable exceptions are C++
which have destructors
Historically, languages intended for beginners, such as BASIC
, have often used garbage collection for heap-allocated variable-length data types, such as strings and lists, so as not to burden programmers with manual memory management. On early microcomputers, with their limited memory and slow processors, BASIC garbage collection could often cause apparently random, inexplicable pauses in the midst of program operation.
Some BASIC interpreters, such as Applesoft BASIC
on the Apple II
family, repeatedly scanned the string descriptors for the string having the highest address in order to compact it toward high memory, resulting in
performance, which could introduce minutes-long pauses in the execution of string-intensive programs. A replacement garbage collector for Applesoft BASIC published in Call-A.P.P.L.E.
(January 1981, pages 40–45, Randy Wigginton
) identified a group of strings in every pass over the heap, which cut collection time dramatically. BASIC.System, released with ProDOS
in 1983, provided a windowing garbage collector for BASIC that reduced most collections to a fraction of a second.
While the Objective-C
traditionally had no garbage collection, with the release of OS X 10.5
in 2007 Apple
introduced garbage collection for Objective-C
2.0, using an in-house developed runtime collector.
However, with the 2012 release of OS X 10.8
, garbage collection was deprecated in favor of LLVM
's automatic reference counter
(ARC) that was introduced with OS X 10.7
. Furthermore, since May 2015 Apple even forbids the usage of garbage collection for new OS X applications in the App Store
, garbage collection has never been introduced due to problems in application responsivity and performance;
instead, iOS uses ARC.
Garbage collection is rarely used on embedded
or real-time systems because of the usual need for very tight control over the use of limited resources. However, garbage collectors compatible with many limited environments have been developed. The Microsoft .NET Micro Framework.NET nanoFramework
and Java Platform, Micro Edition
are embedded software platforms that, like their larger cousins, include garbage collection.
Garbage collectors available in Java JDKs include:
* Concurrent mark sweep collector
* C4 (Continuously Concurrent Compacting Collector)
Compile-time garbage collection
is a form of static analysis
allowing memory to be reused and reclaimed based on invariants known during compilation.
This form of garbage collection has been studied in the Mercury programming language
, and it saw greater usage with the introduction of LLVM
's automatic reference counter
(ARC) into Apple's ecosystem (iOS and OS X) in 2011.
[Apple says Mac app makers must transition to ARC memory management by May](_blank)
by AppleInsider (February 20, 2015)
Incremental, concurrent, and real-time garbage collectors have been developed, such as Baker's
algorithm or Lieberman
In Baker's algorithm, the allocation is done in either half of a single region of memory. When it becomes half full, a garbage collection is performed which moves the live objects into the other half and the remaining objects are implicitly deallocated. The running program (the 'mutator') has to check that any object it references is in the correct half, and if not move it across, while a background task is finding all of the objects.
Generational garbage collection
schemes are based on the empirical observation that most objects die young. In generational garbage collection two or more allocation regions (generations) are kept, which are kept separate based on object's age. New objects are created in the "young" generation that is regularly collected, and when a generation is full, the objects that are still referenced from older regions are copied into the next oldest generation. Occasionally a full scan is performed.
Some high-level language computer architecture
s include hardware support for real-time garbage collection.
Most implementations of real-time garbage collectors use tracing
. Such real-time garbage collectors meet hard real-time
constraints when used with a real-time operating system.
[McCloskey, Bacon, Cheng, Grov]
"Staccato: A Parallel and Concurrent Real-time Compacting Garbage Collector for Multiprocessors"
* Destructor (computer programming)
* International Symposium on Memory Management
* Memory management
* Dead-code elimination
* Smart pointer
* Virtual memory compression
* (511 pages)
* (404 pages)
The Memory Management Reference
* ttp://tinygc.sourceforge.net/ TinyGC - an independent implementation of the BoehmGC APIConservative Garbage Collection Implementation for C LanguageMeixnerGC - an incremental mark and sweep garbage collector for C++ using smart pointers
Category:Articles with example code
Category:Solid-state computer storage