Adding debug statements to your code to write out the fields in an object, or even entire data collections, is a common problem-solving approach. Often you must do this iteratively as you discover that you need more and more information to understand and solve the problem. Although this process can be effective, it can sometimes fail to bear fruit: The volume of debug code can cause the problem to disappear, you might need to add debug to code you don't own, debugging may require you to restart processes, or the overall performance impact of the debug may prevent the application from running.
Memory Analyzer is a cross-platform, open source tool that you can use not only to diagnose memory problems, but also to gain huge insight into the state and behaviour of an entire Java application. By reading in a snapshot dump created by the Java runtime whilst the application is running, Memory Analyzer gives you a way to diagnose tricky problems that debug code might fail to expose.
This article shows how to generate dumps and use them to examine and diagnose the state of your application. With Memory Analyzer, you can inspect threads, objects, bundles, and whole data collections to debug Java code problems that go beyond memory leaks.
Memory Analyzer can currently work with three dump types:
- IBM Portable Heap Dump (PHD): This proprietary IBM format contains only the type and size of each Java object in the process, and the relationships among the objects. This dump-file format is significantly smaller than the other formats and contains the least information. The data is usually sufficient, though, for diagnosing memory leaks and getting a basic understanding of the application's architecture and footprint.
- HPROF binary dump: The HPROF binary format contains all the data present in the IBM PHD format as well as the primitive data held inside the Java objects, and the thread details. You can look at the values held in fields inside the objects and see which methods were being executed at the time the dump was taken. The additional primitive data makes HPROF dumps significantly larger than PHD-format dumps; they are approximately the same size as the used Java heap.
- IBM system dumps: When the IBM Java runtime is being used, the native operating-system dump file — a core file on AIX® or Linux, a minidump on Windows®, or a SVC dump on z/OS® — can be loaded into Memory Analyzer. These dumps contain the entire memory image of the running application — all the information and data in the HPROF format, as well as all of the native-memory and thread information. This is the largest and most comprehensive dump-file format.
Table 1 summarises the differences among the dump-file types:
Table 1. Summary of the dump types' characteristics
|Dump format||Approximate size on disk||Objects, classes, and classloaders||Thread details||Field names||Field and array references||Primitive fields||Primitive array contents||Accurate garbage-collection roots||Native memory and threads|
|IBM PHD||20 percent of Java heap size||Y||With Javacore*||N||Y||N||N||N||N|
|HPROF||Java heap size||Y||Y||Y||Y||Y||Y||Y||N|
|IBM system dumps||Java heap size + 30 percent||Y||Y||Y||Y||Y||Y||Y||Y|
*By loading in both a javacore.txt file (IBM thread dump file) and a heapdump.phd file that were generated at the same time, Memory Analyzer makes thread details available in the IBM PHD format dump.
Both the HPROF and IBM system dump formats can be compressed well, usually to around 20 percent of their original size, using operating-system tools.
Different mechanisms are available for obtaining the various dumps for each of the Java runtimes, providing flexibility that lets you generate snapshot dumps for scenarios beyond those involving
OutOfMemoryErrors. The mechanisms available depend on which vendor's Java runtime you're using.
For all dump types, you must ensure sufficient disk space for the dumps so that they are not truncated. The default location of the dumps is the current working directory of the JVM process. For IBM JVMs, you can change this with the
-Xdump file command-line option. For the HotSpot JVM, you can change it using the
-XX:HeapDumpPath command-line option. See Resources for links to the relevant syntax.
Dumps from the operating system can be used for both IBM and HotSpot JVMs. For the IBM
JVM, you can create dumps with the
jextract tool (shipped
with the JDK) and load them directly into Memory Analyzer; for the HotSpot JVM, you
jmap tool to extract heap dumps from core dumps.
(We discuss both techniques in detail later in this article.) However, on some
operating systems, you must ensure that the process is running with sufficient
ulimits before creating the core dump; otherwise, the core dump will
be truncated and analysis will be limited. If the
ulimits are incorrect, you must modify them and restart the process before gathering a dump. See Resources for links to detailed information on obtaining system dumps from AIX, Linux®, z/OS, and Solaris.
The HotSpot-based Java runtimes generate the HPROF format dump only. You can choose among several interactive methods and one event-based method for generating the dump:
- Interactive methods:
Using a Ctrl+Break: If the
-XX:+HeapDumpOnCtrlBreakcommand-line option is set for the running application, an HPROF format dump is generated along with a thread dump when a Ctrl+Break event, or
SIGQUIT(usually generated using
kill -3), is sent via the console. This option may not be available on some versions, in which case try:
jmaputility tool (see Resources), delivered in the bin directory of the JDK, provides an option to request an HPROF dump from the running process. With Java 5, use:
jmap -dump:format=b pid
With Java 6, use this version, where
liveis optional and results in only the "live" objects being written to the dump-file process ID (PID):
jmap -dump[live,]format=b,file=filename pid
Using the operating system: Use the nondestructive
gcorecommand or the destructive
kill -11commands to produce a core file. Then, extract a heap dump from the core file using
jmap -dump:format=b,file=heap.hprof path to java executable core
- Using the JConsole tool: A
dumpHeapoperation is provided under the
HotSpotDiagnosticMBean in JConsole. This operation requests that a HPROF dump be generated.
- Event-based method:
- On an
OutOfMemoryError: If the
-XX:+HeapDumpOnOutOfMemoryErrorcommand-line option is set for the running application, an HPROF format dump is generated when an
OutOfMemoryErroroccurs. It is ideal to have this in place for production systems, because it is almost always required to diagnose memory issues, and it incurs no ongoing performance overhead. In older releases of HotSpot-based Java runtimes, there's no limit to how many heap dumps are produced on this event per JVM run; in newer releases, a maximum of one heap dump is produced on this event per JVM run.
- On an
The IBM runtimes provide dump and trace engines that can generate either PHD-format or system dumps in a large number of interactive and event-based based scenarios. You can also generate interactive dumps using the Health Center tool or programmatically using a Java API.
- Interactive methods
SIGQUITor Ctrl+Break: When a Ctrl+Break or
SIGQUIT(usually generated using
kill -3) is sent to the IBM runtime, a user event is generated in the IBM dump engine. By default this event only generates a thread dump file (javacore.txt). You can use the
-Xdump:heap:events=useroption to generate a PHD-format dump, or the
-Xdump:system:events=useroption to generate a system dump of the Java application.
Using the operating system to produce a system dump:
gencore(or the destructive
gcore(or the destructive
SVCDUMPor console dump
Using IBM Monitoring and Diagnostics Tools for Java - Health Center: The Health Center tool provides a menu option for requesting either a PHD or a system dump from a running Java process (see Resources).
- Event-based methods. The IBM dump and trace engines provide a flexible set of capabilities for generating PHD and system dumps on a large number of events, from exceptions being thrown to methods being executed. Using them, you should be able to generate dumps for most problem scenarios you want to diagnose:
Using the IBM dump engine: The dump engine provides a large number of events on which you can produce a PHD or system dump. Further, it lets you filter on types of those events in order to exercise finer-grained control over when to generate dumps.
You can see the default events by using the
-Xdump:whatoption. You'll notice, for example, that a heapdump.phd and javacore.txt are produced on the first four
OutOfMemoryErrorexceptions in the JVM.
To gather more data, you can produce a system dump instead of a heap dump on an
-Xdump:heap:none -Xdump:java+system:events=systhrow, filter=java/lang/OutOfMemoryError,range=1..4,request=exclusive+compact+prepwalk
Some exceptions, for example
NullPointerExceptions, are generated commonly in most applications by a wide range of code. This makes it difficult to generate a dump on a particular
NullPointerExceptionof interest. To help you be more specific about which exception to generate the dump on, an extra level of filtering is provided for "throw" and "catch" events that lets you specify the throwing and catching methods, respectively. You do this by adding a
#separator and then adding the throwing or catching method as appropriate. For example, this option produces a system dump when a
NullPointerExceptionis thrown by the
This option produces a system dump when a
NullPointerExceptionis caught by the
In addition to filtering on the events, you can also specify a range of events on which you want dumps to be generated. For example, this option produces a dump only on the fifth occurrence of a
This option uses a range to produce a dump only on the second, third, and fourth occurrences of a
Table 2 summarizes the most useful events and filters:
Table 2. Available dump events
Event Description Available filtering Example
General protection fault (crash)
User generated signal (
VM shutdown, including call to
Generate a system dump on VM shutdown with an exit code between
Class load Class name
Generate a system dump when the
com.ibm.example.Exampleclass is loaded.
Class unload Class name
Generate a system dump when the
com.ibm.example.Exampleclass is unloaded.
An exception being thrown Exception class name
Generate a system dump when a
An exception being caught Exception class name
Generate a system dump when a
A Java exception is about to be thrown by the JVM. (This is different from the
throwevent because it is only triggered for error conditions detected internally in the JVM.)
Exception class name
Generate a system dump when an
A Java object is allocated Size of object being allocated
Generate a system dump when an object larger than 5MB is allocated.
- Using the IBM trace engine:
The trace engine allows PHD and system dumps to be triggered on method entry or exit for any Java method running in the application. You accomplish this by using the
triggerkeyword to the
-Xtracecommand-line options that control the IBM trace engine. The syntax for the trigger option is:
Adding the following command-line option to the application produces a system dump when the
Example.trigger()method is called:
This command-line option produces a PHD dump when the
Example.trigger()method is called:
It is recommended, though, that you set a range so that you don't create dumps every time the method is called. This example ignores the first five calls to
Example.trigger()and then triggers one dump:
Note that an empty term is used for the
exitActionin this example because we're triggering the dumps on method entry only.
- Programmatic methods:
The IBM runtimes also provide a
systemDump()methods. They generate thread dumps, PHD dumps, and system dumps, respectively.
As well as the methods for obtaining dumps that are provided by the runtimes themselves, Memory Analyzer also provides an Acquire Heap Dump option, shown in Figure 1, that allows you to trigger and load a snapshot dump from a Java process running on the same machine as Memory Analyzer:
Figure 1. Using the Acquire Heap Dump function in Memory Analyzer
On HotSpot-based runtimes, Memory Analyzer generates the dump using
jmap. For the IBM runtimes, the dump is generated using the Java "late attach" functionality and programmatic API. Java 6 SR6 is required for the function to work, because earlier releases do not contain the "late attach" function.
For IBM system dumps, the dump must be postprocessed using the
jextract tool shipped with the JDK:
jextract is run on the same physical machine that
produced the dump, using
jextract from the same JDK installation that produced the dump, and with read access to the same libraries that
java process was running with. Given that
jextract can consume significant CPU cycles processing the dump, this may be unacceptable in some production systems. In this case, the dump should be processed on the closest matching system, such as a preproduction test system. The Service Refresh (SR) and Fix Pack (FP) versions of the Java runtimes should match.
jextract produces a ZIP file that includes the original core dump, a processed representation of the dump, the Java executable, and the libraries used by the
java process. You can delete the original (unzipped) core dump after running
jextract. The ZIP file is what you should load into Memory Analyzer.
You can extract a PHD dump from a
jextracted system dump by loading the ZIP into
jdmpview and executing the
heapdump command (see Resources).
Memory Analyzer can diagnose
OutOfMemoryErrors by looking for areas of the application that are either leaking memory or have a footprint requirement that's too large for the available memory. Memory Analyzer does automatic leak detection and generates a Leak Suspects report (see Resources).
The additional data that's available in the HPROF and IBM system dumps, particularly the field names and field values — along with the capabilities of the Inspector view and Object Query Language (OQL) — also make it possible to diagnose a wider range of problem types than "What's using all of the memory?". For example, you can ascertain the occupancy and load factor of collections to see if they are efficiently sized, or look at the hostname and port associated with a
ConnectException to see what connection the application was trying to create.
When any object is selected in Memory Analyzer, the Inspector view shows the available information relating to that object, including the class hierarchy, attributes, and statics. The Attributes panel shows the instance fields and values associated with the object, and the Statics panel shows the static fields and values associated with the class.
The Inspector view shown in Figure 2 for a simple
java.net.URL object lets you see details about that object, including the type of protocol the URL is for and the destination:
Figure 2. The Statics, Attributes, and Class Hierarchy panels in the Inspector view
In Figure 2, you can see in the Attributes panel that the URL object refers to a JAR file (the protocol field) located on the local file system (in the location specified by the path and file fields).
OQL can be used to query a dump using custom, SQL-like queries. This topic could be an article in itself, so we'll just highlight a few examples. For more details, consult the Help contents on OQL available from within Memory Analyzer.
OQL is particularly useful for following a path down the outgoing references and fields of a set of objects to a particular field. For example, if class
A has a field
foo of type
B, and class
B has a field
bar that is a
String, then a simple query to find all of those
Strings would be:
SELECT aliasA.foo.bar.toString() FROM A aliasA
We give class
A an alias,
aliasA, which we then reference in the
SELECT clause. This query strictly selects only from instances of class
A. If we wanted to select from all instances of class
A as well as any subclasses, we'd use:
SELECT aliasA.foo.bar.toString() FROM INSTANCEOF A aliasA
Here is a more complicated example with
SELECT k, k.capacity FROM java.nio.DirectByteBuffer k WHERE ((k.viewedBuffer=null)and(inbounds(k).length>1))
In this case, we want to get the capacity field of any
DirectByteBuffer, which gives the native memory held by that object. We also want to filter out any
DirectByteBuffers that have a null
viewedBuffer field (because those are just views into other
DirectByteBuffers) and more than one inbound reference (so that we don't look at those pending clean-up with their phantom reference — that is, we want only "live"
With Memory Analyzer, you can compare tables generated by queries. The tables can either be from the same dump, letting you see whether
String objects from one view are present in a collection object seen in another view, or across separate dumps, letting you look for changes in data, for example growth of object collections.
To run a comparison, you add the relevant tables to the Compare Basket and then request that the entries in the basket be compared. First find and select the entry for the table in the Navigation History, then select Add to Compare Basket from the context menu, as shown in Figure 3:
Figure 3. Adding tables from the Navigation History view to the Compare Basket
Once you have two entries in the Compare Basket, you can run the comparison using the Compare-the-results button (the red exclamation mark) in the top right-hand corner of the panel, as shown in Figure 4:
Figure 4. Comparing the results of the entries in the Compare Basket
Another important use of Memory Analyzer is to find which components are using most of the heap, even in situations without a memory leak. If memory usage can be reduced, then the system's capacity or performance can be improved, allowing more sessions or less time spent garbage collecting.
The Top Components report is the first step. It splits memory usage by component in the system, analyzes the usage in each component, and looks for wasteful practices. Objects dominated (retained by) another object can be said to be owned by that dominator. The Top Components report lists all the objects that are not owned by another object. These are the top dominators of the heap. The top dominators are then divided by classloader using the class of the objects, and all those top dominators and objects owned by them are allocated to appropriate classloaders. You analyse further by selecting one of the classloaders in the report to open a new classloader-specific component report.
For each component, the Collections objects are analyzed. The Collections classes, as
java.util.*, are a great time-saver for programmers, providing well-tested implementations of lists, sets, and maps. The average application can have millions of collections, so wasted space in collections can be significant.
Empty collections are one common cause of wasted memory.
HashSets are created with a default-size backing array of perhaps 10 entries, ready to hold entries. It's surprisingly common in applications for a collection to be created but no objects stored in it. This can rapidly consume memory. For example, with 100,000 empty collections, the backing arrays alone could consume 100,000 * (24+10*4) bytes = 6MB.
The Empty Collection report looks at the standard collection classes and extensions thereof and analyzes them by size of collection. It then produces a table for each collection sorted by size of collection, with the most frequent size first. If a large proportion of instances of a type of a collection are empty, then the report flags this as a possible memory waste.
One solution is to delay allocating the collection until an entry is needed to be inserted. Another is to allocate a collection with a default size of 0 or 1, letting it grow if required at some runtime cost. A third is to trim the collection to size after the initialization phase is complete.
A related area is collections with only a few entries and a lot of wasted space. The Collection Fill Ratio section shows, for each collection type, the number of instances of that collection with a particular fill ratio. This reveals collections with a large proportion of empty space.
Strings and character arrays occupy a large amount of space in a typical business
application, so they are another area worthy of analysis. This section of the component report analyses strings for common content. Strings are immutable. Strings constants with the same value are guaranteed by the VM specification to use the same instance. Dynamically built strings have no such guarantee, and two
Strings built by, for example, reading data from a database or disk that have the same value will have separate instances and separate backing character arrays. If these strings are kept, then this can be significant.
You can solve this problem by using
String.intern() or maintaining a user hash set or hash map.
String.substring() is implemented in the Java language by building a new
String sharing the original character array. This is efficient if the original string is still needed. If only a small substring is needed, then — because the whole character array is retained — some space is wasted. The Waste In Char Arrays query shows the amount of wasted space in character arrays that are only referenced by strings.
Modern applications are divided into components, often based on classloaders, to provide some degree of isolation between parts of the application. Components can be updated by stopping the usage of one component classloader and loading a new version of the component using a new classloader. In time, the old version is freed by garbage collection, assuming the application contains no external references to classes or objects or the classloader.
The Class Loader Explorer query shows all the classloaders in the system, and so works for all applications. It shows the classes loaded by a classloader and also the parent chain of the classloader so that classloading problems can be understood. By inspection, you can see if multiple copies of a classloader exist. If next to no instances of classes are defined by a classloader, then it is likely that the classloader is idle.
The duplicate-classes query shows class names that have been loaded by multiple classloaders. This could be an indication of a classloader memory leak. It takes just one reference to an object held elsewhere in the system, such as in a registry, for a classloader leak to occur. The object holds a reference to its class, and the class to the classloader, and the classloader to all the defined classes.
A common classloading framework is the OSGi framework. One implementation is Eclipse Equinox, used for Eclipse-based applications to separate the plug-ins, and also used for WebSphere® Application Server 6.1 and later versions. When trying to understand the state of an application, it is useful to know the state of all the bundles. The Eclipse Equinox Bundle Explorer query, shown in Figure 5, does just that:
Figure 5. Eclipse Bundle Explorer
A system or HPROF dump has all the objects and fields. The Bundle Explorer shows all the bundles in the system, together with their states and dependencies, dependents, and services. It can show bundles that are unexpectedly active and so using more resources.
As indicated in Table 1, a dump can include thread details that can provide unique insights into what was happening at the time of the dump. This can include all active threads stacks, all of the frames for each thread, and most important, some or all of the active Java locals on those frames.
The Thread Overview view, shown in Figure 6, shows every thread in the JVM as well as various attributes of that thread, such as its retained heap size, context classloader, priority, state, and native ID:
Figure 6. Thread Overview
The retained heap size is particularly useful in cases in which no Java heap problem per se occurs during an
OutOfMemoryError, but rather the sum of the retained heaps of the threads is "too much." In this case, the JVM may be undersized, the thread pool sizes may be too large, or the average or maximum Java heap "load" of a thread is too high.
The Thread Stacks view, shown in Figure 7, shows every thread, its stack, stack frames, and Java locals on those stack frames:
Figure 7. Thread Stacks view
In the example in Figure 7, a thread of the type
java.lang.Thread and with the name
main — the main thread in a simple command-line program — is expanded. Each stack frame of the thread is displayed, and those with available Java locals are expandable. In this case, a
String has been passed as an argument from
Play.method2, and the contents of the string,
user1, is highlighted in the red circle. You can imagine the power of being able to reconstruct or reverse-engineer what was happening at the time of the dump based on what was happening in each thread stack frame and on which objects.
Note that because of runtime optimization, not all related objects such as method parameters or object instances will be available (although those objects will be in the dump), but those objects that are actively being "worked on" usually will be.
When exceptions are generated in the application, additional complications can make it more difficult to analyse the exception's cause. Two examples of this dilemma are:
- The use of logging mechanisms means that the exception is lost or the exception message is removed.
- The exception is producing a message that contains insufficient information.
In the first case, either the exception message or the entire exception is completely lost, making it difficult to know that the problem exists or to gain basic information about it. In the second case, the exception has been logged and the exception message and stack trace is available, but it doesn't contain the necessary information to resolve the exception's cause.
Because Memory Analyzer has access to the fields inside the objects, it's possible to find the exception message from the exception object. In some cases, it's also possible to extract additional data that's not in the original exception.
One way to locate the exceptions present in the snapshot dump is to use the OQL capability in Memory Analyzer to locate objects of interest in the dump. For example, this query finds all of the exception objects:
SELECT * FROM INSTANCEOF java.lang.Exception exceptions
This next query produces a list of all of the exceptions, from which you can use the Inspector view to look at the fields inside each exception. Knowing that the field that contains the exception message is the
detailMessage field, you can also modify the query to extract the exception messages directly and display them immediately as part of the results table:
SELECT exceptions.@displayName, exceptions.detailMessage.toString() FROM INSTANCEOF java.lang.Exception exceptions
The preceding query produces the output shown in Figure 8:
Figure 8. Output of the OQL query for exceptions, including exception messages
Figure 8 shows every exception that is still present in the application, and the message that would have been shown when the exception was thrown.
Although finding the exception object from the dump allows you to recover the exception message, sometimes the exception message is too generic or vague to enable you to understand the problem's cause. A good example is the
java.net.ConnectException. When trying to make a socket connection to a host that is not accepting connections, you get the following message:
java.net.ConnectException: Connection refused: connect at java.net.PlainSocketImpl.socketConnect(Native Method) at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:352) at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:214) at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:201) at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:377) at java.net.Socket.connect(Socket.java:530) at java.net.Socket.connect(Socket.java:480) at java.net.Socket.(Socket.java:377) at java.net.Socket.(Socket.java:220)
This message is sufficient if you have access to the code that's creating the socket and can see from the code which hostname and port are being used. In the case of more-complex code in which the hostname and port values are subject to change because they are obtained from external sources (user input values, databases, and the like), the message doesn't help you to understand why the connection is being refused.
The stack trace should include a socket object that contains the useful data, and if we can find the socket object in the snapshot dump using Memory Analyzer, then we can find out which hostname and port refused the connection.
The easiest way to do this is to generate a dump when the exception is thrown. This can be done on the IBM runtimes using the following
-Xdump option set:
This option generates an IBM system dump on the first occurrence of a
ConnectException generated by the
After loading the generated snapshot dump into Memory Analyzer, we can use the Open Query Browser > Java Basics > Thread Stacks option to list the threads and the objects associated with each method in the stack trace of the threads.
By expanding the current thread and the method frames in the thread, you can look at the objects associated with those methods. In the case of a
java.net.ConnectException, the most interesting method is
java.net.Socket.connect(). Expanding this method frame shows a reference to a
java.net.Socket object in memory. This is the socket connection we were trying to make.
Socket object is selected, the fields are shown in the Inspector view, as you can see in Figure 9:
Figure 9. Inspector view for the
The information in Figure 9 isn't too useful, because the real implementation of the
Socket is in the
impl field. You can inspect the contents of the
impl object by either expanding the
Socket object and selecting the
impl java.net.SocksSocketImpl line in the main panel, or by right-clicking on the
impl field in the Inspector view and selecting Go Into. Now the fields for
SocksSocketImpl are visible in the Inspector view, as shown in Figure 10:
Figure 10. Inspector view for the
The view shown in Figure 10 gives access to the
port fields. In this case, the port is
100, but the address field points to a
java.net.Inet4Address object. Following the same process to look into the fields of the
Inet4Address objects shows the results displayed in Figure 11:
Figure 11. Inspector view for the
You can see that the
hostName is set to
Here are a few tips and tricks that may be useful:
- Don't forget that Memory Analyzer itself may run out of memory. For the Eclipse MAT, edit
-Xmxin the MemoryAnalyzer.ini file. For the ISA version, edit the ISA Install/rcp/eclipse/plugins/com.ibm.rcp.j2se.../jvm.properties file.
- If you're still running out of memory on a 32-bit version of Memory Analyzer, use the 64-bit version of Eclipse MAT or try the headless mode (see Resources). (The ISA tool currently does not currently support 64-bit.)
- Memory Analyzer writes "swap" files in the directory of the dump, which lessens the dump's reload time. These files can be zipped, sent to another machine, and placed in the same directory as the dump, making a complete reload of the dump unnecessary.
- If a dump's size does not correlate with the garbage collector at the time of the dump, consult the Unreachable Objects Histogram link in the Overview tab. The Java heap may have had a lot of garbage (for example, if a tenured collection hadn't run for some time) that Memory Analyzer removed.
- If two objects
Bdo not have direct references to each other, but both have outgoing references to some set of objects
C, then the Retained Heap of the set
Cwill not be included in either of the retained sets of
B, but rather in the retained set of the dominator of both
B. In some situations,
Bmay be temporarily observing the set
C, which is actually the progeny of
A. In this case, you can right-click on
Aand select Java Basics > Customized Retained Set and use the address of
Bas the exclude (
- You can load multiple dumps at once and compare them. Open the Histogram of the more recent dump, click the Compare button at the top, and choose the baseline dump.
- When you are exploring a reference tree, be aware that references can refer, directly or indirectly, back to a "parent" reference, so that you could enter an exploration loop or cycle (for example, in a linked list). Be aware of the object addresses. Also, be aware that if the class name of an object is preceded by the word
class, that you are exploring the static instance of that class.
Stringvalue displayed in most views is limited to 1,024 characters. If you need the whole
String, right-click on the object and select Copy > Save value to file.
- Most views have an export option, and most HTML results are created on the file system, so that data can be exported for sharing or further transformation. Relatedly, you can press Ctrl+C on any selection of rows in a grid to copy a textual representation of those rows to your clipboard.
Memory Analyzer was originally developed as "a fast and feature-rich Java heap analyzer that helps you find memory leaks and reduce memory consumption," as it's described on Eclipse.org. But its capabilities clearly stretch far beyond that description. In addition to their role in diagnosing "normal" memory problems, snapshot dumps can be used as an alternative to, or in addition to, other types of problem-determination techniques such as tracing and patching. Particularly with HPROF dumps and IBM system dumps, Memory Analyzer gives you memory contents such as primitives and the field names from the original source code. Using the various views covered in this article, you can explore or reverse-engineer the problem at hand, including overall footprint and memory efficiency, Eclipse bundles and classloader relationships, thread data usage and stack frame locals, exceptions, and more. OQL and the Memory Analyzer plug-in model also allow you to inspect the dump more easily using a query language and programmatic methods that can help in automating common analysis.
Webcast: Debugging Java Applications with Memory Analyzer and the IBM Extensions for Memory Analyzer: Chris Bailey and Kevin Grigorenko lead this session on how to generate dumps and use them to examine your application: find memory leaks, inspect threads, and look at the IBM product structures using the IBM Extensions for Memory Analyzer.
Memory Analyzer blog and Memory Analyzer wiki: Check out these resources on Eclipse.org.
on troubleshooting Java applications: Follow Chris Bailey's developerWorks blog.
Using the -Xdump option: See this section of the IBM Java Diagnostics Guide 6 for information on
Java HotSpot VM Options: Here you'll find the syntax for changing the location of a dump file generated with the HotSpot JVM.
- How to obtain system dumps for:
- AIX: Enabling full AIX core files.
- Linux: Setting up and checking your Linux environment.
- z/OS: Obtaining SVC dumps and Setting up dumps.
- Solaris: Troubleshooting Guide for Java SE 6 with HotSpotVM.
jmap: Read the official
- Using dump agents: Learn more about configuring the IBM dump engine to generate system dumps and heap dumps.
Using system dumps and the dump viewer: Read about the IBM
Automated Heap Dump Analysis: Finding Memory Leaks with One Click: Learn more about automatic leak detection with Memory Analyzer.
- How to run on 64bit VM while the native SWT are 32bit: See how to run the Eclipse Memory Analyzer Tool in 64-bit headless mode.
Browse the technology bookstore for books on these and other technical topics.
developerWorks Java technology zone: Find hundreds of articles about every aspect of Java programming.
- IBM Extensions for
Memory Analyzer: The IBM Extensions for Memory Analyzer offer both additional
capabilities for debugging generic Java applications, and capabilities
for debugging specific IBM software products.
Get products and technologies
IBM Monitoring and Diagnostics Tools for Java - Memory Analyzer: Get Memory Analyzer as part of the IBM Support Assistant.
Eclipse Memory Analyzer Tool and the IBM DTFJ Plug-in: Download the Eclipse MAT and install the DTFJ plug-in to use MAT with IBM PHD and system dumps.
IBM Monitoring and Diagnostics Tools for Java - Health Center: You can use Health Center to request either a PHD or system dump from a running Java process.
- Get involved in the developerWorks community.
Chris Bailey is a member of the IBM Java Technology Center team at the Hursley Park Development Lab in the United Kingdom. As the technical architect for the IBM Java service and support organization, he is responsible for enabling users of the IBM SDK for Java to deliver successful application deployments. Chris is also involved in gathering and assessing new requirements, delivering new debugging capabilities and tools, making improvements in documentation, and improving the quality of the IBM SDK for Java.
Andrew Johnson is a Chartered Engineer and an Advisory Software Engineer at the IBM Java Technology Center in Hursley, England. He joined IBM in 1988 after receiving a B.A. in Electronic Engineering from Cambridge University. Since 1996 he has worked on Java virtual machines, Just-in-time compilers, and tools for diagnosing Java problems. He wrote an adapter for the Eclipse Memory Analyzer to read dumps from IBM VMs and is now a committer on that project.
Kevin Grigorenko is a software engineer on the WebSphere Application Server SWAT team, which provides worldwide, on-site, and remote supplemental product-defect support, particularly in critical customer-support situations. He focuses on problem determination for WebSphere Application Server and related stack products and operating systems, including the IBM and Oracle JVMs, AIX, Linux, Solaris, Windows, z/OS, HP/UX, and i5/OS.