mm/compaction: more trace to understand when/why compaction start/finish
authorJoonsoo Kim <iamjoonsoo.kim@lge.com>
Wed, 11 Feb 2015 23:27:06 +0000 (15:27 -0800)
committerLinus Torvalds <torvalds@linux-foundation.org>
Thu, 12 Feb 2015 01:06:04 +0000 (17:06 -0800)
It is not well analyzed that when/why compaction start/finish or not.
With these new tracepoints, we can know much more about start/finish
reason of compaction.  I can find following bug with these tracepoint.

http://www.spinics.net/lists/linux-mm/msg81582.html

Signed-off-by: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Mel Gorman <mgorman@suse.de>
Cc: David Rientjes <rientjes@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
include/linux/compaction.h
include/trace/events/compaction.h
mm/compaction.c

index db64cae065305f8b88901026203d109adf6c3a1b..501d7513aac1bba793e65d06c92bee9dde5d9205 100644 (file)
@@ -12,6 +12,9 @@
 #define COMPACT_PARTIAL                3
 /* The full zone was compacted */
 #define COMPACT_COMPLETE       4
+/* For more detailed tracepoint output */
+#define COMPACT_NO_SUITABLE_PAGE       5
+#define COMPACT_NOT_SUITABLE_ZONE      6
 /* When adding new state, please change compaction_status_string, too */
 
 /* Used to signal whether compaction detected need_sched() or lock contention */
index 139020b55612a5c2f8f8d1c51149fac4cbd96d0c..d46535801f63032bb505a2e41805f67c98e5186f 100644 (file)
@@ -164,6 +164,80 @@ TRACE_EVENT(mm_compaction_end,
                compaction_status_string[__entry->status])
 );
 
+TRACE_EVENT(mm_compaction_try_to_compact_pages,
+
+       TP_PROTO(
+               int order,
+               gfp_t gfp_mask,
+               enum migrate_mode mode),
+
+       TP_ARGS(order, gfp_mask, mode),
+
+       TP_STRUCT__entry(
+               __field(int, order)
+               __field(gfp_t, gfp_mask)
+               __field(enum migrate_mode, mode)
+       ),
+
+       TP_fast_assign(
+               __entry->order = order;
+               __entry->gfp_mask = gfp_mask;
+               __entry->mode = mode;
+       ),
+
+       TP_printk("order=%d gfp_mask=0x%x mode=%d",
+               __entry->order,
+               __entry->gfp_mask,
+               (int)__entry->mode)
+);
+
+DECLARE_EVENT_CLASS(mm_compaction_suitable_template,
+
+       TP_PROTO(struct zone *zone,
+               int order,
+               int ret),
+
+       TP_ARGS(zone, order, ret),
+
+       TP_STRUCT__entry(
+               __field(int, nid)
+               __field(char *, name)
+               __field(int, order)
+               __field(int, ret)
+       ),
+
+       TP_fast_assign(
+               __entry->nid = zone_to_nid(zone);
+               __entry->name = (char *)zone->name;
+               __entry->order = order;
+               __entry->ret = ret;
+       ),
+
+       TP_printk("node=%d zone=%-8s order=%d ret=%s",
+               __entry->nid,
+               __entry->name,
+               __entry->order,
+               compaction_status_string[__entry->ret])
+);
+
+DEFINE_EVENT(mm_compaction_suitable_template, mm_compaction_finished,
+
+       TP_PROTO(struct zone *zone,
+               int order,
+               int ret),
+
+       TP_ARGS(zone, order, ret)
+);
+
+DEFINE_EVENT(mm_compaction_suitable_template, mm_compaction_suitable,
+
+       TP_PROTO(struct zone *zone,
+               int order,
+               int ret),
+
+       TP_ARGS(zone, order, ret)
+);
+
 #endif /* _TRACE_COMPACTION_H */
 
 /* This part must be outside protection */
index b12df9fe10b4e59c488ba2a2484a7e9c8d3c2fee..b6ede459c1bbf5a70ebeadf0803ecc5784fa82cb 100644 (file)
@@ -41,6 +41,8 @@ static const char *const compaction_status_string[] = {
        "continue",
        "partial",
        "complete",
+       "no_suitable_page",
+       "not_suitable_zone",
 };
 #endif
 
@@ -1049,7 +1051,7 @@ static isolate_migrate_t isolate_migratepages(struct zone *zone,
        return cc->nr_migratepages ? ISOLATE_SUCCESS : ISOLATE_NONE;
 }
 
-static int compact_finished(struct zone *zone, struct compact_control *cc,
+static int __compact_finished(struct zone *zone, struct compact_control *cc,
                            const int migratetype)
 {
        unsigned int order;
@@ -1104,7 +1106,20 @@ static int compact_finished(struct zone *zone, struct compact_control *cc,
                        return COMPACT_PARTIAL;
        }
 
-       return COMPACT_CONTINUE;
+       return COMPACT_NO_SUITABLE_PAGE;
+}
+
+static int compact_finished(struct zone *zone, struct compact_control *cc,
+                           const int migratetype)
+{
+       int ret;
+
+       ret = __compact_finished(zone, cc, migratetype);
+       trace_mm_compaction_finished(zone, cc->order, ret);
+       if (ret == COMPACT_NO_SUITABLE_PAGE)
+               ret = COMPACT_CONTINUE;
+
+       return ret;
 }
 
 /*
@@ -1114,7 +1129,7 @@ static int compact_finished(struct zone *zone, struct compact_control *cc,
  *   COMPACT_PARTIAL  - If the allocation would succeed without compaction
  *   COMPACT_CONTINUE - If compaction should run now
  */
-unsigned long compaction_suitable(struct zone *zone, int order,
+static unsigned long __compaction_suitable(struct zone *zone, int order,
                                        int alloc_flags, int classzone_idx)
 {
        int fragindex;
@@ -1158,11 +1173,24 @@ unsigned long compaction_suitable(struct zone *zone, int order,
         */
        fragindex = fragmentation_index(zone, order);
        if (fragindex >= 0 && fragindex <= sysctl_extfrag_threshold)
-               return COMPACT_SKIPPED;
+               return COMPACT_NOT_SUITABLE_ZONE;
 
        return COMPACT_CONTINUE;
 }
 
+unsigned long compaction_suitable(struct zone *zone, int order,
+                                       int alloc_flags, int classzone_idx)
+{
+       unsigned long ret;
+
+       ret = __compaction_suitable(zone, order, alloc_flags, classzone_idx);
+       trace_mm_compaction_suitable(zone, order, ret);
+       if (ret == COMPACT_NOT_SUITABLE_ZONE)
+               ret = COMPACT_SKIPPED;
+
+       return ret;
+}
+
 static int compact_zone(struct zone *zone, struct compact_control *cc)
 {
        int ret;
@@ -1376,6 +1404,8 @@ unsigned long try_to_compact_pages(gfp_t gfp_mask, unsigned int order,
        if (!order || !may_enter_fs || !may_perform_io)
                return COMPACT_SKIPPED;
 
+       trace_mm_compaction_try_to_compact_pages(order, gfp_mask, mode);
+
        /* Compact each zone in the list */
        for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx,
                                                                ac->nodemask) {
This page took 0.029421 seconds and 5 git commands to generate.