2
0
mirror of https://github.com/edk2-porting/linux-next.git synced 2024-12-19 10:44:14 +08:00
linux-next/include/trace/events/pagemap.h
Mel Gorman 24b7e5819a mm: pagemap: avoid unnecessary overhead when tracepoints are deactivated
This was formerly the series "Improve sequential read throughput" which
noted some major differences in performance of tiobench since 3.0.
While there are a number of factors, two that dominated were the
introduction of the fair zone allocation policy and changes to CFQ.

The behaviour of fair zone allocation policy makes more sense than
tiobench as a benchmark and CFQ defaults were not changed due to
insufficient benchmarking.

This series is what's left.  It's one functional fix to the fair zone
allocation policy when used on NUMA machines and a reduction of overhead
in general.  tiobench was used for the comparison despite its flaws as
an IO benchmark as in this case we are primarily interested in the
overhead of page allocator and page reclaim activity.

On UMA, it makes little difference to overhead

          3.16.0-rc3   3.16.0-rc3
             vanilla lowercost-v5
User          383.61      386.77
System        403.83      401.74
Elapsed      5411.50     5413.11

On a 4-socket NUMA machine it's a bit more noticable

          3.16.0-rc3   3.16.0-rc3
             vanilla lowercost-v5
User          746.94      802.00
System      65336.22    40852.33
Elapsed     27553.52    27368.46

This patch (of 6):

The LRU insertion and activate tracepoints take PFN as a parameter
forcing the overhead to the caller.  Move the overhead to the tracepoint
fast-assign method to ensure the cost is only incurred when the
tracepoint is active.

Signed-off-by: Mel Gorman <mgorman@suse.de>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2014-08-06 18:01:20 -07:00

88 lines
2.1 KiB
C

#undef TRACE_SYSTEM
#define TRACE_SYSTEM pagemap
#if !defined(_TRACE_PAGEMAP_H) || defined(TRACE_HEADER_MULTI_READ)
#define _TRACE_PAGEMAP_H
#include <linux/tracepoint.h>
#include <linux/mm.h>
#define PAGEMAP_MAPPED 0x0001u
#define PAGEMAP_ANONYMOUS 0x0002u
#define PAGEMAP_FILE 0x0004u
#define PAGEMAP_SWAPCACHE 0x0008u
#define PAGEMAP_SWAPBACKED 0x0010u
#define PAGEMAP_MAPPEDDISK 0x0020u
#define PAGEMAP_BUFFERS 0x0040u
#define trace_pagemap_flags(page) ( \
(PageAnon(page) ? PAGEMAP_ANONYMOUS : PAGEMAP_FILE) | \
(page_mapped(page) ? PAGEMAP_MAPPED : 0) | \
(PageSwapCache(page) ? PAGEMAP_SWAPCACHE : 0) | \
(PageSwapBacked(page) ? PAGEMAP_SWAPBACKED : 0) | \
(PageMappedToDisk(page) ? PAGEMAP_MAPPEDDISK : 0) | \
(page_has_private(page) ? PAGEMAP_BUFFERS : 0) \
)
TRACE_EVENT(mm_lru_insertion,
TP_PROTO(
struct page *page,
int lru
),
TP_ARGS(page, lru),
TP_STRUCT__entry(
__field(struct page *, page )
__field(unsigned long, pfn )
__field(int, lru )
__field(unsigned long, flags )
),
TP_fast_assign(
__entry->page = page;
__entry->pfn = page_to_pfn(page);
__entry->lru = lru;
__entry->flags = trace_pagemap_flags(page);
),
/* Flag format is based on page-types.c formatting for pagemap */
TP_printk("page=%p pfn=%lu lru=%d flags=%s%s%s%s%s%s",
__entry->page,
__entry->pfn,
__entry->lru,
__entry->flags & PAGEMAP_MAPPED ? "M" : " ",
__entry->flags & PAGEMAP_ANONYMOUS ? "a" : "f",
__entry->flags & PAGEMAP_SWAPCACHE ? "s" : " ",
__entry->flags & PAGEMAP_SWAPBACKED ? "b" : " ",
__entry->flags & PAGEMAP_MAPPEDDISK ? "d" : " ",
__entry->flags & PAGEMAP_BUFFERS ? "B" : " ")
);
TRACE_EVENT(mm_lru_activate,
TP_PROTO(struct page *page),
TP_ARGS(page),
TP_STRUCT__entry(
__field(struct page *, page )
__field(unsigned long, pfn )
),
TP_fast_assign(
__entry->page = page;
__entry->pfn = page_to_pfn(page);
),
/* Flag format is based on page-types.c formatting for pagemap */
TP_printk("page=%p pfn=%lu", __entry->page, __entry->pfn)
);
#endif /* _TRACE_PAGEMAP_H */
/* This part must be outside protection */
#include <trace/define_trace.h>