05 Jun, 2014

40 commits

  • This introduces memblock_alloc_range() which allocates memblock from the
    specified range of physical address. I would like to use this function
    to specify the location of CMA.

    Signed-off-by: Akinobu Mita
    Cc: Marek Szyprowski
    Cc: Konrad Rzeszutek Wilk
    Cc: David Woodhouse
    Cc: Don Dutile
    Cc: Thomas Gleixner
    Cc: Ingo Molnar
    Cc: "H. Peter Anvin"
    Cc: Andi Kleen
    Cc: Yinghai Lu
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Akinobu Mita
     
  • This adds support for the DMA Contiguous Memory Allocator for
    intel-iommu. This change enables dma_alloc_coherent() to allocate big
    contiguous memory.

    It is achieved in the same way as nommu_dma_ops currently does, i.e.
    trying to allocate memory by dma_alloc_from_contiguous() and
    alloc_pages() is used as a fallback.

    Signed-off-by: Akinobu Mita
    Cc: Marek Szyprowski
    Cc: Konrad Rzeszutek Wilk
    Cc: David Woodhouse
    Cc: Don Dutile
    Cc: Thomas Gleixner
    Cc: Ingo Molnar
    Cc: "H. Peter Anvin"
    Cc: Andi Kleen
    Cc: Yinghai Lu
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Akinobu Mita
     
  • The DMA Contiguous Memory Allocator support on x86 is disabled when
    swiotlb config option is enabled. So DMA CMA is always disabled on
    x86_64 because swiotlb is always enabled. This attempts to support for
    DMA CMA with enabling swiotlb config option.

    The contiguous memory allocator on x86 is integrated in the function
    dma_generic_alloc_coherent() which is .alloc callback in nommu_dma_ops
    for dma_alloc_coherent().

    x86_swiotlb_alloc_coherent() which is .alloc callback in swiotlb_dma_ops
    tries to allocate with dma_generic_alloc_coherent() firstly and then
    swiotlb_alloc_coherent() is called as a fallback.

    The main part of supporting DMA CMA with swiotlb is that changing
    x86_swiotlb_free_coherent() which is .free callback in swiotlb_dma_ops
    for dma_free_coherent() so that it can distinguish memory allocated by
    dma_generic_alloc_coherent() from one allocated by
    swiotlb_alloc_coherent() and release it with dma_generic_free_coherent()
    which can handle contiguous memory. This change requires making
    is_swiotlb_buffer() global function.

    This also needs to change .free callback in the dma_map_ops for amd_gart
    and sta2x11, because these dma_ops are also using
    dma_generic_alloc_coherent().

    Signed-off-by: Akinobu Mita
    Acked-by: Marek Szyprowski
    Acked-by: Konrad Rzeszutek Wilk
    Cc: David Woodhouse
    Cc: Don Dutile
    Cc: Thomas Gleixner
    Cc: Ingo Molnar
    Cc: "H. Peter Anvin"
    Cc: Andi Kleen
    Cc: Yinghai Lu
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Akinobu Mita
     
  • This patchset enhances the DMA Contiguous Memory Allocator on x86.

    Currently the DMA CMA is only supported with pci-nommu dma_map_ops and
    furthermore it can't be enabled on x86_64. But I would like to allocate
    big contiguous memory with dma_alloc_coherent() and tell it to the device
    that requires it, regardless of which dma mapping implementation is
    actually used in the system.

    So this makes it work with swiotlb and intel-iommu dma_map_ops, too. And
    this also extends "cma=" kernel parameter to specify placement constraint
    by the physical address range of memory allocations. For example, CMA
    allocates memory below 4GB by "cma=64M@0-4G", it is required for the
    devices only supporting 32-bit addressing on 64-bit systems without iommu.

    This patch (of 5):

    Calling dma_alloc_coherent() with __GFP_ZERO must return zeroed memory.

    But when the contiguous memory allocator (CMA) is enabled on x86 and the
    memory region is allocated by dma_alloc_from_contiguous(), it doesn't
    return zeroed memory. Because dma_generic_alloc_coherent() forgot to fill
    the memory region with zero if it was allocated by
    dma_alloc_from_contiguous()

    Most implementations of dma_alloc_coherent() return zeroed memory
    regardless of whether __GFP_ZERO is specified. So this fixes it by
    unconditionally zeroing the allocated memory region.

    Alternatively, we could fix dma_alloc_from_contiguous() to return zeroed
    out memory and remove memset() from all caller of it. But we can't simply
    remove the memset on arm because __dma_clear_buffer() is used there for
    ensuring cache flushing and it is used in many places. Of course we can
    do redundant memset in dma_alloc_from_contiguous(), but I think this patch
    is less impact for fixing this problem.

    Signed-off-by: Akinobu Mita
    Cc: Marek Szyprowski
    Cc: Konrad Rzeszutek Wilk
    Cc: David Woodhouse
    Cc: Don Dutile
    Cc: Thomas Gleixner
    Cc: Ingo Molnar
    Cc: "H. Peter Anvin"
    Cc: Andi Kleen
    Cc: Yinghai Lu
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Akinobu Mita
     
  • For single threaded workloads, we can avoid flushing and iterating through
    the entire list of tasks, making the whole function a lot faster,
    requiring only a single atomic read for the mm_users.

    Signed-off-by: Davidlohr Bueso
    Suggested-by: Oleg Nesterov
    Cc: Aswin Chandramouleeswaran
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Davidlohr Bueso
     
  • Introduce a CONFIG_DEBUG_VM_VMACACHE option to enable counting the cache
    hit rate -- exported in /proc/vmstat.

    Any updates to the caching scheme needs this kind of data, thus it can
    save some work re-implementing the counting all the time.

    Signed-off-by: Davidlohr Bueso
    Cc: Aswin Chandramouleeswaran
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Davidlohr Bueso
     
  • Prior to this change, we would decide whether to force scan a LRU during
    reclaim if that LRU itself was too small for the current priority.
    However, this can lead to the file LRU getting force scanned even if
    there are a lot of anonymous pages we can reclaim, leading to hot file
    pages getting needlessly reclaimed.

    To address this, we instead only force scan when none of the reclaimable
    LRUs are big enough.

    Gives huge improvements with zswap. For example, when doing -j20 kernel
    build in a 500MB container with zswap enabled, runtime (in seconds) is
    greatly reduced:

    x without this change
    + with this change
    N Min Max Median Avg Stddev
    x 5 700.997 790.076 763.928 754.05 39.59493
    + 5 141.634 197.899 155.706 161.9 21.270224
    Difference at 95.0% confidence
    -592.15 +/- 46.3521
    -78.5293% +/- 6.14709%
    (Student's t, pooled s = 31.7819)

    Should also give some improvements in regular (non-zswap) swap cases.

    Yes, hughd found significant speedup using regular swap, with several
    memcgs under pressure; and it should also be effective in the non-memcg
    case, whenever one or another zone LRU is forced too small.

    Signed-off-by: Suleiman Souhlal
    Signed-off-by: Hugh Dickins
    Cc: Suleiman Souhlal
    Cc: Mel Gorman
    Acked-by: Rik van Riel
    Acked-by: Rafael Aquini
    Cc: Michal Hocko
    Cc: Yuanhan Liu
    Cc: Seth Jennings
    Cc: Bob Liu
    Cc: Minchan Kim
    Cc: Luigi Semenzato
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Suleiman Souhlal
     
  • clear_refs_write() is called earlier than clear_soft_dirty() and it is
    more natural to clear VM_SOFTDIRTY (which belongs to VMA entry but not
    PTEs) that early instead of clearing it a way deeper inside call chain.

    Signed-off-by: Cyrill Gorcunov
    Cc: Pavel Emelyanov
    Cc: Hugh Dickins
    Cc: Naoya Horiguchi
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Cyrill Gorcunov
     
  • pte_file_mksoft_dirty operates with argument passed by a value and
    returns modified result thus we need to assign @ptfile here, otherwise
    itis a no-op which may lead to loss of the softdirty bit.

    Signed-off-by: Cyrill Gorcunov
    Cc: Pavel Emelyanov
    Cc: Hugh Dickins
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Cyrill Gorcunov
     
  • Hugh reported:

    | I noticed your soft_dirty work in install_file_pte(): which looked
    | good at first, until I realized that it's propagating the soft_dirty
    | of a pte it's about to zap completely, to the unrelated entry it's
    | about to insert in its place. Which seems very odd to me.

    Indeed this code ends up being nop in result -- pte_file_mksoft_dirty()
    operates with pte_t argument and returns new pte_t which were never used
    after. After looking more I think what we need is to soft-dirtify all
    newely remapped file pages because it should look like a new mapping for
    memory tracker.

    Signed-off-by: Cyrill Gorcunov
    Reported-by: Hugh Dickins
    Cc: Pavel Emelyanov
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Cyrill Gorcunov
     
  • Currently to allocate a page that should be charged to kmemcg (e.g.
    threadinfo), we pass __GFP_KMEMCG flag to the page allocator. The page
    allocated is then to be freed by free_memcg_kmem_pages. Apart from
    looking asymmetrical, this also requires intrusion to the general
    allocation path. So let's introduce separate functions that will
    alloc/free pages charged to kmemcg.

    The new functions are called alloc_kmem_pages and free_kmem_pages. They
    should be used when the caller actually would like to use kmalloc, but
    has to fall back to the page allocator for the allocation is large.
    They only differ from alloc_pages and free_pages in that besides
    allocating or freeing pages they also charge them to the kmem resource
    counter of the current memory cgroup.

    [sfr@canb.auug.org.au: export kmalloc_order() to modules]
    Signed-off-by: Vladimir Davydov
    Acked-by: Greg Thelen
    Cc: Johannes Weiner
    Acked-by: Michal Hocko
    Cc: Glauber Costa
    Cc: Christoph Lameter
    Cc: Pekka Enberg
    Signed-off-by: Stephen Rothwell
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Vladimir Davydov
     
  • We have only a few places where we actually want to charge kmem so
    instead of intruding into the general page allocation path with
    __GFP_KMEMCG it's better to explictly charge kmem there. All kmem
    charges will be easier to follow that way.

    This is a step towards removing __GFP_KMEMCG. It removes __GFP_KMEMCG
    from memcg caches' allocflags. Instead it makes slab allocation path
    call memcg_charge_kmem directly getting memcg to charge from the cache's
    memcg params.

    This also eliminates any possibility of misaccounting an allocation
    going from one memcg's cache to another memcg, because now we always
    charge slabs against the memcg the cache belongs to. That's why this
    patch removes the big comment to memcg_kmem_get_cache.

    Signed-off-by: Vladimir Davydov
    Acked-by: Greg Thelen
    Cc: Johannes Weiner
    Acked-by: Michal Hocko
    Cc: Glauber Costa
    Cc: Christoph Lameter
    Cc: Pekka Enberg
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Vladimir Davydov
     
  • There used to be only one path out of __slab_alloc(), and ALLOC_SLOWPATH
    got bumped in that exit path. Now there are two, and a bunch of gotos.
    ALLOC_SLOWPATH can now get set more than once during a single call to
    __slab_alloc() which is pretty bogus. Here's the sequence:

    1. Enter __slab_alloc(), fall through all the way to the
    stat(s, ALLOC_SLOWPATH);
    2. hit 'if (!freelist)', and bump DEACTIVATE_BYPASS, jump to
    new_slab (goto #1)
    3. Hit 'if (c->partial)', bump CPU_PARTIAL_ALLOC, goto redo
    (goto #2)
    4. Fall through in the same path we did before all the way to
    stat(s, ALLOC_SLOWPATH)
    5. bump ALLOC_REFILL stat, then return

    Doing this is obviously bogus. It keeps us from being able to
    accurately compare ALLOC_SLOWPATH vs. ALLOC_FASTPATH. It also means
    that the total number of allocs always exceeds the total number of
    frees.

    This patch moves stat(s, ALLOC_SLOWPATH) to be called from the same
    place that __slab_alloc() is. This makes it much less likely that
    ALLOC_SLOWPATH will get botched again in the spaghetti-code inside
    __slab_alloc().

    Signed-off-by: Dave Hansen
    Acked-by: Christoph Lameter
    Acked-by: David Rientjes
    Cc: Pekka Enberg
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Dave Hansen
     
  • When the slab or slub allocators cannot allocate additional slab pages,
    they emit diagnostic information to the kernel log such as current
    number of slabs, number of objects, active objects, etc. This is always
    coupled with a page allocation failure warning since it is controlled by
    !__GFP_NOWARN.

    Suppress this out of memory warning if the allocator is configured
    without debug supported. The page allocation failure warning will
    indicate it is a failed slab allocation, the order, and the gfp mask, so
    this is only useful to diagnose allocator issues.

    Since CONFIG_SLUB_DEBUG is already enabled by default for the slub
    allocator, there is no functional change with this patch. If debug is
    disabled, however, the warnings are now suppressed.

    Signed-off-by: David Rientjes
    Cc: Pekka Enberg
    Acked-by: Christoph Lameter
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    David Rientjes
     
  • Inspired by Joe Perches suggestion in ntfs logging clean-up.

    Signed-off-by: Fabian Frederick
    Acked-by: Christoph Lameter
    Cc: Joe Perches
    Cc: Pekka Enberg
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • All printk(KERN_foo converted to pr_foo()

    Default printk converted to pr_warn()

    Coalesce format fragments

    Signed-off-by: Fabian Frederick
    Acked-by: Christoph Lameter
    Cc: Joe Perches
    Cc: Pekka Enberg
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • On system with 2TiB ram, current x86_64 have 128M as section size, and
    one memory_block only include one section. So will have 16400 entries
    under /sys/devices/system/memory/.

    Current code try to use block id to find block pointer in /sys for any
    section, and reuse that block pointer. that finding will take some time
    even after commit 7c243c7168dc ("mm: speedup in __early_pfn_to_nid")
    that will skip the search in that case during booting up.

    So solution could be increase block size just like SGI UV system did.
    (harded code to 2g).

    This patch is trying to probe the block size to make it match mmio remap
    size. for example, Intel Nehalem later system will have memory range [0,
    TOML), [4g, TOMH]. If the memory hole is 2g and total is 128g, TOM will
    be 2g, and TOM2 will be 130g.

    We could use 2g as block size instead of default 128M. That will reduce
    number of entries in /sys/devices/system/memory/

    On system 6TiB system will reduce boot time by 35 seconds.

    Signed-off-by: Yinghai Lu
    Cc: Ingo Molnar
    Cc: "H. Peter Anvin"
    Cc: Thomas Gleixner
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Yinghai Lu
     
  • _PAGE_NUMA is currently an alias of _PROT_PROTNONE to trap NUMA hinting
    faults on x86. Care is taken such that _PAGE_NUMA is used only in
    situations where the VMA flags distinguish between NUMA hinting faults
    and prot_none faults. This decision was x86-specific and conceptually
    it is difficult requiring special casing to distinguish between PROTNONE
    and NUMA ptes based on context.

    Fundamentally, we only need the _PAGE_NUMA bit to tell the difference
    between an entry that is really unmapped and a page that is protected
    for NUMA hinting faults as if the PTE is not present then a fault will
    be trapped.

    Swap PTEs on x86-64 use the bits after _PAGE_GLOBAL for the offset.
    This patch shrinks the maximum possible swap size and uses the bit to
    uniquely distinguish between NUMA hinting ptes and swap ptes.

    Signed-off-by: Mel Gorman
    Cc: David Vrabel
    Cc: Ingo Molnar
    Cc: Peter Anvin
    Cc: Fengguang Wu
    Cc: Linus Torvalds
    Cc: Steven Noonan
    Cc: Rik van Riel
    Cc: Peter Zijlstra
    Cc: Andrea Arcangeli
    Cc: Dave Hansen
    Cc: Srikar Dronamraju
    Cc: Cyrill Gorcunov
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Mel Gorman
     
  • 32-bit support for NUMA is an oddity on its own but with automatic NUMA
    balancing on top there is a reasonable risk that the CPUPID information
    cannot be stored in the page flags. This patch removes support for
    automatic NUMA support on 32-bit x86.

    Signed-off-by: Mel Gorman
    Cc: David Vrabel
    Cc: Ingo Molnar
    Cc: Peter Anvin
    Cc: Fengguang Wu
    Cc: Linus Torvalds
    Cc: Steven Noonan
    Cc: Rik van Riel
    Cc: Peter Zijlstra
    Cc: Andrea Arcangeli
    Cc: Dave Hansen
    Cc: Srikar Dronamraju
    Cc: Cyrill Gorcunov
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Mel Gorman
     
  • Description by Jan Kara:
    "A lot of older filesystems don't properly flush volatile disk caches
    on fsync(2) which can lead to loss of fsynced data after power failure.

    This patch makes generic_file_fsync() issue proper cache flush to fix the
    problem. Sysadmin can use /sys/devices/.../cache_type to tell the system
    it should not send the cache flush."

    [akpm@linux-foundation.org: nuke ifdef]
    [akpm@linux-foundation.org: fix warning]
    Signed-off-by: Fabian Frederick
    Suggested-by: Jan Kara
    Suggested-by: Christoph Hellwig
    Cc: Jan Kara
    Cc: Christoph Hellwig
    Cc: Alexander Viro
    Cc: "Theodore Ts'o"
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • Function parameters comment fixing.

    Signed-off-by: Fabian Frederick
    Cc: Eric Van Hensbergen
    Cc: Randy Dunlap
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • v9fs_sysfs_init is only called by __init init_v9fs

    Signed-off-by: Fabian Frederick
    Cc: Eric Van Hensbergen
    Cc: Ron Minnich
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • dlm_recovery_ctxt.received is unused.

    ocfs2_should_refresh_lock_res() can only return 0 or 1, so the error
    handling code in ocfs2_super_lock() is unneeded.

    Signed-off-by: joyce.xue
    Cc: Joel Becker
    Cc: Mark Fasheh
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Xue jiufei
     
  • Ocfs2 cluster size may be 1MB, which has 20 bits. When resize, the
    input new clusters is mostly the number of clusters in a group
    descriptor(32256).

    Since the input clusters is defined as type int, so it will overflow
    when shift left 20 bits and then lead to incorrect global bitmap i_size.

    Signed-off-by: Joseph Qi
    Cc: Mark Fasheh
    Cc: Joel Becker
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Joseph Qi
     
  • Parameters new_clusters and first_new_cluster are not used in
    ocfs2_update_last_group_and_inode, so remove them.

    Signed-off-by: Joseph Qi
    Reviewed-by: joyce.xue
    Cc: Mark Fasheh
    Cc: Joel Becker
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Joseph Qi
     
  • We found a race situation when dlm recovery and node joining occurs
    simultaneously if the network state is bad.

    N1 N4

    start joining dlm and send
    query join to all live nodes
    set joining node to N1, return OK
    send query join to other
    live nodes and it may take
    a while

    call dlm_send_join_assert()
    to send assert join message
    when N2 is down, so keep
    trying to send message to N2
    until find N2 is down

    send assert join message to
    N3, but connection is down
    with N3, so it may take a
    while
    become the recovery master for N2
    and send begin reco message to other
    nodes in domain map but no N1
    connection with N3 is rebuild,
    then send assert join to N4
    call dlm_assert_joined_handler(),
    add N1 to domain_map

    dlm recovery done, send finalize message
    to nodes in domain map, including N1
    receiving finalize message,
    trigger the BUG() because
    recovery master mismatch.

    Signed-off-by: joyce.xue
    Cc: Mark Fasheh
    Cc: Joel Becker
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Xue jiufei
     
  • Revert commit 75f82eaa502c ("ocfs2: fix NULL pointer dereference when
    dismount and ocfs2rec simultaneously") because it may cause a umount
    hang while shutting down the truncate log.

    fix NULL pointer dereference when dismount and ocfs2rec simultaneously

    The situation is as followes:
    ocfs2_dismout_volume
    -> ocfs2_recovery_exit
    -> free osb->recovery_map
    -> ocfs2_truncate_shutdown
    -> lock global bitmap inode
    -> ocfs2_wait_for_recovery
    -> check whether osb->recovery_map->rm_used is zero

    Because osb->recovery_map is already freed, rm_used can be any other
    values, so it may yield umount hang.

    To prevent NULL pointer dereference while getting sys_root_inode, we use
    a osb_tl_disable flag to disable schedule osb_truncate_log_wq after
    truncate log shutdown.

    Signed-off-by: joyce.xue
    Cc: Mark Fasheh
    Cc: Joel Becker
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Xue jiufei
     
  • ocfs_info_foo() and ocfs2_get_request_ptr functions are only used in ioctl.c

    Signed-off-by: Fabian Frederick
    Cc: Mark Fasheh
    Cc: Joel Becker
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • We found there is a conversion deadlock when the owner of lockres
    happened to crash before send DLM_PROXY_AST_MSG for a downconverting
    lock. The situation is as follows:

    Node1 Node2 Node3
    the owner of lockresA
    lock_1 granted at EX mode
    and call ocfs2_cluster_unlock
    to decrease ex_holders.
    converting lock_3 from
    NL to EX
    send DLM_PROXY_AST_MSG
    to Node1, asking Node 1
    to downconvert.
    receiving DLM_PROXY_AST_MSG,
    thread ocfs2dc send
    DLM_CONVERT_LOCK_MSG
    to Node2 to downconvert
    lock_1(EX->NL).
    lock_1 can be granted and
    put it into pending_asts
    list, return DLM_NORMAL.
    then something happened
    and Node2 crashed.
    received DLM_NORMAL, waiting
    for DLM_PROXY_AST_MSG.
    selected as the recovery
    master, receving migrate
    lock from Node1, queue
    lock_1 to the tail of
    converting list.

    After dlm recovery, converting list in the master of lockresA(Node3)
    will be: converting list head lock_3(NL->EX) lock_1(EXNL).
    Requested mode of lock_3 is not compatible with the granted mode of
    lock_1, so it can not be granted. and lock_1 can not downconvert
    because covnerting queue is strictly FIFO. So a deadlock is created.
    We think function dlm_process_recovery_data() should queue_ast for
    lock_1 or alter the order of lock_1 and lock_3, so dlm_thread can
    process lock_1 first. And if there are multiple downconverting locks,
    they must convert form PR to NL, so no need to sort them.

    Signed-off-by: joyce.xue
    Cc: Mark Fasheh
    Cc: Joel Becker
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Xue jiufei
     
  • Once JBD2_ABORT is set, ocfs2_commit_cache will fail in
    ocfs2_commit_thread. Then it will get into a loop with mass logs. This
    will meaninglessly consume a larger number of resource and may lead to
    the system hanging. So limit printk in this case.

    [akpm@linux-foundation.org: document the msleep]
    Signed-off-by: Joseph Qi
    Cc: Mark Fasheh
    Cc: Joel Becker
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Joseph Qi
     
  • There are two standard techniques for dereferencing structures pointed
    to by void *: cast to the right type each time they're used, or assign
    to local variables of the right type.

    But there's no need to do *both*.

    Signed-off-by: George Spelvin
    Cc: Mark Fasheh
    Acked-by: Joel Becker
    Reviewed-by: Jie Liu
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    George Spelvin
     
  • Replace strncpy(size 63) by defined value.

    Signed-off-by: Fabian Frederick
    Cc: Joel Becker
    Cc: Mark Fasheh
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • Static values are automatically initialized to NULL.

    Signed-off-by: Fabian Frederick
    Cc: Joel Becker
    Cc: Mark Fasheh
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • Direct conversion of one KERN_DEBUG message without DEBUG definition
    (suggested by Josh Triplett)

    That message will now be disabled by default. (see
    Documentation/CodingStyle Chapter 13)

    Signed-off-by: Fabian Frederick
    Reviewed-by: Josh Triplett
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • Add ODEBUG: prefix to pr_fmt

    Signed-off-by: Fabian Frederick
    Reviewed-by: Josh Triplett
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • Convert all printk to pr_foo() except KERN_DEBUG (see
    Documentation/CodingStyle Chapter 13)

    Signed-off-by: Fabian Frederick
    Reviewed-by: Josh Triplett
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • Add pr_fmt based on module name.

    Signed-off-by: Fabian Frederick
    Cc: Joel Becker
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • Signed-off-by: Fabian Frederick
    Cc: Joel Becker
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • Fix function parameter documentation

    EXPORT_SYMBOLS moved after corresponding functions

    Small coding style and checkpatch warning fixes

    Signed-off-by: Fabian Frederick
    Acked-by: Joel Becker
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Fabian Frederick
     
  • __uc32_ioremap_pfn_caller() should return NULL when the pfn is found to be
    invalid.

    From a recommendation by Guan Xuetao.

    Cc: Guan Xuetao
    Cc: Fabian Frederick
    Signed-off-by: Andrew Morton
    Signed-off-by: Linus Torvalds

    Andrew Morton