[PATCH] lockdep: undo mm/slab.c annotation
undo existing mm/slab.c lock-validator annotations, in preparation of a new, less intrusive annotation patch. Signed-off-by: Ingo Molnar <mingo@elte.hu> Signed-off-by: Arjan van de Ven <arjan@linux.intel.com> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
parent
0e2ffbf650
commit
873623dfab
33
mm/slab.c
33
mm/slab.c
|
@ -1021,8 +1021,7 @@ static void drain_alien_cache(struct kmem_cache *cachep,
|
|||
}
|
||||
}
|
||||
|
||||
static inline int cache_free_alien(struct kmem_cache *cachep, void *objp,
|
||||
int nesting)
|
||||
static inline int cache_free_alien(struct kmem_cache *cachep, void *objp)
|
||||
{
|
||||
struct slab *slabp = virt_to_slab(objp);
|
||||
int nodeid = slabp->nodeid;
|
||||
|
@ -1040,7 +1039,7 @@ static inline int cache_free_alien(struct kmem_cache *cachep, void *objp,
|
|||
STATS_INC_NODEFREES(cachep);
|
||||
if (l3->alien && l3->alien[nodeid]) {
|
||||
alien = l3->alien[nodeid];
|
||||
spin_lock_nested(&alien->lock, nesting);
|
||||
spin_lock(&alien->lock);
|
||||
if (unlikely(alien->avail == alien->limit)) {
|
||||
STATS_INC_ACOVERFLOW(cachep);
|
||||
__drain_alien_cache(cachep, alien, nodeid);
|
||||
|
@ -1069,8 +1068,7 @@ static inline void free_alien_cache(struct array_cache **ac_ptr)
|
|||
{
|
||||
}
|
||||
|
||||
static inline int cache_free_alien(struct kmem_cache *cachep, void *objp,
|
||||
int nesting)
|
||||
static inline int cache_free_alien(struct kmem_cache *cachep, void *objp)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
|
@ -1760,8 +1758,6 @@ static void slab_destroy_objs(struct kmem_cache *cachep, struct slab *slabp)
|
|||
}
|
||||
#endif
|
||||
|
||||
static void __cache_free(struct kmem_cache *cachep, void *objp, int nesting);
|
||||
|
||||
/**
|
||||
* slab_destroy - destroy and release all objects in a slab
|
||||
* @cachep: cache pointer being destroyed
|
||||
|
@ -1785,17 +1781,8 @@ static void slab_destroy(struct kmem_cache *cachep, struct slab *slabp)
|
|||
call_rcu(&slab_rcu->head, kmem_rcu_free);
|
||||
} else {
|
||||
kmem_freepages(cachep, addr);
|
||||
if (OFF_SLAB(cachep)) {
|
||||
unsigned long flags;
|
||||
|
||||
/*
|
||||
* lockdep: we may nest inside an already held
|
||||
* ac->lock, so pass in a nesting flag:
|
||||
*/
|
||||
local_irq_save(flags);
|
||||
__cache_free(cachep->slabp_cache, slabp, 1);
|
||||
local_irq_restore(flags);
|
||||
}
|
||||
if (OFF_SLAB(cachep))
|
||||
kmem_cache_free(cachep->slabp_cache, slabp);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -3135,7 +3122,7 @@ static void cache_flusharray(struct kmem_cache *cachep, struct array_cache *ac)
|
|||
#endif
|
||||
check_irq_off();
|
||||
l3 = cachep->nodelists[node];
|
||||
spin_lock_nested(&l3->list_lock, SINGLE_DEPTH_NESTING);
|
||||
spin_lock(&l3->list_lock);
|
||||
if (l3->shared) {
|
||||
struct array_cache *shared_array = l3->shared;
|
||||
int max = shared_array->limit - shared_array->avail;
|
||||
|
@ -3178,14 +3165,14 @@ static void cache_flusharray(struct kmem_cache *cachep, struct array_cache *ac)
|
|||
* Release an obj back to its cache. If the obj has a constructed state, it must
|
||||
* be in this state _before_ it is released. Called with disabled ints.
|
||||
*/
|
||||
static void __cache_free(struct kmem_cache *cachep, void *objp, int nesting)
|
||||
static inline void __cache_free(struct kmem_cache *cachep, void *objp)
|
||||
{
|
||||
struct array_cache *ac = cpu_cache_get(cachep);
|
||||
|
||||
check_irq_off();
|
||||
objp = cache_free_debugcheck(cachep, objp, __builtin_return_address(0));
|
||||
|
||||
if (cache_free_alien(cachep, objp, nesting))
|
||||
if (cache_free_alien(cachep, objp))
|
||||
return;
|
||||
|
||||
if (likely(ac->avail < ac->limit)) {
|
||||
|
@ -3424,7 +3411,7 @@ void kmem_cache_free(struct kmem_cache *cachep, void *objp)
|
|||
BUG_ON(virt_to_cache(objp) != cachep);
|
||||
|
||||
local_irq_save(flags);
|
||||
__cache_free(cachep, objp, 0);
|
||||
__cache_free(cachep, objp);
|
||||
local_irq_restore(flags);
|
||||
}
|
||||
EXPORT_SYMBOL(kmem_cache_free);
|
||||
|
@ -3449,7 +3436,7 @@ void kfree(const void *objp)
|
|||
kfree_debugcheck(objp);
|
||||
c = virt_to_cache(objp);
|
||||
debug_check_no_locks_freed(objp, obj_size(c));
|
||||
__cache_free(c, (void *)objp, 0);
|
||||
__cache_free(c, (void *)objp);
|
||||
local_irq_restore(flags);
|
||||
}
|
||||
EXPORT_SYMBOL(kfree);
|
||||
|
|
Loading…
Reference in New Issue
Block a user