[PATCH v2 26/28] memcg: per-memcg kmem shrinking
Glauber Costa
glommer at parallels.com
Mon Apr 1 09:14:00 UTC 2013
On 04/01/2013 01:01 PM, Kamezawa Hiroyuki wrote:
> (2013/04/01 17:48), Glauber Costa wrote:
>>>> +static int memcg_try_charge_kmem(struct mem_cgroup *memcg, gfp_t gfp, u64 size)
>>>> +{
>>>> + int retries = MEM_CGROUP_RECLAIM_RETRIES;
>>>
>>> I'm not sure this retry numbers, for anon/file LRUs is suitable for kmem.
>>>
>> Suggestions ?
>>
>
> I think you did tests.
>
Indeed. And in my tests, 2 or 3 retries are already enough to seal the
fate of this.
I though it was safer to go with the same number, though, exactly not to
be too biased by my specific test environments.
I am fine with >= 3.
Michal, you have input here?
>>>> + struct res_counter *fail_res;
>>>> + int ret;
>>>> +
>>>> + do {
>>>> + ret = res_counter_charge(&memcg->kmem, size, &fail_res);
>>>> + if (!ret)
>>>> + return ret;
>>>> +
>>>> + if (!(gfp & __GFP_WAIT))
>>>> + return ret;
>>>> +
>>>> + /*
>>>> + * We will try to shrink kernel memory present in caches. We
>>>> + * are sure that we can wait, so we will. The duration of our
>>>> + * wait is determined by congestion, the same way as vmscan.c
>>>> + *
>>>> + * If we are in FS context, though, then although we can wait,
>>>> + * we cannot call the shrinkers. Most fs shrinkers (which
>>>> + * comprises most of our kmem data) will not run without
>>>> + * __GFP_FS since they can deadlock. The solution is to
>>>> + * synchronously run that in a different context.
>>>> + */
>>>> + if (!(gfp & __GFP_FS)) {
>>>> + /*
>>>> + * we are already short on memory, every queue
>>>> + * allocation is likely to fail
>>>> + */
>>>> + memcg_stop_kmem_account();
>>>> + schedule_work(&memcg->kmemcg_shrink_work);
>>>> + flush_work(&memcg->kmemcg_shrink_work);
>>>> + memcg_resume_kmem_account();
>>>> + } else if (!try_to_free_mem_cgroup_kmem(memcg, gfp))
>>>> + congestion_wait(BLK_RW_ASYNC, HZ/10);
>>>
>>> Why congestion_wait() ? I think calling congestion_wait() in vmscan.c is
>>> a part of memory-reclaim logic but I don't think the caller should do
>>> this kind of voluteer wait without good reason..
>>>
>>>
>>
>> Although it is not the case with dentries (or inodes, since only
>> non-dirty inodes goes to the lru list), some objects we are freeing may
>> need time to be written back to disk. This is the case for instance with
>> the buffer heads and bio's. They will not be actively shrunk in
>> shrinkers, but it is my understanding that they will be released. Inodes
>> as well, may have time to be written back and become non-dirty.
>>
>> In practice, in my tests, this would almost-always fail after a retry if
>> we don't wait, and almost always succeed in a retry if we do wait.
>>
>> Am I missing something in this interpretation ?
>>
>
> Ah, sorry. Can't we put this wait into try_to_free_mem_cgroup_kmem().
>
That I believe we can easily do.
More information about the Containers
mailing list