Wednesday, October 19, 2022

[389-users] Re: 389ds memory limits in kubernetes

Hi,

With limit 6Gi in 256Gi host, no warning or error in spal_meminfo_get.

[19/Oct/2022:08:17:29.636006003 +0000] - NOTICE - bdb_start_autotune - found 6291456k physical memory
[19/Oct/2022:08:17:29.636254777 +0000] - NOTICE - bdb_start_autotune - found 4555076k available
[19/Oct/2022:08:17:29.636386921 +0000] - NOTICE - bdb_start_autotune - cache autosizing: db cache: 393216k
[19/Oct/2022:08:17:29.636513687 +0000] - NOTICE - bdb_start_autotune - total cache size: 322122547 B;

bash from 389ds Pod

bash-5.1# cat /sys/fs/cgroup/memory/memory.limit_in_bytes
6442450944

bash-5.1# ps_mem

 Private  +   Shared  =  RAM used       Program
 ...
 28.0 MiB +   3.2 MiB =  31.2 MiB       dscontainer
  2.3 GiB +   1.8 MiB =   2.3 GiB       ns-slapd
---------------------------------
                          2.3 GiB
=================================

Private memory from process ns-slapd get increased

should we try to set DS_MEMORY_PERCENTAGE parameter?

Regards

El 17/10/2022 a las 13:42, Viktor Ashirov escribiรณ:
Hi,

On Mon, Oct 17, 2022 at 1:14 PM Juan R Moral <jmoral@ujaen.es> wrote:
Hello,

389ds docker image acts different in hosts with different amount of memory.

For example in minikube with 8Gi nodes and limit 6Gi work wells. But in
a kubernetes cluster with 256Gi nodes and limit 6Gi Pod is contantly killed.
Could you please share the startup log from the container from both minikube and kubernetes cluster? It prints memory autosizing values.
I noticed on my 128Gi node that it fails to fetch cgroups memory information and defaults to a large db cache value, even though I limited the amount of memory to 512Mi:

bash-5.2# cat /sys/fs/cgroup/memory.max
536870912

[17/Oct/2022:11:26:59.870355709 +0000] - WARN - spal_meminfo_get - cgroups v1 or v2 unable to be read - may not be on this platform ...
[17/Oct/2022:11:26:59.873439133 +0000] - WARN - spal_meminfo_get - cgroups v1 or v2 unable to be read - may not be on this platform ...
[17/Oct/2022:11:26:59.876112258 +0000] - WARN - spal_meminfo_get - cgroups v1 or v2 unable to be read - may not be on this platform ...
[17/Oct/2022:11:26:59.878714281 +0000] - NOTICE - bdb_start_autotune - found 131811248k physical memory
[17/Oct/2022:11:26:59.881188385 +0000] - NOTICE - bdb_start_autotune - found 127746920k available
[17/Oct/2022:11:26:59.883783867 +0000] - NOTICE - bdb_start_autotune - cache autosizing: db cache: 1572864k
[17/Oct/2022:11:26:59.886336572 +0000] - NOTICE - bdb_start_autotune - total cache size: 1610612736 B;
[17/Oct/2022:11:26:59.889198905 +0000] - WARN - spal_meminfo_get - cgroups v1 or v2 unable to be read - may not be on this platform ...

We might have a bug in the autosizing code.

Thanks.

Regards,


_______________________________________________
389-users mailing list -- 389-users@lists.fedoraproject.org
To unsubscribe send an email to 389-users-leave@lists.fedoraproject.org
Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives: https://lists.fedoraproject.org/archives/list/389-users@lists.fedoraproject.org
Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue


--
Viktor

_______________________________________________  389-users mailing list -- 389-users@lists.fedoraproject.org  To unsubscribe send an email to 389-users-leave@lists.fedoraproject.org  Fedora Code of Conduct: https://docs.fedoraproject.org/en-US/project/code-of-conduct/  List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines  List Archives: https://lists.fedoraproject.org/archives/list/389-users@lists.fedoraproject.org  Do not reply to spam, report it: https://pagure.io/fedora-infrastructure/new_issue  

No comments:

Post a Comment