Hi,
On Wed, Oct 19, 2022 at 11:49 AM Juan R Moral <jmoral(a)ujaen.es> wrote:
Hi,
With limit 6Gi in 256Gi host, no warning or error in spal_meminfo_get.
[19/Oct/2022:08:17:29.636006003 +0000] - NOTICE - bdb_start_autotune -
found 6291456k physical memory
[19/Oct/2022:08:17:29.636254777 +0000] - NOTICE - bdb_start_autotune -
found 4555076k available
[19/Oct/2022:08:17:29.636386921 +0000] - NOTICE - bdb_start_autotune -
cache autosizing: db cache: 393216k
[19/Oct/2022:08:17:29.636513687 +0000] - NOTICE - bdb_start_autotune -
total cache size: 322122547 B;
bash from 389ds Pod
I noticed that bash is older in your container. Could you please confirm
the version of 389-ds-base that you have in your container?
(run inside the 389ds Pod) rpm -q 389-ds-base
bash-5.1# cat /sys/fs/cgroup/memory/memory.limit_in_bytes
6442450944
bash-5.1# ps_mem
Private + Shared = RAM used Program
...
28.0 MiB + 3.2 MiB = 31.2 MiB dscontainer
2.3 GiB + 1.8 MiB = 2.3 GiB ns-slapd
---------------------------------
2.3 GiB
=================================
Private memory from process ns-slapd get increased
should we try to set DS_MEMORY_PERCENTAGE parameter?
You can try, but I have a suspicion that it's a memory leak. 4k is a small
number of entries, they should fit in 6Gi of RAM.
Just to confirm the reproducer:
You have 4k entries, you run ldapsearch command from the first message in a
loop, and over several hours the resident memory of ns-slapd increases?
Thanks.
Regards
El 17/10/2022 a las 13:42, Viktor Ashirov escribió:
Hi,
On Mon, Oct 17, 2022 at 1:14 PM Juan R Moral <jmoral(a)ujaen.es> wrote:
> Hello,
>
> 389ds docker image acts different in hosts with different amount of
> memory.
>
> For example in minikube with 8Gi nodes and limit 6Gi work wells. But in
> a kubernetes cluster with 256Gi nodes and limit 6Gi Pod is contantly
> killed.
>
Could you please share the startup log from the container from both
minikube and kubernetes cluster? It prints memory autosizing values.
I noticed on my 128Gi node that it fails to fetch cgroups memory
information and defaults to a large db cache value, even though I limited
the amount of memory to 512Mi:
bash-5.2# cat /sys/fs/cgroup/memory.max
536870912
[17/Oct/2022:11:26:59.870355709 +0000] - WARN - spal_meminfo_get - cgroups
v1 or v2 unable to be read - may not be on this platform ...
[17/Oct/2022:11:26:59.873439133 +0000] - WARN - spal_meminfo_get - cgroups
v1 or v2 unable to be read - may not be on this platform ...
[17/Oct/2022:11:26:59.876112258 +0000] - WARN - spal_meminfo_get - cgroups
v1 or v2 unable to be read - may not be on this platform ...
[17/Oct/2022:11:26:59.878714281 +0000] - NOTICE - bdb_start_autotune -
found 131811248k physical memory
[17/Oct/2022:11:26:59.881188385 +0000] - NOTICE - bdb_start_autotune -
found 127746920k available
[17/Oct/2022:11:26:59.883783867 +0000] - NOTICE - bdb_start_autotune -
cache autosizing: db cache: 1572864k
[17/Oct/2022:11:26:59.886336572 +0000] - NOTICE - bdb_start_autotune -
total cache size: 1610612736 B;
[17/Oct/2022:11:26:59.889198905 +0000] - WARN - spal_meminfo_get - cgroups
v1 or v2 unable to be read - may not be on this platform ...
We might have a bug in the autosizing code.
Thanks.
>
> Regards,
>
>
> _______________________________________________
> 389-users mailing list -- 389-users(a)lists.fedoraproject.org
> To unsubscribe send an email to 389-users-leave(a)lists.fedoraproject.org
> Fedora Code of Conduct:
>
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
> List Guidelines:
https://fedoraproject.org/wiki/Mailing_list_guidelines
> List Archives:
>
https://lists.fedoraproject.org/archives/list/389-users@lists.fedoraproje...
> Do not reply to spam, report it:
>
https://pagure.io/fedora-infrastructure/new_issue
>
--
Viktor
_______________________________________________
389-users mailing list -- 389-users(a)lists.fedoraproject.org
To unsubscribe send an email to 389-users-leave(a)lists.fedoraproject.org
Fedora Code of Conduct:
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines:
https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives:
https://lists.fedoraproject.org/archives/list/389-users@lists.fedoraproje...
Do not reply to spam, report it:
https://pagure.io/fedora-infrastructure/new_issue
_______________________________________________
389-users mailing list -- 389-users(a)lists.fedoraproject.org
To unsubscribe send an email to 389-users-leave(a)lists.fedoraproject.org
Fedora Code of Conduct:
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines:
https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives:
https://lists.fedoraproject.org/archives/list/389-users@lists.fedoraproje...
Do not reply to spam, report it:
https://pagure.io/fedora-infrastructure/new_issue
--
Viktor