[OpenSIPS-Users] cluster presence
volga629 at networklab.ca
volga629 at networklab.ca
Thu Aug 16 00:12:00 EDT 2018
Hello Bogdan,
I was monitoring and trasfered config to dev machine where 0 load or
connection and it look like some sort of mini crash, but I can't tell
100%.
Here are log, the first line mean INVITE 200 OK reply then all those
messages start showing up and 200 OK never reach end point.
Aug 15 20:12:14 aitossbc03 /usr/sbin/opensips[29772]: OnReply_Route3:
[INVITE] Direction: [FS ~> Client] and source IP pbx ip
Aug 15 20:12:14 aitossbc03 /usr/sbin/opensips[29765]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already scheduled
100 ms ago (now 571950 ms), delaying execution
Aug 15 20:12:14 aitossbc03 /usr/sbin/opensips[29765]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already scheduled
200 ms ago (now 572050 ms), delaying execution
Aug 15 20:12:14 aitossbc03 /usr/sbin/opensips[29765]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already scheduled
300 ms ago (now 572150 ms), delaying execution
Aug 15 20:12:15 aitossbc03 /usr/sbin/opensips[29765]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already scheduled
400 ms ago (now 572250 ms), delaying execution
Aug 15 20:12:15 aitossbc03 /usr/sbin/opensips[29765]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already scheduled
500 ms ago (now 572350 ms), delaying execution
Aug 15 20:12:15 aitossbc03 /usr/sbin/opensips[29765]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already scheduled
600 ms ago (now 572450 ms), delaying execution
Aug 15 20:12:15 aitossbc03 /usr/sbin/opensips[29765]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already scheduled
700 ms ago (now 572550 ms), delaying execution
Aug 15 20:12:15 aitossbc03 /usr/sbin/opensips[29765]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already scheduled
800 ms ago (now 572650 ms), delaying execution
Aug 15 20:12:15 aitossbc03 /usr/sbin/opensips[29765]:
WARNING:core:utimer_ticker: utimer task <tm-utimer> already scheduled
900 ms ago (now 572750 ms), delaying execution
On Thu, Aug 9, 2018 at 6:54 AM, Bogdan-Andrei Iancu
<bogdan at opensips.org> wrote:
> Hi Volga,
>
> The logs report a serious heavy execution of certain timer routines
> in OpenSIPS, like the presence cleanup takes more than 191
> secs....most probably due long lasting DB queries ??? or because of
> heavy load in OpenSIPS that leads to starvation in handling the timer
> jobs - what is the internal load of OpenSIPS ? (use the 'load:' class
> of statistics to check it)
>
> Regards,
>
> Bogdan-Andrei Iancu
>
> OpenSIPS Founder and Developer
> http://www.opensips-solutions.com
> OpenSIPS Bootcamp 2018
> http://opensips.org/training/OpenSIPS_Bootcamp_2018/
>
> On 07/01/2018 07:28 AM, volga629 at networklab.ca wrote:
>> Hello Bogdan,
>> I checked database connection and it looks normal ping less then sec
>> to database.
>> I monitored load on database nodes and didn't notcied any extra load
>> on them.
>> Right now I see 2 types of messages on all cluster nodes.
>>
>>
>> Jun 30 22:56:45 aitossbc01 /usr/sbin/opensips[20245]:
>> WARNING:core:timer_ticker: timer task <presence-dbupdate> already
>> scheduled 100410 ms ago (now 700510 ms), skipping execution
>> Jun 30 22:56:45 aitossbc01 /usr/sbin/opensips[20245]:
>> WARNING:core:timer_ticker: timer task <presence-pclean> already
>> scheduled 191700 ms ago (now 700510 ms), delaying execution
>>
>> and
>>
>> Jun 30 23:14:36 aitossbc02 /usr/sbin/opensips[9376]:
>> WARNING:core:utimer_ticker: utimer task <tm-utimer> already
>> scheduled 3880 ms ago (now 113380 ms), delaying execution
>> Jun 30 23:14:36 aitossbc02 /usr/sbin/opensips[9376]:
>> WARNING:core:utimer_ticker: utimer task <tm-utimer> already
>> scheduled 5880 ms ago (now 113380 ms), delaying execution
>>
>> Also that not affecting call or call quality.
>> Only filling up logs.
>>
>> volga629
>>
>> On Thu, Jun 14, 2018 at 6:48 AM, Bogdan-Andrei Iancu
>> <bogdan at opensips.org> wrote:
>>> Hi Volga,
>>>
>>> How large is the presentity data set in your system ? I'm asking as
>>> the routine that seems to be slow queries the presentity table in
>>> order to get the expired presentities - this is done with or
>>> without clustering. Still, in clustering, all the nodes are doing
>>> this query, putting extra stress on the DB.
>>>
>>> Now, for each expire presentity, OpenSIPS has to send a NOTIFY to
>>> its subscribers - and here, having clustering enable, it is a
>>> difference. IF you have 3 nodes, so 3 shared tags, OpenSIPS will
>>> do 3 queries (one per tag) in order to fetch from active_watchers
>>> the subscribers with the tag looking for the presentity. So, who
>>> large is the subscriber's data set ?
>>>
>>> Do you notice any extra DB load when the cleanup timer kicks in ?
>>>
>>> Regards,
>>>
>>> Bogdan-Andrei Iancu
>>>
>>> OpenSIPS Founder and Developer
>>> http://www.opensips-solutions.com
>>> OpenSIPS Summit 2018
>>> http://www.opensips.org/events/Summit-2018Amsterdam
>>>
>>> On 06/07/2018 03:43 PM, volga629 at networklab.ca wrote:
>>>> Hello Bogdan-Andrei,
>>>> Yes those messages start showing up when cluster enabled.
>>>> Standalone mode all works no issues.
>>>> Right now in opensips cluster we have 2 active one backup. In
>>>> PgSQL 5 nodes 3 active 2 backup. In Mongodb cluster: 2 mongos
>>>> 3 config 2 shred.
>>>> We use postgres BDR cluster and MongoDB Cluster.
>>>>
>>>> Here are configuration
>>>>
>>>> #### Presence
>>>> loadmodule "presence.so"
>>>> loadmodule "presence_mwi.so"
>>>> loadmodule "presence_xml.so"
>>>> loadmodule "presence_dialoginfo.so"
>>>> loadmodule "presence_callinfo.so"
>>>> loadmodule "pua.so"
>>>> loadmodule "pua_dialoginfo.so"
>>>> loadmodule "xcap.so"
>>>> modparam("presence|xcap|pua","db_url","postgres://URI/opensips_prod01")
>>>> modparam("presence","server_address","sip:proxy at PUBLIC IP:5082")
>>>> modparam("presence", "notify_offline_body", 1)
>>>> modparam("presence", "fallback2db", 1)
>>>> modparam("presence", "clean_period", 30)
>>>> modparam("presence", "mix_dialog_presence", 1)
>>>> modparam("presence", "cluster_id", 1)
>>>> modparam("presence", "cluster_sharing_tags", "A=active")
>>>> modparam("presence", "cluster_federation_mode", 1)
>>>> modparam("presence", "cluster_pres_events" ,"presence ,
>>>> dialog;sla")
>>>> modparam("presence_xml", "force_active", 1)
>>>> modparam("presence_xml", "pidf_manipulation", 1)
>>>> modparam("pua_dialoginfo", "presence_server", "sip:proxy at PUBLIC
>>>> IP:5082")
>>>>
>>>>
>>>> volga629
>>>>
>>>>
>>>> On Thu, Jun 7, 2018 at 7:32 AM, Bogdan-Andrei Iancu
>>>> <bogdan at opensips.org> wrote:
>>>>> Hi Slava,
>>>>>
>>>>> What is the presence clustering configuration you have here ?
>>>>> also, what is the DB setup in regards to the cluster ?
>>>>>
>>>>> Also, did you start getting those errors only after enabling the
>>>>> clustering support? have you run an opensips individual
>>>>> presence node to see if you still get them ?
>>>>>
>>>>> Regards,
>>>>>
>>>>> Bogdan-Andrei Iancu
>>>>>
>>>>> OpenSIPS Founder and Developer
>>>>> http://www.opensips-solutions.com
>>>>> OpenSIPS Summit 2018
>>>>> http://www.opensips.org/events/Summit-2018Amsterdam
>>>>>
>>>>> On 06/06/2018 05:20 PM, volga629 at networklab.ca wrote:
>>>>>> Hello Everyone,
>>>>>> I am trying put togher 3 nodes ( 2 active 1 backup) presence
>>>>>> cluster and log filled with messages regard cleanup
>>>>>> timer.
>>>>>> Any help thank you.
>>>>>>
>>>>>> opensips-2.4.1.b044f11ee-16.fc27.x86_64
>>>>>>
>>>>>>
>>>>>> Jun 6 09:15:12 sbc01 /usr/sbin/opensips[4584]:
>>>>>> WARNING:core:timer_ticker: timer task <presence-pclean>
>>>>>> already scheduled for 5201330 ms (now 5653910 ms), it
>>>>>> may overlap..
>>>>>> Jun 6 09:15:13 sbc01 /usr/sbin/opensips[4584]:
>>>>>> WARNING:core:timer_ticker: timer task <presence-pclean>
>>>>>> already scheduled for 5201330 ms (now 5654900 ms), it
>>>>>> may overlap..
>>>>>>
>>>>>>
>>>>>> Slava.
>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> Users mailing list
>>>>>> Users at lists.opensips.org
>>>>>> http://lists.opensips.org/cgi-bin/mailman/listinfo/users
>>>>>
>>>>>
>>>>
>>>
>>>
>>
>
>
More information about the Users
mailing list