mmm_control problem

Asked by ophion

hi
we are installing the mysql-mmm in other environment. The mysql-masters, mysql-slaver, mysql-mmm-agents and mysql-mmm-monitor is working. if I run
           $> mmm_control show
                      bd01(xxx.xxx.xxx.xxx) master/ONLINE. Roles:
                      bd01_slave(xxx.xxx.xxx.xxx) slave/ONLINE. Roles:
                      bd02(xxx.xxx.xxx.xxx) master/ONLINE. Roles: writer(xxx.xxx.xxx.xxx)
                      bd02_slave(xxx.xxx.xxx.xxx) slave/ONLINE. Roles:

          $> mmm_control checks
                      bdk02_slave ping [last change: 2011/09/08 11:22:06] OK
                      bd02_slave mysql [last change: 2011/09/08 11:22:06] OK
                      bd02_slave rep_threads [last change: 2011/09/08 11:22:06] OK
                      bd02_slave rep_backlog [last change: 2011/09/08 11:22:06] OK: Backlog is null
                      bd01 ping [last change: 2011/09/08 11:22:06] OK
                      bd01 mysql [last change: 2011/09/08 11:22:06] OK
                      bd01 rep_threads [last change: 2011/09/08 11:22:06] OK
                      bd01 rep_backlog [last change: 2011/09/08 11:22:06] OK: Backlog is null
                      bd01_slave ping [last change: 2011/09/08 11:22:06] OK
                      bd01_slave mysql [last change: 2011/09/08 11:22:06] OK
                      bd01_slave rep_threads [last change: 2011/09/08 11:22:06] OK
                      bd01_slave rep_backlog [last change: 2011/09/08 11:22:06] OK: Backlog is null
                      bd02 ping [last change: 2011/09/08 11:22:06] OK
                      bd02 mysql [last change: 2011/09/08 11:22:06] OK
                      bd02 rep_threads [last change: 2011/09/08 11:22:06] OK
                      bd02 rep_backlog [last change: 2011/09/08 11:22:06] OK: Backlog is null

so it seems that all it's working fine. But if I try to connect to the writer the connetion answers:
                      ERROR 2003 (HY000): Can't connect to MySQL server on 'xxx.xxx.xxx.xxx' (113)
thus we don't have a connection. I can't understood why the mmm_control says that all it's OK. That's a very big problem.

Does anyone have an idea of ​​what may be happening?

Question information

Language:
English Edit question
Status:
Solved
For:
mysql-mmm Edit question
Assignee:
No assignee Edit question
Solved by:
ophion
Solved:
Last query:
Last reply:
Revision history for this message
Walter Heck (walterheck) said :
#1

Go to the machine that is supposed to have that specific role and
exceute a 'ip addr show' command. you should see the virtual IP there.
If not, that is your problem :)

Walter

On Thu, Sep 8, 2011 at 19:41, ophion
<email address hidden> wrote:
> New question #170573 on mysql-mmm:
> https://answers.launchpad.net/mysql-mmm/+question/170573
>
> hi
> we are installing the mysql-mmm in other environment. The mysql-masters, mysql-slaver, mysql-mmm-agents and mysql-mmm-monitor is working. if I run
>           $>  mmm_control show
>                      bd01(xxx.xxx.xxx.xxx) master/ONLINE. Roles:
>                      bd01_slave(xxx.xxx.xxx.xxx) slave/ONLINE. Roles:
>                      bd02(xxx.xxx.xxx.xxx) master/ONLINE. Roles: writer(xxx.xxx.xxx.xxx)
>                      bd02_slave(xxx.xxx.xxx.xxx) slave/ONLINE. Roles:
>
>          $>  mmm_control checks
>                      bdk02_slave  ping         [last change: 2011/09/08 11:22:06]  OK
>                      bd02_slave  mysql        [last change: 2011/09/08 11:22:06]  OK
>                      bd02_slave  rep_threads  [last change: 2011/09/08 11:22:06]  OK
>                      bd02_slave  rep_backlog  [last change: 2011/09/08 11:22:06]  OK: Backlog is null
>                      bd01        ping         [last change: 2011/09/08 11:22:06]  OK
>                      bd01        mysql        [last change: 2011/09/08 11:22:06]  OK
>                      bd01        rep_threads  [last change: 2011/09/08 11:22:06]  OK
>                      bd01        rep_backlog  [last change: 2011/09/08 11:22:06]  OK: Backlog is null
>                      bd01_slave  ping         [last change: 2011/09/08 11:22:06]  OK
>                      bd01_slave  mysql        [last change: 2011/09/08 11:22:06]  OK
>                      bd01_slave  rep_threads  [last change: 2011/09/08 11:22:06]  OK
>                      bd01_slave  rep_backlog  [last change: 2011/09/08 11:22:06]  OK: Backlog is null
>                      bd02        ping         [last change: 2011/09/08 11:22:06]  OK
>                      bd02        mysql        [last change: 2011/09/08 11:22:06]  OK
>                      bd02        rep_threads  [last change: 2011/09/08 11:22:06]  OK
>                      bd02        rep_backlog  [last change: 2011/09/08 11:22:06]  OK: Backlog is null
>
> so it seems that all it's working fine. But if I try to connect to the writer the connetion answers:
>                      ERROR 2003 (HY000): Can't connect to MySQL server on 'xxx.xxx.xxx.xxx' (113)
> thus we don't have a connection. I can't understood why the mmm_control says that all it's OK. That's a very big problem.
>
> Does anyone have an idea of what may be happening?
>
>
>
> --
> You received this question notification because you are an answer
> contact for mysql-mmm.
>

--
Walter Heck

--
follow @walterheck on twitter to see what I'm up to!
--
Check out my new startup: Server Monitoring as a Service @ http://tribily.com
Follow @tribily on Twitter and/or 'Like' our Facebook page at
http://www.facebook.com/tribily

Revision history for this message
ophion (danical) said :
#2

Yes thats our problem
          $> ip addr show
                          ......................
                 bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP
                           link/ether 68:b5:99:71:95:72 brd ff:ff:ff:ff:ff:ff
                           inet 192.168.xxx.xxx/24 brd 192.168.xxx.255 scope global bond0
                           inet6 fe80::6ab5:99ff:fe71:9572/64 scope link
                                 valid_lft forever preferred_lft forever
                          ......................

the inet it's the local IP of the host but our writer IP it is no there.
Have you got any idea ? how can I solve the problem?
Thanks

Revision history for this message
Walter Heck (walterheck) said :
#3

your problem is most likely in your config files. Can you post
mmm_common.conf and mmm_mon.conf here? Don't black out all of teh ip
addresses please (just public ones) so we can see if the problem is
with those.

cheers,

Walter

On Fri, Sep 9, 2011 at 10:01, ophion
<email address hidden> wrote:
> Question #170573 on mysql-mmm changed:
> https://answers.launchpad.net/mysql-mmm/+question/170573
>
> ophion posted a new comment:
> Yes thats our problem
>          $>  ip addr show
>                          ......................
>                 bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP
>                           link/ether 68:b5:99:71:95:72 brd ff:ff:ff:ff:ff:ff
>                           inet 192.168.xxx.xxx/24 brd 192.168.xxx.255 scope global bond0
>                           inet6 fe80::6ab5:99ff:fe71:9572/64 scope link
>                                 valid_lft forever preferred_lft forever
>                          ......................
>
> the inet it's the local IP of the host but our writer IP it is no there.
> Have you got any idea ? how can I solve the problem?
> Thanks
>
> --
> You received this question notification because you are an answer
> contact for mysql-mmm.
>

--
Walter Heck

--
follow @walterheck on twitter to see what I'm up to!
--
Check out my new startup: Server Monitoring as a Service @ http://tribily.com
Follow @tribily on Twitter and/or 'Like' our Facebook page at
http://www.facebook.com/tribily

Revision history for this message
ophion (danical) said :
#4

The mmm_common.conf is :

active_master_role writer

<host default>
    cluster_interface bond0
    pid_path /var/run/mysql-mmm/mmm_agentd.pid
    bin_path /usr/libexec/mysql-mmm/
    replication_user rp
    replication_password xxxxxxxxx
    agent_user agent
    agent_password xxxxxxxx
</host>

<host bulk01>
    ip 192.168.72.52
    mode master
    peer bulk02
</host>

<host bulk02>
    ip 192.168.72.57
    mode master
    peer bulk01
</host>

<role writer>
    hosts bulk01, bulk02
    ips 192.168.72.63
    mode exclusive
</role>

and the mmm_mon.conf:

include mmm_common.conf

<monitor>
        ip 127.0.0.1
        pid_path /var/run/mysql-mmm/mmm_mond.pid
        bin_path /usr/libexec/mysql-mmm
        status_path /var/lib/mysql-mmm/mmm_mond.status
        ping_ips 192.168.72.52, 192.168.72.57
        auto_set_online 60
</monitor>

<host default>
        monitor_user monitor
        monitor_password xxxxxxxx
</host>

debug 0

Revision history for this message
Walter Heck (walterheck) said :
#5

Seems fine, except maybe the bonded interface being used as the
cluster interface. Maybe that might be causing problems with the
assignment of the virtual IP.
Another option is to set debug =1 in teh config file of either the
agent or the monitor and see what happens.

Cheers,

Walter

On Fri, Sep 9, 2011 at 12:41, ophion
<email address hidden> wrote:
> Question #170573 on mysql-mmm changed:
> https://answers.launchpad.net/mysql-mmm/+question/170573
>
>    Status: Answered => Open
>
> ophion is still having a problem:
> The mmm_common.conf is :
>
> active_master_role      writer
>
> <host default>
>    cluster_interface       bond0
>    pid_path                /var/run/mysql-mmm/mmm_agentd.pid
>    bin_path                /usr/libexec/mysql-mmm/
>    replication_user        rp
>    replication_password    xxxxxxxxx
>    agent_user              agent
>    agent_password          xxxxxxxx
> </host>
>
>
> <host bulk01>
>    ip      192.168.72.52
>    mode    master
>    peer    bulk02
> </host>
>
> <host bulk02>
>    ip      192.168.72.57
>    mode    master
>    peer    bulk01
> </host>
>
> <role writer>
>    hosts   bulk01, bulk02
>    ips     192.168.72.63
>    mode    exclusive
> </role>
>
> and the mmm_mon.conf:
>
> include mmm_common.conf
>
> <monitor>
>        ip                        127.0.0.1
>        pid_path            /var/run/mysql-mmm/mmm_mond.pid
>        bin_path            /usr/libexec/mysql-mmm
>        status_path         /var/lib/mysql-mmm/mmm_mond.status
>        ping_ips            192.168.72.52, 192.168.72.57
>        auto_set_online     60
> </monitor>
>
> <host default>
>        monitor_user             monitor
>        monitor_password    xxxxxxxx
> </host>
>
> debug 0
>
> --
> You received this question notification because you are an answer
> contact for mysql-mmm.
>

--
Walter Heck

--
follow @walterheck on twitter to see what I'm up to!
--
Check out my new startup: Server Monitoring as a Service @ http://tribily.com
Follow @tribily on Twitter and/or 'Like' our Facebook page at
http://www.facebook.com/tribily

Revision history for this message
ophion (danical) said :
#6

It seems that the problem was in the firewall. I open the RHN CLUSTER ports and this solved the problems.

I haven't found any reference in the documentation indicating that these ports are needed, so I don't know if the problem has been solved by opening the ports or the RHN CLUSTER has run something that has been fixed it. Any idea?

My only concern is why is the command "mmm_control show" said everything was OK.