Re: Mimic prometheus plugin -no socket could be created

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




To have prometheus plugin working you HAVE to tell it to listen to an IPV4 address ...like this 

ceph config set mgr mgr/prometheus/server_addr 0.0.0.0

On Fri, 24 Aug 2018 at 12:44, Jones de Andrade <johannesrs@xxxxxxxxx> wrote:
Hi all.

I'm new to ceph, and after having serious problems in ceph stages 0, 1 and 2 that I could solve myself, now it seems that I have hit a wall harder than my head. :)

When I run salt-run state.orch ceph.stage.deploy, i monitor I see it going up to here:

#######
[14/71]   ceph.sysctl on
          node01....................................... ✓ (0.5s)
          node02........................................ ✓ (0.7s)
          node03....................................... ✓ (0.6s)
          node04......................................... ✓ (0.5s)
          node05....................................... ✓ (0.6s)
          node06.......................................... ✓ (0.5s)

[15/71]   ceph.osd on
          node01...................................... ❌ (0.7s)
          node02........................................ ❌ (0.7s)
          node03....................................... ❌ (0.7s)
          node04......................................... ❌ (0.6s)
          node05....................................... ❌ (0.6s)
          node06.......................................... ❌ (0.7s)

Ended stage: ceph.stage.deploy succeeded=14/71 failed=1/71 time=624.7s

Failures summary:

ceph.osd (/srv/salt/ceph/osd):
  node02:
    deploy OSDs: Module function osd.deploy threw an exception. Exception: Mine on node02 for cephdisks.list
  node03:
    deploy OSDs: Module function osd.deploy threw an exception. Exception: Mine on node03 for cephdisks.list
  node01:
    deploy OSDs: Module function osd.deploy threw an exception. Exception: Mine on node01 for cephdisks.list
  node04:
    deploy OSDs: Module function osd.deploy threw an exception. Exception: Mine on node04 for cephdisks.list
  node05:
    deploy OSDs: Module function osd.deploy threw an exception. Exception: Mine on node05 for cephdisks.list
  node06:
    deploy OSDs: Module function osd.deploy threw an exception. Exception: Mine on node06 for cephdisks.list
#######

Since this is a first attempt in 6 simple test machines, we are going to put the mon, osds, etc, in all nodes at first. Only the master is left in a single machine (node01) by now.

As they are simple machines, they have a single hdd, which is partitioned as follows (the hda4 partition is unmounted and left for the ceph system):

###########
# lsblk
NAME   MAJ:MIN RM   SIZE RO TYPE MOUNTPOINT
sda      8:0    0 465,8G  0 disk
├─sda1   8:1    0   500M  0 part /boot/efi
├─sda2   8:2    0    16G  0 part [SWAP]
├─sda3   8:3    0  49,3G  0 part /
└─sda4   8:4    0   400G  0 part
sr0     11:0    1   3,7G  0 rom

# salt -I 'roles:storage' cephdisks.list
node01:
node02:
node03:
node04:
node05:
node06:

# salt -I 'roles:storage' pillar.get ceph
node02:
    ----------
    storage:
        ----------
        osds:
            ----------
            /dev/sda4:
                ----------
                format:
                    bluestore
                standalone:
                    True
(and so on for all 6 machines)
##########

Finally and just in case, my policy.cfg file reads:

#########
#cluster-unassigned/cluster/*.sls
cluster-ceph/cluster/*.sls
profile-default/cluster/*.sls
profile-default/stack/default/ceph/minions/*yml
config/stack/default/global.yml
config/stack/default/ceph/cluster.yml
role-master/cluster/node01.sls
role-admin/cluster/*.sls
role-mon/cluster/*.sls
role-mgr/cluster/*.sls
role-mds/cluster/*.sls
role-ganesha/cluster/*.sls
role-client-nfs/cluster/*.sls
role-client-cephfs/cluster/*.sls
##########

Please, could someone help me and shed some light on this issue?

Thanks a lot in advance,

Regasrds,

Jones



On Thu, Aug 23, 2018 at 2:46 PM John Spray <jspray@xxxxxxxxxx> wrote:
On Thu, Aug 23, 2018 at 5:18 PM Steven Vacaroaia <stef97@xxxxxxxxx> wrote:
>
> Hi All,
>
> I am trying to enable prometheus plugin with no success due to "no socket could be created"
>
> The instructions for enabling the plugin are very straightforward and simple
>
> Note
> My ultimate goal is to use Prometheus with Cephmetrics
> Some of you suggested to deploy ceph-exporter but why do we need to do that when there is a plugin already ?
>
>
> How can I troubleshoot this further ?
>
> nhandled exception from module 'prometheus' while running on mgr.mon01: error('No socket could be created',)
> Aug 23 12:03:06 mon01 ceph-mgr: 2018-08-23 12:03:06.615 7fadab50e700 -1 prometheus.serve:
> Aug 23 12:03:06 mon01 ceph-mgr: 2018-08-23 12:03:06.615 7fadab50e700 -1 Traceback (most recent call last):
> Aug 23 12:03:06 mon01 ceph-mgr: File "/usr/lib64/ceph/mgr/prometheus/module.py", line 720, in serve
> Aug 23 12:03:06 mon01 ceph-mgr: cherrypy.engine.start()
> Aug 23 12:03:06 mon01 ceph-mgr: File "/usr/lib/python2.7/site-packages/cherrypy/process/wspbus.py", line 250, in start
> Aug 23 12:03:06 mon01 ceph-mgr: raise e_info
> Aug 23 12:03:06 mon01 ceph-mgr: ChannelFailures: error('No socket could be created',)

The things I usually check if a process can't create a socket are:
 - is there anything on the same node already listening on that port?
 - are there security policies (e.g. selinux) that might be preventing it?

John

>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux