Planet Ceph

Aggregated news from external sources

  • June 6, 2017
    Ceph部署mon出现0.0.0.0地址

    前言 最近在群里两次看到出现mon地址不对的问题,都是显示0.0.0.0:0地址,如下所示: [root@lab8106 ceph]# ceph -s cluster 3137d009-e41e-41f0-b8f8-5cb574502572 health HEALTH_ERR 1 mons down, quorum 0,1,2 lab8106,node8107,lab104 monmap e2: 4 mons at {lab104=192.168.10.4:6789/0,lab8106=192.168.8.106:6789/0,lab8107=0.0.0.0:0/2,node8107=192.168.8.107:6789/0} 这个之前偶尔会看到有出现这个问题,但是自己一直没碰到过,想看下是什么情况下触发的,在征得这个cepher的同意后,登录上他的环境检查了一下,发现是主机名引起的这个问题 问题复现 在部署的过程中,已经规划好了主机名,而又去修改了这个机器的主机名的情况下就会出现这个问题比如我的这个机器,开始规划好lab8107主机名是这个,然后再lab8107上执行hostname node8107,就会触发这个问题 这个在deploy的部署输出日志中可以看得到 [lab8107][WARNIN] ********************************************************************************[lab8107][WARNIN] provided hostname must match remote hostname[lab8107][WARNIN] provided hostname: lab8107[lab8107][WARNIN] remote hostname: node8107[lab8107][WARNIN] monitors may not reach quorum and create-keys will not complete[lab8107][WARNIN] ******************************************************************************** 可以看到 provided hostname: …Read more

  • June 1, 2017
    Centos7升级内核后无法启动解决办法

    前言 这个问题存在有一段时间了,之前做的centos7的ISO,在进行内核的升级以后就存在这个问题: 系统盘在板载sata口上是可以正常启动新内核并且能识别面板硬盘 系统盘插在面板口上新内核无法启动,调试发现无法找到系统盘 系统盘插在面板上默认的3.10内核可以正常启动 暂时的解决办法就是让系统插在板载的sata口上,因为当时没找到具体的解决办法,在这个问题持续了一段时间后,最近再次搜索资料的时候,把问题定位在了initramfs内的驱动的问题,并且对问题进行了解决 解决过程 查询initramfs的驱动 [root@lab103 lab103]# lsinitrd -k 3.10.0-327.el7.x86_64|grep mpt[23]sasdrwxr-xr-x 2 root root 0 Apr 17 12:05 usr/lib/modules/3.10.0-327.el7.x86_64/kernel/drivers/scsi/mpt2sas-rw-r–r– 1 root root 337793 Nov 20 2015 usr/lib/modules/3.10.0-327.el7.x86_64/kernel/drivers/scsi/mpt2sas/mpt2sas.ko 可以看到在3.10内核的时候是mpt2sas驱动 可以在4.x内核中看到新版的内核已经把mpt2sas升级为mpt3sas /lib/modules/4.4.46/kernel/drivers/scsi/mpt3sas/mpt3sas.ko 查询initramfs内的模块 lsinitrd -k 4.4.46|grep mpt[23]sas 可以看到并没有输出,说明initramfs并没有把这个驱动打进去 这个地方有两种方式来解决 方法一: 修改 /etc/dracut.conf文件,增加字段 add_drivers+=”mpt3sas” 重新生成initramfs dracut -f /boot/initramfs-4.4.46.img 4.4.46 方法二: 强制加载驱动 dracut –force –add-drivers mpt3sas …Read more

  • May 27, 2017
    A tool to rebalance uneven Ceph pools

    The algorithm to fix uneven CRUSH distributions in Ceph was implemented as the crush optimize subcommand. Given the output of ceph report, crush analyze can show buckets that are over/under filled: $ ceph report > ceph_report.json $ crush analyze –crushmap ceph_report.json –pool 3 ~id~ ~weight~ ~PGs~ ~over/under filled %~ ~name~ cloud3-1363 -6 419424 1084 7.90 …Read more

  • May 23, 2017
    Freebsd10.2安装包升级pkg引起环境破坏的解决

    前言 freebsd10.2环境在安装一个新软件包的时候提示升级pkg到1.10.1,然后点击了升级,然后整个pkg环境就无法使用了 记录 升级完了软件包以后第一个错误提示 FreeBSD: /usr/local/lib/libpkg.so.3: Undefined symbol “utimensat” 这个是因为这个库是在freebsd的10.3当中才有的库,而我的环境是10.2的环境 网上有一个解决办法 更新源 # cat /usr/local/etc/pkg/repos/FreeBSD.confFreeBSD: { url: “pkg+http://pkg.FreeBSD.org/${ABI}/release_2”, enabled: yes} 检查当前版本 # pkg –version1.10.1 更新缓存 # pkg update 卸载 # pkg delete -f pkg 重新安装 # pkg install -y pkg# pkg2ng 检查版本 # pkg –version1.5.4 这个在我的环境下没有生效 还有一个办法 有个pkg-static命令可以使用,,然后/var/cache/pkg里边缓存的包。执行命令: # pkg-static install -f /var/cache/pkg/pkg-1.5.4.txz“` 这个在我的环境下报错“`bashroot@mkiso:/usr/ports/ports-mgmt/pkg # …Read more

  • May 12, 2017
    An algorithm to fix uneven CRUSH distributions in Ceph

    The current CRUSH implementation in Ceph does not always provide an even distribution. The most common cause of unevenness is when only a few thousands PGs, or less, are mapped. This is not enough samples and the variations can be as high as 25%. For instance, when there are two OSDs with the same weight, …Read more

  • May 11, 2017
    Ceph space lost due to overweight CRUSH items

    When a CRUSH bucket contains five Ceph OSDs with the following weights: weight osd.0 5 osd.1 1 osd.2 1 osd.3 1 osd.4 1 20% of the space in osd.0 will never be used by a pool with two replicas. The osd.0 gets 55% of the values for the first replica (i.e 5 / 9), as …Read more

  • May 8, 2017
    OpenStack Summit Boston: Deterministic Storage Performance

    Deterministic Storage Performance – ‘The AWS Way’ for capacity based QoS with OpenStack and Ceph Date: 08/05/17 Video: http://sebastien-han.fr/viewer/web/viewer.html?val=http://www.sebastien-han.fr/down/Deterministic_Storage_Performance_-_The_AWS_Way_for_capacity_based_QoS_with_OpenStack_and_Ceph_-_OS_SUMMIT17.pdf Source: Sebastian Han (OpenStack Summit Boston: Deterministic Storage Performance)

  • May 6, 2017
    Recovering from a complete node failure

    Recovering an entire OSD node A Ceph Recovery Story Note: This will be a very lengthy and detail account of my experience. If you want to skip it, please just scroll down to the TL:DR section at the bottom I wanted to share with everyone a situation that happened to me over the weekend. This …Read more

  • May 5, 2017
    Sneak preview: Upcoming Ceph Management Features

    Despite the number of disruptive changes that we went through in the past few weeks, e.g. moving our code base from Mercurial to git, relocating our infrastructure to a new data center, refactoring our code base for version 3.0, our developers have been busy working on expanding the Ceph management capabilities in openATTIC. I’d like …Read more

  • May 4, 2017
    Ceph full ratio and uneven CRUSH distributions

    A common CRUSH rule in Ceph is step chooseleaf firstn 0 type host meaning Placement Groups (PGs) will place replicas on different hosts so the cluster can sustain the failure of any host without losing data. The missing replicas are then restored from the surviving replicas (via a process called “backfilling”) and placed on the …Read more

  • May 3, 2017
    Ceph OSD从filestore 转换到 bluestore的方法

    前言 前段时间看到豪迈的公众号上提到了这个离线转换工具,最近看到群里有人问,找了下没什么相关文档,就自己写了一个,供参考 实践步骤 获取代码并安装 git clone https://github.com/ceph/ceph.gitcd cephgit submodule update –init –recursive./make-distrpm -bb ceph.spec 生成rpm安装包后进行安装,这个过程就不讲太多,根据各种文档安装上最新的版本即可,这个代码合进去时间并不久,大概是上个月才合进去的 配置集群 首先配置一个filestore的集群,这个也是很简单的,我的环境配置一个单主机三个OSD的集群 [root@lab8106 ceph]# ceph -s cluster 3daaf51a-eeba-43a6-9f58-c26c5796f928 health HEALTH_WARN mon.lab8106 low disk space monmap e2: 1 mons at {lab8106=192.168.8.106:6789/0} election epoch 4, quorum 0 lab8106 mgr active: lab8106 osdmap e16: 3 osds: 3 up, 3 in pgmap v34: 64 …Read more

  • May 3, 2017
    openATTIC 2.0.20 has been released

    It is our great pleasure to announce the release of openATTIC version 2.0.20. This is a minor bugfix release, which also provides a number of small selected improvements, e.g. in the WebUI (styling, usability), installation and logging (now adds PID and process name to logs). Furthermore, we updated our documentation – especially the installation instructions …Read more

Careers