Jewel版cephfs当rm -rf文件时报错"No space left on device"
目录一、概述
原因:cephfs的stray满了
stray就相当于回收站,stray有文件数限制,默认是10万,定时删除,但删除的很慢。导致stray一直处于满的状态。
临时的解决方法:
ceph tell mds.node01 injectargs --mds_bal_fragment_size_max 1000000
然后把mds_bal_fragment_size_max写到配置文件里。
大功告成了。
永久解决方法
https://cloud.tencent.com/developer/article/1177839
(没试过)
二、详情
因为线上环境直接重启肯定不好。所以还是一边监测一边调整参数。
查看当前mds的状态
$ ceph daemonperf mds.node3
-----mds------ --mds_server-- ---objecter--- -----mds_cache----- ---mds_log----
rlat inos caps|hsr hcs hcr |writ read actv|recd recy stry purg|segs evts subm|
0 26k 129 | 0 0 0 | 0 0 0 | 0 0 6.2k 0 | 31 136 0
0 26k 129 | 0 0 0 | 0 0 0 | 0 0 6.2k 0 | 31 136 0
其中node3是自己active的mds的名称。
修改参数
$ ceph daemon mds.node3 config show | grep mds_bal_fragment_size_max
"mds_bal_fragment_size_max": "100000",
#修改mds_bal_fragment_size_max
$ ceph tell mds.node01 injectargs --mds_bal_fragment_size_max 1000000
查询命令还可以使用。"ceph --show-config"
同理修改其他配置:(有一点效果,但不推荐)
ceph tell mds.node03 injectargs --filer_max_purge_ops 10000
ceph tell mds.node03 injectargs --mds_max_purge_files 256
修改完后再次查看mds状态,会发现purge的数据频率会略微提升。
依据
源码文件StrayManager.cc
/*
* Note that there are compromises to how throttling
* is implemented here, in the interests of simplicity:
* * If insufficient ops are available to execute
* the next item on the queue, we block even if
* there are items further down the queue requiring
* fewer ops which might be executable
* * The ops considered "in use" by a purge will be
* an overestimate over the period of execution, as
* we count filer_max_purge_ops and ops for old backtraces
* as in use throughout, even though towards the end
* of the purge the actual ops in flight will be
* lower.
* * The ops limit may be exceeded if the number of ops
* required by a single inode is greater than the
* limit, for example directories with very many
* fragments.
*/
/*......*/
// Work out a limit based on n_pgs / n_mdss, multiplied by the user's
// preference for how many ops per PG
max_purge_ops = uint64_t(((double)pg_count / (double)mds_count) *
g_conf->mds_max_purge_ops_per_pg);
// User may also specify a hard limit, apply this if so.
if (g_conf->mds_max_purge_ops) {
max_purge_ops = MIN(max_purge_ops, g_conf->mds_max_purge_ops);
}
三、参考网址
https://www.cnblogs.com/hlc-123/p/10700725.html
https://www.cnblogs.com/zphj1987/p/13575384.html
http://blog.wjin.org/posts/cephfs-remove-file-error.html
https://cloud.tencent.com/developer/article/1177839
https://docs.ceph.com/en/latest/cephfs/mds-config-ref/#
https://cloud.tencent.com/developer/article/1181883
https://tracker.ceph.com/issues/15920?journal_id=71023