Dmitriy Lysenko
2016-08-08 08:26:42 UTC
Good day.
My CephFS switched to read only
This problem was previously on Hammer, but i recreated cephfs, upgraded to Jewel and problem was solved, but appeared after some time.
ceph.log
2016-08-07 18:11:31.226960 mon.0 192.168.13.100:6789/0 148601 : cluster [INF] HEALTH_WARN; mds0: MDS in read-only mode
ceph-mds.log:
2016-08-07 18:10:58.699731 7f9fa2ba6700 1 mds.0.cache.dir(10000000afe) commit error -22 v 1
2016-08-07 18:10:58.699755 7f9fa2ba6700 -1 log_channel(cluster) log [ERR] : failed to commit dir 10000000afe object, errno -22
2016-08-07 18:10:58.699763 7f9fa2ba6700 -1 mds.0.2271 unhandled write error (22) Invalid argument, force readonly...
2016-08-07 18:10:58.699773 7f9fa2ba6700 1 mds.0.cache force file system read-only
2016-08-07 18:10:58.699777 7f9fa2ba6700 0 log_channel(cluster) log [WRN] : force file system read-only
I founded this object:
$ rados --pool metadata ls | grep 10000000afe
10000000afe.00000000
and successfully got it:
$ rados --pool metadata get 10000000afe.00000000 obj
$ echo $?
0
How to switchout MDS from readonly mode?
Are there any tools to test the CephFS system for errors?
$ ceph -v
ceph version 10.2.2 (45107e21c568dd033c2f0a3107dec8f0b0e58374)
$ ceph fs ls
name: cephfs, metadata pool: metadata, data pools: [data ]
$ ceph mds stat
e2283: 1/1/1 up {0=drop-03=up:active}, 3 up:standby
$ ceph osd lspools
0 data,1 metadata,6 one,
$ ceph osd dump | grep 'replicated size'
pool 0 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 256 pgp_num 256 last_change 45647 crash_replay_interval 45 min_read_recency_for_promote 1 min_write_recency_for_promote 1 stripe_width 0
pool 1 'metadata' replicated size 2 min_size 1 crush_ruleset 1 object_hash rjenkins pg_num 256 pgp_num 256 last_change 45649 min_read_recency_for_promote 1 min_write_recency_for_promote 1 stripe_width 0
pool 6 'one' replicated size 3 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 53462 flags hashpspool min_read_recency_for_promote 1 min_write_recency_for_promote 1 stripe_width 0
Thank you for help.
--
Dmitry Lysenko
ISP Sovtest, Kursk, Russia
jabber: ***@jabber.sovtest.ru
My CephFS switched to read only
This problem was previously on Hammer, but i recreated cephfs, upgraded to Jewel and problem was solved, but appeared after some time.
ceph.log
2016-08-07 18:11:31.226960 mon.0 192.168.13.100:6789/0 148601 : cluster [INF] HEALTH_WARN; mds0: MDS in read-only mode
ceph-mds.log:
2016-08-07 18:10:58.699731 7f9fa2ba6700 1 mds.0.cache.dir(10000000afe) commit error -22 v 1
2016-08-07 18:10:58.699755 7f9fa2ba6700 -1 log_channel(cluster) log [ERR] : failed to commit dir 10000000afe object, errno -22
2016-08-07 18:10:58.699763 7f9fa2ba6700 -1 mds.0.2271 unhandled write error (22) Invalid argument, force readonly...
2016-08-07 18:10:58.699773 7f9fa2ba6700 1 mds.0.cache force file system read-only
2016-08-07 18:10:58.699777 7f9fa2ba6700 0 log_channel(cluster) log [WRN] : force file system read-only
I founded this object:
$ rados --pool metadata ls | grep 10000000afe
10000000afe.00000000
and successfully got it:
$ rados --pool metadata get 10000000afe.00000000 obj
$ echo $?
0
How to switchout MDS from readonly mode?
Are there any tools to test the CephFS system for errors?
$ ceph -v
ceph version 10.2.2 (45107e21c568dd033c2f0a3107dec8f0b0e58374)
$ ceph fs ls
name: cephfs, metadata pool: metadata, data pools: [data ]
$ ceph mds stat
e2283: 1/1/1 up {0=drop-03=up:active}, 3 up:standby
$ ceph osd lspools
0 data,1 metadata,6 one,
$ ceph osd dump | grep 'replicated size'
pool 0 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 256 pgp_num 256 last_change 45647 crash_replay_interval 45 min_read_recency_for_promote 1 min_write_recency_for_promote 1 stripe_width 0
pool 1 'metadata' replicated size 2 min_size 1 crush_ruleset 1 object_hash rjenkins pg_num 256 pgp_num 256 last_change 45649 min_read_recency_for_promote 1 min_write_recency_for_promote 1 stripe_width 0
pool 6 'one' replicated size 3 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 512 pgp_num 512 last_change 53462 flags hashpspool min_read_recency_for_promote 1 min_write_recency_for_promote 1 stripe_width 0
Thank you for help.
--
Dmitry Lysenko
ISP Sovtest, Kursk, Russia
jabber: ***@jabber.sovtest.ru