- OUTPUT TRUNCATED TO THE LAST 204800 bytes
- da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.0 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.792874 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.792885 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.792896 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.792908 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.4 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.792919 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.5 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.792930 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.6 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.792941 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.7 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.792953 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.8 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.792964 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.9 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.792975 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.a -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.792986 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.b -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.792997 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.c -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793009 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.d -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793020 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.e -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793031 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.f -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793042 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.10 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793054 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.11 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793065 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.12 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793077 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.13 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793088 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.14 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793099 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.15 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793110 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.16 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793122 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.17 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793133 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.18 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793144 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.19 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793155 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1a -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793167 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1b -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793178 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1c -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793189 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1d -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793200 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1e -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793212 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1f -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793223 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.20 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793234 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.21 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793246 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.22 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793257 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.23 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793268 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.24 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793280 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.25 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793291 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.26 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793302 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.27 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793313 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.28 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793325 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.29 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793336 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2a -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793348 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2b -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793359 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2c -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793371 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2d -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793382 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2e -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793393 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2f -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793404 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.30 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793416 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.31 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793427 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.32 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793438 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.33 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793450 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.34 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793461 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.35 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793473 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.36 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793484 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.37 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793496 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.38 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793508 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.39 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793519 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3a -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793569 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3b -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793581 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3c -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793593 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3d -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793604 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3e -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793624 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3f -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793638 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.0 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793649 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793663 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793675 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793686 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.4 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793698 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.5 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793709 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.6 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793721 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.7 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793732 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.8 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793744 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.9 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793755 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.a -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793767 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.b -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793778 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.c -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793789 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.d -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793801 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.e -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793812 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.f -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793824 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.10 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793835 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.11 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793847 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.12 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793858 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.13 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793869 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.14 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793881 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.15 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793892 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.16 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793904 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.17 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793915 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.18 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793926 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.19 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793938 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1a -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793949 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1b -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793960 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1c -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793972 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1d -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793983 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1e -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.793994 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1f -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794006 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.20 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794017 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.21 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794028 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.22 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794040 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.23 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794051 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.24 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794062 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.25 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794074 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.26 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794085 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.27 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794096 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.28 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794108 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.29 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794119 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2a -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794130 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2b -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794142 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2c -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794153 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2d -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794165 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2e -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794176 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2f -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794187 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.30 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794198 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.31 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794210 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.32 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794221 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.33 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794232 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.34 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794243 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.35 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794255 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.36 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794266 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.37 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794277 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.38 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794288 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.39 -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794300 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3a -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794311 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3b -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794322 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3c -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794333 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3d -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794345 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3e -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794356 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3f -> no osds in epoch 9, skipping
- 2015-01-19 19:50:53.794362 7f28da472700 10 mon.--public-addr@0(leader).pg v10 send_pg_creates to 192 pgs
- 2015-01-19 19:50:53.794365 7f28da472700 10 mon.--public-addr@0(leader).pg v10 update_logger
- 2015-01-19 19:50:53.794446 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(mdsmap 1..1) refresh
- 2015-01-19 19:50:53.794507 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..9) refresh
- 2015-01-19 19:50:53.794569 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..10) refresh
- 2015-01-19 19:50:53.794575 7f28da472700 10 mon.--public-addr@0(leader).log v10 update_from_paxos
- 2015-01-19 19:50:53.794578 7f28da472700 10 mon.--public-addr@0(leader).log v10 update_from_paxos version 10 summary v 10
- 2015-01-19 19:50:53.794653 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(monmap 1..1) refresh
- 2015-01-19 19:50:53.794719 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) refresh
- 2015-01-19 19:50:53.794725 7f28da472700 10 mon.--public-addr@0(leader).auth v2 update_from_paxos
- 2015-01-19 19:50:53.794729 7f28da472700 10 mon.--public-addr@0(leader).pg v10 map_pg_creates to 192 pgs -- no change
- 2015-01-19 19:50:53.794735 7f28da472700 10 mon.--public-addr@0(leader).pg v10 send_pg_creates to 192 pgs
- 2015-01-19 19:50:53.794741 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..33) commit_proposal
- 2015-01-19 19:50:53.794746 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..33) commit_proposal proposal 0x42a82a0 took 0.004959 to finish
- 2015-01-19 19:50:53.794761 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..10) _active
- 2015-01-19 19:50:53.794765 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..10) remove_legacy_versions
- 2015-01-19 19:50:53.794786 7f28da472700 7 mon.--public-addr@0(leader).paxosservice(pgmap 1..10) _active creating new pending
- 2015-01-19 19:50:53.794803 7f28da472700 10 mon.--public-addr@0(leader).pg v10 create_pending v 11
- 2015-01-19 19:50:53.794808 7f28da472700 10 mon.--public-addr@0(leader).pg v10 check_osd_map already seen 9 >= 9
- 2015-01-19 19:50:53.794811 7f28da472700 10 mon.--public-addr@0(leader).pg v10 update_logger
- 2015-01-19 19:50:53.794845 7f28da472700 0 log [INF] : pgmap v10: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
- 2015-01-19 19:50:53.794884 7f28da472700 1 -- 127.0.0.1:6789/0 --> mon.0 127.0.0.1:6789/0 -- log(1 entries) v1 -- ?+0 0x3ed9f80
- 2015-01-19 19:50:53.794904 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(1 entries) v1 local
- 2015-01-19 19:50:53.794937 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..33) finish_round state 1 proposals left 0
- 2015-01-19 19:50:53.795008 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:53.795033 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:53.795058 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:53.795065 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..10) dispatch log(1 entries) v1 from mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:53.795076 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..33) is_readable now=2015-01-19 19:50:53.795114 lease_expire=0.000000 has v0 lc 33
- 2015-01-19 19:50:53.795132 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 preprocess_query log(1 entries) v1 from mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:53.795143 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 preprocess_log log(1 entries) v1 from mon.0
- 2015-01-19 19:50:53.795150 7f28d9c71700 20 is_capable service=log command= write on cap allow *
- 2015-01-19 19:50:53.795153 7f28d9c71700 20 allow so far , doing grant allow *
- 2015-01-19 19:50:53.795156 7f28d9c71700 20 allow all
- 2015-01-19 19:50:53.795170 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 prepare_update log(1 entries) v1 from mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:53.795179 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 prepare_log log(1 entries) v1 from mon.0
- 2015-01-19 19:50:53.795184 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 logging 2015-01-19 19:50:53.789842 mon.0 127.0.0.1:6789/0 25 : [INF] osdmap e9: 0 osds: 0 up, 0 in
- 2015-01-19 19:50:53.795207 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..10) setting proposal_timer 0x3c583f0 with delay of 0.0950649
- 2015-01-19 19:50:53.795234 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3edca40
- 2015-01-19 19:50:53.795252 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(1 entries) v1 ==== 0+0+0 (0 0 0) 0x3ed9f80 con 0x3d00c60
- 2015-01-19 19:50:53.795268 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:53.795271 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:53.795281 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:53.795285 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..10) dispatch log(1 entries) v1 from mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:53.795293 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..33) is_readable now=2015-01-19 19:50:53.795294 lease_expire=0.000000 has v0 lc 33
- 2015-01-19 19:50:53.795305 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 preprocess_query log(1 entries) v1 from mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:53.795313 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 preprocess_log log(1 entries) v1 from mon.0
- 2015-01-19 19:50:53.795320 7f28d9c71700 20 is_capable service=log command= write on cap allow *
- 2015-01-19 19:50:53.795323 7f28d9c71700 20 allow so far , doing grant allow *
- 2015-01-19 19:50:53.795325 7f28d9c71700 20 allow all
- 2015-01-19 19:50:53.795333 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 prepare_update log(1 entries) v1 from mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:53.795341 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 prepare_log log(1 entries) v1 from mon.0
- 2015-01-19 19:50:53.795346 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 logging 2015-01-19 19:50:53.794850 mon.0 127.0.0.1:6789/0 26 : [INF] pgmap v10: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
- 2015-01-19 19:50:53.795368 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..10) proposal_timer already set
- 2015-01-19 19:50:53.795374 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3ed9f80
- 2015-01-19 19:50:53.798643 7f28e027f700 2 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01b80).reader couldn't read tag, (0) Success
- 2015-01-19 19:50:53.798710 7f28e027f700 2 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01b80).fault (0) Success
- 2015-01-19 19:50:53.798780 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01b80).fault on lossy channel, failing
- 2015-01-19 19:50:53.798797 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01b80).stop
- 2015-01-19 19:50:53.798815 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).unregister_pipe
- 2015-01-19 19:50:53.798829 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).discard_queue
- 2015-01-19 19:50:53.798863 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).writer finishing
- 2015-01-19 19:50:53.798857 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).reader done
- 2015-01-19 19:50:53.798877 7f28d9c71700 10 mon.--public-addr@0(leader) e1 ms_handle_reset 0x3d01b80 127.0.0.1:0/1011105
- 2015-01-19 19:50:53.798900 7f28d846e700 10 -- 127.0.0.1:6789/0 queue_reap 0x3d71400
- 2015-01-19 19:50:53.798915 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).writer done
- 2015-01-19 19:50:53.798910 7f28d9c71700 10 mon.--public-addr@0(leader) e1 reset/close on session client.? 127.0.0.1:0/1011105
- 2015-01-19 19:50:53.798926 7f28d9c71700 10 mon.--public-addr@0(leader) e1 remove_session MonSession: client.? 127.0.0.1:0/1011105 is openallow * client.? 127.0.0.1:0/1011105
- 2015-01-19 19:50:53.798932 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper
- 2015-01-19 19:50:53.798957 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper reaping pipe 0x3d71400 127.0.0.1:0/1011105
- 2015-01-19 19:50:53.798966 7f28dac73700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).discard_queue
- 2015-01-19 19:50:53.798984 7f28dac73700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).unregister_pipe - not registered
- 2015-01-19 19:50:53.798999 7f28dac73700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).join
- 2015-01-19 19:50:53.799030 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper reaped pipe 0x3d71400 127.0.0.1:0/1011105
- 2015-01-19 19:50:53.799044 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper deleted pipe 0x3d71400
- 2015-01-19 19:50:53.799051 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper done
- 2015-01-19 19:50:53.890369 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..10) propose_pending
- 2015-01-19 19:50:53.890402 7f28da472700 10 mon.--public-addr@0(leader).log v10 encode_full log v 10
- 2015-01-19 19:50:53.890517 7f28da472700 10 mon.--public-addr@0(leader).log v10 encode_pending v11
- 2015-01-19 19:50:53.890553 7f28da472700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..33) queue_proposal bl 6214 bytes; ctx = 0x3c58400
- 2015-01-19 19:50:53.890569 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..33) propose_queued 34 6214 bytes
- 2015-01-19 19:50:53.890574 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..33) begin for 34 6214 bytes
- 2015-01-19 19:50:53.891309 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..33) commit 34
- 2015-01-19 19:50:53.891968 7f28da472700 10 mon.--public-addr@0(leader) e1 refresh_from_paxos
- 2015-01-19 19:50:53.892045 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..10) refresh
- 2015-01-19 19:50:53.892096 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(mdsmap 1..1) refresh
- 2015-01-19 19:50:53.892143 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..9) refresh
- 2015-01-19 19:50:53.892190 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) refresh
- 2015-01-19 19:50:53.892197 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos
- 2015-01-19 19:50:53.892201 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos version 11 summary v 10
- 2015-01-19 19:50:53.892220 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos latest full 10
- 2015-01-19 19:50:53.892245 7f28da472700 7 mon.--public-addr@0(leader).log v11 update_from_paxos applying incremental log 11 2015-01-19 19:50:53.789842 mon.0 127.0.0.1:6789/0 25 : [INF] osdmap e9: 0 osds: 0 up, 0 in
- 2015-01-19 19:50:53.892289 7f28da472700 7 mon.--public-addr@0(leader).log v11 update_from_paxos applying incremental log 11 2015-01-19 19:50:53.794850 mon.0 127.0.0.1:6789/0 26 : [INF] pgmap v10: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
- 2015-01-19 19:50:53.789842 mon.0 127.0.0.1:6789/0 25 : [INF] osdmap e9: 0 osds: 0 up, 0 in
- 2015-01-19 19:50:53.794850 mon.0 127.0.0.1:6789/0 26 : [INF] pgmap v10: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
- 2015-01-19 19:50:53.892338 7f28da472700 10 mon.--public-addr@0(leader).log v11 check_subs
- 2015-01-19 19:50:53.892402 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(monmap 1..1) refresh
- 2015-01-19 19:50:53.892453 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) refresh
- 2015-01-19 19:50:53.892460 7f28da472700 10 mon.--public-addr@0(leader).auth v2 update_from_paxos
- 2015-01-19 19:50:53.892465 7f28da472700 10 mon.--public-addr@0(leader).pg v10 map_pg_creates to 192 pgs -- no change
- 2015-01-19 19:50:53.892469 7f28da472700 10 mon.--public-addr@0(leader).pg v10 send_pg_creates to 192 pgs
- 2015-01-19 19:50:53.892474 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..34) commit_proposal
- 2015-01-19 19:50:53.892478 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..34) commit_proposal proposal 0x42afd20 took 0.001912 to finish
- 2015-01-19 19:50:53.892485 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) _active
- 2015-01-19 19:50:53.892488 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) remove_legacy_versions
- 2015-01-19 19:50:53.892502 7f28da472700 7 mon.--public-addr@0(leader).paxosservice(logm 1..11) _active creating new pending
- 2015-01-19 19:50:53.892514 7f28da472700 10 mon.--public-addr@0(leader).log v11 create_pending v 12
- 2015-01-19 19:50:53.892519 7f28da472700 7 mon.--public-addr@0(leader).log v11 _updated_log for mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:53.892532 7f28da472700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:6789/0 -- log(last 25) v1 -- ?+0 0x3e91500 con 0x3d00c60
- 2015-01-19 19:50:53.892541 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(last 25) v1 local
- 2015-01-19 19:50:53.892569 7f28da472700 7 mon.--public-addr@0(leader).log v11 _updated_log for mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:53.892577 7f28da472700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:6789/0 -- log(last 26) v1 -- ?+0 0x3e92680 con 0x3d00c60
- 2015-01-19 19:50:53.892584 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(last 26) v1 local
- 2015-01-19 19:50:53.892611 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..34) finish_round state 1 proposals left 0
- 2015-01-19 19:50:53.892597 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(last 25) v1 ==== 0+0+0 (0 0 0) 0x3e91500 con 0x3d00c60
- 2015-01-19 19:50:53.892674 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:53.892681 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:53.892701 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:53.892727 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e91500
- 2015-01-19 19:50:53.892744 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(last 26) v1 ==== 0+0+0 (0 0 0) 0x3e92680 con 0x3d00c60
- 2015-01-19 19:50:53.892763 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:53.892767 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:53.892777 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:53.892796 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e92680
- 2015-01-19 19:50:54.009019 7f28d9470700 20 accepter.accepter poll got 1
- 2015-01-19 19:50:54.009046 7f28d9470700 10 accepter.pfd.revents=1
- 2015-01-19 19:50:54.009056 7f28d9470700 10 accepter.accepted incoming on sd 21
- 2015-01-19 19:50:54.009106 7f28d9470700 20 accepter.accepter calling poll
- 2015-01-19 19:50:54.009113 7f28e027f700 10 -- 127.0.0.1:6789/0 >> :/0 pipe(0x3d70f00 sd=21 :0 s=0 pgs=0 cs=0 l=0 c=0x3d01ce0).accept
- 2015-01-19 19:50:54.009236 7f28e027f700 1 -- 127.0.0.1:6789/0 >> :/0 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=0 c=0x3d01ce0).accept sd=21 127.0.0.1:56336/0
- 2015-01-19 19:50:54.009374 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=0 c=0x3d01ce0).accept peer addr is 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.009419 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=0 c=0x3d01ce0).accept got peer connect_seq 0 global_seq 1
- 2015-01-19 19:50:54.009441 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d01ce0).accept of host_type 8, policy.lossy=1 policy.server=1 policy.standby=0 policy.resetcheck=0
- 2015-01-19 19:50:54.009455 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d01ce0).accept my proto 15, their proto 15
- 2015-01-19 19:50:54.009466 7f28e027f700 10 mon.--public-addr@0(leader) e1 ms_verify_authorizer 127.0.0.1:0/1011140 client protocol 0
- 2015-01-19 19:50:54.009475 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d01ce0).accept: setting up session_security.
- 2015-01-19 19:50:54.009486 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d01ce0).accept new session
- 2015-01-19 19:50:54.009495 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).accept success, connect_seq = 1, sending READY
- 2015-01-19 19:50:54.009504 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).accept features 52776558133247
- 2015-01-19 19:50:54.009532 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).register_pipe
- 2015-01-19 19:50:54.009572 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).discard_requeued_up_to 0
- 2015-01-19 19:50:54.009587 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).accept starting writer, state open
- 2015-01-19 19:50:54.009634 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).accept done
- 2015-01-19 19:50:54.009662 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.009670 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.009703 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
- 2015-01-19 19:50:54.009810 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got KEEPALIVE2 2015-01-19 19:50:54.009760
- 2015-01-19 19:50:54.009841 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.009846 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.009861 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_keepalive2 15 2015-01-19 19:50:54.009760
- 2015-01-19 19:50:54.009925 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.009944 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
- 2015-01-19 19:50:54.009976 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got MSG
- 2015-01-19 19:50:54.009997 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got envelope type=17 src client.? front=60 data=0 off 0
- 2015-01-19 19:50:54.010014 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 60 bytes from policy throttler 0/104857600
- 2015-01-19 19:50:54.010027 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 60 from dispatch throttler 0/104857600
- 2015-01-19 19:50:54.010046 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got front 60
- 2015-01-19 19:50:54.010059 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).aborted = 0
- 2015-01-19 19:50:54.010068 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got 60 + 0 + 0 byte message
- 2015-01-19 19:50:54.010099 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got message 1 0x3edcc80 auth(proto 0 30 bytes epoch 0) v1
- 2015-01-19 19:50:54.010114 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3edcc80 prio 127
- 2015-01-19 19:50:54.010138 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.010143 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.010160 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_ack 1
- 2015-01-19 19:50:54.010176 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.010186 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
- 2015-01-19 19:50:54.010165 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.? 127.0.0.1:0/1011140 1 ==== auth(proto 0 30 bytes epoch 0) v1 ==== 60+0+0 (900162395 0 0) 0x3edcc80 con 0x3d01ce0
- 2015-01-19 19:50:54.010212 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.010219 7f28d9c71700 10 mon.--public-addr@0(leader) e1 do not have session, making new one
- 2015-01-19 19:50:54.010227 7f28d9c71700 10 mon.--public-addr@0(leader) e1 ms_dispatch new session MonSession: client.? 127.0.0.1:0/1011140 is open for client.? 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.010240 7f28d9c71700 10 mon.--public-addr@0(leader) e1 setting timeout on session
- 2015-01-19 19:50:54.010244 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps
- 2015-01-19 19:50:54.010249 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) dispatch auth(proto 0 30 bytes epoch 0) v1 from client.? 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.010261 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..34) is_readable now=2015-01-19 19:50:54.010263 lease_expire=0.000000 has v0 lc 34
- 2015-01-19 19:50:54.010280 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 preprocess_query auth(proto 0 30 bytes epoch 0) v1 from client.? 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.010291 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 prep_auth() blob_size=30
- 2015-01-19 19:50:54.010324 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 AuthMonitor::assign_global_id m=auth(proto 0 30 bytes epoch 0) v1 mon=0/1 last_allocated=4113 max_global_id=4196
- 2015-01-19 19:50:54.010331 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 next_global_id should be 4114
- 2015-01-19 19:50:54.010361 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011140 -- mon_map v1 -- ?+0 0x3cc1680 con 0x3d01ce0
- 2015-01-19 19:50:54.010375 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_map v1 remote, 127.0.0.1:0/1011140, have pipe.
- 2015-01-19 19:50:54.010418 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011140 -- auth_reply(proto 1 0 (0) Success) v1 -- ?+0 0x3ce9c00 con 0x3d01ce0
- 2015-01-19 19:50:54.010408 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.010469 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 1 features 52776558133247 0x3cc1680 mon_map v1
- 2015-01-19 19:50:54.010513 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 1): sig = 0
- 2015-01-19 19:50:54.010541 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 1 0x3cc1680
- 2015-01-19 19:50:54.010548 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message auth_reply(proto 1 0 (0) Success) v1 remote, 127.0.0.1:0/1011140, have pipe.
- 2015-01-19 19:50:54.010578 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 60 to dispatch throttler 60/104857600
- 2015-01-19 19:50:54.010592 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3edcc80
- 2015-01-19 19:50:54.010630 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.010658 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 2 features 52776558133247 0x3ce9c00 auth_reply(proto 1 0 (0) Success) v1
- 2015-01-19 19:50:54.010701 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 2): sig = 0
- 2015-01-19 19:50:54.010719 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 2 0x3ce9c00
- 2015-01-19 19:50:54.010764 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.010783 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
- 2015-01-19 19:50:54.011437 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
- 2015-01-19 19:50:54.011464 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 1
- 2015-01-19 19:50:54.011476 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.011486 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
- 2015-01-19 19:50:54.011495 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 2
- 2015-01-19 19:50:54.011502 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.011511 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got MSG
- 2015-01-19 19:50:54.011521 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got envelope type=15 src client.? front=23 data=0 off 0
- 2015-01-19 19:50:54.011533 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 23 bytes from policy throttler 0/104857600
- 2015-01-19 19:50:54.011545 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 23 from dispatch throttler 0/104857600
- 2015-01-19 19:50:54.011559 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got front 23
- 2015-01-19 19:50:54.011570 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).aborted = 0
- 2015-01-19 19:50:54.011578 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got 23 + 0 + 0 byte message
- 2015-01-19 19:50:54.011609 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got message 2 0x3e91a40 mon_subscribe({monmap=0+}) v2
- 2015-01-19 19:50:54.011636 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3e91a40 prio 127
- 2015-01-19 19:50:54.011674 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.011689 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got MSG
- 2015-01-19 19:50:54.011679 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.011699 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got envelope type=15 src client.4114 front=42 data=0 off 0
- 2015-01-19 19:50:54.011702 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_ack 2
- 2015-01-19 19:50:54.011709 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 42 bytes from policy throttler 23/104857600
- 2015-01-19 19:50:54.011693 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.? 127.0.0.1:0/1011140 2 ==== mon_subscribe({monmap=0+}) v2 ==== 23+0+0 (1620593354 0 0) 0x3e91a40 con 0x3d01ce0
- 2015-01-19 19:50:54.011718 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 42 from dispatch throttler 23/104857600
- 2015-01-19 19:50:54.011722 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.011726 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.011732 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got front 42
- 2015-01-19 19:50:54.011730 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
- 2015-01-19 19:50:54.011733 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011140 is openallow * for client.? 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.011741 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).aborted = 0
- 2015-01-19 19:50:54.011750 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.011749 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got 42 + 0 + 0 byte message
- 2015-01-19 19:50:54.011754 7f28d9c71700 10 mon.--public-addr@0(leader) e1 handle_subscribe mon_subscribe({monmap=0+}) v2
- 2015-01-19 19:50:54.011768 7f28d9c71700 10 mon.--public-addr@0(leader) e1 check_sub monmap next 0 have 1
- 2015-01-19 19:50:54.011773 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got message 3 0x3e90a80 mon_subscribe({monmap=2+,osdmap=0}) v2
- 2015-01-19 19:50:54.011785 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3e90a80 prio 127
- 2015-01-19 19:50:54.011785 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011140 -- mon_map v1 -- ?+0 0x3cc0b40 con 0x3d01ce0
- 2015-01-19 19:50:54.011799 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.011803 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.011809 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got MSG
- 2015-01-19 19:50:54.011818 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_ack 3
- 2015-01-19 19:50:54.011819 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got envelope type=15 src client.4114 front=42 data=0 off 0
- 2015-01-19 19:50:54.011823 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_map v1 remote, 127.0.0.1:0/1011140, have pipe.
- 2015-01-19 19:50:54.011829 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 42 bytes from policy throttler 65/104857600
- 2015-01-19 19:50:54.011837 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 42 from dispatch throttler 65/104857600
- 2015-01-19 19:50:54.011844 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.? 127.0.0.1:0/1011140 -- mon_subscribe_ack(300s) v1 -- ?+0 0x3e92680
- 2015-01-19 19:50:54.011850 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got front 42
- 2015-01-19 19:50:54.011848 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 3 features 52776558133247 0x3cc0b40 mon_map v1
- 2015-01-19 19:50:54.011871 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 3): sig = 0
- 2015-01-19 19:50:54.011878 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).aborted = 0
- 2015-01-19 19:50:54.011885 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 3 0x3cc0b40
- 2015-01-19 19:50:54.011888 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got 42 + 0 + 0 byte message
- 2015-01-19 19:50:54.011889 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_subscribe_ack(300s) v1 remote, 127.0.0.1:0/1011140, have pipe.
- 2015-01-19 19:50:54.011902 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 23 to dispatch throttler 107/104857600
- 2015-01-19 19:50:54.011912 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e91a40
- 2015-01-19 19:50:54.011908 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got message 4 0x3e91c00 mon_subscribe({monmap=2+,osdmap=0}) v2
- 2015-01-19 19:50:54.011935 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3e91c00 prio 127
- 2015-01-19 19:50:54.011930 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4114 127.0.0.1:0/1011140 3 ==== mon_subscribe({monmap=2+,osdmap=0}) v2 ==== 42+0+0 (2069260600 0 0) 0x3e90a80 con 0x3d01ce0
- 2015-01-19 19:50:54.011946 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.011948 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011140 is openallow * for client.? 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.011956 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.011959 7f28d9c71700 10 mon.--public-addr@0(leader) e1 handle_subscribe mon_subscribe({monmap=2+,osdmap=0}) v2
- 2015-01-19 19:50:54.011955 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.011966 7f28d9c71700 10 mon.--public-addr@0(leader) e1 check_sub monmap next 2 have 1
- 2015-01-19 19:50:54.011974 7f28d9c71700 20 is_capable service=osd command= read on cap allow *
- 2015-01-19 19:50:54.011977 7f28d9c71700 20 allow so far , doing grant allow *
- 2015-01-19 19:50:54.011967 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.011979 7f28d9c71700 20 allow all
- 2015-01-19 19:50:54.011982 7f28d9c71700 10 mon.--public-addr@0(leader).osd e9 check_sub 0x3c45240 next 0 (onetime)
- 2015-01-19 19:50:54.011981 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_ack 4
- 2015-01-19 19:50:54.011996 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 4 features 52776558133247 0x3e92680 mon_subscribe_ack(300s) v1
- 2015-01-19 19:50:54.012015 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 4): sig = 0
- 2015-01-19 19:50:54.012024 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 4 0x3e92680
- 2015-01-19 19:50:54.012053 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.012063 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
- 2015-01-19 19:50:54.012113 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.? 127.0.0.1:0/1011140 -- osd_map(9..9 src has 1..9) v3 -- ?+0 0x3edcc80
- 2015-01-19 19:50:54.012131 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message osd_map(9..9 src has 1..9) v3 remote, 127.0.0.1:0/1011140, have pipe.
- 2015-01-19 19:50:54.012149 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.012152 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.4114 127.0.0.1:0/1011140 -- mon_subscribe_ack(300s) v1 -- ?+0 0x3e91a40
- 2015-01-19 19:50:54.012163 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 5 features 52776558133247 0x3edcc80 osd_map(9..9 src has 1..9) v3
- 2015-01-19 19:50:54.012190 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 5): sig = 0
- 2015-01-19 19:50:54.012203 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 5 0x3edcc80
- 2015-01-19 19:50:54.012208 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_subscribe_ack(300s) v1 remote, 127.0.0.1:0/1011140, have pipe.
- 2015-01-19 19:50:54.012224 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 42 to dispatch throttler 84/104857600
- 2015-01-19 19:50:54.012231 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e90a80
- 2015-01-19 19:50:54.012245 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.012245 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4114 127.0.0.1:0/1011140 4 ==== mon_subscribe({monmap=2+,osdmap=0}) v2 ==== 42+0+0 (2069260600 0 0) 0x3e91c00 con 0x3d01ce0
- 2015-01-19 19:50:54.012259 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.012261 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011140 is openallow * for client.? 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.012259 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 6 features 52776558133247 0x3e91a40 mon_subscribe_ack(300s) v1
- 2015-01-19 19:50:54.012269 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.012272 7f28d9c71700 10 mon.--public-addr@0(leader) e1 handle_subscribe mon_subscribe({monmap=2+,osdmap=0}) v2
- 2015-01-19 19:50:54.012279 7f28d9c71700 10 mon.--public-addr@0(leader) e1 check_sub monmap next 2 have 1
- 2015-01-19 19:50:54.012276 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 6): sig = 0
- 2015-01-19 19:50:54.012285 7f28d9c71700 20 is_capable service=osd command= read on cap allow *
- 2015-01-19 19:50:54.012288 7f28d9c71700 20 allow so far , doing grant allow *
- 2015-01-19 19:50:54.012290 7f28d9c71700 20 allow all
- 2015-01-19 19:50:54.012285 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 6 0x3e91a40
- 2015-01-19 19:50:54.012292 7f28d9c71700 10 mon.--public-addr@0(leader).osd e9 check_sub 0x3c44bc0 next 0 (onetime)
- 2015-01-19 19:50:54.012312 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.012321 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
- 2015-01-19 19:50:54.012368 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.? 127.0.0.1:0/1011140 -- osd_map(9..9 src has 1..9) v3 -- ?+0 0x3cdca40
- 2015-01-19 19:50:54.012385 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message osd_map(9..9 src has 1..9) v3 remote, 127.0.0.1:0/1011140, have pipe.
- 2015-01-19 19:50:54.012408 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.012410 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.4114 127.0.0.1:0/1011140 -- mon_subscribe_ack(300s) v1 -- ?+0 0x3e90a80
- 2015-01-19 19:50:54.012420 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 7 features 52776558133247 0x3cdca40 osd_map(9..9 src has 1..9) v3
- 2015-01-19 19:50:54.012443 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 7): sig = 0
- 2015-01-19 19:50:54.012457 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 7 0x3cdca40
- 2015-01-19 19:50:54.012462 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_subscribe_ack(300s) v1 remote, 127.0.0.1:0/1011140, have pipe.
- 2015-01-19 19:50:54.012480 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 42 to dispatch throttler 42/104857600
- 2015-01-19 19:50:54.012490 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e91c00
- 2015-01-19 19:50:54.012503 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.012518 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 8 features 52776558133247 0x3e90a80 mon_subscribe_ack(300s) v1
- 2015-01-19 19:50:54.012533 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 8): sig = 0
- 2015-01-19 19:50:54.012542 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 8 0x3e90a80
- 2015-01-19 19:50:54.012566 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.012575 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
- 2015-01-19 19:50:54.015276 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
- 2015-01-19 19:50:54.015303 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 3
- 2015-01-19 19:50:54.015312 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.015321 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
- 2015-01-19 19:50:54.015329 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 4
- 2015-01-19 19:50:54.015335 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.015342 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
- 2015-01-19 19:50:54.015364 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 5
- 2015-01-19 19:50:54.015371 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.015379 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
- 2015-01-19 19:50:54.015388 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 6
- 2015-01-19 19:50:54.015396 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.015404 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
- 2015-01-19 19:50:54.015428 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 7
- 2015-01-19 19:50:54.015439 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.015452 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
- 2015-01-19 19:50:54.015464 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 8
- 2015-01-19 19:50:54.015473 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.015485 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got MSG
- 2015-01-19 19:50:54.015498 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got envelope type=50 src client.4114 front=80 data=0 off 0
- 2015-01-19 19:50:54.015514 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 80 bytes from policy throttler 0/104857600
- 2015-01-19 19:50:54.015528 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 80 from dispatch throttler 0/104857600
- 2015-01-19 19:50:54.015547 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got front 80
- 2015-01-19 19:50:54.015561 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).aborted = 0
- 2015-01-19 19:50:54.015572 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got 80 + 0 + 0 byte message
- 2015-01-19 19:50:54.015613 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got message 5 0x3cc1680 mon_command({"prefix": "get_command_descriptions"} v 0) v1
- 2015-01-19 19:50:54.015649 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3cc1680 prio 127
- 2015-01-19 19:50:54.015685 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.015689 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.015698 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4114 127.0.0.1:0/1011140 5 ==== mon_command({"prefix": "get_command_descriptions"} v 0) v1 ==== 80+0+0 (450363176 0 0) 0x3cc1680 con 0x3d01ce0
- 2015-01-19 19:50:54.015709 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_ack 5
- 2015-01-19 19:50:54.015718 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.015720 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011140 is openallow * for client.? 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.015727 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.015723 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.015730 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
- 2015-01-19 19:50:54.020479 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011140 -- mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1 -- ?+30001 0x3cc1860 con 0x3d01ce0
- 2015-01-19 19:50:54.020511 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1 remote, 127.0.0.1:0/1011140, have pipe.
- 2015-01-19 19:50:54.020540 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 80 to dispatch throttler 80/104857600
- 2015-01-19 19:50:54.020552 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3cc1680
- 2015-01-19 19:50:54.020569 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.020626 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 9 features 52776558133247 0x3cc1860 mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1
- 2015-01-19 19:50:54.020658 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 9): sig = 0
- 2015-01-19 19:50:54.020671 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 9 0x3cc1860
- 2015-01-19 19:50:54.020726 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.020741 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
- 2015-01-19 19:50:54.119996 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
- 2015-01-19 19:50:54.120041 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 9
- 2015-01-19 19:50:54.120055 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.120069 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got MSG
- 2015-01-19 19:50:54.120088 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got envelope type=50 src client.4114 front=104 data=0 off 0
- 2015-01-19 19:50:54.120106 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 104 bytes from policy throttler 0/104857600
- 2015-01-19 19:50:54.120120 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 104 from dispatch throttler 0/104857600
- 2015-01-19 19:50:54.120142 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got front 104
- 2015-01-19 19:50:54.120160 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).aborted = 0
- 2015-01-19 19:50:54.120171 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got 104 + 0 + 0 byte message
- 2015-01-19 19:50:54.120214 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got message 6 0x3cc1e00 mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1
- 2015-01-19 19:50:54.120233 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3cc1e00 prio 127
- 2015-01-19 19:50:54.120260 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
- 2015-01-19 19:50:54.120279 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.120288 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4114 127.0.0.1:0/1011140 6 ==== mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 ==== 104+0+0 (4242270301 0 0) 0x3cc1e00 con 0x3d01ce0
- 2015-01-19 19:50:54.120300 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_ack 6
- 2015-01-19 19:50:54.120306 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.120309 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011140 is openallow * for client.? 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.120316 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.120313 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.120320 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
- 2015-01-19 19:50:54.120444 7f28d9c71700 0 mon.--public-addr@0(leader) e1 handle_command mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1
- 2015-01-19 19:50:54.120494 7f28d9c71700 20 is_capable service=osd command=osd erasure-code-profile rm read write on cap allow *
- 2015-01-19 19:50:54.120500 7f28d9c71700 20 allow so far , doing grant allow *
- 2015-01-19 19:50:54.120502 7f28d9c71700 20 allow all
- 2015-01-19 19:50:54.120503 7f28d9c71700 10 mon.--public-addr@0(leader) e1 _allowed_command capable
- 2015-01-19 19:50:54.120506 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..9) dispatch mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 from client.4114 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.120516 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..34) is_readable now=2015-01-19 19:50:54.120517 lease_expire=0.000000 has v0 lc 34
- 2015-01-19 19:50:54.120525 7f28d9c71700 10 mon.--public-addr@0(leader).osd e9 preprocess_query mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 from client.4114 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.120673 7f28d9c71700 7 mon.--public-addr@0(leader).osd e9 prepare_update mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 from client.4114 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.120819 7f28d9c71700 10 mon.--public-addr@0(leader).osd e9 should_propose
- 2015-01-19 19:50:54.120829 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..9) setting proposal_timer 0x3c583e0 with delay of 0.05
- 2015-01-19 19:50:54.120862 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 104 to dispatch throttler 104/104857600
- 2015-01-19 19:50:54.120872 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3cc1e00
- 2015-01-19 19:50:54.170935 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..9) propose_pending
- 2015-01-19 19:50:54.170952 7f28da472700 10 mon.--public-addr@0(leader).osd e9 encode_pending e 10
- 2015-01-19 19:50:54.171009 7f28da472700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..34) queue_proposal bl 287 bytes; ctx = 0x3c583f0
- 2015-01-19 19:50:54.171021 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..34) propose_queued 35 287 bytes
- 2015-01-19 19:50:54.171026 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..34) begin for 35 287 bytes
- 2015-01-19 19:50:54.171406 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..34) commit 35
- 2015-01-19 19:50:54.171710 7f28da472700 10 mon.--public-addr@0(leader) e1 refresh_from_paxos
- 2015-01-19 19:50:54.171836 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..10) refresh
- 2015-01-19 19:50:54.171885 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(mdsmap 1..1) refresh
- 2015-01-19 19:50:54.171928 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) refresh
- 2015-01-19 19:50:54.171934 7f28da472700 15 mon.--public-addr@0(leader).osd e9 update_from_paxos paxos e 10, my e 9
- 2015-01-19 19:50:54.171987 7f28da472700 7 mon.--public-addr@0(leader).osd e9 update_from_paxos applying incremental 10
- 2015-01-19 19:50:54.172094 7f28da472700 1 mon.--public-addr@0(leader).osd e10 e10: 0 osds: 0 up, 0 in
- 2015-01-19 19:50:54.172313 7f28da472700 10 mon.--public-addr@0(leader).pg v10 check_osd_map -- osdmap not readable, waiting
- 2015-01-19 19:50:54.172326 7f28da472700 10 mon.--public-addr@0(leader).osd e10 check_subs
- 2015-01-19 19:50:54.172332 7f28da472700 10 mon.--public-addr@0(leader).osd e10 share_map_with_random_osd no up osds, don't share with anyone
- 2015-01-19 19:50:54.172336 7f28da472700 10 mon.--public-addr@0(leader).osd e10 update_logger
- 2015-01-19 19:50:54.172406 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) refresh
- 2015-01-19 19:50:54.172414 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos
- 2015-01-19 19:50:54.172417 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos version 11 summary v 11
- 2015-01-19 19:50:54.172462 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(monmap 1..1) refresh
- 2015-01-19 19:50:54.172511 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) refresh
- 2015-01-19 19:50:54.172517 7f28da472700 10 mon.--public-addr@0(leader).auth v2 update_from_paxos
- 2015-01-19 19:50:54.172524 7f28da472700 10 mon.--public-addr@0(leader).pg v10 map_pg_creates to 192 pgs osdmap epoch 10
- 2015-01-19 19:50:54.172546 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.0 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172559 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172568 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172577 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172589 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.4 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172597 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.5 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172605 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.6 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172613 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.7 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172628 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.8 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172637 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.9 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172662 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.a -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172676 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.b -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172690 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.c -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172702 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.d -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172715 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.e -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172728 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.f -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172741 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.10 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172754 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.11 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172767 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.12 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172780 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.13 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172793 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.14 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172805 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.15 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172818 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.16 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172831 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.17 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172843 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.18 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172856 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.19 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172870 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1a -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172883 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1b -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172895 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1c -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172907 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1d -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172920 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1e -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172932 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1f -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172945 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.20 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172957 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.21 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172970 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.22 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172982 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.23 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.172994 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.24 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173007 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.25 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173019 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.26 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173032 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.27 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173045 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.28 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173057 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.29 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173069 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2a -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173081 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2b -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173093 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2c -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173105 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2d -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173117 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2e -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173130 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2f -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173143 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.30 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173155 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.31 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173167 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.32 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173179 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.33 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173191 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.34 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173204 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.35 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173216 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.36 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173228 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.37 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173240 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.38 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173252 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.39 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173264 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3a -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173277 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3b -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173289 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3c -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173300 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3d -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173313 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3e -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173325 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3f -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173337 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.0 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173349 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173362 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173374 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173388 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.4 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173400 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.5 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173412 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.6 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173424 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.7 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173439 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.8 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173451 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.9 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173463 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.a -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173475 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.b -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173487 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.c -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173500 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.d -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173512 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.e -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173524 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.f -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173536 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.10 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173554 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.11 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173566 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.12 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173578 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.13 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173590 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.14 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173603 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.15 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173615 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.16 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173637 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.17 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173649 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.18 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173664 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.19 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173676 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1a -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173688 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1b -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173701 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1c -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173713 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1d -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173726 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1e -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173738 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1f -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173751 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.20 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173763 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.21 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173776 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.22 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173788 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.23 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173802 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.24 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173814 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.25 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173827 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.26 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173839 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.27 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173852 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.28 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173865 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.29 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173919 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2a -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173933 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2b -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173946 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2c -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173959 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2d -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173972 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2e -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.173985 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2f -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174000 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.30 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174012 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.31 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174024 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.32 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174037 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.33 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174051 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.34 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174064 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.35 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174077 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.36 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174089 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.37 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174102 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.38 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174115 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.39 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174127 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3a -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174140 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3b -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174153 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3c -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174165 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3d -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174178 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3e -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174190 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3f -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174203 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.0 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174216 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174229 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174242 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174254 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.4 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174267 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.5 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174279 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.6 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174292 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.7 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174306 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.8 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174318 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.9 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174331 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.a -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174343 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.b -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174356 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.c -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174368 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.d -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174380 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.e -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174393 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.f -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174406 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.10 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174423 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.11 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174436 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.12 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174448 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.13 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174461 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.14 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174473 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.15 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174486 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.16 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174498 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.17 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174511 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.18 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174523 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.19 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174535 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1a -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174548 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1b -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174560 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1c -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174572 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1d -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174585 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1e -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174597 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1f -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174609 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.20 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174630 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.21 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174643 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.22 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174656 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.23 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174669 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.24 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174681 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.25 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174694 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.26 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174707 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.27 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174720 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.28 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174733 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.29 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174746 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2a -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174758 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2b -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174771 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2c -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174783 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2d -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174796 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2e -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174809 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2f -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174821 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.30 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174837 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.31 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174849 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.32 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174862 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.33 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174874 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.34 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174886 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.35 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174899 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.36 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174911 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.37 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174923 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.38 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174936 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.39 -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174948 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3a -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174960 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3b -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174973 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3c -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174985 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3d -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.174997 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3e -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.175009 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3f -> no osds in epoch 10, skipping
- 2015-01-19 19:50:54.175016 7f28da472700 10 mon.--public-addr@0(leader).pg v10 send_pg_creates to 192 pgs
- 2015-01-19 19:50:54.175026 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) commit_proposal
- 2015-01-19 19:50:54.175032 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) commit_proposal proposal 0x42af850 took 0.004014 to finish
- 2015-01-19 19:50:54.175043 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) _active
- 2015-01-19 19:50:54.175046 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) remove_legacy_versions
- 2015-01-19 19:50:54.175074 7f28da472700 7 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) _active creating new pending
- 2015-01-19 19:50:54.175102 7f28da472700 10 mon.--public-addr@0(leader).osd e10 create_pending e 11
- 2015-01-19 19:50:54.175191 7f28da472700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011140 -- mon_command_ack([{"prefix": "osd erasure-code-profile rm", "name": "myprofile"}]=0 v10) v1 -- ?+0 0x3cc10e0 con 0x3d01ce0
- 2015-01-19 19:50:54.175215 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message mon_command_ack([{"prefix": "osd erasure-code-profile rm", "name": "myprofile"}]=0 v10) v1 remote, 127.0.0.1:0/1011140, have pipe.
- 2015-01-19 19:50:54.175254 7f28da472700 5 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) is_readable now=2015-01-19 19:50:54.175256 lease_expire=0.000000 has v0 lc 35
- 2015-01-19 19:50:54.175271 7f28da472700 10 mon.--public-addr@0(leader).pg v10 check_osd_map applying osdmap e10 to pg_map
- 2015-01-19 19:50:54.175280 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.175319 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 10 features 52776558133247 0x3cc10e0 mon_command_ack([{"prefix": "osd erasure-code-profile rm", "name": "myprofile"}]=0 v10) v1
- 2015-01-19 19:50:54.175337 7f28da472700 10 mon.--public-addr@0(leader).pg v10 register_new_pgs checking pg pools for osdmap epoch 10, last_pg_scan 9
- 2015-01-19 19:50:54.175343 7f28da472700 10 mon.--public-addr@0(leader).pg v10 no change in pool 0 replicated size 3 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 1 flags hashpspool crash_replay_interval 45 stripe_width 0
- 2015-01-19 19:50:54.175344 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 10): sig = 0
- 2015-01-19 19:50:54.175350 7f28da472700 10 mon.--public-addr@0(leader).pg v10 no change in pool 1 replicated size 3 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 1 flags hashpspool stripe_width 0
- 2015-01-19 19:50:54.175355 7f28da472700 10 mon.--public-addr@0(leader).pg v10 no change in pool 2 replicated size 3 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 1 flags hashpspool stripe_width 0
- 2015-01-19 19:50:54.175355 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 10 0x3cc10e0
- 2015-01-19 19:50:54.175406 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
- 2015-01-19 19:50:54.175416 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
- 2015-01-19 19:50:54.175432 7f28da472700 10 mon.--public-addr@0(leader).pg v10 register_new_pgs registered 0 new pgs, removed 0 uncreated pgs
- 2015-01-19 19:50:54.175438 7f28da472700 10 mon.--public-addr@0(leader).pg v10 check_down_pgs
- 2015-01-19 19:50:54.175450 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..10) propose_pending
- 2015-01-19 19:50:54.175454 7f28da472700 10 mon.--public-addr@0(leader).pg v10 encode_pending v 11
- 2015-01-19 19:50:54.175503 7f28da472700 5 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) queue_proposal bl 422 bytes; ctx = 0x3c58420
- 2015-01-19 19:50:54.175519 7f28da472700 5 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) propose_new_value not active; proposal queued
- 2015-01-19 19:50:54.175528 7f28da472700 10 mon.--public-addr@0(leader).pg v10 map_pg_creates to 192 pgs -- no change
- 2015-01-19 19:50:54.175530 7f28da472700 10 mon.--public-addr@0(leader).pg v10 send_pg_creates to 192 pgs
- 2015-01-19 19:50:54.175534 7f28da472700 10 mon.--public-addr@0(leader).osd e10 update_logger
- 2015-01-19 19:50:54.175542 7f28da472700 0 log [INF] : osdmap e10: 0 osds: 0 up, 0 in
- 2015-01-19 19:50:54.175566 7f28da472700 1 -- 127.0.0.1:6789/0 --> mon.0 127.0.0.1:6789/0 -- log(1 entries) v1 -- ?+0 0x3edca40
- 2015-01-19 19:50:54.175574 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(1 entries) v1 local
- 2015-01-19 19:50:54.175594 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..35) finish_round state 1 proposals left 1
- 2015-01-19 19:50:54.175600 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..35) propose_queued 36 422 bytes
- 2015-01-19 19:50:54.175603 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) begin for 36 422 bytes
- 2015-01-19 19:50:54.175636 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(1 entries) v1 ==== 0+0+0 (0 0 0) 0x3edca40 con 0x3d00c60
- 2015-01-19 19:50:54.175870 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) commit 36
- 2015-01-19 19:50:54.176266 7f28da472700 10 mon.--public-addr@0(leader) e1 refresh_from_paxos
- 2015-01-19 19:50:54.176385 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..11) refresh
- 2015-01-19 19:50:54.176399 7f28da472700 10 mon.--public-addr@0(leader).pg v10 update_from_paxos read_incremental
- 2015-01-19 19:50:54.176457 7f28da472700 10 mon.--public-addr@0(leader).pg v11 read_pgmap_meta
- 2015-01-19 19:50:54.176595 7f28da472700 10 mon.--public-addr@0(leader).pg v11 map_pg_creates to 192 pgs -- no change
- 2015-01-19 19:50:54.176606 7f28da472700 10 mon.--public-addr@0(leader).pg v11 send_pg_creates to 192 pgs
- 2015-01-19 19:50:54.176610 7f28da472700 10 mon.--public-addr@0(leader).pg v11 update_logger
- 2015-01-19 19:50:54.176704 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(mdsmap 1..1) refresh
- 2015-01-19 19:50:54.176775 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) refresh
- 2015-01-19 19:50:54.176847 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) refresh
- 2015-01-19 19:50:54.176858 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos
- 2015-01-19 19:50:54.176862 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos version 11 summary v 11
- 2015-01-19 19:50:54.176932 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(monmap 1..1) refresh
- 2015-01-19 19:50:54.177008 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) refresh
- 2015-01-19 19:50:54.177018 7f28da472700 10 mon.--public-addr@0(leader).auth v2 update_from_paxos
- 2015-01-19 19:50:54.177028 7f28da472700 10 mon.--public-addr@0(leader).pg v11 map_pg_creates to 192 pgs -- no change
- 2015-01-19 19:50:54.177033 7f28da472700 10 mon.--public-addr@0(leader).pg v11 send_pg_creates to 192 pgs
- 2015-01-19 19:50:54.177045 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..36) commit_proposal
- 2015-01-19 19:50:54.177051 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..36) commit_proposal proposal 0x42ead10 took 0.001534 to finish
- 2015-01-19 19:50:54.177061 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..11) _active
- 2015-01-19 19:50:54.177065 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..11) remove_legacy_versions
- 2015-01-19 19:50:54.177088 7f28da472700 7 mon.--public-addr@0(leader).paxosservice(pgmap 1..11) _active creating new pending
- 2015-01-19 19:50:54.177105 7f28da472700 10 mon.--public-addr@0(leader).pg v11 create_pending v 12
- 2015-01-19 19:50:54.177110 7f28da472700 10 mon.--public-addr@0(leader).pg v11 check_osd_map already seen 10 >= 10
- 2015-01-19 19:50:54.177113 7f28da472700 10 mon.--public-addr@0(leader).pg v11 update_logger
- 2015-01-19 19:50:54.177148 7f28da472700 0 log [INF] : pgmap v11: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
- 2015-01-19 19:50:54.177185 7f28da472700 1 -- 127.0.0.1:6789/0 --> mon.0 127.0.0.1:6789/0 -- log(1 entries) v1 -- ?+0 0x3ed9d40
- 2015-01-19 19:50:54.177199 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(1 entries) v1 local
- 2015-01-19 19:50:54.177232 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..36) finish_round state 1 proposals left 0
- 2015-01-19 19:50:54.177302 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.177328 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:54.177355 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.177362 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) dispatch log(1 entries) v1 from mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:54.177375 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..36) is_readable now=2015-01-19 19:50:54.177376 lease_expire=0.000000 has v0 lc 36
- 2015-01-19 19:50:54.177391 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 preprocess_query log(1 entries) v1 from mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:54.177400 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 preprocess_log log(1 entries) v1 from mon.0
- 2015-01-19 19:50:54.177408 7f28d9c71700 20 is_capable service=log command= write on cap allow *
- 2015-01-19 19:50:54.177412 7f28d9c71700 20 allow so far , doing grant allow *
- 2015-01-19 19:50:54.177414 7f28d9c71700 20 allow all
- 2015-01-19 19:50:54.177431 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 prepare_update log(1 entries) v1 from mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:54.177440 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 prepare_log log(1 entries) v1 from mon.0
- 2015-01-19 19:50:54.177446 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 logging 2015-01-19 19:50:54.175544 mon.0 127.0.0.1:6789/0 27 : [INF] osdmap e10: 0 osds: 0 up, 0 in
- 2015-01-19 19:50:54.177471 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) setting proposal_timer 0x3c583d0 with delay of 0.0984108
- 2015-01-19 19:50:54.177500 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3edca40
- 2015-01-19 19:50:54.177532 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(1 entries) v1 ==== 0+0+0 (0 0 0) 0x3ed9d40 con 0x3d00c60
- 2015-01-19 19:50:54.177550 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.177554 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:54.177565 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.177569 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) dispatch log(1 entries) v1 from mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:54.177578 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..36) is_readable now=2015-01-19 19:50:54.177579 lease_expire=0.000000 has v0 lc 36
- 2015-01-19 19:50:54.177590 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 preprocess_query log(1 entries) v1 from mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:54.177598 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 preprocess_log log(1 entries) v1 from mon.0
- 2015-01-19 19:50:54.177605 7f28d9c71700 20 is_capable service=log command= write on cap allow *
- 2015-01-19 19:50:54.177608 7f28d9c71700 20 allow so far , doing grant allow *
- 2015-01-19 19:50:54.177610 7f28d9c71700 20 allow all
- 2015-01-19 19:50:54.177635 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 prepare_update log(1 entries) v1 from mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:54.177646 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 prepare_log log(1 entries) v1 from mon.0
- 2015-01-19 19:50:54.177651 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 logging 2015-01-19 19:50:54.177152 mon.0 127.0.0.1:6789/0 28 : [INF] pgmap v11: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
- 2015-01-19 19:50:54.177675 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) proposal_timer already set
- 2015-01-19 19:50:54.177682 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3ed9d40
- 2015-01-19 19:50:54.184487 7f28e027f700 2 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader couldn't read tag, (0) Success
- 2015-01-19 19:50:54.184553 7f28e027f700 2 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).fault (0) Success
- 2015-01-19 19:50:54.184660 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).fault on lossy channel, failing
- 2015-01-19 19:50:54.184682 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).stop
- 2015-01-19 19:50:54.184705 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).unregister_pipe
- 2015-01-19 19:50:54.184722 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).discard_queue
- 2015-01-19 19:50:54.184754 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).reader done
- 2015-01-19 19:50:54.184763 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).writer finishing
- 2015-01-19 19:50:54.184803 7f28d846e700 10 -- 127.0.0.1:6789/0 queue_reap 0x3d70f00
- 2015-01-19 19:50:54.184787 7f28d9c71700 10 mon.--public-addr@0(leader) e1 ms_handle_reset 0x3d01ce0 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.184818 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).writer done
- 2015-01-19 19:50:54.184820 7f28d9c71700 10 mon.--public-addr@0(leader) e1 reset/close on session client.? 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.184835 7f28d9c71700 10 mon.--public-addr@0(leader) e1 remove_session MonSession: client.? 127.0.0.1:0/1011140 is openallow * client.? 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.184862 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper
- 2015-01-19 19:50:54.184889 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper reaping pipe 0x3d70f00 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.184899 7f28dac73700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).discard_queue
- 2015-01-19 19:50:54.184918 7f28dac73700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).unregister_pipe - not registered
- 2015-01-19 19:50:54.184945 7f28dac73700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).join
- 2015-01-19 19:50:54.184977 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper reaped pipe 0x3d70f00 127.0.0.1:0/1011140
- 2015-01-19 19:50:54.184992 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper deleted pipe 0x3d70f00
- 2015-01-19 19:50:54.185002 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper done
- 2015-01-19 19:50:54.275975 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) propose_pending
- 2015-01-19 19:50:54.276007 7f28da472700 10 mon.--public-addr@0(leader).log v11 encode_full log v 11
- 2015-01-19 19:50:54.276134 7f28da472700 10 mon.--public-addr@0(leader).log v11 encode_pending v12
- 2015-01-19 19:50:54.276172 7f28da472700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..36) queue_proposal bl 6666 bytes; ctx = 0x3c583e0
- 2015-01-19 19:50:54.276189 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..36) propose_queued 37 6666 bytes
- 2015-01-19 19:50:54.276194 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..36) begin for 37 6666 bytes
- 2015-01-19 19:50:54.276945 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..36) commit 37
- 2015-01-19 19:50:54.277667 7f28da472700 10 mon.--public-addr@0(leader) e1 refresh_from_paxos
- 2015-01-19 19:50:54.277768 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..11) refresh
- 2015-01-19 19:50:54.277831 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(mdsmap 1..1) refresh
- 2015-01-19 19:50:54.277883 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) refresh
- 2015-01-19 19:50:54.277934 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..12) refresh
- 2015-01-19 19:50:54.277941 7f28da472700 10 mon.--public-addr@0(leader).log v12 update_from_paxos
- 2015-01-19 19:50:54.277944 7f28da472700 10 mon.--public-addr@0(leader).log v12 update_from_paxos version 12 summary v 11
- 2015-01-19 19:50:54.277965 7f28da472700 10 mon.--public-addr@0(leader).log v12 update_from_paxos latest full 11
- 2015-01-19 19:50:54.277993 7f28da472700 7 mon.--public-addr@0(leader).log v12 update_from_paxos applying incremental log 12 2015-01-19 19:50:54.175544 mon.0 127.0.0.1:6789/0 27 : [INF] osdmap e10: 0 osds: 0 up, 0 in
- 2015-01-19 19:50:54.278043 7f28da472700 7 mon.--public-addr@0(leader).log v12 update_from_paxos applying incremental log 12 2015-01-19 19:50:54.177152 mon.0 127.0.0.1:6789/0 28 : [INF] pgmap v11: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
- 2015-01-19 19:50:54.175544 mon.0 127.0.0.1:6789/0 27 : [INF] osdmap e10: 0 osds: 0 up, 0 in
- 2015-01-19 19:50:54.177152 mon.0 127.0.0.1:6789/0 28 : [INF] pgmap v11: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
- 2015-01-19 19:50:54.278099 7f28da472700 10 mon.--public-addr@0(leader).log v12 check_subs
- 2015-01-19 19:50:54.278171 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(monmap 1..1) refresh
- 2015-01-19 19:50:54.278227 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) refresh
- 2015-01-19 19:50:54.278235 7f28da472700 10 mon.--public-addr@0(leader).auth v2 update_from_paxos
- 2015-01-19 19:50:54.278241 7f28da472700 10 mon.--public-addr@0(leader).pg v11 map_pg_creates to 192 pgs -- no change
- 2015-01-19 19:50:54.278245 7f28da472700 10 mon.--public-addr@0(leader).pg v11 send_pg_creates to 192 pgs
- 2015-01-19 19:50:54.278255 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..37) commit_proposal
- 2015-01-19 19:50:54.278259 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..37) commit_proposal proposal 0x42ea6f0 took 0.002073 to finish
- 2015-01-19 19:50:54.278267 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..12) _active
- 2015-01-19 19:50:54.278270 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..12) remove_legacy_versions
- 2015-01-19 19:50:54.278287 7f28da472700 7 mon.--public-addr@0(leader).paxosservice(logm 1..12) _active creating new pending
- 2015-01-19 19:50:54.278301 7f28da472700 10 mon.--public-addr@0(leader).log v12 create_pending v 13
- 2015-01-19 19:50:54.278305 7f28da472700 7 mon.--public-addr@0(leader).log v12 _updated_log for mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:54.278319 7f28da472700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:6789/0 -- log(last 27) v1 -- ?+0 0x3e91f80 con 0x3d00c60
- 2015-01-19 19:50:54.278329 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(last 27) v1 local
- 2015-01-19 19:50:54.278363 7f28da472700 7 mon.--public-addr@0(leader).log v12 _updated_log for mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:54.278376 7f28da472700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:6789/0 -- log(last 28) v1 -- ?+0 0x3e92140 con 0x3d00c60
- 2015-01-19 19:50:54.278384 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(last 28) v1 local
- 2015-01-19 19:50:54.278399 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..37) finish_round state 1 proposals left 0
- 2015-01-19 19:50:54.278384 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(last 27) v1 ==== 0+0+0 (0 0 0) 0x3e91f80 con 0x3d00c60
- 2015-01-19 19:50:54.278420 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.278427 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:54.278444 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.278474 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e91f80
- 2015-01-19 19:50:54.278499 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(last 28) v1 ==== 0+0+0 (0 0 0) 0x3e92140 con 0x3d00c60
- 2015-01-19 19:50:54.278521 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.278525 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
- 2015-01-19 19:50:54.278539 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.278562 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e92140
- 2015-01-19 19:50:54.390213 7f28d9470700 20 accepter.accepter poll got 1
- 2015-01-19 19:50:54.390240 7f28d9470700 10 accepter.pfd.revents=1
- 2015-01-19 19:50:54.390251 7f28d9470700 10 accepter.accepted incoming on sd 21
- 2015-01-19 19:50:54.390304 7f28d9470700 20 accepter.accepter calling poll
- 2015-01-19 19:50:54.390308 7f28e027f700 10 -- 127.0.0.1:6789/0 >> :/0 pipe(0x3d71180 sd=21 :0 s=0 pgs=0 cs=0 l=0 c=0x3d02680).accept
- 2015-01-19 19:50:54.390431 7f28e027f700 1 -- 127.0.0.1:6789/0 >> :/0 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=0 c=0x3d02680).accept sd=21 127.0.0.1:56337/0
- 2015-01-19 19:50:54.390605 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=0 c=0x3d02680).accept peer addr is 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.390675 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=0 c=0x3d02680).accept got peer connect_seq 0 global_seq 1
- 2015-01-19 19:50:54.390695 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d02680).accept of host_type 8, policy.lossy=1 policy.server=1 policy.standby=0 policy.resetcheck=0
- 2015-01-19 19:50:54.390705 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d02680).accept my proto 15, their proto 15
- 2015-01-19 19:50:54.390713 7f28e027f700 10 mon.--public-addr@0(leader) e1 ms_verify_authorizer 127.0.0.1:0/1011174 client protocol 0
- 2015-01-19 19:50:54.390721 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d02680).accept: setting up session_security.
- 2015-01-19 19:50:54.390729 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d02680).accept new session
- 2015-01-19 19:50:54.390735 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).accept success, connect_seq = 1, sending READY
- 2015-01-19 19:50:54.390741 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).accept features 52776558133247
- 2015-01-19 19:50:54.390769 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).register_pipe
- 2015-01-19 19:50:54.390802 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).discard_requeued_up_to 0
- 2015-01-19 19:50:54.390813 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).accept starting writer, state open
- 2015-01-19 19:50:54.390842 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).accept done
- 2015-01-19 19:50:54.390858 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.390885 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.390924 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.390944 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got KEEPALIVE2 2015-01-19 19:50:54.390916
- 2015-01-19 19:50:54.390965 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.390975 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.390994 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_keepalive2 15 2015-01-19 19:50:54.390916
- 2015-01-19 19:50:54.391060 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.391069 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got MSG
- 2015-01-19 19:50:54.391079 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.391087 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got envelope type=17 src client.? front=60 data=0 off 0
- 2015-01-19 19:50:54.391098 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 60 bytes from policy throttler 0/104857600
- 2015-01-19 19:50:54.391109 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 60 from dispatch throttler 0/104857600
- 2015-01-19 19:50:54.391123 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got front 60
- 2015-01-19 19:50:54.391132 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).aborted = 0
- 2015-01-19 19:50:54.391138 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got 60 + 0 + 0 byte message
- 2015-01-19 19:50:54.391165 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got message 1 0x3cdca40 auth(proto 0 30 bytes epoch 0) v1
- 2015-01-19 19:50:54.391178 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3cdca40 prio 127
- 2015-01-19 19:50:54.391196 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.391203 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.391207 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.? 127.0.0.1:0/1011174 1 ==== auth(proto 0 30 bytes epoch 0) v1 ==== 60+0+0 (900162395 0 0) 0x3cdca40 con 0x3d02680
- 2015-01-19 19:50:54.391231 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.391226 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_ack 1
- 2015-01-19 19:50:54.391235 7f28d9c71700 10 mon.--public-addr@0(leader) e1 do not have session, making new one
- 2015-01-19 19:50:54.391241 7f28d9c71700 10 mon.--public-addr@0(leader) e1 ms_dispatch new session MonSession: client.? 127.0.0.1:0/1011174 is open for client.? 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.391241 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.391252 7f28d9c71700 10 mon.--public-addr@0(leader) e1 setting timeout on session
- 2015-01-19 19:50:54.391255 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps
- 2015-01-19 19:50:54.391253 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.391258 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) dispatch auth(proto 0 30 bytes epoch 0) v1 from client.? 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.391267 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..37) is_readable now=2015-01-19 19:50:54.391268 lease_expire=0.000000 has v0 lc 37
- 2015-01-19 19:50:54.391278 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 preprocess_query auth(proto 0 30 bytes epoch 0) v1 from client.? 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.391284 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 prep_auth() blob_size=30
- 2015-01-19 19:50:54.391302 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 AuthMonitor::assign_global_id m=auth(proto 0 30 bytes epoch 0) v1 mon=0/1 last_allocated=4114 max_global_id=4196
- 2015-01-19 19:50:54.391307 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 next_global_id should be 4115
- 2015-01-19 19:50:54.391324 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011174 -- mon_map v1 -- ?+0 0x3cc1680 con 0x3d02680
- 2015-01-19 19:50:54.391332 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_map v1 remote, 127.0.0.1:0/1011174, have pipe.
- 2015-01-19 19:50:54.391351 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.391359 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011174 -- auth_reply(proto 1 0 (0) Success) v1 -- ?+0 0x3ce9400 con 0x3d02680
- 2015-01-19 19:50:54.391369 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 1 features 52776558133247 0x3cc1680 mon_map v1
- 2015-01-19 19:50:54.391389 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 1): sig = 0
- 2015-01-19 19:50:54.391401 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 1 0x3cc1680
- 2015-01-19 19:50:54.391405 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message auth_reply(proto 1 0 (0) Success) v1 remote, 127.0.0.1:0/1011174, have pipe.
- 2015-01-19 19:50:54.391425 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 60 to dispatch throttler 60/104857600
- 2015-01-19 19:50:54.391431 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3cdca40
- 2015-01-19 19:50:54.391441 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.391460 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 2 features 52776558133247 0x3ce9400 auth_reply(proto 1 0 (0) Success) v1
- 2015-01-19 19:50:54.391496 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 2): sig = 0
- 2015-01-19 19:50:54.391518 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 2 0x3ce9400
- 2015-01-19 19:50:54.391555 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.391573 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.392094 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
- 2015-01-19 19:50:54.392121 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 1
- 2015-01-19 19:50:54.392130 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.392138 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
- 2015-01-19 19:50:54.392145 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 2
- 2015-01-19 19:50:54.392151 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.392158 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got MSG
- 2015-01-19 19:50:54.392165 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got envelope type=15 src client.? front=23 data=0 off 0
- 2015-01-19 19:50:54.392175 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 23 bytes from policy throttler 0/104857600
- 2015-01-19 19:50:54.392183 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 23 from dispatch throttler 0/104857600
- 2015-01-19 19:50:54.392195 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got front 23
- 2015-01-19 19:50:54.392203 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).aborted = 0
- 2015-01-19 19:50:54.392210 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got 23 + 0 + 0 byte message
- 2015-01-19 19:50:54.392236 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got message 2 0x3e90a80 mon_subscribe({monmap=0+}) v2
- 2015-01-19 19:50:54.392249 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3e90a80 prio 127
- 2015-01-19 19:50:54.392267 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.392270 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.392277 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.? 127.0.0.1:0/1011174 2 ==== mon_subscribe({monmap=0+}) v2 ==== 23+0+0 (1620593354 0 0) 0x3e90a80 con 0x3d02680
- 2015-01-19 19:50:54.392289 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_ack 2
- 2015-01-19 19:50:54.392296 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.392299 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011174 is openallow * for client.? 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.392307 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.392305 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.392310 7f28d9c71700 10 mon.--public-addr@0(leader) e1 handle_subscribe mon_subscribe({monmap=0+}) v2
- 2015-01-19 19:50:54.392312 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.392321 7f28d9c71700 10 mon.--public-addr@0(leader) e1 check_sub monmap next 0 have 1
- 2015-01-19 19:50:54.392337 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011174 -- mon_map v1 -- ?+0 0x3cc0f00 con 0x3d02680
- 2015-01-19 19:50:54.392347 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_map v1 remote, 127.0.0.1:0/1011174, have pipe.
- 2015-01-19 19:50:54.392365 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.392366 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.? 127.0.0.1:0/1011174 -- mon_subscribe_ack(300s) v1 -- ?+0 0x3e92140
- 2015-01-19 19:50:54.392378 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 3 features 52776558133247 0x3cc0f00 mon_map v1
- 2015-01-19 19:50:54.392395 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 3): sig = 0
- 2015-01-19 19:50:54.392407 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 3 0x3cc0f00
- 2015-01-19 19:50:54.392414 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got MSG
- 2015-01-19 19:50:54.392413 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_subscribe_ack(300s) v1 remote, 127.0.0.1:0/1011174, have pipe.
- 2015-01-19 19:50:54.392433 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 23 to dispatch throttler 23/104857600
- 2015-01-19 19:50:54.392443 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e90a80
- 2015-01-19 19:50:54.392450 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.392449 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got envelope type=15 src client.4115 front=42 data=0 off 0
- 2015-01-19 19:50:54.392465 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 4 features 52776558133247 0x3e92140 mon_subscribe_ack(300s) v1
- 2015-01-19 19:50:54.392468 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 42 bytes from policy throttler 0/104857600
- 2015-01-19 19:50:54.392480 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 4): sig = 0
- 2015-01-19 19:50:54.392481 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 42 from dispatch throttler 0/104857600
- 2015-01-19 19:50:54.392487 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 4 0x3e92140
- 2015-01-19 19:50:54.392514 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.392519 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got front 42
- 2015-01-19 19:50:54.392527 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.392536 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).aborted = 0
- 2015-01-19 19:50:54.392546 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got 42 + 0 + 0 byte message
- 2015-01-19 19:50:54.392579 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got message 3 0x3e91a40 mon_subscribe({monmap=2+,osdmap=0}) v2
- 2015-01-19 19:50:54.392602 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3e91a40 prio 127
- 2015-01-19 19:50:54.392640 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.392636 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.392651 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_ack 3
- 2015-01-19 19:50:54.392645 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4115 127.0.0.1:0/1011174 3 ==== mon_subscribe({monmap=2+,osdmap=0}) v2 ==== 42+0+0 (2069260600 0 0) 0x3e91a40 con 0x3d02680
- 2015-01-19 19:50:54.392656 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got MSG
- 2015-01-19 19:50:54.392664 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.392661 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.392668 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.392666 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011174 is openallow * for client.? 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.392677 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.392669 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got envelope type=15 src client.4115 front=42 data=0 off 0
- 2015-01-19 19:50:54.392680 7f28d9c71700 10 mon.--public-addr@0(leader) e1 handle_subscribe mon_subscribe({monmap=2+,osdmap=0}) v2
- 2015-01-19 19:50:54.392688 7f28d9c71700 10 mon.--public-addr@0(leader) e1 check_sub monmap next 2 have 1
- 2015-01-19 19:50:54.392685 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 42 bytes from policy throttler 42/104857600
- 2015-01-19 19:50:54.392696 7f28d9c71700 20 is_capable service=osd command= read on cap allow *
- 2015-01-19 19:50:54.392700 7f28d9c71700 20 allow so far , doing grant allow *
- 2015-01-19 19:50:54.392702 7f28d9c71700 20 allow all
- 2015-01-19 19:50:54.392697 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 42 from dispatch throttler 42/104857600
- 2015-01-19 19:50:54.392706 7f28d9c71700 10 mon.--public-addr@0(leader).osd e10 check_sub 0x3c443c0 next 0 (onetime)
- 2015-01-19 19:50:54.392716 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got front 42
- 2015-01-19 19:50:54.392727 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).aborted = 0
- 2015-01-19 19:50:54.392733 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got 42 + 0 + 0 byte message
- 2015-01-19 19:50:54.392754 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got message 4 0x3e92680 mon_subscribe({monmap=2+,osdmap=0}) v2
- 2015-01-19 19:50:54.392765 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3e92680 prio 127
- 2015-01-19 19:50:54.392777 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.392780 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.392789 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_ack 4
- 2015-01-19 19:50:54.392796 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.392803 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.392806 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.? 127.0.0.1:0/1011174 -- osd_map(10..10 src has 1..10) v3 -- ?+0 0x3cdca40
- 2015-01-19 19:50:54.392820 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message osd_map(10..10 src has 1..10) v3 remote, 127.0.0.1:0/1011174, have pipe.
- 2015-01-19 19:50:54.392838 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.392841 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.4115 127.0.0.1:0/1011174 -- mon_subscribe_ack(300s) v1 -- ?+0 0x3e90a80
- 2015-01-19 19:50:54.392848 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 5 features 52776558133247 0x3cdca40 osd_map(10..10 src has 1..10) v3
- 2015-01-19 19:50:54.392873 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 5): sig = 0
- 2015-01-19 19:50:54.392884 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 5 0x3cdca40
- 2015-01-19 19:50:54.392889 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_subscribe_ack(300s) v1 remote, 127.0.0.1:0/1011174, have pipe.
- 2015-01-19 19:50:54.392901 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 42 to dispatch throttler 84/104857600
- 2015-01-19 19:50:54.392906 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e91a40
- 2015-01-19 19:50:54.392918 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4115 127.0.0.1:0/1011174 4 ==== mon_subscribe({monmap=2+,osdmap=0}) v2 ==== 42+0+0 (2069260600 0 0) 0x3e92680 con 0x3d02680
- 2015-01-19 19:50:54.392929 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.392932 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011174 is openallow * for client.? 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.392929 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.392938 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.392941 7f28d9c71700 10 mon.--public-addr@0(leader) e1 handle_subscribe mon_subscribe({monmap=2+,osdmap=0}) v2
- 2015-01-19 19:50:54.392947 7f28d9c71700 10 mon.--public-addr@0(leader) e1 check_sub monmap next 2 have 1
- 2015-01-19 19:50:54.392952 7f28d9c71700 20 is_capable service=osd command= read on cap allow *
- 2015-01-19 19:50:54.392955 7f28d9c71700 20 allow so far , doing grant allow *
- 2015-01-19 19:50:54.392956 7f28d9c71700 20 allow all
- 2015-01-19 19:50:54.392958 7f28d9c71700 10 mon.--public-addr@0(leader).osd e10 check_sub 0x3c44d00 next 0 (onetime)
- 2015-01-19 19:50:54.392942 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 6 features 52776558133247 0x3e90a80 mon_subscribe_ack(300s) v1
- 2015-01-19 19:50:54.392986 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 6): sig = 0
- 2015-01-19 19:50:54.392994 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 6 0x3e90a80
- 2015-01-19 19:50:54.393017 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.393025 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.393022 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.? 127.0.0.1:0/1011174 -- osd_map(10..10 src has 1..10) v3 -- ?+0 0x3cd8900
- 2015-01-19 19:50:54.393034 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message osd_map(10..10 src has 1..10) v3 remote, 127.0.0.1:0/1011174, have pipe.
- 2015-01-19 19:50:54.393052 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.393054 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.4115 127.0.0.1:0/1011174 -- mon_subscribe_ack(300s) v1 -- ?+0 0x3e91a40
- 2015-01-19 19:50:54.393062 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 7 features 52776558133247 0x3cd8900 osd_map(10..10 src has 1..10) v3
- 2015-01-19 19:50:54.393083 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 7): sig = 0
- 2015-01-19 19:50:54.393095 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 7 0x3cd8900
- 2015-01-19 19:50:54.393099 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_subscribe_ack(300s) v1 remote, 127.0.0.1:0/1011174, have pipe.
- 2015-01-19 19:50:54.393112 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 42 to dispatch throttler 42/104857600
- 2015-01-19 19:50:54.393117 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e92680
- 2015-01-19 19:50:54.393130 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.393141 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 8 features 52776558133247 0x3e91a40 mon_subscribe_ack(300s) v1
- 2015-01-19 19:50:54.393154 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 8): sig = 0
- 2015-01-19 19:50:54.393162 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 8 0x3e91a40
- 2015-01-19 19:50:54.393181 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.393189 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.396483 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
- 2015-01-19 19:50:54.396509 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 3
- 2015-01-19 19:50:54.396517 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.396526 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
- 2015-01-19 19:50:54.396534 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 4
- 2015-01-19 19:50:54.396540 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.396547 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
- 2015-01-19 19:50:54.396553 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 5
- 2015-01-19 19:50:54.396559 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.396566 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
- 2015-01-19 19:50:54.396572 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 6
- 2015-01-19 19:50:54.396579 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.396585 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
- 2015-01-19 19:50:54.396592 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 7
- 2015-01-19 19:50:54.396597 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.396604 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
- 2015-01-19 19:50:54.396611 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 8
- 2015-01-19 19:50:54.396626 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.396634 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got MSG
- 2015-01-19 19:50:54.396642 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got envelope type=50 src client.4115 front=80 data=0 off 0
- 2015-01-19 19:50:54.396657 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 80 bytes from policy throttler 0/104857600
- 2015-01-19 19:50:54.396665 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 80 from dispatch throttler 0/104857600
- 2015-01-19 19:50:54.396676 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got front 80
- 2015-01-19 19:50:54.396685 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).aborted = 0
- 2015-01-19 19:50:54.396691 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got 80 + 0 + 0 byte message
- 2015-01-19 19:50:54.396715 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got message 5 0x3cc1680 mon_command({"prefix": "get_command_descriptions"} v 0) v1
- 2015-01-19 19:50:54.396726 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3cc1680 prio 127
- 2015-01-19 19:50:54.396740 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.396745 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.396753 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4115 127.0.0.1:0/1011174 5 ==== mon_command({"prefix": "get_command_descriptions"} v 0) v1 ==== 80+0+0 (450363176 0 0) 0x3cc1680 con 0x3d02680
- 2015-01-19 19:50:54.396772 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.396770 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_ack 5
- 2015-01-19 19:50:54.396775 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011174 is openallow * for client.? 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.396781 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.396784 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.396790 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.399684 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011174 -- mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1 -- ?+30001 0x3cc1c20 con 0x3d02680
- 2015-01-19 19:50:54.399702 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1 remote, 127.0.0.1:0/1011174, have pipe.
- 2015-01-19 19:50:54.399719 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 80 to dispatch throttler 80/104857600
- 2015-01-19 19:50:54.399726 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3cc1680
- 2015-01-19 19:50:54.399743 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.399772 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 9 features 52776558133247 0x3cc1c20 mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1
- 2015-01-19 19:50:54.399800 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 9): sig = 0
- 2015-01-19 19:50:54.399813 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 9 0x3cc1c20
- 2015-01-19 19:50:54.399868 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.399882 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.500019 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
- 2015-01-19 19:50:54.500045 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 9
- 2015-01-19 19:50:54.500053 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.500061 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got MSG
- 2015-01-19 19:50:54.500069 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got envelope type=50 src client.4115 front=104 data=0 off 0
- 2015-01-19 19:50:54.500078 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 104 bytes from policy throttler 0/104857600
- 2015-01-19 19:50:54.500086 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 104 from dispatch throttler 0/104857600
- 2015-01-19 19:50:54.500097 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got front 104
- 2015-01-19 19:50:54.500105 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).aborted = 0
- 2015-01-19 19:50:54.500111 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got 104 + 0 + 0 byte message
- 2015-01-19 19:50:54.500134 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got message 6 0x3cc1860 mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1
- 2015-01-19 19:50:54.500146 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3cc1860 prio 127
- 2015-01-19 19:50:54.500162 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
- 2015-01-19 19:50:54.500165 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.500175 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4115 127.0.0.1:0/1011174 6 ==== mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 ==== 104+0+0 (4242270301 0 0) 0x3cc1860 con 0x3d02680
- 2015-01-19 19:50:54.500189 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_ack 6
- 2015-01-19 19:50:54.500197 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
- 2015-01-19 19:50:54.500200 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011174 is openallow * for client.? 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.500207 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
- 2015-01-19 19:50:54.500203 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.500211 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.500337 7f28d9c71700 0 mon.--public-addr@0(leader) e1 handle_command mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1
- 2015-01-19 19:50:54.500388 7f28d9c71700 20 is_capable service=osd command=osd erasure-code-profile rm read write on cap allow *
- 2015-01-19 19:50:54.500394 7f28d9c71700 20 allow so far , doing grant allow *
- 2015-01-19 19:50:54.500396 7f28d9c71700 20 allow all
- 2015-01-19 19:50:54.500398 7f28d9c71700 10 mon.--public-addr@0(leader) e1 _allowed_command capable
- 2015-01-19 19:50:54.500401 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) dispatch mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 from client.4115 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.500411 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..37) is_readable now=2015-01-19 19:50:54.500411 lease_expire=0.000000 has v0 lc 37
- 2015-01-19 19:50:54.500419 7f28d9c71700 10 mon.--public-addr@0(leader).osd e10 preprocess_query mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 from client.4115 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.500553 7f28d9c71700 7 mon.--public-addr@0(leader).osd e10 prepare_update mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 from client.4115 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.500716 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011174 -- mon_command_ack([{"prefix": "osd erasure-code-profile rm", "name": "myprofile"}]=0 erasure-code-profile myprofile does not exist v10) v1 -- ?+0 0x3cc1680 con 0x3d02680
- 2015-01-19 19:50:54.500733 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_command_ack([{"prefix": "osd erasure-code-profile rm", "name": "myprofile"}]=0 erasure-code-profile myprofile does not exist v10) v1 remote, 127.0.0.1:0/1011174, have pipe.
- 2015-01-19 19:50:54.500749 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.500753 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 104 to dispatch throttler 104/104857600
- 2015-01-19 19:50:54.500760 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3cc1860
- 2015-01-19 19:50:54.500760 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 10 features 52776558133247 0x3cc1680 mon_command_ack([{"prefix": "osd erasure-code-profile rm", "name": "myprofile"}]=0 erasure-code-profile myprofile does not exist v10) v1
- 2015-01-19 19:50:54.500783 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 10): sig = 0
- 2015-01-19 19:50:54.500798 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 10 0x3cc1680
- 2015-01-19 19:50:54.500843 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
- 2015-01-19 19:50:54.500856 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
- 2015-01-19 19:50:54.504876 7f28e027f700 2 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader couldn't read tag, (0) Success
- 2015-01-19 19:50:54.504943 7f28e027f700 2 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).fault (0) Success
- 2015-01-19 19:50:54.505046 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).fault on lossy channel, failing
- 2015-01-19 19:50:54.505065 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).stop
- 2015-01-19 19:50:54.505085 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).unregister_pipe
- 2015-01-19 19:50:54.505102 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).discard_queue
- 2015-01-19 19:50:54.505126 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).reader done
- 2015-01-19 19:50:54.505130 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).writer finishing
- 2015-01-19 19:50:54.505150 7f28d846e700 10 -- 127.0.0.1:6789/0 queue_reap 0x3d71180
- 2015-01-19 19:50:54.505145 7f28d9c71700 10 mon.--public-addr@0(leader) e1 ms_handle_reset 0x3d02680 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.505160 7f28d9c71700 10 mon.--public-addr@0(leader) e1 reset/close on session client.? 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.505160 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).writer done
- 2015-01-19 19:50:54.505166 7f28d9c71700 10 mon.--public-addr@0(leader) e1 remove_session MonSession: client.? 127.0.0.1:0/1011174 is openallow * client.? 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.505171 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper
- 2015-01-19 19:50:54.505194 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper reaping pipe 0x3d71180 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.505208 7f28dac73700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).discard_queue
- 2015-01-19 19:50:54.505239 7f28dac73700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).unregister_pipe - not registered
- 2015-01-19 19:50:54.505254 7f28dac73700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).join
- 2015-01-19 19:50:54.505287 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper reaped pipe 0x3d71180 127.0.0.1:0/1011174
- 2015-01-19 19:50:54.505305 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper deleted pipe 0x3d71180
- 2015-01-19 19:50:54.505312 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper done
- call_TEST_functions: 82: return 1
- run: 27: return 1
- main: 110: code=1
- main: 112: teardown osd-erasure-code-profile
- teardown: 24: local dir=osd-erasure-code-profile
- teardown: 25: kill_daemons osd-erasure-code-profile
- kill_daemons: 60: local dir=osd-erasure-code-profile
- kkill_daemons: 59: find osd-erasure-code-profile
- kkill_daemons: 59: grep pidfile
- kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
- kkill_daemons: 62: cat osd-erasure-code-profile/--public-addr/pidfile
- kill_daemons: 62: pid=10528
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 10528
- kill_daemons: 65: sleep 0
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 10528
- kill_daemons: 65: sleep 1
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 10528
- kill_daemons: 64: break
- teardown: 26: rm -fr osd-erasure-code-profile
- main: 113: return 1
- FAIL: ./src/test/mon/osd-pool-create.sh.log
- main: 105: setup osd-pool-create
- setup: 18: local dir=osd-pool-create
- setup: 19: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- find: 'osd-pool-create': No such file or directory
- teardown: 26: rm -fr osd-pool-create
- setup: 20: mkdir osd-pool-create
- main: 106: local code
- main: 107: run osd-pool-create
- run: 20: local dir=osd-pool-create
- run: 22: export CEPH_ARGS
- rrun: 23: uuidgen
- run: 23: CEPH_ARGS+='--fsid=b22579e5-88d7-496a-9f15-a6b0c601bb8f --auth-supported=none '
- run: 24: CEPH_ARGS+='--mon-host=127.0.0.1 '
- rrun: 26: set
- rrun: 26: sed -n -e 's/^\(TEST_[0-9a-z_]*\) .*/\1/p'
- run: 26: FUNCTIONS='TEST_default_deprectated_0
- TEST_default_deprectated_1
- TEST_default_deprectated_2
- TEST_erasure_code_pool
- TEST_erasure_code_profile_default
- TEST_erasure_code_profile_default_pending
- TEST_erasure_crush_rule
- TEST_erasure_crush_rule_pending
- TEST_erasure_crush_stripe_width
- TEST_erasure_crush_stripe_width_padded
- TEST_erasure_invalid_profile
- TEST_replicated_pool
- TEST_replicated_pool_with_ruleset
- TEST_simple_crush_rule_pending'
- run: 27: for TEST_function in '$FUNCTIONS'
- run: 28: setup osd-pool-create
- setup: 18: local dir=osd-pool-create
- setup: 19: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- teardown: 26: rm -fr osd-pool-create
- setup: 20: mkdir osd-pool-create
- run: 29: TEST_default_deprectated_0 osd-pool-create
- TEST_default_deprectated_0: 35: local dir=osd-pool-create
- TEST_default_deprectated_0: 37: expected=66
- TEST_default_deprectated_0: 38: run_mon osd-pool-create a --public-addr 127.0.0.1 --osd_pool_default_crush_replicated_ruleset 66
- run_mon: 30: local dir=osd-pool-create
- run_mon: 31: shift
- run_mon: 32: local id=a
- run_mon: 33: shift
- run_mon: 34: dir+=/a
- run_mon: 37: ./ceph-mon --id a --mkfs --mon-data=osd-pool-create/a --run-dir=osd-pool-create/a --public-addr 127.0.0.1 --osd_pool_default_crush_replicated_ruleset 66
- ./ceph-mon: renaming mon.noname-a 127.0.0.1:6789/0 to mon.a
- ./ceph-mon: set fsid to b22579e5-88d7-496a-9f15-a6b0c601bb8f
- ./ceph-mon: created monfs at osd-pool-create/a for mon.a
- run_mon: 43: ./ceph-mon --id a --paxos-propose-interval=0.1 --osd-pool-default-erasure-code-directory=.libs --debug-mon 20 --debug-ms 20 --debug-paxos 20 --mon-advanced-debug-mode --chdir= --mon-data=osd-pool-create/a --log-file=osd-pool-create/a/log --mon-cluster-log-file=osd-pool-create/a/log --run-dir=osd-pool-create/a --pid-file=osd-pool-create/a/pidfile --public-addr 127.0.0.1 --osd_pool_default_crush_replicated_ruleset 66
- TEST_default_deprectated_0: 40: ./ceph --format json osd dump
- TEST_default_deprectated_0: 40: grep '"crush_ruleset":66'
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- {"epoch":1,"fsid":"b22579e5-88d7-496a-9f15-a6b0c601bb8f","created":"2015-01-19 19:45:56.846784","modified":"2015-01-19 19:45:56.846784","flags":"","cluster_snapshot":"","pool_max":2,"max_osd":0,"pools":[{"pool":0,"pool_name":"data","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":66,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":45,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":1,"pool_name":"metadata","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":66,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":2,"pool_name":"rbd","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":66,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0}],"osds":[],"osd_xinfo":[],"pg_temp":[],"primary_temp":[],"blacklist":[],"erasure_code_profiles":{"default":{"directory":".libs","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}}}
- TEST_default_deprectated_0: 41: grep 'osd_pool_default_crush_rule is deprecated ' osd-pool-create/a/log
- run: 30: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
- kkill_daemons: 62: cat osd-pool-create/a/pidfile
- kill_daemons: 62: pid=13234
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13234
- kill_daemons: 65: sleep 0
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13234
- kill_daemons: 65: sleep 1
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13234
- kill_daemons: 64: break
- teardown: 26: rm -fr osd-pool-create
- run: 27: for TEST_function in '$FUNCTIONS'
- run: 28: setup osd-pool-create
- setup: 18: local dir=osd-pool-create
- setup: 19: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- find: 'osd-pool-create': No such file or directory
- teardown: 26: rm -fr osd-pool-create
- setup: 20: mkdir osd-pool-create
- run: 29: TEST_default_deprectated_1 osd-pool-create
- TEST_default_deprectated_1: 45: local dir=osd-pool-create
- TEST_default_deprectated_1: 47: expected=55
- TEST_default_deprectated_1: 48: run_mon osd-pool-create a --public-addr 127.0.0.1 --osd_pool_default_crush_rule 55
- run_mon: 30: local dir=osd-pool-create
- run_mon: 31: shift
- run_mon: 32: local id=a
- run_mon: 33: shift
- run_mon: 34: dir+=/a
- run_mon: 37: ./ceph-mon --id a --mkfs --mon-data=osd-pool-create/a --run-dir=osd-pool-create/a --public-addr 127.0.0.1 --osd_pool_default_crush_rule 55
- ./ceph-mon: renaming mon.noname-a 127.0.0.1:6789/0 to mon.a
- ./ceph-mon: set fsid to b22579e5-88d7-496a-9f15-a6b0c601bb8f
- ./ceph-mon: created monfs at osd-pool-create/a for mon.a
- run_mon: 43: ./ceph-mon --id a --paxos-propose-interval=0.1 --osd-pool-default-erasure-code-directory=.libs --debug-mon 20 --debug-ms 20 --debug-paxos 20 --mon-advanced-debug-mode --chdir= --mon-data=osd-pool-create/a --log-file=osd-pool-create/a/log --mon-cluster-log-file=osd-pool-create/a/log --run-dir=osd-pool-create/a --pid-file=osd-pool-create/a/pidfile --public-addr 127.0.0.1 --osd_pool_default_crush_rule 55
- TEST_default_deprectated_1: 50: ./ceph --format json osd dump
- TEST_default_deprectated_1: 50: grep '"crush_ruleset":55'
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- {"epoch":1,"fsid":"b22579e5-88d7-496a-9f15-a6b0c601bb8f","created":"2015-01-19 19:45:58.235083","modified":"2015-01-19 19:45:58.235083","flags":"","cluster_snapshot":"","pool_max":2,"max_osd":0,"pools":[{"pool":0,"pool_name":"data","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":55,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":45,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":1,"pool_name":"metadata","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":55,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":2,"pool_name":"rbd","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":55,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0}],"osds":[],"osd_xinfo":[],"pg_temp":[],"primary_temp":[],"blacklist":[],"erasure_code_profiles":{"default":{"directory":".libs","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}}}
- TEST_default_deprectated_1: 51: grep 'osd_pool_default_crush_rule is deprecated ' osd-pool-create/a/log
- 2015-01-19 19:45:58.234604 7f8575a797c0 0 osd_pool_default_crush_rule is deprecated use osd_pool_default_crush_replicated_ruleset instead
- run: 30: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
- kkill_daemons: 62: cat osd-pool-create/a/pidfile
- kill_daemons: 62: pid=13296
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13296
- kill_daemons: 65: sleep 0
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13296
- kill_daemons: 65: sleep 1
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13296
- kill_daemons: 64: break
- teardown: 26: rm -fr osd-pool-create
- run: 27: for TEST_function in '$FUNCTIONS'
- run: 28: setup osd-pool-create
- setup: 18: local dir=osd-pool-create
- setup: 19: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- find: 'osd-pool-create': No such file or directory
- teardown: 26: rm -fr osd-pool-create
- setup: 20: mkdir osd-pool-create
- run: 29: TEST_default_deprectated_2 osd-pool-create
- TEST_default_deprectated_2: 55: local dir=osd-pool-create
- TEST_default_deprectated_2: 56: expected=77
- TEST_default_deprectated_2: 57: unexpected=33
- TEST_default_deprectated_2: 58: run_mon osd-pool-create a --public-addr 127.0.0.1 --osd_pool_default_crush_rule 77 --osd_pool_default_crush_replicated_ruleset 33
- run_mon: 30: local dir=osd-pool-create
- run_mon: 31: shift
- run_mon: 32: local id=a
- run_mon: 33: shift
- run_mon: 34: dir+=/a
- run_mon: 37: ./ceph-mon --id a --mkfs --mon-data=osd-pool-create/a --run-dir=osd-pool-create/a --public-addr 127.0.0.1 --osd_pool_default_crush_rule 77 --osd_pool_default_crush_replicated_ruleset 33
- ./ceph-mon: renaming mon.noname-a 127.0.0.1:6789/0 to mon.a
- ./ceph-mon: set fsid to b22579e5-88d7-496a-9f15-a6b0c601bb8f
- ./ceph-mon: created monfs at osd-pool-create/a for mon.a
- run_mon: 43: ./ceph-mon --id a --paxos-propose-interval=0.1 --osd-pool-default-erasure-code-directory=.libs --debug-mon 20 --debug-ms 20 --debug-paxos 20 --mon-advanced-debug-mode --chdir= --mon-data=osd-pool-create/a --log-file=osd-pool-create/a/log --mon-cluster-log-file=osd-pool-create/a/log --run-dir=osd-pool-create/a --pid-file=osd-pool-create/a/pidfile --public-addr 127.0.0.1 --osd_pool_default_crush_rule 77 --osd_pool_default_crush_replicated_ruleset 33
- TEST_default_deprectated_2: 61: ./ceph --format json osd dump
- TEST_default_deprectated_2: 61: grep '"crush_ruleset":77'
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- {"epoch":1,"fsid":"b22579e5-88d7-496a-9f15-a6b0c601bb8f","created":"2015-01-19 19:45:59.624973","modified":"2015-01-19 19:45:59.624973","flags":"","cluster_snapshot":"","pool_max":2,"max_osd":0,"pools":[{"pool":0,"pool_name":"data","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":77,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":45,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":1,"pool_name":"metadata","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":77,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":2,"pool_name":"rbd","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":77,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0}],"osds":[],"osd_xinfo":[],"pg_temp":[],"primary_temp":[],"blacklist":[],"erasure_code_profiles":{"default":{"directory":".libs","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}}}
- TEST_default_deprectated_2: 62: ./ceph --format json osd dump
- TEST_default_deprectated_2: 62: grep '"crush_ruleset":33'
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- TEST_default_deprectated_2: 63: grep 'osd_pool_default_crush_rule is deprecated ' osd-pool-create/a/log
- 2015-01-19 19:45:59.624390 7fe4b28d57c0 0 osd_pool_default_crush_rule is deprecated use osd_pool_default_crush_replicated_ruleset instead
- run: 30: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
- kkill_daemons: 62: cat osd-pool-create/a/pidfile
- kill_daemons: 62: pid=13358
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13358
- kill_daemons: 65: sleep 0
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13358
- kill_daemons: 65: sleep 1
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13358
- kill_daemons: 64: break
- teardown: 26: rm -fr osd-pool-create
- run: 27: for TEST_function in '$FUNCTIONS'
- run: 28: setup osd-pool-create
- setup: 18: local dir=osd-pool-create
- setup: 19: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- find: 'osd-pool-create': No such file or directory
- teardown: 26: rm -fr osd-pool-create
- setup: 20: mkdir osd-pool-create
- run: 29: TEST_erasure_code_pool osd-pool-create
- TEST_erasure_code_pool: 179: local dir=osd-pool-create
- TEST_erasure_code_pool: 180: run_mon osd-pool-create a --public-addr 127.0.0.1
- run_mon: 30: local dir=osd-pool-create
- run_mon: 31: shift
- run_mon: 32: local id=a
- run_mon: 33: shift
- run_mon: 34: dir+=/a
- run_mon: 37: ./ceph-mon --id a --mkfs --mon-data=osd-pool-create/a --run-dir=osd-pool-create/a --public-addr 127.0.0.1
- ./ceph-mon: renaming mon.noname-a 127.0.0.1:6789/0 to mon.a
- ./ceph-mon: set fsid to b22579e5-88d7-496a-9f15-a6b0c601bb8f
- ./ceph-mon: created monfs at osd-pool-create/a for mon.a
- run_mon: 43: ./ceph-mon --id a --paxos-propose-interval=0.1 --osd-pool-default-erasure-code-directory=.libs --debug-mon 20 --debug-ms 20 --debug-paxos 20 --mon-advanced-debug-mode --chdir= --mon-data=osd-pool-create/a --log-file=osd-pool-create/a/log --mon-cluster-log-file=osd-pool-create/a/log --run-dir=osd-pool-create/a --pid-file=osd-pool-create/a/pidfile --public-addr 127.0.0.1
- TEST_erasure_code_pool: 181: ./ceph --format json osd dump
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- TEST_erasure_code_pool: 182: local 'expected="erasure_code_profile":"default"'
- TEST_erasure_code_pool: 183: grep '"erasure_code_profile":"default"' osd-pool-create/osd.json
- TEST_erasure_code_pool: 184: ./ceph osd pool create erasurecodes 12 12 erasure
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- pool 'erasurecodes' created
- TEST_erasure_code_pool: 185: ./ceph --format json osd dump
- TEST_erasure_code_pool: 185: tee osd-pool-create/osd.json
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- {"epoch":3,"fsid":"b22579e5-88d7-496a-9f15-a6b0c601bb8f","created":"2015-01-19 19:46:01.351748","modified":"2015-01-19 19:46:02.069213","flags":"","cluster_snapshot":"","pool_max":3,"max_osd":0,"pools":[{"pool":0,"pool_name":"data","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":0,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":45,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":1,"pool_name":"metadata","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":0,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":2,"pool_name":"rbd","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":0,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":3,"pool_name":"erasurecodes","flags":1,"flags_names":"hashpspool","type":3,"size":3,"min_size":2,"crush_ruleset":1,"object_hash":2,"pg_num":12,"pg_placement_num":12,"crash_replay_interval":0,"last_change":"3","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"default","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":4096}],"osds":[],"osd_xinfo":[],"pg_temp":[],"primary_temp":[],"blacklist":[],"erasure_code_profiles":{"default":{"directory":".libs","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}}}
- TEST_erasure_code_pool: 186: grep '"erasure_code_profile":"default"' osd-pool-create/osd.json
- TEST_erasure_code_pool: 188: ./ceph osd pool create erasurecodes 12 12 erasure
- TEST_erasure_code_pool: 189: grep 'already exists'
- pool 'erasurecodes' already exists
- TEST_erasure_code_pool: 190: ./ceph osd pool create erasurecodes 12 12
- TEST_erasure_code_pool: 191: grep 'cannot change to type replicated'
- Error EINVAL: pool 'erasurecodes' cannot change to type replicated
- run: 30: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
- kkill_daemons: 62: cat osd-pool-create/a/pidfile
- kill_daemons: 62: pid=13456
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13456
- kill_daemons: 65: sleep 0
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13456
- kill_daemons: 65: sleep 1
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13456
- kill_daemons: 64: break
- teardown: 26: rm -fr osd-pool-create
- run: 27: for TEST_function in '$FUNCTIONS'
- run: 28: setup osd-pool-create
- setup: 18: local dir=osd-pool-create
- setup: 19: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- find: 'osd-pool-create': No such file or directory
- teardown: 26: rm -fr osd-pool-create
- setup: 20: mkdir osd-pool-create
- run: 29: TEST_erasure_code_profile_default osd-pool-create
- TEST_erasure_code_profile_default: 127: local dir=osd-pool-create
- TEST_erasure_code_profile_default: 128: run_mon osd-pool-create a --public-addr 127.0.0.1
- run_mon: 30: local dir=osd-pool-create
- run_mon: 31: shift
- run_mon: 32: local id=a
- run_mon: 33: shift
- run_mon: 34: dir+=/a
- run_mon: 37: ./ceph-mon --id a --mkfs --mon-data=osd-pool-create/a --run-dir=osd-pool-create/a --public-addr 127.0.0.1
- ./ceph-mon: renaming mon.noname-a 127.0.0.1:6789/0 to mon.a
- ./ceph-mon: set fsid to b22579e5-88d7-496a-9f15-a6b0c601bb8f
- ./ceph-mon: created monfs at osd-pool-create/a for mon.a
- run_mon: 43: ./ceph-mon --id a --paxos-propose-interval=0.1 --osd-pool-default-erasure-code-directory=.libs --debug-mon 20 --debug-ms 20 --debug-paxos 20 --mon-advanced-debug-mode --chdir= --mon-data=osd-pool-create/a --log-file=osd-pool-create/a/log --mon-cluster-log-file=osd-pool-create/a/log --run-dir=osd-pool-create/a --pid-file=osd-pool-create/a/pidfile --public-addr 127.0.0.1
- TEST_erasure_code_profile_default: 129: ./ceph osd erasure-code-profile rm default
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- TEST_erasure_code_profile_default: 130: ./ceph osd erasure-code-profile ls
- TEST_erasure_code_profile_default: 130: grep default
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- TEST_erasure_code_profile_default: 131: ./ceph osd pool create 12 12 erasure default
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- erasure not valid: erasure doesn't represent an int
- pool '12' created
- TEST_erasure_code_profile_default: 132: ./ceph osd erasure-code-profile ls
- TEST_erasure_code_profile_default: 132: grep default
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- default
- run: 30: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
- kkill_daemons: 62: cat osd-pool-create/a/pidfile
- kill_daemons: 62: pid=13661
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13661
- kill_daemons: 65: sleep 0
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13661
- kill_daemons: 65: sleep 1
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13661
- kill_daemons: 64: break
- teardown: 26: rm -fr osd-pool-create
- run: 27: for TEST_function in '$FUNCTIONS'
- run: 28: setup osd-pool-create
- setup: 18: local dir=osd-pool-create
- setup: 19: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- find: 'osd-pool-create': No such file or directory
- teardown: 26: rm -fr osd-pool-create
- setup: 20: mkdir osd-pool-create
- run: 29: TEST_erasure_code_profile_default_pending osd-pool-create
- TEST_erasure_code_profile_default_pending: 136: local dir=osd-pool-create
- TEST_erasure_code_profile_default_pending: 137: run_mon osd-pool-create a --public-addr 127.0.0.1
- run_mon: 30: local dir=osd-pool-create
- run_mon: 31: shift
- run_mon: 32: local id=a
- run_mon: 33: shift
- run_mon: 34: dir+=/a
- run_mon: 37: ./ceph-mon --id a --mkfs --mon-data=osd-pool-create/a --run-dir=osd-pool-create/a --public-addr 127.0.0.1
- ./ceph-mon: renaming mon.noname-a 127.0.0.1:6789/0 to mon.a
- ./ceph-mon: set fsid to b22579e5-88d7-496a-9f15-a6b0c601bb8f
- ./ceph-mon: created monfs at osd-pool-create/a for mon.a
- run_mon: 43: ./ceph-mon --id a --paxos-propose-interval=0.1 --osd-pool-default-erasure-code-directory=.libs --debug-mon 20 --debug-ms 20 --debug-paxos 20 --mon-advanced-debug-mode --chdir= --mon-data=osd-pool-create/a --log-file=osd-pool-create/a/log --mon-cluster-log-file=osd-pool-create/a/log --run-dir=osd-pool-create/a --pid-file=osd-pool-create/a/pidfile --public-addr 127.0.0.1
- TEST_erasure_code_profile_default_pending: 138: ./ceph osd erasure-code-profile rm default
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- TEST_erasure_code_profile_default_pending: 139: ./ceph osd erasure-code-profile ls
- TEST_erasure_code_profile_default_pending: 139: grep default
- *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
- TTEST_erasure_code_profile_default_pending: 141: echo '{"prefix":"osdmonitor_prepare_command","prepare":"osd erasure-code-profile set","name":"default"}'
- TTEST_erasure_code_profile_default_pending: 141: nc -U osd-pool-create/a/ceph-mon.a.asok
- ./test/mon/osd-pool-create.sh: line 141: nc: command not found
- TTEST_erasure_code_profile_default_pending: 141: cut --bytes=5-
- TEST_erasure_code_profile_default_pending: 141: result=
- TEST_erasure_code_profile_default_pending: 142: test = true
- ./test/mon/osd-pool-create.sh: line 142: test: =: unary operator expected
- TEST_erasure_code_profile_default_pending: 142: return 1
- run: 29: return 1
- main: 110: code=1
- main: 112: teardown osd-pool-create
- teardown: 24: local dir=osd-pool-create
- teardown: 25: kill_daemons osd-pool-create
- kill_daemons: 60: local dir=osd-pool-create
- kkill_daemons: 59: find osd-pool-create
- kkill_daemons: 59: grep pidfile
- kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
- kkill_daemons: 62: cat osd-pool-create/a/pidfile
- kill_daemons: 62: pid=13826
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13826
- kill_daemons: 65: sleep 0
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13826
- kill_daemons: 65: sleep 1
- kill_daemons: 63: for try in 0 1 1 1 2 3
- kill_daemons: 64: kill -9 13826
- kill_daemons: 64: break
- teardown: 26: rm -fr osd-pool-create
- main: 113: return 1