1. OUTPUT TRUNCATED TO THE LAST 204800 bytes
  2. da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.0 -> no osds in epoch 9, skipping
  3. 2015-01-19 19:50:53.792874 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1 -> no osds in epoch 9, skipping
  4. 2015-01-19 19:50:53.792885 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2 -> no osds in epoch 9, skipping
  5. 2015-01-19 19:50:53.792896 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3 -> no osds in epoch 9, skipping
  6. 2015-01-19 19:50:53.792908 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.4 -> no osds in epoch 9, skipping
  7. 2015-01-19 19:50:53.792919 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.5 -> no osds in epoch 9, skipping
  8. 2015-01-19 19:50:53.792930 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.6 -> no osds in epoch 9, skipping
  9. 2015-01-19 19:50:53.792941 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.7 -> no osds in epoch 9, skipping
  10. 2015-01-19 19:50:53.792953 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.8 -> no osds in epoch 9, skipping
  11. 2015-01-19 19:50:53.792964 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.9 -> no osds in epoch 9, skipping
  12. 2015-01-19 19:50:53.792975 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.a -> no osds in epoch 9, skipping
  13. 2015-01-19 19:50:53.792986 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.b -> no osds in epoch 9, skipping
  14. 2015-01-19 19:50:53.792997 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.c -> no osds in epoch 9, skipping
  15. 2015-01-19 19:50:53.793009 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.d -> no osds in epoch 9, skipping
  16. 2015-01-19 19:50:53.793020 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.e -> no osds in epoch 9, skipping
  17. 2015-01-19 19:50:53.793031 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.f -> no osds in epoch 9, skipping
  18. 2015-01-19 19:50:53.793042 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.10 -> no osds in epoch 9, skipping
  19. 2015-01-19 19:50:53.793054 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.11 -> no osds in epoch 9, skipping
  20. 2015-01-19 19:50:53.793065 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.12 -> no osds in epoch 9, skipping
  21. 2015-01-19 19:50:53.793077 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.13 -> no osds in epoch 9, skipping
  22. 2015-01-19 19:50:53.793088 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.14 -> no osds in epoch 9, skipping
  23. 2015-01-19 19:50:53.793099 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.15 -> no osds in epoch 9, skipping
  24. 2015-01-19 19:50:53.793110 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.16 -> no osds in epoch 9, skipping
  25. 2015-01-19 19:50:53.793122 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.17 -> no osds in epoch 9, skipping
  26. 2015-01-19 19:50:53.793133 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.18 -> no osds in epoch 9, skipping
  27. 2015-01-19 19:50:53.793144 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.19 -> no osds in epoch 9, skipping
  28. 2015-01-19 19:50:53.793155 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1a -> no osds in epoch 9, skipping
  29. 2015-01-19 19:50:53.793167 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1b -> no osds in epoch 9, skipping
  30. 2015-01-19 19:50:53.793178 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1c -> no osds in epoch 9, skipping
  31. 2015-01-19 19:50:53.793189 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1d -> no osds in epoch 9, skipping
  32. 2015-01-19 19:50:53.793200 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1e -> no osds in epoch 9, skipping
  33. 2015-01-19 19:50:53.793212 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1f -> no osds in epoch 9, skipping
  34. 2015-01-19 19:50:53.793223 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.20 -> no osds in epoch 9, skipping
  35. 2015-01-19 19:50:53.793234 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.21 -> no osds in epoch 9, skipping
  36. 2015-01-19 19:50:53.793246 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.22 -> no osds in epoch 9, skipping
  37. 2015-01-19 19:50:53.793257 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.23 -> no osds in epoch 9, skipping
  38. 2015-01-19 19:50:53.793268 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.24 -> no osds in epoch 9, skipping
  39. 2015-01-19 19:50:53.793280 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.25 -> no osds in epoch 9, skipping
  40. 2015-01-19 19:50:53.793291 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.26 -> no osds in epoch 9, skipping
  41. 2015-01-19 19:50:53.793302 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.27 -> no osds in epoch 9, skipping
  42. 2015-01-19 19:50:53.793313 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.28 -> no osds in epoch 9, skipping
  43. 2015-01-19 19:50:53.793325 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.29 -> no osds in epoch 9, skipping
  44. 2015-01-19 19:50:53.793336 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2a -> no osds in epoch 9, skipping
  45. 2015-01-19 19:50:53.793348 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2b -> no osds in epoch 9, skipping
  46. 2015-01-19 19:50:53.793359 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2c -> no osds in epoch 9, skipping
  47. 2015-01-19 19:50:53.793371 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2d -> no osds in epoch 9, skipping
  48. 2015-01-19 19:50:53.793382 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2e -> no osds in epoch 9, skipping
  49. 2015-01-19 19:50:53.793393 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2f -> no osds in epoch 9, skipping
  50. 2015-01-19 19:50:53.793404 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.30 -> no osds in epoch 9, skipping
  51. 2015-01-19 19:50:53.793416 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.31 -> no osds in epoch 9, skipping
  52. 2015-01-19 19:50:53.793427 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.32 -> no osds in epoch 9, skipping
  53. 2015-01-19 19:50:53.793438 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.33 -> no osds in epoch 9, skipping
  54. 2015-01-19 19:50:53.793450 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.34 -> no osds in epoch 9, skipping
  55. 2015-01-19 19:50:53.793461 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.35 -> no osds in epoch 9, skipping
  56. 2015-01-19 19:50:53.793473 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.36 -> no osds in epoch 9, skipping
  57. 2015-01-19 19:50:53.793484 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.37 -> no osds in epoch 9, skipping
  58. 2015-01-19 19:50:53.793496 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.38 -> no osds in epoch 9, skipping
  59. 2015-01-19 19:50:53.793508 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.39 -> no osds in epoch 9, skipping
  60. 2015-01-19 19:50:53.793519 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3a -> no osds in epoch 9, skipping
  61. 2015-01-19 19:50:53.793569 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3b -> no osds in epoch 9, skipping
  62. 2015-01-19 19:50:53.793581 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3c -> no osds in epoch 9, skipping
  63. 2015-01-19 19:50:53.793593 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3d -> no osds in epoch 9, skipping
  64. 2015-01-19 19:50:53.793604 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3e -> no osds in epoch 9, skipping
  65. 2015-01-19 19:50:53.793624 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3f -> no osds in epoch 9, skipping
  66. 2015-01-19 19:50:53.793638 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.0 -> no osds in epoch 9, skipping
  67. 2015-01-19 19:50:53.793649 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1 -> no osds in epoch 9, skipping
  68. 2015-01-19 19:50:53.793663 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2 -> no osds in epoch 9, skipping
  69. 2015-01-19 19:50:53.793675 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3 -> no osds in epoch 9, skipping
  70. 2015-01-19 19:50:53.793686 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.4 -> no osds in epoch 9, skipping
  71. 2015-01-19 19:50:53.793698 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.5 -> no osds in epoch 9, skipping
  72. 2015-01-19 19:50:53.793709 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.6 -> no osds in epoch 9, skipping
  73. 2015-01-19 19:50:53.793721 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.7 -> no osds in epoch 9, skipping
  74. 2015-01-19 19:50:53.793732 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.8 -> no osds in epoch 9, skipping
  75. 2015-01-19 19:50:53.793744 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.9 -> no osds in epoch 9, skipping
  76. 2015-01-19 19:50:53.793755 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.a -> no osds in epoch 9, skipping
  77. 2015-01-19 19:50:53.793767 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.b -> no osds in epoch 9, skipping
  78. 2015-01-19 19:50:53.793778 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.c -> no osds in epoch 9, skipping
  79. 2015-01-19 19:50:53.793789 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.d -> no osds in epoch 9, skipping
  80. 2015-01-19 19:50:53.793801 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.e -> no osds in epoch 9, skipping
  81. 2015-01-19 19:50:53.793812 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.f -> no osds in epoch 9, skipping
  82. 2015-01-19 19:50:53.793824 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.10 -> no osds in epoch 9, skipping
  83. 2015-01-19 19:50:53.793835 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.11 -> no osds in epoch 9, skipping
  84. 2015-01-19 19:50:53.793847 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.12 -> no osds in epoch 9, skipping
  85. 2015-01-19 19:50:53.793858 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.13 -> no osds in epoch 9, skipping
  86. 2015-01-19 19:50:53.793869 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.14 -> no osds in epoch 9, skipping
  87. 2015-01-19 19:50:53.793881 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.15 -> no osds in epoch 9, skipping
  88. 2015-01-19 19:50:53.793892 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.16 -> no osds in epoch 9, skipping
  89. 2015-01-19 19:50:53.793904 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.17 -> no osds in epoch 9, skipping
  90. 2015-01-19 19:50:53.793915 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.18 -> no osds in epoch 9, skipping
  91. 2015-01-19 19:50:53.793926 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.19 -> no osds in epoch 9, skipping
  92. 2015-01-19 19:50:53.793938 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1a -> no osds in epoch 9, skipping
  93. 2015-01-19 19:50:53.793949 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1b -> no osds in epoch 9, skipping
  94. 2015-01-19 19:50:53.793960 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1c -> no osds in epoch 9, skipping
  95. 2015-01-19 19:50:53.793972 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1d -> no osds in epoch 9, skipping
  96. 2015-01-19 19:50:53.793983 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1e -> no osds in epoch 9, skipping
  97. 2015-01-19 19:50:53.793994 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1f -> no osds in epoch 9, skipping
  98. 2015-01-19 19:50:53.794006 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.20 -> no osds in epoch 9, skipping
  99. 2015-01-19 19:50:53.794017 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.21 -> no osds in epoch 9, skipping
  100. 2015-01-19 19:50:53.794028 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.22 -> no osds in epoch 9, skipping
  101. 2015-01-19 19:50:53.794040 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.23 -> no osds in epoch 9, skipping
  102. 2015-01-19 19:50:53.794051 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.24 -> no osds in epoch 9, skipping
  103. 2015-01-19 19:50:53.794062 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.25 -> no osds in epoch 9, skipping
  104. 2015-01-19 19:50:53.794074 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.26 -> no osds in epoch 9, skipping
  105. 2015-01-19 19:50:53.794085 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.27 -> no osds in epoch 9, skipping
  106. 2015-01-19 19:50:53.794096 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.28 -> no osds in epoch 9, skipping
  107. 2015-01-19 19:50:53.794108 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.29 -> no osds in epoch 9, skipping
  108. 2015-01-19 19:50:53.794119 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2a -> no osds in epoch 9, skipping
  109. 2015-01-19 19:50:53.794130 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2b -> no osds in epoch 9, skipping
  110. 2015-01-19 19:50:53.794142 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2c -> no osds in epoch 9, skipping
  111. 2015-01-19 19:50:53.794153 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2d -> no osds in epoch 9, skipping
  112. 2015-01-19 19:50:53.794165 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2e -> no osds in epoch 9, skipping
  113. 2015-01-19 19:50:53.794176 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2f -> no osds in epoch 9, skipping
  114. 2015-01-19 19:50:53.794187 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.30 -> no osds in epoch 9, skipping
  115. 2015-01-19 19:50:53.794198 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.31 -> no osds in epoch 9, skipping
  116. 2015-01-19 19:50:53.794210 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.32 -> no osds in epoch 9, skipping
  117. 2015-01-19 19:50:53.794221 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.33 -> no osds in epoch 9, skipping
  118. 2015-01-19 19:50:53.794232 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.34 -> no osds in epoch 9, skipping
  119. 2015-01-19 19:50:53.794243 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.35 -> no osds in epoch 9, skipping
  120. 2015-01-19 19:50:53.794255 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.36 -> no osds in epoch 9, skipping
  121. 2015-01-19 19:50:53.794266 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.37 -> no osds in epoch 9, skipping
  122. 2015-01-19 19:50:53.794277 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.38 -> no osds in epoch 9, skipping
  123. 2015-01-19 19:50:53.794288 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.39 -> no osds in epoch 9, skipping
  124. 2015-01-19 19:50:53.794300 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3a -> no osds in epoch 9, skipping
  125. 2015-01-19 19:50:53.794311 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3b -> no osds in epoch 9, skipping
  126. 2015-01-19 19:50:53.794322 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3c -> no osds in epoch 9, skipping
  127. 2015-01-19 19:50:53.794333 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3d -> no osds in epoch 9, skipping
  128. 2015-01-19 19:50:53.794345 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3e -> no osds in epoch 9, skipping
  129. 2015-01-19 19:50:53.794356 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3f -> no osds in epoch 9, skipping
  130. 2015-01-19 19:50:53.794362 7f28da472700 10 mon.--public-addr@0(leader).pg v10 send_pg_creates to 192 pgs
  131. 2015-01-19 19:50:53.794365 7f28da472700 10 mon.--public-addr@0(leader).pg v10 update_logger
  132. 2015-01-19 19:50:53.794446 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(mdsmap 1..1) refresh
  133. 2015-01-19 19:50:53.794507 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..9) refresh
  134. 2015-01-19 19:50:53.794569 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..10) refresh
  135. 2015-01-19 19:50:53.794575 7f28da472700 10 mon.--public-addr@0(leader).log v10 update_from_paxos
  136. 2015-01-19 19:50:53.794578 7f28da472700 10 mon.--public-addr@0(leader).log v10 update_from_paxos version 10 summary v 10
  137. 2015-01-19 19:50:53.794653 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(monmap 1..1) refresh
  138. 2015-01-19 19:50:53.794719 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) refresh
  139. 2015-01-19 19:50:53.794725 7f28da472700 10 mon.--public-addr@0(leader).auth v2 update_from_paxos
  140. 2015-01-19 19:50:53.794729 7f28da472700 10 mon.--public-addr@0(leader).pg v10 map_pg_creates to 192 pgs -- no change
  141. 2015-01-19 19:50:53.794735 7f28da472700 10 mon.--public-addr@0(leader).pg v10 send_pg_creates to 192 pgs
  142. 2015-01-19 19:50:53.794741 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..33) commit_proposal
  143. 2015-01-19 19:50:53.794746 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..33) commit_proposal proposal 0x42a82a0 took 0.004959 to finish
  144. 2015-01-19 19:50:53.794761 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..10) _active
  145. 2015-01-19 19:50:53.794765 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..10) remove_legacy_versions
  146. 2015-01-19 19:50:53.794786 7f28da472700 7 mon.--public-addr@0(leader).paxosservice(pgmap 1..10) _active creating new pending
  147. 2015-01-19 19:50:53.794803 7f28da472700 10 mon.--public-addr@0(leader).pg v10 create_pending v 11
  148. 2015-01-19 19:50:53.794808 7f28da472700 10 mon.--public-addr@0(leader).pg v10 check_osd_map already seen 9 >= 9
  149. 2015-01-19 19:50:53.794811 7f28da472700 10 mon.--public-addr@0(leader).pg v10 update_logger
  150. 2015-01-19 19:50:53.794845 7f28da472700 0 log [INF] : pgmap v10: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
  151. 2015-01-19 19:50:53.794884 7f28da472700 1 -- 127.0.0.1:6789/0 --> mon.0 127.0.0.1:6789/0 -- log(1 entries) v1 -- ?+0 0x3ed9f80
  152. 2015-01-19 19:50:53.794904 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(1 entries) v1 local
  153. 2015-01-19 19:50:53.794937 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..33) finish_round state 1 proposals left 0
  154. 2015-01-19 19:50:53.795008 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  155. 2015-01-19 19:50:53.795033 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
  156. 2015-01-19 19:50:53.795058 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  157. 2015-01-19 19:50:53.795065 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..10) dispatch log(1 entries) v1 from mon.0 127.0.0.1:6789/0
  158. 2015-01-19 19:50:53.795076 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..33) is_readable now=2015-01-19 19:50:53.795114 lease_expire=0.000000 has v0 lc 33
  159. 2015-01-19 19:50:53.795132 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 preprocess_query log(1 entries) v1 from mon.0 127.0.0.1:6789/0
  160. 2015-01-19 19:50:53.795143 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 preprocess_log log(1 entries) v1 from mon.0
  161. 2015-01-19 19:50:53.795150 7f28d9c71700 20 is_capable service=log command= write on cap allow *
  162. 2015-01-19 19:50:53.795153 7f28d9c71700 20 allow so far , doing grant allow *
  163. 2015-01-19 19:50:53.795156 7f28d9c71700 20 allow all
  164. 2015-01-19 19:50:53.795170 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 prepare_update log(1 entries) v1 from mon.0 127.0.0.1:6789/0
  165. 2015-01-19 19:50:53.795179 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 prepare_log log(1 entries) v1 from mon.0
  166. 2015-01-19 19:50:53.795184 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 logging 2015-01-19 19:50:53.789842 mon.0 127.0.0.1:6789/0 25 : [INF] osdmap e9: 0 osds: 0 up, 0 in
  167. 2015-01-19 19:50:53.795207 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..10) setting proposal_timer 0x3c583f0 with delay of 0.0950649
  168. 2015-01-19 19:50:53.795234 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3edca40
  169. 2015-01-19 19:50:53.795252 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(1 entries) v1 ==== 0+0+0 (0 0 0) 0x3ed9f80 con 0x3d00c60
  170. 2015-01-19 19:50:53.795268 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  171. 2015-01-19 19:50:53.795271 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
  172. 2015-01-19 19:50:53.795281 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  173. 2015-01-19 19:50:53.795285 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..10) dispatch log(1 entries) v1 from mon.0 127.0.0.1:6789/0
  174. 2015-01-19 19:50:53.795293 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..33) is_readable now=2015-01-19 19:50:53.795294 lease_expire=0.000000 has v0 lc 33
  175. 2015-01-19 19:50:53.795305 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 preprocess_query log(1 entries) v1 from mon.0 127.0.0.1:6789/0
  176. 2015-01-19 19:50:53.795313 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 preprocess_log log(1 entries) v1 from mon.0
  177. 2015-01-19 19:50:53.795320 7f28d9c71700 20 is_capable service=log command= write on cap allow *
  178. 2015-01-19 19:50:53.795323 7f28d9c71700 20 allow so far , doing grant allow *
  179. 2015-01-19 19:50:53.795325 7f28d9c71700 20 allow all
  180. 2015-01-19 19:50:53.795333 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 prepare_update log(1 entries) v1 from mon.0 127.0.0.1:6789/0
  181. 2015-01-19 19:50:53.795341 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 prepare_log log(1 entries) v1 from mon.0
  182. 2015-01-19 19:50:53.795346 7f28d9c71700 10 mon.--public-addr@0(leader).log v10 logging 2015-01-19 19:50:53.794850 mon.0 127.0.0.1:6789/0 26 : [INF] pgmap v10: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
  183. 2015-01-19 19:50:53.795368 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..10) proposal_timer already set
  184. 2015-01-19 19:50:53.795374 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3ed9f80
  185. 2015-01-19 19:50:53.798643 7f28e027f700 2 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01b80).reader couldn't read tag, (0) Success
  186. 2015-01-19 19:50:53.798710 7f28e027f700 2 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01b80).fault (0) Success
  187. 2015-01-19 19:50:53.798780 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01b80).fault on lossy channel, failing
  188. 2015-01-19 19:50:53.798797 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01b80).stop
  189. 2015-01-19 19:50:53.798815 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).unregister_pipe
  190. 2015-01-19 19:50:53.798829 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).discard_queue
  191. 2015-01-19 19:50:53.798863 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).writer finishing
  192. 2015-01-19 19:50:53.798857 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).reader done
  193. 2015-01-19 19:50:53.798877 7f28d9c71700 10 mon.--public-addr@0(leader) e1 ms_handle_reset 0x3d01b80 127.0.0.1:0/1011105
  194. 2015-01-19 19:50:53.798900 7f28d846e700 10 -- 127.0.0.1:6789/0 queue_reap 0x3d71400
  195. 2015-01-19 19:50:53.798915 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).writer done
  196. 2015-01-19 19:50:53.798910 7f28d9c71700 10 mon.--public-addr@0(leader) e1 reset/close on session client.? 127.0.0.1:0/1011105
  197. 2015-01-19 19:50:53.798926 7f28d9c71700 10 mon.--public-addr@0(leader) e1 remove_session MonSession: client.? 127.0.0.1:0/1011105 is openallow * client.? 127.0.0.1:0/1011105
  198. 2015-01-19 19:50:53.798932 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper
  199. 2015-01-19 19:50:53.798957 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper reaping pipe 0x3d71400 127.0.0.1:0/1011105
  200. 2015-01-19 19:50:53.798966 7f28dac73700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).discard_queue
  201. 2015-01-19 19:50:53.798984 7f28dac73700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).unregister_pipe - not registered
  202. 2015-01-19 19:50:53.798999 7f28dac73700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011105 pipe(0x3d71400 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01b80).join
  203. 2015-01-19 19:50:53.799030 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper reaped pipe 0x3d71400 127.0.0.1:0/1011105
  204. 2015-01-19 19:50:53.799044 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper deleted pipe 0x3d71400
  205. 2015-01-19 19:50:53.799051 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper done
  206. 2015-01-19 19:50:53.890369 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..10) propose_pending
  207. 2015-01-19 19:50:53.890402 7f28da472700 10 mon.--public-addr@0(leader).log v10 encode_full log v 10
  208. 2015-01-19 19:50:53.890517 7f28da472700 10 mon.--public-addr@0(leader).log v10 encode_pending v11
  209. 2015-01-19 19:50:53.890553 7f28da472700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..33) queue_proposal bl 6214 bytes; ctx = 0x3c58400
  210. 2015-01-19 19:50:53.890569 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..33) propose_queued 34 6214 bytes
  211. 2015-01-19 19:50:53.890574 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..33) begin for 34 6214 bytes
  212. 2015-01-19 19:50:53.891309 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..33) commit 34
  213. 2015-01-19 19:50:53.891968 7f28da472700 10 mon.--public-addr@0(leader) e1 refresh_from_paxos
  214. 2015-01-19 19:50:53.892045 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..10) refresh
  215. 2015-01-19 19:50:53.892096 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(mdsmap 1..1) refresh
  216. 2015-01-19 19:50:53.892143 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..9) refresh
  217. 2015-01-19 19:50:53.892190 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) refresh
  218. 2015-01-19 19:50:53.892197 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos
  219. 2015-01-19 19:50:53.892201 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos version 11 summary v 10
  220. 2015-01-19 19:50:53.892220 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos latest full 10
  221. 2015-01-19 19:50:53.892245 7f28da472700 7 mon.--public-addr@0(leader).log v11 update_from_paxos applying incremental log 11 2015-01-19 19:50:53.789842 mon.0 127.0.0.1:6789/0 25 : [INF] osdmap e9: 0 osds: 0 up, 0 in
  222. 2015-01-19 19:50:53.892289 7f28da472700 7 mon.--public-addr@0(leader).log v11 update_from_paxos applying incremental log 11 2015-01-19 19:50:53.794850 mon.0 127.0.0.1:6789/0 26 : [INF] pgmap v10: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
  223. 2015-01-19 19:50:53.789842 mon.0 127.0.0.1:6789/0 25 : [INF] osdmap e9: 0 osds: 0 up, 0 in
  224. 2015-01-19 19:50:53.794850 mon.0 127.0.0.1:6789/0 26 : [INF] pgmap v10: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
  225. 2015-01-19 19:50:53.892338 7f28da472700 10 mon.--public-addr@0(leader).log v11 check_subs
  226. 2015-01-19 19:50:53.892402 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(monmap 1..1) refresh
  227. 2015-01-19 19:50:53.892453 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) refresh
  228. 2015-01-19 19:50:53.892460 7f28da472700 10 mon.--public-addr@0(leader).auth v2 update_from_paxos
  229. 2015-01-19 19:50:53.892465 7f28da472700 10 mon.--public-addr@0(leader).pg v10 map_pg_creates to 192 pgs -- no change
  230. 2015-01-19 19:50:53.892469 7f28da472700 10 mon.--public-addr@0(leader).pg v10 send_pg_creates to 192 pgs
  231. 2015-01-19 19:50:53.892474 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..34) commit_proposal
  232. 2015-01-19 19:50:53.892478 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..34) commit_proposal proposal 0x42afd20 took 0.001912 to finish
  233. 2015-01-19 19:50:53.892485 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) _active
  234. 2015-01-19 19:50:53.892488 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) remove_legacy_versions
  235. 2015-01-19 19:50:53.892502 7f28da472700 7 mon.--public-addr@0(leader).paxosservice(logm 1..11) _active creating new pending
  236. 2015-01-19 19:50:53.892514 7f28da472700 10 mon.--public-addr@0(leader).log v11 create_pending v 12
  237. 2015-01-19 19:50:53.892519 7f28da472700 7 mon.--public-addr@0(leader).log v11 _updated_log for mon.0 127.0.0.1:6789/0
  238. 2015-01-19 19:50:53.892532 7f28da472700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:6789/0 -- log(last 25) v1 -- ?+0 0x3e91500 con 0x3d00c60
  239. 2015-01-19 19:50:53.892541 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(last 25) v1 local
  240. 2015-01-19 19:50:53.892569 7f28da472700 7 mon.--public-addr@0(leader).log v11 _updated_log for mon.0 127.0.0.1:6789/0
  241. 2015-01-19 19:50:53.892577 7f28da472700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:6789/0 -- log(last 26) v1 -- ?+0 0x3e92680 con 0x3d00c60
  242. 2015-01-19 19:50:53.892584 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(last 26) v1 local
  243. 2015-01-19 19:50:53.892611 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..34) finish_round state 1 proposals left 0
  244. 2015-01-19 19:50:53.892597 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(last 25) v1 ==== 0+0+0 (0 0 0) 0x3e91500 con 0x3d00c60
  245. 2015-01-19 19:50:53.892674 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  246. 2015-01-19 19:50:53.892681 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
  247. 2015-01-19 19:50:53.892701 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  248. 2015-01-19 19:50:53.892727 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e91500
  249. 2015-01-19 19:50:53.892744 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(last 26) v1 ==== 0+0+0 (0 0 0) 0x3e92680 con 0x3d00c60
  250. 2015-01-19 19:50:53.892763 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  251. 2015-01-19 19:50:53.892767 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
  252. 2015-01-19 19:50:53.892777 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  253. 2015-01-19 19:50:53.892796 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e92680
  254. 2015-01-19 19:50:54.009019 7f28d9470700 20 accepter.accepter poll got 1
  255. 2015-01-19 19:50:54.009046 7f28d9470700 10 accepter.pfd.revents=1
  256. 2015-01-19 19:50:54.009056 7f28d9470700 10 accepter.accepted incoming on sd 21
  257. 2015-01-19 19:50:54.009106 7f28d9470700 20 accepter.accepter calling poll
  258. 2015-01-19 19:50:54.009113 7f28e027f700 10 -- 127.0.0.1:6789/0 >> :/0 pipe(0x3d70f00 sd=21 :0 s=0 pgs=0 cs=0 l=0 c=0x3d01ce0).accept
  259. 2015-01-19 19:50:54.009236 7f28e027f700 1 -- 127.0.0.1:6789/0 >> :/0 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=0 c=0x3d01ce0).accept sd=21 127.0.0.1:56336/0
  260. 2015-01-19 19:50:54.009374 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=0 c=0x3d01ce0).accept peer addr is 127.0.0.1:0/1011140
  261. 2015-01-19 19:50:54.009419 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=0 c=0x3d01ce0).accept got peer connect_seq 0 global_seq 1
  262. 2015-01-19 19:50:54.009441 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d01ce0).accept of host_type 8, policy.lossy=1 policy.server=1 policy.standby=0 policy.resetcheck=0
  263. 2015-01-19 19:50:54.009455 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d01ce0).accept my proto 15, their proto 15
  264. 2015-01-19 19:50:54.009466 7f28e027f700 10 mon.--public-addr@0(leader) e1 ms_verify_authorizer 127.0.0.1:0/1011140 client protocol 0
  265. 2015-01-19 19:50:54.009475 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d01ce0).accept: setting up session_security.
  266. 2015-01-19 19:50:54.009486 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d01ce0).accept new session
  267. 2015-01-19 19:50:54.009495 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).accept success, connect_seq = 1, sending READY
  268. 2015-01-19 19:50:54.009504 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).accept features 52776558133247
  269. 2015-01-19 19:50:54.009532 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).register_pipe
  270. 2015-01-19 19:50:54.009572 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).discard_requeued_up_to 0
  271. 2015-01-19 19:50:54.009587 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).accept starting writer, state open
  272. 2015-01-19 19:50:54.009634 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).accept done
  273. 2015-01-19 19:50:54.009662 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  274. 2015-01-19 19:50:54.009670 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  275. 2015-01-19 19:50:54.009703 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
  276. 2015-01-19 19:50:54.009810 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got KEEPALIVE2 2015-01-19 19:50:54.009760
  277. 2015-01-19 19:50:54.009841 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  278. 2015-01-19 19:50:54.009846 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  279. 2015-01-19 19:50:54.009861 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_keepalive2 15 2015-01-19 19:50:54.009760
  280. 2015-01-19 19:50:54.009925 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  281. 2015-01-19 19:50:54.009944 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
  282. 2015-01-19 19:50:54.009976 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got MSG
  283. 2015-01-19 19:50:54.009997 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got envelope type=17 src client.? front=60 data=0 off 0
  284. 2015-01-19 19:50:54.010014 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 60 bytes from policy throttler 0/104857600
  285. 2015-01-19 19:50:54.010027 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 60 from dispatch throttler 0/104857600
  286. 2015-01-19 19:50:54.010046 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got front 60
  287. 2015-01-19 19:50:54.010059 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).aborted = 0
  288. 2015-01-19 19:50:54.010068 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got 60 + 0 + 0 byte message
  289. 2015-01-19 19:50:54.010099 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got message 1 0x3edcc80 auth(proto 0 30 bytes epoch 0) v1
  290. 2015-01-19 19:50:54.010114 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3edcc80 prio 127
  291. 2015-01-19 19:50:54.010138 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  292. 2015-01-19 19:50:54.010143 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  293. 2015-01-19 19:50:54.010160 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_ack 1
  294. 2015-01-19 19:50:54.010176 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  295. 2015-01-19 19:50:54.010186 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
  296. 2015-01-19 19:50:54.010165 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.? 127.0.0.1:0/1011140 1 ==== auth(proto 0 30 bytes epoch 0) v1 ==== 60+0+0 (900162395 0 0) 0x3edcc80 con 0x3d01ce0
  297. 2015-01-19 19:50:54.010212 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  298. 2015-01-19 19:50:54.010219 7f28d9c71700 10 mon.--public-addr@0(leader) e1 do not have session, making new one
  299. 2015-01-19 19:50:54.010227 7f28d9c71700 10 mon.--public-addr@0(leader) e1 ms_dispatch new session MonSession: client.? 127.0.0.1:0/1011140 is open for client.? 127.0.0.1:0/1011140
  300. 2015-01-19 19:50:54.010240 7f28d9c71700 10 mon.--public-addr@0(leader) e1 setting timeout on session
  301. 2015-01-19 19:50:54.010244 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps
  302. 2015-01-19 19:50:54.010249 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) dispatch auth(proto 0 30 bytes epoch 0) v1 from client.? 127.0.0.1:0/1011140
  303. 2015-01-19 19:50:54.010261 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..34) is_readable now=2015-01-19 19:50:54.010263 lease_expire=0.000000 has v0 lc 34
  304. 2015-01-19 19:50:54.010280 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 preprocess_query auth(proto 0 30 bytes epoch 0) v1 from client.? 127.0.0.1:0/1011140
  305. 2015-01-19 19:50:54.010291 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 prep_auth() blob_size=30
  306. 2015-01-19 19:50:54.010324 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 AuthMonitor::assign_global_id m=auth(proto 0 30 bytes epoch 0) v1 mon=0/1 last_allocated=4113 max_global_id=4196
  307. 2015-01-19 19:50:54.010331 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 next_global_id should be 4114
  308. 2015-01-19 19:50:54.010361 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011140 -- mon_map v1 -- ?+0 0x3cc1680 con 0x3d01ce0
  309. 2015-01-19 19:50:54.010375 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_map v1 remote, 127.0.0.1:0/1011140, have pipe.
  310. 2015-01-19 19:50:54.010418 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011140 -- auth_reply(proto 1 0 (0) Success) v1 -- ?+0 0x3ce9c00 con 0x3d01ce0
  311. 2015-01-19 19:50:54.010408 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  312. 2015-01-19 19:50:54.010469 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 1 features 52776558133247 0x3cc1680 mon_map v1
  313. 2015-01-19 19:50:54.010513 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 1): sig = 0
  314. 2015-01-19 19:50:54.010541 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 1 0x3cc1680
  315. 2015-01-19 19:50:54.010548 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message auth_reply(proto 1 0 (0) Success) v1 remote, 127.0.0.1:0/1011140, have pipe.
  316. 2015-01-19 19:50:54.010578 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 60 to dispatch throttler 60/104857600
  317. 2015-01-19 19:50:54.010592 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3edcc80
  318. 2015-01-19 19:50:54.010630 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  319. 2015-01-19 19:50:54.010658 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 2 features 52776558133247 0x3ce9c00 auth_reply(proto 1 0 (0) Success) v1
  320. 2015-01-19 19:50:54.010701 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 2): sig = 0
  321. 2015-01-19 19:50:54.010719 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 2 0x3ce9c00
  322. 2015-01-19 19:50:54.010764 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  323. 2015-01-19 19:50:54.010783 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
  324. 2015-01-19 19:50:54.011437 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
  325. 2015-01-19 19:50:54.011464 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 1
  326. 2015-01-19 19:50:54.011476 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  327. 2015-01-19 19:50:54.011486 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
  328. 2015-01-19 19:50:54.011495 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 2
  329. 2015-01-19 19:50:54.011502 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  330. 2015-01-19 19:50:54.011511 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got MSG
  331. 2015-01-19 19:50:54.011521 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got envelope type=15 src client.? front=23 data=0 off 0
  332. 2015-01-19 19:50:54.011533 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 23 bytes from policy throttler 0/104857600
  333. 2015-01-19 19:50:54.011545 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 23 from dispatch throttler 0/104857600
  334. 2015-01-19 19:50:54.011559 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got front 23
  335. 2015-01-19 19:50:54.011570 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).aborted = 0
  336. 2015-01-19 19:50:54.011578 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got 23 + 0 + 0 byte message
  337. 2015-01-19 19:50:54.011609 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got message 2 0x3e91a40 mon_subscribe({monmap=0+}) v2
  338. 2015-01-19 19:50:54.011636 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3e91a40 prio 127
  339. 2015-01-19 19:50:54.011674 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  340. 2015-01-19 19:50:54.011689 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got MSG
  341. 2015-01-19 19:50:54.011679 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  342. 2015-01-19 19:50:54.011699 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got envelope type=15 src client.4114 front=42 data=0 off 0
  343. 2015-01-19 19:50:54.011702 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_ack 2
  344. 2015-01-19 19:50:54.011709 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 42 bytes from policy throttler 23/104857600
  345. 2015-01-19 19:50:54.011693 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.? 127.0.0.1:0/1011140 2 ==== mon_subscribe({monmap=0+}) v2 ==== 23+0+0 (1620593354 0 0) 0x3e91a40 con 0x3d01ce0
  346. 2015-01-19 19:50:54.011718 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 42 from dispatch throttler 23/104857600
  347. 2015-01-19 19:50:54.011722 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  348. 2015-01-19 19:50:54.011726 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  349. 2015-01-19 19:50:54.011732 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got front 42
  350. 2015-01-19 19:50:54.011730 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
  351. 2015-01-19 19:50:54.011733 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011140 is openallow * for client.? 127.0.0.1:0/1011140
  352. 2015-01-19 19:50:54.011741 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).aborted = 0
  353. 2015-01-19 19:50:54.011750 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  354. 2015-01-19 19:50:54.011749 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got 42 + 0 + 0 byte message
  355. 2015-01-19 19:50:54.011754 7f28d9c71700 10 mon.--public-addr@0(leader) e1 handle_subscribe mon_subscribe({monmap=0+}) v2
  356. 2015-01-19 19:50:54.011768 7f28d9c71700 10 mon.--public-addr@0(leader) e1 check_sub monmap next 0 have 1
  357. 2015-01-19 19:50:54.011773 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got message 3 0x3e90a80 mon_subscribe({monmap=2+,osdmap=0}) v2
  358. 2015-01-19 19:50:54.011785 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3e90a80 prio 127
  359. 2015-01-19 19:50:54.011785 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011140 -- mon_map v1 -- ?+0 0x3cc0b40 con 0x3d01ce0
  360. 2015-01-19 19:50:54.011799 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  361. 2015-01-19 19:50:54.011803 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  362. 2015-01-19 19:50:54.011809 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got MSG
  363. 2015-01-19 19:50:54.011818 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_ack 3
  364. 2015-01-19 19:50:54.011819 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got envelope type=15 src client.4114 front=42 data=0 off 0
  365. 2015-01-19 19:50:54.011823 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_map v1 remote, 127.0.0.1:0/1011140, have pipe.
  366. 2015-01-19 19:50:54.011829 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 42 bytes from policy throttler 65/104857600
  367. 2015-01-19 19:50:54.011837 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 42 from dispatch throttler 65/104857600
  368. 2015-01-19 19:50:54.011844 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.? 127.0.0.1:0/1011140 -- mon_subscribe_ack(300s) v1 -- ?+0 0x3e92680
  369. 2015-01-19 19:50:54.011850 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got front 42
  370. 2015-01-19 19:50:54.011848 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 3 features 52776558133247 0x3cc0b40 mon_map v1
  371. 2015-01-19 19:50:54.011871 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 3): sig = 0
  372. 2015-01-19 19:50:54.011878 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).aborted = 0
  373. 2015-01-19 19:50:54.011885 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 3 0x3cc0b40
  374. 2015-01-19 19:50:54.011888 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got 42 + 0 + 0 byte message
  375. 2015-01-19 19:50:54.011889 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_subscribe_ack(300s) v1 remote, 127.0.0.1:0/1011140, have pipe.
  376. 2015-01-19 19:50:54.011902 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 23 to dispatch throttler 107/104857600
  377. 2015-01-19 19:50:54.011912 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e91a40
  378. 2015-01-19 19:50:54.011908 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got message 4 0x3e91c00 mon_subscribe({monmap=2+,osdmap=0}) v2
  379. 2015-01-19 19:50:54.011935 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3e91c00 prio 127
  380. 2015-01-19 19:50:54.011930 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4114 127.0.0.1:0/1011140 3 ==== mon_subscribe({monmap=2+,osdmap=0}) v2 ==== 42+0+0 (2069260600 0 0) 0x3e90a80 con 0x3d01ce0
  381. 2015-01-19 19:50:54.011946 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  382. 2015-01-19 19:50:54.011948 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011140 is openallow * for client.? 127.0.0.1:0/1011140
  383. 2015-01-19 19:50:54.011956 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  384. 2015-01-19 19:50:54.011959 7f28d9c71700 10 mon.--public-addr@0(leader) e1 handle_subscribe mon_subscribe({monmap=2+,osdmap=0}) v2
  385. 2015-01-19 19:50:54.011955 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  386. 2015-01-19 19:50:54.011966 7f28d9c71700 10 mon.--public-addr@0(leader) e1 check_sub monmap next 2 have 1
  387. 2015-01-19 19:50:54.011974 7f28d9c71700 20 is_capable service=osd command= read on cap allow *
  388. 2015-01-19 19:50:54.011977 7f28d9c71700 20 allow so far , doing grant allow *
  389. 2015-01-19 19:50:54.011967 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  390. 2015-01-19 19:50:54.011979 7f28d9c71700 20 allow all
  391. 2015-01-19 19:50:54.011982 7f28d9c71700 10 mon.--public-addr@0(leader).osd e9 check_sub 0x3c45240 next 0 (onetime)
  392. 2015-01-19 19:50:54.011981 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_ack 4
  393. 2015-01-19 19:50:54.011996 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 4 features 52776558133247 0x3e92680 mon_subscribe_ack(300s) v1
  394. 2015-01-19 19:50:54.012015 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 4): sig = 0
  395. 2015-01-19 19:50:54.012024 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 4 0x3e92680
  396. 2015-01-19 19:50:54.012053 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  397. 2015-01-19 19:50:54.012063 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
  398. 2015-01-19 19:50:54.012113 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.? 127.0.0.1:0/1011140 -- osd_map(9..9 src has 1..9) v3 -- ?+0 0x3edcc80
  399. 2015-01-19 19:50:54.012131 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message osd_map(9..9 src has 1..9) v3 remote, 127.0.0.1:0/1011140, have pipe.
  400. 2015-01-19 19:50:54.012149 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  401. 2015-01-19 19:50:54.012152 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.4114 127.0.0.1:0/1011140 -- mon_subscribe_ack(300s) v1 -- ?+0 0x3e91a40
  402. 2015-01-19 19:50:54.012163 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 5 features 52776558133247 0x3edcc80 osd_map(9..9 src has 1..9) v3
  403. 2015-01-19 19:50:54.012190 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 5): sig = 0
  404. 2015-01-19 19:50:54.012203 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 5 0x3edcc80
  405. 2015-01-19 19:50:54.012208 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_subscribe_ack(300s) v1 remote, 127.0.0.1:0/1011140, have pipe.
  406. 2015-01-19 19:50:54.012224 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 42 to dispatch throttler 84/104857600
  407. 2015-01-19 19:50:54.012231 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e90a80
  408. 2015-01-19 19:50:54.012245 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  409. 2015-01-19 19:50:54.012245 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4114 127.0.0.1:0/1011140 4 ==== mon_subscribe({monmap=2+,osdmap=0}) v2 ==== 42+0+0 (2069260600 0 0) 0x3e91c00 con 0x3d01ce0
  410. 2015-01-19 19:50:54.012259 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  411. 2015-01-19 19:50:54.012261 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011140 is openallow * for client.? 127.0.0.1:0/1011140
  412. 2015-01-19 19:50:54.012259 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 6 features 52776558133247 0x3e91a40 mon_subscribe_ack(300s) v1
  413. 2015-01-19 19:50:54.012269 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  414. 2015-01-19 19:50:54.012272 7f28d9c71700 10 mon.--public-addr@0(leader) e1 handle_subscribe mon_subscribe({monmap=2+,osdmap=0}) v2
  415. 2015-01-19 19:50:54.012279 7f28d9c71700 10 mon.--public-addr@0(leader) e1 check_sub monmap next 2 have 1
  416. 2015-01-19 19:50:54.012276 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 6): sig = 0
  417. 2015-01-19 19:50:54.012285 7f28d9c71700 20 is_capable service=osd command= read on cap allow *
  418. 2015-01-19 19:50:54.012288 7f28d9c71700 20 allow so far , doing grant allow *
  419. 2015-01-19 19:50:54.012290 7f28d9c71700 20 allow all
  420. 2015-01-19 19:50:54.012285 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 6 0x3e91a40
  421. 2015-01-19 19:50:54.012292 7f28d9c71700 10 mon.--public-addr@0(leader).osd e9 check_sub 0x3c44bc0 next 0 (onetime)
  422. 2015-01-19 19:50:54.012312 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  423. 2015-01-19 19:50:54.012321 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
  424. 2015-01-19 19:50:54.012368 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.? 127.0.0.1:0/1011140 -- osd_map(9..9 src has 1..9) v3 -- ?+0 0x3cdca40
  425. 2015-01-19 19:50:54.012385 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message osd_map(9..9 src has 1..9) v3 remote, 127.0.0.1:0/1011140, have pipe.
  426. 2015-01-19 19:50:54.012408 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  427. 2015-01-19 19:50:54.012410 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.4114 127.0.0.1:0/1011140 -- mon_subscribe_ack(300s) v1 -- ?+0 0x3e90a80
  428. 2015-01-19 19:50:54.012420 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 7 features 52776558133247 0x3cdca40 osd_map(9..9 src has 1..9) v3
  429. 2015-01-19 19:50:54.012443 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 7): sig = 0
  430. 2015-01-19 19:50:54.012457 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 7 0x3cdca40
  431. 2015-01-19 19:50:54.012462 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_subscribe_ack(300s) v1 remote, 127.0.0.1:0/1011140, have pipe.
  432. 2015-01-19 19:50:54.012480 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 42 to dispatch throttler 42/104857600
  433. 2015-01-19 19:50:54.012490 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e91c00
  434. 2015-01-19 19:50:54.012503 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  435. 2015-01-19 19:50:54.012518 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 8 features 52776558133247 0x3e90a80 mon_subscribe_ack(300s) v1
  436. 2015-01-19 19:50:54.012533 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 8): sig = 0
  437. 2015-01-19 19:50:54.012542 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 8 0x3e90a80
  438. 2015-01-19 19:50:54.012566 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  439. 2015-01-19 19:50:54.012575 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
  440. 2015-01-19 19:50:54.015276 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
  441. 2015-01-19 19:50:54.015303 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 3
  442. 2015-01-19 19:50:54.015312 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  443. 2015-01-19 19:50:54.015321 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
  444. 2015-01-19 19:50:54.015329 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 4
  445. 2015-01-19 19:50:54.015335 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  446. 2015-01-19 19:50:54.015342 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
  447. 2015-01-19 19:50:54.015364 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 5
  448. 2015-01-19 19:50:54.015371 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  449. 2015-01-19 19:50:54.015379 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
  450. 2015-01-19 19:50:54.015388 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 6
  451. 2015-01-19 19:50:54.015396 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  452. 2015-01-19 19:50:54.015404 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
  453. 2015-01-19 19:50:54.015428 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 7
  454. 2015-01-19 19:50:54.015439 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  455. 2015-01-19 19:50:54.015452 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
  456. 2015-01-19 19:50:54.015464 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 8
  457. 2015-01-19 19:50:54.015473 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  458. 2015-01-19 19:50:54.015485 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got MSG
  459. 2015-01-19 19:50:54.015498 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got envelope type=50 src client.4114 front=80 data=0 off 0
  460. 2015-01-19 19:50:54.015514 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 80 bytes from policy throttler 0/104857600
  461. 2015-01-19 19:50:54.015528 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 80 from dispatch throttler 0/104857600
  462. 2015-01-19 19:50:54.015547 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got front 80
  463. 2015-01-19 19:50:54.015561 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).aborted = 0
  464. 2015-01-19 19:50:54.015572 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got 80 + 0 + 0 byte message
  465. 2015-01-19 19:50:54.015613 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got message 5 0x3cc1680 mon_command({"prefix": "get_command_descriptions"} v 0) v1
  466. 2015-01-19 19:50:54.015649 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3cc1680 prio 127
  467. 2015-01-19 19:50:54.015685 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  468. 2015-01-19 19:50:54.015689 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  469. 2015-01-19 19:50:54.015698 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4114 127.0.0.1:0/1011140 5 ==== mon_command({"prefix": "get_command_descriptions"} v 0) v1 ==== 80+0+0 (450363176 0 0) 0x3cc1680 con 0x3d01ce0
  470. 2015-01-19 19:50:54.015709 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_ack 5
  471. 2015-01-19 19:50:54.015718 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  472. 2015-01-19 19:50:54.015720 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011140 is openallow * for client.? 127.0.0.1:0/1011140
  473. 2015-01-19 19:50:54.015727 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  474. 2015-01-19 19:50:54.015723 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  475. 2015-01-19 19:50:54.015730 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
  476. 2015-01-19 19:50:54.020479 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011140 -- mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1 -- ?+30001 0x3cc1860 con 0x3d01ce0
  477. 2015-01-19 19:50:54.020511 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1 remote, 127.0.0.1:0/1011140, have pipe.
  478. 2015-01-19 19:50:54.020540 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 80 to dispatch throttler 80/104857600
  479. 2015-01-19 19:50:54.020552 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3cc1680
  480. 2015-01-19 19:50:54.020569 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  481. 2015-01-19 19:50:54.020626 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 9 features 52776558133247 0x3cc1860 mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1
  482. 2015-01-19 19:50:54.020658 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 9): sig = 0
  483. 2015-01-19 19:50:54.020671 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 9 0x3cc1860
  484. 2015-01-19 19:50:54.020726 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  485. 2015-01-19 19:50:54.020741 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
  486. 2015-01-19 19:50:54.119996 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ACK
  487. 2015-01-19 19:50:54.120041 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got ack seq 9
  488. 2015-01-19 19:50:54.120055 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  489. 2015-01-19 19:50:54.120069 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got MSG
  490. 2015-01-19 19:50:54.120088 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got envelope type=50 src client.4114 front=104 data=0 off 0
  491. 2015-01-19 19:50:54.120106 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 104 bytes from policy throttler 0/104857600
  492. 2015-01-19 19:50:54.120120 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader wants 104 from dispatch throttler 0/104857600
  493. 2015-01-19 19:50:54.120142 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got front 104
  494. 2015-01-19 19:50:54.120160 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).aborted = 0
  495. 2015-01-19 19:50:54.120171 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got 104 + 0 + 0 byte message
  496. 2015-01-19 19:50:54.120214 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader got message 6 0x3cc1e00 mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1
  497. 2015-01-19 19:50:54.120233 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3cc1e00 prio 127
  498. 2015-01-19 19:50:54.120260 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader reading tag...
  499. 2015-01-19 19:50:54.120279 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  500. 2015-01-19 19:50:54.120288 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4114 127.0.0.1:0/1011140 6 ==== mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 ==== 104+0+0 (4242270301 0 0) 0x3cc1e00 con 0x3d01ce0
  501. 2015-01-19 19:50:54.120300 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).write_ack 6
  502. 2015-01-19 19:50:54.120306 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  503. 2015-01-19 19:50:54.120309 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011140 is openallow * for client.? 127.0.0.1:0/1011140
  504. 2015-01-19 19:50:54.120316 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  505. 2015-01-19 19:50:54.120313 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  506. 2015-01-19 19:50:54.120320 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
  507. 2015-01-19 19:50:54.120444 7f28d9c71700 0 mon.--public-addr@0(leader) e1 handle_command mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1
  508. 2015-01-19 19:50:54.120494 7f28d9c71700 20 is_capable service=osd command=osd erasure-code-profile rm read write on cap allow *
  509. 2015-01-19 19:50:54.120500 7f28d9c71700 20 allow so far , doing grant allow *
  510. 2015-01-19 19:50:54.120502 7f28d9c71700 20 allow all
  511. 2015-01-19 19:50:54.120503 7f28d9c71700 10 mon.--public-addr@0(leader) e1 _allowed_command capable
  512. 2015-01-19 19:50:54.120506 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..9) dispatch mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 from client.4114 127.0.0.1:0/1011140
  513. 2015-01-19 19:50:54.120516 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..34) is_readable now=2015-01-19 19:50:54.120517 lease_expire=0.000000 has v0 lc 34
  514. 2015-01-19 19:50:54.120525 7f28d9c71700 10 mon.--public-addr@0(leader).osd e9 preprocess_query mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 from client.4114 127.0.0.1:0/1011140
  515. 2015-01-19 19:50:54.120673 7f28d9c71700 7 mon.--public-addr@0(leader).osd e9 prepare_update mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 from client.4114 127.0.0.1:0/1011140
  516. 2015-01-19 19:50:54.120819 7f28d9c71700 10 mon.--public-addr@0(leader).osd e9 should_propose
  517. 2015-01-19 19:50:54.120829 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..9) setting proposal_timer 0x3c583e0 with delay of 0.05
  518. 2015-01-19 19:50:54.120862 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 104 to dispatch throttler 104/104857600
  519. 2015-01-19 19:50:54.120872 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3cc1e00
  520. 2015-01-19 19:50:54.170935 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..9) propose_pending
  521. 2015-01-19 19:50:54.170952 7f28da472700 10 mon.--public-addr@0(leader).osd e9 encode_pending e 10
  522. 2015-01-19 19:50:54.171009 7f28da472700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..34) queue_proposal bl 287 bytes; ctx = 0x3c583f0
  523. 2015-01-19 19:50:54.171021 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..34) propose_queued 35 287 bytes
  524. 2015-01-19 19:50:54.171026 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..34) begin for 35 287 bytes
  525. 2015-01-19 19:50:54.171406 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..34) commit 35
  526. 2015-01-19 19:50:54.171710 7f28da472700 10 mon.--public-addr@0(leader) e1 refresh_from_paxos
  527. 2015-01-19 19:50:54.171836 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..10) refresh
  528. 2015-01-19 19:50:54.171885 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(mdsmap 1..1) refresh
  529. 2015-01-19 19:50:54.171928 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) refresh
  530. 2015-01-19 19:50:54.171934 7f28da472700 15 mon.--public-addr@0(leader).osd e9 update_from_paxos paxos e 10, my e 9
  531. 2015-01-19 19:50:54.171987 7f28da472700 7 mon.--public-addr@0(leader).osd e9 update_from_paxos applying incremental 10
  532. 2015-01-19 19:50:54.172094 7f28da472700 1 mon.--public-addr@0(leader).osd e10 e10: 0 osds: 0 up, 0 in
  533. 2015-01-19 19:50:54.172313 7f28da472700 10 mon.--public-addr@0(leader).pg v10 check_osd_map -- osdmap not readable, waiting
  534. 2015-01-19 19:50:54.172326 7f28da472700 10 mon.--public-addr@0(leader).osd e10 check_subs
  535. 2015-01-19 19:50:54.172332 7f28da472700 10 mon.--public-addr@0(leader).osd e10 share_map_with_random_osd no up osds, don't share with anyone
  536. 2015-01-19 19:50:54.172336 7f28da472700 10 mon.--public-addr@0(leader).osd e10 update_logger
  537. 2015-01-19 19:50:54.172406 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) refresh
  538. 2015-01-19 19:50:54.172414 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos
  539. 2015-01-19 19:50:54.172417 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos version 11 summary v 11
  540. 2015-01-19 19:50:54.172462 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(monmap 1..1) refresh
  541. 2015-01-19 19:50:54.172511 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) refresh
  542. 2015-01-19 19:50:54.172517 7f28da472700 10 mon.--public-addr@0(leader).auth v2 update_from_paxos
  543. 2015-01-19 19:50:54.172524 7f28da472700 10 mon.--public-addr@0(leader).pg v10 map_pg_creates to 192 pgs osdmap epoch 10
  544. 2015-01-19 19:50:54.172546 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.0 -> no osds in epoch 10, skipping
  545. 2015-01-19 19:50:54.172559 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1 -> no osds in epoch 10, skipping
  546. 2015-01-19 19:50:54.172568 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2 -> no osds in epoch 10, skipping
  547. 2015-01-19 19:50:54.172577 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3 -> no osds in epoch 10, skipping
  548. 2015-01-19 19:50:54.172589 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.4 -> no osds in epoch 10, skipping
  549. 2015-01-19 19:50:54.172597 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.5 -> no osds in epoch 10, skipping
  550. 2015-01-19 19:50:54.172605 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.6 -> no osds in epoch 10, skipping
  551. 2015-01-19 19:50:54.172613 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.7 -> no osds in epoch 10, skipping
  552. 2015-01-19 19:50:54.172628 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.8 -> no osds in epoch 10, skipping
  553. 2015-01-19 19:50:54.172637 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.9 -> no osds in epoch 10, skipping
  554. 2015-01-19 19:50:54.172662 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.a -> no osds in epoch 10, skipping
  555. 2015-01-19 19:50:54.172676 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.b -> no osds in epoch 10, skipping
  556. 2015-01-19 19:50:54.172690 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.c -> no osds in epoch 10, skipping
  557. 2015-01-19 19:50:54.172702 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.d -> no osds in epoch 10, skipping
  558. 2015-01-19 19:50:54.172715 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.e -> no osds in epoch 10, skipping
  559. 2015-01-19 19:50:54.172728 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.f -> no osds in epoch 10, skipping
  560. 2015-01-19 19:50:54.172741 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.10 -> no osds in epoch 10, skipping
  561. 2015-01-19 19:50:54.172754 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.11 -> no osds in epoch 10, skipping
  562. 2015-01-19 19:50:54.172767 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.12 -> no osds in epoch 10, skipping
  563. 2015-01-19 19:50:54.172780 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.13 -> no osds in epoch 10, skipping
  564. 2015-01-19 19:50:54.172793 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.14 -> no osds in epoch 10, skipping
  565. 2015-01-19 19:50:54.172805 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.15 -> no osds in epoch 10, skipping
  566. 2015-01-19 19:50:54.172818 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.16 -> no osds in epoch 10, skipping
  567. 2015-01-19 19:50:54.172831 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.17 -> no osds in epoch 10, skipping
  568. 2015-01-19 19:50:54.172843 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.18 -> no osds in epoch 10, skipping
  569. 2015-01-19 19:50:54.172856 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.19 -> no osds in epoch 10, skipping
  570. 2015-01-19 19:50:54.172870 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1a -> no osds in epoch 10, skipping
  571. 2015-01-19 19:50:54.172883 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1b -> no osds in epoch 10, skipping
  572. 2015-01-19 19:50:54.172895 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1c -> no osds in epoch 10, skipping
  573. 2015-01-19 19:50:54.172907 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1d -> no osds in epoch 10, skipping
  574. 2015-01-19 19:50:54.172920 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1e -> no osds in epoch 10, skipping
  575. 2015-01-19 19:50:54.172932 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.1f -> no osds in epoch 10, skipping
  576. 2015-01-19 19:50:54.172945 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.20 -> no osds in epoch 10, skipping
  577. 2015-01-19 19:50:54.172957 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.21 -> no osds in epoch 10, skipping
  578. 2015-01-19 19:50:54.172970 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.22 -> no osds in epoch 10, skipping
  579. 2015-01-19 19:50:54.172982 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.23 -> no osds in epoch 10, skipping
  580. 2015-01-19 19:50:54.172994 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.24 -> no osds in epoch 10, skipping
  581. 2015-01-19 19:50:54.173007 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.25 -> no osds in epoch 10, skipping
  582. 2015-01-19 19:50:54.173019 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.26 -> no osds in epoch 10, skipping
  583. 2015-01-19 19:50:54.173032 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.27 -> no osds in epoch 10, skipping
  584. 2015-01-19 19:50:54.173045 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.28 -> no osds in epoch 10, skipping
  585. 2015-01-19 19:50:54.173057 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.29 -> no osds in epoch 10, skipping
  586. 2015-01-19 19:50:54.173069 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2a -> no osds in epoch 10, skipping
  587. 2015-01-19 19:50:54.173081 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2b -> no osds in epoch 10, skipping
  588. 2015-01-19 19:50:54.173093 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2c -> no osds in epoch 10, skipping
  589. 2015-01-19 19:50:54.173105 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2d -> no osds in epoch 10, skipping
  590. 2015-01-19 19:50:54.173117 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2e -> no osds in epoch 10, skipping
  591. 2015-01-19 19:50:54.173130 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.2f -> no osds in epoch 10, skipping
  592. 2015-01-19 19:50:54.173143 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.30 -> no osds in epoch 10, skipping
  593. 2015-01-19 19:50:54.173155 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.31 -> no osds in epoch 10, skipping
  594. 2015-01-19 19:50:54.173167 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.32 -> no osds in epoch 10, skipping
  595. 2015-01-19 19:50:54.173179 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.33 -> no osds in epoch 10, skipping
  596. 2015-01-19 19:50:54.173191 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.34 -> no osds in epoch 10, skipping
  597. 2015-01-19 19:50:54.173204 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.35 -> no osds in epoch 10, skipping
  598. 2015-01-19 19:50:54.173216 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.36 -> no osds in epoch 10, skipping
  599. 2015-01-19 19:50:54.173228 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.37 -> no osds in epoch 10, skipping
  600. 2015-01-19 19:50:54.173240 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.38 -> no osds in epoch 10, skipping
  601. 2015-01-19 19:50:54.173252 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.39 -> no osds in epoch 10, skipping
  602. 2015-01-19 19:50:54.173264 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3a -> no osds in epoch 10, skipping
  603. 2015-01-19 19:50:54.173277 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3b -> no osds in epoch 10, skipping
  604. 2015-01-19 19:50:54.173289 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3c -> no osds in epoch 10, skipping
  605. 2015-01-19 19:50:54.173300 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3d -> no osds in epoch 10, skipping
  606. 2015-01-19 19:50:54.173313 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3e -> no osds in epoch 10, skipping
  607. 2015-01-19 19:50:54.173325 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 0.3f -> no osds in epoch 10, skipping
  608. 2015-01-19 19:50:54.173337 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.0 -> no osds in epoch 10, skipping
  609. 2015-01-19 19:50:54.173349 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1 -> no osds in epoch 10, skipping
  610. 2015-01-19 19:50:54.173362 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2 -> no osds in epoch 10, skipping
  611. 2015-01-19 19:50:54.173374 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3 -> no osds in epoch 10, skipping
  612. 2015-01-19 19:50:54.173388 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.4 -> no osds in epoch 10, skipping
  613. 2015-01-19 19:50:54.173400 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.5 -> no osds in epoch 10, skipping
  614. 2015-01-19 19:50:54.173412 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.6 -> no osds in epoch 10, skipping
  615. 2015-01-19 19:50:54.173424 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.7 -> no osds in epoch 10, skipping
  616. 2015-01-19 19:50:54.173439 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.8 -> no osds in epoch 10, skipping
  617. 2015-01-19 19:50:54.173451 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.9 -> no osds in epoch 10, skipping
  618. 2015-01-19 19:50:54.173463 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.a -> no osds in epoch 10, skipping
  619. 2015-01-19 19:50:54.173475 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.b -> no osds in epoch 10, skipping
  620. 2015-01-19 19:50:54.173487 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.c -> no osds in epoch 10, skipping
  621. 2015-01-19 19:50:54.173500 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.d -> no osds in epoch 10, skipping
  622. 2015-01-19 19:50:54.173512 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.e -> no osds in epoch 10, skipping
  623. 2015-01-19 19:50:54.173524 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.f -> no osds in epoch 10, skipping
  624. 2015-01-19 19:50:54.173536 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.10 -> no osds in epoch 10, skipping
  625. 2015-01-19 19:50:54.173554 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.11 -> no osds in epoch 10, skipping
  626. 2015-01-19 19:50:54.173566 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.12 -> no osds in epoch 10, skipping
  627. 2015-01-19 19:50:54.173578 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.13 -> no osds in epoch 10, skipping
  628. 2015-01-19 19:50:54.173590 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.14 -> no osds in epoch 10, skipping
  629. 2015-01-19 19:50:54.173603 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.15 -> no osds in epoch 10, skipping
  630. 2015-01-19 19:50:54.173615 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.16 -> no osds in epoch 10, skipping
  631. 2015-01-19 19:50:54.173637 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.17 -> no osds in epoch 10, skipping
  632. 2015-01-19 19:50:54.173649 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.18 -> no osds in epoch 10, skipping
  633. 2015-01-19 19:50:54.173664 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.19 -> no osds in epoch 10, skipping
  634. 2015-01-19 19:50:54.173676 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1a -> no osds in epoch 10, skipping
  635. 2015-01-19 19:50:54.173688 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1b -> no osds in epoch 10, skipping
  636. 2015-01-19 19:50:54.173701 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1c -> no osds in epoch 10, skipping
  637. 2015-01-19 19:50:54.173713 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1d -> no osds in epoch 10, skipping
  638. 2015-01-19 19:50:54.173726 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1e -> no osds in epoch 10, skipping
  639. 2015-01-19 19:50:54.173738 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.1f -> no osds in epoch 10, skipping
  640. 2015-01-19 19:50:54.173751 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.20 -> no osds in epoch 10, skipping
  641. 2015-01-19 19:50:54.173763 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.21 -> no osds in epoch 10, skipping
  642. 2015-01-19 19:50:54.173776 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.22 -> no osds in epoch 10, skipping
  643. 2015-01-19 19:50:54.173788 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.23 -> no osds in epoch 10, skipping
  644. 2015-01-19 19:50:54.173802 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.24 -> no osds in epoch 10, skipping
  645. 2015-01-19 19:50:54.173814 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.25 -> no osds in epoch 10, skipping
  646. 2015-01-19 19:50:54.173827 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.26 -> no osds in epoch 10, skipping
  647. 2015-01-19 19:50:54.173839 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.27 -> no osds in epoch 10, skipping
  648. 2015-01-19 19:50:54.173852 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.28 -> no osds in epoch 10, skipping
  649. 2015-01-19 19:50:54.173865 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.29 -> no osds in epoch 10, skipping
  650. 2015-01-19 19:50:54.173919 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2a -> no osds in epoch 10, skipping
  651. 2015-01-19 19:50:54.173933 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2b -> no osds in epoch 10, skipping
  652. 2015-01-19 19:50:54.173946 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2c -> no osds in epoch 10, skipping
  653. 2015-01-19 19:50:54.173959 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2d -> no osds in epoch 10, skipping
  654. 2015-01-19 19:50:54.173972 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2e -> no osds in epoch 10, skipping
  655. 2015-01-19 19:50:54.173985 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.2f -> no osds in epoch 10, skipping
  656. 2015-01-19 19:50:54.174000 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.30 -> no osds in epoch 10, skipping
  657. 2015-01-19 19:50:54.174012 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.31 -> no osds in epoch 10, skipping
  658. 2015-01-19 19:50:54.174024 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.32 -> no osds in epoch 10, skipping
  659. 2015-01-19 19:50:54.174037 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.33 -> no osds in epoch 10, skipping
  660. 2015-01-19 19:50:54.174051 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.34 -> no osds in epoch 10, skipping
  661. 2015-01-19 19:50:54.174064 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.35 -> no osds in epoch 10, skipping
  662. 2015-01-19 19:50:54.174077 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.36 -> no osds in epoch 10, skipping
  663. 2015-01-19 19:50:54.174089 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.37 -> no osds in epoch 10, skipping
  664. 2015-01-19 19:50:54.174102 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.38 -> no osds in epoch 10, skipping
  665. 2015-01-19 19:50:54.174115 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.39 -> no osds in epoch 10, skipping
  666. 2015-01-19 19:50:54.174127 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3a -> no osds in epoch 10, skipping
  667. 2015-01-19 19:50:54.174140 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3b -> no osds in epoch 10, skipping
  668. 2015-01-19 19:50:54.174153 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3c -> no osds in epoch 10, skipping
  669. 2015-01-19 19:50:54.174165 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3d -> no osds in epoch 10, skipping
  670. 2015-01-19 19:50:54.174178 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3e -> no osds in epoch 10, skipping
  671. 2015-01-19 19:50:54.174190 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 1.3f -> no osds in epoch 10, skipping
  672. 2015-01-19 19:50:54.174203 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.0 -> no osds in epoch 10, skipping
  673. 2015-01-19 19:50:54.174216 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1 -> no osds in epoch 10, skipping
  674. 2015-01-19 19:50:54.174229 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2 -> no osds in epoch 10, skipping
  675. 2015-01-19 19:50:54.174242 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3 -> no osds in epoch 10, skipping
  676. 2015-01-19 19:50:54.174254 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.4 -> no osds in epoch 10, skipping
  677. 2015-01-19 19:50:54.174267 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.5 -> no osds in epoch 10, skipping
  678. 2015-01-19 19:50:54.174279 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.6 -> no osds in epoch 10, skipping
  679. 2015-01-19 19:50:54.174292 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.7 -> no osds in epoch 10, skipping
  680. 2015-01-19 19:50:54.174306 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.8 -> no osds in epoch 10, skipping
  681. 2015-01-19 19:50:54.174318 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.9 -> no osds in epoch 10, skipping
  682. 2015-01-19 19:50:54.174331 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.a -> no osds in epoch 10, skipping
  683. 2015-01-19 19:50:54.174343 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.b -> no osds in epoch 10, skipping
  684. 2015-01-19 19:50:54.174356 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.c -> no osds in epoch 10, skipping
  685. 2015-01-19 19:50:54.174368 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.d -> no osds in epoch 10, skipping
  686. 2015-01-19 19:50:54.174380 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.e -> no osds in epoch 10, skipping
  687. 2015-01-19 19:50:54.174393 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.f -> no osds in epoch 10, skipping
  688. 2015-01-19 19:50:54.174406 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.10 -> no osds in epoch 10, skipping
  689. 2015-01-19 19:50:54.174423 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.11 -> no osds in epoch 10, skipping
  690. 2015-01-19 19:50:54.174436 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.12 -> no osds in epoch 10, skipping
  691. 2015-01-19 19:50:54.174448 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.13 -> no osds in epoch 10, skipping
  692. 2015-01-19 19:50:54.174461 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.14 -> no osds in epoch 10, skipping
  693. 2015-01-19 19:50:54.174473 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.15 -> no osds in epoch 10, skipping
  694. 2015-01-19 19:50:54.174486 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.16 -> no osds in epoch 10, skipping
  695. 2015-01-19 19:50:54.174498 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.17 -> no osds in epoch 10, skipping
  696. 2015-01-19 19:50:54.174511 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.18 -> no osds in epoch 10, skipping
  697. 2015-01-19 19:50:54.174523 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.19 -> no osds in epoch 10, skipping
  698. 2015-01-19 19:50:54.174535 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1a -> no osds in epoch 10, skipping
  699. 2015-01-19 19:50:54.174548 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1b -> no osds in epoch 10, skipping
  700. 2015-01-19 19:50:54.174560 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1c -> no osds in epoch 10, skipping
  701. 2015-01-19 19:50:54.174572 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1d -> no osds in epoch 10, skipping
  702. 2015-01-19 19:50:54.174585 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1e -> no osds in epoch 10, skipping
  703. 2015-01-19 19:50:54.174597 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.1f -> no osds in epoch 10, skipping
  704. 2015-01-19 19:50:54.174609 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.20 -> no osds in epoch 10, skipping
  705. 2015-01-19 19:50:54.174630 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.21 -> no osds in epoch 10, skipping
  706. 2015-01-19 19:50:54.174643 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.22 -> no osds in epoch 10, skipping
  707. 2015-01-19 19:50:54.174656 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.23 -> no osds in epoch 10, skipping
  708. 2015-01-19 19:50:54.174669 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.24 -> no osds in epoch 10, skipping
  709. 2015-01-19 19:50:54.174681 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.25 -> no osds in epoch 10, skipping
  710. 2015-01-19 19:50:54.174694 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.26 -> no osds in epoch 10, skipping
  711. 2015-01-19 19:50:54.174707 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.27 -> no osds in epoch 10, skipping
  712. 2015-01-19 19:50:54.174720 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.28 -> no osds in epoch 10, skipping
  713. 2015-01-19 19:50:54.174733 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.29 -> no osds in epoch 10, skipping
  714. 2015-01-19 19:50:54.174746 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2a -> no osds in epoch 10, skipping
  715. 2015-01-19 19:50:54.174758 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2b -> no osds in epoch 10, skipping
  716. 2015-01-19 19:50:54.174771 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2c -> no osds in epoch 10, skipping
  717. 2015-01-19 19:50:54.174783 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2d -> no osds in epoch 10, skipping
  718. 2015-01-19 19:50:54.174796 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2e -> no osds in epoch 10, skipping
  719. 2015-01-19 19:50:54.174809 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.2f -> no osds in epoch 10, skipping
  720. 2015-01-19 19:50:54.174821 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.30 -> no osds in epoch 10, skipping
  721. 2015-01-19 19:50:54.174837 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.31 -> no osds in epoch 10, skipping
  722. 2015-01-19 19:50:54.174849 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.32 -> no osds in epoch 10, skipping
  723. 2015-01-19 19:50:54.174862 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.33 -> no osds in epoch 10, skipping
  724. 2015-01-19 19:50:54.174874 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.34 -> no osds in epoch 10, skipping
  725. 2015-01-19 19:50:54.174886 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.35 -> no osds in epoch 10, skipping
  726. 2015-01-19 19:50:54.174899 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.36 -> no osds in epoch 10, skipping
  727. 2015-01-19 19:50:54.174911 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.37 -> no osds in epoch 10, skipping
  728. 2015-01-19 19:50:54.174923 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.38 -> no osds in epoch 10, skipping
  729. 2015-01-19 19:50:54.174936 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.39 -> no osds in epoch 10, skipping
  730. 2015-01-19 19:50:54.174948 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3a -> no osds in epoch 10, skipping
  731. 2015-01-19 19:50:54.174960 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3b -> no osds in epoch 10, skipping
  732. 2015-01-19 19:50:54.174973 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3c -> no osds in epoch 10, skipping
  733. 2015-01-19 19:50:54.174985 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3d -> no osds in epoch 10, skipping
  734. 2015-01-19 19:50:54.174997 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3e -> no osds in epoch 10, skipping
  735. 2015-01-19 19:50:54.175009 7f28da472700 20 mon.--public-addr@0(leader).pg v10 map_pg_creates 2.3f -> no osds in epoch 10, skipping
  736. 2015-01-19 19:50:54.175016 7f28da472700 10 mon.--public-addr@0(leader).pg v10 send_pg_creates to 192 pgs
  737. 2015-01-19 19:50:54.175026 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) commit_proposal
  738. 2015-01-19 19:50:54.175032 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) commit_proposal proposal 0x42af850 took 0.004014 to finish
  739. 2015-01-19 19:50:54.175043 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) _active
  740. 2015-01-19 19:50:54.175046 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) remove_legacy_versions
  741. 2015-01-19 19:50:54.175074 7f28da472700 7 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) _active creating new pending
  742. 2015-01-19 19:50:54.175102 7f28da472700 10 mon.--public-addr@0(leader).osd e10 create_pending e 11
  743. 2015-01-19 19:50:54.175191 7f28da472700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011140 -- mon_command_ack([{"prefix": "osd erasure-code-profile rm", "name": "myprofile"}]=0 v10) v1 -- ?+0 0x3cc10e0 con 0x3d01ce0
  744. 2015-01-19 19:50:54.175215 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message mon_command_ack([{"prefix": "osd erasure-code-profile rm", "name": "myprofile"}]=0 v10) v1 remote, 127.0.0.1:0/1011140, have pipe.
  745. 2015-01-19 19:50:54.175254 7f28da472700 5 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) is_readable now=2015-01-19 19:50:54.175256 lease_expire=0.000000 has v0 lc 35
  746. 2015-01-19 19:50:54.175271 7f28da472700 10 mon.--public-addr@0(leader).pg v10 check_osd_map applying osdmap e10 to pg_map
  747. 2015-01-19 19:50:54.175280 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  748. 2015-01-19 19:50:54.175319 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer encoding 10 features 52776558133247 0x3cc10e0 mon_command_ack([{"prefix": "osd erasure-code-profile rm", "name": "myprofile"}]=0 v10) v1
  749. 2015-01-19 19:50:54.175337 7f28da472700 10 mon.--public-addr@0(leader).pg v10 register_new_pgs checking pg pools for osdmap epoch 10, last_pg_scan 9
  750. 2015-01-19 19:50:54.175343 7f28da472700 10 mon.--public-addr@0(leader).pg v10 no change in pool 0 replicated size 3 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 1 flags hashpspool crash_replay_interval 45 stripe_width 0
  751. 2015-01-19 19:50:54.175344 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer signed seq # 10): sig = 0
  752. 2015-01-19 19:50:54.175350 7f28da472700 10 mon.--public-addr@0(leader).pg v10 no change in pool 1 replicated size 3 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 1 flags hashpspool stripe_width 0
  753. 2015-01-19 19:50:54.175355 7f28da472700 10 mon.--public-addr@0(leader).pg v10 no change in pool 2 replicated size 3 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 1 flags hashpspool stripe_width 0
  754. 2015-01-19 19:50:54.175355 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sending 10 0x3cc10e0
  755. 2015-01-19 19:50:54.175406 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer: state = open policy.server=1
  756. 2015-01-19 19:50:54.175416 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).writer sleeping
  757. 2015-01-19 19:50:54.175432 7f28da472700 10 mon.--public-addr@0(leader).pg v10 register_new_pgs registered 0 new pgs, removed 0 uncreated pgs
  758. 2015-01-19 19:50:54.175438 7f28da472700 10 mon.--public-addr@0(leader).pg v10 check_down_pgs
  759. 2015-01-19 19:50:54.175450 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..10) propose_pending
  760. 2015-01-19 19:50:54.175454 7f28da472700 10 mon.--public-addr@0(leader).pg v10 encode_pending v 11
  761. 2015-01-19 19:50:54.175503 7f28da472700 5 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) queue_proposal bl 422 bytes; ctx = 0x3c58420
  762. 2015-01-19 19:50:54.175519 7f28da472700 5 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) propose_new_value not active; proposal queued
  763. 2015-01-19 19:50:54.175528 7f28da472700 10 mon.--public-addr@0(leader).pg v10 map_pg_creates to 192 pgs -- no change
  764. 2015-01-19 19:50:54.175530 7f28da472700 10 mon.--public-addr@0(leader).pg v10 send_pg_creates to 192 pgs
  765. 2015-01-19 19:50:54.175534 7f28da472700 10 mon.--public-addr@0(leader).osd e10 update_logger
  766. 2015-01-19 19:50:54.175542 7f28da472700 0 log [INF] : osdmap e10: 0 osds: 0 up, 0 in
  767. 2015-01-19 19:50:54.175566 7f28da472700 1 -- 127.0.0.1:6789/0 --> mon.0 127.0.0.1:6789/0 -- log(1 entries) v1 -- ?+0 0x3edca40
  768. 2015-01-19 19:50:54.175574 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(1 entries) v1 local
  769. 2015-01-19 19:50:54.175594 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..35) finish_round state 1 proposals left 1
  770. 2015-01-19 19:50:54.175600 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..35) propose_queued 36 422 bytes
  771. 2015-01-19 19:50:54.175603 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) begin for 36 422 bytes
  772. 2015-01-19 19:50:54.175636 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(1 entries) v1 ==== 0+0+0 (0 0 0) 0x3edca40 con 0x3d00c60
  773. 2015-01-19 19:50:54.175870 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..35) commit 36
  774. 2015-01-19 19:50:54.176266 7f28da472700 10 mon.--public-addr@0(leader) e1 refresh_from_paxos
  775. 2015-01-19 19:50:54.176385 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..11) refresh
  776. 2015-01-19 19:50:54.176399 7f28da472700 10 mon.--public-addr@0(leader).pg v10 update_from_paxos read_incremental
  777. 2015-01-19 19:50:54.176457 7f28da472700 10 mon.--public-addr@0(leader).pg v11 read_pgmap_meta
  778. 2015-01-19 19:50:54.176595 7f28da472700 10 mon.--public-addr@0(leader).pg v11 map_pg_creates to 192 pgs -- no change
  779. 2015-01-19 19:50:54.176606 7f28da472700 10 mon.--public-addr@0(leader).pg v11 send_pg_creates to 192 pgs
  780. 2015-01-19 19:50:54.176610 7f28da472700 10 mon.--public-addr@0(leader).pg v11 update_logger
  781. 2015-01-19 19:50:54.176704 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(mdsmap 1..1) refresh
  782. 2015-01-19 19:50:54.176775 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) refresh
  783. 2015-01-19 19:50:54.176847 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) refresh
  784. 2015-01-19 19:50:54.176858 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos
  785. 2015-01-19 19:50:54.176862 7f28da472700 10 mon.--public-addr@0(leader).log v11 update_from_paxos version 11 summary v 11
  786. 2015-01-19 19:50:54.176932 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(monmap 1..1) refresh
  787. 2015-01-19 19:50:54.177008 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) refresh
  788. 2015-01-19 19:50:54.177018 7f28da472700 10 mon.--public-addr@0(leader).auth v2 update_from_paxos
  789. 2015-01-19 19:50:54.177028 7f28da472700 10 mon.--public-addr@0(leader).pg v11 map_pg_creates to 192 pgs -- no change
  790. 2015-01-19 19:50:54.177033 7f28da472700 10 mon.--public-addr@0(leader).pg v11 send_pg_creates to 192 pgs
  791. 2015-01-19 19:50:54.177045 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..36) commit_proposal
  792. 2015-01-19 19:50:54.177051 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..36) commit_proposal proposal 0x42ead10 took 0.001534 to finish
  793. 2015-01-19 19:50:54.177061 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..11) _active
  794. 2015-01-19 19:50:54.177065 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..11) remove_legacy_versions
  795. 2015-01-19 19:50:54.177088 7f28da472700 7 mon.--public-addr@0(leader).paxosservice(pgmap 1..11) _active creating new pending
  796. 2015-01-19 19:50:54.177105 7f28da472700 10 mon.--public-addr@0(leader).pg v11 create_pending v 12
  797. 2015-01-19 19:50:54.177110 7f28da472700 10 mon.--public-addr@0(leader).pg v11 check_osd_map already seen 10 >= 10
  798. 2015-01-19 19:50:54.177113 7f28da472700 10 mon.--public-addr@0(leader).pg v11 update_logger
  799. 2015-01-19 19:50:54.177148 7f28da472700 0 log [INF] : pgmap v11: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
  800. 2015-01-19 19:50:54.177185 7f28da472700 1 -- 127.0.0.1:6789/0 --> mon.0 127.0.0.1:6789/0 -- log(1 entries) v1 -- ?+0 0x3ed9d40
  801. 2015-01-19 19:50:54.177199 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(1 entries) v1 local
  802. 2015-01-19 19:50:54.177232 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..36) finish_round state 1 proposals left 0
  803. 2015-01-19 19:50:54.177302 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  804. 2015-01-19 19:50:54.177328 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
  805. 2015-01-19 19:50:54.177355 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  806. 2015-01-19 19:50:54.177362 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) dispatch log(1 entries) v1 from mon.0 127.0.0.1:6789/0
  807. 2015-01-19 19:50:54.177375 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..36) is_readable now=2015-01-19 19:50:54.177376 lease_expire=0.000000 has v0 lc 36
  808. 2015-01-19 19:50:54.177391 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 preprocess_query log(1 entries) v1 from mon.0 127.0.0.1:6789/0
  809. 2015-01-19 19:50:54.177400 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 preprocess_log log(1 entries) v1 from mon.0
  810. 2015-01-19 19:50:54.177408 7f28d9c71700 20 is_capable service=log command= write on cap allow *
  811. 2015-01-19 19:50:54.177412 7f28d9c71700 20 allow so far , doing grant allow *
  812. 2015-01-19 19:50:54.177414 7f28d9c71700 20 allow all
  813. 2015-01-19 19:50:54.177431 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 prepare_update log(1 entries) v1 from mon.0 127.0.0.1:6789/0
  814. 2015-01-19 19:50:54.177440 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 prepare_log log(1 entries) v1 from mon.0
  815. 2015-01-19 19:50:54.177446 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 logging 2015-01-19 19:50:54.175544 mon.0 127.0.0.1:6789/0 27 : [INF] osdmap e10: 0 osds: 0 up, 0 in
  816. 2015-01-19 19:50:54.177471 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) setting proposal_timer 0x3c583d0 with delay of 0.0984108
  817. 2015-01-19 19:50:54.177500 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3edca40
  818. 2015-01-19 19:50:54.177532 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(1 entries) v1 ==== 0+0+0 (0 0 0) 0x3ed9d40 con 0x3d00c60
  819. 2015-01-19 19:50:54.177550 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  820. 2015-01-19 19:50:54.177554 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
  821. 2015-01-19 19:50:54.177565 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  822. 2015-01-19 19:50:54.177569 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) dispatch log(1 entries) v1 from mon.0 127.0.0.1:6789/0
  823. 2015-01-19 19:50:54.177578 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..36) is_readable now=2015-01-19 19:50:54.177579 lease_expire=0.000000 has v0 lc 36
  824. 2015-01-19 19:50:54.177590 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 preprocess_query log(1 entries) v1 from mon.0 127.0.0.1:6789/0
  825. 2015-01-19 19:50:54.177598 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 preprocess_log log(1 entries) v1 from mon.0
  826. 2015-01-19 19:50:54.177605 7f28d9c71700 20 is_capable service=log command= write on cap allow *
  827. 2015-01-19 19:50:54.177608 7f28d9c71700 20 allow so far , doing grant allow *
  828. 2015-01-19 19:50:54.177610 7f28d9c71700 20 allow all
  829. 2015-01-19 19:50:54.177635 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 prepare_update log(1 entries) v1 from mon.0 127.0.0.1:6789/0
  830. 2015-01-19 19:50:54.177646 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 prepare_log log(1 entries) v1 from mon.0
  831. 2015-01-19 19:50:54.177651 7f28d9c71700 10 mon.--public-addr@0(leader).log v11 logging 2015-01-19 19:50:54.177152 mon.0 127.0.0.1:6789/0 28 : [INF] pgmap v11: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
  832. 2015-01-19 19:50:54.177675 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) proposal_timer already set
  833. 2015-01-19 19:50:54.177682 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3ed9d40
  834. 2015-01-19 19:50:54.184487 7f28e027f700 2 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).reader couldn't read tag, (0) Success
  835. 2015-01-19 19:50:54.184553 7f28e027f700 2 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).fault (0) Success
  836. 2015-01-19 19:50:54.184660 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).fault on lossy channel, failing
  837. 2015-01-19 19:50:54.184682 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d01ce0).stop
  838. 2015-01-19 19:50:54.184705 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).unregister_pipe
  839. 2015-01-19 19:50:54.184722 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).discard_queue
  840. 2015-01-19 19:50:54.184754 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).reader done
  841. 2015-01-19 19:50:54.184763 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).writer finishing
  842. 2015-01-19 19:50:54.184803 7f28d846e700 10 -- 127.0.0.1:6789/0 queue_reap 0x3d70f00
  843. 2015-01-19 19:50:54.184787 7f28d9c71700 10 mon.--public-addr@0(leader) e1 ms_handle_reset 0x3d01ce0 127.0.0.1:0/1011140
  844. 2015-01-19 19:50:54.184818 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).writer done
  845. 2015-01-19 19:50:54.184820 7f28d9c71700 10 mon.--public-addr@0(leader) e1 reset/close on session client.? 127.0.0.1:0/1011140
  846. 2015-01-19 19:50:54.184835 7f28d9c71700 10 mon.--public-addr@0(leader) e1 remove_session MonSession: client.? 127.0.0.1:0/1011140 is openallow * client.? 127.0.0.1:0/1011140
  847. 2015-01-19 19:50:54.184862 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper
  848. 2015-01-19 19:50:54.184889 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper reaping pipe 0x3d70f00 127.0.0.1:0/1011140
  849. 2015-01-19 19:50:54.184899 7f28dac73700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).discard_queue
  850. 2015-01-19 19:50:54.184918 7f28dac73700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).unregister_pipe - not registered
  851. 2015-01-19 19:50:54.184945 7f28dac73700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011140 pipe(0x3d70f00 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d01ce0).join
  852. 2015-01-19 19:50:54.184977 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper reaped pipe 0x3d70f00 127.0.0.1:0/1011140
  853. 2015-01-19 19:50:54.184992 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper deleted pipe 0x3d70f00
  854. 2015-01-19 19:50:54.185002 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper done
  855. 2015-01-19 19:50:54.275975 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..11) propose_pending
  856. 2015-01-19 19:50:54.276007 7f28da472700 10 mon.--public-addr@0(leader).log v11 encode_full log v 11
  857. 2015-01-19 19:50:54.276134 7f28da472700 10 mon.--public-addr@0(leader).log v11 encode_pending v12
  858. 2015-01-19 19:50:54.276172 7f28da472700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..36) queue_proposal bl 6666 bytes; ctx = 0x3c583e0
  859. 2015-01-19 19:50:54.276189 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..36) propose_queued 37 6666 bytes
  860. 2015-01-19 19:50:54.276194 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..36) begin for 37 6666 bytes
  861. 2015-01-19 19:50:54.276945 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..36) commit 37
  862. 2015-01-19 19:50:54.277667 7f28da472700 10 mon.--public-addr@0(leader) e1 refresh_from_paxos
  863. 2015-01-19 19:50:54.277768 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(pgmap 1..11) refresh
  864. 2015-01-19 19:50:54.277831 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(mdsmap 1..1) refresh
  865. 2015-01-19 19:50:54.277883 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) refresh
  866. 2015-01-19 19:50:54.277934 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..12) refresh
  867. 2015-01-19 19:50:54.277941 7f28da472700 10 mon.--public-addr@0(leader).log v12 update_from_paxos
  868. 2015-01-19 19:50:54.277944 7f28da472700 10 mon.--public-addr@0(leader).log v12 update_from_paxos version 12 summary v 11
  869. 2015-01-19 19:50:54.277965 7f28da472700 10 mon.--public-addr@0(leader).log v12 update_from_paxos latest full 11
  870. 2015-01-19 19:50:54.277993 7f28da472700 7 mon.--public-addr@0(leader).log v12 update_from_paxos applying incremental log 12 2015-01-19 19:50:54.175544 mon.0 127.0.0.1:6789/0 27 : [INF] osdmap e10: 0 osds: 0 up, 0 in
  871. 2015-01-19 19:50:54.278043 7f28da472700 7 mon.--public-addr@0(leader).log v12 update_from_paxos applying incremental log 12 2015-01-19 19:50:54.177152 mon.0 127.0.0.1:6789/0 28 : [INF] pgmap v11: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
  872. 2015-01-19 19:50:54.175544 mon.0 127.0.0.1:6789/0 27 : [INF] osdmap e10: 0 osds: 0 up, 0 in
  873. 2015-01-19 19:50:54.177152 mon.0 127.0.0.1:6789/0 28 : [INF] pgmap v11: 192 pgs: 192 creating; 0 bytes data, 0 kB used, 0 kB / 0 kB avail
  874. 2015-01-19 19:50:54.278099 7f28da472700 10 mon.--public-addr@0(leader).log v12 check_subs
  875. 2015-01-19 19:50:54.278171 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(monmap 1..1) refresh
  876. 2015-01-19 19:50:54.278227 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) refresh
  877. 2015-01-19 19:50:54.278235 7f28da472700 10 mon.--public-addr@0(leader).auth v2 update_from_paxos
  878. 2015-01-19 19:50:54.278241 7f28da472700 10 mon.--public-addr@0(leader).pg v11 map_pg_creates to 192 pgs -- no change
  879. 2015-01-19 19:50:54.278245 7f28da472700 10 mon.--public-addr@0(leader).pg v11 send_pg_creates to 192 pgs
  880. 2015-01-19 19:50:54.278255 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..37) commit_proposal
  881. 2015-01-19 19:50:54.278259 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos updating c 1..37) commit_proposal proposal 0x42ea6f0 took 0.002073 to finish
  882. 2015-01-19 19:50:54.278267 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..12) _active
  883. 2015-01-19 19:50:54.278270 7f28da472700 10 mon.--public-addr@0(leader).paxosservice(logm 1..12) remove_legacy_versions
  884. 2015-01-19 19:50:54.278287 7f28da472700 7 mon.--public-addr@0(leader).paxosservice(logm 1..12) _active creating new pending
  885. 2015-01-19 19:50:54.278301 7f28da472700 10 mon.--public-addr@0(leader).log v12 create_pending v 13
  886. 2015-01-19 19:50:54.278305 7f28da472700 7 mon.--public-addr@0(leader).log v12 _updated_log for mon.0 127.0.0.1:6789/0
  887. 2015-01-19 19:50:54.278319 7f28da472700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:6789/0 -- log(last 27) v1 -- ?+0 0x3e91f80 con 0x3d00c60
  888. 2015-01-19 19:50:54.278329 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(last 27) v1 local
  889. 2015-01-19 19:50:54.278363 7f28da472700 7 mon.--public-addr@0(leader).log v12 _updated_log for mon.0 127.0.0.1:6789/0
  890. 2015-01-19 19:50:54.278376 7f28da472700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:6789/0 -- log(last 28) v1 -- ?+0 0x3e92140 con 0x3d00c60
  891. 2015-01-19 19:50:54.278384 7f28da472700 20 -- 127.0.0.1:6789/0 submit_message log(last 28) v1 local
  892. 2015-01-19 19:50:54.278399 7f28da472700 10 mon.--public-addr@0(leader).paxos(paxos active c 1..37) finish_round state 1 proposals left 0
  893. 2015-01-19 19:50:54.278384 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(last 27) v1 ==== 0+0+0 (0 0 0) 0x3e91f80 con 0x3d00c60
  894. 2015-01-19 19:50:54.278420 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  895. 2015-01-19 19:50:54.278427 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
  896. 2015-01-19 19:50:54.278444 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  897. 2015-01-19 19:50:54.278474 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e91f80
  898. 2015-01-19 19:50:54.278499 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== mon.0 127.0.0.1:6789/0 0 ==== log(last 28) v1 ==== 0+0+0 (0 0 0) 0x3e92140 con 0x3d00c60
  899. 2015-01-19 19:50:54.278521 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  900. 2015-01-19 19:50:54.278525 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: mon.0 127.0.0.1:6789/0 is openallow * for mon.0 127.0.0.1:6789/0
  901. 2015-01-19 19:50:54.278539 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  902. 2015-01-19 19:50:54.278562 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e92140
  903. 2015-01-19 19:50:54.390213 7f28d9470700 20 accepter.accepter poll got 1
  904. 2015-01-19 19:50:54.390240 7f28d9470700 10 accepter.pfd.revents=1
  905. 2015-01-19 19:50:54.390251 7f28d9470700 10 accepter.accepted incoming on sd 21
  906. 2015-01-19 19:50:54.390304 7f28d9470700 20 accepter.accepter calling poll
  907. 2015-01-19 19:50:54.390308 7f28e027f700 10 -- 127.0.0.1:6789/0 >> :/0 pipe(0x3d71180 sd=21 :0 s=0 pgs=0 cs=0 l=0 c=0x3d02680).accept
  908. 2015-01-19 19:50:54.390431 7f28e027f700 1 -- 127.0.0.1:6789/0 >> :/0 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=0 c=0x3d02680).accept sd=21 127.0.0.1:56337/0
  909. 2015-01-19 19:50:54.390605 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=0 c=0x3d02680).accept peer addr is 127.0.0.1:0/1011174
  910. 2015-01-19 19:50:54.390675 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=0 c=0x3d02680).accept got peer connect_seq 0 global_seq 1
  911. 2015-01-19 19:50:54.390695 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d02680).accept of host_type 8, policy.lossy=1 policy.server=1 policy.standby=0 policy.resetcheck=0
  912. 2015-01-19 19:50:54.390705 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d02680).accept my proto 15, their proto 15
  913. 2015-01-19 19:50:54.390713 7f28e027f700 10 mon.--public-addr@0(leader) e1 ms_verify_authorizer 127.0.0.1:0/1011174 client protocol 0
  914. 2015-01-19 19:50:54.390721 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d02680).accept: setting up session_security.
  915. 2015-01-19 19:50:54.390729 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=0 pgs=0 cs=0 l=1 c=0x3d02680).accept new session
  916. 2015-01-19 19:50:54.390735 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).accept success, connect_seq = 1, sending READY
  917. 2015-01-19 19:50:54.390741 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).accept features 52776558133247
  918. 2015-01-19 19:50:54.390769 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).register_pipe
  919. 2015-01-19 19:50:54.390802 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).discard_requeued_up_to 0
  920. 2015-01-19 19:50:54.390813 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).accept starting writer, state open
  921. 2015-01-19 19:50:54.390842 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).accept done
  922. 2015-01-19 19:50:54.390858 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  923. 2015-01-19 19:50:54.390885 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  924. 2015-01-19 19:50:54.390924 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  925. 2015-01-19 19:50:54.390944 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got KEEPALIVE2 2015-01-19 19:50:54.390916
  926. 2015-01-19 19:50:54.390965 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  927. 2015-01-19 19:50:54.390975 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  928. 2015-01-19 19:50:54.390994 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_keepalive2 15 2015-01-19 19:50:54.390916
  929. 2015-01-19 19:50:54.391060 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  930. 2015-01-19 19:50:54.391069 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got MSG
  931. 2015-01-19 19:50:54.391079 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  932. 2015-01-19 19:50:54.391087 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got envelope type=17 src client.? front=60 data=0 off 0
  933. 2015-01-19 19:50:54.391098 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 60 bytes from policy throttler 0/104857600
  934. 2015-01-19 19:50:54.391109 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 60 from dispatch throttler 0/104857600
  935. 2015-01-19 19:50:54.391123 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got front 60
  936. 2015-01-19 19:50:54.391132 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).aborted = 0
  937. 2015-01-19 19:50:54.391138 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got 60 + 0 + 0 byte message
  938. 2015-01-19 19:50:54.391165 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got message 1 0x3cdca40 auth(proto 0 30 bytes epoch 0) v1
  939. 2015-01-19 19:50:54.391178 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3cdca40 prio 127
  940. 2015-01-19 19:50:54.391196 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  941. 2015-01-19 19:50:54.391203 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  942. 2015-01-19 19:50:54.391207 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.? 127.0.0.1:0/1011174 1 ==== auth(proto 0 30 bytes epoch 0) v1 ==== 60+0+0 (900162395 0 0) 0x3cdca40 con 0x3d02680
  943. 2015-01-19 19:50:54.391231 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  944. 2015-01-19 19:50:54.391226 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_ack 1
  945. 2015-01-19 19:50:54.391235 7f28d9c71700 10 mon.--public-addr@0(leader) e1 do not have session, making new one
  946. 2015-01-19 19:50:54.391241 7f28d9c71700 10 mon.--public-addr@0(leader) e1 ms_dispatch new session MonSession: client.? 127.0.0.1:0/1011174 is open for client.? 127.0.0.1:0/1011174
  947. 2015-01-19 19:50:54.391241 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  948. 2015-01-19 19:50:54.391252 7f28d9c71700 10 mon.--public-addr@0(leader) e1 setting timeout on session
  949. 2015-01-19 19:50:54.391255 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps
  950. 2015-01-19 19:50:54.391253 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  951. 2015-01-19 19:50:54.391258 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(auth 1..2) dispatch auth(proto 0 30 bytes epoch 0) v1 from client.? 127.0.0.1:0/1011174
  952. 2015-01-19 19:50:54.391267 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..37) is_readable now=2015-01-19 19:50:54.391268 lease_expire=0.000000 has v0 lc 37
  953. 2015-01-19 19:50:54.391278 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 preprocess_query auth(proto 0 30 bytes epoch 0) v1 from client.? 127.0.0.1:0/1011174
  954. 2015-01-19 19:50:54.391284 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 prep_auth() blob_size=30
  955. 2015-01-19 19:50:54.391302 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 AuthMonitor::assign_global_id m=auth(proto 0 30 bytes epoch 0) v1 mon=0/1 last_allocated=4114 max_global_id=4196
  956. 2015-01-19 19:50:54.391307 7f28d9c71700 10 mon.--public-addr@0(leader).auth v2 next_global_id should be 4115
  957. 2015-01-19 19:50:54.391324 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011174 -- mon_map v1 -- ?+0 0x3cc1680 con 0x3d02680
  958. 2015-01-19 19:50:54.391332 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_map v1 remote, 127.0.0.1:0/1011174, have pipe.
  959. 2015-01-19 19:50:54.391351 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  960. 2015-01-19 19:50:54.391359 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011174 -- auth_reply(proto 1 0 (0) Success) v1 -- ?+0 0x3ce9400 con 0x3d02680
  961. 2015-01-19 19:50:54.391369 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 1 features 52776558133247 0x3cc1680 mon_map v1
  962. 2015-01-19 19:50:54.391389 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 1): sig = 0
  963. 2015-01-19 19:50:54.391401 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 1 0x3cc1680
  964. 2015-01-19 19:50:54.391405 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message auth_reply(proto 1 0 (0) Success) v1 remote, 127.0.0.1:0/1011174, have pipe.
  965. 2015-01-19 19:50:54.391425 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 60 to dispatch throttler 60/104857600
  966. 2015-01-19 19:50:54.391431 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3cdca40
  967. 2015-01-19 19:50:54.391441 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  968. 2015-01-19 19:50:54.391460 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 2 features 52776558133247 0x3ce9400 auth_reply(proto 1 0 (0) Success) v1
  969. 2015-01-19 19:50:54.391496 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 2): sig = 0
  970. 2015-01-19 19:50:54.391518 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 2 0x3ce9400
  971. 2015-01-19 19:50:54.391555 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  972. 2015-01-19 19:50:54.391573 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  973. 2015-01-19 19:50:54.392094 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
  974. 2015-01-19 19:50:54.392121 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 1
  975. 2015-01-19 19:50:54.392130 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  976. 2015-01-19 19:50:54.392138 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
  977. 2015-01-19 19:50:54.392145 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 2
  978. 2015-01-19 19:50:54.392151 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  979. 2015-01-19 19:50:54.392158 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got MSG
  980. 2015-01-19 19:50:54.392165 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got envelope type=15 src client.? front=23 data=0 off 0
  981. 2015-01-19 19:50:54.392175 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 23 bytes from policy throttler 0/104857600
  982. 2015-01-19 19:50:54.392183 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 23 from dispatch throttler 0/104857600
  983. 2015-01-19 19:50:54.392195 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got front 23
  984. 2015-01-19 19:50:54.392203 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).aborted = 0
  985. 2015-01-19 19:50:54.392210 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got 23 + 0 + 0 byte message
  986. 2015-01-19 19:50:54.392236 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got message 2 0x3e90a80 mon_subscribe({monmap=0+}) v2
  987. 2015-01-19 19:50:54.392249 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3e90a80 prio 127
  988. 2015-01-19 19:50:54.392267 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  989. 2015-01-19 19:50:54.392270 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  990. 2015-01-19 19:50:54.392277 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.? 127.0.0.1:0/1011174 2 ==== mon_subscribe({monmap=0+}) v2 ==== 23+0+0 (1620593354 0 0) 0x3e90a80 con 0x3d02680
  991. 2015-01-19 19:50:54.392289 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_ack 2
  992. 2015-01-19 19:50:54.392296 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  993. 2015-01-19 19:50:54.392299 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011174 is openallow * for client.? 127.0.0.1:0/1011174
  994. 2015-01-19 19:50:54.392307 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  995. 2015-01-19 19:50:54.392305 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  996. 2015-01-19 19:50:54.392310 7f28d9c71700 10 mon.--public-addr@0(leader) e1 handle_subscribe mon_subscribe({monmap=0+}) v2
  997. 2015-01-19 19:50:54.392312 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  998. 2015-01-19 19:50:54.392321 7f28d9c71700 10 mon.--public-addr@0(leader) e1 check_sub monmap next 0 have 1
  999. 2015-01-19 19:50:54.392337 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011174 -- mon_map v1 -- ?+0 0x3cc0f00 con 0x3d02680
  1000. 2015-01-19 19:50:54.392347 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_map v1 remote, 127.0.0.1:0/1011174, have pipe.
  1001. 2015-01-19 19:50:54.392365 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1002. 2015-01-19 19:50:54.392366 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.? 127.0.0.1:0/1011174 -- mon_subscribe_ack(300s) v1 -- ?+0 0x3e92140
  1003. 2015-01-19 19:50:54.392378 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 3 features 52776558133247 0x3cc0f00 mon_map v1
  1004. 2015-01-19 19:50:54.392395 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 3): sig = 0
  1005. 2015-01-19 19:50:54.392407 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 3 0x3cc0f00
  1006. 2015-01-19 19:50:54.392414 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got MSG
  1007. 2015-01-19 19:50:54.392413 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_subscribe_ack(300s) v1 remote, 127.0.0.1:0/1011174, have pipe.
  1008. 2015-01-19 19:50:54.392433 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 23 to dispatch throttler 23/104857600
  1009. 2015-01-19 19:50:54.392443 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e90a80
  1010. 2015-01-19 19:50:54.392450 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1011. 2015-01-19 19:50:54.392449 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got envelope type=15 src client.4115 front=42 data=0 off 0
  1012. 2015-01-19 19:50:54.392465 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 4 features 52776558133247 0x3e92140 mon_subscribe_ack(300s) v1
  1013. 2015-01-19 19:50:54.392468 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 42 bytes from policy throttler 0/104857600
  1014. 2015-01-19 19:50:54.392480 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 4): sig = 0
  1015. 2015-01-19 19:50:54.392481 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 42 from dispatch throttler 0/104857600
  1016. 2015-01-19 19:50:54.392487 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 4 0x3e92140
  1017. 2015-01-19 19:50:54.392514 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1018. 2015-01-19 19:50:54.392519 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got front 42
  1019. 2015-01-19 19:50:54.392527 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  1020. 2015-01-19 19:50:54.392536 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).aborted = 0
  1021. 2015-01-19 19:50:54.392546 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got 42 + 0 + 0 byte message
  1022. 2015-01-19 19:50:54.392579 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got message 3 0x3e91a40 mon_subscribe({monmap=2+,osdmap=0}) v2
  1023. 2015-01-19 19:50:54.392602 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3e91a40 prio 127
  1024. 2015-01-19 19:50:54.392640 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1025. 2015-01-19 19:50:54.392636 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  1026. 2015-01-19 19:50:54.392651 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_ack 3
  1027. 2015-01-19 19:50:54.392645 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4115 127.0.0.1:0/1011174 3 ==== mon_subscribe({monmap=2+,osdmap=0}) v2 ==== 42+0+0 (2069260600 0 0) 0x3e91a40 con 0x3d02680
  1028. 2015-01-19 19:50:54.392656 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got MSG
  1029. 2015-01-19 19:50:54.392664 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  1030. 2015-01-19 19:50:54.392661 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1031. 2015-01-19 19:50:54.392668 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  1032. 2015-01-19 19:50:54.392666 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011174 is openallow * for client.? 127.0.0.1:0/1011174
  1033. 2015-01-19 19:50:54.392677 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  1034. 2015-01-19 19:50:54.392669 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got envelope type=15 src client.4115 front=42 data=0 off 0
  1035. 2015-01-19 19:50:54.392680 7f28d9c71700 10 mon.--public-addr@0(leader) e1 handle_subscribe mon_subscribe({monmap=2+,osdmap=0}) v2
  1036. 2015-01-19 19:50:54.392688 7f28d9c71700 10 mon.--public-addr@0(leader) e1 check_sub monmap next 2 have 1
  1037. 2015-01-19 19:50:54.392685 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 42 bytes from policy throttler 42/104857600
  1038. 2015-01-19 19:50:54.392696 7f28d9c71700 20 is_capable service=osd command= read on cap allow *
  1039. 2015-01-19 19:50:54.392700 7f28d9c71700 20 allow so far , doing grant allow *
  1040. 2015-01-19 19:50:54.392702 7f28d9c71700 20 allow all
  1041. 2015-01-19 19:50:54.392697 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 42 from dispatch throttler 42/104857600
  1042. 2015-01-19 19:50:54.392706 7f28d9c71700 10 mon.--public-addr@0(leader).osd e10 check_sub 0x3c443c0 next 0 (onetime)
  1043. 2015-01-19 19:50:54.392716 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got front 42
  1044. 2015-01-19 19:50:54.392727 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).aborted = 0
  1045. 2015-01-19 19:50:54.392733 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got 42 + 0 + 0 byte message
  1046. 2015-01-19 19:50:54.392754 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got message 4 0x3e92680 mon_subscribe({monmap=2+,osdmap=0}) v2
  1047. 2015-01-19 19:50:54.392765 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3e92680 prio 127
  1048. 2015-01-19 19:50:54.392777 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  1049. 2015-01-19 19:50:54.392780 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1050. 2015-01-19 19:50:54.392789 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_ack 4
  1051. 2015-01-19 19:50:54.392796 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1052. 2015-01-19 19:50:54.392803 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  1053. 2015-01-19 19:50:54.392806 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.? 127.0.0.1:0/1011174 -- osd_map(10..10 src has 1..10) v3 -- ?+0 0x3cdca40
  1054. 2015-01-19 19:50:54.392820 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message osd_map(10..10 src has 1..10) v3 remote, 127.0.0.1:0/1011174, have pipe.
  1055. 2015-01-19 19:50:54.392838 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1056. 2015-01-19 19:50:54.392841 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.4115 127.0.0.1:0/1011174 -- mon_subscribe_ack(300s) v1 -- ?+0 0x3e90a80
  1057. 2015-01-19 19:50:54.392848 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 5 features 52776558133247 0x3cdca40 osd_map(10..10 src has 1..10) v3
  1058. 2015-01-19 19:50:54.392873 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 5): sig = 0
  1059. 2015-01-19 19:50:54.392884 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 5 0x3cdca40
  1060. 2015-01-19 19:50:54.392889 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_subscribe_ack(300s) v1 remote, 127.0.0.1:0/1011174, have pipe.
  1061. 2015-01-19 19:50:54.392901 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 42 to dispatch throttler 84/104857600
  1062. 2015-01-19 19:50:54.392906 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e91a40
  1063. 2015-01-19 19:50:54.392918 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4115 127.0.0.1:0/1011174 4 ==== mon_subscribe({monmap=2+,osdmap=0}) v2 ==== 42+0+0 (2069260600 0 0) 0x3e92680 con 0x3d02680
  1064. 2015-01-19 19:50:54.392929 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  1065. 2015-01-19 19:50:54.392932 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011174 is openallow * for client.? 127.0.0.1:0/1011174
  1066. 2015-01-19 19:50:54.392929 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1067. 2015-01-19 19:50:54.392938 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  1068. 2015-01-19 19:50:54.392941 7f28d9c71700 10 mon.--public-addr@0(leader) e1 handle_subscribe mon_subscribe({monmap=2+,osdmap=0}) v2
  1069. 2015-01-19 19:50:54.392947 7f28d9c71700 10 mon.--public-addr@0(leader) e1 check_sub monmap next 2 have 1
  1070. 2015-01-19 19:50:54.392952 7f28d9c71700 20 is_capable service=osd command= read on cap allow *
  1071. 2015-01-19 19:50:54.392955 7f28d9c71700 20 allow so far , doing grant allow *
  1072. 2015-01-19 19:50:54.392956 7f28d9c71700 20 allow all
  1073. 2015-01-19 19:50:54.392958 7f28d9c71700 10 mon.--public-addr@0(leader).osd e10 check_sub 0x3c44d00 next 0 (onetime)
  1074. 2015-01-19 19:50:54.392942 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 6 features 52776558133247 0x3e90a80 mon_subscribe_ack(300s) v1
  1075. 2015-01-19 19:50:54.392986 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 6): sig = 0
  1076. 2015-01-19 19:50:54.392994 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 6 0x3e90a80
  1077. 2015-01-19 19:50:54.393017 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1078. 2015-01-19 19:50:54.393025 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  1079. 2015-01-19 19:50:54.393022 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.? 127.0.0.1:0/1011174 -- osd_map(10..10 src has 1..10) v3 -- ?+0 0x3cd8900
  1080. 2015-01-19 19:50:54.393034 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message osd_map(10..10 src has 1..10) v3 remote, 127.0.0.1:0/1011174, have pipe.
  1081. 2015-01-19 19:50:54.393052 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1082. 2015-01-19 19:50:54.393054 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> client.4115 127.0.0.1:0/1011174 -- mon_subscribe_ack(300s) v1 -- ?+0 0x3e91a40
  1083. 2015-01-19 19:50:54.393062 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 7 features 52776558133247 0x3cd8900 osd_map(10..10 src has 1..10) v3
  1084. 2015-01-19 19:50:54.393083 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 7): sig = 0
  1085. 2015-01-19 19:50:54.393095 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 7 0x3cd8900
  1086. 2015-01-19 19:50:54.393099 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_subscribe_ack(300s) v1 remote, 127.0.0.1:0/1011174, have pipe.
  1087. 2015-01-19 19:50:54.393112 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 42 to dispatch throttler 42/104857600
  1088. 2015-01-19 19:50:54.393117 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3e92680
  1089. 2015-01-19 19:50:54.393130 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1090. 2015-01-19 19:50:54.393141 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 8 features 52776558133247 0x3e91a40 mon_subscribe_ack(300s) v1
  1091. 2015-01-19 19:50:54.393154 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 8): sig = 0
  1092. 2015-01-19 19:50:54.393162 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 8 0x3e91a40
  1093. 2015-01-19 19:50:54.393181 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1094. 2015-01-19 19:50:54.393189 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  1095. 2015-01-19 19:50:54.396483 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
  1096. 2015-01-19 19:50:54.396509 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 3
  1097. 2015-01-19 19:50:54.396517 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  1098. 2015-01-19 19:50:54.396526 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
  1099. 2015-01-19 19:50:54.396534 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 4
  1100. 2015-01-19 19:50:54.396540 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  1101. 2015-01-19 19:50:54.396547 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
  1102. 2015-01-19 19:50:54.396553 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 5
  1103. 2015-01-19 19:50:54.396559 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  1104. 2015-01-19 19:50:54.396566 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
  1105. 2015-01-19 19:50:54.396572 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 6
  1106. 2015-01-19 19:50:54.396579 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  1107. 2015-01-19 19:50:54.396585 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
  1108. 2015-01-19 19:50:54.396592 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 7
  1109. 2015-01-19 19:50:54.396597 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  1110. 2015-01-19 19:50:54.396604 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
  1111. 2015-01-19 19:50:54.396611 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 8
  1112. 2015-01-19 19:50:54.396626 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  1113. 2015-01-19 19:50:54.396634 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got MSG
  1114. 2015-01-19 19:50:54.396642 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got envelope type=50 src client.4115 front=80 data=0 off 0
  1115. 2015-01-19 19:50:54.396657 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 80 bytes from policy throttler 0/104857600
  1116. 2015-01-19 19:50:54.396665 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 80 from dispatch throttler 0/104857600
  1117. 2015-01-19 19:50:54.396676 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got front 80
  1118. 2015-01-19 19:50:54.396685 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).aborted = 0
  1119. 2015-01-19 19:50:54.396691 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got 80 + 0 + 0 byte message
  1120. 2015-01-19 19:50:54.396715 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got message 5 0x3cc1680 mon_command({"prefix": "get_command_descriptions"} v 0) v1
  1121. 2015-01-19 19:50:54.396726 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3cc1680 prio 127
  1122. 2015-01-19 19:50:54.396740 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  1123. 2015-01-19 19:50:54.396745 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1124. 2015-01-19 19:50:54.396753 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4115 127.0.0.1:0/1011174 5 ==== mon_command({"prefix": "get_command_descriptions"} v 0) v1 ==== 80+0+0 (450363176 0 0) 0x3cc1680 con 0x3d02680
  1125. 2015-01-19 19:50:54.396772 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  1126. 2015-01-19 19:50:54.396770 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_ack 5
  1127. 2015-01-19 19:50:54.396775 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011174 is openallow * for client.? 127.0.0.1:0/1011174
  1128. 2015-01-19 19:50:54.396781 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  1129. 2015-01-19 19:50:54.396784 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1130. 2015-01-19 19:50:54.396790 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  1131. 2015-01-19 19:50:54.399684 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011174 -- mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1 -- ?+30001 0x3cc1c20 con 0x3d02680
  1132. 2015-01-19 19:50:54.399702 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1 remote, 127.0.0.1:0/1011174, have pipe.
  1133. 2015-01-19 19:50:54.399719 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 80 to dispatch throttler 80/104857600
  1134. 2015-01-19 19:50:54.399726 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3cc1680
  1135. 2015-01-19 19:50:54.399743 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1136. 2015-01-19 19:50:54.399772 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 9 features 52776558133247 0x3cc1c20 mon_command_ack([{"prefix": "get_command_descriptions"}]=0 v0) v1
  1137. 2015-01-19 19:50:54.399800 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 9): sig = 0
  1138. 2015-01-19 19:50:54.399813 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 9 0x3cc1c20
  1139. 2015-01-19 19:50:54.399868 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1140. 2015-01-19 19:50:54.399882 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  1141. 2015-01-19 19:50:54.500019 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ACK
  1142. 2015-01-19 19:50:54.500045 7f28e027f700 15 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got ack seq 9
  1143. 2015-01-19 19:50:54.500053 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  1144. 2015-01-19 19:50:54.500061 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got MSG
  1145. 2015-01-19 19:50:54.500069 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got envelope type=50 src client.4115 front=104 data=0 off 0
  1146. 2015-01-19 19:50:54.500078 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 104 bytes from policy throttler 0/104857600
  1147. 2015-01-19 19:50:54.500086 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader wants 104 from dispatch throttler 0/104857600
  1148. 2015-01-19 19:50:54.500097 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got front 104
  1149. 2015-01-19 19:50:54.500105 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).aborted = 0
  1150. 2015-01-19 19:50:54.500111 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got 104 + 0 + 0 byte message
  1151. 2015-01-19 19:50:54.500134 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader got message 6 0x3cc1860 mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1
  1152. 2015-01-19 19:50:54.500146 7f28e027f700 20 -- 127.0.0.1:6789/0 queue 0x3cc1860 prio 127
  1153. 2015-01-19 19:50:54.500162 7f28e027f700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader reading tag...
  1154. 2015-01-19 19:50:54.500165 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1155. 2015-01-19 19:50:54.500175 7f28d9c71700 1 -- 127.0.0.1:6789/0 <== client.4115 127.0.0.1:0/1011174 6 ==== mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 ==== 104+0+0 (4242270301 0 0) 0x3cc1860 con 0x3d02680
  1156. 2015-01-19 19:50:54.500189 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).write_ack 6
  1157. 2015-01-19 19:50:54.500197 7f28d9c71700 20 mon.--public-addr@0(leader) e1 have connection
  1158. 2015-01-19 19:50:54.500200 7f28d9c71700 20 mon.--public-addr@0(leader) e1 ms_dispatch existing session MonSession: client.? 127.0.0.1:0/1011174 is openallow * for client.? 127.0.0.1:0/1011174
  1159. 2015-01-19 19:50:54.500207 7f28d9c71700 20 mon.--public-addr@0(leader) e1 caps allow *
  1160. 2015-01-19 19:50:54.500203 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1161. 2015-01-19 19:50:54.500211 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  1162. 2015-01-19 19:50:54.500337 7f28d9c71700 0 mon.--public-addr@0(leader) e1 handle_command mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1
  1163. 2015-01-19 19:50:54.500388 7f28d9c71700 20 is_capable service=osd command=osd erasure-code-profile rm read write on cap allow *
  1164. 2015-01-19 19:50:54.500394 7f28d9c71700 20 allow so far , doing grant allow *
  1165. 2015-01-19 19:50:54.500396 7f28d9c71700 20 allow all
  1166. 2015-01-19 19:50:54.500398 7f28d9c71700 10 mon.--public-addr@0(leader) e1 _allowed_command capable
  1167. 2015-01-19 19:50:54.500401 7f28d9c71700 10 mon.--public-addr@0(leader).paxosservice(osdmap 1..10) dispatch mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 from client.4115 127.0.0.1:0/1011174
  1168. 2015-01-19 19:50:54.500411 7f28d9c71700 5 mon.--public-addr@0(leader).paxos(paxos active c 1..37) is_readable now=2015-01-19 19:50:54.500411 lease_expire=0.000000 has v0 lc 37
  1169. 2015-01-19 19:50:54.500419 7f28d9c71700 10 mon.--public-addr@0(leader).osd e10 preprocess_query mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 from client.4115 127.0.0.1:0/1011174
  1170. 2015-01-19 19:50:54.500553 7f28d9c71700 7 mon.--public-addr@0(leader).osd e10 prepare_update mon_command({"prefix": "osd erasure-code-profile rm", "name": "myprofile"} v 0) v1 from client.4115 127.0.0.1:0/1011174
  1171. 2015-01-19 19:50:54.500716 7f28d9c71700 1 -- 127.0.0.1:6789/0 --> 127.0.0.1:0/1011174 -- mon_command_ack([{"prefix": "osd erasure-code-profile rm", "name": "myprofile"}]=0 erasure-code-profile myprofile does not exist v10) v1 -- ?+0 0x3cc1680 con 0x3d02680
  1172. 2015-01-19 19:50:54.500733 7f28d9c71700 20 -- 127.0.0.1:6789/0 submit_message mon_command_ack([{"prefix": "osd erasure-code-profile rm", "name": "myprofile"}]=0 erasure-code-profile myprofile does not exist v10) v1 remote, 127.0.0.1:0/1011174, have pipe.
  1173. 2015-01-19 19:50:54.500749 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1174. 2015-01-19 19:50:54.500753 7f28d9c71700 10 -- 127.0.0.1:6789/0 dispatch_throttle_release 104 to dispatch throttler 104/104857600
  1175. 2015-01-19 19:50:54.500760 7f28d9c71700 20 -- 127.0.0.1:6789/0 done calling dispatch on 0x3cc1860
  1176. 2015-01-19 19:50:54.500760 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer encoding 10 features 52776558133247 0x3cc1680 mon_command_ack([{"prefix": "osd erasure-code-profile rm", "name": "myprofile"}]=0 erasure-code-profile myprofile does not exist v10) v1
  1177. 2015-01-19 19:50:54.500783 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer signed seq # 10): sig = 0
  1178. 2015-01-19 19:50:54.500798 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sending 10 0x3cc1680
  1179. 2015-01-19 19:50:54.500843 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer: state = open policy.server=1
  1180. 2015-01-19 19:50:54.500856 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).writer sleeping
  1181. 2015-01-19 19:50:54.504876 7f28e027f700 2 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).reader couldn't read tag, (0) Success
  1182. 2015-01-19 19:50:54.504943 7f28e027f700 2 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).fault (0) Success
  1183. 2015-01-19 19:50:54.505046 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).fault on lossy channel, failing
  1184. 2015-01-19 19:50:54.505065 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=2 pgs=1 cs=1 l=1 c=0x3d02680).stop
  1185. 2015-01-19 19:50:54.505085 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).unregister_pipe
  1186. 2015-01-19 19:50:54.505102 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).discard_queue
  1187. 2015-01-19 19:50:54.505126 7f28e027f700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).reader done
  1188. 2015-01-19 19:50:54.505130 7f28d846e700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).writer finishing
  1189. 2015-01-19 19:50:54.505150 7f28d846e700 10 -- 127.0.0.1:6789/0 queue_reap 0x3d71180
  1190. 2015-01-19 19:50:54.505145 7f28d9c71700 10 mon.--public-addr@0(leader) e1 ms_handle_reset 0x3d02680 127.0.0.1:0/1011174
  1191. 2015-01-19 19:50:54.505160 7f28d9c71700 10 mon.--public-addr@0(leader) e1 reset/close on session client.? 127.0.0.1:0/1011174
  1192. 2015-01-19 19:50:54.505160 7f28d846e700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).writer done
  1193. 2015-01-19 19:50:54.505166 7f28d9c71700 10 mon.--public-addr@0(leader) e1 remove_session MonSession: client.? 127.0.0.1:0/1011174 is openallow * client.? 127.0.0.1:0/1011174
  1194. 2015-01-19 19:50:54.505171 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper
  1195. 2015-01-19 19:50:54.505194 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper reaping pipe 0x3d71180 127.0.0.1:0/1011174
  1196. 2015-01-19 19:50:54.505208 7f28dac73700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).discard_queue
  1197. 2015-01-19 19:50:54.505239 7f28dac73700 10 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).unregister_pipe - not registered
  1198. 2015-01-19 19:50:54.505254 7f28dac73700 20 -- 127.0.0.1:6789/0 >> 127.0.0.1:0/1011174 pipe(0x3d71180 sd=21 :6789 s=4 pgs=1 cs=1 l=1 c=0x3d02680).join
  1199. 2015-01-19 19:50:54.505287 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper reaped pipe 0x3d71180 127.0.0.1:0/1011174
  1200. 2015-01-19 19:50:54.505305 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper deleted pipe 0x3d71180
  1201. 2015-01-19 19:50:54.505312 7f28dac73700 10 -- 127.0.0.1:6789/0 reaper done
  1202. call_TEST_functions: 82: return 1
  1203. run: 27: return 1
  1204. main: 110: code=1
  1205. main: 112: teardown osd-erasure-code-profile
  1206. teardown: 24: local dir=osd-erasure-code-profile
  1207. teardown: 25: kill_daemons osd-erasure-code-profile
  1208. kill_daemons: 60: local dir=osd-erasure-code-profile
  1209. kkill_daemons: 59: find osd-erasure-code-profile
  1210. kkill_daemons: 59: grep pidfile
  1211. kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
  1212. kkill_daemons: 62: cat osd-erasure-code-profile/--public-addr/pidfile
  1213. kill_daemons: 62: pid=10528
  1214. kill_daemons: 63: for try in 0 1 1 1 2 3
  1215. kill_daemons: 64: kill -9 10528
  1216. kill_daemons: 65: sleep 0
  1217. kill_daemons: 63: for try in 0 1 1 1 2 3
  1218. kill_daemons: 64: kill -9 10528
  1219. kill_daemons: 65: sleep 1
  1220. kill_daemons: 63: for try in 0 1 1 1 2 3
  1221. kill_daemons: 64: kill -9 10528
  1222. kill_daemons: 64: break
  1223. teardown: 26: rm -fr osd-erasure-code-profile
  1224. main: 113: return 1
  1225. FAIL: ./src/test/mon/osd-pool-create.sh.log
  1226. main: 105: setup osd-pool-create
  1227. setup: 18: local dir=osd-pool-create
  1228. setup: 19: teardown osd-pool-create
  1229. teardown: 24: local dir=osd-pool-create
  1230. teardown: 25: kill_daemons osd-pool-create
  1231. kill_daemons: 60: local dir=osd-pool-create
  1232. kkill_daemons: 59: find osd-pool-create
  1233. kkill_daemons: 59: grep pidfile
  1234. find: 'osd-pool-create': No such file or directory
  1235. teardown: 26: rm -fr osd-pool-create
  1236. setup: 20: mkdir osd-pool-create
  1237. main: 106: local code
  1238. main: 107: run osd-pool-create
  1239. run: 20: local dir=osd-pool-create
  1240. run: 22: export CEPH_ARGS
  1241. rrun: 23: uuidgen
  1242. run: 23: CEPH_ARGS+='--fsid=b22579e5-88d7-496a-9f15-a6b0c601bb8f --auth-supported=none '
  1243. run: 24: CEPH_ARGS+='--mon-host=127.0.0.1 '
  1244. rrun: 26: set
  1245. rrun: 26: sed -n -e 's/^\(TEST_[0-9a-z_]*\) .*/\1/p'
  1246. run: 26: FUNCTIONS='TEST_default_deprectated_0
  1247. TEST_default_deprectated_1
  1248. TEST_default_deprectated_2
  1249. TEST_erasure_code_pool
  1250. TEST_erasure_code_profile_default
  1251. TEST_erasure_code_profile_default_pending
  1252. TEST_erasure_crush_rule
  1253. TEST_erasure_crush_rule_pending
  1254. TEST_erasure_crush_stripe_width
  1255. TEST_erasure_crush_stripe_width_padded
  1256. TEST_erasure_invalid_profile
  1257. TEST_replicated_pool
  1258. TEST_replicated_pool_with_ruleset
  1259. TEST_simple_crush_rule_pending'
  1260. run: 27: for TEST_function in '$FUNCTIONS'
  1261. run: 28: setup osd-pool-create
  1262. setup: 18: local dir=osd-pool-create
  1263. setup: 19: teardown osd-pool-create
  1264. teardown: 24: local dir=osd-pool-create
  1265. teardown: 25: kill_daemons osd-pool-create
  1266. kill_daemons: 60: local dir=osd-pool-create
  1267. kkill_daemons: 59: find osd-pool-create
  1268. kkill_daemons: 59: grep pidfile
  1269. teardown: 26: rm -fr osd-pool-create
  1270. setup: 20: mkdir osd-pool-create
  1271. run: 29: TEST_default_deprectated_0 osd-pool-create
  1272. TEST_default_deprectated_0: 35: local dir=osd-pool-create
  1273. TEST_default_deprectated_0: 37: expected=66
  1274. TEST_default_deprectated_0: 38: run_mon osd-pool-create a --public-addr 127.0.0.1 --osd_pool_default_crush_replicated_ruleset 66
  1275. run_mon: 30: local dir=osd-pool-create
  1276. run_mon: 31: shift
  1277. run_mon: 32: local id=a
  1278. run_mon: 33: shift
  1279. run_mon: 34: dir+=/a
  1280. run_mon: 37: ./ceph-mon --id a --mkfs --mon-data=osd-pool-create/a --run-dir=osd-pool-create/a --public-addr 127.0.0.1 --osd_pool_default_crush_replicated_ruleset 66
  1281. ./ceph-mon: renaming mon.noname-a 127.0.0.1:6789/0 to mon.a
  1282. ./ceph-mon: set fsid to b22579e5-88d7-496a-9f15-a6b0c601bb8f
  1283. ./ceph-mon: created monfs at osd-pool-create/a for mon.a
  1284. run_mon: 43: ./ceph-mon --id a --paxos-propose-interval=0.1 --osd-pool-default-erasure-code-directory=.libs --debug-mon 20 --debug-ms 20 --debug-paxos 20 --mon-advanced-debug-mode --chdir= --mon-data=osd-pool-create/a --log-file=osd-pool-create/a/log --mon-cluster-log-file=osd-pool-create/a/log --run-dir=osd-pool-create/a --pid-file=osd-pool-create/a/pidfile --public-addr 127.0.0.1 --osd_pool_default_crush_replicated_ruleset 66
  1285. TEST_default_deprectated_0: 40: ./ceph --format json osd dump
  1286. TEST_default_deprectated_0: 40: grep '"crush_ruleset":66'
  1287. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1288. {"epoch":1,"fsid":"b22579e5-88d7-496a-9f15-a6b0c601bb8f","created":"2015-01-19 19:45:56.846784","modified":"2015-01-19 19:45:56.846784","flags":"","cluster_snapshot":"","pool_max":2,"max_osd":0,"pools":[{"pool":0,"pool_name":"data","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":66,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":45,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":1,"pool_name":"metadata","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":66,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":2,"pool_name":"rbd","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":66,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0}],"osds":[],"osd_xinfo":[],"pg_temp":[],"primary_temp":[],"blacklist":[],"erasure_code_profiles":{"default":{"directory":".libs","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}}}
  1289. TEST_default_deprectated_0: 41: grep 'osd_pool_default_crush_rule is deprecated ' osd-pool-create/a/log
  1290. run: 30: teardown osd-pool-create
  1291. teardown: 24: local dir=osd-pool-create
  1292. teardown: 25: kill_daemons osd-pool-create
  1293. kill_daemons: 60: local dir=osd-pool-create
  1294. kkill_daemons: 59: find osd-pool-create
  1295. kkill_daemons: 59: grep pidfile
  1296. kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
  1297. kkill_daemons: 62: cat osd-pool-create/a/pidfile
  1298. kill_daemons: 62: pid=13234
  1299. kill_daemons: 63: for try in 0 1 1 1 2 3
  1300. kill_daemons: 64: kill -9 13234
  1301. kill_daemons: 65: sleep 0
  1302. kill_daemons: 63: for try in 0 1 1 1 2 3
  1303. kill_daemons: 64: kill -9 13234
  1304. kill_daemons: 65: sleep 1
  1305. kill_daemons: 63: for try in 0 1 1 1 2 3
  1306. kill_daemons: 64: kill -9 13234
  1307. kill_daemons: 64: break
  1308. teardown: 26: rm -fr osd-pool-create
  1309. run: 27: for TEST_function in '$FUNCTIONS'
  1310. run: 28: setup osd-pool-create
  1311. setup: 18: local dir=osd-pool-create
  1312. setup: 19: teardown osd-pool-create
  1313. teardown: 24: local dir=osd-pool-create
  1314. teardown: 25: kill_daemons osd-pool-create
  1315. kill_daemons: 60: local dir=osd-pool-create
  1316. kkill_daemons: 59: find osd-pool-create
  1317. kkill_daemons: 59: grep pidfile
  1318. find: 'osd-pool-create': No such file or directory
  1319. teardown: 26: rm -fr osd-pool-create
  1320. setup: 20: mkdir osd-pool-create
  1321. run: 29: TEST_default_deprectated_1 osd-pool-create
  1322. TEST_default_deprectated_1: 45: local dir=osd-pool-create
  1323. TEST_default_deprectated_1: 47: expected=55
  1324. TEST_default_deprectated_1: 48: run_mon osd-pool-create a --public-addr 127.0.0.1 --osd_pool_default_crush_rule 55
  1325. run_mon: 30: local dir=osd-pool-create
  1326. run_mon: 31: shift
  1327. run_mon: 32: local id=a
  1328. run_mon: 33: shift
  1329. run_mon: 34: dir+=/a
  1330. run_mon: 37: ./ceph-mon --id a --mkfs --mon-data=osd-pool-create/a --run-dir=osd-pool-create/a --public-addr 127.0.0.1 --osd_pool_default_crush_rule 55
  1331. ./ceph-mon: renaming mon.noname-a 127.0.0.1:6789/0 to mon.a
  1332. ./ceph-mon: set fsid to b22579e5-88d7-496a-9f15-a6b0c601bb8f
  1333. ./ceph-mon: created monfs at osd-pool-create/a for mon.a
  1334. run_mon: 43: ./ceph-mon --id a --paxos-propose-interval=0.1 --osd-pool-default-erasure-code-directory=.libs --debug-mon 20 --debug-ms 20 --debug-paxos 20 --mon-advanced-debug-mode --chdir= --mon-data=osd-pool-create/a --log-file=osd-pool-create/a/log --mon-cluster-log-file=osd-pool-create/a/log --run-dir=osd-pool-create/a --pid-file=osd-pool-create/a/pidfile --public-addr 127.0.0.1 --osd_pool_default_crush_rule 55
  1335. TEST_default_deprectated_1: 50: ./ceph --format json osd dump
  1336. TEST_default_deprectated_1: 50: grep '"crush_ruleset":55'
  1337. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1338. {"epoch":1,"fsid":"b22579e5-88d7-496a-9f15-a6b0c601bb8f","created":"2015-01-19 19:45:58.235083","modified":"2015-01-19 19:45:58.235083","flags":"","cluster_snapshot":"","pool_max":2,"max_osd":0,"pools":[{"pool":0,"pool_name":"data","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":55,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":45,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":1,"pool_name":"metadata","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":55,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":2,"pool_name":"rbd","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":55,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0}],"osds":[],"osd_xinfo":[],"pg_temp":[],"primary_temp":[],"blacklist":[],"erasure_code_profiles":{"default":{"directory":".libs","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}}}
  1339. TEST_default_deprectated_1: 51: grep 'osd_pool_default_crush_rule is deprecated ' osd-pool-create/a/log
  1340. 2015-01-19 19:45:58.234604 7f8575a797c0 0 osd_pool_default_crush_rule is deprecated use osd_pool_default_crush_replicated_ruleset instead
  1341. run: 30: teardown osd-pool-create
  1342. teardown: 24: local dir=osd-pool-create
  1343. teardown: 25: kill_daemons osd-pool-create
  1344. kill_daemons: 60: local dir=osd-pool-create
  1345. kkill_daemons: 59: find osd-pool-create
  1346. kkill_daemons: 59: grep pidfile
  1347. kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
  1348. kkill_daemons: 62: cat osd-pool-create/a/pidfile
  1349. kill_daemons: 62: pid=13296
  1350. kill_daemons: 63: for try in 0 1 1 1 2 3
  1351. kill_daemons: 64: kill -9 13296
  1352. kill_daemons: 65: sleep 0
  1353. kill_daemons: 63: for try in 0 1 1 1 2 3
  1354. kill_daemons: 64: kill -9 13296
  1355. kill_daemons: 65: sleep 1
  1356. kill_daemons: 63: for try in 0 1 1 1 2 3
  1357. kill_daemons: 64: kill -9 13296
  1358. kill_daemons: 64: break
  1359. teardown: 26: rm -fr osd-pool-create
  1360. run: 27: for TEST_function in '$FUNCTIONS'
  1361. run: 28: setup osd-pool-create
  1362. setup: 18: local dir=osd-pool-create
  1363. setup: 19: teardown osd-pool-create
  1364. teardown: 24: local dir=osd-pool-create
  1365. teardown: 25: kill_daemons osd-pool-create
  1366. kill_daemons: 60: local dir=osd-pool-create
  1367. kkill_daemons: 59: find osd-pool-create
  1368. kkill_daemons: 59: grep pidfile
  1369. find: 'osd-pool-create': No such file or directory
  1370. teardown: 26: rm -fr osd-pool-create
  1371. setup: 20: mkdir osd-pool-create
  1372. run: 29: TEST_default_deprectated_2 osd-pool-create
  1373. TEST_default_deprectated_2: 55: local dir=osd-pool-create
  1374. TEST_default_deprectated_2: 56: expected=77
  1375. TEST_default_deprectated_2: 57: unexpected=33
  1376. TEST_default_deprectated_2: 58: run_mon osd-pool-create a --public-addr 127.0.0.1 --osd_pool_default_crush_rule 77 --osd_pool_default_crush_replicated_ruleset 33
  1377. run_mon: 30: local dir=osd-pool-create
  1378. run_mon: 31: shift
  1379. run_mon: 32: local id=a
  1380. run_mon: 33: shift
  1381. run_mon: 34: dir+=/a
  1382. run_mon: 37: ./ceph-mon --id a --mkfs --mon-data=osd-pool-create/a --run-dir=osd-pool-create/a --public-addr 127.0.0.1 --osd_pool_default_crush_rule 77 --osd_pool_default_crush_replicated_ruleset 33
  1383. ./ceph-mon: renaming mon.noname-a 127.0.0.1:6789/0 to mon.a
  1384. ./ceph-mon: set fsid to b22579e5-88d7-496a-9f15-a6b0c601bb8f
  1385. ./ceph-mon: created monfs at osd-pool-create/a for mon.a
  1386. run_mon: 43: ./ceph-mon --id a --paxos-propose-interval=0.1 --osd-pool-default-erasure-code-directory=.libs --debug-mon 20 --debug-ms 20 --debug-paxos 20 --mon-advanced-debug-mode --chdir= --mon-data=osd-pool-create/a --log-file=osd-pool-create/a/log --mon-cluster-log-file=osd-pool-create/a/log --run-dir=osd-pool-create/a --pid-file=osd-pool-create/a/pidfile --public-addr 127.0.0.1 --osd_pool_default_crush_rule 77 --osd_pool_default_crush_replicated_ruleset 33
  1387. TEST_default_deprectated_2: 61: ./ceph --format json osd dump
  1388. TEST_default_deprectated_2: 61: grep '"crush_ruleset":77'
  1389. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1390. {"epoch":1,"fsid":"b22579e5-88d7-496a-9f15-a6b0c601bb8f","created":"2015-01-19 19:45:59.624973","modified":"2015-01-19 19:45:59.624973","flags":"","cluster_snapshot":"","pool_max":2,"max_osd":0,"pools":[{"pool":0,"pool_name":"data","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":77,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":45,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":1,"pool_name":"metadata","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":77,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":2,"pool_name":"rbd","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":77,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0}],"osds":[],"osd_xinfo":[],"pg_temp":[],"primary_temp":[],"blacklist":[],"erasure_code_profiles":{"default":{"directory":".libs","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}}}
  1391. TEST_default_deprectated_2: 62: ./ceph --format json osd dump
  1392. TEST_default_deprectated_2: 62: grep '"crush_ruleset":33'
  1393. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1394. TEST_default_deprectated_2: 63: grep 'osd_pool_default_crush_rule is deprecated ' osd-pool-create/a/log
  1395. 2015-01-19 19:45:59.624390 7fe4b28d57c0 0 osd_pool_default_crush_rule is deprecated use osd_pool_default_crush_replicated_ruleset instead
  1396. run: 30: teardown osd-pool-create
  1397. teardown: 24: local dir=osd-pool-create
  1398. teardown: 25: kill_daemons osd-pool-create
  1399. kill_daemons: 60: local dir=osd-pool-create
  1400. kkill_daemons: 59: find osd-pool-create
  1401. kkill_daemons: 59: grep pidfile
  1402. kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
  1403. kkill_daemons: 62: cat osd-pool-create/a/pidfile
  1404. kill_daemons: 62: pid=13358
  1405. kill_daemons: 63: for try in 0 1 1 1 2 3
  1406. kill_daemons: 64: kill -9 13358
  1407. kill_daemons: 65: sleep 0
  1408. kill_daemons: 63: for try in 0 1 1 1 2 3
  1409. kill_daemons: 64: kill -9 13358
  1410. kill_daemons: 65: sleep 1
  1411. kill_daemons: 63: for try in 0 1 1 1 2 3
  1412. kill_daemons: 64: kill -9 13358
  1413. kill_daemons: 64: break
  1414. teardown: 26: rm -fr osd-pool-create
  1415. run: 27: for TEST_function in '$FUNCTIONS'
  1416. run: 28: setup osd-pool-create
  1417. setup: 18: local dir=osd-pool-create
  1418. setup: 19: teardown osd-pool-create
  1419. teardown: 24: local dir=osd-pool-create
  1420. teardown: 25: kill_daemons osd-pool-create
  1421. kill_daemons: 60: local dir=osd-pool-create
  1422. kkill_daemons: 59: find osd-pool-create
  1423. kkill_daemons: 59: grep pidfile
  1424. find: 'osd-pool-create': No such file or directory
  1425. teardown: 26: rm -fr osd-pool-create
  1426. setup: 20: mkdir osd-pool-create
  1427. run: 29: TEST_erasure_code_pool osd-pool-create
  1428. TEST_erasure_code_pool: 179: local dir=osd-pool-create
  1429. TEST_erasure_code_pool: 180: run_mon osd-pool-create a --public-addr 127.0.0.1
  1430. run_mon: 30: local dir=osd-pool-create
  1431. run_mon: 31: shift
  1432. run_mon: 32: local id=a
  1433. run_mon: 33: shift
  1434. run_mon: 34: dir+=/a
  1435. run_mon: 37: ./ceph-mon --id a --mkfs --mon-data=osd-pool-create/a --run-dir=osd-pool-create/a --public-addr 127.0.0.1
  1436. ./ceph-mon: renaming mon.noname-a 127.0.0.1:6789/0 to mon.a
  1437. ./ceph-mon: set fsid to b22579e5-88d7-496a-9f15-a6b0c601bb8f
  1438. ./ceph-mon: created monfs at osd-pool-create/a for mon.a
  1439. run_mon: 43: ./ceph-mon --id a --paxos-propose-interval=0.1 --osd-pool-default-erasure-code-directory=.libs --debug-mon 20 --debug-ms 20 --debug-paxos 20 --mon-advanced-debug-mode --chdir= --mon-data=osd-pool-create/a --log-file=osd-pool-create/a/log --mon-cluster-log-file=osd-pool-create/a/log --run-dir=osd-pool-create/a --pid-file=osd-pool-create/a/pidfile --public-addr 127.0.0.1
  1440. TEST_erasure_code_pool: 181: ./ceph --format json osd dump
  1441. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1442. TEST_erasure_code_pool: 182: local 'expected="erasure_code_profile":"default"'
  1443. TEST_erasure_code_pool: 183: grep '"erasure_code_profile":"default"' osd-pool-create/osd.json
  1444. TEST_erasure_code_pool: 184: ./ceph osd pool create erasurecodes 12 12 erasure
  1445. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1446. pool 'erasurecodes' created
  1447. TEST_erasure_code_pool: 185: ./ceph --format json osd dump
  1448. TEST_erasure_code_pool: 185: tee osd-pool-create/osd.json
  1449. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1450. {"epoch":3,"fsid":"b22579e5-88d7-496a-9f15-a6b0c601bb8f","created":"2015-01-19 19:46:01.351748","modified":"2015-01-19 19:46:02.069213","flags":"","cluster_snapshot":"","pool_max":3,"max_osd":0,"pools":[{"pool":0,"pool_name":"data","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":0,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":45,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":1,"pool_name":"metadata","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":0,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":2,"pool_name":"rbd","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_ruleset":0,"object_hash":2,"pg_num":64,"pg_placement_num":64,"crash_replay_interval":0,"last_change":"1","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":0,"cache_target_full_ratio_micro":0,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":0},{"pool":3,"pool_name":"erasurecodes","flags":1,"flags_names":"hashpspool","type":3,"size":3,"min_size":2,"crush_ruleset":1,"object_hash":2,"pg_num":12,"pg_placement_num":12,"crash_replay_interval":0,"last_change":"3","last_force_op_resend":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"default","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"stripe_width":4096}],"osds":[],"osd_xinfo":[],"pg_temp":[],"primary_temp":[],"blacklist":[],"erasure_code_profiles":{"default":{"directory":".libs","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}}}
  1451. TEST_erasure_code_pool: 186: grep '"erasure_code_profile":"default"' osd-pool-create/osd.json
  1452. TEST_erasure_code_pool: 188: ./ceph osd pool create erasurecodes 12 12 erasure
  1453. TEST_erasure_code_pool: 189: grep 'already exists'
  1454. pool 'erasurecodes' already exists
  1455. TEST_erasure_code_pool: 190: ./ceph osd pool create erasurecodes 12 12
  1456. TEST_erasure_code_pool: 191: grep 'cannot change to type replicated'
  1457. Error EINVAL: pool 'erasurecodes' cannot change to type replicated
  1458. run: 30: teardown osd-pool-create
  1459. teardown: 24: local dir=osd-pool-create
  1460. teardown: 25: kill_daemons osd-pool-create
  1461. kill_daemons: 60: local dir=osd-pool-create
  1462. kkill_daemons: 59: find osd-pool-create
  1463. kkill_daemons: 59: grep pidfile
  1464. kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
  1465. kkill_daemons: 62: cat osd-pool-create/a/pidfile
  1466. kill_daemons: 62: pid=13456
  1467. kill_daemons: 63: for try in 0 1 1 1 2 3
  1468. kill_daemons: 64: kill -9 13456
  1469. kill_daemons: 65: sleep 0
  1470. kill_daemons: 63: for try in 0 1 1 1 2 3
  1471. kill_daemons: 64: kill -9 13456
  1472. kill_daemons: 65: sleep 1
  1473. kill_daemons: 63: for try in 0 1 1 1 2 3
  1474. kill_daemons: 64: kill -9 13456
  1475. kill_daemons: 64: break
  1476. teardown: 26: rm -fr osd-pool-create
  1477. run: 27: for TEST_function in '$FUNCTIONS'
  1478. run: 28: setup osd-pool-create
  1479. setup: 18: local dir=osd-pool-create
  1480. setup: 19: teardown osd-pool-create
  1481. teardown: 24: local dir=osd-pool-create
  1482. teardown: 25: kill_daemons osd-pool-create
  1483. kill_daemons: 60: local dir=osd-pool-create
  1484. kkill_daemons: 59: find osd-pool-create
  1485. kkill_daemons: 59: grep pidfile
  1486. find: 'osd-pool-create': No such file or directory
  1487. teardown: 26: rm -fr osd-pool-create
  1488. setup: 20: mkdir osd-pool-create
  1489. run: 29: TEST_erasure_code_profile_default osd-pool-create
  1490. TEST_erasure_code_profile_default: 127: local dir=osd-pool-create
  1491. TEST_erasure_code_profile_default: 128: run_mon osd-pool-create a --public-addr 127.0.0.1
  1492. run_mon: 30: local dir=osd-pool-create
  1493. run_mon: 31: shift
  1494. run_mon: 32: local id=a
  1495. run_mon: 33: shift
  1496. run_mon: 34: dir+=/a
  1497. run_mon: 37: ./ceph-mon --id a --mkfs --mon-data=osd-pool-create/a --run-dir=osd-pool-create/a --public-addr 127.0.0.1
  1498. ./ceph-mon: renaming mon.noname-a 127.0.0.1:6789/0 to mon.a
  1499. ./ceph-mon: set fsid to b22579e5-88d7-496a-9f15-a6b0c601bb8f
  1500. ./ceph-mon: created monfs at osd-pool-create/a for mon.a
  1501. run_mon: 43: ./ceph-mon --id a --paxos-propose-interval=0.1 --osd-pool-default-erasure-code-directory=.libs --debug-mon 20 --debug-ms 20 --debug-paxos 20 --mon-advanced-debug-mode --chdir= --mon-data=osd-pool-create/a --log-file=osd-pool-create/a/log --mon-cluster-log-file=osd-pool-create/a/log --run-dir=osd-pool-create/a --pid-file=osd-pool-create/a/pidfile --public-addr 127.0.0.1
  1502. TEST_erasure_code_profile_default: 129: ./ceph osd erasure-code-profile rm default
  1503. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1504. TEST_erasure_code_profile_default: 130: ./ceph osd erasure-code-profile ls
  1505. TEST_erasure_code_profile_default: 130: grep default
  1506. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1507. TEST_erasure_code_profile_default: 131: ./ceph osd pool create 12 12 erasure default
  1508. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1509. erasure not valid: erasure doesn't represent an int
  1510. pool '12' created
  1511. TEST_erasure_code_profile_default: 132: ./ceph osd erasure-code-profile ls
  1512. TEST_erasure_code_profile_default: 132: grep default
  1513. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1514. default
  1515. run: 30: teardown osd-pool-create
  1516. teardown: 24: local dir=osd-pool-create
  1517. teardown: 25: kill_daemons osd-pool-create
  1518. kill_daemons: 60: local dir=osd-pool-create
  1519. kkill_daemons: 59: find osd-pool-create
  1520. kkill_daemons: 59: grep pidfile
  1521. kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
  1522. kkill_daemons: 62: cat osd-pool-create/a/pidfile
  1523. kill_daemons: 62: pid=13661
  1524. kill_daemons: 63: for try in 0 1 1 1 2 3
  1525. kill_daemons: 64: kill -9 13661
  1526. kill_daemons: 65: sleep 0
  1527. kill_daemons: 63: for try in 0 1 1 1 2 3
  1528. kill_daemons: 64: kill -9 13661
  1529. kill_daemons: 65: sleep 1
  1530. kill_daemons: 63: for try in 0 1 1 1 2 3
  1531. kill_daemons: 64: kill -9 13661
  1532. kill_daemons: 64: break
  1533. teardown: 26: rm -fr osd-pool-create
  1534. run: 27: for TEST_function in '$FUNCTIONS'
  1535. run: 28: setup osd-pool-create
  1536. setup: 18: local dir=osd-pool-create
  1537. setup: 19: teardown osd-pool-create
  1538. teardown: 24: local dir=osd-pool-create
  1539. teardown: 25: kill_daemons osd-pool-create
  1540. kill_daemons: 60: local dir=osd-pool-create
  1541. kkill_daemons: 59: find osd-pool-create
  1542. kkill_daemons: 59: grep pidfile
  1543. find: 'osd-pool-create': No such file or directory
  1544. teardown: 26: rm -fr osd-pool-create
  1545. setup: 20: mkdir osd-pool-create
  1546. run: 29: TEST_erasure_code_profile_default_pending osd-pool-create
  1547. TEST_erasure_code_profile_default_pending: 136: local dir=osd-pool-create
  1548. TEST_erasure_code_profile_default_pending: 137: run_mon osd-pool-create a --public-addr 127.0.0.1
  1549. run_mon: 30: local dir=osd-pool-create
  1550. run_mon: 31: shift
  1551. run_mon: 32: local id=a
  1552. run_mon: 33: shift
  1553. run_mon: 34: dir+=/a
  1554. run_mon: 37: ./ceph-mon --id a --mkfs --mon-data=osd-pool-create/a --run-dir=osd-pool-create/a --public-addr 127.0.0.1
  1555. ./ceph-mon: renaming mon.noname-a 127.0.0.1:6789/0 to mon.a
  1556. ./ceph-mon: set fsid to b22579e5-88d7-496a-9f15-a6b0c601bb8f
  1557. ./ceph-mon: created monfs at osd-pool-create/a for mon.a
  1558. run_mon: 43: ./ceph-mon --id a --paxos-propose-interval=0.1 --osd-pool-default-erasure-code-directory=.libs --debug-mon 20 --debug-ms 20 --debug-paxos 20 --mon-advanced-debug-mode --chdir= --mon-data=osd-pool-create/a --log-file=osd-pool-create/a/log --mon-cluster-log-file=osd-pool-create/a/log --run-dir=osd-pool-create/a --pid-file=osd-pool-create/a/pidfile --public-addr 127.0.0.1
  1559. TEST_erasure_code_profile_default_pending: 138: ./ceph osd erasure-code-profile rm default
  1560. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1561. TEST_erasure_code_profile_default_pending: 139: ./ceph osd erasure-code-profile ls
  1562. TEST_erasure_code_profile_default_pending: 139: grep default
  1563. *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH ***
  1564. TTEST_erasure_code_profile_default_pending: 141: echo '{"prefix":"osdmonitor_prepare_command","prepare":"osd erasure-code-profile set","name":"default"}'
  1565. TTEST_erasure_code_profile_default_pending: 141: nc -U osd-pool-create/a/ceph-mon.a.asok
  1566. ./test/mon/osd-pool-create.sh: line 141: nc: command not found
  1567. TTEST_erasure_code_profile_default_pending: 141: cut --bytes=5-
  1568. TEST_erasure_code_profile_default_pending: 141: result=
  1569. TEST_erasure_code_profile_default_pending: 142: test = true
  1570. ./test/mon/osd-pool-create.sh: line 142: test: =: unary operator expected
  1571. TEST_erasure_code_profile_default_pending: 142: return 1
  1572. run: 29: return 1
  1573. main: 110: code=1
  1574. main: 112: teardown osd-pool-create
  1575. teardown: 24: local dir=osd-pool-create
  1576. teardown: 25: kill_daemons osd-pool-create
  1577. kill_daemons: 60: local dir=osd-pool-create
  1578. kkill_daemons: 59: find osd-pool-create
  1579. kkill_daemons: 59: grep pidfile
  1580. kill_daemons: 61: for pidfile in '$(find $dir | grep pidfile)'
  1581. kkill_daemons: 62: cat osd-pool-create/a/pidfile
  1582. kill_daemons: 62: pid=13826
  1583. kill_daemons: 63: for try in 0 1 1 1 2 3
  1584. kill_daemons: 64: kill -9 13826
  1585. kill_daemons: 65: sleep 0
  1586. kill_daemons: 63: for try in 0 1 1 1 2 3
  1587. kill_daemons: 64: kill -9 13826
  1588. kill_daemons: 65: sleep 1
  1589. kill_daemons: 63: for try in 0 1 1 1 2 3
  1590. kill_daemons: 64: kill -9 13826
  1591. kill_daemons: 64: break
  1592. teardown: 26: rm -fr osd-pool-create
  1593. main: 113: return 1