Uploaded image for project: 'Core Server'
  1. Core Server
  2. SERVER-3125

replset synchronisation performance problem (storing files)

    • Type: Icon: Bug Bug
    • Resolution: Done
    • Priority: Icon: Critical - P2 Critical - P2
    • None
    • Affects Version/s: 1.8.1
    • Component/s: Replication
    • None
    • Environment:
      linux 2.6.38 kernel, 4GB ram, 2GB swap for each node, 3x nodex, 4th computer with application in java. connection between computers - 100mbps LAN, default oplog size (no oplogSize directive in configuration file)
    • Linux

      My java app is storing messages and files in mongo.

      Secondary nodes couldn't receive data enough fast to replicate changes.

      Set name: testreplset
      Majority up: yes
      Member id Up cctime Last heartbeat Votes Priority State Messages optime skew
      172.17.0.251:27017 0 1 74 hrs 1 sec ago 1 1 RECOVERING error RS102 too stale to catch up 4dd6539b:81
      172.17.0.252:27017 (me) 1 1 74 hrs 1 1 RECOVERING error RS102 too stale to catch up 4dd6546d:7c
      172.17.0.253:27017 2 1 74 hrs 1 sec ago 1 1 PRIMARY 4dd65853:8

      12:28:59 [startReplSets] replSet STARTUP2
      12:28:59 [rs Manager] replSet can't see a majority, will not try to elect self
      12:28:59 [replica set sync] replSet SECONDARY
      12:28:59 [ReplSetHealthPollTask] replSet info 172.17.0.253:27017 is up
      12:28:59 [ReplSetHealthPollTask] replSet member 172.17.0.253:27017 SECONDARY
      12:29:00 [rs Manager] replSet info electSelf 1
      12:29:00 [rs Manager] replSet couldn't elect self, only received -9999 votes
      12:29:01 [conn2] replSet info voting yea for 2
      12:29:01 [ReplSetHealthPollTask] replSet member 172.17.0.253:27017 PRIMARY
      12:29:01 [ReplSetHealthPollTask] replSet info 172.17.0.251:27017 is up
      12:29:01 [ReplSetHealthPollTask] replSet member 172.17.0.251:27017 SECONDARY
      12:29:10 [replica set sync] replSet syncThread: 10278 dbclient error communicating with server: 172.17.0.253:27017
      12:29:11 [ReplSetHealthPollTask] replSet info 172.17.0.253:27017 is down (or slow to respond): DBClientBase::findOne: transport error: 172.17.0.253:27017 query: { replSetHeartbeat: "testreplset", v: 1, pv: 1, checkEmpty: false, from: " 12:29:12 [rs Manager] replSet not trying to elect self as responded yea to someone else recently
      12:29:18 [conn6] replSet info voting yea for 2
      12:29:18 [rs Manager] replSet not trying to elect self as responded yea to someone else recently
      12:29:19 [ReplSetHealthPollTask] replSet info 172.17.0.253:27017 is up
      Fri May 20 13:45:44 [ReplSetHealthPollTask] replSet member 172.17.0.251:27017 RECOVERING
      13:48:55 [replica set sync] replSet error RS102 too stale to catch up, at least from 172.17.0.253:27017
      13:48:55 [replica set sync] replSet our last optime : May 20 13:45:49 4dd6546d:7c
      13:48:55 [replica set sync] replSet oldest at 172.17.0.253:27017 : May 20 13:45:52 4dd65470:19
      13:48:55 [replica set sync] replSet See http://www.mongodb.org/display/DOCS/Resyncing+a+Very+Stale+Replica+Set+Member
      13:48:55 [replica set sync] replSet error RS102 too stale to catch up
      13:48:55 [replica set sync] replSet RECOVERING
      13:49:46 [ReplSetHealthPollTask] replSet member 172.17.0.251:27017 SECONDARY
      13:49:48 [ReplSetHealthPollTask] replSet member 172.17.0.251:27017 RECOVERING
      13:50:55 [replica set sync] replSet error RS102 too stale to catch up, at least from 172.17.0.253:27017
      13:50:55 [replica set sync] replSet our last optime : May 20 13:45:49 4dd6546d:7c
      13:50:55 [replica set sync] replSet oldest at 172.17.0.253:27017 : May 20 13:47:42 4dd654de:32
      13:50:55 [replica set sync] replSet See http://www.mongodb.org/display/DOCS/Resyncing+a+Very+Stale+Replica+Set+Member
      13:50:55 [replica set sync] replSet error RS102 too stale to catch up
      13:52:55 [replica set sync] replSet SECONDARY
      13:52:56 [replica set sync] replSet error RS102 too stale to catch up, at least from 172.17.0.253:27017
      13:52:56 [replica set sync] replSet our last optime : May 20 13:45:49 4dd6546d:7c
      13:52:56 [replica set sync] replSet oldest at 172.17.0.253:27017 : May 20 13:49:51 4dd6555f:162
      13:52:56 [replica set sync] replSet See http://www.mongodb.org/display/DOCS/Resyncing+a+Very+Stale+Replica+Set+Member
      13:52:56 [replica set sync] replSet error RS102 too stale to catch up
      13:52:56 [replica set sync] replSet RECOVERING
      13:54:56 [replica set sync] replSet error RS102 too stale to catch up, at least from 172.17.0.253:27017
      13:54:56 [replica set sync] replSet our last optime : May 20 13:45:49 4dd6546d:7c
      13:54:56 [replica set sync] replSet oldest at 172.17.0.253:27017 : May 20 13:51:43 4dd655cf:47
      13:54:56 [replica set sync] replSet See http://www.mongodb.org/display/DOCS/Resyncing+a+Very+Stale+Replica+Set+Member
      13:54:56 [replica set sync] replSet error RS102 too stale to catch up

            Assignee:
            kristina Kristina Chodorow (Inactive)
            Reporter:
            msz MartinS
            Votes:
            0 Vote for this issue
            Watchers:
            3 Start watching this issue

              Created:
              Updated:
              Resolved: