privacore-open-source-searc.../Repair.cpp
Ivan Skytte Jørgensen beeddcf35d Got rid of gb-include.h
2018-07-26 17:29:51 +02:00

1717 lines
48 KiB
C++

// Copyright 2007, Gigablast Inc.
#include "Repair.h"
#include "Rdb.h"
#include "Spider.h"
#include "Pages.h"
#include "Hostdb.h"
#include "ip.h"
#include "Spider.h"
#include "SpiderLoop.h"
#include "Process.h"
#include "Posdb.h"
#include "Clusterdb.h"
#include "Linkdb.h"
#include "XmlDoc.h"
#include "File.h"
#include "RdbMerge.h"
#include "Collectiondb.h"
#include "UrlBlockCheck.h"
#include "SpiderdbSqlite.h"
#include "max_niceness.h"
#include "Conf.h"
#include "Mem.h"
#include "Errno.h"
#include <fcntl.h>
#include <unistd.h>
repair_mode_t g_repairMode = REPAIR_MODE_NONE;
// the global instance
Repair g_repair;
static void formRepairdatFilename(char dst[1024]) {
sprintf(dst, "%s/repair.dat", g_hostdb.m_dir);
}
static int32_t getMinRepairMode() {
if(g_repairMode < g_hostdb.getMinRepairMode())
return g_repairMode;
if(g_hostdb.getNumHosts() == 1)
return g_repairMode;
return g_hostdb.getMinRepairMode();
}
static int32_t getMinRepairModeBesides0() {
if(g_repairMode != 0 && g_repairMode < g_hostdb.getMinRepairModeBesides0())
return g_repairMode;
if(g_hostdb.getNumHosts() == 1)
return g_repairMode;
return g_hostdb.getMinRepairModeBesides0();
}
static Rdb **getSecondaryRdbs ( int32_t *nsr ) {
static Rdb *s_rdbs[50];
static int32_t s_nsr = 0;
static bool s_init = false;
if ( ! s_init ) {
s_init = true;
s_nsr = 0;
s_rdbs[s_nsr++] = g_titledb2.getRdb ();
s_rdbs[s_nsr++] = g_posdb2.getRdb ();
s_rdbs[s_nsr++] = g_clusterdb2.getRdb ();
s_rdbs[s_nsr++] = g_linkdb2.getRdb ();
s_rdbs[s_nsr++] = g_tagdb2.getRdb ();
}
*nsr = s_nsr;
return s_rdbs;
}
Repair::Repair() {
// Coverity
m_docId = 0;
m_totalMem = 0;
m_stage = 0;
m_updated = false;
m_nextTitledbKey = 0;
m_endKey = 0;
m_SAVE_START = 0;
m_lastDocId = 0;
m_prevDocId = 0;
m_completedFirstScan = false;
m_completedSpiderdbScan = false;
m_lastTitledbKey = 0;
m_lastSpiderdbKey = 0;
m_recsScanned = 0;
m_recsetErrors = 0;
m_recsCorruptErrors = 0;
m_recsDupDocIds = 0;
m_recsNegativeKeys = 0;
m_recsWrongGroupId = 0;
m_recsInjected = 0;
m_nonIndexableExtensions = 0;
m_urlBlocked = 0;
m_urlUnwanted = 0;
m_spiderRecsScanned = 0;
m_spiderRecSetErrors = 0;
m_spiderRecNotAssigned = 0;
m_spiderRecBadTLD = 0;
m_rebuildTitledb = false;
m_rebuildPosdb = false;
m_rebuildClusterdb = false;
m_rebuildSpiderdb = false;
m_rebuildSpiderdbSmall = false;
m_rebuildSitedb = false;
m_rebuildLinkdb = false;
m_rebuildTagdb = false;
m_fullRebuild = true;
m_rebuildRoots = true;
m_rebuildNonRoots = true;
m_collnum = 0;
m_colli = 0;
m_numColls = 0;
m_SAVE_END = 0;
m_cr=NULL;
m_startTime = 0;
m_isSuspended = false;
m_numOutstandingInjects = 0;
m_msg5InUse = false;
m_saveRepairState = false;
m_isRetrying = false;
memset(m_collOffs, 0, sizeof(m_collOffs));
memset(m_collLens, 0, sizeof(m_collLens));
}
bool Repair::init ( ) {
//logf(LOG_DEBUG,"repair: TODO: alloc s_docs[] on demand to save mem");
m_msg5InUse = false;
m_isSuspended = false;
m_saveRepairState = false;
m_isRetrying = false;
if (!g_loop.registerSleepCallback(1, NULL, repairWrapper, "Repair::repairWrapper")) {
log(LOG_WARN, "repair: Failed register callback.");
return false;
}
return true;
}
bool Repair::isRepairActive() const {
return g_repairMode >= REPAIR_MODE_4;
}
// . called once a second by a sleep callback
// . this is responsible for advancing from one g_repairMode to the next
void Repair::repairWrapper(int fd, void *state) {
g_errno = 0;
// . all hosts should have their g_conf.m_repairMode parm set
// . it is global now, not collection based, since we need to
// lock down titledb for the scan and there could be recs from
// the collection we are repairing in titledb's rdbtree, which,
// when dumped, would mess up our scan.
if ( ! g_conf.m_repairingEnabled ) return;
// if it got turned back on after being suspended, start where
// we left off, this is how we re-enter Repair::loop()
if ( g_repair.m_isSuspended && g_repairMode == REPAIR_MODE_4 ) {
// unsuspend it
g_repair.m_isSuspended = false;
log("repair: Resuming repair scan after suspension.");
// try to read another title rec, or whatever
g_repair.loop();
return;
}
// if we are in retry mode
if ( g_repair.m_isRetrying && g_repairMode == REPAIR_MODE_4 ) {
// reset it
g_repair.m_isRetrying = false;
// try to read another title rec, or whatever
g_repair.loop();
return;
}
//
// ok, repairing is enabled at this point
//
static bool s_oldConfSpideringEnabled = false;
// are we just starting?
if ( g_repairMode == REPAIR_MODE_NONE ) {
// turn spiders off since repairing is enabled
s_oldConfSpideringEnabled = g_conf.m_spideringEnabled;
g_conf.m_spideringEnabled = false;
g_repair.m_startTime = gettimeofdayInMilliseconds();
// enter repair mode level 1
g_repairMode = REPAIR_MODE_1;
log("repair: Waiting for all writing operations to stop.");
}
// we can only enter repairMode 2 once all "writing" has stopped
if ( g_repairMode == REPAIR_MODE_1 ) {
// wait for all merging to stop just to be on the safe side
if ( g_merge.isMerging() ) return;
// wait until no spiders out
if ( g_spiderLoop.getNumSpidersOut() > 0 ) return;
// wait for ny outstanding unlinks or renames to finish
if ( BigFile::anyOngoingUnlinksOrRenames() ) return;
// . make sure all Msg4s are done and have completely added all
// recs they were supposed to
// . PROBLEM: if resuming a repair after re-starting, we can
// not turn on repairing
// . SOLVED: saveState() for msg4 uses different filename
if ( hasAddsInQueue() ) return;
// . ok, go to level 2
// . we can only get to level *3* once PingServer.cpp sees
// that all hosts in the cluster are in level 2. that way we
// guarantee not to add or delete any recs from any rdb,
// because that could damage the repair. PingServer will
// call g_repair.allHostsRead() when they all report they
// have a repair mode of 2.
g_repairMode = REPAIR_MODE_2;
log("repair: All oustanding writing operations stopped. ");
log("repair: Waiting for all other hosts to stop, too.");
}
// we can only enter mode 3 once all hosts are in 2 or higher
if ( g_repairMode == REPAIR_MODE_2 ) {
// we are still waiting on some guy if this is <= 1
if ( getMinRepairMode() < REPAIR_MODE_2 ) return;
// . this will return true if everything is saved to disk that
// needs to be, otherwise false if waiting on an rdb to finish
// saving
// . do this after all hosts are done writing, otherwise
// they might add data to our rdbs!
if ( !saveAllRdbs()) return;
//log("repair: Initializing the new Rdbs and scan parameters.");
// reset scan info BEFORE calling Repair::load()
g_repair.resetForNewCollection();
// before calling loop for the first time, init the scan,
// this will block and only return when it is done
g_repair.initScan();
// on error this sets g_repairingEnabled to false
if ( ! g_conf.m_repairingEnabled ) return;
// save "addsinprogress" file now so that the file will be
// saved as essentially an empty file at this point.
saveAddsInProgress ( NULL );
// hey, everyone is done "writing"
g_repairMode = REPAIR_MODE_3;
log("repair: All data saved and clock synced.");
log("repair: Waiting for all hosts to save and sync clocks.");
}
if ( g_repairMode == REPAIR_MODE_3 ) {
// wait for others to save everything
if ( getMinRepairMode() < REPAIR_MODE_3 ) return;
// start the loop
log("repair: All hosts saved.");
log("repair: Loading repair-addsinprogress.dat");
// . tell Msg4 to load state using the new filename now
// . load "repair-addsinprogress" file
loadAddsInProgress ( "repair-" );
//log("repair: Scanning titledb file #%" PRId32".", g_repair.m_fn );
log("repair: Starting repair scan.");
// advance
g_repairMode = REPAIR_MODE_4;
// only trigger repair loop on spider hosts
if (g_hostdb.getMyHost()->m_spiderEnabled) {
// now start calling the loop. returns false if blocks
if (!g_repair.loop()) return;
} else {
// assume we have completed scan
g_repair.m_completedFirstScan = true;
g_repair.m_completedSpiderdbScan = true;
}
}
// we can only enter mode 4 once we have completed the repairs
// and have dumped all the in-memory data to disk
if ( g_repairMode == REPAIR_MODE_4 ) {
// wait for scan loops to complete
if ( ! g_repair.m_completedFirstScan ) return;
if ( ! g_repair.m_completedSpiderdbScan ) return;
log("repair: Scan completed.");
log("repair: Waiting for other hosts to complete scan.");
// ok, we are ready to update the data files
g_repairMode = REPAIR_MODE_5;
}
// we can only enter mode 5 once all hosts are in 4 or higher
if ( g_repairMode == REPAIR_MODE_5 ) {
// if add queues still adding, wait, otherwise they will not
// be able to add to our rebuild collection
if ( hasAddsInQueue() ) return;
log("repair: All adds have been flushed.");
log("repair: Waiting for all other hosts to flush out their add operations.");
// update repair mode
g_repairMode = REPAIR_MODE_6;
}
if ( g_repairMode == REPAIR_MODE_6 ) {
// wait for everyone to get to mode 6 before we dump, otherwise
// data might arrive in the middle of the dumping and it stays
// in the in-memory RdbTree!
if ( getMinRepairMode() < REPAIR_MODE_6 ) return;
// we might have to dump again
g_repair.dumpLoop();
// are we done dumping?
if ( ! g_repair.dumpsCompleted() ) return;
// wait for all merging to stop just to be on the safe side
if ( g_merge.isMerging() ) return;
// wait for ny outstanding unlinks or renames to finish
if ( BigFile::anyOngoingUnlinksOrRenames() ) return;
log("repair: Final dump completed.");
log("repair: Updating rdbs to use newly repaired data.");
// everyone is ready
g_repairMode = REPAIR_MODE_7;
}
// we can only enter mode 6 once we are done updating the original
// rdbs with the rebuilt/repaired data. we move the old rdb data files
// into the trash and replace it with the new data.
if ( g_repairMode == REPAIR_MODE_7 ) {
// wait for autosave...
if ( g_process.m_mode ) return; // = SAVE_MODE;
// save to disk so it zeroes out indexdbRebuild-saved.dat
// which should have 0 records in it cuz we dumped it above
// in g_repair.dumpLoop()
if ( !saveAllRdbs()) return;
// . this blocks and gets the job done
// . this will move the old *.dat and *-saved.dat files into
// a subdir in the trash subdir
// . it will rename the rebuilt files to remove the "Rebuild"
// from their filenames
// . it will then restart the primary rdbs using those newly
// rebuilt and renamed files
// . this will not allow itself to be called more than once
// per scan/repair process
g_repair.updateRdbs();
log("repair: resetting secondary rdbs.");
// . only do this after indexdbRebuild-saved.dat has had a
// chance to save to "zero-out" its file on disk
// . all done with these guys, free their mem
g_repair.resetSecondaryRdbs();
// save "repair-addsinprogress" now so that the file will
// be saved as essentially an empty file at this point.
saveAddsInProgress ( "repair-" );
// reset it again in case it gets saved again later
g_repair.resetForNewCollection();
// unlink the repair.dat file, in case we core and are unable
// to save the freshly-reset repair.dat file
log("repair: unlinking repair.dat");
char tmp[1024];
formRepairdatFilename(tmp);
::unlink ( tmp );
// do not save it again! we just unlinked it!!
g_repair.m_saveRepairState = false;
log("repair: Waiting for other hosts to complete update.");
// ready to reset
g_repairMode = REPAIR_MODE_8;
}
// go back to mode 0 once all hosts have reached mode 8
if ( g_repairMode == REPAIR_MODE_8 ) {
// nobody can be in <8 (they might be 0!)
if ( getMinRepairModeBesides0() != REPAIR_MODE_8 ) return;
log("repair: Exiting repair mode. took %" PRId64" ms",
gettimeofdayInMilliseconds() - g_repair.m_startTime);
// turn it off to prevent going back to mode 1 again
g_conf.m_repairingEnabled = false;
// restore spider config
g_conf.m_spideringEnabled = s_oldConfSpideringEnabled;
s_oldConfSpideringEnabled = false;
// ok reset
g_repairMode = REPAIR_MODE_NONE;
}
}
void Repair::resetForNewCollection ( ) {
m_stage = 0;
m_lastDocId = 0;
m_prevDocId = 0;
m_completedFirstScan = false;
m_completedSpiderdbScan = false;
}
void Repair::initScan ( ) {
// reset some stuff for the titledb scan
m_nextTitledbKey.setMin();
m_lastSpiderdbKey.setMin();
m_endKey.setMax();
m_titleRecList.reset();
// all Repair::updateRdbs() to be called
m_updated = false;
// titledb scan stats
m_recsScanned = 0;
m_recsNegativeKeys = 0;
m_recsetErrors = 0;
m_recsCorruptErrors = 0;
m_recsDupDocIds = 0;
m_recsWrongGroupId = 0;
m_nonIndexableExtensions = 0;
m_urlBlocked = 0;
m_urlUnwanted = 0;
m_spiderRecsScanned = 0;
m_spiderRecSetErrors = 0;
m_spiderRecNotAssigned = 0;
m_spiderRecBadTLD = 0;
m_rebuildTitledb = g_conf.m_rebuildTitledb;
m_rebuildPosdb = g_conf.m_rebuildPosdb;
m_rebuildClusterdb = g_conf.m_rebuildClusterdb;
m_rebuildSpiderdb = g_conf.m_rebuildSpiderdb;
m_rebuildSpiderdbSmall = g_conf.m_rebuildSpiderdbSmall;
m_rebuildLinkdb = g_conf.m_rebuildLinkdb;
m_fullRebuild = g_conf.m_fullRebuild;
m_rebuildRoots = g_conf.m_rebuildRoots;
m_rebuildNonRoots = g_conf.m_rebuildNonRoots;
m_numOutstandingInjects = 0;
if ( m_fullRebuild ) {
// why rebuild titledb? its the base. no we need to
// rebuild it for new event displays.
m_rebuildTitledb = true;
m_rebuildSpiderdb = false;
m_rebuildSpiderdbSmall = false;
m_rebuildPosdb = true;
m_rebuildClusterdb = true;
m_rebuildLinkdb = true;
}
// . set the list of ptrs to the collections we have to repair
// . should be comma or space separated in g_conf.m_collsToRepair
// . none listed means to repair all collections
char *s = g_conf.m_collsToRepair.getBufStart();
char *cbuf = g_conf.m_collsToRepair.getBufStart();
char emptyStr[1]; emptyStr[0] = '\0';
if ( ! s ) s = emptyStr;
if ( ! cbuf ) cbuf = emptyStr;
// reset the list of ptrs to colls to repair
m_numColls = 0;
// scan through the collections in the string, if there are any
collLoop:
// skip non alnum chars
while ( *s && !is_alnum_a(*s) ) s++;
// if not at the end of the string, grab the collection
if ( *s ) {
m_collOffs[m_numColls] = s - cbuf;
// hold it
char *begin = s;
// find the length
while ( *s && *s != ',' && !is_wspace_a(*s) ) s++;
// store that, too
m_collLens[m_numColls] = s - begin;
// advance the number of collections
m_numColls++;
// get the next collection if under 100 collections still
if ( m_numColls < maxCollections ) goto collLoop;
}
// split the mem we have available among the rdbs
m_totalMem = g_conf.m_repairMem;
// 30MB min
if ( m_totalMem < 30000000 ) m_totalMem = 30000000;
//
// try to get some more mem.
//
// weight factors
float weight = 0;
if ( m_rebuildTitledb ) weight += 100.0;
if ( m_rebuildPosdb ) weight += 100.0;
if ( m_rebuildClusterdb ) weight += 1.0;
if ( m_rebuildSpiderdb ) weight += 5.0;
if ( m_rebuildSpiderdbSmall) weight += 5.0;
if ( m_rebuildLinkdb ) weight += 20.0;
if ( m_rebuildTagdb ) weight += 5.0;
// assign memory based on weight
int32_t titledbMem = 0;
int32_t posdbMem = 0;
int32_t clusterdbMem = 0;
int32_t spiderdbMem = 0;
int32_t linkdbMem = 0;
float tt = (float)m_totalMem;
if ( m_rebuildTitledb ) titledbMem = (int32_t)((100.0 * tt)/weight);
if ( m_rebuildPosdb ) posdbMem = (int32_t)((100.0 * tt)/weight);
if ( m_rebuildClusterdb ) clusterdbMem = (int32_t)(( 1.0 * tt)/weight);
if ( m_rebuildSpiderdb ) spiderdbMem = (int32_t)(( 5.0 * tt)/weight);
if ( m_rebuildSpiderdbSmall ) spiderdbMem = (int32_t)(( 5.0 * tt)/weight);
if ( m_rebuildLinkdb ) linkdbMem = (int32_t)(( 20.0 * tt)/weight);
if ( m_numColls <= 0 ) {
log("rebuild: Rebuild had no collection specified. You need to enter a collection or list of collections.");
goto hadError;
}
// init secondary rdbs
if ( m_rebuildTitledb ) {
if ( ! g_titledb2.init2 ( titledbMem ) ) goto hadError;
// clean tree in case loaded from saved file
Rdb *r = g_titledb2.getRdb();
if ( r ) r->cleanTree();
}
if ( m_rebuildPosdb ) {
if ( ! g_posdb2.init2 ( posdbMem ) ) goto hadError;
// clean tree in case loaded from saved file
Rdb *r = g_posdb2.getRdb();
if ( r ) r->cleanTree();
}
if ( m_rebuildClusterdb )
if ( ! g_clusterdb2.init2 ( clusterdbMem ) ) goto hadError;
if ( m_rebuildLinkdb )
if ( ! g_linkdb2.init2 ( linkdbMem ) ) goto hadError;
g_errno = 0;
// reset current coll we are repairing
m_colli = -1;
m_completedFirstScan = false;
// . tell it to advance to the next collection
// . this will call addColl() on the appropriate Rdbs
// . it will call addColl() on the primary rdbs for m_fullRebuild
getNextCollToRepair();
// if could not get any, bail
if ( ! m_cr ) goto hadError;
g_errno = 0;
// load the old repair state if on disk, this will block
load();
// now we can save if we need to
m_saveRepairState = true;
// if error loading, ignore it
g_errno = 0;
return;
// on any init2() error, reset all and return true
hadError:
int32_t saved_errno = g_errno;
// all done with these guys
resetSecondaryRdbs();
// pull back g_errno
g_errno = saved_errno;
log("repair: Had error in repair init. %s. Exiting.",
mstrerror(g_errno));
// back to step 0
g_repairMode = REPAIR_MODE_NONE;
m_colli = -1;
g_conf.m_repairingEnabled = false;
return;
}
// . sets m_coll/m_collLen to the next collection to repair
// . sets m_coll to NULL when none are left (we are done)
void Repair::getNextCollToRepair ( ) {
// . advance index into collections
// . can be index into m_colls or into g_collectiondb
m_colli++;
// ptr to first coll
if ( m_numColls ) {
if ( m_colli >= m_numColls ) {
//m_coll = NULL;
//m_collLen = 0;
return;
}
char *buf = g_conf.m_collsToRepair.getBufStart();
char *coll = buf + m_collOffs [m_colli];
int collLen = m_collLens[m_colli];
m_cr = g_collectiondb.getRec (coll, collLen);
// if DNE, set m_coll to NULL to stop repairing
if ( ! m_cr ) { g_errno = ENOCOLLREC; return; }
}
// otherwise, we are repairing every collection by default
else {
m_cr = NULL;
// loop m_colli over all the possible collnums
while ( ! m_cr && m_colli < g_collectiondb.getNumRecs() )
m_cr = g_collectiondb.getRec(++m_colli);
if ( ! m_cr ) {
g_errno = ENOCOLLREC;
return;
}
}
// collection cannot be deleted while we are in repair mode...
m_collnum = m_cr->m_collnum;
log("repair: now rebuilding for collection '%s' (%i)"
, m_cr->m_coll
, (int)m_collnum
);
char *coll = m_cr->m_coll;
// add collection to secondary rdbs
if ( m_rebuildTitledb ) {
if ( //! g_titledb2.addColl ( m_coll ) &&
! g_titledb2.getRdb()->addRdbBase1(coll) &&
g_errno != EEXIST ) goto hadError;
}
if ( m_rebuildPosdb ) {
if ( ! g_posdb2.getRdb()->addRdbBase1 ( coll ) &&
g_errno != EEXIST ) goto hadError;
}
if ( m_rebuildClusterdb ) {
if ( ! g_clusterdb2.getRdb()->addRdbBase1 ( coll ) &&
g_errno != EEXIST ) goto hadError;
}
if ( m_rebuildLinkdb ) {
if ( ! g_linkdb2.getRdb()->addRdbBase1 ( coll ) &&
g_errno != EEXIST ) goto hadError;
}
return;
hadError:
log("repair: Had error getting next coll to repair: %s. Exiting.",
mstrerror(g_errno));
}
void Repair::loopWrapper(void *state, RdbList *list, Msg5 *msg5) {
Repair *THIS = (Repair *)state;
THIS->m_msg5InUse = false;
THIS->loop();
}
enum {
STAGE_TITLEDB_0 = 0 ,
STAGE_TITLEDB_1 ,
STAGE_TITLEDB_2 ,
STAGE_TITLEDB_3 ,
STAGE_TITLEDB_4 ,
STAGE_SPIDERDB_0
};
bool Repair::save ( ) {
// do not do a blocking save for auto save if
// we never entere repair mode
if ( ! m_saveRepairState ) return true;
// log it
log("repair: saving repair.dat");
char tmp[1024];
formRepairdatFilename(tmp);
File ff;
ff.set ( tmp );
if ( ! ff.open ( O_RDWR | O_CREAT | O_TRUNC ) ) {
log(LOG_WARN, "repair: Could not open %s : %s", ff.getFilename(), mstrerror(g_errno));
return false;
}
g_errno = 0;
int32_t size = &m_SAVE_END - &m_SAVE_START;
int bytes_written = ff.write(&m_SAVE_START, size, 0 );
ff.close();
if(bytes_written!=size) {
log(LOG_WARN, "repair: Could not write to %s : %s", ff.getFilename(), mstrerror(g_errno));
ff.unlink();
return false;
}
return true;
}
bool Repair::load ( ) {
char tmp[1024];
formRepairdatFilename(tmp);
File ff;
ff.set ( tmp );
logf(LOG_INIT,"repair: Loading %s to resume repair.",tmp);
if ( ! ff.open ( O_RDONLY ) ) {
log(LOG_WARN, "repair: Could not open %s : %s", ff.getFilename(), mstrerror(g_errno));
return false;
}
int32_t size = &m_SAVE_END - &m_SAVE_START;
if(ff.getFileSize() != size) {
log(LOG_WARN, "repair: %s exists but has wrong size", ff.getFilename());
ff.unlink();
return false;
}
g_errno = 0;
int bytes_read =ff.read(&m_SAVE_START, size, 0);
ff.close();
if(bytes_read!=size) {
log(LOG_WARN, "repair: Could not read from %s : %s", ff.getFilename(), mstrerror(g_errno));
ff.unlink();
}
// resume titledb scan?
m_nextTitledbKey = m_lastTitledbKey;
// reinstate the valuable vars
m_cr = g_collectiondb.getRec(m_collnum);
m_stage = STAGE_TITLEDB_0;
if ( m_completedFirstScan ) m_stage = STAGE_SPIDERDB_0;
return true;
}
// . this is the main repair loop
// . this is repsonsible for calling all the repair functions
// . all repair callbacks given come back into this loop
// . returns false if blocked, true otherwise
// . sets g_errno on error
bool Repair::loop() {
logTrace(g_conf.m_logTraceRepairs,"BEGIN");
// was repairing turned off all of a sudden?
if(!g_conf.m_repairingEnabled) {
//log("repair: suspending repair.");
// when it gets turned back on, the sleep callback above
// will notice it was suspended and call loop() again to
// resume where we left off...
m_isSuspended = true;
logTrace(g_conf.m_logTraceRepairs,"END, repair suspended");
return true;
}
// if we re-entered this loop from doneWithIndexDocWrapper
// do not launch another msg5 if it is currently out!
if(m_msg5InUse) {
logTrace(g_conf.m_logTraceRepairs,"END, waiting for msg5");
return false;
}
// set this to on
g_process.m_repairNeedsSave = true;
//loop over the rdblist until something blocks or we hit m_maxRepairinjections limit
for(;;) {
if(g_process.isShuttingDown())
return true;
if(m_stage == STAGE_TITLEDB_0) {
logTrace(g_conf.m_logTraceRepairs,"STAGE_TITLEDB_0 - scanRecs");
m_stage++;
if(!scanRecs())
return false;
}
if(m_stage == STAGE_TITLEDB_1) {
logTrace(g_conf.m_logTraceRepairs,"STAGE_TITLEDB_1 - gotScanRecList");
m_stage++;
if(!gotScanRecList())
return false;
}
if(m_stage == STAGE_TITLEDB_2) {
logTrace(g_conf.m_logTraceRepairs,"STAGE_TITLEDB_2");
m_stage++;
}
// get the site rec to see if it is banned first, before injecting it
if(m_stage == STAGE_TITLEDB_3) {
logTrace(g_conf.m_logTraceRepairs,"STAGE_TITLEDB_3");
// if we have maxed out our injects, wait for one to come back
if(m_numOutstandingInjects >= g_conf.m_maxRepairinjections)
return false;
m_stage++;
logTrace(g_conf.m_logTraceRepairs,"injectTitleRec");
bool status = injectTitleRec();
logTrace(g_conf.m_logTraceRepairs,"injectTitleRec returned %s", status?"true":"false");
// try to launch another
if(m_numOutstandingInjects < g_conf.m_maxRepairinjections) {
m_stage = STAGE_TITLEDB_0;
logTrace(g_conf.m_logTraceRepairs,"Still have more free repair spiders, loop.");
continue;
}
// if we are full and it blocked... wait now
if(! status) {
logTrace(g_conf.m_logTraceRepairs,"END, return false. Full queue and blocked.");
return false;
}
}
if(m_stage == STAGE_TITLEDB_4) {
logTrace(g_conf.m_logTraceRepairs,"STAGE_TITLEDB_4");
m_stage++;
}
// if we are not done with the titledb scan loop back up
if(!m_completedFirstScan) {
m_stage = STAGE_TITLEDB_0;
logTrace(g_conf.m_logTraceRepairs,"loop, set STAGE_TITLEDB_0");
continue;
}
break;
}
// if we are waiting for injects to come back, return
if(m_numOutstandingInjects > 0) {
// tell injection complete wrapper to call us back, otherwise
// we never end up moving on to the spider phase
logTrace(g_conf.m_logTraceRepairs,"return false. Have %" PRId32" outstanding injects", m_numOutstandingInjects);
return false;
}
// reset list
m_titleRecList.reset();
// in order for dump to work we must be in mode 4 because
// Rdb::dumpTree() checks that
g_repairMode = REPAIR_MODE_4;
// force dump to disk of the newly rebuilt rdbs, because we need to
// make sure their trees are empty when the primary rdbs assume
// the data and map files of the secondary rdbs. i don't want to
// have to mess with tree data as well.
// if we do not complete the dump here it will be monitored above
// in the sleep wrapper, repairWrapper(), and that will call
// Repair::loop() (this function) again when the dump is done
// and we will be able to advance past this m_stage
// . dump the trees of all secondary rdbs that need it
//dumpLoop();
// are we done dumping?
//if ( ! dumpsCompleted() ) return false;
log(LOG_INFO,"repair: Repair finished. %" PRId64 " titledb records scanned; errors:"
" set=%" PRId64
" corrupt=%" PRId64
" dup=%" PRId64
" negative=%" PRId64
" wrong-group=%" PRId64
" injected=%" PRId64
" m_nonIndexableExtensions=%" PRId64
" m_urlBlocked=%" PRId64
" m_urlUnwanted=%" PRId64,
m_recsScanned,
m_recsetErrors,
m_recsCorruptErrors,
m_recsDupDocIds,
m_recsNegativeKeys,
m_recsWrongGroupId,
m_recsInjected,
m_nonIndexableExtensions,
m_urlBlocked,
m_urlUnwanted
);
// we are all done with the repair loop
logTrace(g_conf.m_logTraceRepairs,"END");
return true;
}
// this blocks
void Repair::updateRdbs ( ) {
if ( m_updated ) return;
// do not double call
m_updated = true;
// . replace old rdbs with the new ones
// . these calls must all block otherwise things will get out of sync
Rdb *rdb1;
Rdb *rdb2;
if ( m_rebuildTitledb ) {
rdb1 = g_titledb.getRdb ();
rdb2 = g_titledb2.getRdb();
rdb1->updateToRebuildFiles ( rdb2 , m_cr->m_coll );
}
if ( m_rebuildPosdb ) {
rdb1 = g_posdb.getRdb();
rdb2 = g_posdb2.getRdb();
rdb1->updateToRebuildFiles ( rdb2 , m_cr->m_coll );
}
if ( m_rebuildClusterdb ) {
rdb1 = g_clusterdb.getRdb();
rdb2 = g_clusterdb2.getRdb();
rdb1->updateToRebuildFiles ( rdb2 , m_cr->m_coll );
}
if ( m_rebuildLinkdb ) {
rdb1 = g_linkdb.getRdb();
rdb2 = g_linkdb2.getRdb();
rdb1->updateToRebuildFiles ( rdb2 , m_cr->m_coll );
}
if(m_rebuildSpiderdb || m_rebuildSpiderdbSmall)
SpiderdbSqlite::swapinSecondarySpiderdb(m_collnum, m_cr->m_coll);
}
void Repair::resetSecondaryRdbs ( ) {
int32_t nsr;
Rdb **rdbs = getSecondaryRdbs ( &nsr );
for ( int32_t i = 0 ; i < nsr ; i++ ) {
Rdb *rdb = rdbs[i];
rdb->reset();
}
//g_spiderdb2 has alraedy closed the the collection as a side-effect of SpiderdbSqlite::swapinSecondarySpiderdb()
}
bool Repair::dumpLoop ( ) {
int32_t nsr;
Rdb **rdbs = getSecondaryRdbs ( &nsr );
for ( int32_t i = 0 ; i < nsr ; i++ ) {
Rdb *rdb = rdbs[i];
rdb->submitRdbDumpJob(true);
}
g_errno = 0;
// . register sleep wrapper to check when dumping is done
// . it will call Repair::loop() when done
return false;
}
bool Repair::dumpsCompleted ( ) {
int32_t nsr;
Rdb **rdbs = getSecondaryRdbs ( &nsr );
for ( int32_t i = 0 ; i < nsr ; i++ ) {
Rdb *rdb = rdbs[i];
// anything in tree/buckets?
if ( rdb->getNumUsedNodes() ) return false;
// still dumping?
if (rdb->isDumping()) return false;
}
// no more dump activity
return true;
}
// . this is only called from repairLoop()
// . returns false if blocked, true otherwise
// . grab the next scan record
bool Repair::scanRecs ( ) {
// just the tree?
RdbBase *base = g_titledb.getRdb()->getBase ( m_collnum );
// always clear last bit of g_nextKey
m_nextTitledbKey.n0 &= 0xfffffffffffffffeLL;
// for saving
m_lastTitledbKey = m_nextTitledbKey;
log(LOG_DEBUG,"repair: nextKey=%s endKey=%s coll=%s collnum=%" PRId32" bnf=%" PRId32,
KEYSTR(&m_nextTitledbKey,sizeof(key96_t)),
KEYSTR(&m_endKey,sizeof(key96_t)),
m_cr->m_coll,
(int32_t)m_collnum,
(int32_t)base->getNumFiles());
// sanity check
if ( m_msg5InUse ) {
g_process.shutdownAbort(true); }
// when building we can get the rec from the twin in case of data corruption on disk
bool fixErrors = true;
// get the list of recs
g_errno = 0;
if ( m_msg5.getList ( RDB_TITLEDB ,
m_collnum ,
&m_titleRecList ,
&m_nextTitledbKey,
&m_endKey , // should be maxed!
1024 , // min rec sizes
true , // include tree?
0 , // startFileNum
-1 , // m_numFiles
this , // state
loopWrapper , // callback
MAX_NICENESS , // niceness
fixErrors , // do error correction?
-1 , // maxRetries
false)) // isRealMerge
return true;
m_msg5InUse = true;
return false;
}
// . this is only called from repairLoop()
// . returns false if blocked, true otherwise
bool Repair::gotScanRecList ( ) {
if ( g_errno == ECORRUPTDATA ) {
log("repair: Encountered corruption1 in titledb. NextKey=%s",
KEYSTR(&m_nextTitledbKey,sizeof(key96_t)));
// advance one if positive, must always start on a neg
if(!KEYNEG(m_nextTitledbKey))
m_nextTitledbKey++;
// count as error
m_recsCorruptErrors++;
}
// was there an error? list will probably be empty
if ( g_errno ) {
log("repair: Got error reading title rec: %s.",
mstrerror(g_errno));
// keep retrying, might be OOM
m_stage = STAGE_TITLEDB_0 ;
// sleep 1 second and retry
m_isRetrying = true;
// exit the loop code, Repair::loop() will be re-called
return false;
}
// all done with this bigfile if this list is empty
if ( m_titleRecList.isEmpty() ) {
m_completedFirstScan = true;
logf(LOG_INFO,"repair: Completed titledb scan of %" PRId64" records.",m_recsScanned);
//logf(LOG_INFO,"repair: Starting spiderdb scan.");
m_stage = STAGE_SPIDERDB_0;
// force spider scan completed now too!
m_completedSpiderdbScan = true;
return true;
}
key96_t tkey = m_titleRecList.getCurrentKey();
int64_t docId = Titledb::getDocId ( &tkey );
// save the current docid
m_docId = docId;
// count the title recs we scan
m_recsScanned++;
// skip if bad... CORRUPTION
if ( tkey < m_nextTitledbKey ) {
log("repair: Encountered corruption2 in titledb. key=%s < NextKey=%s FirstDocId=%" PRIu64".",
KEYSTR(&tkey,sizeof(key96_t)),
KEYSTR(&m_nextTitledbKey,sizeof(key96_t)),
docId);
m_nextTitledbKey++;
// advance one if positive, must always start on a negative key
if(!KEYNEG(m_nextTitledbKey))
m_nextTitledbKey++;
m_stage = STAGE_TITLEDB_0;
return true;
}
else {
// advance m_nextTitledbKey to get next titleRec
m_nextTitledbKey = m_titleRecList.getCurrentKey();
m_nextTitledbKey++;
// advance one if positive, must always start on a negative key
if(!KEYNEG(m_nextTitledbKey))
m_nextTitledbKey++;
}
// are we the host this url is meant for?
uint32_t shardNum = getShardNum (RDB_TITLEDB , &tkey );
if ( shardNum != getMyShardNum() ) {
m_recsWrongGroupId++;
m_stage = STAGE_TITLEDB_0;
return true;
}
// is it a negative titledb key?
if(KEYNEG(tkey)) {
m_recsNegativeKeys++;
m_stage = STAGE_TITLEDB_0;
return true;
}
return true;
}
void Repair::doneWithIndexDoc(XmlDoc *xd) {
logTrace(g_conf.m_logTraceRepairs,"BEGIN");
// preserve
int32_t saved_errno = g_errno;
// nuke it
mdelete ( xd , sizeof(XmlDoc) , "xdprnuke");
delete ( xd );
// reduce the count
g_repair.m_numOutstandingInjects--;
// error?
if ( saved_errno ) {
g_repair.m_recsetErrors++;
g_repair.m_stage = STAGE_TITLEDB_0; // 0
return;
}
logTrace(g_conf.m_logTraceRepairs,"END");
}
void Repair::doneWithIndexDocWrapper(void *state) {
logTrace(g_conf.m_logTraceRepairs,"BEGIN");
// clean up
doneWithIndexDoc ( (XmlDoc *)state );
// and re-enter the loop to get next title rec
g_repair.loop();
logTrace(g_conf.m_logTraceRepairs,"END");
}
bool Repair::injectTitleRec ( ) {
logTrace(g_conf.m_logTraceRepairs,"BEGIN");
// scan for our docid in the title rec list
char *titleRec = NULL;
int32_t titleRecSize = 0;
// scan the titleRecs in the list
for ( ; ! m_titleRecList.isExhausted() ; m_titleRecList.skipCurrentRecord ( ) ) {
// get the rec
char *rec = m_titleRecList.getCurrentRec();
int32_t recSize = m_titleRecList.getCurrentRecSize();
// get that key
key96_t *k = (key96_t *)rec;
// skip negative recs, first one should not be negative however
if(KEYNEG(*k)) continue;
// get docid of that guy
int64_t dd = Titledb::getDocId(k);
// compare that
if ( m_docId != dd ) continue;
// we got it!
titleRec = rec;
titleRecSize = recSize;
break;
}
if(m_rebuildSpiderdbSmall) {
logTrace(g_conf.m_logTraceRepairs,"Jumping to injectTitleRecSmall");
return injectTitleRecSmall(titleRec,titleRecSize);
}
XmlDoc *xd = NULL;
try { xd = new ( XmlDoc ); }
catch(std::bad_alloc&) {
g_errno = ENOMEM;
m_recsetErrors++;
m_stage = STAGE_TITLEDB_0; // 0
return true;
}
mnew ( xd , sizeof(XmlDoc),"xmldocpr");
if (!xd->set2(titleRec, -1, m_cr->m_coll, MAX_NICENESS)) {
m_recsetErrors++;
m_stage = STAGE_TITLEDB_0; // 0
logTrace(g_conf.m_logTraceRepairs,"END, return true. XmlDoc->set2 failed");
return true;
}
// set callback
xd->setCallback ( xd , doneWithIndexDocWrapper );
// clear any error involved with cache, it doesn't matter so much
g_errno = 0;
// invalidate certain things to recompute!
// we are now setting from docid
xd->m_tagRecValid = false;
// rebuild the title rec! otherwise we re-add the old one!!!!!!!
xd->m_titleRecBufValid = false;
// free it since set2() should have uncompressed it!
//mfree ( titleRec , titleRecSize, "repair" );
// and so xd doesn't free it
xd->m_titleRecBuf.purge();
// use the ptr_utf8Content that we have
xd->m_recycleContent = true;
// rebuild the content hash since we change that function sometimes
xd->m_contentHash32Valid = false;
// claim it, so "tr" is not overwritten
m_numOutstandingInjects++;
bool addToSecondaryRdbs = true;
xd->m_usePosdb = m_rebuildPosdb;
xd->m_useClusterdb = m_rebuildClusterdb;
xd->m_useLinkdb = m_rebuildLinkdb;
xd->m_useSpiderdb = m_rebuildSpiderdb;
xd->m_useTitledb = m_rebuildTitledb;
xd->m_useSecondaryRdbs = addToSecondaryRdbs;
// always use tagdb because if we update the sitenuminlinks
// or whatever, we want to add that to tagdb
xd->m_useTagdb = true;
if (!g_conf.m_rebuildAddOutlinks) {
xd->m_spiderLinks2 = 0;
}
// not if rebuilding link info though! we assume the old link info is
// bad...
if ( m_rebuildLinkdb ) {
xd->m_useTagdb = false;
}
if ( ! g_conf.m_rebuildRecycleLinkInfo ) {
// then recompute link info as well!
xd->m_linkInfo1Valid = false;
// make null to be safe
xd->ptr_linkInfo1 = NULL;
xd->size_linkInfo1 = 0;
}
// . also lookup site rank again!
// . this will use the value in tagdb if less than 48 hours otherwise
// it will recompute it
// . CRAP! this makes the data undeletable if siterank changes!
// so we have to be able to re-save our title rec with the new
// site rank info...
if ( xd->m_useTitledb ) {
// save for logging
xd->m_logLangId = xd->m_langId;
xd->m_logSiteNumInlinks = xd->m_siteNumInlinks;
// recompute site, no more domain sites allowed
xd->m_siteValid = false;
xd->ptr_site = NULL;
xd->size_site = 0;
// recalculate the sitenuminlinks
xd->m_siteNumInlinksValid = false;
// recalculate the langid
xd->m_langIdValid = false;
// recalcualte and store the link info
xd->m_linkInfo1Valid = false;
// make null to be safe
xd->ptr_linkInfo1 = NULL;
xd->size_linkInfo1 = 0;
// re-get the tag rec from tagdb
xd->m_tagRecValid = false;
xd->m_tagRecDataValid = false;
}
// don't recalculate isAdult flag when only rebuilding spiderdb
if (xd->m_useSpiderdb && !xd->m_useClusterdb && !xd->m_useTitledb) {
xd->m_isAdultValid = true;
}
if (!g_conf.m_rebuildUseTitleRecTagRec) {
xd->m_tagRecDataValid = false;
}
xd->m_priority = -1;
xd->m_priorityValid = true;
xd->m_contentValid = true;
xd->m_content = xd->ptr_utf8Content;
xd->m_contentLen = xd->size_utf8Content - 1;
// don't check block list
xd->m_blockedDocValid = true;
xd->m_blockedDoc = false;
// don't check site median page temperature
xd->m_calledServiceSiteMedianPageTemperature = true;
// . get the meta list to add
// . sets m_usePosdb, m_useTitledb, etc.
logTrace(g_conf.m_logTraceRepairs,"Calling indexDoc");
bool status = xd->indexDoc ( );
// blocked?
if ( ! status )
{
logTrace(g_conf.m_logTraceRepairs,"END, return false. XmlDoc->indexDoc blocked");
return false;
}
// give it back
doneWithIndexDoc ( xd );
logTrace(g_conf.m_logTraceRepairs,"END, return true");
return true;
}
namespace {
//state for when we create spider requests just with the titlerec first-url (and not the links in the documents)
struct SmallInjectState {
Msg4 msg4;
SpiderRequest sreq;
};
}
bool Repair::injectTitleRecSmall(char *titleRec, int32_t titleRecSize) {
logTrace(g_conf.m_logTraceRepairs,"BEGIN");
//decompress+decode xmldoc
XmlDoc xd;
if (!xd.set2(titleRec, titleRecSize, m_cr->m_coll, MAX_NICENESS)) {
m_recsetErrors++;
m_stage = STAGE_TITLEDB_0;
logTrace(g_conf.m_logTraceRepairs,"END, return true. XmlDoc->set2 failed");
return true;
}
//get url and check if it is still wanted
const Url *url = xd.getFirstUrl();
if(url->hasNonIndexableExtension(TITLEREC_CURRENT_VERSION)) {
m_nonIndexableExtensions++;
logTrace(g_conf.m_logTraceRepairs,"END, return true. hasNonIndexableExtension");
return true;
}
if(isUrlBlocked(*url)) {
m_urlBlocked++;
logTrace(g_conf.m_logTraceRepairs,"END, return true. isUrlBlocked");
return true;
}
if(isUrlUnwanted(*url)) {
m_urlUnwanted++;
logTrace(g_conf.m_logTraceRepairs,"END, return true. isUrlUnwanted");
return true;
}
//set up state on heap, and prepare spiderrequest for msg4-out
SmallInjectState *sis;
try {
sis = new SmallInjectState();
} catch(std::bad_alloc&) {
m_recsetErrors++; //sort of
logTrace(g_conf.m_logTraceRepairs,"END, return true. std::bad_alloc");
return true;
}
sis->sreq.reset();
strcpy(sis->sreq.m_url, url->getUrl());
sis->sreq.setKey(*xd.getFirstIp(), 0, false);
sis->sreq.m_firstIp = *xd.getFirstIp();
sis->sreq.m_hostHash32 = url->getHostHash32();
sis->sreq.m_domHash32 = url->getDomainHash32();
sis->sreq.m_siteHash32 = url->getHostHash32();
sis->sreq.m_addedTime = xd.m_firstIndexedDate;
if(xd.m_siteNumInlinksValid)
sis->sreq.m_siteNumInlinks = xd.m_siteNumInlinks;
if(sis->msg4.addMetaList((const char*)&sis->sreq,sis->sreq.getRecSize(), m_collnum, sis,smallInjectCallback, RDB2_SPIDERDB2_DEPRECATED)) {
//failed or immediateley succeeded
delete sis;
logTrace(g_conf.m_logTraceRepairs,"END, return true. addMetaList");
return true;
}
//blocked
logTrace(g_conf.m_logTraceRepairs,"END, return true. blocked");
return false;
}
void Repair::smallInjectCallback(void *state) {
SmallInjectState *sis = reinterpret_cast<SmallInjectState*>(state);
delete sis;
g_repair.loop();
}
// . returns false if fails cuz buffer cannot be grown (oom)
// . this is called by Parms.cpp
bool Repair::printRepairStatus(SafeBuf *sb) {
// default is a repairMode of 0, "not running"
const char *status = "not running";
if ( g_repairMode == REPAIR_MODE_NONE && g_conf.m_repairingEnabled )
status = "waiting for previous rebuild to complete";
if ( g_repairMode == REPAIR_MODE_1 )
status = "waiting for spiders or merge to stop";
if ( g_repairMode == REPAIR_MODE_2 )
status = "waiting for all hosts in network to stop "
"spidering and merging";
if ( g_repairMode == REPAIR_MODE_3 )
status = "waiting for all hosts to save";
if ( g_repairMode == REPAIR_MODE_4 ) {
if ( m_completedFirstScan )
status = "scanning old spiderdb";
else
status = "scanning old records";
}
if ( g_repairMode == REPAIR_MODE_5 )
status = "waiting for final dump to complete";
if ( g_repairMode == REPAIR_MODE_6 )
status = "waiting for others to finish scan and dump";
if ( g_repairMode == REPAIR_MODE_7 )
status = "updating rdbs with new data";
if ( g_repairMode == REPAIR_MODE_8 )
status = "waiting for all hosts to complete update";
// the titledb scan stats (phase 1)
int64_t ns = m_recsScanned ;
int64_t nr = g_titledb.getRdb()->getNumTotalRecs() ;
float ratio = nr ? ((float)ns * 100.0) / (float)nr : 0.0;
int64_t errors =
m_recsetErrors +
m_recsCorruptErrors +
m_recsDupDocIds ;
const char *newColl = " &nbsp; ";
const char *oldColl = " &nbsp; ";
if ( m_cr ) oldColl = m_cr->m_coll;
const Host *mh = g_hostdb.getMinRepairModeHost();
int32_t minHostId = -1;
char minIpBuf[64];
minIpBuf[0] = '\0';
int16_t minPort = 80;
if ( mh ) {
minHostId = mh->m_hostId;
int32_t minHostIp = g_hostdb.getBestIp ( mh );
iptoa(minHostIp,minIpBuf);
minPort = mh->getInternalHttpPort();
}
// now show the rebuild status
sb->safePrintf (
"<table class=\"main\" id=\"repairstatustable\">\n"
"<tr class=\"level1\"><th colspan=2>Rebuild Status</th></tr>\n"
"<tr class=\"level2\"><td colspan=2>"
"<font size=-2>"
"Use this to rebuild a database or to reindex "
"all pages to pick up new link text. Or to "
"reindex all pages to pick up new site rank info "
"from tagdb. To pick up "
"new link text you should rebuild titledb and posdb. "
"If unsure, just do a full rebuild, but it will "
"require about 2GB more than the disk used before "
"the rebuild, so at its peak the rebuild will use "
"a little more than double the disk space you "
"are using now. Also you will want to set "
"recycle link text to false to pick up the new link "
"text. However, if you just want to pick up "
"new sitenuminlinks tags in tagdb to get more "
"accurate siteranks for each result, then you can "
"leave the recycle link text set to true."
""
"<br><br>"
"All spidering for all collections will be disabled "
"when the rebuild is in progress. But you should "
"still be able to conduct searches on the original "
"index. You can pause "
"the rebuild by disabling <i>rebuild mode enabled"
"</i>. Each shard should save its rebuid state so "
"you can safely shut shards down when rebuilding "
"and they should resume on startup. When the rebuild "
"completes it moves the original files to the trash "
"subdirectory and replaces them with the newly "
"rebuilt files."
"</font>"
"</td></tr>"
// status (see list of above statuses)
"<tr><td width=50%%>status</td>"
"<td>%s</td></tr>\n"
"<tr><td width=50%%>rebuild mode"
"</td>"
"<td>%" PRId32"</td></tr>\n"
"<tr>"
"<td width=50%%>min rebuild mode</td>"
"<td>%" PRId32"</td></tr>\n"
"<tr>"
"<td width=50%%>host ID with min rebuild mode"
"</td>"
"<td><a href=\"http://%s:%hu/admin/rebuild\">"
"%" PRId32"</a></td></tr>\n"
"<tr><td>old collection</td>"
"<td>%s</td></tr>"
"<tr><td>new collection</td>"
"<td>%s</td></tr>"
,
status ,
(int32_t)g_repairMode,
(int32_t)g_hostdb.getMinRepairMode(),
minIpBuf, // ip string
minPort, // port
(int32_t)minHostId,
oldColl ,
newColl
);
sb->safePrintf (
// docs done, includes overwritten title recs
"<tr class=\"bg0\"><td>titledb recs scanned</td>"
"<td>%" PRId64" of up to %" PRId64 " (~%.2f%%)</td></tr>\n"
"<tr class=\"bg0\"><td>titledb rec error count</td>"
"<td>%" PRId64"</td></tr>\n"
// sub errors
"<tr class=\"bg0\"><td> &nbsp; set errors</td>"
"<td>%" PRId64"</td></tr>\n"
"<tr class=\"bg0\"><td> &nbsp; corrupt errors</td>"
"<td>%" PRId64"</td></tr>\n"
"<tr class=\"bg0\"><td> &nbsp; dup docid errors</td>"
"<td>%" PRId64"</td></tr>\n"
"<tr class=\"bg0\"><td> &nbsp; negative keys</td>"
"<td>%" PRId64"</td></tr>\n"
"<tr class=\"bg0\"><td> &nbsp; wrong shard</td>"
"<td>%" PRId64"</td></tr>\n"
"<tr class=\"bg0\"><td> &nbsp; non-indexable extension</td>"
"<td>%" PRId64"</td></tr>\n"
"<tr class=\"bg0\"><td> &nbsp; Blocked URL</td>"
"<td>%" PRId64"</td></tr>\n"
"<tr class=\"bg0\"><td> &nbsp; Unwanted</td>"
"<td>%" PRId64"</td></tr>\n"
,
ns ,
nr ,
ratio ,
errors ,
m_recsetErrors ,
m_recsCorruptErrors ,
m_recsDupDocIds ,
m_recsNegativeKeys ,
m_recsWrongGroupId,
m_nonIndexableExtensions,
m_urlBlocked,
m_urlUnwanted
);
int32_t nsr;
Rdb **rdbs = getSecondaryRdbs ( &nsr );
// . count the recs in each secondary rdb
// . those are the rdbs we are adding the recs to
for ( int32_t i = 0 ; i < nsr ; i++ ) {
const char *bg = DARK_BLUE;
Rdb *rdb = rdbs[i];
int64_t tr = rdb->getNumTotalRecs();
// skip if init2() as not called on it b/c the
// m_dbname will be 0
if ( tr == 0 ) continue;
sb->safePrintf(
"<tr bgcolor=#%s><td>%s2 recs</td>"
"<td>%" PRId64"</td></tr>\n" ,
bg,
rdb->getDbname(),
rdb->getNumTotalRecs());
}
// close up that table
sb->safePrintf("</table>\n<br>");
// print a table
sb->safePrintf (
"<table class=\"main\" width=\"100%%\" id=\"repairstatustable2\">\n"
// current collection being repaired
"<tr class=\"level1\"><th colspan=2>Rebuild Settings In Use</th></tr>"
// . print parms for this repair
// . they may differ than current controls because
// the current controls were changed after the
// repair started
"<tr>"
"<td width=50%%>full rebuild</td>"
"<td>%s</td></tr>\n"
"<tr><td>rebuild titledb</td>"
"<td>%s</td></tr>\n"
"<tr><td>rebuild posdb</td>"
"<td>%s</td></tr>\n"
"<tr><td>rebuild clusterdb</td>"
"<td>%s</td></tr>\n"
"<tr><td>rebuild spiderdb</td>"
"<td>%s</td></tr>\n"
"<tr><td>rebuild spiderdb (small)</td>"
"<td>%s</td></tr>\n"
"<tr><td>rebuild linkdb</td>"
"<td>%s</td></tr>\n"
"<tr><td>rebuild root urls</td>"
"<td>%s</td></tr>\n"
"<tr>"
"<td>rebuild non-root urls</td>"
"<td>%s</td></tr>\n"
"</table>\n"
"<br>\n"
,
m_fullRebuild ? "Y":"N",
m_rebuildTitledb ? "Y":"N",
m_rebuildPosdb ? "Y":"N",
m_rebuildClusterdb ? "Y":"N",
m_rebuildSpiderdb ? "Y":"N",
m_rebuildSpiderdbSmall ? "Y":"N",
m_rebuildLinkdb ? "Y":"N",
m_rebuildRoots ? "Y":"N",
m_rebuildNonRoots ? "Y":"N"
);
return true;
}
static bool s_savingAll = false;
// . return false if blocked, true otherwise
// . will call the callback when all have been saved
// . used by Repair.cpp to save all rdbs before doing repair work
bool Repair::saveAllRdbs() {
// only call once
if (s_savingAll) {
return false;
}
// set it
s_savingAll = true;
// TODO: why is this called like 100x per second when a merge is
// going on? why don't we sleep longer in between?
for (int32_t i = 0; i < g_process.m_numRdbs; i++) {
Rdb *rdb = g_process.m_rdbs[i];
if (!rdb->isInitialized() || rdb->getRdbId() == RDB_DOLEDB) {
continue;
}
// save/close it
rdb->saveTree(true, NULL, doneSavingRdb);
}
// return if still waiting on one to close
if (anyRdbNeedsSave()) {
return false;
}
// all done
return true;
}
// return false if one or more is still not closed yet
bool Repair::anyRdbNeedsSave() {
for (int32_t i = 0; i < g_process.m_numRdbs; i++) {
Rdb *rdb = g_process.m_rdbs[i];
if (rdb->getRdbId() == RDB_DOLEDB) {
continue;
}
if (rdb->needsSave()) {
return true;
}
}
s_savingAll = false;
return false;
}
// returns false if waiting on some to save
void Repair::doneSavingRdb(void *state) {
if (!anyRdbNeedsSave()) {
return;
}
// all done
s_savingAll = false;
}