Skip to content
Snippets Groups Projects
Commit c1484857 authored by jsclose's avatar jsclose
Browse files

fixed tests

parent 0f11d91c
Branches
No related tags found
No related merge requests found
No preview for this file type
No preview for this file type
QueueTest 0 → 100755
File added
File added
File added
......@@ -33,7 +33,9 @@ int main ( int argc, char *argv[] )
unordered_map < size_t, int > *duplicateUrlMap = new unordered_map < size_t, int >( );
UrlFrontier *urlFrontier = new UrlFrontier( );
ProducerConsumerQueue < DocIndex * > *IndexerQueue = new ProducerConsumerQueue < DocIndex * >( );
Indexer indexer( IndexerQueue );
ProducerConsumerQueue< unordered_map<string , DocIndex * > > *AnchorQueue = new ProducerConsumerQueue< unordered_map<string , DocIndex * > >( );
Indexer indexer( IndexerQueue , AnchorQueue);
string path = util::GetCurrentWorkingDir() +"/crawler/tests/testSeeds.txt";
/*
seeds = util::getFileMap( path );
......@@ -72,8 +74,10 @@ int main ( int argc, char *argv[] )
urlFrontier->Push(url2);
indexer.StartThread( );
Crawler crawler( mode, urlFrontier, IndexerQueue );
crawler.SpawnSpiders( numberOfSpiders );
Crawler crawler( mode, urlFrontier, IndexerQueue, AnchorQueue );
atomic_bool *alive = new atomic_bool(true);
crawler.SpawnSpiders( numberOfSpiders , alive);
crawler.WaitOnAllSpiders( );
indexer.WaitForFinish( );
......
......@@ -31,10 +31,11 @@ int main ( int argc, char *argv[] )
int numberOfSpiders = 1;
unordered_map < size_t, int > *duplicateUrlMap = new unordered_map < size_t, int >( );
UrlFrontier *urlFrontier = new UrlFrontier( );
ProducerConsumerQueue< unordered_map<string , DocIndex * > > *AnchorQueue = new ProducerConsumerQueue< unordered_map<string , DocIndex * > >( );
ProducerConsumerQueue < DocIndex * > *IndexerQueue = new ProducerConsumerQueue < DocIndex * >( );
Indexer indexer( IndexerQueue );
Indexer indexer( IndexerQueue, AnchorQueue );
string path = util::GetCurrentWorkingDir() +"/crawler/tests/testSeeds.txt";
......@@ -52,8 +53,10 @@ int main ( int argc, char *argv[] )
indexer.StartThread( );
Crawler crawler( mode, urlFrontier, IndexerQueue );
crawler.SpawnSpiders( numberOfSpiders );
Crawler crawler( mode, urlFrontier, IndexerQueue, AnchorQueue );
atomic_bool *alive = new atomic_bool(true);
crawler.SpawnSpiders( numberOfSpiders , alive );
crawler.WaitOnAllSpiders( );
indexer.WaitForFinish( );
......
No preview for this file type
File added
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Please register or to comment