mirror of
https://github.com/RetroShare/RetroShare.git
synced 2024-12-11 17:04:31 -05:00
c9f99c0be9
git-svn-id: http://svn.code.sf.net/p/retroshare/code/trunk@5030 b45a01b8-16f6-495d-af2f-9b41ad6348cc
1312 lines
37 KiB
C++
1312 lines
37 KiB
C++
/*
|
|
* libretroshare/src/ft: ftdatamultiplex.h
|
|
*
|
|
* File Transfer for RetroShare.
|
|
*
|
|
* Copyright 2008 by Robert Fernie.
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License Version 2 as published by the Free Software Foundation.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307
|
|
* USA.
|
|
*
|
|
* Please report all bugs and problems to "retroshare@lunamutt.com".
|
|
*
|
|
*/
|
|
|
|
/*
|
|
* ftDataMultiplexModule.
|
|
*
|
|
* This multiplexes the data from PQInterface to the ftTransferModules.
|
|
*/
|
|
|
|
#include "ft/ftdatamultiplex.h"
|
|
#include "ft/fttransfermodule.h"
|
|
#include "ft/ftfilecreator.h"
|
|
#include "ft/ftfileprovider.h"
|
|
#include "ft/ftsearch.h"
|
|
#include "util/rsdir.h"
|
|
#include <retroshare/rsturtle.h>
|
|
|
|
/* For Thread Behaviour */
|
|
const uint32_t DMULTIPLEX_MIN = 10; /* 1ms sleep */
|
|
const uint32_t DMULTIPLEX_MAX = 1000; /* 1 sec sleep */
|
|
const double DMULTIPLEX_RELAX = 0.5; /* ??? */
|
|
|
|
static const uint32_t MAX_CHECKING_CHUNK_WAIT_DELAY = 120 ; //! TTL for an inactive chunk
|
|
const uint32_t MAX_SIMULTANEOUS_CRC_REQUESTS = 20 ;
|
|
|
|
/******
|
|
* #define MPLEX_DEBUG 1
|
|
*****/
|
|
|
|
ftClient::ftClient(ftTransferModule *module, ftFileCreator *creator)
|
|
:mModule(module), mCreator(creator)
|
|
{
|
|
return;
|
|
}
|
|
|
|
const uint32_t FT_DATA = 0x0001; // data cuhnk to be stored
|
|
const uint32_t FT_DATA_REQ = 0x0002; // data request to be treated
|
|
const uint32_t FT_CLIENT_CHUNK_MAP_REQ = 0x0003; // chunk map request to be treated by client
|
|
const uint32_t FT_SERVER_CHUNK_MAP_REQ = 0x0004; // chunk map request to be treated by server
|
|
const uint32_t FT_CRC32MAP_REQ = 0x0005; // crc32 map request to be treated by server
|
|
const uint32_t FT_CLIENT_CHUNK_CRC_REQ = 0x0006; // chunk sha1 crc request to be treated
|
|
|
|
ftRequest::ftRequest(uint32_t type, std::string peerId, std::string hash, uint64_t size, uint64_t offset, uint32_t chunk, void *data)
|
|
:mType(type), mPeerId(peerId), mHash(hash), mSize(size),
|
|
mOffset(offset), mChunk(chunk), mData(data)
|
|
{
|
|
return;
|
|
}
|
|
|
|
ftDataMultiplex::ftDataMultiplex(std::string ownId, ftDataSend *server, ftSearch *search)
|
|
:RsQueueThread(DMULTIPLEX_MIN, DMULTIPLEX_MAX, DMULTIPLEX_RELAX), dataMtx("ftDataMultiplex"),
|
|
mDataSend(server), mSearch(search), mOwnId(ownId)
|
|
{
|
|
return;
|
|
}
|
|
|
|
bool ftDataMultiplex::addTransferModule(ftTransferModule *mod, ftFileCreator *f)
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
std::map<std::string, ftClient>::iterator it;
|
|
if (mClients.end() != (it = mClients.find(mod->hash())))
|
|
{
|
|
/* error */
|
|
return false;
|
|
}
|
|
mClients[mod->hash()] = ftClient(mod, f);
|
|
|
|
return true;
|
|
}
|
|
|
|
bool ftDataMultiplex::removeTransferModule(std::string hash)
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
std::map<std::string, ftClient>::iterator it;
|
|
if (mClients.end() == (it = mClients.find(hash)))
|
|
{
|
|
/* error */
|
|
return false;
|
|
}
|
|
mClients.erase(it);
|
|
|
|
// This is very important to delete the hash from servers as well, because
|
|
// after removing the transfer module, ftController will delete the fileCreator.
|
|
// If the file creator is also a server in use, then it will cause a crash
|
|
// at the next server request.
|
|
//
|
|
// With the current action, the next server request will re-create the server as
|
|
// a ftFileProvider.
|
|
//
|
|
std::map<std::string, ftFileProvider*>::iterator sit = mServers.find(hash) ;
|
|
|
|
if(sit != mServers.end())
|
|
mServers.erase(sit);
|
|
|
|
return true;
|
|
}
|
|
|
|
|
|
bool ftDataMultiplex::FileUploads(std::list<std::string> &hashs)
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
std::map<std::string, ftFileProvider *>::iterator sit;
|
|
for(sit = mServers.begin(); sit != mServers.end(); sit++)
|
|
{
|
|
hashs.push_back(sit->first);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool ftDataMultiplex::FileDownloads(std::list<std::string> &hashs)
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
std::map<std::string, ftClient>::iterator cit;
|
|
for(cit = mClients.begin(); cit != mClients.end(); cit++)
|
|
{
|
|
hashs.push_back(cit->first);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
|
|
bool ftDataMultiplex::FileDetails(const std::string &hash, uint32_t hintsflag, FileInfo &info)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::FileDetails(";
|
|
std::cerr << hash << ", " << hintsflag << ")";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
if(hintsflag & RS_FILE_HINTS_DOWNLOAD)
|
|
{
|
|
std::map<std::string, ftClient>::iterator cit;
|
|
if (mClients.end() != (cit = mClients.find(hash)))
|
|
{
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::FileDetails()";
|
|
std::cerr << " Found ftFileCreator!";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
|
|
//(cit->second).mModule->FileDetails(info);
|
|
(cit->second).mCreator->FileDetails(info);
|
|
return true;
|
|
}
|
|
}
|
|
|
|
if(hintsflag & RS_FILE_HINTS_UPLOAD)
|
|
{
|
|
std::map<std::string, ftFileProvider *>::iterator sit;
|
|
sit = mServers.find(hash);
|
|
if (sit != mServers.end())
|
|
{
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::FileDetails()";
|
|
std::cerr << " Found ftFileProvider!";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
|
|
(sit->second)->FileDetails(info);
|
|
return true;
|
|
}
|
|
}
|
|
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::FileDetails()";
|
|
std::cerr << " Found nothing";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
|
|
return false;
|
|
}
|
|
|
|
/* data interface */
|
|
|
|
/*************** SEND INTERFACE (calls ftDataSend) *******************/
|
|
|
|
/* Client Send */
|
|
bool ftDataMultiplex::sendDataRequest(const std::string& peerId, const std::string& hash, uint64_t size, uint64_t offset, uint32_t chunksize)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::sendDataRequest() Client Send";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
return mDataSend->sendDataRequest(peerId,hash,size,offset,chunksize);
|
|
}
|
|
|
|
/* Server Send */
|
|
bool ftDataMultiplex::sendData(const std::string& peerId, const std::string& hash, uint64_t size, uint64_t offset, uint32_t chunksize, void *data)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::sendData() Server Send";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
return mDataSend->sendData(peerId,hash,size,offset,chunksize,data);
|
|
}
|
|
|
|
|
|
/*************** RECV INTERFACE (provides ftDataRecv) ****************/
|
|
|
|
/* Client Recv */
|
|
bool ftDataMultiplex::recvData(const std::string& peerId, const std::string& hash, uint64_t size, uint64_t offset, uint32_t chunksize, void *data)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::recvData() Client Recv";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
/* Store in Queue */
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
mRequestQueue.push_back(ftRequest(FT_DATA,peerId,hash,size,offset,chunksize,data));
|
|
|
|
return true;
|
|
}
|
|
|
|
|
|
/* Server Recv */
|
|
bool ftDataMultiplex::recvDataRequest(const std::string& peerId, const std::string& hash, uint64_t size, uint64_t offset, uint32_t chunksize)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::recvDataRequest() Server Recv";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
/* Store in Queue */
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
mRequestQueue.push_back(
|
|
ftRequest(FT_DATA_REQ,peerId,hash,size,offset,chunksize,NULL));
|
|
|
|
return true;
|
|
}
|
|
|
|
bool ftDataMultiplex::recvChunkMapRequest(const std::string& peerId, const std::string& hash,bool is_client)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::recvChunkMapRequest() Server Recv";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
/* Store in Queue */
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
if(is_client)
|
|
mRequestQueue.push_back(ftRequest(FT_CLIENT_CHUNK_MAP_REQ,peerId,hash,0,0,0,NULL));
|
|
else
|
|
mRequestQueue.push_back(ftRequest(FT_SERVER_CHUNK_MAP_REQ,peerId,hash,0,0,0,NULL));
|
|
|
|
return true;
|
|
}
|
|
|
|
bool ftDataMultiplex::recvCRC32MapRequest(const std::string& peerId, const std::string& hash)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::recvChunkMapRequest() Server Recv";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
/* Store in Queue */
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
mRequestQueue.push_back(ftRequest(FT_CRC32MAP_REQ,peerId,hash,0,0,0,NULL));
|
|
|
|
return true;
|
|
}
|
|
bool ftDataMultiplex::recvSingleChunkCrcRequest(const std::string& peerId, const std::string& hash,uint32_t chunk_number)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::recvChunkMapRequest() Server Recv";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
/* Store in Queue */
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
mRequestQueue.push_back(ftRequest(FT_CLIENT_CHUNK_CRC_REQ,peerId,hash,0,0,chunk_number,NULL));
|
|
|
|
return true;
|
|
}
|
|
class CRC32Thread: public RsThread
|
|
{
|
|
public:
|
|
CRC32Thread(ftDataMultiplex *dataplex,const std::string& peerId,const std::string& hash)
|
|
: _plex(dataplex),_finished(false),_peerId(peerId),_hash(hash) {}
|
|
|
|
virtual void run()
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "CRC32Thread is running for file " << _hash << std::endl;
|
|
#endif
|
|
_plex->computeAndSendCRC32Map(_peerId,_hash) ;
|
|
_finished = true ;
|
|
}
|
|
bool finished() { return _finished ; }
|
|
private:
|
|
ftDataMultiplex *_plex ;
|
|
bool _finished ;
|
|
std::string _peerId ;
|
|
std::string _hash ;
|
|
};
|
|
|
|
/*********** BACKGROUND THREAD OPERATIONS ***********/
|
|
bool ftDataMultiplex::workQueued()
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
if (mRequestQueue.size() > 0)
|
|
{
|
|
return true;
|
|
}
|
|
|
|
if (mSearchQueue.size() > 0)
|
|
{
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool ftDataMultiplex::doWork()
|
|
{
|
|
bool doRequests = true;
|
|
time_t now = time(NULL) ;
|
|
|
|
/* Handle All the current Requests */
|
|
while(doRequests)
|
|
{
|
|
ftRequest req;
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
if (mRequestQueue.size() == 0)
|
|
{
|
|
doRequests = false;
|
|
continue;
|
|
}
|
|
|
|
req = mRequestQueue.front();
|
|
mRequestQueue.pop_front();
|
|
}
|
|
|
|
/* MUTEX FREE */
|
|
|
|
switch(req.mType)
|
|
{
|
|
case FT_DATA:
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::doWork() Handling FT_DATA";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
handleRecvData(req.mPeerId, req.mHash, req.mSize, req.mOffset, req.mChunk, req.mData);
|
|
break;
|
|
|
|
case FT_DATA_REQ:
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::doWork() Handling FT_DATA_REQ";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
handleRecvDataRequest(req.mPeerId, req.mHash, req.mSize, req.mOffset, req.mChunk);
|
|
break;
|
|
|
|
case FT_CLIENT_CHUNK_MAP_REQ:
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::doWork() Handling FT_CLIENT_CHUNK_MAP_REQ";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
handleRecvClientChunkMapRequest(req.mPeerId,req.mHash) ;
|
|
break ;
|
|
|
|
case FT_SERVER_CHUNK_MAP_REQ:
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::doWork() Handling FT_CLIENT_CHUNK_MAP_REQ";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
handleRecvServerChunkMapRequest(req.mPeerId,req.mHash) ;
|
|
break ;
|
|
|
|
case FT_CRC32MAP_REQ:
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::doWork() Handling FT_CLIENT_CRC32_MAP_REQ";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
handleRecvCRC32MapRequest(req.mPeerId,req.mHash) ;
|
|
break ;
|
|
|
|
case FT_CLIENT_CHUNK_CRC_REQ:
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::doWork() Handling FT_CLIENT_CHUNK_CRC_REQ";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
handleRecvChunkCrcRequest(req.mPeerId,req.mHash,req.mChunk) ;
|
|
break ;
|
|
|
|
default:
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::doWork() Ignoring UNKNOWN";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
break;
|
|
}
|
|
}
|
|
|
|
// Look for potentially finished CRC32Map threads, and destroys them.
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
for(std::list<CRC32Thread*>::iterator lit(_crc32map_threads.begin());lit!=_crc32map_threads.end();)
|
|
if((*lit)->finished())
|
|
{
|
|
std::cerr << "ftDataMultiplex::doWork: thread " << *lit << " ended. Deleting it." << std::endl;
|
|
(*lit)->join() ;
|
|
delete (*lit) ;
|
|
std::list<CRC32Thread*>::iterator tmp(lit) ;
|
|
++lit ;
|
|
_crc32map_threads.erase(tmp) ;
|
|
}
|
|
else
|
|
{
|
|
std::cerr << "ftDataMultiplex::doWork: thread " << *lit << " still working. Not quitting it." << std::endl;
|
|
++lit ;
|
|
}
|
|
|
|
// Take the opportunity to cleanup the list, so that it cannot grow indefinitely
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::doWork: Cleaning up list of cached maps." << std::endl ;
|
|
#endif
|
|
|
|
// Keep CRC32 maps in cache for 30 mins max.
|
|
//
|
|
for(std::map<std::string,std::pair<time_t,CRC32Map> >::iterator it = _cached_crc32maps.begin();it!=_cached_crc32maps.end();)
|
|
if(it->second.first + 30*60 < now)
|
|
{
|
|
std::cerr << "Removing cached map for file " << it->first << " that was kept for too long now." << std::endl;
|
|
|
|
std::map<std::string,std::pair<time_t,CRC32Map> >::iterator tmp(it) ;
|
|
++it ;
|
|
_cached_crc32maps.erase(tmp) ;
|
|
}
|
|
else
|
|
++it ;
|
|
}
|
|
|
|
/* Only Handle One Search Per Period....
|
|
* Lower Priority
|
|
*/
|
|
ftRequest req;
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
if (mSearchQueue.size() == 0)
|
|
{
|
|
/* Finished */
|
|
return true;
|
|
}
|
|
|
|
req = mSearchQueue.front();
|
|
mSearchQueue.pop_front();
|
|
}
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::doWork() Handling Search Request";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
if(handleSearchRequest(req.mPeerId, req.mHash))
|
|
handleRecvDataRequest(req.mPeerId, req.mHash, req.mSize, req.mOffset, req.mChunk) ;
|
|
|
|
|
|
return true;
|
|
}
|
|
|
|
bool ftDataMultiplex::recvSingleChunkCrc(const std::string& peerId, const std::string& hash,uint32_t chunk_number,const Sha1CheckSum& crc)
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::recvSingleChunkCrc() Received crc of file " << hash << ", from peer id " << peerId << ", chunk " << chunk_number << ", crc=" << crc.toStdString() << std::endl;
|
|
#endif
|
|
|
|
std::map<std::string, ftClient>::iterator it = mClients.find(hash);
|
|
|
|
if(it == mClients.end())
|
|
{
|
|
std::cerr << "ftDataMultiplex::recvSingleChunkCrc() ERROR: No matching Client for CRC. This is an error. " << hash << " !" << std::endl;
|
|
/* error */
|
|
return false;
|
|
}
|
|
|
|
// store in the cache as well
|
|
|
|
Sha1CacheEntry& sha1cache(_cached_sha1maps[hash]) ;
|
|
|
|
if(sha1cache._map.size() == 0)
|
|
sha1cache._map = Sha1Map(it->second.mCreator->fileSize(),ChunkMap::CHUNKMAP_FIXED_CHUNK_SIZE) ;
|
|
|
|
sha1cache._map.set(chunk_number,crc) ;
|
|
|
|
// remove this chunk from the request list as well.
|
|
|
|
std::map<uint32_t,ChunkCheckSumSourceList>::iterator it2(sha1cache._to_ask.find(chunk_number)) ;
|
|
|
|
if(it2 != sha1cache._to_ask.end())
|
|
sha1cache._to_ask.erase(it2) ;
|
|
|
|
sha1cache._received.push_back(chunk_number) ;
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::recvSingleChunkCrc() stored in cache. " << std::endl;
|
|
#endif
|
|
|
|
return true ;
|
|
}
|
|
|
|
bool ftDataMultiplex::dispatchReceivedChunkCheckSum()
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
uint32_t MAX_CHECKSUM_CHECK_PER_FILE = 25 ;
|
|
|
|
for(std::map<std::string,Sha1CacheEntry>::iterator it(_cached_sha1maps.begin());it!=_cached_sha1maps.end();++it)
|
|
{
|
|
ftFileCreator *client = NULL ;
|
|
|
|
for(uint32_t n=0;n<MAX_CHECKSUM_CHECK_PER_FILE && n < it->second._received.size();)
|
|
{
|
|
if(client == NULL)
|
|
{
|
|
std::map<std::string, ftClient>::iterator itc = mClients.find(it->first);
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::dispatchReceivedChunkCheckSum(): treating hash " << it->first << std::endl;
|
|
#endif
|
|
|
|
if(itc == mClients.end())
|
|
{
|
|
std::cerr << "ftDataMultiplex::dispatchReceivedChunkCheckSum() ERROR: No matching Client for hash. This is an error. Hash=" << it->first << std::endl;
|
|
/* error */
|
|
break ;
|
|
}
|
|
else
|
|
client = itc->second.mCreator ;
|
|
}
|
|
int chunk_number = it->second._received[n] ;
|
|
|
|
if(!it->second._map.isSet(chunk_number))
|
|
{
|
|
std::cerr << "ftDataMultiplex::dispatchReceivedChunkCheckSum() ERROR: chunk " << chunk_number << " is supposed to be initialized but it was not received !!" << std::endl;
|
|
++n ;
|
|
continue ;
|
|
}
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::dispatchReceivedChunkCheckSum(): checking chunk " << chunk_number << " with hash " << it->second._map[chunk_number].toStdString() << std::endl;
|
|
#endif
|
|
client->verifyChunk(chunk_number,it->second._map[chunk_number]) ;
|
|
|
|
it->second._received[n] = it->second._received.back() ;
|
|
it->second._received.pop_back() ;
|
|
}
|
|
}
|
|
return true ;
|
|
}
|
|
|
|
bool ftDataMultiplex::recvCRC32Map(const std::string& /*peerId*/, const std::string& hash,const CRC32Map& crc_map)
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
std::map<std::string, ftClient>::iterator it = mClients.find(hash);
|
|
|
|
if(it == mClients.end())
|
|
{
|
|
std::cerr << "ftDataMultiplex::recvCRCMap() ERROR: No matching Client for CRC32map. This is an error. " << hash << " !" << std::endl;
|
|
/* error */
|
|
return false;
|
|
}
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::recvCRCMap() Passing crc map of file " << hash << ", to FT Module" << std::endl;
|
|
#endif
|
|
|
|
(it->second).mModule->addCRC32Map(crc_map);
|
|
return true ;
|
|
}
|
|
|
|
// A chunk map has arrived. It can be two different situations:
|
|
// - an uploader has sent his chunk map, so we need to store it in the corresponding ftFileProvider
|
|
// - a source for a download has sent his chunk map, so we need to send it to the corresponding ftFileCreator.
|
|
//
|
|
bool ftDataMultiplex::recvChunkMap(const std::string& peerId, const std::string& hash,const CompressedChunkMap& compressed_map,bool client)
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
if(client) // is the chunk map for a client, or for a server ?
|
|
{
|
|
std::map<std::string, ftClient>::iterator it = mClients.find(hash);
|
|
|
|
if(it == mClients.end())
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::recvChunkMap() ERROR: No matching Client for hash " << hash << " !";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
/* error */
|
|
return false;
|
|
}
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::recvChunkMap() Passing map of file " << hash << ", to FT Module";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
|
|
(it->second).mCreator->setSourceMap(peerId, compressed_map);
|
|
return true ;
|
|
}
|
|
else
|
|
{
|
|
std::map<std::string, ftFileProvider *>::iterator it = mServers.find(hash) ;
|
|
|
|
if(it == mServers.end())
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvChunkMap() ERROR: No matching file Provider for hash " << hash ;
|
|
std::cerr << std::endl;
|
|
#endif
|
|
return false;
|
|
}
|
|
|
|
it->second->setClientMap(peerId, compressed_map);
|
|
return true ;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
|
|
bool ftDataMultiplex::handleRecvCRC32MapRequest(const std::string& peerId, const std::string& hash)
|
|
{
|
|
bool found = false ;
|
|
CRC32Map cmap ;
|
|
|
|
// 1 - look into cache
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvChunkMapReq() : source " << peerId << " asked for CRC32 map for file " << hash << std::endl;
|
|
#endif
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
std::map<std::string,std::pair<time_t,CRC32Map> >::iterator it = _cached_crc32maps.find(hash) ;
|
|
|
|
if(it != _cached_crc32maps.end())
|
|
{
|
|
cmap = it->second.second ;
|
|
it->second.first = time(NULL) ; // update time stamp
|
|
found = true ;
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvChunkMapReq() : CRC32 map found in cache !!" << std::endl;
|
|
#endif
|
|
|
|
}
|
|
}
|
|
|
|
if(found)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "File CRC32 map was obtained successfully. Sending it." << std::endl ;
|
|
#endif
|
|
|
|
mDataSend->sendCRC32Map(peerId,hash,cmap);
|
|
return true ;
|
|
}
|
|
else
|
|
{
|
|
std::cerr << "File CRC32 Not found. Computing it." << std::endl ;
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
if(_crc32map_threads.size() > 1)
|
|
{
|
|
std::cerr << "Too many threads already computing CRC32Maps (2 is the current maximum)! Giving up." << std::endl;
|
|
return false ;
|
|
}
|
|
}
|
|
|
|
CRC32Thread *thread = new CRC32Thread(this,peerId,hash);
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
_crc32map_threads.push_back(thread) ;
|
|
}
|
|
thread->start() ;
|
|
return true ;
|
|
}
|
|
}
|
|
|
|
bool ftDataMultiplex::computeAndSendCRC32Map(const std::string& peerId, const std::string& hash)
|
|
{
|
|
bool found ;
|
|
std::map<std::string, ftFileProvider *>::iterator it ;
|
|
std::string filename ;
|
|
uint64_t filesize =0;
|
|
|
|
// 1 - look into the list of servers
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
it = mServers.find(hash) ;
|
|
|
|
if(it == mServers.end())
|
|
found = false ;
|
|
}
|
|
|
|
// 2 - if not found, create a server.
|
|
//
|
|
if(!found)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvChunkMapReq() ERROR: No matching file Provider for hash " << hash ;
|
|
std::cerr << std::endl;
|
|
#endif
|
|
if(!handleSearchRequest(peerId,hash))
|
|
return false ;
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvChunkMapReq() A new file Provider has been made up for hash " << hash ;
|
|
std::cerr << std::endl;
|
|
#endif
|
|
}
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
it = mServers.find(hash) ;
|
|
|
|
if(it == mServers.end()) // handleSearchRequest should have filled mServers[hash], but we have been off-mutex since,
|
|
{
|
|
std::cerr << "Could definitely not find a provider for file " << hash << ". Maybe the file does not exist?" << std::endl;
|
|
return false ; // so it's safer to check again.
|
|
}
|
|
else
|
|
{
|
|
filesize = it->second->fileSize() ;
|
|
filename = it->second->fileName() ;
|
|
}
|
|
}
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "Computing CRC32Map for file " << filename << ", hash=" << hash << ", size=" << filesize << std::endl;
|
|
#endif
|
|
|
|
FILE *fd = RsDirUtil::rs_fopen(filename.c_str(),"rb") ;
|
|
|
|
if(fd == NULL)
|
|
{
|
|
std::cerr << "Could not open file " << filename << " for read!! CRC32Map computation cancelled." << std::endl ;
|
|
return false ;
|
|
}
|
|
|
|
CRC32Map cmap ;
|
|
if(!RsDirUtil::crc32File(fd,filesize,ChunkMap::CHUNKMAP_FIXED_CHUNK_SIZE,cmap))
|
|
{
|
|
std::cerr << "CRC32Map computation failed." << std::endl ;
|
|
fclose(fd) ;
|
|
return false ;
|
|
}
|
|
fclose(fd) ;
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
std::cerr << "File CRC32 was successfully computed. Storing it into cache." << std::endl ;
|
|
|
|
_cached_crc32maps[hash] = std::pair<time_t,CRC32Map>(time(NULL),cmap) ;
|
|
}
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "File CRC32 was successfully computed. Sending it." << std::endl ;
|
|
#endif
|
|
mDataSend->sendCRC32Map(peerId,hash,cmap);
|
|
|
|
return true ;
|
|
}
|
|
|
|
bool ftDataMultiplex::handleRecvClientChunkMapRequest(const std::string& peerId, const std::string& hash)
|
|
{
|
|
CompressedChunkMap cmap ;
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
std::map<std::string, ftClient>::iterator it = mClients.find(hash);
|
|
|
|
if(it == mClients.end())
|
|
{
|
|
// If we can't find the client, it's not a problem. Chunk maps from
|
|
// clients are not essential, as they are only used for display.
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvServerChunkMapRequest() ERROR: No matching Client for hash " << hash ;
|
|
std::cerr << ". Performing local search." << std::endl;
|
|
#endif
|
|
return false;
|
|
}
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvServerChunkMapRequest() Sending map of file " << hash << ", to peer " << peerId << std::endl;
|
|
#endif
|
|
|
|
(it->second).mCreator->getAvailabilityMap(cmap);
|
|
}
|
|
|
|
mDataSend->sendChunkMap(peerId,hash,cmap,false);
|
|
|
|
return true ;
|
|
}
|
|
|
|
bool ftDataMultiplex::handleRecvChunkCrcRequest(const std::string& peerId, const std::string& hash, uint32_t chunk_number)
|
|
{
|
|
// look into the sha1sum cache
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvChunkMapReq() looking for chunk " << chunk_number << " for hash " << hash << std::endl;
|
|
#endif
|
|
|
|
Sha1CheckSum crc ;
|
|
bool found = false ;
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
Sha1CacheEntry& sha1cache(_cached_sha1maps[hash]) ;
|
|
sha1cache.last_activity = time(NULL) ; // update time_stamp
|
|
|
|
if(sha1cache._map.size() > 0 && sha1cache._map.isSet(chunk_number))
|
|
{
|
|
crc = sha1cache._map[chunk_number] ;
|
|
found = true ;
|
|
}
|
|
}
|
|
|
|
if(found)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvChunkMapReq() found in cache ! Sending " << crc.toStdString() << std::endl;
|
|
#endif
|
|
mDataSend->sendSingleChunkCRC(peerId,hash,chunk_number,crc);
|
|
return true ;
|
|
}
|
|
|
|
std::map<std::string, ftFileProvider *>::iterator it ;
|
|
std::string filename ;
|
|
uint64_t filesize =0;
|
|
found = true ;
|
|
|
|
// 1 - look into the list of servers.Not clients ! Clients dont' have verified data.
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
it = mServers.find(hash) ;
|
|
|
|
if(it == mServers.end())
|
|
found = false ;
|
|
}
|
|
|
|
// 2 - if not found, create a server.
|
|
//
|
|
if(!found)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvChunkMapReq() ERROR: No matching file Provider for hash " << hash ;
|
|
std::cerr << std::endl;
|
|
#endif
|
|
if(!handleSearchRequest(peerId,hash))
|
|
return false ;
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvChunkMapReq() A new file Provider has been made up for hash " << hash ;
|
|
std::cerr << std::endl;
|
|
#endif
|
|
}
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
it = mServers.find(hash) ;
|
|
|
|
if(it == mServers.end()) // handleSearchRequest should have filled mServers[hash], but we have been off-mutex since,
|
|
{
|
|
std::cerr << "Could definitely not find a provider for file " << hash << ". Maybe the file does not exist?" << std::endl;
|
|
return false ; // so it's safer to check again.
|
|
}
|
|
else
|
|
{
|
|
filesize = it->second->fileSize() ;
|
|
filename = it->second->fileName() ;
|
|
}
|
|
}
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "Computing Sha1 for chunk " << chunk_number<< " of file " << filename << ", hash=" << hash << ", size=" << filesize << std::endl;
|
|
#endif
|
|
|
|
unsigned char *buf = new unsigned char[ChunkMap::CHUNKMAP_FIXED_CHUNK_SIZE] ;
|
|
FILE *fd = RsDirUtil::rs_fopen(filename.c_str(),"rb") ;
|
|
|
|
if(fd == NULL)
|
|
{
|
|
std::cerr << "Cannot read file " << filename << ". Something's wrong!" << std::endl;
|
|
delete buf ;
|
|
return false ;
|
|
}
|
|
uint32_t len ;
|
|
if(fseeko64(fd,(uint64_t)chunk_number * (uint64_t)ChunkMap::CHUNKMAP_FIXED_CHUNK_SIZE,SEEK_SET)!=0 || 0==(len = fread(buf,1,ChunkMap::CHUNKMAP_FIXED_CHUNK_SIZE,fd)))
|
|
{
|
|
std::cerr << "Cannot fseek/read from file " << filename << " at position " << (uint64_t)chunk_number * (uint64_t)ChunkMap::CHUNKMAP_FIXED_CHUNK_SIZE << std::endl;
|
|
fclose(fd) ;
|
|
}
|
|
fclose(fd) ;
|
|
|
|
crc = RsDirUtil::sha1sum(buf,len) ;
|
|
|
|
// update cache
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
Sha1CacheEntry& sha1cache(_cached_sha1maps[hash]) ;
|
|
|
|
if(sha1cache._map.size() == 0)
|
|
sha1cache._map = Sha1Map(filesize,ChunkMap::CHUNKMAP_FIXED_CHUNK_SIZE) ;
|
|
|
|
sha1cache._map.set(chunk_number,crc) ;
|
|
}
|
|
std::cerr << "Sending CRC of chunk " << chunk_number<< " of file " << filename << ", hash=" << hash << ", size=" << filesize << ", crc=" << crc.toStdString() << std::endl;
|
|
|
|
mDataSend->sendSingleChunkCRC(peerId,hash,chunk_number,crc);
|
|
return true ;
|
|
}
|
|
|
|
bool ftDataMultiplex::handleRecvServerChunkMapRequest(const std::string& peerId, const std::string& hash)
|
|
{
|
|
CompressedChunkMap cmap ;
|
|
std::map<std::string, ftFileProvider *>::iterator it ;
|
|
bool found = true ;
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
it = mServers.find(hash) ;
|
|
|
|
if(it == mServers.end())
|
|
found = false ;
|
|
}
|
|
|
|
if(!found)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvChunkMapReq() ERROR: No matching file Provider for hash " << hash ;
|
|
std::cerr << std::endl;
|
|
#endif
|
|
if(!handleSearchRequest(peerId,hash))
|
|
return false ;
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvChunkMapReq() A new file Provider has been made up for hash " << hash ;
|
|
std::cerr << std::endl;
|
|
#endif
|
|
}
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
it = mServers.find(hash) ;
|
|
|
|
if(it == mServers.end()) // handleSearchRequest should have filled mServers[hash], but we have been off-mutex since,
|
|
return false ; // so it's safer to check again.
|
|
else
|
|
it->second->getAvailabilityMap(cmap);
|
|
}
|
|
|
|
mDataSend->sendChunkMap(peerId,hash,cmap,true);
|
|
|
|
return true;
|
|
}
|
|
|
|
bool ftDataMultiplex::handleRecvData(const std::string& peerId,
|
|
const std::string& hash, uint64_t /*size*/,
|
|
uint64_t offset, uint32_t chunksize, void *data)
|
|
{
|
|
ftTransferModule *transfer_module = NULL ;
|
|
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
std::map<std::string, ftClient>::iterator it;
|
|
if (mClients.end() == (it = mClients.find(hash)))
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvData() ERROR: No matching Client!";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
/* error */
|
|
return false;
|
|
}
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvData() Passing to Module";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
|
|
transfer_module = (it->second).mModule ;
|
|
}
|
|
transfer_module->recvFileData(peerId, offset, chunksize, data);
|
|
|
|
return true;
|
|
}
|
|
|
|
|
|
/* called by ftTransferModule */
|
|
bool ftDataMultiplex::handleRecvDataRequest(const std::string& peerId, const std::string& hash, uint64_t size, uint64_t offset, uint32_t chunksize)
|
|
{
|
|
/**** Find Files *****/
|
|
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
std::map<std::string, ftClient>::iterator cit;
|
|
if (mOwnId == peerId)
|
|
{
|
|
/* own requests must be passed to Servers */
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvData() OwnId, so skip Clients...";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
}
|
|
else if (mClients.end() != (cit = mClients.find(hash)))
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvData() Matched to a Client.";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
locked_handleServerRequest((cit->second).mCreator, peerId, hash, size, offset, chunksize);
|
|
return true;
|
|
}
|
|
|
|
std::map<std::string, ftFileProvider *>::iterator sit;
|
|
if (mServers.end() != (sit = mServers.find(hash)))
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvData() Matched to a Provider.";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
locked_handleServerRequest(sit->second, peerId, hash, size, offset, chunksize);
|
|
return true;
|
|
}
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleRecvData() No Match... adding to Search Queue.";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
|
|
/* Add to Search Queue */
|
|
mSearchQueue.push_back( ftRequest(FT_DATA_REQ, peerId, hash, size, offset, chunksize, NULL));
|
|
|
|
return true;
|
|
}
|
|
|
|
bool ftDataMultiplex::locked_handleServerRequest(ftFileProvider *provider,
|
|
std::string peerId, std::string hash, uint64_t size,
|
|
uint64_t offset, uint32_t chunksize)
|
|
{
|
|
if(chunksize > std::min(size,uint64_t(10*1024*1024)))
|
|
{
|
|
std::cerr << "Warning: peer " << peerId << " is asking a large chunk (s=" << chunksize << ") for hash " << hash << ", filesize=" << size << ". This is unexpected." << std::endl ;
|
|
return false ;
|
|
}
|
|
void *data = malloc(chunksize);
|
|
|
|
if(data == NULL)
|
|
{
|
|
std::cerr << "WARNING: Could not allocate data for a chunksize of " << chunksize << std::endl ;
|
|
return false ;
|
|
}
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::locked_handleServerRequest()";
|
|
std::cerr << "\t peer: " << peerId << " hash: " << hash;
|
|
std::cerr << " size: " << size;
|
|
std::cerr << std::endl;
|
|
std::cerr << "\t offset: " << offset;
|
|
std::cerr << " chunksize: " << chunksize << " data: " << data;
|
|
std::cerr << std::endl;
|
|
#endif
|
|
|
|
if (provider->getFileData(peerId,offset, chunksize, data))
|
|
{
|
|
/* send data out */
|
|
sendData(peerId, hash, size, offset, chunksize, data);
|
|
return true;
|
|
}
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::locked_handleServerRequest()";
|
|
std::cerr << " FAILED";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
free(data);
|
|
|
|
return false;
|
|
}
|
|
|
|
bool ftDataMultiplex::getClientChunkMap(const std::string& upload_hash,const std::string& peerId,CompressedChunkMap& cmap)
|
|
{
|
|
bool too_old = false;
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
std::map<std::string,ftFileProvider *>::iterator sit = mServers.find(upload_hash);
|
|
|
|
if(mServers.end() == sit)
|
|
return false ;
|
|
|
|
sit->second->getClientMap(peerId,cmap,too_old) ;
|
|
}
|
|
|
|
// If the map is too old then we should ask an other map to the peer.
|
|
//
|
|
if(too_old)
|
|
sendChunkMapRequest(peerId,upload_hash,true);
|
|
|
|
return true ;
|
|
}
|
|
|
|
bool ftDataMultiplex::sendChunkMapRequest(const std::string& peer_id,const std::string& hash,bool is_client)
|
|
{
|
|
return mDataSend->sendChunkMapRequest(peer_id,hash,is_client);
|
|
}
|
|
bool ftDataMultiplex::sendCRC32MapRequest(const std::string& peer_id,const std::string& hash)
|
|
{
|
|
return mDataSend->sendCRC32MapRequest(peer_id,hash);
|
|
}
|
|
bool ftDataMultiplex::sendSingleChunkCRCRequests(const std::string& hash, const std::vector<std::pair<uint32_t,std::list<std::string> > >& to_ask)
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
// Put all requested chunks in the request queue.
|
|
|
|
Sha1CacheEntry& ce(_cached_sha1maps[hash]) ;
|
|
|
|
for(uint32_t i=0;i<to_ask.size();++i)
|
|
{
|
|
ChunkCheckSumSourceList& list(ce._to_ask[to_ask[i].first]) ;
|
|
|
|
for(std::list<std::string>::const_iterator it(to_ask[i].second.begin());it!=to_ask[i].second.end();++it)
|
|
list[*it] = 0 ;
|
|
}
|
|
return true ;
|
|
}
|
|
|
|
void ftDataMultiplex::handlePendingCrcRequests()
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
time_t now = time(NULL) ;
|
|
uint32_t n=0 ;
|
|
|
|
for(std::map<std::string,Sha1CacheEntry>::iterator it(_cached_sha1maps.begin());it!=_cached_sha1maps.end();++it)
|
|
for(std::map<uint32_t,ChunkCheckSumSourceList>::iterator it2(it->second._to_ask.begin());it2!=it->second._to_ask.end();++it2)
|
|
for(std::map<std::string,time_t>::iterator it3(it2->second.begin());it3!=it2->second.end();++it3)
|
|
if(it3->second + MAX_CHECKING_CHUNK_WAIT_DELAY < now) // do nothing, otherwise, ask again
|
|
{
|
|
mDataSend->sendSingleChunkCRCRequest(it3->first,it->first,it2->first);
|
|
it3->second = now ;
|
|
|
|
if(++n > MAX_SIMULTANEOUS_CRC_REQUESTS)
|
|
return ;
|
|
}
|
|
}
|
|
|
|
void ftDataMultiplex::deleteUnusedServers()
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
//scan the uploads list in ftdatamultiplex and delete the items which time out
|
|
time_t now = time(NULL);
|
|
|
|
for(std::map<std::string, ftFileProvider *>::iterator sit(mServers.begin());sit != mServers.end();)
|
|
if(sit->second->purgeOldPeers(now,10))
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::deleteUnusedServers(): provider " << (void*)sit->second << " has no active peers. Removing. Now=" << now << std::endl ;
|
|
#endif
|
|
// We don't delete servers that are clients at the same time !
|
|
if(dynamic_cast<ftFileCreator*>(sit->second) == NULL)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::deleteUnusedServers(): deleting file provider " << (void*)sit->second << std::endl ;
|
|
#endif
|
|
delete sit->second;
|
|
}
|
|
#ifdef MPLEX_DEBUG
|
|
else
|
|
std::cerr << "ftDataMultiplex::deleteUnusedServers(): " << (void*)sit->second << " was not deleted because it's also a file creator." << std::endl ;
|
|
#endif
|
|
|
|
std::map<std::string, ftFileProvider *>::iterator tmp(sit);
|
|
++tmp ;
|
|
|
|
mServers.erase(sit);
|
|
|
|
sit = tmp ;
|
|
}
|
|
else
|
|
++sit ;
|
|
}
|
|
|
|
bool ftDataMultiplex::handleSearchRequest(const std::string& peerId, const std::string& hash)
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleSearchRequest(";
|
|
std::cerr << peerId << ", " << hash << "...)";
|
|
std::cerr << std::endl;
|
|
#endif
|
|
|
|
/*
|
|
* Do Actual search
|
|
* Could be Cache File, Local or Extra
|
|
* (anywhere but remote really)
|
|
*
|
|
* the network wide and browsable flags are needed, otherwise results get filtered.
|
|
* For tunnel creation, the check of browsable/network wide flag is already done, so
|
|
* if we get a file download packet here, the source is already allowed to download it.
|
|
*
|
|
*/
|
|
|
|
FileInfo info;
|
|
uint32_t hintflags = RS_FILE_HINTS_EXTRA | RS_FILE_HINTS_LOCAL | RS_FILE_HINTS_SPEC_ONLY ;
|
|
|
|
if(rsTurtle->isTurtlePeer(peerId))
|
|
hintflags |= RS_FILE_HINTS_NETWORK_WIDE ;
|
|
else
|
|
hintflags |= RS_FILE_HINTS_BROWSABLE | RS_FILE_HINTS_CACHE ;
|
|
|
|
if (mSearch->search(hash, hintflags, info))
|
|
{
|
|
|
|
/* setup a new provider */
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
// We might already have a file provider, if two requests have got stacked in the request queue. So let's
|
|
// check that before.
|
|
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << "ftDataMultiplex::handleSearchRequest(";
|
|
std::cerr << " Found Local File, sharing...";
|
|
#endif
|
|
std::map<std::string,ftFileProvider*>::const_iterator it = mServers.find(hash) ;
|
|
ftFileProvider *provider ;
|
|
|
|
if(it == mServers.end())
|
|
{
|
|
provider = new ftFileProvider(info.path, info.size, hash);
|
|
mServers[hash] = provider;
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << " created new file provider " << (void*)provider << std::endl;
|
|
#endif
|
|
}
|
|
else
|
|
{
|
|
#ifdef MPLEX_DEBUG
|
|
std::cerr << " re-using existing file provider " << (void*)it->second << std::endl;
|
|
#endif
|
|
}
|
|
|
|
return true;
|
|
}
|
|
// Now check wether the required file is actually being downloaded. In such a case,
|
|
// setup the file provider to be the file creator itself. Warning: this server should not
|
|
// be deleted when not used anymore. We need to restrict this to client peers that are
|
|
// not ourself, since the file transfer also handles the local cache traffic (this
|
|
// is something to be changed soon!!)
|
|
//
|
|
|
|
if(peerId != mOwnId)
|
|
{
|
|
RsStackMutex stack(dataMtx); /******* LOCK MUTEX ******/
|
|
|
|
std::map<std::string,ftClient>::const_iterator it(mClients.find(hash)) ;
|
|
|
|
if(it != mClients.end())
|
|
{
|
|
mServers[hash] = it->second.mCreator ;
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
|
|
|
|
|
|
|