mirror of
https://gitee.com/spark-store-project/spark-store
synced 2025-12-15 05:12:04 +08:00
支持图标跟随主题,改进详情页,加入更好的多线程下载
新的下载会重试一个线程上的错误,一个线程崩溃次数过多会转移到队列里等待重新安排,其他的暂时没写
This commit is contained in:
255
src/spkdownload.cpp
Normal file
255
src/spkdownload.cpp
Normal file
@@ -0,0 +1,255 @@
|
||||
|
||||
#include "spkdownload.h"
|
||||
#include "spkutils.h"
|
||||
#include <QEventLoop>
|
||||
|
||||
SpkDownloadMgr::SpkDownloadMgr(QObject *parent)
|
||||
{
|
||||
mDestFolder = CFG->value("download/dir", QDir::homePath() + "/.local/spark-store/downloads")
|
||||
.toString();
|
||||
|
||||
QDir dest(mDestFolder);
|
||||
if(!dest.exists())
|
||||
QDir().mkdir(mDestFolder);
|
||||
|
||||
// Distribution servers
|
||||
QString srvPaths = CFG->value("download/servers", "https://d.store.deepinos.org/").toString();
|
||||
mServers = srvPaths.split(";;");
|
||||
|
||||
mCurrentDownloadId = -1;
|
||||
mActiveWorkerCount = 0;
|
||||
}
|
||||
|
||||
SpkDownloadMgr::~SpkDownloadMgr()
|
||||
{
|
||||
// TODO
|
||||
}
|
||||
|
||||
void SpkDownloadMgr::SetNewServers(QList<QString> servers)
|
||||
{
|
||||
mServers = servers;
|
||||
}
|
||||
|
||||
SpkDownloadMgr::RemoteFileInfo SpkDownloadMgr::GetRemoteFileInfo(QUrl url)
|
||||
{
|
||||
QEventLoop event;
|
||||
QNetworkRequest request;
|
||||
request.setUrl(QUrl(url));
|
||||
|
||||
// Use a HEAD request to get file's actual size.
|
||||
request.setAttribute(QNetworkRequest::FollowRedirectsAttribute, true);
|
||||
QNetworkReply *reply = STORE->SendCustomHeadRequest(request);
|
||||
connect(reply, &QNetworkReply::finished, &event, &QEventLoop::quit);
|
||||
event.exec();
|
||||
|
||||
RemoteFileInfo ret;
|
||||
if(reply->hasRawHeader("Content-Length"))
|
||||
ret.Size = reply->header(QNetworkRequest::ContentLengthHeader).toUInt();
|
||||
ret.SupportPartialDownload = (reply->rawHeader("Accept-Ranges") == QByteArrayLiteral("bytes"));
|
||||
|
||||
reply->deleteLater();
|
||||
return ret;
|
||||
}
|
||||
|
||||
void SpkDownloadMgr::SetDestinationFolder(QString path)
|
||||
{
|
||||
QDir dir(path);
|
||||
if(!dir.exists())
|
||||
{
|
||||
if(!QDir().mkdir(path))
|
||||
{
|
||||
sErrPop(tr("Cannot create download destination folder %1! It is now /tmp/.").arg(path));
|
||||
mDestFolder = "/tmp/";
|
||||
return;
|
||||
}
|
||||
}
|
||||
mDestFolder = path;
|
||||
}
|
||||
|
||||
bool SpkDownloadMgr::StartNewDownload(QString path, int downloadId)
|
||||
{
|
||||
if(mCurrentDownloadId != -1)
|
||||
return false; // Already downloading something
|
||||
|
||||
// Try get the file size first. If one server fails then go to next server
|
||||
RemoteFileInfo info;
|
||||
for(int i = 0; i < mServers.size() && info.Size == -1; i++)
|
||||
{
|
||||
info = GetRemoteFileInfo(mServers[i] + path);
|
||||
// TODO: Mark dead servers as unusable so they don't get scheduled first?
|
||||
}
|
||||
if(info.Size == -1) return false; // If all servers failed then we say it's a failure
|
||||
|
||||
mCurrentRemoteFileInfo = info;
|
||||
mActiveWorkerCount = 0;
|
||||
|
||||
// Create the destination file.
|
||||
mDestFile.close();
|
||||
mDestFile.setFileName(mDestFolder + '/' + SpkUtils::CutFileName(path));
|
||||
if(!mDestFile.open(QFile::ReadWrite))
|
||||
return false;
|
||||
|
||||
mCurrentRemotePath = path;
|
||||
|
||||
// Schedule tasks onto different servers if it's supported
|
||||
if(info.SupportPartialDownload && mServers.size() > 1)
|
||||
{
|
||||
int blockSize = info.Size / mServers.size();
|
||||
int i;
|
||||
for(i = 0; i < mServers.size() - 1; i++)
|
||||
{
|
||||
DownloadWorker worker
|
||||
{
|
||||
.BeginOffset = i * blockSize,
|
||||
.BytesNeeded = blockSize,
|
||||
.BytesRecvd = 0
|
||||
};
|
||||
worker.Reply =
|
||||
STORE->SendDownloadRequest(mServers[i] + path,
|
||||
worker.BeginOffset,
|
||||
worker.BeginOffset + worker.BytesNeeded);
|
||||
worker.Reply->setProperty("workerId", i);
|
||||
mScheduledWorkers.append(worker);
|
||||
}
|
||||
// Last one
|
||||
DownloadWorker worker
|
||||
{
|
||||
.BeginOffset = i * blockSize,
|
||||
.BytesNeeded = info.Size - i * blockSize,
|
||||
.BytesRecvd = 0
|
||||
};
|
||||
worker.Reply =
|
||||
STORE->SendDownloadRequest(mServers[i] + path,
|
||||
worker.BeginOffset,
|
||||
worker.BeginOffset + worker.BytesNeeded);
|
||||
worker.Reply->setProperty("workerId", i);
|
||||
mScheduledWorkers.append(worker);
|
||||
}
|
||||
else
|
||||
{
|
||||
DownloadWorker worker { .BeginOffset = 0, .BytesNeeded = info.Size, .BytesRecvd = 0 };
|
||||
worker.Reply = STORE->SendDownloadRequest(mServers[0] + path);
|
||||
mScheduledWorkers.append(worker);
|
||||
}
|
||||
|
||||
// Link the worker's replies with the manager
|
||||
for(auto &i : mScheduledWorkers)
|
||||
{
|
||||
LinkReplyWithMe(i.Reply);
|
||||
i.Reply->setProperty("failCount", 0); // Used for fail retry algorithm
|
||||
}
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
bool SpkDownloadMgr::PauseCurrentDownload()
|
||||
{
|
||||
// UNIMPLEMENTED
|
||||
return false;
|
||||
}
|
||||
|
||||
bool SpkDownloadMgr::CancelCurrentDownload()
|
||||
{
|
||||
// UNIMPLEMENTED
|
||||
return false;
|
||||
}
|
||||
|
||||
void SpkDownloadMgr::WorkerFinish()
|
||||
{
|
||||
QNetworkReply *reply = static_cast<QNetworkReply*>(sender());
|
||||
int id = reply->property("workerId").toInt();
|
||||
DownloadWorker &worker = mScheduledWorkers[id];
|
||||
|
||||
mActiveWorkerCount--;
|
||||
|
||||
if(reply->error() == QNetworkReply::NetworkError::NoError)
|
||||
{
|
||||
// Finished successfully, destroy associated stuff
|
||||
reply->deleteLater();
|
||||
worker.Reply = nullptr;
|
||||
|
||||
// Try schedule fail retries here.
|
||||
TryScheduleFailureRetries(id);
|
||||
|
||||
// Check if we're finished.
|
||||
if(mActiveWorkerCount == 0)
|
||||
{
|
||||
emit DownloadStopped(Success, mCurrentDownloadId);
|
||||
mScheduledWorkers.clear();
|
||||
mFailureRetryQueue.clear();
|
||||
mCurrentDownloadId = -1;
|
||||
}
|
||||
}
|
||||
else
|
||||
{
|
||||
// Failed here! Update our offset and required bytes count etc.
|
||||
worker.BeginOffset += worker.BytesRecvd;
|
||||
worker.BytesNeeded -= worker.BytesRecvd;
|
||||
worker.BytesRecvd = 0;
|
||||
|
||||
if(reply->property("failCount").toInt() > MaximumThreadRetryCount)
|
||||
{
|
||||
// Failed too many times, this server is probably down or really bad condition.
|
||||
// Schedule it on other servers.
|
||||
reply->deleteLater();
|
||||
worker.Reply = nullptr;
|
||||
mFailureRetryQueue.enqueue(worker);
|
||||
return;
|
||||
}
|
||||
|
||||
// We can still retry.
|
||||
worker.Reply =
|
||||
STORE->SendDownloadRequest(mServers[id] + mCurrentRemotePath,
|
||||
worker.BeginOffset,
|
||||
worker.BeginOffset + worker.BytesNeeded);
|
||||
LinkReplyWithMe(worker.Reply);
|
||||
worker.Reply->setProperty("failCount", reply->property("failCount").toInt() + 1);
|
||||
reply->deleteLater();
|
||||
}
|
||||
}
|
||||
|
||||
void SpkDownloadMgr::WorkerDownloadProgress()
|
||||
{
|
||||
QNetworkReply *reply = static_cast<QNetworkReply*>(sender());
|
||||
DownloadWorker &worker = mScheduledWorkers[reply->property("workerId").toInt()];
|
||||
auto replyData = reply->readAll();
|
||||
|
||||
mDestFile.seek(worker.BeginOffset + worker.BytesRecvd);
|
||||
mDestFile.write(replyData);
|
||||
worker.BytesRecvd += replyData.size();
|
||||
}
|
||||
|
||||
void SpkDownloadMgr::LinkReplyWithMe(QNetworkReply *reply)
|
||||
{
|
||||
mActiveWorkerCount++; // Each time you spin up a request you must do this so it's ok to do it here
|
||||
connect(reply, &QNetworkReply::readyRead, this, &SpkDownloadMgr::WorkerDownloadProgress);
|
||||
connect(reply, &QNetworkReply::finished, this, &SpkDownloadMgr::WorkerFinish);
|
||||
}
|
||||
|
||||
void SpkDownloadMgr::TryScheduleFailureRetries()
|
||||
{
|
||||
if(mFailureRetryQueue.isEmpty())
|
||||
return;
|
||||
|
||||
for(int i = 0; i < mScheduledWorkers.size() && !mFailureRetryQueue.isEmpty(); i++)
|
||||
{
|
||||
TryScheduleFailureRetries(i);
|
||||
}
|
||||
}
|
||||
|
||||
void SpkDownloadMgr::TryScheduleFailureRetries(int i)
|
||||
{
|
||||
if(mFailureRetryQueue.isEmpty())
|
||||
return;
|
||||
if(mScheduledWorkers[i].Reply == nullptr)
|
||||
{
|
||||
// Schedule it here, it has finished its job
|
||||
mScheduledWorkers[i] = mFailureRetryQueue.dequeue();
|
||||
DownloadWorker &worker = mScheduledWorkers[i];
|
||||
mScheduledWorkers[i].Reply =
|
||||
STORE->SendDownloadRequest(mServers[i] + mCurrentRemotePath,
|
||||
worker.BeginOffset,
|
||||
worker.BeginOffset + worker.BytesNeeded);
|
||||
LinkReplyWithMe(mScheduledWorkers[i].Reply);
|
||||
}
|
||||
}
|
||||
Reference in New Issue
Block a user