|
|
|
/*
|
|
|
|
This file is part of KMail, the KDE mail client.
|
|
|
|
Copyright (c) 2000 Don Sanders <sanders@kde.org>
|
|
|
|
|
|
|
|
Based on popaccount by:
|
|
|
|
Stefan Taferner <taferner@kde.org>
|
|
|
|
Markus Wuebben <markus.wuebben@kde.org>
|
|
|
|
|
|
|
|
KMail is free software; you can redistribute it and/or modify it
|
|
|
|
under the terms of the GNU General Public License, version 2, as
|
|
|
|
published by the Free Software Foundation.
|
|
|
|
|
|
|
|
KMail is distributed in the hope that it will be useful, but
|
|
|
|
WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
General Public License for more details.
|
|
|
|
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
|
|
along with this program; if not, write to the Free Software
|
|
|
|
Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
|
|
#include <config.h>
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#include "popaccount.h"
|
|
|
|
|
|
|
|
#include "broadcaststatus.h"
|
|
|
|
using KPIM::BroadcastStatus;
|
|
|
|
#include "progressmanager.h"
|
|
|
|
#include "kmfoldermgr.h"
|
|
|
|
#include "kmfiltermgr.h"
|
|
|
|
#include "kmpopfiltercnfrmdlg.h"
|
|
|
|
#include "protocols.h"
|
|
|
|
#include "kmglobal.h"
|
|
|
|
#include "util.h"
|
|
|
|
#include "accountmanager.h"
|
|
|
|
|
|
|
|
#include <kdebug.h>
|
|
|
|
#include <kstandarddirs.h>
|
|
|
|
#include <klocale.h>
|
|
|
|
#include <kmessagebox.h>
|
|
|
|
#include <kmainwindow.h>
|
|
|
|
#include <kio/scheduler.h>
|
|
|
|
#include <kio/passdlg.h>
|
|
|
|
#include <kconfig.h>
|
|
|
|
using KIO::MetaData;
|
|
|
|
|
|
|
|
#include <tqstylesheet.h>
|
|
|
|
|
|
|
|
static const unsigned short int pop3DefaultPort = 110;
|
|
|
|
|
|
|
|
namespace KMail {
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
PopAccount::PopAccount(AccountManager* aOwner, const TQString& aAccountName, uint id)
|
|
|
|
: NetworkAccount(aOwner, aAccountName, id),
|
|
|
|
headerIt(headersOnServer),
|
|
|
|
processMsgsTimer( 0, "processMsgsTimer" )
|
|
|
|
{
|
|
|
|
init();
|
|
|
|
job = 0;
|
|
|
|
mSlave = 0;
|
|
|
|
mPort = defaultPort();
|
|
|
|
stage = Idle;
|
|
|
|
indexOfCurrentMsg = -1;
|
|
|
|
curMsgStrm = 0;
|
|
|
|
processingDelay = 2*100;
|
|
|
|
mProcessing = false;
|
|
|
|
dataCounter = 0;
|
|
|
|
mUidsOfSeenMsgsDict.setAutoDelete( false );
|
|
|
|
mUidsOfNextSeenMsgsDict.setAutoDelete( false );
|
|
|
|
|
|
|
|
headersOnServer.setAutoDelete(true);
|
|
|
|
connect(&processMsgsTimer,TQT_SIGNAL(timeout()),TQT_SLOT(slotProcessPendingMsgs()));
|
|
|
|
KIO::Scheduler::connect(
|
|
|
|
TQT_SIGNAL(slaveError(KIO::Slave *, int, const TQString &)),
|
|
|
|
this, TQT_SLOT(slotSlaveError(KIO::Slave *, int, const TQString &)));
|
|
|
|
|
|
|
|
mHeaderDeleteUids.clear();
|
|
|
|
mHeaderDownUids.clear();
|
|
|
|
mHeaderLaterUids.clear();
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
PopAccount::~PopAccount()
|
|
|
|
{
|
|
|
|
if (job) {
|
|
|
|
job->kill();
|
|
|
|
mMsgsPendingDownload.clear();
|
|
|
|
processRemainingQueuedMessages();
|
|
|
|
saveUidList();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
TQString PopAccount::type(void) const
|
|
|
|
{
|
|
|
|
return "pop";
|
|
|
|
}
|
|
|
|
|
|
|
|
TQString PopAccount::protocol() const {
|
|
|
|
return useSSL() ? POP_SSL_PROTOCOL : POP_PROTOCOL;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned short int PopAccount::defaultPort() const {
|
|
|
|
return pop3DefaultPort;
|
|
|
|
}
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::init(void)
|
|
|
|
{
|
|
|
|
NetworkAccount::init();
|
|
|
|
|
|
|
|
mUsePipelining = false;
|
|
|
|
mLeaveOnServer = false;
|
|
|
|
mLeaveOnServerDays = -1;
|
|
|
|
mLeaveOnServerCount = -1;
|
|
|
|
mLeaveOnServerSize = -1;
|
|
|
|
mFilterOnServer = false;
|
|
|
|
//tz todo
|
|
|
|
mFilterOnServerCheckSize = 50000;
|
|
|
|
}
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::pseudoAssign( const KMAccount * a ) {
|
|
|
|
slotAbortRequested();
|
|
|
|
NetworkAccount::pseudoAssign( a );
|
|
|
|
|
|
|
|
const PopAccount * p = dynamic_cast<const PopAccount*>( a );
|
|
|
|
if ( !p ) return;
|
|
|
|
|
|
|
|
setUsePipelining( p->usePipelining() );
|
|
|
|
setLeaveOnServer( p->leaveOnServer() );
|
|
|
|
setLeaveOnServerDays( p->leaveOnServerDays() );
|
|
|
|
setLeaveOnServerCount( p->leaveOnServerCount() );
|
|
|
|
setLeaveOnServerSize( p->leaveOnServerSize() );
|
|
|
|
setFilterOnServer( p->filterOnServer() );
|
|
|
|
setFilterOnServerCheckSize( p->filterOnServerCheckSize() );
|
|
|
|
}
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::processNewMail(bool _interactive)
|
|
|
|
{
|
|
|
|
if (stage == Idle) {
|
|
|
|
|
|
|
|
if ( (mAskAgain || passwd().isEmpty() || mLogin.isEmpty()) &&
|
|
|
|
mAuth != "GSSAPI" ) {
|
|
|
|
TQString passwd = NetworkAccount::passwd();
|
|
|
|
bool b = storePasswd();
|
|
|
|
if (KIO::PasswordDialog::getNameAndPassword(mLogin, passwd, &b,
|
|
|
|
i18n("You need to supply a username and a password to access this "
|
|
|
|
"mailbox."), false, TQString(), mName, i18n("Account:"))
|
|
|
|
!= TQDialog::Accepted)
|
|
|
|
{
|
|
|
|
checkDone( false, CheckAborted );
|
|
|
|
return;
|
|
|
|
} else {
|
|
|
|
setPasswd( passwd, b );
|
|
|
|
if ( b ) {
|
|
|
|
kmkernel->acctMgr()->writeConfig( true );
|
|
|
|
}
|
|
|
|
mAskAgain = false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TQString seenUidList = locateLocal( "data", "kmail/" + mLogin + ":" + "@" +
|
|
|
|
mHost + ":" + TQString("%1").tqarg(mPort) );
|
|
|
|
KConfig config( seenUidList );
|
|
|
|
TQStringList uidsOfSeenMsgs = config.readListEntry( "seenUidList" );
|
|
|
|
TQValueList<int> timeOfSeenMsgs = config.readIntListEntry( "seenUidTimeList" );
|
|
|
|
mUidsOfSeenMsgsDict.clear();
|
|
|
|
mUidsOfSeenMsgsDict.resize( KMail::nextPrime( ( uidsOfSeenMsgs.count() * 11 ) / 10 ) );
|
|
|
|
int idx = 1;
|
|
|
|
for ( TQStringList::ConstIterator it = uidsOfSeenMsgs.begin();
|
|
|
|
it != uidsOfSeenMsgs.end(); ++it, idx++ ) {
|
|
|
|
// we use mUidsOfSeenMsgsDict to just provide fast random access to the
|
|
|
|
// keys, so we can store the index(+1) that corresponds to the index of
|
|
|
|
// mTimeOfSeenMsgsVector for use in PopAccount::slotData()
|
|
|
|
mUidsOfSeenMsgsDict.insert( *it, (const int *)idx );
|
|
|
|
}
|
|
|
|
mTimeOfSeenMsgsVector.clear();
|
|
|
|
mTimeOfSeenMsgsVector.reserve( timeOfSeenMsgs.size() );
|
|
|
|
for ( TQValueList<int>::ConstIterator it = timeOfSeenMsgs.begin();
|
|
|
|
it != timeOfSeenMsgs.end(); ++it) {
|
|
|
|
mTimeOfSeenMsgsVector.append( *it );
|
|
|
|
}
|
|
|
|
// If the counts differ then the config file has presumably been tampered
|
|
|
|
// with and so to avoid possible unwanted message deletion we'll treat
|
|
|
|
// them all as newly seen by clearing the seen times vector
|
|
|
|
if ( mTimeOfSeenMsgsVector.count() != mUidsOfSeenMsgsDict.count() )
|
|
|
|
mTimeOfSeenMsgsVector.clear();
|
|
|
|
TQStringList downloadLater = config.readListEntry( "downloadLater" );
|
|
|
|
for ( TQStringList::Iterator it = downloadLater.begin(); it != downloadLater.end(); ++it ) {
|
|
|
|
mHeaderLaterUids.insert( *it, true );
|
|
|
|
}
|
|
|
|
mUidsOfNextSeenMsgsDict.clear();
|
|
|
|
mTimeOfNextSeenMsgsMap.clear();
|
|
|
|
mSizeOfNextSeenMsgsDict.clear();
|
|
|
|
|
|
|
|
interactive = _interactive;
|
|
|
|
mUidlFinished = false;
|
|
|
|
startJob();
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
checkDone( false, CheckIgnored );
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::readConfig(KConfig& config)
|
|
|
|
{
|
|
|
|
NetworkAccount::readConfig(config);
|
|
|
|
|
|
|
|
mUsePipelining = config.readNumEntry("pipelining", false);
|
|
|
|
mLeaveOnServer = config.readNumEntry("leave-on-server", false);
|
|
|
|
mLeaveOnServerDays = config.readNumEntry("leave-on-server-days", -1);
|
|
|
|
mLeaveOnServerCount = config.readNumEntry("leave-on-server-count", -1);
|
|
|
|
mLeaveOnServerSize = config.readNumEntry("leave-on-server-size", -1);
|
|
|
|
mFilterOnServer = config.readNumEntry("filter-on-server", false);
|
|
|
|
mFilterOnServerCheckSize = config.readUnsignedNumEntry("filter-os-check-size", 50000);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::writeConfig(KConfig& config)
|
|
|
|
{
|
|
|
|
NetworkAccount::writeConfig(config);
|
|
|
|
|
|
|
|
config.writeEntry("pipelining", mUsePipelining);
|
|
|
|
config.writeEntry("leave-on-server", mLeaveOnServer);
|
|
|
|
config.writeEntry("leave-on-server-days", mLeaveOnServerDays);
|
|
|
|
config.writeEntry("leave-on-server-count", mLeaveOnServerCount);
|
|
|
|
config.writeEntry("leave-on-server-size", mLeaveOnServerSize);
|
|
|
|
config.writeEntry("filter-on-server", mFilterOnServer);
|
|
|
|
config.writeEntry("filter-os-check-size", mFilterOnServerCheckSize);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::setUsePipelining(bool b)
|
|
|
|
{
|
|
|
|
mUsePipelining = b;
|
|
|
|
}
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::setLeaveOnServer(bool b)
|
|
|
|
{
|
|
|
|
mLeaveOnServer = b;
|
|
|
|
}
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::setLeaveOnServerDays(int days)
|
|
|
|
{
|
|
|
|
mLeaveOnServerDays = days;
|
|
|
|
}
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::setLeaveOnServerCount(int count)
|
|
|
|
{
|
|
|
|
mLeaveOnServerCount = count;
|
|
|
|
}
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::setLeaveOnServerSize(int size)
|
|
|
|
{
|
|
|
|
mLeaveOnServerSize = size;
|
|
|
|
}
|
|
|
|
|
|
|
|
//---------------------------------------------------------------------------
|
|
|
|
void PopAccount::setFilterOnServer(bool b)
|
|
|
|
{
|
|
|
|
mFilterOnServer = b;
|
|
|
|
}
|
|
|
|
|
|
|
|
//---------------------------------------------------------------------------
|
|
|
|
void PopAccount::setFilterOnServerCheckSize(unsigned int aSize)
|
|
|
|
{
|
|
|
|
mFilterOnServerCheckSize = aSize;
|
|
|
|
}
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::connectJob() {
|
|
|
|
KIO::Scheduler::assignJobToSlave(mSlave, job);
|
|
|
|
connect(job, TQT_SIGNAL( data( KIO::Job*, const TQByteArray &)),
|
|
|
|
TQT_SLOT( slotData( KIO::Job*, const TQByteArray &)));
|
|
|
|
connect(job, TQT_SIGNAL( result( KIO::Job * ) ),
|
|
|
|
TQT_SLOT( slotResult( KIO::Job * ) ) );
|
|
|
|
connect(job, TQT_SIGNAL(infoMessage( KIO::Job*, const TQString & )),
|
|
|
|
TQT_SLOT( slotMsgRetrieved(KIO::Job*, const TQString &)));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::slotCancel()
|
|
|
|
{
|
|
|
|
mMsgsPendingDownload.clear();
|
|
|
|
processRemainingQueuedMessages();
|
|
|
|
saveUidList();
|
|
|
|
slotJobFinished();
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::slotProcessPendingMsgs()
|
|
|
|
{
|
|
|
|
if (mProcessing) // not reentrant
|
|
|
|
return;
|
|
|
|
mProcessing = true;
|
|
|
|
|
|
|
|
bool addedOk;
|
|
|
|
TQValueList<KMMessage*>::Iterator cur = msgsAwaitingProcessing.begin();
|
|
|
|
TQStringList::Iterator curId = msgIdsAwaitingProcessing.begin();
|
|
|
|
TQStringList::Iterator curUid = msgUidsAwaitingProcessing.begin();
|
|
|
|
|
|
|
|
while (cur != msgsAwaitingProcessing.end()) {
|
|
|
|
// note we can actually end up processing events in processNewMsg
|
|
|
|
// this happens when send receipts is turned on
|
|
|
|
// hence the check for re-entry at the start of this method.
|
|
|
|
// -sanders Update processNewMsg should no longer process events
|
|
|
|
|
|
|
|
addedOk = processNewMsg(*cur); //added ok? Error displayed if not.
|
|
|
|
|
|
|
|
if (!addedOk) {
|
|
|
|
mMsgsPendingDownload.clear();
|
|
|
|
msgIdsAwaitingProcessing.clear();
|
|
|
|
msgUidsAwaitingProcessing.clear();
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
idsOfMsgsToDelete.append( *curId );
|
|
|
|
mUidsOfNextSeenMsgsDict.insert( *curUid, (const int *)1 );
|
|
|
|
mTimeOfNextSeenMsgsMap.insert( *curUid, time(0) );
|
|
|
|
}
|
|
|
|
++cur;
|
|
|
|
++curId;
|
|
|
|
++curUid;
|
|
|
|
}
|
|
|
|
|
|
|
|
msgsAwaitingProcessing.clear();
|
|
|
|
msgIdsAwaitingProcessing.clear();
|
|
|
|
msgUidsAwaitingProcessing.clear();
|
|
|
|
mProcessing = false;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::slotAbortRequested()
|
|
|
|
{
|
|
|
|
if (stage == Idle) return;
|
|
|
|
if ( mMailCheckProgressItem )
|
|
|
|
disconnect( mMailCheckProgressItem, TQT_SIGNAL( progressItemCanceled( KPIM::ProgressItem* ) ),
|
|
|
|
this, TQT_SLOT( slotAbortRequested() ) );
|
|
|
|
stage = Quit;
|
|
|
|
if (job) job->kill();
|
|
|
|
job = 0;
|
|
|
|
mSlave = 0;
|
|
|
|
slotCancel();
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::startJob()
|
|
|
|
{
|
|
|
|
// Run the precommand
|
|
|
|
if (!runPrecommand(precommand()))
|
|
|
|
{
|
|
|
|
KMessageBox::sorry(0,
|
|
|
|
i18n("Could not execute precommand: %1").tqarg(precommand()),
|
|
|
|
i18n("KMail Error Message"));
|
|
|
|
checkDone( false, CheckError );
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
// end precommand code
|
|
|
|
|
|
|
|
KURL url = getUrl();
|
|
|
|
|
|
|
|
if ( !url.isValid() ) {
|
|
|
|
KMessageBox::error(0, i18n("Source URL is malformed"),
|
|
|
|
i18n("Kioslave Error Message") );
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
mMsgsPendingDownload.clear();
|
|
|
|
idsOfMsgs.clear();
|
|
|
|
mUidForIdMap.clear();
|
|
|
|
idsOfMsgsToDelete.clear();
|
|
|
|
idsOfForcedDeletes.clear();
|
|
|
|
|
|
|
|
//delete any headers if there are some this have to be done because of check again
|
|
|
|
headersOnServer.clear();
|
|
|
|
headers = false;
|
|
|
|
indexOfCurrentMsg = -1;
|
|
|
|
|
|
|
|
Q_ASSERT( !mMailCheckProgressItem );
|
|
|
|
TQString escapedName = TQStyleSheet::escape( mName );
|
|
|
|
mMailCheckProgressItem = KPIM::ProgressManager::createProgressItem(
|
|
|
|
"MailCheck" + mName,
|
|
|
|
escapedName,
|
|
|
|
i18n("Preparing transmission from \"%1\"...").tqarg( escapedName ),
|
|
|
|
true, // can be canceled
|
|
|
|
useSSL() || useTLS() );
|
|
|
|
connect( mMailCheckProgressItem, TQT_SIGNAL( progressItemCanceled( KPIM::ProgressItem* ) ),
|
|
|
|
this, TQT_SLOT( slotAbortRequested() ) );
|
|
|
|
|
|
|
|
numBytes = 0;
|
|
|
|
numBytesRead = 0;
|
|
|
|
stage = List;
|
|
|
|
mSlave = KIO::Scheduler::getConnectedSlave( url, slaveConfig() );
|
|
|
|
if (!mSlave)
|
|
|
|
{
|
|
|
|
slotSlaveError(0, KIO::ERR_CANNOT_LAUNCH_PROCESS, url.protocol());
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
url.setPath(TQString("/index"));
|
|
|
|
job = KIO::get( url, false, false );
|
|
|
|
connectJob();
|
|
|
|
}
|
|
|
|
|
|
|
|
MetaData PopAccount::slaveConfig() const {
|
|
|
|
MetaData m = NetworkAccount::slaveConfig();
|
|
|
|
|
|
|
|
m.insert("progress", "off");
|
|
|
|
m.insert("pipelining", (mUsePipelining) ? "on" : "off");
|
|
|
|
if (mAuth == "PLAIN" || mAuth == "LOGIN" || mAuth == "CRAM-MD5" ||
|
|
|
|
mAuth == "DIGEST-MD5" || mAuth == "NTLM" || mAuth == "GSSAPI") {
|
|
|
|
m.insert("auth", "SASL");
|
|
|
|
m.insert("sasl", mAuth);
|
|
|
|
} else if ( mAuth == "*" )
|
|
|
|
m.insert("auth", "USER");
|
|
|
|
else
|
|
|
|
m.insert("auth", mAuth);
|
|
|
|
|
|
|
|
return m;
|
|
|
|
}
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
// one message is finished
|
|
|
|
// add data to a KMMessage
|
|
|
|
void PopAccount::slotMsgRetrieved(KIO::Job*, const TQString & infoMsg)
|
|
|
|
{
|
|
|
|
if (infoMsg != "message complete") return;
|
|
|
|
KMMessage *msg = new KMMessage;
|
|
|
|
msg->setComplete(true);
|
|
|
|
// Make sure to use LF as line ending to make the processing easier
|
|
|
|
// when piping through external programs
|
|
|
|
uint newSize = Util::crlf2lf( curMsgData.data(), curMsgData.size() );
|
|
|
|
curMsgData.resize( newSize );
|
|
|
|
msg->fromByteArray( curMsgData , true );
|
|
|
|
if (stage == Head)
|
|
|
|
{
|
|
|
|
int size = mMsgsPendingDownload[ headerIt.current()->id() ];
|
|
|
|
kdDebug(5006) << "Size of Message: " << size << endl;
|
|
|
|
msg->setMsgLength( size );
|
|
|
|
headerIt.current()->setHeader(msg);
|
|
|
|
++headerIt;
|
|
|
|
slotGetNextHdr();
|
|
|
|
} else {
|
|
|
|
//kdDebug(5006) << kfuncinfo << "stage == Retr" << endl;
|
|
|
|
//kdDebug(5006) << "curMsgData.size() = " << curMsgData.size() << endl;
|
|
|
|
msg->setMsgLength( curMsgData.size() );
|
|
|
|
msgsAwaitingProcessing.append(msg);
|
|
|
|
msgIdsAwaitingProcessing.append(idsOfMsgs[indexOfCurrentMsg]);
|
|
|
|
msgUidsAwaitingProcessing.append( mUidForIdMap[idsOfMsgs[indexOfCurrentMsg]] );
|
|
|
|
slotGetNextMsg();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
// finit state machine to cycle trow the stages
|
|
|
|
void PopAccount::slotJobFinished() {
|
|
|
|
TQStringList emptyList;
|
|
|
|
if (stage == List) {
|
|
|
|
kdDebug(5006) << k_funcinfo << "stage == List" << endl;
|
|
|
|
// set the initial size of mUidsOfNextSeenMsgsDict to the number of
|
|
|
|
// messages on the server + 10%
|
|
|
|
mUidsOfNextSeenMsgsDict.resize( KMail::nextPrime( ( idsOfMsgs.count() * 11 ) / 10 ) );
|
|
|
|
KURL url = getUrl();
|
|
|
|
url.setPath(TQString("/uidl"));
|
|
|
|
job = KIO::get( url, false, false );
|
|
|
|
connectJob();
|
|
|
|
stage = Uidl;
|
|
|
|
}
|
|
|
|
else if (stage == Uidl) {
|
|
|
|
kdDebug(5006) << k_funcinfo << "stage == Uidl" << endl;
|
|
|
|
mUidlFinished = true;
|
|
|
|
|
|
|
|
if ( mLeaveOnServer && mUidForIdMap.isEmpty() &&
|
|
|
|
mUidsOfNextSeenMsgsDict.isEmpty() && !idsOfMsgs.isEmpty() ) {
|
|
|
|
KMessageBox::sorry(0, i18n("Your POP3 server (Account: %1) does not support "
|
|
|
|
"the UIDL command: this command is required to determine, in a reliable way, "
|
|
|
|
"which of the mails on the server KMail has already seen before;\n"
|
|
|
|
"the feature to leave the mails on the server will therefore not "
|
|
|
|
"work properly.").tqarg(NetworkAccount::name()) );
|
|
|
|
// An attempt to work around buggy pop servers, these seem to be popular.
|
|
|
|
mUidsOfNextSeenMsgsDict = mUidsOfSeenMsgsDict;
|
|
|
|
}
|
|
|
|
|
|
|
|
//check if filter on server
|
|
|
|
if (mFilterOnServer == true) {
|
|
|
|
TQMap<TQString, int>::Iterator hids;
|
|
|
|
for ( hids = mMsgsPendingDownload.begin();
|
|
|
|
hids != mMsgsPendingDownload.end(); hids++ ) {
|
|
|
|
kdDebug(5006) << "Length: " << hids.data() << endl;
|
|
|
|
//check for mails bigger mFilterOnServerCheckSize
|
|
|
|
if ( (unsigned int)hids.data() >= mFilterOnServerCheckSize ) {
|
|
|
|
kdDebug(5006) << "bigger than " << mFilterOnServerCheckSize << endl;
|
|
|
|
headersOnServer.append(new KMPopHeaders( hids.key(),
|
|
|
|
mUidForIdMap[hids.key()],
|
|
|
|
Later));//TODO
|
|
|
|
//set Action if already known
|
|
|
|
if( mHeaderDeleteUids.contains( headersOnServer.current()->uid() ) ) {
|
|
|
|
headersOnServer.current()->setAction(Delete);
|
|
|
|
}
|
|
|
|
else if( mHeaderDownUids.contains( headersOnServer.current()->uid() ) ) {
|
|
|
|
headersOnServer.current()->setAction(Down);
|
|
|
|
}
|
|
|
|
else if( mHeaderLaterUids.contains( headersOnServer.current()->uid() ) ) {
|
|
|
|
headersOnServer.current()->setAction(Later);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// delete the uids so that you don't get them twice in the list
|
|
|
|
mHeaderDeleteUids.clear();
|
|
|
|
mHeaderDownUids.clear();
|
|
|
|
mHeaderLaterUids.clear();
|
|
|
|
}
|
|
|
|
// kdDebug(5006) << "Num of Msgs to Filter: " << headersOnServer.count() << endl;
|
|
|
|
// if there are mails which should be checkedc download the headers
|
|
|
|
if ((headersOnServer.count() > 0) && (mFilterOnServer == true)) {
|
|
|
|
headerIt.toFirst();
|
|
|
|
KURL url = getUrl();
|
|
|
|
TQString headerIds;
|
|
|
|
while (headerIt.current())
|
|
|
|
{
|
|
|
|
headerIds += headerIt.current()->id();
|
|
|
|
if (!headerIt.atLast()) headerIds += ",";
|
|
|
|
++headerIt;
|
|
|
|
}
|
|
|
|
headerIt.toFirst();
|
|
|
|
url.setPath(TQString("/headers/") + headerIds);
|
|
|
|
job = KIO::get( url, false, false );
|
|
|
|
connectJob();
|
|
|
|
slotGetNextHdr();
|
|
|
|
stage = Head;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
stage = Retr;
|
|
|
|
numMsgs = mMsgsPendingDownload.count();
|
|
|
|
numBytesToRead = 0;
|
|
|
|
TQMap<TQString, int>::Iterator len;
|
|
|
|
for ( len = mMsgsPendingDownload.begin();
|
|
|
|
len != mMsgsPendingDownload.end(); len++ )
|
|
|
|
numBytesToRead += len.data();
|
|
|
|
idsOfMsgs = TQStringList( mMsgsPendingDownload.keys() );
|
|
|
|
KURL url = getUrl();
|
|
|
|
url.setPath( "/download/" + idsOfMsgs.join(",") );
|
|
|
|
job = KIO::get( url, false, false );
|
|
|
|
connectJob();
|
|
|
|
slotGetNextMsg();
|
|
|
|
processMsgsTimer.start(processingDelay);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else if (stage == Head) {
|
|
|
|
kdDebug(5006) << k_funcinfo << "stage == Head" << endl;
|
|
|
|
|
|
|
|
// All headers have been downloaded, check which mail you want to get
|
|
|
|
// data is in list headersOnServer
|
|
|
|
|
|
|
|
// check if headers apply to a filter
|
|
|
|
// if set the action of the filter
|
|
|
|
KMPopFilterAction action;
|
|
|
|
bool dlgPopup = false;
|
|
|
|
for (headersOnServer.first(); headersOnServer.current(); headersOnServer.next()) {
|
|
|
|
action = (KMPopFilterAction)kmkernel->popFilterMgr()->process(headersOnServer.current()->header());
|
|
|
|
//debug todo
|
|
|
|
switch ( action ) {
|
|
|
|
case NoAction:
|
|
|
|
kdDebug(5006) << "PopFilterAction = NoAction" << endl;
|
|
|
|
break;
|
|
|
|
case Later:
|
|
|
|
kdDebug(5006) << "PopFilterAction = Later" << endl;
|
|
|
|
break;
|
|
|
|
case Delete:
|
|
|
|
kdDebug(5006) << "PopFilterAction = Delete" << endl;
|
|
|
|
break;
|
|
|
|
case Down:
|
|
|
|
kdDebug(5006) << "PopFilterAction = Down" << endl;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
kdDebug(5006) << "PopFilterAction = default oops!" << endl;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
switch ( action ) {
|
|
|
|
case NoAction:
|
|
|
|
//kdDebug(5006) << "PopFilterAction = NoAction" << endl;
|
|
|
|
dlgPopup = true;
|
|
|
|
break;
|
|
|
|
case Later:
|
|
|
|
if (kmkernel->popFilterMgr()->showLaterMsgs())
|
|
|
|
dlgPopup = true;
|
|
|
|
// fall through
|
|
|
|
default:
|
|
|
|
headersOnServer.current()->setAction(action);
|
|
|
|
headersOnServer.current()->setRuleMatched(true);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// if there are some messages which are not coverd by a filter
|
|
|
|
// show the dialog
|
|
|
|
headers = true;
|
|
|
|
if (dlgPopup) {
|
|
|
|
KMPopFilterCnfrmDlg dlg(&headersOnServer, this->name(), kmkernel->popFilterMgr()->showLaterMsgs());
|
|
|
|
dlg.exec();
|
|
|
|
}
|
|
|
|
|
|
|
|
for (headersOnServer.first(); headersOnServer.current(); headersOnServer.next()) {
|
|
|
|
if (headersOnServer.current()->action() == Delete ||
|
|
|
|
headersOnServer.current()->action() == Later) {
|
|
|
|
//remove entries from the lists when the mails should not be downloaded
|
|
|
|
//(deleted or downloaded later)
|
|
|
|
if ( mMsgsPendingDownload.contains( headersOnServer.current()->id() ) ) {
|
|
|
|
mMsgsPendingDownload.remove( headersOnServer.current()->id() );
|
|
|
|
}
|
|
|
|
if (headersOnServer.current()->action() == Delete) {
|
|
|
|
mHeaderDeleteUids.insert(headersOnServer.current()->uid(), true);
|
|
|
|
mUidsOfNextSeenMsgsDict.insert( headersOnServer.current()->uid(),
|
|
|
|
(const int *)1 );
|
|
|
|
idsOfMsgsToDelete.append(headersOnServer.current()->id());
|
|
|
|
mTimeOfNextSeenMsgsMap.insert( headersOnServer.current()->uid(),
|
|
|
|
time(0) );
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
mHeaderLaterUids.insert(headersOnServer.current()->uid(), true);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else if (headersOnServer.current()->action() == Down) {
|
|
|
|
mHeaderDownUids.insert(headersOnServer.current()->uid(), true);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
headersOnServer.clear();
|
|
|
|
stage = Retr;
|
|
|
|
numMsgs = mMsgsPendingDownload.count();
|
|
|
|
numBytesToRead = 0;
|
|
|
|
TQMap<TQString, int>::Iterator len;
|
|
|
|
for (len = mMsgsPendingDownload.begin();
|
|
|
|
len != mMsgsPendingDownload.end(); len++)
|
|
|
|
numBytesToRead += len.data();
|
|
|
|
idsOfMsgs = TQStringList( mMsgsPendingDownload.keys() );
|
|
|
|
KURL url = getUrl();
|
|
|
|
url.setPath( "/download/" + idsOfMsgs.join(",") );
|
|
|
|
job = KIO::get( url, false, false );
|
|
|
|
connectJob();
|
|
|
|
slotGetNextMsg();
|
|
|
|
processMsgsTimer.start(processingDelay);
|
|
|
|
}
|
|
|
|
else if (stage == Retr) {
|
|
|
|
if ( mMailCheckProgressItem )
|
|
|
|
mMailCheckProgressItem->setProgress( 100 );
|
|
|
|
processRemainingQueuedMessages();
|
|
|
|
|
|
|
|
mHeaderDeleteUids.clear();
|
|
|
|
mHeaderDownUids.clear();
|
|
|
|
mHeaderLaterUids.clear();
|
|
|
|
|
|
|
|
kmkernel->folderMgr()->syncAllFolders();
|
|
|
|
|
|
|
|
KURL url = getUrl();
|
|
|
|
TQMap< TQPair<time_t, TQString>, int > idsToSave;
|
|
|
|
idsToSave.clear();
|
|
|
|
// Check if we want to keep any messages
|
|
|
|
if ( mLeaveOnServer && !idsOfMsgsToDelete.isEmpty() ) {
|
|
|
|
// Keep all messages on server
|
|
|
|
if ( mLeaveOnServerDays == -1 && mLeaveOnServerCount <= 0 &&
|
|
|
|
mLeaveOnServerSize <= 0)
|
|
|
|
idsOfMsgsToDelete.clear();
|
|
|
|
// Delete old messages
|
|
|
|
else if ( mLeaveOnServerDays > 0 && !mTimeOfNextSeenMsgsMap.isEmpty() ) {
|
|
|
|
time_t timeLimit = time(0) - (86400 * mLeaveOnServerDays);
|
|
|
|
kdDebug() << "timeLimit is " << timeLimit << endl;
|
|
|
|
TQStringList::Iterator cur = idsOfMsgsToDelete.begin();
|
|
|
|
for ( ; cur != idsOfMsgsToDelete.end(); ++cur) {
|
|
|
|
time_t msgTime = mTimeOfNextSeenMsgsMap[mUidForIdMap[*cur]];
|
|
|
|
kdDebug() << "id: " << *cur << " msgTime: " << msgTime << endl;
|
|
|
|
if (msgTime >= timeLimit ||
|
|
|
|
!mTimeOfNextSeenMsgsMap[mUidForIdMap[*cur]]) {
|
|
|
|
kdDebug() << "Saving msg id " << *cur << endl;
|
|
|
|
TQPair<time_t, TQString> msg(msgTime, *cur);
|
|
|
|
idsToSave.insert( msg, 1 );
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Delete more old messages if there are more than mLeaveOnServerCount
|
|
|
|
if ( mLeaveOnServerCount > 0 ) {
|
|
|
|
int numToDelete = idsToSave.count() - mLeaveOnServerCount;
|
|
|
|
kdDebug() << "numToDelete is " << numToDelete << endl;
|
|
|
|
if ( numToDelete > 0 && (unsigned)numToDelete < idsToSave.count() ) {
|
|
|
|
TQMap< TQPair<time_t, TQString>, int >::Iterator cur = idsToSave.begin();
|
|
|
|
for ( int deleted = 0; deleted < numToDelete && cur != idsToSave.end()
|
|
|
|
; deleted++, cur++ ) {
|
|
|
|
kdDebug() << "deleting msg id " << cur.key().second << endl;
|
|
|
|
idsToSave.remove( cur );
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else if ( numToDelete > 0 && (unsigned)numToDelete >= idsToSave.count() )
|
|
|
|
idsToSave.clear();
|
|
|
|
}
|
|
|
|
// Delete more old messages until we're under mLeaveOnServerSize MBs
|
|
|
|
if ( mLeaveOnServerSize > 0 ) {
|
|
|
|
double sizeOnServer = 0;
|
|
|
|
TQMap< TQPair<time_t, TQString>, int >::Iterator cur = idsToSave.begin();
|
|
|
|
for ( ; cur != idsToSave.end(); cur++ ) {
|
|
|
|
sizeOnServer +=
|
|
|
|
*mSizeOfNextSeenMsgsDict[ mUidForIdMap[ cur.key().second ] ];
|
|
|
|
}
|
|
|
|
kdDebug() << "sizeOnServer is " << sizeOnServer/(1024*1024) << "MB" << endl;
|
|
|
|
long limitInBytes = mLeaveOnServerSize * ( 1024 * 1024 );
|
|
|
|
for ( cur = idsToSave.begin(); cur != idsToSave.end()
|
|
|
|
&& sizeOnServer > limitInBytes; cur++ ) {
|
|
|
|
sizeOnServer -=
|
|
|
|
*mSizeOfNextSeenMsgsDict[ mUidForIdMap[ cur.key().second ] ];
|
|
|
|
idsToSave.remove( cur );
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Save msgs from deletion
|
|
|
|
TQMap< TQPair<time_t, TQString>, int >::Iterator it = idsToSave.begin();
|
|
|
|
kdDebug() << "Going to save " << idsToSave.count() << endl;
|
|
|
|
for ( ; it != idsToSave.end(); ++it ) {
|
|
|
|
kdDebug() << "saving msg id " << it.key().second << endl;
|
|
|
|
idsOfMsgsToDelete.remove( it.key().second );
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if ( !idsOfForcedDeletes.isEmpty() ) {
|
|
|
|
idsOfMsgsToDelete += idsOfForcedDeletes;
|
|
|
|
idsOfForcedDeletes.clear();
|
|
|
|
}
|
|
|
|
|
|
|
|
// If there are messages to delete then delete them
|
|
|
|
if ( !idsOfMsgsToDelete.isEmpty() ) {
|
|
|
|
stage = Dele;
|
|
|
|
if ( mMailCheckProgressItem )
|
|
|
|
mMailCheckProgressItem->seStatus(
|
|
|
|
i18n( "Fetched 1 message from %1. Deleting messages from server...",
|
|
|
|
"Fetched %n messages from %1. Deleting messages from server...",
|
|
|
|
numMsgs )
|
|
|
|
.tqarg( mHost ) );
|
|
|
|
url.setPath("/remove/" + idsOfMsgsToDelete.join(","));
|
|
|
|
kdDebug(5006) << "url: " << url.prettyURL() << endl;
|
|
|
|
} else {
|
|
|
|
stage = Quit;
|
|
|
|
if ( mMailCheckProgressItem )
|
|
|
|
mMailCheckProgressItem->seStatus(
|
|
|
|
i18n( "Fetched 1 message from %1. Terminating transmission...",
|
|
|
|
"Fetched %n messages from %1. Terminating transmission...",
|
|
|
|
numMsgs )
|
|
|
|
.tqarg( mHost ) );
|
|
|
|
url.setPath(TQString("/commit"));
|
|
|
|
kdDebug(5006) << "url: " << url.prettyURL() << endl;
|
|
|
|
}
|
|
|
|
job = KIO::get( url, false, false );
|
|
|
|
connectJob();
|
|
|
|
}
|
|
|
|
else if (stage == Dele) {
|
|
|
|
kdDebug(5006) << k_funcinfo << "stage == Dele" << endl;
|
|
|
|
// remove the uids of all messages which have been deleted
|
|
|
|
for ( TQStringList::ConstIterator it = idsOfMsgsToDelete.begin();
|
|
|
|
it != idsOfMsgsToDelete.end(); ++it ) {
|
|
|
|
mUidsOfNextSeenMsgsDict.remove( mUidForIdMap[*it] );
|
|
|
|
}
|
|
|
|
idsOfMsgsToDelete.clear();
|
|
|
|
if ( mMailCheckProgressItem )
|
|
|
|
mMailCheckProgressItem->seStatus(
|
|
|
|
i18n( "Fetched 1 message from %1. Terminating transmission...",
|
|
|
|
"Fetched %n messages from %1. Terminating transmission...",
|
|
|
|
numMsgs )
|
|
|
|
.tqarg( mHost ) );
|
|
|
|
KURL url = getUrl();
|
|
|
|
url.setPath(TQString("/commit"));
|
|
|
|
job = KIO::get( url, false, false );
|
|
|
|
stage = Quit;
|
|
|
|
connectJob();
|
|
|
|
}
|
|
|
|
else if (stage == Quit) {
|
|
|
|
kdDebug(5006) << k_funcinfo << "stage == Quit" << endl;
|
|
|
|
saveUidList();
|
|
|
|
job = 0;
|
|
|
|
if (mSlave) KIO::Scheduler::disconnectSlave(mSlave);
|
|
|
|
mSlave = 0;
|
|
|
|
stage = Idle;
|
|
|
|
if( mMailCheckProgressItem ) { // do this only once...
|
|
|
|
bool canceled = !kmkernel || kmkernel->mailCheckAborted() || mMailCheckProgressItem->canceled();
|
|
|
|
int numMessages = canceled ? indexOfCurrentMsg : idsOfMsgs.count();
|
|
|
|
BroadcastStatus::instance()->seStatusMsgTransmissionCompleted(
|
|
|
|
this->name(), numMessages, numBytes, numBytesRead, numBytesToRead, mLeaveOnServer, mMailCheckProgressItem );
|
|
|
|
// set mMailCheckProgressItem = 0 before calling setComplete() to prevent
|
|
|
|
// a race condition
|
|
|
|
ProgressItem *savedMailCheckProgressItem = mMailCheckProgressItem;
|
|
|
|
mMailCheckProgressItem = 0;
|
|
|
|
savedMailCheckProgressItem->setComplete(); // that will delete it
|
|
|
|
checkDone( ( numMessages > 0 ), canceled ? CheckAborted : CheckOK );
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::processRemainingQueuedMessages()
|
|
|
|
{
|
|
|
|
kdDebug(5006) << k_funcinfo << endl;
|
|
|
|
slotProcessPendingMsgs(); // Force processing of any messages still in the queue
|
|
|
|
processMsgsTimer.stop();
|
|
|
|
|
|
|
|
stage = Quit;
|
|
|
|
if ( kmkernel && kmkernel->folderMgr() ) {
|
|
|
|
kmkernel->folderMgr()->syncAllFolders();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::saveUidList()
|
|
|
|
{
|
|
|
|
kdDebug(5006) << k_funcinfo << endl;
|
|
|
|
// Don't update the seen uid list unless we successfully got
|
|
|
|
// a new list from the server
|
|
|
|
if (!mUidlFinished) return;
|
|
|
|
|
|
|
|
TQStringList uidsOfNextSeenMsgs;
|
|
|
|
TQValueList<int> seenUidTimeList;
|
|
|
|
TQDictIterator<int> it( mUidsOfNextSeenMsgsDict );
|
|
|
|
for( ; it.current(); ++it ) {
|
|
|
|
uidsOfNextSeenMsgs.append( it.currentKey() );
|
|
|
|
seenUidTimeList.append( mTimeOfNextSeenMsgsMap[it.currentKey()] );
|
|
|
|
}
|
|
|
|
TQString seenUidList = locateLocal( "data", "kmail/" + mLogin + ":" + "@" +
|
|
|
|
mHost + ":" + TQString("%1").tqarg(mPort) );
|
|
|
|
KConfig config( seenUidList );
|
|
|
|
config.writeEntry( "seenUidList", uidsOfNextSeenMsgs );
|
|
|
|
config.writeEntry( "seenUidTimeList", seenUidTimeList );
|
|
|
|
config.writeEntry( "downloadLater", TQStringList( mHeaderLaterUids.keys() ) );
|
|
|
|
config.sync();
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::slotGetNextMsg()
|
|
|
|
{
|
|
|
|
TQMap<TQString, int>::Iterator next = mMsgsPendingDownload.begin();
|
|
|
|
|
|
|
|
curMsgData.resize(0);
|
|
|
|
numMsgBytesRead = 0;
|
|
|
|
curMsgLen = 0;
|
|
|
|
delete curMsgStrm;
|
|
|
|
curMsgStrm = 0;
|
|
|
|
|
|
|
|
if ( next != mMsgsPendingDownload.end() ) {
|
|
|
|
// get the next message
|
|
|
|
int nextLen = next.data();
|
|
|
|
curMsgStrm = new TQDataStream( curMsgData, IO_WriteOnly );
|
|
|
|
curMsgLen = nextLen;
|
|
|
|
++indexOfCurrentMsg;
|
|
|
|
kdDebug(5006) << TQString("Length of message about to get %1").tqarg( nextLen ) << endl;
|
|
|
|
mMsgsPendingDownload.remove( next.key() );
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::slotData( KIO::Job* job, const TQByteArray &data)
|
|
|
|
{
|
|
|
|
if (data.size() == 0) {
|
|
|
|
kdDebug(5006) << "Data: <End>" << endl;
|
|
|
|
if ((stage == Retr) && (numMsgBytesRead < curMsgLen))
|
|
|
|
numBytesRead += curMsgLen - numMsgBytesRead;
|
|
|
|
else if (stage == Head){
|
|
|
|
kdDebug(5006) << "Head: <End>" << endl;
|
|
|
|
}
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
int oldNumMsgBytesRead = numMsgBytesRead;
|
|
|
|
if (stage == Retr) {
|
|
|
|
headers = false;
|
|
|
|
curMsgStrm->writeRawBytes( data.data(), data.size() );
|
|
|
|
numMsgBytesRead += data.size();
|
|
|
|
if (numMsgBytesRead > curMsgLen)
|
|
|
|
numMsgBytesRead = curMsgLen;
|
|
|
|
numBytesRead += numMsgBytesRead - oldNumMsgBytesRead;
|
|
|
|
dataCounter++;
|
|
|
|
if ( mMailCheckProgressItem &&
|
|
|
|
( dataCounter % 5 == 0 ||
|
|
|
|
( indexOfCurrentMsg + 1 == numMsgs && numMsgBytesRead == curMsgLen ) ) )
|
|
|
|
{
|
|
|
|
TQString msg;
|
|
|
|
if (numBytes != numBytesToRead && mLeaveOnServer)
|
|
|
|
{
|
|
|
|
msg = i18n("Fetching message %1 of %2 (%3 of %4 KB) for %5@%6 "
|
|
|
|
"(%7 KB remain on the server).")
|
|
|
|
.tqarg(indexOfCurrentMsg+1).tqarg(numMsgs).tqarg(numBytesRead/1024)
|
|
|
|
.tqarg(numBytesToRead/1024).tqarg(mLogin).tqarg(mHost).tqarg(numBytes/1024);
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
msg = i18n("Fetching message %1 of %2 (%3 of %4 KB) for %5@%6.")
|
|
|
|
.tqarg(indexOfCurrentMsg+1).tqarg(numMsgs).tqarg(numBytesRead/1024)
|
|
|
|
.tqarg(numBytesToRead/1024).tqarg(mLogin).tqarg(mHost);
|
|
|
|
}
|
|
|
|
mMailCheckProgressItem->seStatus( msg );
|
|
|
|
mMailCheckProgressItem->setProgress(
|
|
|
|
(numBytesToRead <= 100) ? 50 // We never know what the server tells us
|
|
|
|
// This way of dividing is required for > 21MB of mail
|
|
|
|
: (numBytesRead / (numBytesToRead / 100)) );
|
|
|
|
}
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (stage == Head) {
|
|
|
|
curMsgStrm->writeRawBytes( data.data(), data.size() );
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
// otherwise stage is List Or Uidl
|
|
|
|
TQString qdata = data;
|
|
|
|
qdata = qdata.simplifyWhiteSpace(); // Workaround for Maillennium POP3/UNIBOX
|
|
|
|
int spc = qdata.find( ' ' );
|
|
|
|
if ( stage == List ) {
|
|
|
|
if ( spc > 0 ) {
|
|
|
|
TQString length = qdata.mid(spc+1);
|
|
|
|
if (length.find(' ') != -1) length.truncate(length.find(' '));
|
|
|
|
int len = length.toInt();
|
|
|
|
numBytes += len;
|
|
|
|
TQString id = qdata.left(spc);
|
|
|
|
idsOfMsgs.append( id );
|
|
|
|
mMsgsPendingDownload.insert( id, len );
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
stage = Idle;
|
|
|
|
if ( job ) job->kill();
|
|
|
|
job = 0;
|
|
|
|
mSlave = 0;
|
|
|
|
KMessageBox::error( 0, i18n( "Unable to complete LIST operation." ),
|
|
|
|
i18n( "Invalid Response From Server") );
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else { // stage == Uidl
|
|
|
|
Q_ASSERT ( stage == Uidl);
|
|
|
|
|
|
|
|
TQString id;
|
|
|
|
TQString uid;
|
|
|
|
|
|
|
|
if ( spc <= 0 ) {
|
|
|
|
// an invalid uidl line. we might just need to skip it, but
|
|
|
|
// some servers generate invalid uids with valid ids. in that
|
|
|
|
// case we will just make up a uid - which will cause us to
|
|
|
|
// not cache the document, but we will be able to interoperate
|
|
|
|
|
|
|
|
int testid = atoi ( qdata.ascii() );
|
|
|
|
if ( testid < 1 ) {
|
|
|
|
// we'll just have to skip this
|
|
|
|
kdDebug(5006) << "PopAccount::slotData skipping UIDL entry due to parse error "
|
|
|
|
<< endl << qdata.ascii() << endl;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
id.setNum (testid, 10);
|
|
|
|
|
|
|
|
TQString datestring, serialstring;
|
|
|
|
|
|
|
|
serialstring.setNum ( ++dataCounter, 10 );
|
|
|
|
datestring.setNum ( time(NULL),10 );
|
|
|
|
uid = TQString( "uidlgen" ) + datestring + TQString( "." ) + serialstring;
|
|
|
|
kdDebug(5006) << "PopAccount::slotData message " << id.ascii()
|
|
|
|
<< "%d has bad UIDL, cannot keep a copy on server" << endl;
|
|
|
|
idsOfForcedDeletes.append( id );
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
id = qdata.left( spc );
|
|
|
|
uid = qdata.mid( spc + 1 );
|
|
|
|
}
|
|
|
|
|
|
|
|
int *size = new int; //malloc(size_of(int));
|
|
|
|
*size = mMsgsPendingDownload[id];
|
|
|
|
mSizeOfNextSeenMsgsDict.insert( uid, size );
|
|
|
|
if ( mUidsOfSeenMsgsDict.find( uid ) != 0 ) {
|
|
|
|
if ( mMsgsPendingDownload.contains( id ) ) {
|
|
|
|
mMsgsPendingDownload.remove( id );
|
|
|
|
}
|
|
|
|
else
|
|
|
|
kdDebug(5006) << "PopAccount::slotData synchronization failure." << endl;
|
|
|
|
idsOfMsgsToDelete.append( id );
|
|
|
|
mUidsOfNextSeenMsgsDict.insert( uid, (const int *)1 );
|
|
|
|
if ( mTimeOfSeenMsgsVector.empty() ) {
|
|
|
|
mTimeOfNextSeenMsgsMap.insert( uid, time(0) );
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// cast the int* with a long to can convert it to a int, BTW
|
|
|
|
// works with g++-4.0 and amd64
|
|
|
|
mTimeOfNextSeenMsgsMap.insert( uid, mTimeOfSeenMsgsVector[(int)( long )
|
|
|
|
mUidsOfSeenMsgsDict[uid] - 1] );
|
|
|
|
}
|
|
|
|
}
|
|
|
|
mUidForIdMap.insert( id, uid );
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::slotResult( KIO::Job* )
|
|
|
|
{
|
|
|
|
if (!job) return;
|
|
|
|
if ( job->error() )
|
|
|
|
{
|
|
|
|
if (interactive) {
|
|
|
|
if (headers) { // nothing to be done for headers
|
|
|
|
idsOfMsgs.clear();
|
|
|
|
}
|
|
|
|
if (stage == Head && job->error() == KIO::ERR_COULD_NOT_READ)
|
|
|
|
{
|
|
|
|
KMessageBox::error(0, i18n("Your server does not support the "
|
|
|
|
"TOP command. Therefore it is not possible to fetch the headers "
|
|
|
|
"of large emails first, before downloading them."));
|
|
|
|
slotCancel();
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
// force the dialog to be shown next time the account is checked
|
|
|
|
if (!mStorePasswd) mPasswd = "";
|
|
|
|
job->showErrorDialog();
|
|
|
|
}
|
|
|
|
slotCancel();
|
|
|
|
}
|
|
|
|
else
|
|
|
|
slotJobFinished();
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::slotSlaveError(KIO::Slave *aSlave, int error,
|
|
|
|
const TQString &errorMsg)
|
|
|
|
{
|
|
|
|
if (aSlave != mSlave) return;
|
|
|
|
if (error == KIO::ERR_SLAVE_DIED) mSlave = 0;
|
|
|
|
|
|
|
|
// explicitely disconnect the slave if the connection went down
|
|
|
|
if ( error == KIO::ERR_CONNECTION_BROKEN && mSlave ) {
|
|
|
|
KIO::Scheduler::disconnectSlave( mSlave );
|
|
|
|
mSlave = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (interactive && kmkernel) {
|
|
|
|
KMessageBox::error(kmkernel->mainWin(), KIO::buildErrorString(error, errorMsg));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
stage = Quit;
|
|
|
|
if (error == KIO::ERR_COULD_NOT_LOGIN && !mStorePasswd)
|
|
|
|
mAskAgain = true;
|
|
|
|
/* We need a timer, otherwise slotSlaveError of the next account is also
|
|
|
|
executed, if it reuses the slave, because the slave member variable
|
|
|
|
is changed too early */
|
|
|
|
TQTimer::singleShot(0, this, TQT_SLOT(slotCancel()));
|
|
|
|
}
|
|
|
|
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
void PopAccount::slotGetNextHdr(){
|
|
|
|
kdDebug(5006) << "slotGetNextHeader" << endl;
|
|
|
|
|
|
|
|
curMsgData.resize(0);
|
|
|
|
delete curMsgStrm;
|
|
|
|
curMsgStrm = 0;
|
|
|
|
|
|
|
|
curMsgStrm = new TQDataStream( curMsgData, IO_WriteOnly );
|
|
|
|
}
|
|
|
|
|
|
|
|
void PopAccount::killAllJobs( bool ) {
|
|
|
|
// must reimpl., but we don't use it yet
|
|
|
|
}
|
|
|
|
|
|
|
|
} // namespace KMail
|
|
|
|
#include "popaccount.moc"
|