Opened at 2012-12-05T03:49:05Z
Closed at 2020-10-30T12:35:44Z
#1886 closed defect (wontfix)
cloud backend: crawlers are running too often
Reported by: | davidsarah | Owned by: | daira |
---|---|---|---|
Priority: | normal | Milestone: | 1.15.0 |
Component: | code-storage | Version: | 1.9.2 |
Keywords: | cloud-backend crawler performance regression blocks-cloud-merge | Cc: | |
Launchpad Bug: |
Description (last modified by daira)
I flogtool tailed the merged cloud backend and leasedb (as of 1819-cloud-merge/022796fb), and observed that the bucket crawler and accounting crawler were running much too often. This was confirmed by looking at the storage status page at /storage, which said that the bucket crawler and accounting crawler would run in 2 minutes and 7 minutes respectively, after they had run a few minutes ago.
The bucket crawler is supposed to have a minimum cycle time of one hour, and the accounting crawler 12 hours.
Change History (12)
comment:1 Changed at 2013-07-04T19:15:45Z by daira
- Description modified (diff)
- Keywords regression blocks-cloud-merge added
- Owner set to daira
- Status changed from new to assigned
comment:2 Changed at 2013-07-22T20:48:52Z by daira
- Milestone changed from 1.11.0 to 1.12.0
comment:3 follow-up: ↓ 4 Changed at 2013-11-14T19:05:07Z by zooko
comment:4 in reply to: ↑ 3 ; follow-up: ↓ 5 Changed at 2013-11-21T22:55:06Z by daira
comment:5 in reply to: ↑ 4 ; follow-up: ↓ 6 Changed at 2013-11-22T05:53:28Z by zooko
Replying to daira:
Replying to zooko:
This would be fixed by #1834.
Well, not unless we removed all use of crawlers, which I don't think #1834 proposed to do.
IIRC, it proposes to remover all crawlers except for this one: #1835, and that one would be executed only in specific circumstances, not executing automatically over-and-over, so this ticket would almost certainly be obsolete.
comment:6 in reply to: ↑ 5 Changed at 2013-11-23T13:15:29Z by daira
Replying to zooko:
IIRC, it proposes to remover all crawlers except for this one: #1835, and that one would be executed only in specific circumstances, not executing automatically over-and-over, so this ticket would almost certainly be obsolete.
Oh, good point. Nevertheless I want to fix this bug before then, because I think it's shallow and could easily be fixed in an earlier release.
comment:7 Changed at 2014-12-03T04:23:17Z by zooko
Hm, I wonder if this could have to do with the out-of-file-descriptors issue that my client ("WAG") is experiencing — #2342.
comment:8 Changed at 2014-12-23T10:41:48Z by dquintela
Hello, first time tahoe user here,
I've testing cloud storage for 3 or 4 days (branch 2237-cloud-backend-s4) into my raspberry-pi, running raspbian. Despite it having very long startup times, I noticed .tahoe/logs/twistd.log filled up of lines like this, that seems indicating the shares crawler is being run to often. This has the nasty side effecting that on amazon billing page I am already with 80000 requests. Rough estimate this could mean 3 to 5 USD per month on an idle storage node alone.
This seems to be related to #1835 and #1886 - sorry for the crossposting.
2014-12-22 15:48:37+0000 [-] Starting factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fod%2F> 2014-12-22 15:48:37+0000 [HTTPPageGetter,client] Stopping factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fod%2F> 2014-12-22 15:48:37+0000 [-] Starting factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Foe%2F> 2014-12-22 15:48:38+0000 [HTTPPageGetter,client] Stopping factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Foe%2F> 2014-12-22 15:48:38+0000 [-] Starting factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fof%2F> 2014-12-22 15:48:38+0000 [HTTPPageGetter,client] Stopping factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fof%2F> 2014-12-22 15:48:38+0000 [-] Starting factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fog%2F> 2014-12-22 15:48:38+0000 [HTTPPageGetter,client] Stopping factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fog%2F> ...
comment:9 Changed at 2016-03-22T05:02:25Z by warner
- Milestone changed from 1.12.0 to 1.13.0
Milestone renamed
comment:10 Changed at 2016-06-28T18:17:14Z by warner
- Milestone changed from 1.13.0 to 1.14.0
renaming milestone
comment:11 Changed at 2020-06-30T14:45:13Z by exarkun
- Milestone changed from 1.14.0 to 1.15.0
Moving open issues out of closed milestones.
comment:12 Changed at 2020-10-30T12:35:44Z by exarkun
- Resolution set to wontfix
- Status changed from assigned to closed
The established line of development on the "cloud backend" branch has been abandoned. This ticket is being closed as part of a batch-ticket cleanup for "cloud backend"-related tickets.
If this is a bug, it is probably genuinely no longer relevant. The "cloud backend" branch is too large and unwieldy to ever be merged into the main line of development (particularly now that the Python 3 porting effort is significantly underway).
If this is a feature, it may be relevant to some future efforts - if they are sufficiently similar to the "cloud backend" effort - but I am still closing it because there are no immediate plans for a new development effort in such a direction.
Tickets related to the "leasedb" are included in this set because the "leasedb" code is in the "cloud backend" branch and fairly well intertwined with the "cloud backend". If there is interest in lease implementation change at some future time then that effort will essentially have to be restarted as well.
This would be fixed by #1834.