#1886 closed defect (wontfix)

cloud backend: crawlers are running too often

Reported by: davidsarah Owned by: daira
Priority: normal Milestone: 1.15.0
Component: code-storage Version: 1.9.2
Keywords: cloud-backend crawler performance regression blocks-cloud-merge Cc:
Launchpad Bug:

Description (last modified by daira)

I flogtool tailed the merged cloud backend and leasedb (as of 1819-cloud-merge/022796fb), and observed that the bucket crawler and accounting crawler were running much too often. This was confirmed by looking at the storage status page at /storage, which said that the bucket crawler and accounting crawler would run in 2 minutes and 7 minutes respectively, after they had run a few minutes ago.

The bucket crawler is supposed to have a minimum cycle time of one hour, and the accounting crawler 12 hours.

Change History (12)

comment:1 Changed at 2013-07-04T19:15:45Z by daira

  • Description modified (diff)
  • Keywords regression blocks-cloud-merge added
  • Owner set to daira
  • Status changed from new to assigned

comment:2 Changed at 2013-07-22T20:48:52Z by daira

  • Milestone changed from 1.11.0 to 1.12.0

comment:3 follow-up: Changed at 2013-11-14T19:05:07Z by zooko

This would be fixed by #1834.

comment:4 in reply to: ↑ 3 ; follow-up: Changed at 2013-11-21T22:55:06Z by daira

Replying to zooko:

This would be fixed by #1834.

Well, not unless we removed all use of crawlers, which I don't think #1834 proposed to do.

comment:5 in reply to: ↑ 4 ; follow-up: Changed at 2013-11-22T05:53:28Z by zooko

Replying to daira:

Replying to zooko:

This would be fixed by #1834.

Well, not unless we removed all use of crawlers, which I don't think #1834 proposed to do.

IIRC, it proposes to remover all crawlers except for this one: #1835, and that one would be executed only in specific circumstances, not executing automatically over-and-over, so this ticket would almost certainly be obsolete.

comment:6 in reply to: ↑ 5 Changed at 2013-11-23T13:15:29Z by daira

Replying to zooko:

IIRC, it proposes to remover all crawlers except for this one: #1835, and that one would be executed only in specific circumstances, not executing automatically over-and-over, so this ticket would almost certainly be obsolete.

Oh, good point. Nevertheless I want to fix this bug before then, because I think it's shallow and could easily be fixed in an earlier release.

comment:7 Changed at 2014-12-03T04:23:17Z by zooko

Hm, I wonder if this could have to do with the out-of-file-descriptors issue that my client ("WAG") is experiencing — #2342.

comment:8 Changed at 2014-12-23T10:41:48Z by dquintela

Hello, first time user tahoe here,

I've testing cloud storage for 3 or 4 days (branch 2237-cloud-backend-s4) into my raspberry-pi, running raspbian. Despite it having very long startup times, I noticed .tahoe/logs/twistd.log filled up of lines like this, that seems indicating the shares crawler is being run to often. This has the nasty side effecting that on amazon billing page I am already with 80000 requests. Rough estimate this could mean 3 to 5 USD per month on an idle storage node alone.

This seems to be related to #1835 and #1886 - sorry for the crossposting.

2014-12-22 15:48:37+0000 [-] Starting factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fod%2F> 2014-12-22 15:48:37+0000 [HTTPPageGetter,client] Stopping factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fod%2F> 2014-12-22 15:48:37+0000 [-] Starting factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Foe%2F> 2014-12-22 15:48:38+0000 [HTTPPageGetter,client] Stopping factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Foe%2F> 2014-12-22 15:48:38+0000 [-] Starting factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fof%2F> 2014-12-22 15:48:38+0000 [HTTPPageGetter,client] Stopping factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fof%2F> 2014-12-22 15:48:38+0000 [-] Starting factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fog%2F> 2014-12-22 15:48:38+0000 [HTTPPageGetter,client] Stopping factory <HTTPClientFactory: http://bucket_identifier.s3.amazonaws.com/?prefix=shares%2Fog%2F> ...

Version 0, edited at 2014-12-23T10:41:48Z by dquintela (next)

comment:9 Changed at 2016-03-22T05:02:25Z by warner

  • Milestone changed from 1.12.0 to 1.13.0

Milestone renamed

comment:10 Changed at 2016-06-28T18:17:14Z by warner

  • Milestone changed from 1.13.0 to 1.14.0

renaming milestone

comment:11 Changed at 2020-06-30T14:45:13Z by exarkun

  • Milestone changed from 1.14.0 to 1.15.0

Moving open issues out of closed milestones.

comment:12 Changed at 2020-10-30T12:35:44Z by exarkun

  • Resolution set to wontfix
  • Status changed from assigned to closed

The established line of development on the "cloud backend" branch has been abandoned. This ticket is being closed as part of a batch-ticket cleanup for "cloud backend"-related tickets.

If this is a bug, it is probably genuinely no longer relevant. The "cloud backend" branch is too large and unwieldy to ever be merged into the main line of development (particularly now that the Python 3 porting effort is significantly underway).

If this is a feature, it may be relevant to some future efforts - if they are sufficiently similar to the "cloud backend" effort - but I am still closing it because there are no immediate plans for a new development effort in such a direction.

Tickets related to the "leasedb" are included in this set because the "leasedb" code is in the "cloud backend" branch and fairly well intertwined with the "cloud backend". If there is interest in lease implementation change at some future time then that effort will essentially have to be restarted as well.

Note: See TracTickets for help on using tickets.