From 4500a1aa6daaf7b38705ef03fa0c99c8fdafa2be Mon Sep 17 00:00:00 2001 From: Nick Martin Date: Tue, 5 Nov 2013 15:29:31 -0800 Subject: [PATCH] update params files and readme for new benchmark. --- .../chat-benchmark/scenarios/README.md | 35 ++++++------------- .../scenarios/bigdata-static.json | 9 ----- .../scenarios/bigdata-updates.json | 10 ------ .../chat-benchmark/scenarios/nodata.json | 7 ---- 4 files changed, 10 insertions(+), 51 deletions(-) delete mode 100644 examples/unfinished/chat-benchmark/scenarios/bigdata-static.json delete mode 100644 examples/unfinished/chat-benchmark/scenarios/bigdata-updates.json delete mode 100644 examples/unfinished/chat-benchmark/scenarios/nodata.json diff --git a/examples/unfinished/chat-benchmark/scenarios/README.md b/examples/unfinished/chat-benchmark/scenarios/README.md index 925fd1fe89..dab01d5da0 100644 --- a/examples/unfinished/chat-benchmark/scenarios/README.md +++ b/examples/unfinished/chat-benchmark/scenarios/README.md @@ -1,27 +1,12 @@ -Parameters for simulation: +Each client creates a room when it starts up and joins that room plus +`roomsPerClient` other rooms. -Each document is randomly placed in a collection, with a random -'bucket' field. Clients sub to 1 bucket in each collection. +Every `talkativePeriodSeconds`, each client has a +`chanceClientIsTalkative` probability of being "talkative". If it is +talkative for that period, it will send one `messageSize` sized message +every `talkativeMessagesPerSecond` for the duration of the period. -- numCollections - how many collections to spread the documents over -- numBuckets - number of buckets per collection. - -- initialDocuments: Inital documents added by the server. Probably - not usefully combined with maxAgeSeconds - -- maxAgeSeconds: How long to leave documents in the database. This, - combined with all the various rates, determines the steady state - database size. In seconds. falsy to disable. - -Per-client action rates: -- insertsPerSecond -- updatesPerSecond -- removesPerSecond - -- documentSize: bytes of randomness per document. - // XXX make this a random distribution? -- documentNumFields: how many fields of randomness per document. - -XXX also max documents? (count and remove N) +Old messages are deleted every `messageHistorySeconds` seconds. Old +rooms are deleted every `roomHistorySeconds`. `roomHistorySeconds` +should line up with the length a particular client dwells on the site +(typically part of the phantomjs script used to drive the clients.) diff --git a/examples/unfinished/chat-benchmark/scenarios/bigdata-static.json b/examples/unfinished/chat-benchmark/scenarios/bigdata-static.json deleted file mode 100644 index 8523c45360..0000000000 --- a/examples/unfinished/chat-benchmark/scenarios/bigdata-static.json +++ /dev/null @@ -1,9 +0,0 @@ -{ - "params": { - "numCollections": 1, - "numBuckets": 1, - "initialDocuments": 1024, - "documentSize": 2048, - "documentNumFields": 64 - } -} diff --git a/examples/unfinished/chat-benchmark/scenarios/bigdata-updates.json b/examples/unfinished/chat-benchmark/scenarios/bigdata-updates.json deleted file mode 100644 index 19e64db5d8..0000000000 --- a/examples/unfinished/chat-benchmark/scenarios/bigdata-updates.json +++ /dev/null @@ -1,10 +0,0 @@ -{ - "params": { - "numCollections": 1, - "numBuckets": 1, - "initialDocuments": 1024, - "updatesPerSecond": 0.2, - "documentSize": 1024, - "documentNumFields": 32 - } -} diff --git a/examples/unfinished/chat-benchmark/scenarios/nodata.json b/examples/unfinished/chat-benchmark/scenarios/nodata.json deleted file mode 100644 index 53460f0e07..0000000000 --- a/examples/unfinished/chat-benchmark/scenarios/nodata.json +++ /dev/null @@ -1,7 +0,0 @@ -{ - "params": { - "numCollections": 1, - "numBuckets": 1, - "initialDocuments": 0 - } -}