diff --git a/Sizing b/Sizing index e69de29..655cedb 100644 --- a/Sizing +++ b/Sizing @@ -0,0 +1,7 @@ +What size server is required will depend very much on your use case. The numbers of documents you want to process at once, their size and content type + + The S&D demo servers are low level KVM VMs, with the Ingest server running 2 vCPUs and 1G RAM and the frontend server running 2vCPUs and 2G RAM. In our tests of bulk operation (which is via the CLI app https://git.law/newroco/searchanddisplace-core/src/branch/master/demo-cli/README.md) using those demo servers we got the following results. + +Using 380 documents, which were a mix of DOCX, ODT and PDF and vary in size up to ~5M and 60+ pages of dense text (contracts!), running a single searcher on each document took 306 seconds to ingest the documents and complete the search and displace action. + +It was noted that both RAM and CPU were being maxed out in the tests. The software architecture utilises queues so there is considerable room for improving that performance just by adding CPU or RAM \ No newline at end of file