Open GoogleCodeExporter opened 9 years ago
This is a good idea, but I think I want to put this behind the per-host stats
interface, as that's what will tell the admin where all the logs are going.
Original comment by mchol...@gmail.com
on 28 Aug 2012 at 4:00
How can you manually calculate settings? Just the manually formal would help a
lot.
Ex.
Daily logs: 3000000
Disk space for logging: 200Gb
Lightning speed search for the entire data logged to disk. When the logging
disk space is full discard oldest logs
HW:
8 cores 3Ghz Intel 64bit
8 hdd 72Gb 15K disk in raid 10
32 Gb RAM
Original comment by jacobrav...@gmail.com
on 5 Mar 2013 at 7:53
Hi there, I was planning on posting a similar question. So here goes what I
would do based on my (little) experience with ELSA:
- Due to your "lightning speed search" requirement, I assume you don't want
ELSA to go through the archived logs at all (these are compressed non-indexed
Mysql tables, so it takes a lot of processing power to get data out of them)
- Also considering the "when disk is full discard oldest logs" I assume you
don't mind losing the logs.
Therefore, first thing I would do is setting the 'archive/percentage' setting
to 0 (so logs are either in an index or nowhere).
- Then based on your specs (200GB) and my ELSA instances DB average values
(about 300 bytes per table row that includes original message, all parsed
'fields' and MySQL index space), I would say you can keep in store up to around
710M records (about 240 days worth of logs online)
So, I would play with a config like this (still not sure about the # of indexes)
log_size_limit: 200000000
num_indexes: 400 (I would try this just to help prevent temp index
consolidation by time, and force Sphinx use more RAM for temp indexes - but
this point in particular needs expert advise)
index_interval: 120 seconds (I believe this will 'extend' the life of the temp
indexes in RAM and therefore help speed up searching)
allowed_mem_percent: 80 (you still keep ~ 7gb free for buffers/other services)
perm_index_size: 10000000 (based on num_indexes this value should be enough to
consume available storage before overwriting indexes)
No idea about how you could take more advantage of the multiple cores...
But hey, I was about to post a similar question, so this is just 'work in
progress' :-)
Good luck.
Original comment by rhatu...@gmail.com
on 5 Mar 2013 at 6:39
Yes, those are excellent tuning settings for your workload. They should
work nicely and provide around 180 days of logs. One other setting you
should set would be allowed_temp_percent, which I would raise to 80. This
allows 80% of 400 temp indexes to be used before they are consolidated into
one permanent index. Otherwise, you would get the default of 40% of 400,
which is a lot less time overall.
On Tue, Mar 5, 2013 at 12:39 PM, <
enterprise-log-search-and-archive@googlecode.com> wrote:
Original comment by mchol...@gmail.com
on 5 Mar 2013 at 10:17
Excellent! Thanks a lot for the advise Martin.
One more on this thread, it's still not 100% clear to me if there is multiple
(small) permanent index consolidation into one single permanent index in ELSA.
Is this the case?
For example I find this useful to cope with days where there is very little
logging, like a weekend, and prevent these under utilised permanent indexes to
force the rotation of other, better used, perm indexes.
Original comment by rhatu...@gmail.com
on 6 Mar 2013 at 8:02
Indexes will get consolidated whenever a table has more rows than
"perm_index_size" and consists of more than a single index, so your weekend
perm indexes should get further consolidated as necessary.
On Wednesday, March 6, 2013, wrote:
Original comment by mchol...@gmail.com
on 6 Mar 2013 at 2:56
That's perfect, it actually clarifies some doubts I had on that area of the
code.
Thanks for the response man.
Original comment by rhatu...@gmail.com
on 6 Mar 2013 at 3:01
[deleted comment]
How to i set the conf's for the following spec?
Daily logs: 24000000 (24 Million logs pr day)
Disk space for logging: 12 TB
Original comment by jacobrav...@gmail.com
on 2 Apr 2013 at 10:54
Figuring an average log size of 300 bytes, 24 million logs per day will take
about 18 GB of disk space per day, so 12 TB will afford a very long retention
time (about two years). You shouldn't need to change the number of indexes
anymore, because now ELSA will continually consolidate indexes as needed.
However, you should set the sphinx/perm_index_size to 80 million from the
default of 10 million to make sure that each of the 200 possible indexes will
store the full amount to get to the roughly 15 billion logs 12 TB should hold.
Original comment by mchol...@gmail.com
on 3 Apr 2013 at 4:32
Original issue reported on code.google.com by
jeffrey....@gmail.com
on 28 Aug 2012 at 4:50