Closed L-M-Sherlock closed 11 months ago
Interesting. Do you have any ideas how to use these findings to improve FSRS? Perhaps we can add same-day reviews to the estimation of S0?
Also, how is the benchmarking going?
Interesting. Do you have any ideas how to use these findings to improve FSRS?
I attempt to include the short-term reviews with float-delta-t in FSRS. But it increased w[9] (stabilization decay) drastically. In other word, the factor
is very large in the early steps.The short-term memory seems to has very different stabilization process in contrast with the long-term memory.
Also, how is the benchmarking going?
It stuck. It will take 100+ hours to run the benchmark (and my computer cannot do anything else during the period). Converting FSRS from a seq-to-one model to a seq-to-seq model could speed it up. But it's not compatible with TimeSeriesSplit.
Try to include same-day reviews in the estimation of S0.
As for benchmarking, maybe it's really time to rewrite everything in Rust? Well, either that, or run the benchmark but reduce num_threads
. That's what I did to be able to use my computer while the benchmark is running. It will be slower, of course, but 100-200 hours isn't as bad as I was expecting. I could do that for you, if you make the dataset downloadable.
OK. I'm developing a new pre-process script for the benchmark, because I need to save my disk size :). t_history
and r_history
will not stored in the dataset files.
t_history
andr_history
will not stored in the dataset files.
Then how will the optimizer work without them? How will they be stored?
Then how will the optimizer work without them?
I find an efficient way to build these features from the raw data.
what do the blue bubbles represent
I have an idea: maybe we could do the same thing we do for S0, but for learning steps? 1) Use only same-day reviews 2) Calculate the average stability for those, using the curve-fitting method that we use for S0 3) Output that value of S0_short_term and tell the user to use it as a learning step
This could be called "Compute optimal learning steps", a new feature.
Using your graph as an example, here the learning step would be 7m
This could be called "Compute optimal learning steps", a new feature.
I was hoping something like this came to be
@L-M-Sherlock were this finding used in the 4.5 release. What were the changes for 4.5 and are there benchmark results
It wasn't. I still didn't find out a clear pattern of short-term memory.
@Expertium, I found that the same-day reviews have a significant impact.