-
I recently came across SigOpt and used their experiment design module. Some further research then brought me here. I'm just getting started as with MOE but can I basically get the same functionality w…
-
Hi,
I have an existing bayesian bandit system that I might like to discover new points to test through MOE.
For a beta distrubution, is there some way to feed it in directly using value_var to prov…
-
As far as I understand, sai brings different concept to dramatically improve value network in unfair situation.
If we consider extrem situation like 9 handicap stone, leela zero is considering the …
-
# 1.3 Elements of Reinforcement Learning
- *Policy*
- A policy defines the learning agent’s way of behaving at a given time.
- Roughly speaking, a policy is a mapping from perceived states of…
-
If you want to become a reviewer for ReScience, please post your information here. The format is:
```
[name](github account link)
Scientific expertise - Language expertise
ORCID: [xxxx](http…
-
**Submitting author:** @pat-alt (Patrick Altmeyer)
**Repository:** https://github.com/pat-alt/LaplaceRedux.jl
**Branch with paper.md** (empty if default branch):
**Version:**
**Editor:** @matbesanco…
-
There was some talk about training a handicap NN in the other issue, however it's already quite long and handi games are offtopic there too. So want to continue that part of the discussion elsewhere. …
Dorus updated
6 years ago
-
Currently ipfs uses a DHT for all content routing. This works quite well for
many use cases and is generally reliable, fast, and durable. The problem we are
now facing is that it does not scale well. …
-
-
Need to figure out which learning paradigm, package(s), and inputs we need.