Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The problem with preconceptions about your parameters is that you might be missing some crazy cool path to your goal, which you might find by randomly exploring your sample space. I remember seeing this same principle in mcmc samplers using uniform priors. Why is this so crazy?


It's predicated on the assumption that a random discovery from a zero-comprehension state is more likely to get you to a goal than an evolution from a state that has at least some correctness.

More generally, it disingenuously disregards the fact that the definition of the problem brings with it an enormous set of preconceptions. Reductio ad absurdum, you should just start training a model on completely random data in search of some unexpected but useful outcome.

Obviously we don't do this; by setting a goal and a context we have already applied constraints, and so this really just devolves into a quantitative argument about the set of initial conditions.

(This is the entire point of the Minsky / Sussman koan.)


> from a zero-comprehension state is more likely to get you to a goal than an evolution from a state that has at least some correctness.

I get that starting from a point with "some correctness" makes sense if you want to use such information (e.g. a standard starting point). However, such information is a preconceived solution to the problem, which might not be that useful after all. The fact is that you indeed might not at all need such information to find an optimal solution to a given problem.

> by setting a goal and a context we have already applied constraints.

I might be missing your point here since the goal and constraints must come from the real world problem to solve which is independent from the method to solve the problem. Unless you're describing p-value hacking your wait out, which is a broader problem.


With exploring, the starting state should only affect which local-maximum you end up in. Therefore you need to make an argument that a random starting state is likely to end up in a higher local-maximum than a non-random starting state.

There is always a starting state; using a random one only means you don't know what it is.


Exactly, but why do so many people seem to have a problem with this? Sounds like a political problem to me instead of a scientific one.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: