This article seems to be trying to disarm tech elitism by diving into the deep end of it, but by and large the humor is pretty forced and the tone condescending.
I'd also recommend that the author remove the attempt at being edgy at the end with the snide reference to the reports of sexual harassment at Uber. Especially the cheeky "too soon?", which simultaneously acknowledges the gravity of the situation for the victims while also dismissing it entirely in favor of an attempted joke.
So it is not only that you don't find it funny (which is a common thing about jokes - many jokes are bad in the eyes of many people), but also that you want someone else to therefore remove this part.
Where does this idea come from that just because one doesn't like something or finds it stupid/silly/inappropriate/not funny, it has to disappear?
Before you get too defensive consider that they said they'd 'recommend' the author remove the joke, not that it has to go. That's how constructive criticism works.
I think the comment can be better interpreted as 'the tone of the article could be improved, in my opinion, if the author would remove the joke about sexual harassment at the end'
Super exciting seeing on-par performance of RL tasks with dramatically less supervision.
Really looking forward to a follow-up where they explore 2.2.4 further. Sampling examples which provide maximal information game seems like it could result in another huge reduction in the amount of human oversight necessary. Could see an adversarial scheme which could learn to sample these examples optimally from the manifold. This kind of thing is powerful in human learning of complex tasks to ask for clarification or feedback in specific places of uncertainty.
I'm very interested in these lectures and am looking forward to digging into it.
I was wondering if you could provide some feedback on whether deep learning would be useful in classifying images that have text or not. For example, looking at a set of images I wish to classify the ones that have text and the ones that don't have text. A dataset could be like this:
Yeah for sure - these images are pretty different in their composition so it should be pretty easy to classify them. How large is your dataset? Do you need to collect one?
With small amounts of data, transfer learning is the most effective approach. There's a great tutorial on retraining inception for your own categories in TensorFlow: https://www.tensorflow.org/how_tos/image_retraining/.
I don't have a dataset at the moment. I would have to build one. I was thinking of about 200 images in the dataset with 100 of text and 100 of non text. Would that be big enough dataset for transfer learning? Please let me know if there is a dataset you know of that I could leverage
That's a good start. I was thinking you could generate unlimited training data by using a game engine. You'd have the actual 3D model for every single frame.
I believe one came out recently, I saw it on GitHub a few days ago. It mainly did landscape and searched for similar images then combined them to form a new one. It also supported image manipulation, eg turning a brown square purse into a green rounded one. I'll see if I can find it and edit the link in.
Assuming it's computation bound, it's a factor of 5400 (~13 doublings in CPU power required to get to real-time, assuming no algorithmic improvements).
If I'm not mistaken, it seems that the current limitation is that it needs to be produced sequentially for a dependent sequence of audio, perhaps some independent sentences can be run simultaneously using copies of the net assuming no memory limitations. I wonder if it's already possible to create an auidobook for instance in reasonable time.
Google never stated they use those to train models as far as I know. It seems that they are primarily used to spare energy when deploying trained models at scale.
Theres no reason they couldn't use them to train, as long as they can account for the lower precision operations. I think it would be much cheaper to train on them, at that scale anyway.
Afaik the Google TPU does inference only, at 8 bits. I don't think it's possible to train a neural network at 8 bit precision at this point in time. FP16 works for training though, and is twice as fast as FP32 on certain nvidia chips
Backpropagation can work with any precision, as long as you use stochastic rounding (so that the rounding errors are not correlated.) Without stochastic rounding even 16 bits will have rounding error bias.
I haven't seen 8bit training implemented in any (public) frameworks yet - that's not to say it's not possible. If it works then that's great, especially for specialised hardware.
That doesn't imply they can run WaveNet yet - for inference this net is sort of worst-case serial. Their TPU ASIC is almost certainly highly parallel, like a GPU - actually has to be that way for energy efficiency (which is it's claimed benefit).
Wavenet actually looks like it could possibly have been designed to run on CPUs in production, at least after they can further optimize it some. Sampling is super slow right now because it requires an enormous number of tiny dependent TF ops and thus kernels that have huge overhead for tiny amounts of work. A custom implementation could probably circumvent that by evaluating all the layers sequentially in local cache on a fast CPU.
Or they just designed it without much concern for production plausibility yet.
At its heart, this is a new training architecture that allows parameter weights to be updated faster in a distributed setting.
The speed-up happens like so: instead of waiting for the full error gradient to propagate through the entire model, nodes can calculate the local gradient immediately and estimate the rest of it.
The full gradient does eventually get propagated, and it is used to fine-tune the estimator, which is a mini-neural net in itself.
Its amazing that this works, and the implications that full back-prop may not always be needed shakes up a lot of assumptions about training deep nets. This paper also continues the trend of this year of using neural nets as estimators/tools to improve the training of other neural nets. (I'm looking at you GANs).
Overall, excited to see where this goes as other researchers explore the possibilities when you throw the back-prop assumption out.