Taking the approach of a scientist, i.e. a data scientist, this is a hard problem to solve. Not only does Facebook have to implement the simple user.delete.all.data(), but think about the number of internal products that are machine learning based. In order to preserve the usefulness of those models and the hundreds of millions in R&D spent to develop them, you have to carefully understand exactly what products and algos will be affected by just deleting a single record.
This is not a science problem and it is not hard to solve: a scientist would delete the data and re-derive their models because they want the best data possible. This is how you "preserve the usefulness of those models," by ensuring that the data has integrity.
facebook has to figure out a way to understand exactly how thousands of records per user, being deleted, will affect it's models.
This is looking through the telescope backwards. You create your models with the data that you have, and if some data is removed then you regenerate.
You can bet your ass they groom their data after detecting stuff like new botnets and fake clicks, and individual user data does not have to be different.
That is an entire community and classification of people that facebook has to figure out how to reconcile the loss of data for, and fix its machine learning models
Again, they already know how to do this, and they do it all the time. There are no nuances to it outside of roadblocks installed by FB themselves.
It's 110% a business problem, and 110% Zuckerberg's choice to make it difficult, and to invent difficulties in bringing it about.
This is not a science problem and it is not hard to solve: a scientist would delete the data and re-derive their models because they want the best data possible. This is how you "preserve the usefulness of those models," by ensuring that the data has integrity.
facebook has to figure out a way to understand exactly how thousands of records per user, being deleted, will affect it's models.
This is looking through the telescope backwards. You create your models with the data that you have, and if some data is removed then you regenerate.
You can bet your ass they groom their data after detecting stuff like new botnets and fake clicks, and individual user data does not have to be different.
That is an entire community and classification of people that facebook has to figure out how to reconcile the loss of data for, and fix its machine learning models
Again, they already know how to do this, and they do it all the time. There are no nuances to it outside of roadblocks installed by FB themselves.
It's 110% a business problem, and 110% Zuckerberg's choice to make it difficult, and to invent difficulties in bringing it about.