Hacker News new | past | comments | ask | show | jobs | submit login

The author states it's about 100GB of data.



Is 100GB of data really enough to warrant this type of architecture?


if you read more closely, this article is not about the amount of data, but the way how it is managed.


Precisely. While "all NYT data since 1851" sounds like a lot, and >8660 days sounds like a long retention period for a Kafka topic, this--like most systems in the world--is not a Big Data application. One of the key insights from the post is that there are interesting architectural considerations that have nothing to do with data size that make immutable logs a good idea.




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: