I felt the same. The most important thing for a new library or language to do in its introduction is to show meaningful examples that solve a problem in the target domain in the canonical way.
I’m excited to look at this but the comparisons are misleading! E.g., NPY has existed since 2007 (I believe it was the first NumPy RFC) and is exceedingly popular; JSON can, of course, represent multi-dimensional arrays (you can’t get much simpler or readable than [[1, 2], [3, 4]]); I also don’t understand “Pythonic experience” for JSON since dictionaries are so ubiquitous; I could go on …
The description of this is kind of confusing but I think the easiest way to understand it is that it is a data processing pipeline of sorts. Take unstructured data and apply transformation and computation. A similar project to this is Towhee (https://github.com/towhee-io/towhee). This project tries to simplify unstructured data processing and provides pretrained models and pipelines from their hub.
"If you are a deep learning engineer who works on scalable deep learning services, you should use DocArray: it can be the basic building block of your system."
I mean, wow. The basic building block of your system. The very nucleus of any scalable deep learning service! But, what is it?
There is no such thing as unstructured data. If it were unstructured it would be noise.
What people call unstructured data is, when they had little to no foresight when designing the original data structures. Any new data fields are just pilled on later and called unstructured
Then again, I guess folks were similarly confused in the past:
https://news.ycombinator.com/from?site=jina.ai
https://news.ycombinator.com/from?site=github.com/jina-ai