Hacker News new | past | comments | ask | show | jobs | submit login

A nice analysis, and it shows how SQL makes it easy to quickly explore data.

However, it seems like the plots of the results of the queries were done manually by writing some code to make each plot.

I can't stop but mention that using Apache Zeppelin Notebook [1] with Postgres interpreter for Zeppelin [2] (Spark SQL should provide comparable analytical capabilities as well, but this comment is not about it) it is possible to show graphical representation of query results without writing a single line of code.

[1] https://zeppelin.apache.org/

[2] https://zeppelin.apache.org/docs/latest/interpreter/postgres...




Author here! For the graphs I used the ggplot library in R, but really any SQL visualization tool (e.g., SQLPad) would have worked as well. Apache Zeppelin Notebook seems like a great tool and I'd be happy to look into it for my future work. Thanks for the recommendation! :)


I just started looking for tools like this and came across plotly and dash: https://medium.com/@plotlygraphs/introducing-dash-5ecf7191b5...

Am I right that zeppelin is similar to this?


I don't get it, don't you need to write code in Zeppelin to have it show plots? You need to decide which data which aggregation.. do you have a link with any doc showing how to plot without writing code? I followed your links but I miss this.


Look at this part of the video from https://zeppelin.apache.org: https://www.youtube.com/watch?v=_PQbVH_aO5E&t=3m21s. It shows plotting directly from SQL.


I've only used zeppelin for spark so far, but it's been a joy to work with.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: