You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
One of my favourite tools of my trade is jq.
It essentially enables you to process json streams with the same power that
sed, awk and grep provide you with for editing line-based formats (csv, tsv,
etc.).
Hopefully a fun read about asynchronous, non-blocking operations in Mojolicious with concurrent events via Mojo::IOLoop
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
On the Unix command line, the message boundary is orchestrated in this manner: Files -> Lines -> Tab Separated Fields. Note that the Tab can be replaced by other characters, but by default it's a Tab.
Almost every tool expects a file as a collection of messages (lines), and that each line may be further separated in fields using a delimiter. The moment you want to get out of this message boundary architecture, is when things get complicated, and you start needing multiline regex... etc.
It's unfortunate that we haven't made use of ASCII FS nor ASCII RS nor ASCII US. It would make ASCII delimited text easier to parse, as these characters are not ever going to be used in actual text. See: https://en.wikipedia.org/wiki/Delimiter#ASCII_delimited_text
The focus of data science tools on the command line, are basically tools that encourage one-liners or support one-liner use case.