Skip Navigation

Is anyone using awk?

Studying and awk came up.

Spent about an hour and I see some useful commands that extend past what "cut" can do. But really when dealing with printf() format statements is anyone using awk scripts for this?

Or is everyone just using their familiar scripting language. I'd reach for Python for the problems being presented as useful for awk.

54 comments
  • awk will always have a soft spot for me, but I can see why not many take the time to learn it. It tends to be needed right there at the border of problem complexity where you are probably better using a full-fledged scripting tool.

    But learning awk is great for that "now you're thinking in pipes" ah-hah moment.

  • awk predates perl as well as python by a pretty large margin (1978); it's useful, of course, for processing things in a pipeline, but as it became obsolete as a general-purpose scripting language, users have had less and less of a reason to learn its syntax in detail -- so nowadays it shows up in one-liners where it could be replaced by a tiny bit of cut.

    I had worked through a good bit of the O'Reilly 'sed & awk' book -- the first programming book I got, after being enticed by shell scripting in general. Once I learned a bit of Python, & got better at vim scripting, though, I started using it less and less; today I barely remember its syntax.

  • Nearly every day. There was a time when I'd reach for Ruby, but in the end, the stability, ubiquity, and portability of the traditional Unix tools - among whom awk is counted - turned out to be more useful. I mainly underuse its power, though; it serves as a column aggregator or re-arranger, for the most part.

  • I use awk all the time. a very common and probably simplest reason I use it is it's ability to handle variable column locations for data.

    if you know you always want the last field you can do something like

    awk '{print $NF}'

    but usually using it as for performing more advanced operations all in one go without having to pipe something three times.

    sure you can do grep cut grep printf, but you can instead do the pattern matching, the field, the formatting, whatever you need to all in one place.

    it's also got a bunch of more advanced usage of course since it's its own language. one of my favorite advanced one liners is one that will recognize if it is going to print a duplicate line anywhere in your output and prevent it. in cases where you don't want to sort your output but you also want to remove duplicates it is extremely helpful and quick rather than running post-processing on the output in another way.

    all that said main reason I use it is because I know it and it's fast, there's nothing you can do in awk that you can't do in Python or whatever else you're more comfortable with. The best tool for the job is the one that gets it done quickly and accurately. unless your environment is limited and it prevents the installation of tools you're more familiar with then there's no real reason to use this over Python.

  • Yes! Awk is great, I use it all the time for text processing problems that are beyond the scope of normal filters but aren't worth writing a whole program for. It's pretty versatile, and you can split expressions up and chain them together when they get too complicated. Try piping the output into sh sometime. It can be messy though and my awk programs tend to be write-only

  • I learned awk when I was studying and I still use it every now and then. It's one of those tools that come in really handy at times. I work in Ruby, but there's still times when scripting or just wanting to process some text when I end up using awk because it makes sense.

  • With embedded like OpenWRT on a router where you only have busybox/ash shell, awk is your primary expansions tool.

  • awk is pretty damn solid. When I was completely rewriting the gravity.sh script from Pi-hole about six years back, it was easily the fastest for parsing and uniquely sorting content from files with a couple million lines. It made things much more usable on Raspberry Pi Zero hardware, since changing to another language like Python was out of the question.

  • Everyday. I’ve got a lot of stuff that uses it. Granted most of it was mostly created a decade ago but with minimal maintenance it works great. The most helpful script is parsing megacli outputs so I can get a heads up on drive failures and rebuilds among other things.

  • Just had to use it today to turn a key file into a single string with line breaks:

    awk 'NF {sub(/\r/, ""); printf "%s\\n",$0;}' id_rsa

  • awk is awesome! I love it, and I do not regret learning how to use it.

    That said, my workflow invariably always shifts from starting with awk to do something simply with a tiny one-liner, to then doing that with perl or python, and sometimes even creating a file to make the by-now multi-line scripts more easily readable.

    I do not recommend starting with awk, if you do not know other languages already such as Python.

    In short, let your intuition guide you.

54 comments