Lecture 4 - Data Wrangling
View lecture View lecture notes
Exercises
-
Take this short interactive regex tutorial.
Self explanatory.
-
Find:
-
The number of words (in
/usr/share/dict/words) that contain at least threeas and don’t have a'sending.sed’sycommand, or thetrprogram, may help you with case insensitivity.cat /usr/share/dict/words | tr "[:upper:]" "[:lower:]" | grep ".*a.*a.*a.*" | grep "[^('s)]$" | wc -lThe output of the command above turns out to be
639.trconverts all uppercase letters into lower case.grepwith.*a.*a.*a.*regex will match 3 or morea's. You can have a look at the regex debugger to try out other test cases. Similarly, wegrepagain with the[^('s)]$regex. This will avoid all words by using the('s)capture group.$marks the end the line which in this case is also the end of the word. Finally,wcwill count all lines with these filtered words. -
What are the three most common last two letters of those words?
cat /usr/share/dict/words | tr "[:upper:]" "[:lower:]" | grep ".*a.*a.*a.*" \ | grep "[^('s)]$" | sed -E "s/^.*(..)$/\1/" | sort | uniq -c | sort -nk1,1 | \ tail -n3 | awk '{ print $1 }'We append extra sub-commands to the previous command. With
sedwe pluck out the last two letters from each word. Then wesortthe results. Then we pipe the STDOUT touniq -cwhich gives back only unique instances of all the lines fed to it along with the count of repititions. Thensort -n k1,1sorts the incoming list by the 1st column. To get the three most common last two letters of those words we use thetail -n3command.awk '{ print $1 }'strips off the 2nd second column which shows the frequency and prints only the first column. The output of command above is as follows.al ia an -
How many of those two-letter combinations are there?
cat /usr/share/dict/words | tr "[:upper:]" "[:lower:]" | grep ".*a.*a.*a.*" \ | grep "[^('s)]$" | sed -E "s/^.*(..)$/\1/" | sort | uniq -c | sort -nk1,1 | \ tail -n3Command is same as above. Only the
awkpart is removed in order to print both columns. The output of the command above is as follows:49 al 51 ia 101 an
-
-
To do in-place substitution it is quite tempting to do something like
sed s/REGEX/SUBSTITUTION/ input.txt > input.txt. However this is a bad idea, why? Is this particular tosed? Useman sedto find out how to accomplish this.The shell makes the file on the right of the redirection symbol
>empty before writing the output of the the expression on the left. Hence,sedscans through an empty file if the above mentioned command is used.To achieve the required functionality we can use the
-ioption which allows the replacement to be done inplace in the text file. Hence the following command can be used:sed -i "s/REGEX/SUBSTITUTION/" input.txt -
Find your average, median, and max system boot time over the last ten boots. Use
journalctlon Linux andlog showon macOS, and look for log timestamps near the beginning and end of each boot. On Linux, they may look something like:Logs begin at ...and
systemd[577]: Startup finished in ...On macOS, look for:
=== system boot:and
Previous shutdown cause: 5Functionality not available on WSL. Will update as soon as I move to a full-fledged Linux system.
-
Look for boot messages that are not shared between your past three reboots (see
journalctl’s-bflag). Break this task down into multiple steps. First, find a way to get just the logs from the past three boots. There may be an applicable flag on the tool you use to extract the boot logs, or you can usesed '0,/STRING/d'to remove all lines previous to one that matchesSTRING. Next, remove any parts of the line that always varies (like the timestamp). Then, de-duplicate the input lines and keep a count of each one (uniqis your friend). And finally, eliminate any line whose count is 3 (since it was shared among all the boots).Functionality not available on WSL. Will update as soon as I move to a full-fledged Linux system.
-
Find an online data set like this one, this one, or maybe one from here. Fetch it using
curland extract out just two columns of numerical data. If you’re fetching HTML data,pupmight be helpful. For JSON data, tryjq. Find the min and max of one column in a single command, and the difference of the sum of each column in another.curl -o data.html \ https://ucr.fbi.gov/crime-in-the-u.s/2016/crime-in-the-u.s.-2016/topic-pages/tables/table-1