wget –spider –no-check-certificate –force-html -r -l2 https://website.of.interest.com 2>&1 | grep ‘^–‘ | awk ‘{ print $3 }’ | grep -v ‘\.\(css\|js\|png\|gif\|jpg\)$’ >> url-list.txt which will give something like: https://website.of.interest.com/en/Category:BRA https://website.of.interest.com/en/Category:Free_software https://website.of.interest.com/en/Category:Freeware https://website.of.interest.com/en/Special:RecentChangesLinked/Software_overview https://website.of.interest.com/en/Special:WhatLinksHere/Software_overview https://website.of.interest.com/en/Talk:Software_overview https://website.of.interest.com/ithelp/index.php?title=Software_overview&action=edit https://website.of.interest.com/ithelp/index.php?title=Software_overview&action=history https://website.of.interest.com/ithelp/index.php?title=Software_overview&action=info https://website.of.interest.com/ithelp/index.php?title=Software_overview&oldid=2586 https://website.of.interest.com/ithelp/index.php?title=Special:Pdfprint&page=Software_overview https://website.of.interest.com/ithelp/index.php?title=Special:UserLogin&returnto=Software+overview https://website.of.interest.com/ithelp/index.php?title=Talk:Software_overview&action=edit&redlink=1

If you have a file with several lines that contains email addresses that you like to extract, you can do that with Linux grep command: File “emails.txt”: somestreng anotherstring 342345 somename.lastname@domain.com where somename.lastname@domain.com is the string you like to extract: you can do this with: grep -EiEio ‘\b[A-Z0-9._%+-]+@[A-Z0-9.-]+\.[A-Z]{2,4}\b’ emails.txt Result: somename.lastname@domain.com

We have a Linux server (Redhat Enterprise 6) running a multi-site installation of Mediawiki. In total, we have today 120 unique wiki’s, each having its one Postgresql 9.1 database. Suddenly the server stopped responding. In Apache webserver, the total number concurrent apache processes went up to Apache’s max-client setting in httpd.conf, and people couldn’t log

I needed to get some quick random numbers that would represent a time on the form HH:SS. In my case the time should be between 10:00 and 12:55. I did this in bash with: #!/bin/bash # If the random number is 1, then LAST-number is set to 0. # Else it would be set to

I had to find a specific Microsoft Spreadsheet among thousands of files on a mapped network drive O: on a Windows 7 Computer. The problem was that searching and indexing was only performed on local disk like C: and D: so I could only search for keyword in local Excel files. It is not optimal,

I upgraded our Open Journal Systems from 2.2.2 to 2.4.2 and afterwards we saw the email logs were gone. This was related to: http://pkp.sfu.ca/support/forum/viewtopic.php?f=8&t=9140 My solution was this: Established a new Postgresql server in a virtual enviroment (vmware, test-machine), and import the old and the new (upgraded) database. # Log in to Prod server ssh