A Bear, A Beer, and Good Taste

Say you go camping. You take a cooler or two of your favorite suds. And then some party animal comes along… and in the dead of night drinks up your stuff.

And say that party crasher is a bear…

And say he knows good beer from bad…

Do you invite him back? ( I think I would… but I’d need an extra case or two…)


Posted 8/18/2004 7:50 AM

Bear drinks 36 cans of favorite beer
BAKER LAKE, Wash. (AP) — Rain-eeeeer …. Bear? When state Fish and Wildlife agents recently found a black bear passed out on the lawn of Baker Lake Resort, there were some clues scattered nearby — dozens of empty cans of Rainier Beer.

The bear apparently got into campers’ coolers and used his claws and teeth to puncture the cans. And not just any cans.

“He drank the Rainier and wouldn’t drink the Busch beer,” said Lisa Broxson, bookkeeper at the campground and cabins resort east of Mount Baker.

Say what? I knew bears had a sensitive nose, but really? Maybe he just didn’t find it…

Fish and Wildlife enforcement Sgt. Bill Heinck said the bear did try one can of Busch, but ignored the rest.

“He didn’t like that (Busch) and consumed, as near as we can tell, about 36 cans of Rainier.”

OMG! Well. What can I say. One taste of Budwiper and he’s not touching another.

I think I’ve gotta get me some ‘bear tested and bear approved’ Rainier… Wonder if they will put that in their advertising?… Maybe that bear can get a job as a beer tester?…

Agents then used a large, humane trap to capture it for relocation, baiting the trap with the usual: doughnuts, honey and, in this case, two open cans of Rainier. That did the trick.

“This is a new one on me,” Heinck said. “I’ve known them to get into cans, but nothing like this. And it definitely had a preference.”

So, when going Bear Hunting, bait your trap with Good Beer. And if you want beer that not even the Bears will drink, get Bud…

I just can’t help visualizing the Super Bowl Commercial of a couple of bears sharing a couple of cases with the game on the in the background as they taste one Bud and spit it out… with the tag line of “based on a true story… honest!” maybe with a Cameo of Sgt. Heinck swearing it is true…

Subscribe to feed

Posted in Humor | Tagged , | 2 Comments

Cloning USHCN Daily Data

This is just a quick little posting on how to mirror or rapidly clone all the data in a particular subdirectory of a web site (and a smaller sidebar on how not to…)

First, you need really really ought to have ;-) a Unix or Linux machine for this to work. It isn’t all that hard to get one or build your own. The one I’m using for this exercise cost me $60 “all up” and took me about 6 hours to assemble and get working (and that was without really trying and with a lot of gratuitous playing along the way – a ‘production set up’ would likely take 1/2 that if the play and beverage and photo breaks were left out…)

Directions for the DIY low cost Linux Machine set up are here:



This script is pretty simple. Really just one line of the “wget” or ‘web get’ command. I’ve put a load of explanatory comments in front of it (lines starting with #) but you really can ignore all of them and just use the one line.

I have the basic command running fine. I had not tested this specific example of the script at the time of the original posting for the simple reason that the machine had been busy and the last ‘enhancement’ (also known as a ‘one line bug fix’) had not been tested, but I was pretty sure it was all correct. I’ve now tested it and it is working as expected.

With that, here’s the “script”. It is in a file named “getushcn” and with execute permissions set by doing a “chmod +x getushcn” (changes the file mode to add the eXecutable bit). I then print it out with the ‘cat getushcn’ command (concatenate a list of files and print out the contents – but with only one file, just print it).

cat getushcn 

# Fetch a mirrored copy of the CDIAC USHCN Daily temperature data.
# wget is the command that does the fetching.  
# It can be fed an http: address or an ftp: address.
# The -w or --wait command specifies a number of seconds to pause 
# between file fetches.  This helps to prevent over pestering a 
# server by nailing a connection constantly; while the 
# --limit-rate={size} limits via an average of pausing between 
# transfers.  Over time this is about the rate of bandwidth used, 
# but on a gaggle of small files can take a while to stabilize, 
# thus the use of both.
# Since CDIAC uses a "parent" link that points "up one" you need 
# to not follow those or you will end up duplicating the whole 
# structure ( I know... don't ask...) thus the -np or 
# --no-parent option.
# The -m or --mirror option sets a bunch of other flags (in effect)
# so as to recursively copy the entire subdirectory of the target 
# given in the address.  Fine, unless they use 'parent' a lot...
# Then you list the http://name.site.domain/directory or 
# ftp://ftp.site.domain/directory to clone
# Long form looks like:
# wget --wait 10 --limit-rate=100k --no-parent --mirror http://cdiac.ornl.gov/ftp/ushcn_daily
# but I think the --commands look silly and are for people who can't
# keep a table of 4000 things that -c does in 3074 Unix / Linux 
# commands, all of them different, in their head at all times ;-) 
# so I use the short forms.  Eventually not typing all those wasted
# letters will give me decades more time to spend on useful things,
# like comparing the merits of salami vs. prosciutto... 

wget -w 10 --limit-rate=100k -np -m http://cdiac.ornl.gov/ftp/ushcn_daily

And yes, you could just type

wget -np -m http://cdiac.ornl.gov/ftp/ushcn_daily

when in the desired directory and be done… which is what I did last night, but without the -np… but where’s the fun in that? Putting things in a file and calling it a script means never having to make the same typo twice ;-)

I launched this last night (minus the-np and the -w and the –limit-rate) and it ran for about 12 hours cloning away, having copied about 11 GB so far of everything both below the “ushcn_daily” directory and above, thanks to them having a “parent” link in their directories… and me not having the –no-parent set. At this point I’m curious about just how long it will take so I’m was letting it finish. On the one hand, that’s a bit rude as I’m sucking up resources for not much reason. OTOH, it is rate limited by my slow pipe, so not a big impact on them, and once the mirror is done, I need never pester their site again as I have a local cached copy and even if I do an update on the whole thing, only the changed files would be re-sent.

So in the interest of showing everyone just how long it will take if you screw up like I did and leave out the –no-parent setting, I’m I was going to let it complete and then ‘fess up’ (and / or tell the horror story… instead of just junking it incomplete and having no value at all).

UPDATE: Due to a power dropout of about 5 minutes, the ‘let it run’ has ended and I see no reason to restart it. Just accept that it takes fractional days to days if you forget that ‘no parent’ flag… END UPDATE.

Most of the time it was using my full speed of about 230 K/s without the bandwidth limiting flags, but the protocol is polite, so doesn’t seem to bother normal browser use much. Meaning those rate limit and wait settings are a nice idea, but likely not really needed. It is the polite thing to do, though. With those flags set, usage can be tailored to any degree desired. At ‘less than 1/2 the capacity’ it isn’t noticed much at all.

From my watching it last night, cloning just the daily data portion took fairly little time. About an hour, I think, but hard to tell for sure.

pi@RaPiM2 ~/ushcnd/cdiac.ornl.gov/ftp/ushcn_daily $ du -ks
1045568	.

It has about 1 GB of data in it, so divide that by your data rate and that’s how long it will take.

Once mirrored, future runs compare date stamps and only re-copy a file when it has changed. (For most files, some small ones just get resent). IFF you want a rolling archive, you will need to take snapshots prior to re-running the script and updating.

As my R.Pi is built on a 64 GB SD card, using 10 GB, or even 20 GB for a full archive of their whole site (however big it ends up being) costs about $1 / GB. So for a very modest sum you can clone the data archive at a point in time. About $1.05 for the USHCN daily directory and that has two copies of the data in it (by State and in a lump).

Note, too, that this will work for any site that uses the FTP or HTTP file transfer protocols. Just change the link


to point to a different sub-directory or site. It is my intent to make one of these for each of the major temperature data archives and then assess the operational work required to keep a current copy and selected archive copies.

As posted, the script will mirror into the current working directory, so make sure you have picked a place you like before running it. In the final production version, I will have a line at the top of the script to assure things go where desired. For me, that will be a directory named /MIRRORS and the first line in the script will be:


That way, whenever I run it, the results will always end up in that same target directory.

I’ve left it out of the posted version so that you can use this more as an interactive tool to make a copy wherever you like. Once the whole process is set up, I will have an automated script launch (sometime in off hours… dead of night or one weekend a year) that will do the mirror into the ‘usual place’ and make any archival backups desired. The future versions for other data sets and the ‘background job’ will also be posted so folks know how to set that up if desired.

But for now, it’s a ‘one off’ script you can run on your own and in any directory.

The wget command has dozens of other options, so whatever you want it to do, there is likely an option for that. Those wanting to do more can type “man wget” at the command prompt in a terminal window on their Unix / Linux machine.

Happy Cloning!

Subscribe to feed

Posted in AGW and GIStemp Issues, GISStemp Technical and Source Code, Tech Bits | Tagged , , , , , , | 7 Comments