Python Bytes and Characters

Last week I lost two days of my life fighting with Python, encoding, and Supervisor. It all starts with the fact that in Python2 doesn’t handle encodings very well. I’ll leave the discussion of how Python2 handles encodings at that because Python2 is dead at the end of 2019 but suffice it to say that in Python2 bytes and characters were mostly handled the same and it wasn’t ideal because oftentimes characters are made up of multiple bytes.

At work we use Supervisor to control hundreds of programs across about ninety servers. It lets us assign processes to a host and then ensure that they are running. One feature that we take heavy advantage of is the ability to run event listeners inside Supervisor to capture things like logging events from our programs and then run those log lines through a monitoring system and eventually to an event management system.

Until last week we had been running Supervisor version 3 on Python 2. According to the documentation for Supervisor event listeners your event listener communicates with Supervisor by listening on stdin and talking back on stdout. Both stdin and stdout are, by default in Python 2 and Python 3, character streams. In Python2 this matters less but in Python3 this matters more. Supervisor will send new events to your event listener by sending you one line that your listener will read using readline. That one line will have a standard formatting that includes a len field and the len field should, according to the documentation, indicate the number of bytes that you should then read from stdin to get the body of the event. Because Python2 didn’t really distinguish between bytes and characters, this totally worked with Supervisor3 on Python2. My event listener looked roughly like this:

stdin = sys.stdin.buffer
while handle in[stdin], [], [])[0]:
    line = handle.readline()
    if (not line):
        raise EOFError("received eof from supervisord")

    # decode the line to utf8
    line = line.decode("utf-8", "backslashreplace")

    # read the parts of the header line
    header = dict([x.split(":") for x in line.split()])
    data_length = int(header.get("len", 0))
    if (data_length == 0):
        return (header, None, None)

    # read more to get the payload
    data =
    if (not data):
        raise EOFError("received eof from supervisord")

    # decode the data to utf8
    data = data.decode("utf-8", "backslashreplace")

    if ('\n' in data):
        # this message has additional data so extract it out
        event, data = data.split('\n', 1)
        event = dict([x.split(":") for x in event.split()])
        return (header, event, data)
        event = dict([x.split(":") for x in data.split()])
        return (header, event, None)

Basically we use the byte stream version of stdin (stdin.buffer), wait for data to come in to stdin then we readline it, get the payload length, and then read that number of bytes from the buffer and repeat for the next event. With Supervisor3 running on Python2 and our event listener running on Python3 this worked great and was in line with how the documentation for Supervisor said that everything should work.

As said, last week we upgraded Supervisor to version 4 and started running it on Python 3. Things suddenly stopped working but very, very randomly. After a lot of digging we discovered that our event listener was reading in less data than it was supposed to. The event listener would say that ten bytes of data were in the buffer and the listener would read ten bytes and then the next time the listener ran readline it would get garbage data.

It turned out that any time we had a program that logged something that contained variable byte characters (i.e. UTF-8 data) that our event listener would break. After digging through the code for Supervisor it was discovered that this was because Supervisor was sending the listener data that was already encoded in UTF-8 and the length it was giving us was characters and not bytes. UTF-8 is a variable length character encoding. Each character can be one or more bytes. For the vast majority of normal characters (e.g. A through Z, 0 through 9) each character is one byte and so we weren’t experiencing a problem. Some of our programs, however, were logging data that went beyond the normal A through Z and those ended up being more than one byte per character thus why the event listener was reading less data than there actually was.

The solution to this was to turn our byte stream into a character stream so that read would work in characters instead of bytes. Thus we ended up with this solution which does not work and you should keep reading to see why:

stdin = io.TextIOWrapper(sys.stdin.buffer, encoding="utf-8")
while handle in[stdin], [], [])[0]:
    line = handle.readline()
    if (not line):
        raise EOFError("received eof from supervisord")

    # read the parts of the header line
    header = dict([x.split(":") for x in line.split()])
    data_length = int(header.get("len", 0))
    if (data_length == 0):
        return (header, None, None)

    # read more to get the payload
    data =
    if (not data):
        raise EOFError("received eof from supervisord")

    if ('\n' in data):
        # this message has additional data so extract it out
        event, data = data.split('\n', 1)
        event = dict([x.split(":") for x in event.split()])
        return (header, event, data)
        event = dict([x.split(":") for x in data.split()])
        return (header, event, None)

It is pretty identical except that we wrap stdin in a Python3 built-in that converts everything off of the stream into UTF-8. Then we don’t need to do the conversion later and when the event listener calls read it reads characters and not bytes. Problem solved?

After I implemented this the next day we were having a different problem. Event listeners on various servers started blocking. Instead of declaring that they were reading garbage data they just stopped reading data entirely.

It turns out most of our programs print data the normal Unix way: write some data, add a line feed (aka LF aka \n) to the end, repeat. Some of our programs, though, were echoing to their logs data that they got from remote sources that might include line endings other than a simple line feed and instead might print the Windows standard “carriage return line feed” or “CR LF” or \r\n.

By default, io.TextIOWrapper implements what Python calls universal newlines. This “feature” annoyingly converts anything that even looks like a new line into \n so that \r\n becomes \n. Now when a program prints a log line that is ten characters long and has \r\n in the middle of it the TextIOWrapper converts that \r\n into \n. Subsequently the stream reader in the event listener receives nine characters and blocks forever waiting for a tenth character that never will appear.

Thanks Python for trying to be helpful but ultimately not doing what I expect. The solution is to add the argument newline='\n' to the io.TextIOWrapper line so that the wrapper passes you the raw data and doesn’t try to mess with newlines.

Now I have a bug report open on the Supervisor project to address this either by fixing Supervisor to send bytes like the documentation says or by fixing the documentation to say “characters” like is implemented. I would prefer that they send me bytes and let me do the encoding conversion since maybe my programs are printing something non-UTF-8 like CP1256 or Mac Roman or whatever. Right now if your program spits out something that isn’t valid UTF-8 they will replace your log line with the word “Undecipherable” and followed by some object representation of the bytes and maybe I know what it is and I’d like to convert it to something I find useful. Let’s wait and see what happens!

Software Projects

Throughout my time at the University of Washington I have had the privilege of being able to build a lot of systems that underpin large portions of the University’s computing infrastructure. I’m starting to pull some of those systems out of the University’s source control, make some modifications to them such that they are generally useful, and then document and publish them here.

To that end I have put together a projects page with some of the things I’ve done worth highlighting. It is far from complete.

However, to this point it has information on the clone system that I built out, plus some PostgreSQL programs and monitoring tips. I’m hoping to get the software deployment system push up soon, since it is mostly done already, as well as the Supervisor command and control system dart that I recently rebuilt at work up there really soon now, too. Then I’ll get around to creating a publicly usable event management system based on the one I created at UW plus the network device monitoring system that I built at UW as well.

So I just wanted to share that finally, after talking about it for a few years, I’m moving forward with some of this sharing stuff. I also still intend to port over a bunch of my old blog posts, too. But that turned out to be much, much harder than it looked and my video games look so nice after sitting at a desk all day. Soon enough.

Python SSL Socket Server

I recently had to build a small server application in Python. It did not need to be anything complicated. It needed to run on about one hundred servers and receive a tiny command to do something and then be done. A web server would have been overkill and a was definitely not available on all of the hundred servers. Writing a socket server in Python is pretty trivial and the documentation includes example code for you, too. The caveat that I had to deal with is that I needed to validate that the client was who they said they were and I wanted to do it with an SSL certificate so that SSL would handle all of the authentication for me. (The authorization would still have to be handled by me.)

The documentation in Python for writing an SSL server is all over the place. With each version of Python 3 the library has changed in some subtle way that deprecates what was previously the preferred way so if you’re going to do this verify that what I’m showing you here is up to date. I’m pretty certain that this code is valid in Python 3.7, though we are running it in a 3.6 environment.

First, the server.

import socketserver
import ssl

class RequestServer(socketserver.ThreadingMixIn, socketserver.TCPServer):
    # faster re-binding
    allow_reuse_address = True

    # make this bigger than five
    request_queue_size = 10

    # kick connections when we exit
    daemon_threads = True

    def __init__(self, server_address, RequestHandlerClass, bind_and_activate=True):
        super().__init__(server_address, RequestHandlerClass, False)

        # create an ssl context that using the cert
        # that requires the client to present a certificate and
        # validates it against uwca.
        ctx = ssl.create_default_context(ssl.Purpose.CLIENT_AUTH)
        ctx.verify_mode = ssl.CERT_REQUIRED

        # replace the socket with an ssl version of itself
        self.socket = ctx.wrap_socket(self.socket, server_side=True)

        # bind the socket and start the server
        if (bind_and_activate):

class RequestHandler(socketserver.StreamRequestHandler):
    def handle(self):
        print("connection from {}:{}".format(self.client_address[0], self.client_address[1]))

            common_name = self._get_certificate_common_name(self.request.getpeercert())
            if (common_name is None or common_name != ""):
                print("rejecting {}".format(common_name))
                self.wfile.write('{"accepted": false}\n'.encode())

            # now we're going to listen to what they have to say
            data = self.rfile.readline().strip()
            print("data: {}".format(data))
            self.wfile.write('{"accepted": true}\n'.encode())
        except BrokenPipeError:
            print("broken pipe from {}:{}".format(self.client_address[0], self.client_address[1]))

    def _get_certificate_common_name(self, cert):
        if (cert is None):
            return None

        for sub in cert.get("subject", ()):
            for key, value in sub:
                if (key == "commonName"):
                    return value

# this is the server. it handles the sockets. it passes requests to the
# listener (the second argument). the server will run in its own thread
# so that we can kill it when we need to
server = RequestServer(("", 3278), RequestHandler)

It listens on port 3278 and it listens for SSL connections. It will tell SSL clients that its hostname is “”. You should use whatever certificate it is that you have laying around for your server to identify itself.

You’ll notice the line that says “load_verify_locations” and the preceding line that says CERT_REQUIRED. This means that all incoming connections must present a client certificate and that certificate must have been signed by the CA indicated by “load_verify_locations”. This server will accept any client certificate signed by the UW Certificate Authority. That is the authentication component.

But I only want to allow connections from a certificate that I deem authorized. This is the authorization component. That’s what the private method called “_get_certificate_common_name” does. When given the certificate details from the client connection it will extract the client’s common name and returns that. We make sure that common name matches something authorized. In this case our server identifies itself as “” and only allows clients that are using that same certificate. (Is this a good idea? Probably not. But I don’t have the infrastructure to maintain lots of certificates for just this purpose. This is effective for me.)

What does a client look like to all of this? Super simple.

import socket
import ssl

ctx = ssl.create_default_context()
ctx.verify_mode = ssl.CERT_REQUIRED
ctx.check_hostname = True

with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as sock:
    with ctx.wrap_socket(sock, server_hostname="") as ssock:
        ssock.connect(("localhost", 3278))
        ssock.sendall(bytes("this is a test\n", "utf-8"))

This verifies that our server is presenting a valid UW Certificate Authority signed certificate. It also presents our server with the certificate with the common name “”. Finally, we tell our client that our server will identify itself as “”. If we didn’t set a “server_hostname” argument then the client would only validate the connection to the server if the server identified itself as “localhost” as that is the hostname we are connecting to. But our server is identifying itself as “” because that’s the certificate that we made it use.

One interesting note to this: I don’t know about the server code (because I haven’t tried) but the client code does NOT work with eventlet, unless I’m doing something wrong. We’ll find out when they respond to my issue request.

Vacation to Hawaii

For the first time since last summer I took an actual vacation where I traveled somewhere. Until now, the only state to which I had never been was Hawaii. Now I can say that I’ve been to all fifty states and that I’ve spent not-insignificant amounts of time in each of them. (Next goals: visit all ten provinces and three territories of Canada and all of the public islands in Hawaii. Maybe after that I’ll try out other countries.) Of course, I took pictures.

On the first day I just stayed in a hostel around Kailua Kona. The sky was clear, so I took this photo of the crescent waxing moon. (Also called an ‘ole kū kolu moon.) The moon had a very pretty halo around it.

A halo around the ‘ole kū kolu moon on March 13th, 2019.

The second day I went to South Point or Ka Lae, the southernmost point in Hawaii and the United States. (Wikipedia disagrees with me, but Palmyra is in a territory, not a state.) Interestingly, South Point is not that interesting. It’s on a cliff and very rocky and popular with people who have fishing rods.

Later in the second day I went into Hawai’i Volcanoes National Park. I stayed in a cabin just off of the national park but needed to enter the national park to check into the cabin. There I wandered through the steam vents along the Kīlauea caldera. It’s just remarkable to see steam just rising from the dirt. Then you can look out and see toxic smoke rising from a gigantic hole in the ground that periodically shoots molten rock in the air. Most of the area around the caldera was closed because of recent eruptions. However, there were no eruptions at the time of my visit and no lava was visible anywhere in the park. Very disappointing but probably safer.

The Pacific Ocean crashes against the cliffs at the southern end of Volcanoes National Park. These cliffs were made in the past seventy years.

On the third day I visited Hilo. Hilo felt like home in Seattle with its continually misty rain and lush greenery. On my tour of Hilo I stopped at the Kaumana Caves Park, a county park that features entrance to a totally uncontrolled lava tube. Seriously, you just walk in to this pitch black cave at your own risk. My headlamp did not suffice and I actually relied on the flashlight functionality of my phone to see deep into the cave. You turn several times and duck through several small passages in the two and a half mile tunnel, completely losing sight of the entrance. That is, it is nothing like the caves in Virginia.

Spelunkers descend into the Kaumana Caves lava tube with their flashlights shining. The plant life hanging from the top of the cave is actually a volcanic glass fiber known as Pele’s Hair.

I stayed at the Inn at Kulaniapia Falls, an “off-grid” inn situated overlooking a waterfall. It was a really pleasurable experience and I would do it again. The staff are lovely, even if they did wake me up at 4am slamming doors in the office below my room. At least they compensated me for that!

Rainbows appeared in the morning light around the waterfall. It went into a large pool that then drained into several smaller waterfalls.

I missed the dinner reservation at Kulaniapia Falls and instead went into Hilo and had a fantastic dinner at Pineapples restaurant.

The fourth day was the most remarkable. I took a guided tour to the top of Mauna Kea, the tallest mountain in entire Hawaiian island chain at 13,803 feet above sea level. Another 17,000 feet or more of the mountain is below sea level making it the tallest mountain the world from its base. Fortunately the snow was gone from the top of the mountain this March.

The shadow of Mauna Kea against the clouds over which it looms.

The tour took us up the Hawai’i island Saddle Road where I snapped this picture of a tree that survived a lava flow while everything around it burned. Incidentally, this island is covered in wild goats and feral pigs and cats brought to in various waves by Polynesians, the British, and the Spanish. Additionally, this tree and the surrounding lava flow is actually pretty representative of the island of Hawai’i. Most of the island is pretty much covered in barren lava rock. The western side of the island receives hardly any rain and the rain is what begins to break down the lava rock into more fertile land facilitating growth. There’s also the fact that this island is still an active volcano so new lava is being laid down all the time.

A lone tree surrounded by lava.

Once at the top of Mauna Kea, you can see the most beautiful sunset that you’ve ever seen with more colors than you ever thought imaginable.

The sun setting into the clouds from the top of Mauna Kea.

The top of Mauna Kea, you may have heard, is covered with telescopes. In fact there are thirteen telescopes and a couple more antennas. Because the University of Hawaii actually destroyed the original summit, native Hawaiians designated a nearby peak as the peak where they would perform their rituals.

The actual summit to Mauna Kea and the trail up to it, open only to native Hawaiians.

After watching the sunset, the tour guides took us down to the Mauna Kea Very Long Baseline Array antenna, part of an array of ten radio antennas spanning the globe. That is where we got a tutorial on the stars and how the original inhabitants of the Hawaiian islands navigated by the stars. These tour guides were the best and shared a ton of information about Hawaii that I don’t know where else I would find.

Even with a half moon, you could still see a ton of stars by the VLBA antenna at the top of Mauna Kea.

And I had a terrifically disappointing journey back to Seattle via Delta where they managed to do everything wrong from start to finish. I also don’t regret renting a Jeep but I would have preferred that Avis had given me a Jeep that had been made in the last two or three years than the crap that they did give me. It didn’t even have support for Bluetooth. Still, I’d go again.

Glass Blowing

I recently took a glass blowing class at Pratt Fine Arts Center. The class was four hours every week for six weeks and we learned how to gather glass from the furnace, blow a bubble, gather again, and shape it into something that looks marginally close to a cup or whatever. Here are some of the things that I put together including two cups, a bowl, a vase-like thing, some ornaments, a pumpkin, and a chili pepper.

A “bowl”.
A “vase”.
Two cups!
The ornament on the left weighs about a half of a pound. The ornament on the right is all blown out of proportion.
This pumpkin is my second favorite. It involved a cast.
This is my favorite. It’s a chili pepper.

Views from the Viaduct

At near midnight on January 11th, 2019, the state of Washington closed the Alaskan Way Viaduct permanently after more than sixty years ferrying cars along the Seattle waterfront. The dull roar that made it impossible to hold a conversation in Victor Steinbrueck Park disappeared and with it a calm that the waterfront has not heard since the early 1950s. On the night of February 1st, the state closed the Battery Street Tunnel, which the viaduct used to feed into. (It was still being feed by ramps from Western Avenue after the viaduct’s closure.) And then on February 2nd the city and state opened up the new 99 Tunnel, the Battery Street Tunnel, and a portion of the upper deck of the viaduct to pedestrian traffic before the new tunnel opened on February 4th. I went for a tour. Turns out that taking pictures of a road while actually on the road is not quite as interesting, I don’t think, but here are a handful of photos.

Revelers walk through the poorly lit Battery Street Tunnel.
The lower deck of the viaduct, while lit, was closed to pedestrians. This is a view from the Seneca Street off ramp.
The Seattle Great Wheel from the viaduct with the Port of Seattle behind it. No, this is not really viaduct related.

Still More Black and White Photography

This week was a bit of a weird one with some travel mixed in with family visiting. As such I did not get to take as many photos that I considered to be any good as I had the previous two weeks. This week was also the week that our class didn’t have an outing slash field trip. So here are some photos from my trip to Los Angeles plus some of Duke and Seattle Center.

A life guard station on the beach in Santa Monica, California.
Despite being part beagle, Duke’s fur is only black and white.
The metal paneling on the outside of the Museum of Popular Culture aka MoPOP formerly known as Experience Music Project aka EMP.
Art work outside the Museum of Popular Culture. The color version isn’t that bad, either.

Backing Up Your Backups

Like anyone who cares about the data that is on his or her computer, I keep backups. I don’t have the backups run automatically, as is the default in primary operating system, macOS. I also don’t do backups very often. But I do want to keep my data safe, encrypted, and off-site but also still easily accessible. As you might guess, backups for me are a complicated, manual affair. This is how that goes.

(Why don’t I run automatic backups? Well, a lot of the work that I do on my laptop is with local virtual machines running inside VMWare Fusion. Virtual disks attached to virtual machines get very big on your actual disk. Small changes to the data on the virtual disk usually results in huge changes to the underlying files that back those virtual disks that need to actually be backed up. Finally, backing up an in-use virtual disk is not conducive to a quality restoration. So before I do a backup I will stop my virtual machines and before manually initiating the backup. Then I go to sleep while the backup runs.)

To start, my primary computer is a laptop with a one terabyte hard disk that is usually about half full and whose contents churn fairly constantly. I also keep a second hard disk to store about two terabytes of assorted other files — mostly old photographs — that I want to archive but no longer need to be on my laptop. All of my backups from the laptop also go to the second hard disk. This disk is kept in my living room so that I may easily connect it to my laptop to access the assorted archived data and also to more easily run the backup process. That’s easy enough.

But after an apartment building next door to mine burned down back in 2009 I started keeping a copy of my backup disk in a fire safe in my apartment. After my fire safe got a crack in the casing I worried that it might not be as reliable as I expected it to be so I also started keeping a second copy of my backup disk in my desk at work. When I moved to Seattle and realized that an earthquake or volcano might wipe out both my apartment and my office in the same event, I also stared keeping a copy of my backups in the cloud.

I actually only have two files that I actually need backed up. That’s right: two. On my unencrypted external disk I have an encrypted sparse bundle image of my laptop’s Time Machine backups and I have an encrypted sparse bundle image containing about two terabytes of the digital detritus — mostly old photographs — collected from my twenty plus years using a computer. Actually, since they’re sparse bundle images those two files are really two directories containing approximately 223,000 files, but, uh, close enough.

So my primary backup disk contains those two sparse bundle images. After I perform a backup those two files change. The next step is to replicate those two sparse bundle images to my backup’s backups. How is this done? For the backup disks that I keep in my fire safe and in my office, this is easy: rsync. With both the original backup disk and the backup backup disk connected to my laptop, I open up Terminal and run this command:

cd /Volumes/original
rsync -rlptgoDzhOHi --stats --numeric-ids --delete-during *.sparsebundle /Volumes/backup

So that’s easy enough. It copies the sparse bundle images from one disk to another. I run the rsync twice, once to each backup backup disk, transport them to their respective locations, and I’ve got my two backups. The cloud backups are a little bit more complicated.

After researching a number of the backup options — S3, Glacier, DropBox, etc. — they just really weren’t feasible on cost. Using S3, for example, to back up 2TB of data would cost me about $45 a month, plus the cost of the data transfer and that starts to push $600 a year. Glacier is nearly impossible to use. DropBox doesn’t let you store more than 2TB per month on their less expensive professional plan and the option that lets you store unlimited data costs $60 per month or $720 per year.

But I did find an option that lets you store unlimited data and doesn’t cost an arm and a leg: Backblaze B2 Cloud Storage. My two terabytes is costing me $10 per month and there is no cost to transfer the data to their system and no cost to restore the data from their system. (And when I ran the upload from my office the only limitation on upload performance was my laptop’s 1Gbps network interface. I was able to push three to four hundred gigabytes every hour.)

Because I’m such a fan of rsync it turns out that there is a similar option for backing up to the cloud: rclone. After I set up my storage space on Backblaze I created an access key for my laptop and configured rclone with it’s incredibly simple “config” command and now I just run this command:

rclone --transfers 32 --progress sync /Volumes/storage/compy.sparsebundle/ b2-lockaby:lockaby/compy.sparsebundle/

“b2-lockaby” is the rclone nickname for my Backblaze bucket. Unfortunately, wildcards for matching files doesn’t work so I have to run this command twice: once for each “file” that I am backing up. Still, it’s trivial.

But there are a few catches to get to this point. First, my backup disks are ALL unencrypted but I require that I only store my data encrypted. That’s why my sparse bundles are encrypted. So when I connect my unencrypted disk I have to then open and mount my encrypted sparse bundles before I can use the data. When I do the rsync and the rclone I unmount the encrypted sparse bundles. For the sparse bundle image full of random data it’s easy to see how to set this up and how this works. But for the Time Machine backup this isn’t as obvious.

If you’re using macOS and you want to back up your hard disk you have two options. The first option is to connect an external disk to the computer and back up to that. If you tell macOS to encrypt the backup it will convert the disk to a FileVault disk. The second option is to connect your computer to a network disk such as one attached to an AirPort Express or AirPort Extreme. If you use the second option and you tell it to encrypt your backups then macOS will create an encrypted sparse bundle image on the network disk.

But Apple, in its shortsighted wisdom, has discontinued the AirPort line. As a result, being able to run a network backup seems like something that is going to cease being supported in the not-too-distant future. So I decided that directly attached Time Machine backups were going to be the future for me. But I obviously don’t want to convert my external disk to FileVault because then I won’t have encrypted sparse bundles that I can upload to the cloud.

The solution is actually pretty easy but not well documented. First, create an encrypted sparse bundle image on your external disk. Next, mount it. Then issue this command:

sudo tmutil setdestination /Volumes/{mounted-disk-image}

Now you’ll see Time Machine try to back up to that mounted image. When you finish doing your backup through Time Machine, unmount the image, rsync and rclone the sparse bundle, unmount the disk, and go back to your daily life.

More Black and White Photography

At the second week of the black and white photography class that I am taking we did a field trip to the Seattle Museum of Flight. I also took some pictures around the University of Washington campus. I also stood in the pouring rain in the pitch black to take some pictures of rain falling past a street light. It’s been a long week.

Rain falls around a streetlight over Aurora Avenue in Seattle.
The covering over the large collection of airplanes at the Seattle Museum of Flight.
The covering over the large collection of airplanes at the Seattle Museum of Flight.
The blades on the jet engine on a Boeing 787 Dreamliner on display at the Seattle Museum of Flight.
These fancy pants jet turbine blades are on the demonstration engine for the Boeing 787 Dreamliner on display at the Seattle Museum of Flight.
The book stacks in the Suzzallo Library on the University of Washington campus.
A series of lights hang over the reading room in the Suzzallo Library on the University of Washington campus.

Black and White Photography

I began taking a black and white photography class at the North Seattle College continuing education program last week. One week in and it’s just been a bunch of people who are seemingly good photographers show me their photographs and saying “try looking at things this way!” But we had an assignment the first week and that assignment was to take some photographs in black and white and then share them. These are my first week photographs.

Starting off my black and white adventure nice and easy, these are simple lines on a heating vent. I used a macro lens to get the strange focus plane.
This is the view from the inside of John Grade’s Wawona sculpture that hangs in the Museum of History and Industry in Seattle.
These lights are above the lobby and the mezzanine of the UW Tower in U-District in Seattle.
I was pretty fascinated by the patterns that they formed with these lights.
The lights stretched all the way from one end of the lobby to the other and past the elevators and across two floors.