Breaking The Physical Limits Of Fonts Save

Breaking the physical limits of fonts

Project README

Breaking the physical limits of fonts

The challenge: in the fewest resources possible, render meaningful text.

  • How small can a font really go?
  • How many bytes of memory would you need (to store it and run it?)
  • How much code would it take to express it?

Lets see just how far we can take this!

Crash course in bitmaps

Computers typically represent images as bitmaps. The term bitmap is not to be confused with the .bmp file extension which is something quite different. It's just a way to store pixels in memory.


Images often contain multiple planes. Each plane refers to one "layer" of the image. Most images contain three planes in the RGB color space. One for red, green and blue. However images with transparency often contain another plane called alpha. You can think of a colored image as actually three (or four for transparency) grayscale images in one.

  • There are other color spaces than RGB. JPEGs use YUV for instance. We're going to focus on RGB since it's what most people are familiar with.

There's two ways to represent this in memory. You can store each plane separately, or you can interleave the planes. When the planes are interleaved, we use the term channels instead. Interleaving is the most common method today.

Imagine we had a 4x4 image, we could represent the three planes like this, where R is red, G is green and B is blue, respectively.

  R R R R
  R R R R
  R R R R
  R R R R

  G G G G
  G G G G
  G G G G
  G G G G

  B B B B
  B B B B
  B B B B
  B B B B

This would be storing of the planes separately.

Where as doing something like this would be interleaving.

  • Each cluster of characters is exactly one pixel
  • The order I've decided to put them in is called RGB order, there's other orders like BGR where blue comes first. The most common order is RGB however.

I've taken some creative freedoms of laying out the pixels in a 2D fashion so it's more obvious how it maps, however computers don't actually have 2D memory, their memory is 1D. So the above 4x4 example would actually be:



The term bpp is used to refer to either how many "bytes" or "bits" are used "per-pixel". You may have seen 24bpp or 3bpp used before, these are the same thing. One means 24 bits per pixel, the other means 3 bytes per pixel. There's always 8 bits in a byte. Since bit is the smaller unit, it will always be the larger number which is how you can tell if bits or bytes are being used as the unit.


The most common format you see today is 24-bit color i.e. 24bpp, or 3bpp. Here's a nice way of showing what that looks like at the bit level for one pixel in RGB order.

bit   0  1  2  3  4  5  6  7  8  9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
pixel R  R  R  R  R  R  R  R  G  G  G  G  G  G  G  G  B  B  B  B  B  B  B  B
  • There's one byte for R, one for G and one for B.
  • A byte can hold a value from 0 to 255.

So if we have a single pixel with the following color:

  • R 255
  • G 80
  • B 100

Then the first byte stores the value 255, second stores the value 80 and the last stores the value 100.

You may have seen hex, hexadecimal or base-16 representations of color before. Like #ff5064. This is a much nicer way of thinking about color. Each two characters refer to each channel of color. In this case; R is 0xff, which is 255 in decimal. G is 0x50, which is 80 in decimal and B is 0x64, which is 100 in decimal.

  • One nice property of hexadecimal representation is that each byte of a color is represented by two characters, that means each character expresses exactly 4 bits. Every 4 bits is called a nibble.


When the pixels are layed out flat like that and each one has multiple channels. It can get really confusing because we can't tell when the next row of pixels in our image starts. That's why we need to know the dimensions of the image and the bpp to make sense of it. The term stride is used to refer to the amount of bytes there are in one scan-line (row of pixels). In our example we have a 4x4 image, each pixel is 3bpp, that means we have a stride of 4*3 bytes, or generally, w*bpp bytes.

  • It's not always true that an image has a stride of w*bpp, sometimes images have "hidden" padding pixels that are meant to keep the image's dimensions a power of some size. As an example, working with power of two's is far easier and faster when scaling images. So you may have an image that is 120x120 pixels (that you can see) but the actual representation is 128x128 and those "padding" pixels are skipped over when previewing the image. We won't worry about that here.

This is a very simple mapping, for any pixel (x, y) the 1D location of it is (y * w + x) * bpp. Think about the math of this for a moment, y is the row of the pixel, we know each row has w pixels, so y * w moves us through the rows, x moves us along that row to the specific x coordinate on that row. However, that only works when we have one byte per pixel, since we interleaved R, G and B, we need to actually move in units of what ever bpp is, so multiplying the whole thing by 3, or bpp in the general case gives us the 1D location of the pixel. More specifically, it gives us the location of the first channel for that pixel. Reading exactly bpp bytes from that location gives us a whole pixel.

The font atlas

The way displays actually display pixels is through the use of three sub-pixels, one red, one green and another blue. If you zoomed in real close to a single pixel what you'd see is something that resembles this depending on your display.

The one we're interested in exploiting is LCD as that's likely the display technology you're reading this on.

There's some caveats of course

  • Not all displays have this subpixel pattern, some might put blue before red resulting in a BGR pattern.
  • If your display is rotated (phone or tablet) this pattern will be rotated too and this font will stop working.
  • Different subpixel patterns and orientations actually require different subpixel rendering of fonts themselves.
  • This will not work on AMOLED displays since they use pentile subpixel patterns. These displays are by far the most common displays on mobile devices too.

This method of exploiting subpixels for additional rendering resolution is called subpixel rendering.

For more information on subpixel font rendering, check out this excellent resource.

Fortunately for us, someone has already had this idea and built something called millitext. Their work is listed here.

They built the following tiny image by hand.

Which if you look at your monitor close enough looks like.

Here it is scaled up 12x.

That brings us to this really tiny image which is actually a font atlas built off his work, each 1x5 pixels represents a character. The atlas was created by hand and layed out like so.


That might be difficult to see, so I've scaled it up 12x

Which works out to exactly 36x5 pixels in size. The PNG is also 4bpp, since it also has an alpha channel, but we will be ignoring that. Assuming we store each pixel as RGB, we'd need exactly 36*5*3 bytes, or 540 bytes to represent this as a bitmap. That's actually surprisingly good already, since the PNG itself is actually:

# wc -c < font.png 

27x larger!

  • This command tells us how many "bytes" are in a file on a Linux PC.
  • This is 14 KiB!

PNG is not well suited for storing things like this since it's already too small. In fact a BMP can do a much better job, look:

# wc -c < font.bmp
  • Edit: It turns out our pesky PNG file contains a ton of metadata that can be stripped making it far smaller, we also have an unnecessary alpha channel. Running the PNG through tools like pngcrush and optipng really push things to the limits.
# wc -c < font-crushed.png

We can still take this further ourselves with a slightly different approach.


The acute of you may have noticed something interesting about the atlas, there's only seven colors in it, these colors in particular:

  1. #ffffff
  2. #ff0000
  3. #00ff00
  4. #0000ff
  5. #00ffff
  6. #ff00ff
  7. #ffff00


When we only have a few colors like this, it's often easier to create a palette and refer to colors in the palette instead of each pixel being the color value itself. Assuming we use the palette above, then each pixel only ever needs to be represented by a single value in the range 0-6.

  • 1-bit can represent 2 possible values (0, 1)
  • 2-bit can represent 4 possible values (0, 1, 2, 3)
  • 3-bit can represent 8 possible values (0, 1, 2, 3, 4, 5, 6, 7)

If we represented each pixel as a 3-bit quantity, where the value of that pixel referred to our palette, we would only need 68 bytes to represent the entire atlas.

  • The data compression folks out there might point out that you can have such a thing as a "fractional bit", the perfect size we actually need here is 2.875 bits. This is often accomplished through something called entropy coding arithmetic coding. However, this is quite complicated and 68 bytes is already incredibly small.


There's an ugly problem with 3-bit encoding though. It does not divide evenly into a byte. A byte is the smallest addressable unit computers can actually deal with. Imagine we have these three pixels:


If each one takes 3-bits, then two bytes would look like this in memory, where - denotes an unused bit.

bit   0  1  2  3  4  5  6  7  8  9 10 11 12 13 14 15
pixel A  A  A  B  B  B  C  C  C  -  -  -  -  -  -  -

See that, the C pixel takes up one bit of the next byte, it's split across the bytes and in fact, as we start adding more pixels, they can be split up and straddle anywhere! Here we just have 1-bit inside another byte, but in practice a pixel can end up 2-bits inside another byte, 3-bits inside another byte, etc.

An easy solution to this problem would be to use a nibble per pixel, since 4 divides evenly into 8, this would keep everything aligned on a byte, allowing exactly 2 pixels per byte, but it also takes our atlas up from 68 bytes to 90 bytes, which is 1.3x larger.

  • We can actually go far smaller if we exploit the fact that the top-half of some characters are the same as their bottom half (mirroring). Similarly, the use of range coding and other very specific compression techniques can probably get this down far more. We'll leave that for a later writeup.

Bit buffer

Fortunately it's still possible to work with 3-bit quantities, it just requires keeping track of which bit in a byte you are at when encoding and decoding.

Included here is a simple class which writes 3-bit quantities into a byte array.

  • To keep this as accessible as possible for readers, the code will be written in JS but can be extended to other languages.
  • All code will assume Little Endian byte order since that's the most common.
class BitBuffer {
  constructor(bytes) { = new Uint8Array(bytes);
    this.offset = 0;
  write(value) {
    for (let i = 0; i < 3; ) {
      // bits remaining
      const remaining = 3 - i;

      // bit offset in the byte i.e remainder of dividing by 8
      const bit_offset = this.offset & 7;

      // byte offset for a given bit offset, i.e divide by 8
      const byte_offset = this.offset >> 3;

      // max number of bits we can write to the current byte
      const wrote = Math.min(remaining, 8 - bit_offset);

      // mask with the correct bit-width
      const mask = ~(0xff << wrote);

      // shift the bits we want to the start of the byte and mask off the rest
      const write_bits = value & mask;

      // destination mask to zero all the bits we're changing first
      const dest_mask = ~(mask << bit_offset);
      value >>= wrote;
      // write it[byte_offset] = ([byte_offset] & dest_mask) | (write_bits << bit_offset);

      // advance
      this.offset += wrote;
      i += wrote;
  to_string() {
    return Array.from(, (byte) => ('0' + (byte & 0xff).toString(16)).slice(-2)).join('');

So, lets load in that atlas PNG, ignore the alpha and encode it into our bit buffer, we'll use png-js for this.

const PNG = require('png-js');
const fs = require('fs');

// this is our palette of colors
const Palette = [
  [0xff, 0xff, 0xff],
  [0xff, 0x00, 0x00],
  [0x00, 0xff, 0x00],
  [0x00, 0x00, 0xff],
  [0x00, 0xff, 0xff],
  [0xff, 0x00, 0xff],
  [0xff, 0xff, 0x00]

// given a color represented as [R, G, B], find the index in palette where that color is
function find_palette_index(color) {
  const [sR, sG, sB] = color;
  for (let i = 0; i < Palette.length; i++) {
    const [aR, aG, aB] = Palette[i];
    if (sR === aR && sG === aG && sB === aB) {
      return i;
  return -1;

// build the bit buffer representation
function build(cb) {
  const data = fs.readFileSync('subpixels.png');
  const image = new PNG(data);
  image.decode(function(pixels) {
    // we need 3 bits per pixel, so w*h*3 gives us the # of bits for our buffer
    // however BitBuffer can only allocate bytes, dividing this by 8 (bits for a byte)
    // gives us the # of bytes, but that division can result in 67.5 ... Math.ceil
    // just rounds up to 68. this will give the right amount of storage for any
    // size atlas.
    let result = new BitBuffer(Math.ceil((image.width * image.height * 3) / 8));
    for (let y = 0; y < image.height; y++) {
      for (let x = 0; x < image.width; x++) {
        // 1D index as described above
        const index = (y * image.width + x) * 4;
        // extract the RGB pixel value, ignore A (alpha)
        const color = Array.from(pixels.slice(index, index + 3));
        // write out 3-bit palette index to the bit buffer

build((result) => console.log(result.to_string()));

After all that work, we now have a single bit buffer containing our atlas in exactly 68 bytes.

To put that in perspective, here's the original PNG

# wc -c < font.png 

We're 222x smaller!

That's not a mistake. We've compressed something down to 0.45% it's original size!

  • Edit: With the appropriate crushing of the original PNG. We're ~6x smaller now.

Now lets convert the representation to a string so we can embed it into our source code. That's essentially what the to_string method does. It reads off the contents of each byte into a single base-16 number.


However, this is still quite long to embed. Fortunately this is because we've limited ourselves to base-16 which has an alphabet of 16 characters. A better encoding method for this is actually base-64, which gives us 4x more characters, so lets change to_string to use that.

to_string() {
  return Buffer.from('base64');

Which then gives us:


Meaning we can now embed this single string in our JS and begin rasterizing text.

Rasterizing text

We want to only decode one glyph at a time to reduce the memory usage, that can be done quite trivially

const Alphabet = '0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZ';
const Atlas = Uint8Array.from(Buffer.from('MFAAAAwDKNbUsky0bVFtTdq2aZJqDdq2Udt2FQBgAJwCheagdS21kFRlW9e1adJqTdugU4kqADBgQA0jKFC0CmthrQA=', 'base64'));
const Palette = [
  [0xff, 0xff, 0xff],
  [0xff, 0x00, 0x00],
  [0x00, 0xff, 0x00],
  [0x00, 0x00, 0xff],
  [0x00, 0xff, 0xff],
  [0xff, 0x00, 0xff],
  [0xff, 0xff, 0x00]

// at the given bit offset |offset| read a 3-bit value from the Atlas
read = (offset) => {
  let value = 0;
  for (let i = 0; i < 3; ) {
    const bit_offset = offset & 7;
    const read = Math.min(3 - i, 8 - bit_offset);
    const read_bits = (Atlas[offset >> 3] >> bit_offset) & (~(0xff << read));
    value |= read_bits << i;
    offset += read;
    i += read;
  return value;

// for a given glyph |g| unpack the palette indices for the 5 vertical pixels
unpack = (g) => {
  return (new Uint8Array(5)).map((_, i) =>
    read(Alphabet.length*3*i + Alphabet.indexOf(g)*3));

// for given glyph |g| decode the 1x5 vertical RGB strip
decode = (g) => {
  const rgb = new Uint8Array(5*3); 
  unpack(g).forEach((value, index) =>
    rgb.set(Palette[value], index*3));
  return rgb;

The decode function here gives us our original 1x5 strip for the given glyph g. What's most impressive is we only need 5 bytes of memory to decode a single character into memory. Similarly, we only need ~1.875 bytes of memory to read such a character to begin with. Giving us an average working set of 6.875 bytes. When you include the 68 bytes used to represent the atlas and 36 bytes to represent the alphabet string, we're capable of drawing text with less than 128 bytes of RAM theoretically.

  • Writing this in assembler or C would allow you to actually see these savings

Now all that's left is a way to compose these vertical strips into an image for the purposes of drawing some text.

print = (t) => {
  const c = t.toUpperCase().replace(/[^\w\d ]/g, '');
  const w = c.length * 2 - 1, h = 5, bpp = 3; // * 2 for whitespace
  const b = new Uint8Array(w * h * bpp);
  [...c].forEach((g, i) => {
    if (g !== ' ') for (let y = 0; y < h; y++) {
      // copy each 1x1 pixel row to the the bitmap
      b.set(decode(g).slice(y * bpp, y * bpp + bpp), (y * w + i * 2) * bpp);
  return {w: w, h: h, data: b};

With that we've broken the physical limits of fonts.

const fs = require('fs');
const result = print("Breaking the physical limits of fonts");

Use some imagemagick to get a readable image in a format that you can actually preview.

# convert -size 73x5 -depth 8 rgb:73x5.bin done.png

Here's the final result

Here it is scaled up 12x

Here it is on a poorly calibrated display.

Here's a much cleaner image on a properly calibrated display.

Open Source Agenda is not affiliated with "Breaking The Physical Limits Of Fonts" Project. README Source: graphitemaster/breaking_the_physical_limits_of_fonts
Open Issues
Last Commit
5 years ago

Open Source Agenda Badge

Open Source Agenda Rating