Typography Classification in Augmented Reality v1.1

#User Experience

As we are progressing in augmented reality space the complexity of information is increasing with the introduction of more features and functions. The existing typographic rules and structure of information in AR is not sufficient to solve for new variables (know more).

Now the text is not just limited to static consumption, new challenges like movement, rotation, rendering of text (frame rate, resolution) are bringing up several issues like perspective distortion, distance reading, distortion of letter shapes etc. However, some of these challenges are not new at all, although the context has changed because of the three-dimensional medium. Example: learnings from highway signages typography can be applied to scenarios where you want to convey quick information to a user wearing AR glasses while walking on the road. In this case, directly translating the guidelines might not work perfectly because there are limitations of text rendering, the brightness of displays and so on. And this is where the typography has to evolve and solve these novel challenges.

Anchoring of Information

Before I jump onto explaining the text classification I would like to explain anchoring of text which controls the behaviour of text in different scenarios.

  • Anchored to head: In this case, the information moves along with the head movement of the user and always stay in front of him/her.

Figure 1: The virtual objects follow the head movement and stay in the view all the time.

  • Anchored to space: The virtual elements are anchored to real-world coordinates in 3D space around the user. Hence the information stays at a particular position and the user sees it only when he/she is looking in its direction.

Figure 2: The virtual object stays at a particular position and doesn’t move with the head movement.

Placement zones

I have divided the user’s view into three regions based on the distance and priority of information that can be displayed in these regions.

Figure 3: The distances are approximations made based on different studies and guidelines for AR displays and ongoing testing by me.

Heads-up Display (HUD) Region

User Interface (UI) Region

Environment (World) region:

Classification of type of text

I have been working on the classification below to understand and define different scenarios and the consideration that should be kept in mind while choosing typefaces, setting text and even designing typefaces for different applications. The classification can also help you pick the right type of methods of rendering (know more) for displaying your text in your AR experience and explore the functionality of variable font in different cases. Know about future possibilities in this video.

Figure 4: Classification table which describes various states and parameters linked to different types of text.

Text in HUD

In this case, the text sticks to the field of the view of the user and moves with the along with the user’s head movement.

Figure 5: Source: Unknown

Text for long reading

It should ideally be placed in the UI region for better reading experience within the range of 5 metres. You can allow the users to move the text to optimize the distance based on their reading preferences.

Figure 6: Credit: Ultraleep

Sticky info text

The text which is anchored to the real-world objects (usually in a close range up to a few meters away from the user) that has a fixed orientation. What it means is that it changes position, direction based on how the user interacts with the real world object it is anchored to(micro-level interactions in real-world).

Figure 7: Credit: Keiichi Matsuda

Signage text

It is similar to the sticky info text which anchored to the real world object but in this case, the user can’t move or change the orientation of the object. The information is anchored to macro-level objects like geo-location, buildings, vehicles etc.

Figure 8: Credit: Keiichi Matsuda

Responsive text

It can be placed in both UI and environment regions where it changes its orientation (perspective) based on the user’s movement or specific programmed behaviours.

Figure 9: Credit: Rohit Parasa

Ticker text

The term ticker comes from news tickers, the element which has moving text information. Quite a useful method to attract attention and show more information in a small area. Eg: notifications in HUD, showing information in supermarkets.

Figure 8: Credit: Keiichi Matsuda

Parameters that make these text types unique

Viewing angle:

Fixed: the text remains right in front of the user and maintains the angle of view when the user is moving towards or around it. In this case, there is no perspective distortion.

Variable: the text stays in a fixed orientation and viewing angle changes based on the position of the user. Perspective distortion of text is a major issue here.

State of Text

Stationary: The most familiar way of consuming text is when it is static and right in front of us. Eg. text on shop signage, billboard etc.

Moving: Cases where the text constantly moving. Mostly used to solve issues like showing more info in limited space or to attract user’s attention. Eg. news and share market ticker strips.

User’s State

Still: When the user is sitting or standing still at a particular location and consuming the information. Most of the text we consume currently is designed keeping in mind the static state of the user, be it reading a book or reading something on screens.

Moving: With augmented reality expanding to the real world where it becomes part of the day to day life this is a new challenge to make the text consumable in when the user is in motion (walking/running). The closest case we have right now is HUDs in cars.

Note: All the information in this article is based on my ongoing research and some aspects might change and get updated as I move further.

Know more about Future of Typography

Get the latest updates on this ongoing AR research directly to you mailbox!