top of page

“Computing in the Third Dimension” and my new role at Anorak Ventures

I’m pleased to announce that I have joined Anorak Ventures as a Partner, working with Managing Partner Greg Castle to invest in and support exceptional founders in emerging technology (more about me here). I’ve described Anorak’s area of specialization as “Computing in the Third Dimension” – in this post I explain what that means, why it’s novel, and how it will impact the future.

Trapped in a box: the two-dimensional computing interface

The history of computing is widely understood as a series of “eras” of increasing power, each with their definitive leaders:

  • The mainframe era, led by IBM

  • The personal computer era, led by Microsoft

  • The Internet era, led by Google

  • The current ubiquitous computing era, led by Apple in devices, Facebook and Google in consumer services, and Amazon in cloud computing

  • The AI era, which is still in its infancy

Each of these eras made computers simultaneously more powerful and less expensive, making computing more accessible. Cheaper silicon birthed the personal computer era, broadband adoption unlocked the Internet era, and the launch of Amazon Web Services in 2006 and the iPhone in 2007 kicked off the ubiquitous computing era. Through these eras, computers have consistently become faster/better/stronger every year: from VisiCalc’s 254-row limit to petabyte-scale data lakes, or from Usenet posts to Skype calls to FaceTime, computers have gained a bigger role in our lives as they have become more powerful and easier to use.

Despite the onward march of technological power, our experiential interfaces with computers have stagnated in a two-dimensional paradigm. The original Apple Macintosh shipped 38 years ago with a mouse, keyboard, monitor, and printer – the same user interface that we use today.

We still work with computers through an interface invented in 1968 and popularized in 1984.

Smartphones introduced the multitouch interface, but still on a two-dimensional screen. Our entire mental model of software revolves around two-dimensional actions like clicking, dragging, and scrolling. Tellingly, the organizing principle of Web design is the “box model,” forcing every element on every website into the confines of a “box.”

But our sensory systems, and our minds that integrate their input, are inherently three-dimensional and spatial. Written text is ~5,000 years old and pictorial art is ~50,000 years old; spatial reasoning is over 50 million years old, and our most highly developed information interface. We can easily walk through a cocktail party and identify the conversations that are interesting to us, or walk through an office and tune into the right conversations to stay informed. Without spatial reasoning – if we simply listened to all of these overlapping conversations in an audio recording – it would sound like an incoherent jumble. Through two pandemic years of sitting on Zoom, staring at each other in little boxes, we’ve each learned for ourselves that two-dimensional computing simply cannot capture or represent the vibrancy of our three-dimensional world.

On two-dimensional computing surfaces, we lose our mental superpowers and our communication superpowers. Our sarcastic remark is misunderstood as sincere; our request for clarification is misunderstood as a passive-aggressive attack. As a result, our physical selves inhabit an entirely different world from our digital selves, and our lives feel strongly bifurcated between “IRL” and “online” interactions.

We want our online interactions to feel "real" -- they can certainly have major consequences in the physical world - but our two-dimensional online interactions rarely have the emotional tenor of our IRL interactions. After two pandemic years limited to primarily online interactions, restaurants, airports, and highways are packed with people seeking the richer texture of the physical world.

The way forward: computing in the third dimension

The good news is we are in the dawn of a major computing transition as important as the advent of the Internet. Computing, having been “trapped” for decades inside the world of structured databases and two-dimensional inputs and output, is stepping out into the physical, three-dimensional, rough-edged world. At Anorak Ventures, we call this trend “Computing in the Third Dimension,” and some of its pillars include:

  • Computers are understanding the physical world with computer vision and artificial intelligence, capturing much deeper insights with far less manual data entry

  • Computers are acting in the physical world with robotics, turning our understanding of the world into tangible outcomes

  • Computers are creating synthetic worlds through virtual reality and augmented reality, creating experiences for users that have all of the vibrancy, communication bandwidth, and emotional timbre of physical-world experiences inside entirely constructed environments

  • Computers are using generative AI to supercharge these synthetic experiences, allowing users to “construct their dreams” with experiences unattainable in the physical world, but sensorily indistinguishable from reality.

In all four of these areas, the common thread is that the interface boundaries between digital and physical experiences are being dissolved, bringing the power of technology into the physical world with unprecedented scale, and bringing the power of the physical world into the technological domain with unprecedented detail and subtlety.

Computer Vision and Artificial Intelligence

Computing has always been a tool for calculation, record-keeping, and analysis, and their correctness has always depended on the correctness of their inputs. In 1864, Charles Babbage, the father of computing, wrote:

“On two occasions I have been asked [by members of Parliament] - ‘Pray, Mr. Babbage, if you put into the machine wrong figures, will the right answers come out?’ I am not able rightly to apprehend the kind of confusion of ideas that could provoke such a question.”

Less eloquently referred to as “garbage in, garbage out,” data entry has long been a critical business function. When data entry was performed by humans, and double- or triple-checked by other humans, only high-business-value data even got digitized in the first place. The IRS digitized its collection operations to catch tax evaders, manufacturers used ERP software to manage their operational planning, and financiers digitized capital markets to gain better visibility and control of their risks and opportunities. But for every occurrence in the world that was digitized, billions or trillions of undigitized interactions went completely unrecorded.

This started to change when computers started doing their own data entry, starting in earnest in the 1970s. Banks started using optical character recognition to automatically record check numbers, and retailers used barcode scanners and later RFID scanners to automate tracking and inventory. These technologies lowered the costs of data acquisition, but only for pre-defined scenarios with standardized data schemas like scanning known and labeled objects at a cash register.

Recent advances in sensing hardware and machine learning have vastly increased the surface area of automatic data capture and analysis. Instead of setting up our world for computers, by adding barcodes and RFID tags to products and placing scanners in employees’ hands, the data acquisition and analysis can run passively without a human in the loop. Target’s security cameras can track a box of diapers from the warehouse to the store to the trunk of your car. The hardware to acquire data, such as cameras and accelerometers, are getting cheaper and more power-efficient, while the machine learning algorithms that analyze this data are getting increasingly powerful and able to extract higher-level insights. This enables new human interfaces like interactive voice and gesture recognition, as well as software that can analyze and react to data without any human interface.

Is this a good thing? Do we need or want to have an analysis of every time we sneezed, every dog that barked at us, or every blade of grass that we walked on? Perhaps not, but Anorak portfolio company SafelyYou is using computer vision to make our world safer for vulnerable populations.

SafelyYou is solving the extremely difficult problem of senior citizens being injured by falls. Falls are the leading cause of death for adults over 65, and even in nursing homes, where assistance is available, falls often go unnoticed because a resident cannot call for help after they have been injured by a fall. SafelyYou monitors a camera installed in the senior’s room, and can detect when they have fallen and immediately summon help. Not only can SafelyYou alert caregivers to a fall, but it can prevent falls – video review showed that one particular resident had fallen three times by sitting on the edge of her bed while watching TV, and simply putting her TV in front of the chair stopped the problem entirely.

It would have been prohibitively expensive, and intolerably intrusive, for a senior to be monitored in their room 24 hours a day by a human being. Computer vision and artificial intelligence is turning the entire physical world into an input surface, allowing vastly more information about the world to be ingested, processed, and acted upon.


Tightly coupled with computer vision/artificial intelligence is robotics. CV/AI is a big step forward in understanding the world; robotics helps us turn that understanding into action.

Robotics is certainly not new – low-intelligence robots have been used for over 50 years in automotive factories to perform spot welding and to move heavy objects into place. Robots have even used computer vision for decades, such as in agricultural sorting to separate out unripe fruit. However, these robots were purpose-built for a single task, and often needed no intelligence or sensing feedback of any kind.

Today’s robots are vastly more versatile than first-generation robotics due to two major trends: the sensing hardware/machine learning trend described earlier, and the increasing power/decreasing cost of actuators: brushless motors, motor controllers, accelerometers, lithium ion batteries, and the inner-loop control software. Today’s robotics do not just mechanically perform an operation again and again - they can sense their environment and choose the right course of action situationally. The most well-known application of this is self-driving cars, but one of the most interesting applications to us is using robotics to conserve valuable natural resources.

Anorak Ventures’ portfolio company Irrigreen has developed a robotic irrigation head that uses hardware and firmware similar to those you would find in an inkjet printer to “print” a precise pattern of water onto the surface of a lawn. Over 30% of America’s municipal water goes towards watering lawns, and close to half of this water is wasted by traditional “plastic stick” lawn sprinklers that can only water in circles and thus have to be wastefully overlapped.

Irrigreen’s robotic lawn sprinkler eliminates waste and overlap by a tight orchestration of software and hardware. After the user configures the shape of their lawn on their smartphone app, the Irrigreen system uses rain forecasts and soil moisture readings to water the lawn precisely as much as needed, adjusting the angle of the head and the water flow rate as the head sweeps out a full circle:

Irrigreen's digital sprinkler head "prints" a precise pattern of water to minimize waste.

Sensing hardware, actuator hardware, controller hardware, embedded software, machine learning, and cloud computing all work together to deliver the experience that the Irrigreen customer sees on their smartphone app. Because of these interlocking pieces, robotics companies like Irrigreen are tremendously complicated to build and operate, but the founding teams who can successfully do so (and it is almost always a team, with diverse skill sets and work experiences) can deliver value that pure software simply cannot.

Robotics is turning the entire physical world into a computing output surface to match the rich input interfaces that computer vision and AI have enabled. In tandem, AI and robotics are allowing computers to, in many cases, even exceed humans in their ability to sense and to act. The AI in your Apple Watch can detect that you’ve fallen down and have an elevated heart rate; a robotic drone can now fly a defibrillator to you and save your life.

Virtual Reality and Augmented Reality

Virtual reality (VR) is a technology that may eventually eclipse the Internet in its impact on societies, economies, and human lives. I’ve written more about my most optimistic hopes for virtual reality and the reasons that I believe it’s poised to massively break into the consumer mainstream.

The long-term goal of VR has always been to convincingly emulate any experience. If a person sees a dog in front of them on their VR headset, and can pet the dog and feel its fur with their haptic glove, and can hear it bark, and can form a friendship with the dog over time… is it functionally any different from a real dog? That’s really a question for the philosophers, such as Robert Nozick and his thought experiments with his Experience Machine.

Philosophy aside, the Experience Machine is already here. Even Meta’s $299 Quest 2 can transport users into virtual worlds by feeding into their three-dimensional spatial faculties rather than as a two-dimensional windowed experience. When I play Beat Saber for even a few minutes, the feeling of being in an infinite space is so strong that I’m surprised (and a little disappointed) when I take off my headset and find myself in an ordinary room. The impact of VR is even stronger in social interactions, where the illusion of presence creates interactions that feel vastly more real than 2D video calls.

Anorak portfolio company Innerworld takes advantage of not only the increased immersion of social VR, but also the added psychological safety of a remote and anonymized connection. Innerworld offers personal coaching through VR using the techniques of cognitive behavioral therapy (CBT), but in a lower-cost, peer-to-peer model available to those who cannot afford a licensed therapist. This model, called Cognitive Behavioral Immersion, is not only more accessible than licensed therapists, but has specific advantages born of the VR delivery model. The sessions are completely anonymous, which could never happen in a physical service model, and this anonymity allows people to openly discuss topics that they find challenging to discuss in person, even with a licensed professional.

VR is Anorak’s first and heaviest focus area: Managing Partner Greg Castle invested in Oculus’ seed round in 2012, and less than two years later, Facebook had acquired the company for $3 billion, making Oculus the first of six unicorns so far in the Anorak portfolio. Oculus created the modern virtual reality renaissance, and we continue to invest heavily in the VR sector (OssoVR, PrismsVR, Rec Room, and many others).

The dawn of generative AI

Rather than a trend already well underway, like AI, robotics, and virtual reality, generative AI is in its absolute infancy, but accelerating explosively. OpenAI’s DALL-E 2 can construct an image from only a text prompt:

DALL-E 2 creation from only the caption: "teddy bears working on new AI research on the moon in the 1980s"
DALL-E 2 creation from only the caption: "teddy bears working on new AI research on the moon in the 1980s"

… while NVIDIA’s Neural Radiance Fields can synthesize a virtual 3D environment from only a few seconds of scanning:

Created by Karen X. Cheng. Link to Tweet

It doesn’t take a large leap of imagination to simply “speak” a virtual world into existence with a short prompt and experience it in VR. People will be able to spend time with their deceased loved ones, live out alternate lives and entire realities, experience historical events as though they were real, and enjoy experiences like space travel that would otherwise would be attainable only to the narrowest elite. Anorak Ventures does not yet have any portfolio companies in generative AI, but we are eager to invest in this sector.

Computing in the Third Dimension and the future of human-computer interaction

After 38 years of the mouse, keyboard, and monitor, computing is finally breaking free of the two-dimensional interface, and the boundaries between the physical and the virtual worlds are rapidly collapsing. In the next five years, we expect to see:

  • Continued improvement in AIs that source proprietary datastreams and derive insights from these datastreams

  • A Cambrian explosion of robotics, both in form factors and applications, to do everything from services to industrial manufacturing to healthcare

  • An increasingly greater amount of our “screen time” dedicated to VR, and VR being the best way to remotely establish the human connection that was so often found lacking in remote work during the COVID-19 pandemic

  • AI-driven flights of fancy that turn our wildest dreams into virtual worlds we can explore and eventually inhabit

I’m extremely excited to join Anorak as Greg’s first partner and look forward to investing in the founders who are building this world. If you are one of these founders, let's get to know each other:

1,029 views0 comments

Recent Posts

See All


bottom of page