The year is 2032. Former LAPD sergeant John Spartan wakes up after three decades in deep freeze sleep to find everything has changed, including driving.

Almost all the cars in the 1993 movie Demolition Man were self-driving, but of course they were concept cars. Today, more than ten leading car makers (including Audi, BMW, GM, Tesla Motors, Volkswagen, and Volvo) are working on self-driving cars. In addition, Google has developed its own custom vehicle which recently racked over 1 million miles (“the equivalent of 75 years of typical U.S. adult driving”) without any major incidents.

Demolition Man cars

Looking at these impressive achievements, many are asking a pertinent question: how long will it be before self-driving cars will be a common occurrence on our roads?

Imagination Technologies is uniquely positioned at the start of a complex automotive food chain that includes semiconductors vendors, system integrators, software developers and car makers. All of these companies are closely working together on the underlying ADAS (advanced driver assistance systems) technologies that will steer the market on the road of adopting the first commercially-available driverless vehicles.

There are currently three main approaches to implement ADAS functionality inside the car.

One direction relies on storing huge amounts of map data which is then used by the car to navigate around a given environment; think of this method as a train riding on a set of invisible tracks. An example of this approach is the Google driverless car which essentially navigates using only on a set of pre-recorded, street-level HD maps and very little sensing. In this instance, the car would rely solely on high-speed connectivity and sensor aggregation, maintaining a constant link to a cloud mega-structure that supplies the navigational coordinates.


In contrast, another technique relies only on computer vision processing and little to no use of pre-recorded maps. This approach replicates a human driver because the car is able to make real-time decisions based on the rich sensors and high-performance processors aboard. This category of vehicles typically includes multiple cameras for a wide field of view and uses specialized high-performance, low-power chips that offer the processing power of a supercomputer to implement ADAS software and hardware algorithms.

A company pioneering this approach has been Mobileye with its powerful and energy-efficient EyeQ SoCs designed specifically for autonomous cars; for example, a MIPS-based Mobileye EyeQ3 SoC handles the processing required for the highway autopilot feature recently enabled on Tesla cars. In this case, Mobileye have been using a multicore, multi-threaded MIPS I-class CPU to handle the data streams coming from the multiple cameras installed on the vehicle.

In the block diagram below, the quad-core, quad-threaded interAptiv CPU inside the EyeQ4 SoC acts as the brain of the chip, directing the flow of information from the cameras and other sensors to the 2.5 TFLOPS of processing muscle represented by the VLIW blocks on the right.

Mobileye EyeQ4 architecture - MIPS interAptiv M5150 CPU

Finally, a third trend involves the reuse of general-purpose SoC processors for autonomous driving. For cars that already come equipped with an embedded GPU for infotainment purposes, developers can take advantage of the compute resources of a graphics engine to run algorithms that execute recognition and tracking of lanes, pedestrians, vehicles, building facades, parking lots and more. For example, Luxoft’s Computer Vision and Augmented Reality solution uses our PowerVR Imaging Framework and additional software to implement ADAS functionality. The software framework is optimized for embedded PowerVR-based hardware that helps implement a wide range of in-car autonomous driving functionality quickly and cost-efficiently.

PowerVR imaging framework - ADAS demo

To address the requirements of next-generation automotive applications, Imagination recently released PowerVR Series7 – a family of GPUs that implement 10-bit support and hardware virtualization, two important features for ADAS developers. Having 10-bit color depth inside the GPU enables an automotive maker to build a high-precision capture-to-display pipeline that includes the image sensor, the video encoder/decoder and the graphics engine.

PowerVR vision platformThis 10-bit capable multimedia subsystem can detect lanes, traffic signs or other road markings more accurately than traditional 8-bit solutions. Implementing hardware virtualization offers the ability to run two or three applications in separate containers reliably and securely. For example, an SoC combining a MIPS Warrior CPU and a PowerVR Series7 GPU can optimally execute both the Rightware Kanzi UI infotainment cluster and the Luxoft ADAS solution concurrently on the same chip, providing best-in-class performance and reliability for both.

No matter the approach, self-driving vehicles represent a bright future for automotive companies, sensor providers and the semiconductor industry in general. These cars can benefit from advanced SoC modules and high-performance microcontrollers featuring hardware memory management, multi-threading and virtualization capabilities. These capabilities allow OEMs to implement more sophisticated software applications, including model-based process control, artificial intelligence and advanced vision computing.

I’m really excited to see Imagination and our partners playing a central role in developing hardware and software solutions for autonomous driving in 2016 and beyond.

For more news and updates on MIPS, follow us on Twitter (@ImaginationPR, @MIPSGuru, @MIPSdev), LinkedIn, Facebook and Google+.