When Frito is driving Joe and Rita away from the cops, Joe happens to gesture with his hand above the car window, where a vending machine he happens to be passing spots the tattoo. Within seconds two harsh beeps sound in the car and a voice says, “You are harboring a fugitive named NOT SURE. Please, pull over and wait for the police to incarcerate your passenger.”
Frito’s car begins slowing down, and the dashboard screen shows a picture of Not Sure’s ID card and big red text zooming in a loop reading “PULL OVER”
The car interface has a column of buttons down the left reading:
At the bottom is a square of icons: car, radiation, person, and the fourth is obscured by something in the foreground. Across the bottom is Frito’s car ID “FRITO’S F’N CAR” which appears to be a label for a system status of “EVERYTHING’S A-OK, BRO”, a button labeled CHECK INGN [sic], another labeled LOUDER, and a big green circle reading GO.
But the car doesn’t wait for him to pull over. With some tiny beeps it slows to a stop by itself. Frito says, “It turned off my battery!” Moments after they flee the car, it is converged upon by a ring of police officers with weapons loaded (including a rocket launcher pointed backward.)Continue reading →
When using the Cookie to train the AI, Matt has a portable translucent touchscreen by which he controls some of virtual Greta’s environment. (Sharp-eyed viewers of the show will note this translucent panel is the same one he uses at home in his revolting virtual wingman hobby, but the interface is completely different.)
The left side of the screen shows a hamburger menu, the Set Time control, a head, some gears, a star, and a bulleted list. (They’re unlabeled.) The main part of the screen is a scrolling stack of controls including Simulated Body, Control System, and Time Adjustment. Each has an large icon, a header with “Full screen” to the right, a subheader, and a time indicator. This could be redesigned to be much more compact and context-rich for expert users like Matt. It’s seen for maybe half a second, though, and it’s not the new, interesting thing, so we’ll skip it.
The right side of the screen has a stack of Smartelligence logos which are alternately used for confirmation and to put the interface to sleep.
When virtual Greta first freaks out about her circumstance and begins to scream in existential terror, Matt reaches to the panel and mutes her. (To put a fine point on it: He’s a charming monster.) In this mode she cannot make a sound, but can hear him just fine. We do not see the interface he uses to enact this. He uses it to assert conversational control over her. Later he reaches out to the same interface to unmute her.
The control he touches is the one on his panel with a head and some gears reversed out of it. The icon doesn’t make sense for that. The animation showing the unmuting shows it flipping from right to left, so does provide a bit of feedback for Matt, but it should be a more fitting icon and be labeled.
Also it’s teeny tiny, but note that the animation starts before he touches it. Is it anticipatory?
EYE-LINK is an interface used between a person at a desktop who uses support tools to help another person who is live “in the field” using Zed-Eyes. The working relationship between the two is very like Vika and Jack in Oblivion, or like the A.I. in Sight.
In this scene, we see EYE-LINK used by a pick-up artist, Matt, who acts as a remote “wingman” for pick-up student Harry. Matt has a group video chat interface open with paying customers eager to lurk, comment, and learn from the master.
Harry wears a hidden camera and microphone. This is the only tech he seems to have on him, only hearing his wingman’s voice, and only able to communicate back to his wingman by talking generally, talking about something he’s looking at, or using pre-arranged signals.
Tap your beer twice if this is more than a little creepy.
A smaller transparent information panel for automated analysis, research, and advice.
An extra, laptop-like screen where Matt leads a group video chat with a paying audience, who are watching and snarkily commenting on the wingman scenario. It seems likely that this is not an official part of the EYE-LINK software.
The Viper is the primary space fighter of the Colonial Fleet. It comes in several varieties, from the Mark II (shown above), to the Mark VII (the latest version). Each is made for a single pilot, and the controls allow the pilot to navigate short distances in space to dogfight with enemy fighters.
Mark II Viper Cockpit
The Mark II Viper is an analog machine with a very simple Dradis, physical gauges, and paper flight plans. It is a very old system. The Dradis sits in the center console with the largest screen real-estate. A smaller needle gauge under the Dradis shows fuel levels, and a standard joystick/foot pedal system provides control over the Viper’s flight systems.
Mark VII Viper Cockpit
The Viper Mk VII is a mostly digital cockpit with a similar Dradis console in the middle (but with a larger screen and more screen-based controls and information). All other displays are digital screens. A few physical buttons are scattered around the top and bottom of the interface. Some controls are pushed down, but none are readable. Groups of buttons are titled with text like “COMMS CIPHER” and “MASTER SYS A”.
Eight buttons around the Dradis console are labeled with complex icons instead of text.
When the Mk VII Vipers encounter Cylons for the first time, the Cylons use a back-door computer virus to completely shut down the Viper’s systems. The screens fuzz out in the same manner as when Apollo gets caught in an EMP burst.
The Viper Mk VII is then completely uncontrollable, and the pilot’s’ joystick-based controls cease to function.
Overall, the Viper Mk II is set up similarly to a WWII P-52 Mustang or early production F-15 Eagle, while the Viper Mk VII is similar to a modern-day F-16 Falcon or F-22 Raptor.
The Viper is a single seat starfighter, and appears to excel in that role. The pilots focus on their ship, and the Raptor pilots following them focus on the big picture. But other items, including color choice, font choice, and location are an issue.
Otherwise, Items appear a little small, and it requires a lot of training to know what to look for on the dashboards. Also, the black lines radiating from the large grouper labels appear to go nowhere and provide no extra context or grouping. Additionally, the controls (outside of the throttle and joystick) require quite a bit of reach from the seat.
Given that the pilots are accelerating at 9+ gs, reaching a critical control in the middle of a fight could be difficult. Hopefully, the designers of the Vipers made sure that ‘fighting’ controls are all within arms reach of the seat, and that the controls requiring more effort are secondary tasks.
Similarly, all-caps text is the hardest to read at a glance, and should be avoided for interfaces like the Viper that require quick targeting and actions in the middle of combat. The other text is very small, and it would be worth doing a deeper evaluation in the cockpit itself to determine if the font size is too small to read from the seat.
If anyone reading this blog has an accurate Viper cockpit prop, we’d be happy to review it!
Fighter pilots in the Battlestar Galactica universe have quick reflexes, excellent vision, and stellar training. They should be allowed to use all of those abilities for besting Cylons in a dogfight, instead of being forced to spend time deciphering their Viper’s interface.
When Coulson hands Tony a case file, it turns out to be an exciting kind of file. For carrying, it’s a large black slab. After Tony grabs it, he grabs the long edges and pulls in opposite directions. One part is a thin translucent screen that fits into an angled slot in the other part, in a laptop-like configuration, right down to a built-in keyboard.
The grip edge
The grip edge of the screen is thicker than the display, so it has a clear, physical affordance as to what part is meant to be gripped and how to pull it free from its casing, and simultaneously what end goes into the base. It’s simple and obvious. The ribbing on the grip unfortunately runs parallel to the direction of pull. It would make for a better grip and a better affordance if the grip was perpendicular to the direction of pull. Minor quibble.
I’d be worried about the ergonomics of an unadjustable display. I’d be worried about the display being easily unseated or dislodged. I’d also be worried about the strength of the join. Since there’s no give, enough force on the display might snap it clean off. But then again this is a world where “vibrium steel” exists, so material critiques may not be diegetically meaningful.
Once he pulls the display from the base, the screen boops and animated amber arcs spin around the screen, signalling him to login via a rectangular panel on the right hand side of the screen. Tony puts his four fingers in the spot and drags down. A small white graphic confirms his biometrics. As a result, a WIMP display appears in grays and amber colors.
One window on the left hand side shows a keypad, and he enters 1-8-5-4. The keypad disappears and a series of thumbnail images—portraits of members of the Avengers initiative—appear in its place. Pepper asks Tony, “What is all this?” Tony replies, saying, “This is, uh…” and in a quick gesture, places his ten fingertips on the screen at the portraits, and then throws his hands outward, off the display.
The portraits slide offscreen to become ceiling-height volumetric windows filled with rich media dossiers on Thor, Steve Rogers, and David Banner. There are videos, portraits, schematics, tables of data, cellular graphics, and maps. There’s a smaller display near the desktop where the “file” rests about the tesseract. (More on this bit in the next post.)
Insert standard complaint here about the eye strain that a translucent display causes, and the apology that yes, I understand it’s an effective and seemingly high-tech way to show actors and screens simultaneously. But I’d be remiss if I didn’t mention it.
The two-part login shows an understanding of multifactor authentication—a first in the survey, so props for that. Tony must provide something he “is”, i.e. his fingerprints, and something he knows, i.e. the passcode. Only then does the top secret information become available.
I have another standard grouse about the screen providing no affordances that content has an alternate view available, and that a secret gesture summons that view. I’d also ordinarily critique the displays for having nearly no visual hierarchy, i.e. no way for your eyes to begin making sense of it, and a lot of pointless-motion noise that pulls your attention in every which way.
But, this beat is about the wonder of the technology, the breadth of information SHIELD in its arsenal, and the surprise of familiar tech becoming epic, so I’m giving it a narrative pass.
Also, OK, Tony’s a universe-class hacker, so maybe he’s just knowledgeable/cocky enough to not need the affordances and turned them off. All that said, in my due diligence: Affordances still matter, people.
The Ford Explorer is an automated vehicle driven on an electrified track through a set route in the park. It has protective covers over its steering wheel, and a set of cameras throughout the car:
Twin cameras at the steering wheel looking out the windshield to give a remote chauffeur or computer system stereoscopic vision
A small camera on the front bumper looking down at the track right in front of the vehicle
Several cameras facing into the cab, giving park operators an opportunity to observe and interact with visitors. (See the subsequent SUV Surveillance post.)
Presumably, there are protective covers over the gas/brake pedal as well, but we never see that area of the interior; evidence comes from when Dr. Grant and Dr. Saddler want to stop and look at the triceratops they don’t even bother to try and reach for the brake pedal, but merely hop out of the SUV.
One notable hybrid interface device, with both physical and digital aspects, is the Drone Programmer. It is used to encode key tasks or functions into the drone. Note that it is seen only briefly—so we’re going off very little information. It facilitates a crucial low-level reprogramming of Drone 172.
This device is a handheld item, grasped on the left, approximately 3 times as wide as it is tall. Several physical buttons are present, but are unused in the film: aside from grasping, all interaction is done through use of a small touchscreen with enough sensitivity to capture fingertip taps on very small elements.
Jack uses the Programmer while the drone is disabled. When he pulls the cord out of the drone, the drone restarts and immediately begins to try and move/understand its surroundings.
When Drone 172 is released from the Programmer cable, it is in a docile and inert state…
As Vika is looking at the radar and verifying visuals on the dispatched drones with Jack, the symbols for drones 166 and 172 begin flashing red. An alert begins sounding, indicating that the two drones are down.
Vika wants to send Jack to drone 166 first. To do this she sends Jack the drone coordinates by pressing and holding the drone symbol for 166 at which time data coordinates are displayed. She then drags the data coordinates with one finger to the Bubbleship symbol and releases. The coordinates immediately display on Jack’s HUD as a target area showing the direction he needs to go. Continue reading →
As Jack begins his preflight check in the Bubbleship, Vika touches the center of the glass surface to power up the desktop that keeps her in contact with Sally on the TET and allows her to assist and monitor Jack as he repairs the drones on the ground.
Using a tool that looks suspiciously identical to the Injection Carbon Reader, the stem line provides electrical current to nerve endings. It is inserted directly into the alien head and then controlled wirelessly via unseen controls on a nearby touch-sensitive slate pad.
The output on the stem line device animates with changes, but it seems the numbers appear near the middle and then slide to fixed positions on the top and bottom. If the point of the display was output, and it is meant to be seen from a distance, wouldn’t a simpler large-number display make more sense? If motion is meant to convey the meaning, like a digital gauge on a multimeter, then the text should be fixed on the graduated background and slide with it.
By stimulating the locus coeruleus of the alien, the Prometheus scientists seek to “trick it into thinking it’s still alive,” even though they have not confirmed the physiology of this 2000 year old alien specimen, or even that (in a quick Googling, the author learned that) this is the panic area of the nervous system. Were the results really that surprising?