Hot Job At Cruise: Director, Program Management Office

Cruise is hiring a Director, Program Management Office. You should apply, or send me your CV and I will refer you!

Great program managers are amazingly effective at reducing stress, increasing performance, and especially at hitting timelines.

For years, I did not believe this, mainly because I hadn’t seen many good program managers in action. Mostly, I had seen engineers, or product managers, or executives corralled into program managements roles, where they performed adequately but not impressively.

I myself have been corralled into that role a few times. I am not a great program manager.

But then I joined Udacity, which had phenomenal program managers, and I realized how effective they could be. Holding everyone accountable, foreseeing the future and addressing upcoming complications, and reporting out progress are all really important for organizational progress.

This specific role at Cruise will, “lead the management, implementation, and reporting of Cruise’s programs.  You will be responsible for creating roadmaps, developing and adhering to timelines, and working cross-functionally to ensure alignment and collaboration.”

Contact me if you’re interested 😊

SimulationCity

Can you tell which half of the image is simulated data and which side is real sensor data?

Waymo announced a new simulation framework recently, both on its own blog and in a feature story with The Verge. The framework is called SimulationCity.

SimulationCity seems awfully reminiscent of CarCraft, the simulation engine that Waymo made famous in 2017. It’s been four years, which is certainly time for a refresh.

The Verge article is a little cagey about the distinction between SimulationCity and CarCraft:

“The company decided it needed a second simulation program after discovering “gaps” in its virtual testing capabilities, said Ben Frankel, senior product manager at the company. Those gaps included using simulation to validate new vehicle platforms, such as the Jaguar I-Pace electric SUV that Waymo has recently begun testing in California, and the company’s semi-trailer trucks outfitted with sensing hardware and the Waymo driver software.”

Waymo is using a new neural network they developed called SurfelGAN (“surface element generative adversarial network”) to better simulate sensor data, especially complex weather conditions like rain, snow, and fog.

Waymo’s blog post features several different videos and GIFs of SimulationCity, and each looks a little different. One video seems focused on behavioral planning, and features an animated Waymo semi-truck on a highway surrounded by moving green rectangular prisms that are meant to represent other vehicles on the road.

Another video seems to be simulating lidar point clouds.

And yet another video shows high-resolution simulated images paired side-by-side with real camera frames. It’s genuinely challenging to figure out which half of the image is simulated and which half is real.

All of that together seems to indicate that SimulationCity is a comprehensive simulation solution, more than a specialized solution for just camera images. I bet they can run perception, localization, prediction, planning, and maybe even control simulations within the framework, at varying speeds. Impressive.

The AV Software Assembly Line

Cruise Origin Self-Driving Ride-Hail and Delivery Vans Are Coming Soon

Protocol asked ten autonomous vehicle executives, “What do people most often get wrong in discussions about autonomous vehicles?” Cruise’s SVP of Engineering, Mo Elshenaway, answered:

“Self-driving is an all-encompassing AI and engineering challenge. It’s easy to see an AV on the streets and think only about the AI models that power them or the compute and sensor suites built as part of it, but there is a virtual software assembly line built alongside the car itself that enables us to meet the unique scale and safety imperative at play here.

To enable AVs to drive superiorly in any given scenario, and continuously evolve and adapt new paradigms, it requires an ecosystem capable of ingesting petabytes of data and hundreds of years worth of compute every day, training and testing models on a continuous loop for multiple times a week software updates that improve performance and ensure safety. The complex network of new tools, testing infrastructure and development platforms that are behind every seamless handling of a construction zone or double-parked car are themselves significant engineering achievements that stand to have an outsized impact beyond AV as they push the boundaries of ML, robotics and more.”

This was probably the biggest surprise upon joining Cruise, which is embarrassing to admit. Cruise has invested tremendously in developing an entire AV software infrastructure that supports the core AV stack. There are front-end engineers working on visualization tools for machine learning scientists, and site reliability engineers ensuring the performance of cloud services. It’s a little bit like an iceberg, 90% of the activity is below the surface of what we might think of as “core AV engineering.”

The rest of the answers in the article are great, too, including Sterling Anderson (Aurora), Jesse Levinson (Zoox), and Raquel Urtasun (Waabi).

Moving From Supervising Robots To Collaborating With Them

Rashed Haq (@rashedhaq) | Twitter

“The crux of the challenge involves making decisions under uncertainty; that is, choosing actions based on often imperfect observations and incomplete knowledge of the world. Autonomous robots have to observe the current state of the world (imperfect observations), understand how this is likely to evolve (incomplete knowledge), and make decisions about the best course of action to pursue in every situation. This cognitive capability is also essential to interpersonal interactions because human communications presuppose an ability to understand the motivations of the participants and subjects of the discussion. As the complexity of human–machine interactions increases and automated systems become more intelligent, we strive to provide computers with comparable communicative and decision-making capabilities. This is what takes robots from machines that humans supervise to machines with which humans can collaborate.

That is from Rashed Haq, VP of Robotics at Cruise, and my VP in particular. He wrote an article for VentureBeat entitled, “The lessons we learn from self-driving will drive our robotics future.”

“Fab-less” Automotive Design

A prototype Adaptive City Mobility City One electric vehicle, with two people standing next to it.

My latest Forbes.com article is about Adaptive City Mobility, a German startup aiming to develop and manufacture an electric fleet vehicle from the ground up. They rely on what founder Paul Leibold calls, “the network economy.”

“They contracted prototyping to Roding, production planning to HÖRMANN Automotive, and series manufacturing to an international Tier 1 automotive supplier. Downstream functions are also handled by partners. A partner manages vehicle leasing, and the digital platform is being developed by Porsche subsidiary MHP.”

Read the whole thing!

Monday Autonomous Vehicle Round-Up

Herd of cattle at a lakeside by Conrad Bühlmayer
  • Argo AI built a self-driving test facility at the Munich airport. These types of test facilities are still necessary for alpha testing, but they’re a lot less important in the US than they were a few years ago.
  • More information on GM’s project to extract lithium from the Salton Sea. The most recent Autonocast episode explores China’s dominance of the battery supply chain.
  • Elon Musk is catching a lot of flack for tweeting, “Didn’t expect [generalized self-driving] to be so hard, but the difficulty is obvious in retrospect.” I think every single automotive company (and me!) that predicted self-driving cars by 2020 could write some version of that tweet. I guess the difference is Tesla actually sold Full Self-Driving packages, so they’re on the hook.
  • Autonomous underwater robots will greatly accelerate the discovery of the more than three million lost shipwrecks around the world, according to the discoverer of The Titanic. “New chapters of human history are to be read.”
  • UL, a major standards publisher, joins the World Economic Forum’s Safe Drive Initiative. Creating the definitive set of self-driving standards remains a big opportunity.
  • Ken Washington leaves the CTO post at Ford for a VP role at Amazon. Ken was VP of advanced engineering and research while I was at Ford. I briefly met him a couple times, including when he visited Udacity. Great hire for Amazon.

Andrej Karpathy’s CVPR Talk, Annotated

Andrej Karpathy, Tesla’s Senior Director of AI, presented Tesla’s recent work at CVPR 2021. CVPR is one of the foremost conferences for academic research into computer vision. Karpathy always does a great job explaining cutting-edge work in an intelligible format (he is an AI researcher with over 350,000 Twitter followers!).

Karpathy’s presentation is about 40 minutes, but it comes at the end of an 8.5 hour session recording. Hence the timestamps start at 7:51:26.

[7:52:46] As a way of emphasizing the importance of automated driving, Karpathy describes human drivers as “meat computers.” I saw some people take offense to this on the Twitter. I think the shortcomings of human drivers are widely acknowledged and this statements wasn’t necessary, but neither was I extremely offended. Human drivers kill a lot of people.

[7:55:14] Karpathy describes Autopilot’s Pedal Misapplication Mitigation (PMM) feature. I’d not heard of this, but I like it. Malcolm Gladwell released a podcast a few years ago hypothesizing that the Toyota recalls of the aughts and early 2010s were largely due to confused drivers flooring the accelerator pedal when they meant to (and thought they were) flooring the brake pedal. Although Consumer Reports disagrees.

[7:57:40] Karpathy notes that Waymo’s approach to self-driving relies on HD maps and lidar, whereas Tesla’s approach relies only on cameras. He claims this makes Tesla’s approach much more scalable, because of the effort required in building and maintaining the HD map. I’m not sure I agree with him about this – a lot of effort goes into automating the mapping process to make it scalable. And even if mapping does prove to be unscalable, lidar has a lot of uses besides localizing to an HD map.

[8:01:20] One reason that Tesla has removed radar from its sensor suite, according to Karpathy, is to liberate engineers to focus on vision. “We prefer to focus all of our infrastructure on this [cameras] and we’re not wasting people working on the radar stack and the sensor fusion stack.” I had not consider the organizational impact of removing the radar sensor.

[8:02:30] Radar signals are really accurate most of the time, but occasionally the radar signal goes haywire, because the radar wave bounces off a bridge or some other irrelevant object. Sorting the signal from the noise is a challenge.

[8:03:25] A good neural network training pipeline has data that is large, clean, and diverse. With that, “Success is guaranteed.”

[8:04:35] Karpathy explains that Tesla generates such a large dataset by using automated techniques that wouldn’t work for a realtime self-driving system. Because the system is labeling data, rather than processing the data in order to drive, the system can run much slower and use extra sensors, in order to get the labeling correct. Humans even help clean the data.

[8:07:10] Karpathy shares a sample of the 221 “triggers” Tesla uses to source interesting data scenarios from the customer fleet. “radar vision mismatch”, “bounding box jitter”, “detection flicker”, “driver enters/exits tunnel”, “objects on the roof (e.g. canoes)”, “brake lights are detected as on but acceleration is positive”, etc.

[8:08:40] Karpathy outlines the process of training a network, deploying it to customers in “shadow mode”, measuring how accurately the model predicts depth, identifying failure cases, and re-training. He says they’ve done 7 rounds of shadow mode. I’m a little surprised the process is that discrete. I would’ve guessed Tesla had a nearly continuous cycle of re-training and re-deploying models.

[8:10:00] Karpathy shows a very high-level schematic of the neural network architecture. There’s a RESnet-style “backbone” that identifies features and then fuses data across all the sensors on the vehicle and then across time. Then the network branches into heads, then “trunks”, then “terminals.” The combined network shares features but also allows engineers interested in specific features (e.g. velocity for vehicles in front of the car) to tune their branches in isolation.

[8:11:30] “You have a team of, I would say, 20 people who are tuning networks full-time, but they’re all cooperating. So, what is the architecture by which you do is an interesting question and I would say continues to be a challenge over time.” In a few different cases now, Karpathy has discussed organizational dynamics within the engineering team and a significant factor in development.

[8:11:50] Karpathy flashes an image and specs of Tesla’s new massive computer. That Karpathy knows enough about computer architecture to even describe what’s going on here is impressive. He also plugs recruiting for their super-computing team.

[8:14:20] In the vein of integration, Karpathy shares that the team gets to design everything from the super-computer, to the in-vehicle FSD chip, to the neural networks. Vertical integration!

[8:16:00] Karpathy shows an example of radar tracking a vehicle and reporting a lot of noise. He explains that maybe they could work on the radar to fix this, but kind of shrugs and says it’s not worth it, since radar isn’t that useful anyway.

[8:19:40] Karpathy references both the validation and simulation processes, but at such a high level I can’t really tell what they’re doing. He mentions unit tests, simulations, track tests, QA drives, and shadow modes.

[8:20:20] Tesla reports FSD has run about 1.7M Autopilot miles with no crashes. Karpathy warns that crashes are inevitable, at Tesla’s scale. He reports that the legacy stack has a crash “every 5M miles or so.” For context, in the US, human drivers experience fatal crashes about every 65M miles. (Do note the distinction between “fatal crashes”, which is the available data for human drivers, and “all crashes” which is the reference Karpathy provides. We would expect “all crashes” to occur much more frequently than “fatal crashes.”)

[8:22:40] Karpathy speculates that training for vision alone basically requires a fleet (and a super-computer), in order to gather sufficient data. He seems like such a nice guy that I wouldn’t even consider this a dig at lidar-reliant autonomous vehicle companies, but rather I chalk this up to a defense to all the criticism that Tesla’s vision-only approach has received.

Well-To-Wheel Emissions

Paul Leinert has a fun story in Reuters about an automotive sustainability model developed by Argonne National Labs. The model is called GREET (“The Greenhouse gases, Regulated Emissions, and Energy use in Technologies Model”) and it seeks to capture the environmental total cost of ownership of a vehicle – including production and operation, even taking into account the fuel sources that generate the electricity that goes into electric vehicles.

The model is…not easy to use. I registered with the Argonne website to download it, only to discover that the main model is a .NET program. There is are some Excel-based versions of the model, which I loaded up in Google Sheets. I couldn’t get it to work – there are 18 tabs in the workbook, with lots of cells to complete. All I want to know is whether I should feel virtuous about or ashamed of my 2004 Toyota Highlander.

The Leinert article in Reuters offers some insight. Leinert calculates that a new Tesla Model 3 is more environmentally damaging to produce than a gas-powered vehicle, but the Model 3 is conversely much more environmentally-friendly to operate. More precisely, the Model 3 only becomes more environmentally-friendly than a comparable gas-powered vehicle after 13,500 miles of operation.

“It was up against a gasoline-fueled Toyota Corolla weighing 2,955 pounds with a fuel efficiency of 33 miles per gallon. It was assumed both vehicles would travel 173,151 miles during their lifetimes.

But if the same Tesla was being driven in Norway, which generates almost all its electricity from renewable hydropower, the break-even point would come after just 8,400 miles.

If the electricity to recharge the EV comes entirely from coal, which generates the majority of the power in countries such as China and Poland, you would have to drive 78,700 miles to reach carbon parity with the Corolla, according to the Reuters analysis of data generated by Argonne’s model.”

I’ve heard different off-the-cuff estimates of these numbers before, and I’m happy to see that Argonne put in the labor to make an accurate estimate.

I do wish their model were easier to use.

But Argonne does have a nice webpage that tells you how environmentally-friendly it is to drive electric vehicles in different states, based on the power sources for electricity generation.

NHTSA Issues Requirements For Reporting Autonomous Crashes

nhtsa-brand.png | NHTSA

Yesterday, seemingly out of the blue, the US National Highway Traffic Safety Administration (NHTSA) issued requirements for reporting collisions involving automated driving systems (ADS) and advanced driver assistance systems (ADAS): “Standing General Order 2021-01: Incident Reporting for Automated Driving Systems (ADS) and Level 2 Advanced Driver Assistance Systems (ADAS)”

The requirements, which go into effect ten days from publication (yesterday), mandate vehicle manufacturers (mostly in the case of ADAS) and operators (mostly in the case of ADS) report crashes within 1 day or as a monthly aggregate, depending on severity.

The reporting uses an existing IT system called the Safety Recall Dashboard to fill out a pretty thorough one-page form.

Completing these forms will be a much more laborious task for a vehicle manufacturers than autonomous vehicle operators, simply due to scale. Even the largest autonomous vehicle companies currently have fleets that top out in the low hundreds. Toyota sold over 2 million cars in the US last year.

My initial scan of the form focused on how easy it would be for a large manufacturer to automate the reporting process. Most of the fields appear automatable, possibly with effort. But some, like “Narrative” and “Pre-Crash Movement” seem a lot harder to automate.

I wonder whether this reporting system will become a backdoor to a national automotive collision database. NHTSA limits the requirements’ scope to Level 2 ADAS for the next 3 years, but of course if a Level 2 system collides with a Level 0 vehicle, then the collision will get reported. Eventually, if enough vehicles have Level 2 ADAS, then most collisions will wind up in this system.

The scope could go far beyond evaluating autonomy. Years ago, I was surprised to learn that the national statistics on automotive collisions are quite thin, because so often drivers don’t report collisions to the police or even their insurers. Now vehicle manufacturers are required to report their customers’ collisions to NHTSA.

That could open a real can of worms in terms of privacy and civil liberties. It could also ensure that everyone is safer on the road, most of all vulnerable road users like pedestrians and cyclists.