Visual Perception for Self-Driving Cars

  • 4.7
Approx. 31 hours to complete

Course Summary

This course teaches visual perception techniques used in self-driving cars. You'll learn how to analyze camera images and lidar data to make driving decisions.

Key Learning Points

  • Understand the fundamentals of visual perception and how it's used in self-driving cars
  • Learn how to analyze camera images and lidar data
  • Gain practical experience working with real-world data

Related Topics for further study


Learning Outcomes

  • Analyze camera images and lidar data for driving decisions
  • Understand the fundamentals of visual perception in self-driving cars
  • Gain practical experience working with real-world data

Prerequisites or good to have knowledge before taking this course

  • Basic knowledge of Python
  • Familiarity with machine learning concepts

Course Difficulty Level

Intermediate

Course Format

  • Online
  • Self-paced

Similar Courses

  • Self-Driving Cars: Advanced Topics
  • Computer Vision Basics

Related Education Paths


Notable People in This Field

  • Elon Musk
  • Andrew Ng

Related Books

Description

Welcome to Visual Perception for Self-Driving Cars, the third course in University of Toronto’s Self-Driving Cars Specialization.

Knowledge

  • Work with the pinhole camera model, and perform intrinsic and extrinsic camera calibration
  • Detect, describe and match image features and design your own convolutional neural networks
  • Apply these methods to visual odometry, object detection and tracking
  • Apply semantic segmentation for drivable surface estimation

Outline

  • Welcome to Course 3: Visual Perception for Self-Driving Cars
  • Welcome to the Self-Driving Cars Specialization!
  • Welcome to the course
  • Meet the Instructor, Steven Waslander
  • Meet the Instructor, Jonathan Kelly
  • Course Prerequisites
  • How to Use Discussion Forums
  • How to Use Supplementary Readings in This Course
  • Recommended Textbooks
  • Module 1: Basics of 3D Computer Vision
  • Lesson 1 Part 1: The Camera Sensor
  • Lesson 1 Part 2: Camera Projective Geometry
  • Lesson 2: Camera Calibration
  • Lesson 3 Part 1: Visual Depth Perception - Stereopsis
  • Lesson 3 Part 2: Visual Depth Perception - Computing the Disparity
  • Lesson 4: Image Filtering
  • Supplementary Reading: The Camera Sensor
  • Supplementary Reading: Camera Calibration
  • Supplementary Reading: Visual Depth Perception
  • Supplementary Reading: Image Filtering
  • Module 1 Graded Quiz
  • Module 2: Visual Features - Detection, Description and Matching
  • Lesson 1: Introduction to Image features and Feature Detectors
  • Lesson 2: Feature Descriptors
  • Lesson 3 Part 1: Feature Matching
  • Lesson 3 Part 2: Feature Matching: Handling Ambiguity in Matching
  • Lesson 4: Outlier Rejection
  • Lesson 5: Visual Odometry
  • Supplementary Reading: Feature Detectors and Descriptors
  • Supplementary Reading: Feature Matching
  • Supplementary Reading: Feature Matching
  • Supplementary Reading: Outlier Rejection
  • Supplementary Reading: Visual Odometry
  • Module 3: Feedforward Neural Networks
  • Lesson 1: Feed Forward Neural Networks
  • Lesson 2: Output Layers and Loss Functions
  • Lesson 3: Neural Network Training with Gradient Descent
  • Lesson 4: Data Splits and Neural Network Performance Evaluation
  • Lesson 5: Neural Network Regularization
  • Lesson 6: Convolutional Neural Networks
  • Supplementary Reading: Feed-Forward Neural Networks
  • Supplementary Reading: Output Layers and Loss Functions
  • Supplementary Reading: Neural Network Training with Gradient Descent
  • Supplementary Reading: Data Splits and Neural Network Performance Evaluation
  • Supplementary Reading: Neural Network Regularization
  • Supplementary Reading: Convolutional Neural Networks
  • Feed-Forward Neural Networks
  • Module 4: 2D Object Detection
  • Lesson 1: The Object Detection Problem
  • Lesson 2: 2D Object detection with Convolutional Neural Networks
  • Lesson 3: Training vs. Inference
  • Lesson 4: Using 2D Object Detectors for Self-Driving Cars
  • Supplementary Reading: The Object Detection Problem
  • Supplementary Reading: 2D Object detection with Convolutional Neural Networks
  • Supplementary Reading: Training vs. Inference
  • Supplementary Reading: Using 2D Object Detectors for Self-Driving Cars
  • Object Detection For Self-Driving Cars
  • Module 5: Semantic Segmentation
  • Lesson 1: The Semantic Segmentation Problem
  • Lesson 2: ConvNets for Semantic Segmentation
  • Lesson 3: Semantic Segmentation for Road Scene Understanding
  • Supplementary Reading: The Semantic Segmentation Problem
  • Supplementary Reading: ConvNets for Semantic Segmentation
  • Supplementary Reading: Semantic Segmentation for Road Scene Understanding
  • Semantic Segmentation For Self-Driving Cars
  • Module 6: Putting it together - Perception of dynamic objects in the drivable region
  • Project Overview: Using CARLA for object detection and segmentation
  • Final Project Hints
  • Final Project Solution [LOCKED]
  • Congratulations for completing the course!

Summary of User Reviews

Check out what users are saying about the Visual Perception for Self-Driving Cars course on Coursera. Many users found the course to be comprehensive and informative.

Key Aspect Users Liked About This Course

Comprehensive and informative

Pros from User Reviews

  • Course content is well-structured and easy to follow
  • Instructors are knowledgeable and provide clear explanations
  • Hands-on assignments and quizzes help to reinforce understanding

Cons from User Reviews

  • Some users found the course too technical and difficult to understand
  • Limited interaction with instructors and other students
  • Course may not be suitable for beginners with no prior knowledge in computer vision or machine learning
English
Available now
Approx. 31 hours to complete
Steven Waslander
University of Toronto
Coursera

Instructor

Steven Waslander

  • 4.7 Raiting
Share
Saved Course list
Cancel
Get Course Update
Computer Courses