<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Autonomous Driving | David Vázquez</title><link>https://david-vazquez.com/tags/autonomous-driving/</link><atom:link href="https://david-vazquez.com/tags/autonomous-driving/index.xml" rel="self" type="application/rss+xml"/><description>Autonomous Driving</description><generator>Hugo Blox Builder (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Wed, 01 Jun 2016 00:00:00 +0000</lastBuildDate><image><url>https://david-vazquez.com/media/icon_hu_a3642885bc94ba2d.png</url><title>Autonomous Driving</title><link>https://david-vazquez.com/tags/autonomous-driving/</link></image><item><title>SYNTHIA</title><link>https://david-vazquez.com/project/synthia/</link><pubDate>Wed, 01 Jun 2016 00:00:00 +0000</pubDate><guid>https://david-vazquez.com/project/synthia/</guid><description>&lt;p&gt;SYNTHIA is a large collection of synthetic images for semantic segmentation of urban scenes, generated using a video game engine. Published at CVPR 2016 and widely adopted in the autonomous driving research community. Licensed for commercial use by Intel, Audi, Huawei, Toyota, and Samsung.&lt;/p&gt;</description></item><item><title>Elektra Autonomous Vehicle</title><link>https://david-vazquez.com/project/elektra/</link><pubDate>Wed, 01 Jan 2014 00:00:00 +0000</pubDate><guid>https://david-vazquez.com/project/elektra/</guid><description>&lt;style&gt;
.section-row {
display: flex;
flex-wrap: wrap;
gap: 2.5rem;
margin: 2.5rem 0;
align-items: flex-start;
}
.section-row.reverse { flex-direction: row-reverse; }
.section-text { flex: 1; min-width: 300px; }
.section-media { flex: 0 0 calc(45% - 1.25rem); }
@media (max-width: 1024px) {
.section-row, .section-row.reverse { flex-direction: column; }
.section-media { flex: 1 1 100%; }
}
.section-slideshow {
position: relative;
border-radius: 8px;
overflow: hidden;
background: #f5f5f5;
}
.section-slideshow-container {
position: relative;
width: 100%;
padding-bottom: 75%;
height: 0;
}
.section-slideshow-image {
position: absolute;
inset: 0;
opacity: 0;
transition: opacity 0.6s ease-in-out;
}
.section-slideshow-image.active { opacity: 1; }
.section-slideshow-image img {
width: 100%;
height: 100%;
object-fit: cover;
display: block;
}
.slideshow-nav {
position: absolute;
bottom: 1rem;
left: 50%;
transform: translateX(-50%);
display: flex;
gap: 0.5rem;
z-index: 10;
}
.slideshow-dot {
width: 10px;
height: 10px;
border-radius: 50%;
background: rgba(255,255,255,0.5);
cursor: pointer;
border: none;
transition: background 0.3s;
}
.slideshow-dot.active { background: white; }
.slideshow-caption {
position: absolute;
bottom: 2.5rem;
left: 0; right: 0;
padding: 0.5rem 1rem;
background: rgba(0,0,0,0.45);
color: white;
font-size: 0.85rem;
text-align: center;
z-index: 5;
}
.elektra-stats {
display: grid;
grid-template-columns: repeat(4, 1fr);
gap: 1.5rem;
margin: 2rem 0;
text-align: center;
}
.elektra-stat {
padding: 1.25rem;
background: #f5f5f5;
border-radius: 8px;
}
.dark .elektra-stat { background: #2a2a2a; }
.elektra-stat-value { font-size: 1.6rem; font-weight: 700; color: #333; }
.dark .elektra-stat-value { color: #eee; }
.elektra-stat-label { font-size: 0.85rem; color: #666; margin-top: 0.4rem; }
.dark .elektra-stat-label { color: #bbb; }
@media (max-width: 640px) { .elektra-stats { grid-template-columns: repeat(2, 1fr); } }
.featured-video-btn {
display: inline-flex;
align-items: center;
gap: 0.5rem;
padding: 0.875rem 1.75rem;
background: rgb(var(--color-primary-600));
color: white;
border-radius: 8px;
cursor: pointer;
font-size: 1rem;
font-weight: 600;
border: none;
margin: 1.25rem 0;
transition: background 0.2s;
}
.featured-video-btn:hover { background: rgb(var(--color-primary-700)); }
.video-links {
display: flex;
flex-wrap: wrap;
gap: 0.75rem;
margin-top: 1.5rem;
}
.video-link-btn {
display: inline-flex;
align-items: center;
gap: 0.4rem;
padding: 0.4rem 0.9rem;
border: 1px solid rgba(0,0,0,0.15);
border-radius: 6px;
font-size: 0.85rem;
font-weight: 500;
cursor: pointer;
background: none;
color: inherit;
font-family: inherit;
transition: border-color 0.2s, background 0.2s;
}
.video-link-btn:hover { border-color: rgb(var(--color-primary-500)); background: rgba(var(--color-primary-50), 0.5); }
.dark .video-link-btn { border-color: rgba(255,255,255,0.15); }
.dark .video-link-btn:hover { border-color: rgb(var(--color-primary-400)); background: rgba(255,255,255,0.05); }
.video-modal {
display: none;
position: fixed;
inset: 0;
background: rgba(0,0,0,0.8);
z-index: 1000;
align-items: center;
justify-content: center;
padding: 2rem;
}
.video-modal.active { display: flex; }
.video-modal-content {
background: white;
border-radius: 12px;
max-width: 900px;
width: 100%;
overflow: hidden;
box-shadow: 0 10px 40px rgba(0,0,0,0.3);
}
.dark .video-modal-content { background: #1a1a1a; }
.video-modal-header {
display: flex;
justify-content: space-between;
align-items: center;
padding: 1rem;
border-bottom: 1px solid #e5e5e5;
}
.dark .video-modal-header { border-bottom-color: #333; }
.video-modal-title { font-size: 1rem; font-weight: 600; margin: 0; }
.video-modal-close {
background: none;
border: none;
font-size: 1.4rem;
cursor: pointer;
color: #666;
line-height: 1;
padding: 0.25rem 0.5rem;
}
.dark .video-modal-close { color: #aaa; }
.video-modal-player {
position: relative;
padding-bottom: 56.25%;
height: 0;
}
.video-modal-player iframe {
position: absolute;
inset: 0;
width: 100%;
height: 100%;
}
&lt;/style&gt;
&lt;div id="video-modal" class="video-modal"&gt;
&lt;div class="video-modal-content"&gt;
&lt;div class="video-modal-header"&gt;
&lt;h3 class="video-modal-title" id="video-modal-title"&gt;Video&lt;/h3&gt;
&lt;button class="video-modal-close" onclick="closeVideoModal()"&gt;✕&lt;/button&gt;
&lt;/div&gt;
&lt;div class="video-modal-player"&gt;
&lt;iframe id="video-modal-player" src="" frameborder="0" allowfullscreen allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture"&gt;&lt;/iframe&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;script&gt;
let slideIndex = {};
let autoPlayTimer = {};
const slideCaptions = {
project: ["Elektra autonomous vehicle platform", "Multidisciplinary team composition"],
perception: ["Real-time stereo vision processing", "3D scene reconstruction", "Free-space detection", "Pedestrian detection"],
synthia: ["SYNTHIA daytime urban scenario", "SYNTHIA nighttime driving"]
};
function initSlideshow(id, n) {
slideIndex[id] = 1;
showSlides(1, id);
autoPlay(id);
}
function currentSlide(n, id) {
clearTimeout(autoPlayTimer[id]);
showSlides(slideIndex[id] = n, id);
autoPlay(id);
}
function showSlides(n, id) {
const el = document.getElementById(id + '-slideshow');
if (!el) return;
const slides = el.querySelectorAll('.section-slideshow-image');
const dots = el.querySelectorAll('.slideshow-dot');
const cap = document.getElementById(id + '-caption');
if (n &gt; slides.length) slideIndex[id] = 1;
if (n &lt; 1) slideIndex[id] = slides.length;
slides.forEach(s =&gt; s.classList.remove('active'));
dots.forEach(d =&gt; d.classList.remove('active'));
if (slides.length) {
slides[slideIndex[id] - 1].classList.add('active');
if (dots.length) dots[slideIndex[id] - 1].classList.add('active');
if (cap &amp;&amp; slideCaptions[id]) cap.textContent = slideCaptions[id][slideIndex[id] - 1];
}
}
function autoPlay(id) {
autoPlayTimer[id] = setTimeout(() =&gt; {
slideIndex[id]++;
showSlides(slideIndex[id], id);
autoPlay(id);
}, 5000);
}
function openVideoModal(videoId, title) {
document.getElementById('video-modal-title').textContent = title;
document.getElementById('video-modal-player').src = `https://www.youtube.com/embed/${videoId}`;
document.getElementById('video-modal').classList.add('active');
document.body.style.overflow = 'hidden';
}
function closeVideoModal() {
document.getElementById('video-modal').classList.remove('active');
document.getElementById('video-modal-player').src = '';
document.body.style.overflow = '';
}
document.addEventListener('click', e =&gt; {
if (e.target === document.getElementById('video-modal')) closeVideoModal();
});
document.addEventListener('keydown', e =&gt; { if (e.key === 'Escape') closeVideoModal(); });
&lt;/script&gt;
&lt;div class="elektra-stats"&gt;
&lt;div class="elektra-stat"&gt;&lt;div class="elektra-stat-value"&gt;20+&lt;/div&gt;&lt;div class="elektra-stat-label"&gt;Top-tier Publications&lt;/div&gt;&lt;/div&gt;
&lt;div class="elektra-stat"&gt;&lt;div class="elektra-stat-value"&gt;8&lt;/div&gt;&lt;div class="elektra-stat-label"&gt;Partner Institutions&lt;/div&gt;&lt;/div&gt;
&lt;div class="elektra-stat"&gt;&lt;div class="elektra-stat-value"&gt;400 FPS&lt;/div&gt;&lt;div class="elektra-stat-label"&gt;Real-time Stixel&lt;/div&gt;&lt;/div&gt;
&lt;div class="elektra-stat"&gt;&lt;div class="elektra-stat-value"&gt;2010s&lt;/div&gt;&lt;div class="elektra-stat-label"&gt;Active Period&lt;/div&gt;&lt;/div&gt;
&lt;/div&gt;
&lt;hr&gt;
&lt;h2 id="autonomous-driving-in-action"&gt;Autonomous Driving in Action&lt;/h2&gt;
&lt;p&gt;Watch the Elektra platform navigate urban roads autonomously — perception, planning, and control integrated end-to-end:&lt;/p&gt;
&lt;p&gt;&lt;button class="featured-video-btn" onclick="openVideoModal('tvZnN65jbCE', 'On-Road Autonomous Driving Demo')"&gt;▶ Watch Autonomous Driving Demo&lt;/button&gt;&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id="project-overview"&gt;Project Overview&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;Elektra&lt;/strong&gt; is an autonomous driving platform and the &lt;strong&gt;Catalan hub of autonomous driving&lt;/strong&gt;, bringing together more than &lt;strong&gt;20 professionals&lt;/strong&gt; from academia and industry. The platform integrates perception, planning, control, and communications to demonstrate production-ready autonomous driving in urban environments.&lt;/p&gt;
&lt;div class="section-row"&gt;
&lt;div class="section-text"&gt;
&lt;p&gt;&lt;strong&gt;Partner institutions:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;CVC-UAB&lt;/strong&gt; — Environment perception &amp;amp; computer vision&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;CAOS-UAB&lt;/strong&gt; — Embedded hardware &amp;amp; GPU optimization&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;UPC-Tarrasa&lt;/strong&gt; — Control &amp;amp; path planning&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;CTTC-UPC&lt;/strong&gt; — Positioning &amp;amp; localization&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;UAB-DEIC&lt;/strong&gt; — Vehicle-to-vehicle communications&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;UAB-CEPHIS&lt;/strong&gt; — Electronics &amp;amp; integration&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;CT Ingenieros&lt;/strong&gt; — Vehicle engineering &amp;amp; drive-by-wire&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Municipality of Sant Quirze&lt;/strong&gt; — Test track facility&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The &lt;strong&gt;Computer Vision Center (CVC)&lt;/strong&gt; led the perception stack — my primary contribution to the project. Validation was performed at the Sant Quirze test track and in urban environments, demonstrating the system across controlled and real-world scenarios.&lt;/p&gt;
&lt;/div&gt;
&lt;div class="section-media"&gt;
&lt;div class="section-slideshow" id="project-slideshow"&gt;
&lt;div class="section-slideshow-container"&gt;
&lt;div class="section-slideshow-image active"&gt;
&lt;img src="elektra-car.png" alt="Elektra autonomous vehicle platform"&gt;
&lt;/div&gt;
&lt;div class="section-slideshow-image"&gt;
&lt;img src="overview.png" alt="Project team and institution overview"&gt;
&lt;/div&gt;
&lt;div class="slideshow-nav"&gt;
&lt;button class="slideshow-dot active" onclick="currentSlide(1, 'project')"&gt;&lt;/button&gt;
&lt;button class="slideshow-dot" onclick="currentSlide(2, 'project')"&gt;&lt;/button&gt;
&lt;/div&gt;
&lt;div class="slideshow-caption" id="project-caption"&gt;Elektra autonomous vehicle platform&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;script&gt;initSlideshow('project', 2);&lt;/script&gt;
&lt;hr&gt;
&lt;h2 id="perception-system"&gt;Perception System&lt;/h2&gt;
&lt;p&gt;I &lt;strong&gt;initiated and led the full perception pipeline&lt;/strong&gt; — from raw sensor data to high-level scene understanding. The system fuses multiple modalities for robust environmental awareness:&lt;/p&gt;
&lt;div class="section-row reverse"&gt;
&lt;div class="section-text"&gt;
&lt;p&gt;&lt;strong&gt;Obstacle &amp;amp; Pedestrian Detection&lt;/strong&gt;
Real-time CNN-based detection running at 400+ FPS on GPU hardware, with multi-scale detection for obstacles at various distances and temporal consistency across frames.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Free Space &amp;amp; Lane Detection&lt;/strong&gt;
Stixel-based 3D scene representation identifies drivable areas and lane boundaries from dense stereo depth. Adaptive thresholding handles varying road conditions in real time.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;3D Reconstruction &amp;amp; SLAM&lt;/strong&gt;
Stereo cameras provide dense depth estimation. Visual odometry and loop closure detection enable robust 6-DOF localization even in GPS-denied environments (tunnels, urban canyons).&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Sensor Fusion&lt;/strong&gt;
Stereo cameras, monocular vision, LIDAR, and IMU are combined for redundant, accurate scene understanding optimized for embedded automotive hardware.&lt;/p&gt;
&lt;/div&gt;
&lt;div class="section-media"&gt;
&lt;div class="section-slideshow" id="perception-slideshow"&gt;
&lt;div class="section-slideshow-container"&gt;
&lt;div class="section-slideshow-image active"&gt;
&lt;img src="image1.png" alt="Real-time stereo vision processing"&gt;
&lt;/div&gt;
&lt;div class="section-slideshow-image"&gt;
&lt;img src="image102.png" alt="3D scene reconstruction"&gt;
&lt;/div&gt;
&lt;div class="section-slideshow-image"&gt;
&lt;img src="image97.png" alt="Free-space detection"&gt;
&lt;/div&gt;
&lt;div class="section-slideshow-image"&gt;
&lt;img src="image104.png" alt="Pedestrian detection"&gt;
&lt;/div&gt;
&lt;div class="slideshow-nav"&gt;
&lt;button class="slideshow-dot active" onclick="currentSlide(1, 'perception')"&gt;&lt;/button&gt;
&lt;button class="slideshow-dot" onclick="currentSlide(2, 'perception')"&gt;&lt;/button&gt;
&lt;button class="slideshow-dot" onclick="currentSlide(3, 'perception')"&gt;&lt;/button&gt;
&lt;button class="slideshow-dot" onclick="currentSlide(4, 'perception')"&gt;&lt;/button&gt;
&lt;/div&gt;
&lt;div class="slideshow-caption" id="perception-caption"&gt;Real-time stereo vision processing&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;script&gt;initSlideshow('perception', 4);&lt;/script&gt;
&lt;hr&gt;
&lt;h2 id="synthia-synthetic-data-for-autonomous-driving"&gt;SYNTHIA: Synthetic Data for Autonomous Driving&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;SYNTHIA&lt;/strong&gt; is a synthetic data generation framework I developed within the Elektra project that creates photorealistic, automatically labeled driving scenarios — addressing the fundamental bottleneck of acquiring large-scale annotated driving data.&lt;/p&gt;
&lt;div class="section-row"&gt;
&lt;div class="section-text"&gt;
&lt;p&gt;&lt;strong&gt;Capabilities:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Multiple environmental conditions: day, night, rain, fog, snow&lt;/li&gt;
&lt;li&gt;Diverse urban scenes: intersections, pedestrian crossings, parked vehicles&lt;/li&gt;
&lt;li&gt;Automatic ground-truth labels for semantic segmentation, depth, and optical flow&lt;/li&gt;
&lt;li&gt;Scalable: thousands of labeled frames in hours&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;Impact:&lt;/strong&gt;
SYNTHIA powered the Elektra perception pipeline, reducing the need for expensive field data collection and enabling systematic testing across conditions that are rare or dangerous to capture in the real world. Results were published at CVPR, ICCV, and ECCV. The dataset was licensed to Intel, Audi, and Huawei.&lt;/p&gt;
&lt;/div&gt;
&lt;div class="section-media"&gt;
&lt;div class="section-slideshow" id="synthia-slideshow"&gt;
&lt;div class="section-slideshow-container"&gt;
&lt;div class="section-slideshow-image active"&gt;
&lt;img src="synthia-360.png" alt="SYNTHIA daytime urban scenario"&gt;
&lt;/div&gt;
&lt;div class="section-slideshow-image"&gt;
&lt;img src="synthia-overview.png" alt="SYNTHIA multi-condition overview"&gt;
&lt;/div&gt;
&lt;div class="slideshow-nav"&gt;
&lt;button class="slideshow-dot active" onclick="currentSlide(1, 'synthia')"&gt;&lt;/button&gt;
&lt;button class="slideshow-dot" onclick="currentSlide(2, 'synthia')"&gt;&lt;/button&gt;
&lt;/div&gt;
&lt;div class="slideshow-caption" id="synthia-caption"&gt;SYNTHIA daytime urban scenario&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;script&gt;initSlideshow('synthia', 2);&lt;/script&gt;
&lt;hr&gt;
&lt;h2 id="publications--impact"&gt;Publications &amp;amp; Impact&lt;/h2&gt;
&lt;p&gt;The Elektra project generated &lt;strong&gt;20+ peer-reviewed publications&lt;/strong&gt; at top venues including CVPR, ICCV, ECCV, IEEE TITS, and IEEE T-IV. Key contributions include:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Stixel-based 3D scene understanding&lt;/strong&gt; — efficient real-time scene representation&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;SYNTHIA dataset&lt;/strong&gt; — synthetic data for autonomous driving, widely used in the community&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Semantic segmentation&lt;/strong&gt; pipelines for urban scene understanding&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Domain adaptation&lt;/strong&gt; methods bridging synthetic and real data&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;&lt;strong&gt;Legacy:&lt;/strong&gt; Elektra proved vision-centric autonomous driving is achievable in real urban conditions and produced benchmark datasets still used by the research community. Alumni of the team now work at leading autonomous driving companies worldwide.&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id="selected-videos"&gt;Selected Videos&lt;/h2&gt;
&lt;div class="video-links"&gt;
&lt;button class="video-link-btn" onclick="openVideoModal('tvZnN65jbCE', 'Autonomous Driving Demo')"&gt;▶ Autonomous Driving Demo&lt;/button&gt;
&lt;button class="video-link-btn" onclick="openVideoModal('FWM-5Ps8zFo', 'Elektra Project Overview')"&gt;▶ Project Overview&lt;/button&gt;
&lt;button class="video-link-btn" onclick="openVideoModal('7u-mMtm1Q9o', 'Person Detection')"&gt;▶ Person Detection&lt;/button&gt;
&lt;/div&gt;</description></item></channel></rss>