Songdo Traffic: High Accuracy Georeferenced Vehicle Trajectories from a Large-Scale Study in a Smart City
Authors/Creators
Description
Overview
The Songdo Traffic dataset delivers precisely georeferenced vehicle trajectories captured through high-altitude bird's-eye view (BeV) drone footage over Songdo International Business District, South Korea. Comprising approximately 700,000 unique trajectories, this resource represents one of the most extensive aerial traffic datasets publicly available, distinguishing itself through exceptional temporal resolution that captures vehicle movements at 29.97 points per second, enabling unprecedented granularity for advanced urban mobility analysis.
📌 Citation: If you use this dataset in your work, kindly acknowledge it by citing the following article:
Robert Fonod, Haechan Cho, Hwasoo Yeo, Nikolas Geroliminis (2025). Advanced computer vision for extracting georeferenced vehicle trajectories from drone imagery, Transportation Research Part C: Emerging Technologies, vol. 178, 105205. DOI: 10.1016/j.trc.2025.105205.
🔗 Companion dataset: For high-resolution annotated images with vehicle bounding boxes supporting aerial detection research, see Songdo Vision: 10.5281/zenodo.13828408.
Dataset Composition
The dataset consists of four primary components:
- Trajectory Data: 80 ZIP archives containing high-resolution vehicle trajectories with georeferenced positions, speeds and acceleration profiles, and other metadata.
- Orthophoto Cut-Outs: High-resolution (8000×8000 pixel) orthophoto images for each monitored intersection, used for georeferencing and visualization.
- Road and Lane Segmentations: CSV files defining lane polygons within road sections, facilitating mapping of vehicle positions to road segments and lanes.
- Sample Videos: A selection of 4K UHD drone video samples capturing intersection footage during the experiment.
Data Collection
The dataset was collected as part of a collaborative multi-drone experiment conducted by KAIST and EPFL in Songdo, South Korea, from October 4–7, 2022.
- A fleet of 10 drones monitored 20 busy intersections, executing advanced flight plans to optimize coverage.
- 4K (3840×2160) RGB video footage was recorded at 29.97 FPS from altitudes of 140–150 meters.
- Each drone flew 10 sessions per day, covering peak morning and afternoon periods.
- The experiment resulted in 12TB of 4K raw video data.
More details on the experimental setup and data processing pipeline are available in [1].
Data Processing
The trajectories were extracted using 🚀 Geo-trax, an advanced deep learning framework designed for high-altitude UAV-based traffic monitoring. This state-of-the-art pipeline integrates vehicle detection, tracking, trajectory stabilization, and georeferencing to extract high-accuracy traffic data from drone footage.
🎥 A demonstration of the Geo-trax framework in operation is available at: https://youtu.be/gOGivL9FFLk
Key Processing Steps:
- Vehicle Detection & Tracking: Vehicles were detected and tracked across frames using a deep learning-based detector and motion-model-based tracking algorithm.
- Trajectory Stabilization: A novel track stabilization method was applied using detected vehicle bounding boxes as exclusion masks in image registration.
- Georeferencing & Coordinate Transformation: Each trajectory was transformed into global (WGS84), local Cartesian, and orthophoto coordinate systems.
- Vehicle Metadata Estimation: In addition to time-stamped vehicle trajectories, various metadata attributes were also extracted, including vehicle dimensions and type, speed, acceleration, class, lane number, road section, and visibility status.
More details on the extraction methodology are available in [1].
File Structure & Formats
1. Trajectory Data (Daily Intersection ZIPs, 16.2 MB ~ 360.2 MB)
The trajectory data is organized into 80 ZIP files, each containing traffic data for a specific intersection and day of the experiment.
File Naming Convention:
YYYY-MM-DD_intersectionID.zip
YYYY-MM-DDrepresents the date of data collection (2022-10-04 to 2022-10-07).intersectionIDis a unique identifier for one of the 20 intersections where data was collected (A, B, C, E, …, U). The letter D is reserved to denote "Drone".
Each ZIP file contains 10 CSV files, each corresponding to an individual flight session:
YYYY-MM-DD_intersectionID.zip
│── YYYY-MM-DD_intersectionID_AM1.csv
├── …
│── YYYY-MM-DD_intersectionID_AM5.csv
│── YYYY-MM-DD_intersectionID_PM1.csv
├── …
└── YYYY-MM-DD_intersectionID_PM5.csv
Here, AM1-AM5 and PM1-PM5 denote morning and afternoon flight sessions, respectively. For example, 2022-10-04_S_AM1.csv contains all extracted trajectories from the first morning session of the first day at the intersection 'S'.
CSV File Example Structure:
Each CSV file contains high-frequency trajectory data, formatted as follows (d.p. = decimal place):
| Dataset Column Name | Format / Units | Data Type | Explanation |
|---|---|---|---|
Vehicle_ID |
1, 2, … | Integer | Unique vehicle identifier within each CSV file |
Local_Time |
hh:mm:ss.sss |
String | Local Korean time (GMT+9) in ISO 8601 format |
Drone_ID |
1, 2, …, 10 | Integer | Unique identifier for the drone capturing the data |
Ortho_X, Ortho_Y |
px (1 d.p.) | Float | Vehicle center coordinates in the orthophoto cut-out image |
Local_X, Local_Y |
m (2 d.p.) | Float | KGD2002 / Central Belt 2010 planar coordinates (EPSG:5186) |
Latitude, Longitude |
° DD (7 d.p.) | Float | WGS84 geographic coordinates in decimal degrees (EPSG:4326) |
Vehicle_Length*, Vehicle_Width* |
m (2 d.p.) | Float | Estimated physical dimensions of the vehicle |
Vehicle_Class |
Categorical (0–3) | Integer | Vehicle type: 0 (car/van), 1 (bus), 2 (truck), 3 (motorcycle) |
Vehicle_Speed* |
km/h (1 d.p.) | Float | Estimated speed computed from trajectory data using Gaussian smoothing |
Vehicle_Acceleration* |
m/s² (2 d.p.) | Float | Estimated acceleration derived from smoothed speed values |
Road_Section* |
N_G |
String | Road section identifier (N = node, G = lane group) |
Lane_Number* |
1, 2, … | Integer | Lane position (1 = leftmost lane in the direction of travel) |
Visibility |
0/1 | Boolean | 1 = fully visible, 0 = partially visible in the camera frame |
* These columns may be empty under certain conditions, see [1] for more details.
2. Orthophoto Cut-Outs & Georeferencing Parameters (orthophotos.zip, 1.8 GB)
For each monitored intersection, we provide high-resolution orthophoto cut-outs derived from a comprehensive, high-accuracy orthophoto mosaic covering the entire experimental area. These 8000 × 8000 pixel PNG images facilitate accurate georeferencing and contextual visualization of vehicle trajectories.
orthophotos/
│── ortho_parameters.txt
│── A.png
│── A_center.txt
│── B.png
│── B_center.txt
│── …
│── U.png
└── U_center.txt
Included Files:
ortho_parameters.txt– Georeferencing parameters of the base (global) orthophoto, defining the mapping between base orthophoto pixel coordinates and real-world coordinates. Contains four values: base longitude (lng₀), base latitude (lat₀), longitude per pixel (Δlng), and latitude per pixel (Δlat), all at the base orthophoto's native resolution.*.png– High-resolution orthophoto cut-out for a given intersection.*_center.txt– Per-intersection offset files defining the center of each cut-out in the base orthophoto's pixel coordinate system.
Reproducing Georeferenced Coordinates from Orthophoto Coordinates
The orthophoto cut-outs provided here (8000 × 8000 px) were downscaled from original 15,000 × 15,000 px cut-outs extracted from the base orthophoto. The Ortho_X and Ortho_Y columns in the trajectory CSVs refer to positions within these 8000 × 8000 px images. However, the center coordinates in *_center.txt and the pixel scale in ortho_parameters.txt are defined at the original 15,000 × 15,000 resolution. To correctly convert orthophoto coordinates to geographic (WGS84) coordinates, the following steps are required:
- Read the base orthophoto parameters from
ortho_parameters.txt: lng₀, lat₀, Δlng, Δlat. - Read the cut-out center from
{ID}_center.txt: center_x, center_y (in base orthophoto pixels). - Compute the cut-out origin (top-left corner) using the original cut-out size (15,000 px):
- origin_x = center_x − 7500
- origin_y = center_y − 7500
- Compute the adjusted geographic origin:
- lng₀' = lng₀ + origin_x × Δlng
- lat₀' = lat₀ + origin_y × Δlat
- Scale the pixel size to account for the downscaling from 15,000 to 8,000 pixels:
- Δlng' = Δlng × (15000 / 8000)
- Δlat' = Δlat × (15000 / 8000)
- Convert orthophoto coordinates to geographic coordinates:
- Longitude = lng₀' + Ortho_X × Δlng'
- Latitude = lat₀' + Ortho_Y × Δlat'
For more details on the orthophoto generation process and the choice of cut-out resolution, refer to [1].
3. Orthophoto Segmentations (segmentations.zip, 24.9 KB)
We provide the road and lane segmentations for each orthophoto cut-out, stored as CSV files where each row defines a lane polygon within a road section.
Each section (N_G) groups lanes moving in the same direction, with lanes numbered sequentially from the innermost outward. The CSV files are structured as follows:
segmentations/
│── A.csv
│── B.csv
│── …
└── U.csv
Each file contains the following columns:
Section: Road section ID (N_Gformat).Lane: Lane number within the section.tlx, tly,blx, bly,brx, bry,trx, try: Polygon corner coordinates.
These segmentations enabled trajectory points to be mapped to specific lanes and sections in our trajectory dataset. Vehicles outside segmented areas (e.g., intersection centers) remain unlabeled. Perspective distortions may also cause misalignments for taller vehicles.
4. Master Frames (master_frames.zip, 248.7 MB)
We include master reference frames and associated transformation parameters used to align stabilized video frames with the orthophoto space. These are helpful for users wishing to reconstruct or audit the georeferencing transformations applied during trajectory extraction.
master_frames/
│── A.png
│── A.txt
│── B.png
│── B.txt
│── …
│── U.png
└── U.txt
Included Files:
*.png– Master reference image for a given intersection.*.txt– Corresponding text file containing the actual transformation parameters (i.g., homography transformation coefficients) that were used to relate the master frame to the orthophoto cut-out coordinate system.
5. Sample Videos (sample_videos.zip, 26.8 GB)
The dataset includes 29 video samples, each capturing the first 60 seconds of drone hovering over its designated intersection during the final session (PM5) on October 7, 2022. These high-resolution 4K videos provide additional context for trajectory analysis and visualization, complementing the orthophoto cut-outs and segmentations.
sample_videos/
│── A_D1_2022-10-07_PM5_60s.mp4
│── A_D2_2022-10-07_PM5_60s.mp4
│── B_D1_2022-10-07_PM5_60s.mp4
│── …
└── U_D10_2022-10-07_PM5_60s.mp4
Additional Files
README.md– Dataset documentation (this file)LICENSE.txt– Creative Commons Attribution 4.0 License
Known Dataset Artifacts and Limitations
While this dataset is designed for high accuracy, users should be aware of the following known artifacts and limitations:
- Trajectory Fragmentation: Trajectories may be fragmented for motorcycles in complex road infrastructure scenarios (pedestrian crossings, bicycle lanes, traffic signals) and for certain underrepresented truck variants. Additional fragmentations occurred when drones experienced technical issues during hovering, necessitating mid-recording splits that naturally resulted in divided trajectories.
- Vehicle ID Ambiguities: The largest
Vehicle_IDin a CSV file does not necessarily indicate the total number of unique vehicles. - Kinematic Estimation Limitations: Speed and acceleration values are derived from raw tracking data and may be affected by minor errors due to detection inaccuracies, stabilization artifacts, and applied interpolation and smoothing techniques.
- Vehicle Dimension Estimation: Estimates may be unreliable for stationary or non-axially moving vehicles and can be affected by bounding box overestimations capturing protruding vehicle parts or shadows.
- Lane and Section Assignment Inaccuracies: Perspective effects may cause vehicles with significant heights, such as trucks or buses, to be misassigned to incorrect lanes or sections in the orthophoto.
- Occasional pedestrian pair misclassifications: Rarely, two pedestrians walking side by side may be briefly mistaken for a motorcycle, but such instances are short-lived and typically removed by the short trajectory filter.
For a comprehensive discussion of dataset limitations and validation procedures, refer to [1].
Acknowledgments
In addition to the funding sources listed in the metadata, the creators express their gratitude to Sohyeong Kim, Yura Tak, and Weijiang Xiong for their invaluable assistance in the data-wrangling process, Artem Vasilev for his dedicated efforts in data annotation, Jasso Espadaler Clapés for his insightful contributions to vehicle dimension estimation, and Muhammad Ahmed for his support in creating the experimental orthophoto for validation purposes and his contributions to road segmentation. We also thank DroMii company for producing the primary orthophoto and Stanford Center at the Incheon Global Campus (SCIGC) for providing the on-board recorded trajectories of their autonomous vehicle. We also thank the research teams of Prof. Simon Oh (Korea University) and Prof. Minju Park (Hannam University) for their assistance during the data collection campaign, including the provision of drone equipment and student support.
Citation & Attribution
Preferred Citation: If you use Songdo Traffic for any purpose, whether academic research, commercial applications, open-source projects, or benchmarking efforts, please cite our accompanying article [1]:
Robert Fonod, Haechan Cho, Hwasoo Yeo, Nikolas Geroliminis (2025). Advanced computer vision for extracting georeferenced vehicle trajectories from drone imagery, Transportation Research Part C: Emerging Technologies, vol. 178, 105205. DOI: 10.1016/j.trc.2025.105205
BibTeX entry:
@article{fonod2025advanced, title = {Advanced computer vision for extracting georeferenced vehicle trajectories from drone imagery}, author = {Fonod, Robert and Cho, Haechan and Yeo, Hwasoo and Geroliminis, Nikolas},
journal = {Transportation Research Part C: Emerging Technologies},
volume = {178},
pages = {105205},
year = {2025},
publisher = {Elsevier},
doi = {10.1016/j.trc.2025.105205},
url = {https://doi.org/10.1016/j.trc.2025.105205} }
Dataset Citation (for archival purposes): Although Zenodo automatically provides a formal citation for this dataset (see below), including citation export in various formats such as BibTeX, we kindly request that you reference the above article as the primary source of this work.
Robert Fonod, Haechan Cho, Hwasoo Yeo, Nikolas Geroliminis (2025). Songdo Traffic: High Accuracy Georeferenced Vehicle Trajectories from a Large-Scale Study in a Smart City (v1). Zenodo. DOI: 10.5281/zenodo.13828384.
Version History
-
v1.0 (Initial release)
Initial public release of the Songdo Traffic dataset, including georeferenced vehicle trajectories, orthophoto cut-outs, road and lane segmentations, and sample videos.
-
v2.0 (Current release)
Added master reference frames used in the georeferencing pipeline and included the previously missing orthophoto georeferencing .txt parameter files, enabling full reproducibility of the coordinate transformations. Updated the dataset description with instructions for reproducing georeferenced coordinates from the provided orthophoto cut-out coordinates, clarifying the relationship between the provided 8000 × 8000 px cut-outs and the original 15,000 × 15,000 px resolution.
Files
2022-10-04_A.zip
Files
(41.0 GB)
| Name | Size | Download all |
|---|---|---|
|
md5:4ed59b8c13c2b0d6cec7f8c175ad8f0e
|
137.1 MB | Preview Download |
|
md5:5f2e7216dadc259cf8b2888313f1a1c8
|
173.6 MB | Preview Download |
|
md5:eadddba9317fb2ad7c6a568ac3b3036f
|
70.1 MB | Preview Download |
|
md5:b73586263f160a7954fc3010ab3867a5
|
84.9 MB | Preview Download |
|
md5:409a98aec19fc30d714e2150795d87c0
|
82.5 MB | Preview Download |
|
md5:f17faf5629a57e230ba896d26ce4c467
|
16.2 MB | Preview Download |
|
md5:5481e1f88ab4c0ade1c48cb77ed5cd8f
|
97.8 MB | Preview Download |
|
md5:d983a71cf9fa7bbe1d7685436f7152e1
|
117.7 MB | Preview Download |
|
md5:513980a2a0f25b1e7c69111fe74ab05d
|
255.2 MB | Preview Download |
|
md5:078115af6192600e98a2437e105f77b8
|
241.1 MB | Preview Download |
|
md5:f5cb14ead890ee7db5903704ed0e96f5
|
286.9 MB | Preview Download |
|
md5:6353032fe63de1e13206645540a70395
|
154.5 MB | Preview Download |
|
md5:571792bfe36a7acb810519f5531e5200
|
151.7 MB | Preview Download |
|
md5:b608ea6e12ea698104c7b468f959625a
|
62.3 MB | Preview Download |
|
md5:d3dbfc91ffde5ae85352532edf7f348f
|
83.9 MB | Preview Download |
|
md5:348115e6aff8f8d6d33f2b5787ec5543
|
95.0 MB | Preview Download |
|
md5:985bdc65078db30f150f207f74eb426f
|
139.6 MB | Preview Download |
|
md5:a98e7a368ea7949e2efc7a5593f736af
|
298.2 MB | Preview Download |
|
md5:48f2267b342d47d76938f5aa7ea78f99
|
139.0 MB | Preview Download |
|
md5:1e3359038edac6c4598cba27162afb0c
|
100.1 MB | Preview Download |
|
md5:faab1e48341f4f863214e7ab8e170a5d
|
157.6 MB | Preview Download |
|
md5:f1fb8fe7660aa002d7acc007ba897504
|
193.9 MB | Preview Download |
|
md5:ad48eb7d682cc220127a9c181afaf1f1
|
80.4 MB | Preview Download |
|
md5:19a8dde36f2fc86b3c5eeb06cb5c8d81
|
110.9 MB | Preview Download |
|
md5:b940bc9a4d7bc8eedd8c14bcd03702bc
|
96.0 MB | Preview Download |
|
md5:08e7e55d20a140a5802c9096742b9c5f
|
22.1 MB | Preview Download |
|
md5:971cdab4656004ba1e9717fd7f44cc2c
|
94.9 MB | Preview Download |
|
md5:34e1ad03fe984257dbdf68ce07e958fc
|
126.7 MB | Preview Download |
|
md5:e7c45a5b9fe465531305bc5b85661275
|
271.1 MB | Preview Download |
|
md5:9659d33df757a71bdb0b0515bbd73a86
|
253.9 MB | Preview Download |
|
md5:c8ef0c57ab63fbfc5e2b4d63112b7376
|
301.8 MB | Preview Download |
|
md5:c941da842661899a933627fa53f950ab
|
162.5 MB | Preview Download |
|
md5:6acfcdeb2147b1e852b41e4a9daf1cb4
|
163.1 MB | Preview Download |
|
md5:3a278dd882a1393e1f69937897682969
|
65.9 MB | Preview Download |
|
md5:a4008d1e24055e724b0f71cc78c32b5a
|
81.2 MB | Preview Download |
|
md5:45f551876d63ef996d8ff7cae02ce456
|
101.8 MB | Preview Download |
|
md5:29b5b4deb975875ffb7bebd672b23fda
|
144.5 MB | Preview Download |
|
md5:8a9b5046d50f72cf60570a54d6c93349
|
360.2 MB | Preview Download |
|
md5:8221e3fe489624f4e6e87762fdf261cd
|
138.9 MB | Preview Download |
|
md5:1e4f58d8c728de2564f717ec87192eb3
|
103.7 MB | Preview Download |
|
md5:2fac0a95f874c43be057c47924baf526
|
148.8 MB | Preview Download |
|
md5:33c4c8dcef44973bb0c26b09e241a634
|
214.7 MB | Preview Download |
|
md5:e4eaea217d8fd101c7c02c027e8bf2c7
|
79.2 MB | Preview Download |
|
md5:b824b6350bfcfec8e23799b5bb186ab4
|
111.0 MB | Preview Download |
|
md5:80cf62f635310696b23f557cb0442545
|
93.8 MB | Preview Download |
|
md5:183f97bd5a9871c240cbbf70b1dcc73e
|
25.1 MB | Preview Download |
|
md5:2ec47bb36fd4eb12b6d2fed1c061ea7e
|
92.3 MB | Preview Download |
|
md5:e8a84dbccd9d83a12ff672d1456e2010
|
124.3 MB | Preview Download |
|
md5:ddbcec7509746145f3a5cb4fa00ecbfe
|
259.9 MB | Preview Download |
|
md5:33e2fc3da9ded524ad93c861d073fbe7
|
251.2 MB | Preview Download |
|
md5:ef0dfa621351841e6c954f3176d33975
|
307.4 MB | Preview Download |
|
md5:d4e3a9742bc6c8db44c588d8a604891c
|
166.0 MB | Preview Download |
|
md5:c4f1fc6c3f3c60c98522fdab986b7570
|
170.9 MB | Preview Download |
|
md5:90122f8ef33807b86b198400c5fbc166
|
66.9 MB | Preview Download |
|
md5:81fc6c254c96569eaa40903f0c0ff009
|
84.2 MB | Preview Download |
|
md5:bea34fe22d9b4ddc251114d77430974b
|
98.8 MB | Preview Download |
|
md5:8330701fbe4d0dc8f50599ea38b1e516
|
134.7 MB | Preview Download |
|
md5:1d78e426ef43768091102f4e16e74f1d
|
312.6 MB | Preview Download |
|
md5:a0545665bdab4f05baa7ffb61021c024
|
251.0 MB | Preview Download |
|
md5:024417b2119b2f6726483c6285aed5aa
|
82.2 MB | Preview Download |
|
md5:cd5d1e59ee25a8e24fa700e4aa607d8f
|
163.6 MB | Preview Download |
|
md5:084abf4275591dd430ed546de4467d39
|
221.0 MB | Preview Download |
|
md5:dfd8e422127c55ccf870218df04825c9
|
75.5 MB | Preview Download |
|
md5:6ad00a6e33610b9bfb0127dcb4d42022
|
110.8 MB | Preview Download |
|
md5:2d061e940effab1c78e0c5f2f1e9f813
|
110.3 MB | Preview Download |
|
md5:1ebdc260b68426f575e28380ab8230db
|
30.7 MB | Preview Download |
|
md5:ad88316f38c7779efb6289dd69ab6f59
|
96.1 MB | Preview Download |
|
md5:39953b7df017ebe5874806c706be8677
|
130.3 MB | Preview Download |
|
md5:6880fa140f801d556627d00401562695
|
291.8 MB | Preview Download |
|
md5:6fd296896a7f7f90cad26fb889c3b749
|
289.3 MB | Preview Download |
|
md5:c549f0f25b5288c80dfb659ef0816ecb
|
321.3 MB | Preview Download |
|
md5:79093027c266e967a0e857f78aec48c9
|
165.5 MB | Preview Download |
|
md5:a105c78239647ab912bf759ddb1ae25d
|
191.0 MB | Preview Download |
|
md5:8934c3d6f5362ddcb9be0c08a08a35a4
|
81.7 MB | Preview Download |
|
md5:dee6e859bbcd80238bf3d09573824407
|
87.0 MB | Preview Download |
|
md5:30a6d175e91b560d47a15b22f094f8b0
|
103.7 MB | Preview Download |
|
md5:e92f12650de857d638f7f9db988eb98f
|
150.2 MB | Preview Download |
|
md5:ca36707ede0baef6f0494306ecf278ee
|
329.6 MB | Preview Download |
|
md5:9d87c25836891e0bdc5beeaeb5f0ea34
|
277.8 MB | Preview Download |
|
md5:292f9c30dbb0ccda67d45fd67f34eb14
|
85.1 MB | Preview Download |
|
md5:527dc6cad772ccb187d5bfe5af738204
|
18.7 kB | Preview Download |
|
md5:4af9a065be7bcc4ac09ac45a9ce1af71
|
248.6 MB | Preview Download |
|
md5:31a2712a456a22b3db8d5513e75da8b2
|
1.8 GB | Preview Download |
|
md5:06104d1fb4bc3e799dba1cc9b46491f6
|
15.8 kB | Preview Download |
|
md5:cbe3b6f6c1dc5e9dce3406b4c1162b6f
|
26.8 GB | Preview Download |
|
md5:c39304c88ddf39f3e0a40b8621d384b9
|
24.9 kB | Preview Download |
Additional details
Funding
- Board of the Swiss Federal Institutes of Technology
- Open Research Data (ORD) Program of the ETH Board
- Swiss National Science Foundation
- NCCR Automation (phase I) 180545
- Innosuisse – Swiss Innovation Agency
- CityDronics 101.645 IP-ENG
- National Research Foundation of Korea
- Grant funded by the Korean government (MSIT) 2022R1A2C1012380
Dates
- Collected
-
2022-10-04/2022-10-07