nature.com

Vme: A Satellite Imagery Dataset and Benchmark for Detecting Vehicles in the Middle East and Beyond

AbstractDetecting vehicles in satellite images is crucial for traffic management, urban planning, and disaster response. However, current models struggle with real-world diversity, particularly across different regions. This challenge is amplified by geographic bias in existing datasets, which often focus on specific areas and overlook regions like the Middle East. To address this gap, we present the Vehicles in the Middle East (VME) dataset, designed explicitly for vehicle detection in high-resolution satellite images from Middle Eastern countries. Sourced from Maxar, the VME dataset spans 54 cities across 12 countries, comprising over 4,000 image tiles and more than 100,000 vehicles, annotated using both manual and semi-automated methods. Additionally, we introduce the largest benchmark dataset for Car Detection in Satellite Imagery (CDSI), combining images from multiple sources to enhance global car detection. Our experiments demonstrate that models trained on existing datasets perform poorly on Middle Eastern images, while the VME dataset significantly improves detection accuracy in this region. Moreover, state-of-the-art models trained on CDSI achieve substantial improvements in global car detection.

Background & SummarySatellite imagery has become an essential instrument for a wide range of applications from agriculture1 and environmental monitoring2 to urban development3,4 and disaster response5. A recent review of object detection in satellite imagery highlights the difficulty of creating a general-purpose model that can handle thousands of diverse object categories and varying real-world conditions6. Instead, the study recommends focusing on task-specific models in narrower application areas, where success is more likely if large, well-annotated datasets are available. Therefore, our study focuses on vehicle detection in satellite imagery, a critical task with diverse real-world applications such as analyzing traffic flow and patterns for traffic management7,8, monitoring parking lot occupancy rates to support urban planning9, and modeling spatial-temporal changes in vehicle counts as a proxy for internal displacement monitoring10. To this end, we first present a novel labeled dataset called Vehicles in the Middle East (VME) to attenuate the under-representation of the region. We then construct the largest benchmark dataset, called Car Detection in Satellite Imagery (CDSI), by consolidating images from multiple existing satellite imagery datasets for enhanced global car detection.Detecting vehicles in satellite imagery is challenging because each vehicle covers only a few pixels, classifying them as tiny objects. As a result, the surrounding context becomes crucial for accurately delineating these small objects. Several studies have been conducted on tiny object detection in satellite imagery11,12,13. A review comprehensively analyzed these methods based on five factors: data augmentation, multi-scale feature learning, context-based detection, training strategy, and GAN-based detection, and showed that these factors play a role in enhancing the detection performance in tiny objects14. Another systematic study on small object detection was conducted by reviewing existing literature on algorithms and datasets15. Two large-scale benchmarks, SODA-D and SODA-A, were constructed for driving scenarios and aerial scenes. Several algorithms were evaluated on top of these benchmarks with in-depth analyses, resulting in discussions on backbone effectiveness, hierarchical feature representation efficiency, and one-stage detector performance for small object detection. In addition, several studies were performed on vehicle and car detection16,17,18,19. These studies17,18 focus on the development of new vehicle detection models, as well as the enhancement of existing ones, utilizing publicly available datasets such as DOTA20, VEDAI21, and xView22, fMoW23, VAID24, and AI-TOD25.However, existing models for vehicle detection face challenges when applied to diverse real-world scenarios involving the analysis of satellite images from previously unexplored geographic regions26. For example, the visual context of a car on the road in Abu Kamal City, Syria (Fig. 1a) and Alexandria City, Egypt (Fig. 1b) present clear differences compared to a car on the road in Sydney, Australia (Fig. 1c) and Mexico City, Mexico (Fig. 1d). A noticeable contrast is evident in the appearance of built structures and land cover, stemming from unique differences in the natural landscape, climate, economic development, urban planning, and architectural design in Middle Eastern countries. This contrast becomes more pronounced thanks to the rapid pace of urban development in Middle Eastern countries driven by large-scale smart city projects as opposed to the more incremental urban upgrades seen in the US and Europe27,28.Fig. 1The distinct visual context of cars on the road in the Middle Eastern cities: (a) Abu Kamal, Syria, and (b) Alexandria, Egypt & other cities around the world: (c) Sydney, Australia22, and (d) Mexico City, Mexico22.Full size imageTherefore, with the prevalence of datasets focusing on specific regions, a gap related to geographic bias has emerged, particularly in the Middle East as highlighted in Fig. 2. To bridge this gap, the VME dataset, collected from Maxar, spans 54 cities in 12 countries in the Middle East and comprises more than 4,000 high-resolution image tiles of 512 × 512 pixels with more than 100k vehicle instances. The ground-truth annotations were generated using a combination of manual annotation and semi-automated techniques through a crowdsourcing company. Additionally, the CDSI dataset constitutes the largest benchmark for car detection by expanding VME with images from other existing satellite imagery datasets, such as xView22, DOTA-v2.020, VEDAI21, DIOR29, and FAIR1M-2.030.Fig. 2The geographical distribution of VME, and xView, denoted as purple and blue circles, respectively. There is no geographical information reported for the remaining datasets.Full size imageWe conduct comprehensive experiments using advanced object detection models, such as TOOD31 and DINO32, and present baseline results on both individual and combined datasets. The VME baseline evaluation demonstrates a remarkable 56.3% improvement in mAP for car detection in the Middle East compared to models trained on existing datasets. Additionally, the model trained on the CDSI dataset, due to its greater diversity and scale, significantly enhances mAP50, with improvements ranging from 19.6% to 84.6% across all models trained on individual datasets. This newly developed dataset serves as a valuable resource for researchers and professionals in remote sensing, promoting progress in vehicle detection and satellite imagery analysis.MethodsThis section provides details about our novel VME dataset such as the different categories, image resolution, area coverage, and annotation format. Then, we elaborate on the new benchmark dataset (CDSI), where we collect car-related objects from the publicly available datasets and combine them with the VME dataset.VME DatasetWe constructed the VME dataset by collecting satellite images of different cities in the Middle Eastern countries such as Syria, Libya, Iraq, Jordan, Egypt, Qatar, Saudi Arabia, United Arab Emirates, Oman, Kuwait, and Bahrain. We included the most popular cities including the capitals of these countries. The city-level geographic distribution of the collected images in the VME dataset is highlighted with purple circles in Fig. 2, which includes underrepresented geographic regions for vehicle detection in satellite imagery, compared to the blue circles representing the distribution of images in the xView dataset. We note that the remaining datasets do not provide any geographical information at the country or city level and, hence, cannot be accurately represented on the map.Image CollectionFor each city in our dataset, we identified the geographic area of interest (AOI) and collected high-resolution satellite images from Maxar Technologies, which provides access to a large archive of the world’s most recent pan-sharpened natural color images at a spatial resolution of up to 30 cm through a paid subscription to their SecureWatch platform. To this end, we searched the archive for satellite images with (i) RGB color, (ii) less than 20% cloud coverage, (iii) a ground sampling distance of at most 50 cm (i.e., images at 30 cm, 40 cm, and 50 cm spatial resolution), and (iv) off-Nadir angle less than 30 degrees.We downloaded a total of 2,714 image snapshots across all 54 city AOIs. The resulting images are large with an average dimension in the range of 22,475 × 24,043. Since this image size is too large for processing and labeling directly, we generated random crops of image tiles with 512 × 512 pixels. This initially yielded a total of 22,125 image tiles. We ensured the resulting tiles did not have any missing or undefined pixels. Furthermore, to keep the annotation budget under control, we manually discarded the tiles that did not have obvious objects to annotate, such as images with completely green or desert areas. As a result of this filtering, we had 4,303 image tiles to be annotated in the next step.Image AnnotationAfter inspecting the taxonomies of the existing satellite imagery datasets for vehicle-related classes, we defined a three-class taxonomy comprising car, bus, and truck classes in our dataset. We decided to collect oriented bounding box (OBB) annotations as certain applications, such as traffic management, can leverage the direction information as well. We employed Co-one (https://www.co-one.co/), an AI and crowdsourcing-based data platform that promises 95% annotation accuracy. The annotation process started with preparing the guideline handbook, which outlined the project overview, technical guidelines, targeted categories with definitions and examples, rules and tips for the annotation process, and the deliverable format. Then, the data annotation process was conducted with a crowdsource of 6000+ people, where each group focused on a specific category. Finally, an annotation review process was implemented to detect mislabeled annotations via a cross-validation system; and an expert was employed to correct such cases. After the annotation quality review process, final annotations were delivered. We provided images in lossless PNG format, and received OBB annotations in YOLO format as text (*.txt) files. Each annotation file contains the image name, and each line in the file represents a targeted object as follows: x1, y1, x2, y2, x3, y3, x4, y4, category_id, where (x1, y1) is the top left, (x2, y2) is the top right, (x3, y3) is the bottom right, and (x4, y4) is the bottom left point of OBB, and category_id indicates the class index as 0, 1, 2 corresponding to car, bus, and truck, respectively. Additionally, we obtained standard horizontal bounding box (HBB) annotations based on the minimum and maximum x and y coordinates of the OBB annotations with their category. To better help the community utilize the dataset, we provide both the oriented and horizontal bounding box annotation files.Final DatasetOut of 4,303 images annotated, 21 images were deemed damaged or corrupted and excluded from the dataset. Hence, the final dataset contains 4,282 images with a total of 113,737 objects comprising 101,564 cars, 5,327 buses, and 6,846 trucks while 241 images do not contain any instances of the target object classes and are tagged as no_label. The distribution of classes is shown in Fig. 3a. Also, Fig. 3b,c,d highlight the area distribution of cars, buses, and trucks in pixels, respectively. It is observed that all of the car instances fall under the small object range (i.e., area (pixels) < 322) defined in the MS-COCO evaluation, specifically within the first half of the range (i.e., area (pixels) < 512) which is considered tiny objects. On the other hand, both the bus and truck instances fall mostly within the small object range (area (pixels) < 322), with almost negligible overlap into the medium object range (322 < area (pixels) < 962). We provide training, validation, and test sets of the dataset following a random split with a ratio of 5/8, 1/8, and 2/8, respectively. Table 1 presents the statistics for all VME categories, outlining the details of each split.Fig. 3Statistical properties of the object categories in the VME dataset. (a) Distribution of VME categories, (b) Area distribution of cars, (c) Area distribution of buses, (d) Area distribution of trucks.Full size imageTable 1 Number of images and annotations in each category across training, validation, and test splits of the VME dataset.Full size tableCDSI DatasetThis section introduces the related object detection datasets in satellite imagery, such as xView, DOTA-v2.0, VEDAI, FAIR1M-2.0, and DIOR. Also, it describes the filtering and consolidation process of the CDSI dataset.Existing DatasetsWe explored a large list of publicly available datasets for object detection to employ in our study. We excluded those with low-altitude, drone-based, and UAV-based datasets and the datasets with high ground sample distance (GSD) ranges or hidden contexts, such as COWC33, PaCaBa34, PSU35, and VisDrone36. Even some of the new datasets are not yet released, e.g., VehSat37 and EAGLE38. The following are the datasets we employed in our study.xView22 is considered one of the largest publicly available datasets containing 846 images collected from Maxar at various locations around the world, as shown in Fig. 2. The images are available with 30cm/pixel spatial resolution, and the average dimension of the images is 3316 × 2911. The dataset has 60 object classes with 1 million object instances annotated using horizontal bounding boxes for all splits, while the ground truth of testing split is not available. The xView repository (https://challenge.xviewdataset.org/data-download) provides the training and validation images in TIF format and the annotations in GeoJSON format.DOTA-v2.020 contains 2,423 images gathered from Google Earth, different satellites supplied by the Resources Satellite Data and Application Center in China, and aerial images supplied by CycloMedia B.V. The size of the images ranges from 800 to 20,000 pixels, and their spatial resolution varies between 0.1m/pixel to 4.5m/pixel. The dataset contains 18 object classes and objects are annotated using both oriented and horizontal bounding boxes. The dataset is presented in three versions, where the final version (v2.0) contains a total of 1,793,658 object instances of all splits, while the ground truth of the testing split is not available. The DOTA images were released with no geographical information. We obtained DOTA-v2.0 from its repository (https://captain-whu.github.io/DOTA/index.html); the images are in PNG format and its annotations are in YOLO (TXT) format. Acquiring DOTA-v2.0 requires the users to download DOTA-v1.0 first, then get the v2.0.VEDAI21 was built specifically for detecting vehicles in satellite imagery, such as boats, planes, tractors, cars, and vans. The dataset provides two sets of 1,246 images in colored and infrared format, each set at a different spatial resolution (12.5cm/pixel or 25cm/pixel), and hence, image dimensions (1024 × 1024 or 512 × 512 pixels). The annotation format used for the dataset is the oriented bounding box. No geographic information is revealed in VEDAI. In our study, we downloaded colored images with a spatial resolution of 25cm/pixel (i.e., image dimensions of 512 × 512 pixels) from the VEDAI repository (https://downloads.greyc.fr/vedai/). The annotations are stored in TXT files, reporting the four corners of OBBs with the category.DIOR29 is another large-scale benchmark dataset for object detection in optical satellite images. It consists of 23,463 images annotated for 20 object categories and 192,512 object instances using horizontal bounding boxes. Spatial resolution of images is between 0.5m/pixel and 30m/pixel. The dataset claims to cover more than 80 countries, but the specific list of countries has not been released. The dataset can be downloaded from DIOR repository (https://gcheng-nwpu.github.io/); which delivers the images in JPG format and the annotation files in PASCAL-VOC (XML) format.FAIR1M-2.030 contains more than 20,000 images with more than 1 million instances of fine-grained object categories. The images are gathered from Google Earth and the Gaofen satellites, with spatial resolutions between 0.3m/pixel and 0.8m/pixel. The object annotations were collected for five main categories and 37 sub-categories using oriented bounding boxes. It is stated that the dataset covers different continents; but the country- or city-level details about the image locations are not published. The dataset can be obtained from FAIR1M repository (https://gaofen-challenge.com/benchmark); its annotation files are presented in PASCAL-VOC (XML) format, and the images are offered in TIF format.Category MappingTo construct a unified benchmark dataset for car detection in satellite imagery, we investigated the taxonomies of the aforementioned datasets. Each dataset labels car-related objects differently, using terms like “small car,” “small vehicle,” “vehicle,” “car,” or “van.” Thus, we visually inspected these categories to ensure they correspond to the same “car” object we are targeting. For instance, “small car” in xView and “small vehicle” in DOTA-v2.0 refer to standard cars, while in DIOR “vehicle” covers a broader range of vehicles (e.g., cars, trucks, buses, and vans), with “car” being a subset of this general category. Figure 4 illustrates the car-related objects across datasets that we target for constructing the CDSI dataset. We conclude that these classes can be mapped to the same object type, i.e., car, with certain conditions such as filtering by typical car size. Specifically, car-related categories were mapped to the car category in CDSI for objects with an HBB area of less than 400 pixels, as detailed in Table 2. To avoid the challenges associated with training an object detection model using only a single class, we ensured the model encountered hard negatives–objects similar in size to cars. To achieve this, we opted to group all other small objects into a single category called “other.” Similarly, instances from all other categories with an HBB area of less than 400 pixels were mapped to the other small object category in CDSI, as reported in Table 2. As a result, the CDSI dataset consists of two classes: “car” and “other.” The details about data processing and filtering steps are explained next.Fig. 4Example images with car-related objects in (a) xView22, (b) DOTA-v2.020, (c) VEDAI21, (d) DIOR29, (e) FAIR1M-2.030, (f) VME (our) datasets.Full size imageTable 2 Statistics of car-related and other small object categories in different datasets.Full size tableData Processing and FilteringEach dataset uses a different annotation style (e.g., OBB, HBB, or both) and adopts different data representation and file format (e.g., XML files with PASCAL-VOC format, TXT files with YOLO format, JSON files with MS-COCO format, etc.). To consolidate all of the datasets, we designed a data processing pipeline, illustrated in Fig. 5, with the following steps:

Annotation standardization: We standardize all the annotations from different datasets to HBB style. Then, we convert standardized annotations into MS-COCO format which is defined by four values in pixels (x_min, y_min, width, height).

Car-related object size filtering: Given that we are interested in a GSD range of 30-50cm per pixel, we assume that an object with an area greater than 400 pixels is unlikely a car. To verify this assumption, we analyze the car size distributions in all datasets as in Fig. 6. This analysis reveals that an area size of less than 400 pixels reports for more than 90% of all car-related object instances across all datasets. Therefore, we decided to filter out all object instances with an area larger than 400 pixels; even if they were originally labeled as cars. During our visual inspection, we discovered that these cases often relate to labeling errors or images with spatial resolutions exceeding the targeted GSD range.

Relabeling small objects: Using the same threshold, we repeat Step 2 to identify all other small object instances with an area less than 400 pixels and label them as “other” category.

Training setups: Depending on the experimental setup, the car-related object instances are merged with the other small object instances to construct car-other setup for the model training. In contrast, only car-related objects are employed to form car setup for the model training (refer to the “Technical Validation” section for details).

Fig. 5Dataset consolidation pipeline and final experimental setups.Full size imageFig. 6Distribution of car sizes in (a) xView, (b) DOTA-v2.0, (c) VEDAI, (d) FAIR1M-2.0, (e) DIOR, and (f) VME (our) datasets.Full size imageFinal DatasetTable 2 provides general information and summary statistics about all datasets (individual or consolidated) before and after the data processing and filtering pipeline. The final combined dataset, i.e., CDSI, contains a total of 23,250 images with 896,760 car-related object instances and 185,619 other small object instances. Note that we also created a version of CDSI, denoted as CDSI*, where we excluded VME dataset from the consolidation process to highlight the contribution of VME dataset. With regards to the training, validation, and test sets, we first created random splits of all images in each individual dataset after filtering with a ratio of 5/8, 1/8, and 2/8 (as in the VME dataset). We then combined the resulting splits from different datasets to form the final data splits for both CDSI and CDSI* datasets. For instance, the CDSI training set is simply a union of training sets of all datasets, and the same rule applies for validation and test sets.Data RecordsThe repository available at Zenodo39 consists of (a) the VME dataset, including satellite images and annotation files, and (b) the scripts and instructions for creating the CDSI dataset.Overview of the repository files and their formatsThe repository is structured into four components as follows:

annotations_OBB: This folder holds TXT files in YOLO format with Oriented Bounding Box (OBB) annotations. Each annotation file is named after the corresponding image name, with each line describing a targeted object as follows: x1, y1, x2, y2, x3, y3, x4, y4, category_id, where (x1, y1) is the top left, (x2, y2) is the top right, (x3, y3) is the bottom right, and (x4, y4) is the bottom left point of OBB, and category_id indicates the class index as 0, 1, 2 corresponding to car, bus, and truck, respectively. The annotation files of images that do not include any of the targeted objects are empty.

annotations_HBB: This folder contains HBB annotations in separate JSON files for training, validation, and test splits, formatted according to the MS-COCO standard defined by four values in pixels (x_min, y_min, width, height).

satellite_images: This folder contains VME images in PNG format, each with a resolution of 512 × 512 pixels.

CDSI_construction_scripts: This directory contains all the necessary instructions for constructing the CDSI dataset, including: (a) guidelines for downloading each dataset from its respective repository, (b) scripts for converting each dataset to the MS-COCO format, located within the corresponding dataset folders, and (c) instructions for combining the datasets. The training, validation, and test splits are provided in the CDSI_construction_scripts/data_utils folder. Each split file lists the images from each dataset used in the car detection experiments for both detectors.

Additional information on the environment setup and required packages is available in the README.md file.Technical ValidationIn this section, we perform a formal assessment of the quality of the VME annotations. Additionally, we provide details on benchmarks conducted across diverse setups and present analytical results to demonstrate the reliability and validity of the VME and CDSI datasets.VME Annotation QualityWe implemented quality control to ensure the accuracy and consistency of the VME dataset annotations. We randomly selected around 5% of images across all 54 cities and resolutions and labeled these images in-house (by the lead author) to establish ground truth. This process yielded 5,664 ground truth annotations in 215 images. Next, we compared these labels with the annotations from the crowdsource-based platform to calculate True Positives (TP), False Positives (FP), and False Negatives (FN). We identified 5,496 TP, 7 FP, and 168 FN annotations. We then used these values to compute precision, recall, and F1 scores as 0.999, 0.970, and 0.984, respectively. Although some objects were missed, the crucial factor is that the identified objects are indeed the targeted ones, making the minimization of False Positives a priority. This process demonstrates that the annotations are highly accurate.Detection BenchmarksThis section describes the benchmark setup and the application of state-of-the-art detection models to evaluate the technical quality and scientific significance of the VME and CDSI datasets. To this end, we explored three different setups to assess how varying number of images and objects (not necessarily cars) in a dataset affects the detection performance. In the first setup, we use the original datasets with their full taxonomy (i.e., all categories) to train object detection models. In the second setup, we use datasets containing only the images with instances of car and other small object categories. And, in the last setup, we use datasets containing only the images with car instances. To facilitate model training in the first setup, we created distinct training, validation, and test splits based on all the images in the original datasets using a ratio of 5/8, 1/8, and 2/8, respectively. In the second and third setups, these initial data splits were reduced to subsets containing only those images with relevant object instances. It is important to note that, at training time, we utilized each dataset’s training and validation sets. However, at test time, we evaluated all trained models on the car-only test sets to obtain comparable performance scores. Table 3 presents the number of images and annotations across data splits and training setups (i.e., original, car-other, and car) for different datasets.Table 3 Statistics of the training, validation, and test splits in each experimental setup across datasets.Full size tableWe conducted experiments using a state-of-the-art framework, Slicing Aided Hyper Inference (SAHI)40. SAHI is developed particularly for small object detection and provides a generalized slicing-aided inference and fine-tuning channel for detecting small objects. In SAHI, various object detectors were examined such as Fully Convolutional One-Stage Object Detection (FCOS)41, VarifocalNet (VFNET)42 and Task-aligned One-stage Object Detection (TOOD)31. In our study, we adopted the best-performing inference setup reported in SAHI, which is Slicing Aided Fine-tuning, Full-Inference, and Patch Overlap (SAHI+FI+PO) setting with TOOD detector from MMDetection library43. Additionally, we performed experiments with a more recent object detector called DINO32 with Swin-L option from the MMDetection library following the SAHI+FI+PO inference setting. We trained a total of 22 models using the TOOD detector with a batch size of 16 for 24 epochs with SGD optimizer. For the original setup, we started training with a learning rate of 0.01 whereas, for the other setups, we started with a learning rate of 0.005. In all training setups, the learning rate was configured to change at epochs 9, 16, 22 with a learning rate decay equal to 0.1. Similarly, we trained a total of 22 models using the DINO Swin-L detector with a batch size of 2 for 36 epochs with AdamW optimizer with an initial learning rate of 0.0001, which was configured to change at epochs 27 and 33 with learning rate decay equal to 0.1. We ran all of our experiments on an NVIDIA A100 80GB GPU.VME BenchmarkAs we introduce our novel dataset for the first time, we perform experiments to provide baseline results. For this purpose, we train and test models with the original VME categories, utilizing both TOOD and DINO Swin-L detectors. Table 4 presents the class-specific and overall results obtained on the original VME test set. TOOD achieved an overall mAP50 score of 58.5% whereas DINO Swin-L achieved 62.7%. Notably, DINO Swin-L outperforms TOOD by 7.2%, with improvements of 6.2%, 5.9%, and 10.2% in mAP50 scores of car, bus, and truck categories, respectively. These baseline results highlight the challenging nature of the vehicle detection task and verifies our dataset’s reliability for this challenging task. Given these results, we believe our novel dataset focused on Middle Eastern cities will play a key role in advancing vehicle detection in similar regions. Table 4 VME baseline results obtained by training and testing the object detection models on the original VME data splits with all object categories as presented in Table 1.Full size tableFigure 7 illustrates some examples of detection results from the baseline model applied to images from the Middle East sampled from the VME dataset. FP and FN are highlighted with yellow and magenta circles, respectively. To prevent clutter, detections for each object category are visualized separately. The results demonstrate the model’s high detection accuracy, with occasional FP detections and rare FN occurrences, reflecting strong recall performance. These findings underscore the model’s robustness while identifying opportunities for reducing FP rates.Fig. 7Detections on VME images employing VME baseline model trained on all categories. Yellow and magenta circles indicate examples of false positives and false negatives, respectively.Full size imageCDSI BenchmarkThis section provides a comprehensive benchmark across various datasets and setups, emphasizing the enhanced value introduced by the CDSI dataset. Additional analyses, including error evaluation and data visualization, further illustrate the strengths and limitations of the CDSI benchmark.Table 5 summarizes the results achieved by both detectors, TOOD and DINO Swin-L, on CDSI and its constituents. Each row corresponds to a model trained on a particular dataset with a specific setup, e.g., all categories, car-other, and car. We evaluate each trained model on its own test set to quantify its in-domain performance as well as on the VME and CDSI test sets to assess its generalization capabilities. As highlighted before, we use car-only test sets in all cases for comparable results, which we discuss next. First, we observe that all the models trained on individual datasets exhibit poor performance on the VME dataset. Furthermore, the car detection performance does not improve even after combining all the existing datasets together (i.e., CDSI*). In essence, the models trained on existing datasets cannot effectively detect cars in images from the Middle East. In Fig. 8, the predictions of the VME car setup model are compared with the predictions of the models trained on xView and DOTA-v2.0 car setup on example images from the Middle East sampled from the VME dataset. The comparison shows that the models trained on xView and DOTA-v2.0 car setup struggle with detecting cars properly sometimes even in easy scenarios like cars on paved roads (top row).Table 5 Experimental results achieved by TOOD and DINO Swin-L detectors trained on various datasets under different setups.Full size tableFig. 8Comparison of detections on VME images employing the model trained on VME car setup versus detections of the models trained on xView and DOTA-v2.0 car setup.Full size imageUpon examining the CDSI dataset, the table presents the evaluation of predicting the CDSI test set across the trained models on each dataset individually, as well as the trained model on the CDSI. The results highlight the significance of the trained model on images from diverse sources, particularly in the context of detecting cars in satellite imagery. Additionally, the findings underscore the impact of incorporating the VME dataset in the trained model on car setup, revealing that the exclusion of VME in the trained model (CDSI*) leads to a decrease in mAP50 by 6% and 4.3% in TOOD and DINO Swin-L, respectively, when predicting on the CDSI test set.To gain a deeper understanding of the significance of combining datasets (CDSI), we employed the Prithvi Foundation Model, a collaboration between IBM and NASA44,45, which was pretrained on large-scale remote sensing data, including Harmonised Landsat Sentinel 2 (HLS). We utilized the IBM-NASA-Geospatial pretrained model with t-SNE (t-Distributed Stochastic Neighbor Embedding)46, an unsupervised non-linear technique for visualizing feature embeddings, to explore how satellite images are represented in low-dimensional space based on their high-dimensional data. The t-SNE visualization helps in understanding the similarity between points, in this case, different satellite images from various datasets. The results, shown in Fig. 9, illustrate that the features of the FAIR1M-2.0 dataset are distinctly separate from the others. Additionally, xView shares some features with DOTA-v2.0 and DIOR, while VME shares certain features with DIOR and VEDAI. This outcome emphasizes the extent of training a model on a combined dataset for car detection (CDSI) and its implications for the field of car detection.Fig. 9t-SNE visualization of the proposed CDSI dataset.Full size imageTo delve deeper into the details of Table 5, we analyze the performance across various setups using individually trained datasets, VME and CDSI, and assess how their predictions performed on their respective car-only test sets, VME car-only test set and CDSI test set. Overall, the car setup performed better with the TOOD detector in most setups, except for VEDAI and DIOR, which produced better results in the car-other setup in terms of mAP50(%). When focusing on TOOD models trained on other datasets in the car setup and evaluated on the VME test set, results reveal poor performance. Notably, low mAP50 scores were observed for models trained on VEDAI (5.1%) and FAIR1M-2.0 (15.8%), likely due to VEDAI’s limited number of images and annotations, which struggled with car detection in images with varied resolutions and higher car densities. Despite FAIR1M-2.0 being the largest dataset in terms of car-related objects and images, Fig. 9 indicates that its image features differ significantly from those of the VME dataset. A similar pattern is seen in all categories and car-other setups for all models. On the other hand, DINO Swin-L shows slight improvement across all trained models, mirroring the pattern observed with TOOD. Notably, the model trained on CDSI in the car-other setup achieved the highest mAP50 score (86.8%) on the VME test set.To investigate the root causes of errors, we perform an analysis of the detection results47 from the DINO Swin-L model trained on VME and CDSI using the car-other setup. Figures 10 and 11 show a breakdown of errors for the car class for VME and CDSI, respectively. The error analysis provides various insights to identify areas for improvement, including: 1) IoU thresholds of 0.75, 2) IoU thresholds of 0.50, 3) post-localization error removal, 4) false positives within supercategories, 5) category confusion, 6) background false positives, and 7) false negatives, represented as C75, C50, Loc, Sim, Oth, BG, and FN, respectively. Note that the area under each Precision-Recall curve is shown in brackets in the legend. In the case of VME (Fig. 10), overall AP at IoU=.75 is 0.432 (C75), and simply lowering IoU=0.5 increases the AP to 0.861 (C50), whereas perfect localization could increase AP to 0.898 (Loc). We observe some error due to the confusion between the car and other categories and removing such class confusions would only raise AP slightly to 0.909 (Oth). However, we see a bigger room for improvement by eliminating background false positives (i.e., confusions with other small background objects), which boosts the AP to 0.99 (BG). Surprisingly, in the case of VME, the model does not suffer too much from false negatives (i.e., missed detections). On the other hand, for the model trained on CDSI (Fig. 11), we see similar trends in general regarding the errors due to category confusions and background false positives. However, resolving such issues can boost AP to a maximum of 0.851 (BG), which means the rest of the errors are missing detections. The missed detections in the model trained on CDSI are due to the diversity in object instances and variations in image characteristics collected from different regions. In summary, both plots illustrate that the errors are dominated by imperfect localization and background confusions.Fig. 10Error analysis for the car category of the DINO Swin-L detector trained on VME using the car-other setup.Full size imageFig. 11Error analysis for the car category of the DINO Swin-L detector trained on CDSI using the car-other setup.Full size imageUsage NotesThe VME dataset and the script for creating the CDSI dataset are available at Zenodo39. VME images are available in resolutions ranging from 30 to 50 cm per pixel. However, the climate conditions in the Middle East, including haze and airborne dust, can affect the clarity of these images. As a result, some images may have a blurry appearance or exhibit reflections.

Code availability

The data preprocessing script for constructing CDSI dataset, which is written in Python, is available on Zenodo39 and GitHub repository (https://github.com/nalemadi/VME_CDSI_dataset_benchmark) under CDSI_construction_scripts folder. The file README.md provides detailed instructions for building the CDSI dataset, which includes downloading the datasets, converting each to MS-COCO format, and explaining the combination mechanism. Each subfolder is named after its corresponding dataset and contains a conversion script to MS-COCO format. All the required Python packages are listed in the requirements.txt file located within the CDSI_construction_scripts folder.

ReferencesNguyen, T. T. et al. Monitoring agriculture areas with satellite images and deep learning. Applied Soft Computing 95, 106565 (2020).

Google Scholar 

Wang, Y., Cai, G., Yang, L., Zhang, N. & Du, M. Monitoring of urban ecological environment including air quality using satellite imagery. Plos one 17, e0266759 (2022).CAS 

PubMed 

PubMed Central 

Google Scholar 

Albert, A., Kaur, J. & Gonzalez, M. C. Using convolutional networks and satellite imagery to identify patterns in urban environments at a large scale. In Proceedings of the 23rd ACM SIGKDD international conference on knowledge discovery and data mining, 1357–1366 (2017).Huang, X. et al. Urban building classification (ubc)-a dataset for individual building detection and classification from satellite imagery. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 1413–1421 (2022).Higuchi, A. Toward more integrated utilizations of geostationary satellite data for disaster management and risk mitigation. Remote Sensing 13, 1553 (2021).ADS 

MATH 

Google Scholar 

Gui, S., Song, S., Qin, R. & Tang, Y. Remote sensing object detection in the deep learning era–a review. Remote Sensing 16, 327 (2024).ADS 

MATH 

Google Scholar 

Drouyer, S. & de Franchis, C. Highway traffic monitoring on medium resolution satellite images. In IGARSS 2019-2019 IEEE International Geoscience and Remote Sensing Symposium, 1228–1231 (IEEE, 2019).Chen, Y., Qin, R., Zhang, G. & Albanwan, H. Spatial temporal analysis of traffic patterns during the covid-19 epidemic by vehicle detection using planet remote-sensing satellite images. Remote Sensing 13, 208 (2021).ADS 

Google Scholar 

Golej, P., Horak, J., Kukuliac, P. & Orlikova, L. Vehicle detection using panchromatic high-resolution satellite images as a support for urban planning. case study of prague’s centre.GeoScape 16 (2022).Rufener, M.-C., Ofli, F., Fatehkia, M. & Weber, I. Estimation of internal displacement in ukraine from satellite-based car detections. Sci. Reports 14, 31638 (2024).Liu, H.-I. et al. A denoising fpn with transformer r-cnn for tiny object detection. IEEE Transactions on Geoscience and Remote Sensing (2024).Verma, T. et al. Soar: Advancements in small body object detection for aerial imagery using state space models and programmable gradients. Preprint at https://doi.org/10.48550/arXiv.2405.01699 (2024).Zhu, J. et al. Transformer based remote sensing object detection with enhanced multispectral feature extraction. IEEE Geoscience and Remote Sensing Letters (2023).Tong, K., Wu, Y. & Zhou, F. Recent advances in small object detection based on deep learning: A review. Image and Vision Computing 97, 103910 (2020).MATH 

Google Scholar 

Cheng, G. et al. Towards large-scale small object detection: Survey and benchmarks. IEEE Transactions on Pattern Analysis and Machine Intelligence (2023).Gao, P., Tian, T., Li, L., Ma, J. & Tian, J. De-cyclegan: An object enhancement network for weak vehicle detection in satellite images. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing 14, 3403–3414 (2021).ADS 

MATH 

Google Scholar 

Du, Q., Celik, T., Wang, Q. & Li, H.-C. Fully convolutional lightweight pyramid network for vehicle detection in aerial images. IEEE Geoscience and Remote Sensing Letters (2021).Li, X. et al. Vehicle detection in very-high-resolution remote sensing images based on an anchor-free detection model with a more precise foveal area. ISPRS International Journal of Geo-Information 10, 549 (2021).ADS 

MATH 

Google Scholar 

Shi, F., Zhang, T. & Zhang, T. Orientation-aware vehicle detection in aerial images via an anchor-free object detection approach. IEEE Transactions on Geoscience and Remote Sensing 59, 5221–5233 (2020).ADS 

MATH 

Google Scholar 

Ding, J. et al. Object detection in aerial images: A large-scale benchmark and challenges. IEEE transactions on pattern analysis and machine intelligence 44, 7778–7796, https://doi.org/10.1109/TPAMI.2021.3117983 (2021).Article 

MATH 

Google Scholar 

Razakarivony, S. & Jurie, F. Vehicle detection in aerial imagery: A small target detection benchmark. Journal of Visual Communication and Image Representation 34, 187–203, https://doi.org/10.1016/j.jvcir.2015.11.002 (2016).Article 

MATH 

Google Scholar 

Lam, D. et al. xView: Objects in context in overhead imagery. Preprint at https://doi.org/10.48550/arXiv.1802.07856 (2018).Christie, G., Fendley, N., Wilson, J. & Mukherjee, R. Functional map of the world. In CVPR (2018).Lin, H.-Y., Tu, K.-C. & Li, C.-Y. Vaid: An aerial image dataset for vehicle detection and classification. IEEE Access 8, 212209–212219 (2020).MATH 

Google Scholar 

Wang, J., Yang, W., Guo, H., Zhang, R. & Xia, G.-S. Tiny object detection in aerial images. In 2020 25th international conference on pattern recognition (ICPR), 3791–3798 (IEEE, 2021).Minetto, R., Segundo, M. P., Rotich, G. & Sarkar, S. Measuring human and economic activity from satellite imagery to support city-scale decision-making during covid-19 pandemic. IEEE Transactions on Big Data 7, 56–68 (2020).PubMed 

Google Scholar 

ZIGURAT Institute of Technology, Z. 7 Impressive Smart City Projects in the Middle East. https://www.e-zigurat.com/en/blog/smart-city-projects-middle-east/ Accessed on 2024-09-17 (2023).George, R. The Rise of Gulf Smart Cities. Wilson Center. Accessed on 2024-09-18 https://www.wilsoncenter.org/article/rise-gulf-smart-cities Accessed on 2024-09-18 (2024).Li, K., Wan, G., Cheng, G., Meng, L. & Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS Journal of Photogrammetry and Remote Sensing 159, 296–307, https://doi.org/10.1016/j.isprsjprs.2019.11.023 (2020).Article 

ADS 

MATH 

Google Scholar 

Sun, X. et al. FAIR1M: A benchmark dataset for fine-grained object recognition in high-resolution remote sensing imagery. ISPRS Journal of Photogrammetry and Remote Sensing 184, 116–130, https://doi.org/10.1016/j.isprsjprs.2021.12.004 (2022).Article 

ADS 

MATH 

Google Scholar 

Feng, C., Zhong, Y., Gao, Y., Scott, M. R. & Huang, W. Tood: Task-aligned one-stage object detection. In 2021 IEEE/CVF International Conference on Computer Vision (ICCV), 3490–3499 (IEEE Computer Society, 2021).Zhang, H. et al. DINO: DETR with improved denoising anchor boxes for end-to-end object detection. In The Eleventh International Conference on Learning Representations (ICLR) (2023).Mundhenk, T. N., Konjevod, G., Sakla, W. A. & Boakye, K. A large contextual dataset for classification, detection and counting of cars with deep learning. In European Conference on Computer Vision, 785–800 (Springer, 2016).Zambanini, S., Loghin, A.-M., Pfeifer, N., Soley, E. M. & Sablatnig, R. Detection of parking cars in stereo satellite images. Remote Sensing 12, 2170 (2020).ADS 

Google Scholar 

Ammar, A., Koubaa, A., Ahmed, M., Saad, A. & Benjdira, B. Vehicle detection from aerial images using deep learning: A comparative study. Electronics 10, 820 (2021).Zhu, P. et al. Detection and tracking meet drones challenge. IEEE Transactions on Pattern Analysis and Machine Intelligence 44, 7380–7399 (2021).MATH 

Google Scholar 

Drouyer, S. VehSat: a large-scale dataset for vehicle detection in satellite images. In IGARSS 2020 - 2020 IEEE International Geoscience and Remote Sensing Symposium, 268–271, https://doi.org/10.1109/IGARSS39084.2020.9323289 ISSN: 2153-7003 (2020).Azimi, S. M., Bahmanyar, R., Henry, C. & Kurz, F. Eagle: Large-scale vehicle detection dataset in real-world scenarios using aerial imagery. In 2020 25th International Conference on Pattern Recognition (ICPR), 6920–6927 (IEEE, 2021).Al-Emadi, N., Weber, I., Yang, Y. & Ofli, F. VME: A Satellite Imagery Dataset and Benchmark for Detecting Vehicles in the Middle East and Beyond. https://doi.org/10.5281/zenodo.14185684 (2024).Akyon, F. C., Altinuc, S. O. & Temizel, A. Slicing aided hyper inference and fine-tuning for small object detection. In 2022 IEEE International Conference on Image Processing (ICIP), 966–970 (IEEE, 2022).Tian, Z., Shen, C., Chen, H. & He, T. FCOS: Fully Convolutional One-Stage Object Detection. In 2019 IEEE/CVF International Conference on Computer Vision (ICCV), 9626–9635, https://doi.org/10.1109/ICCV.2019.00972 (IEEE, 2019).Zhang, H., Wang, Y., Dayoub, F. & Sunderhauf, N. Varifocalnet: An iou-aware dense object detector. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 8514–8523 (2021).Chen, K. et al. MMDetection: Open mmlab detection toolbox and benchmark. Preprint at https://doi.org/10.48550/arXiv.1906.07155 (2019).Jakubik, J. et al. Foundation models for generalist geospatial artificial intelligence. Preprint at https://doi.org/10.48550/arXiv.2310.18660 (2023).Jakubik, J. et al. HLS Foundation, Prithvi-100M, https://doi.org/10.57967/hf/0952 (2023).Van der Maaten, L. & Hinton, G. Visualizing data using t-sne. Journal of machine learning research 9 (2008).Hoiem, D., Chodpathumwan, Y. & Dai, Q. Diagnosing error in object detectors. In European conference on computer vision, 340–353 (Springer, 2012).Download referencesAcknowledgementsThis publication was made possible by GSRA grant, I.D. # GSRA7-1-0421-20022, from the Qatar National Research Fund (a member of Qatar Foundation). We sincerely thank our colleague Masoomali Fatehkia (Qatar Computing Research Institute, HBKU) for assisting with image collection. Ingmar Weber is supported by funding from the Alexander von Humboldt Foundation and its founder, the Federal Ministry of Education and Research (Bundesministerium für Bildung und Forschung).Author informationAuthors and AffiliationsQatar Computing Research Institute, Hamad Bin Khalifa University, Doha, QatarNoora Al-Emadi & Ferda OfliCollege of Science and Engineering, Hamad Bin Khalifa University, Doha, QatarNoora Al-Emadi & Yin YangSaarland Informatics Campus, Saarland University, Saarbrücken, GermanyIngmar WeberAuthorsNoora Al-EmadiView author publicationsYou can also search for this author inPubMed Google ScholarIngmar WeberView author publicationsYou can also search for this author inPubMed Google ScholarYin YangView author publicationsYou can also search for this author inPubMed Google ScholarFerda OfliView author publicationsYou can also search for this author inPubMed Google ScholarContributionsN.A. conceived the dataset collection and preparation, dataset validation, and pre-processing, conducted the experiments, and wrote the manuscript. I.W. and F.O. facilitated access to Middle East satellite imagery. F.O. provided analysis techniques. I.W., Y.Y., and F.O. supervised and guided the study. All authors reviewed the manuscript.Corresponding authorCorrespondence to

Noora Al-Emadi.Ethics declarations

Competing interests

The authors declare no competing interests.

Additional informationPublisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.

Reprints and permissionsAbout this articleCite this articleAl-Emadi, N., Weber, I., Yang, Y. et al. VME: A Satellite Imagery Dataset and Benchmark for Detecting Vehicles in the Middle East and Beyond.

Sci Data 12, 500 (2025). https://doi.org/10.1038/s41597-025-04567-yDownload citationReceived: 09 October 2024Accepted: 30 January 2025Published: 25 March 2025DOI: https://doi.org/10.1038/s41597-025-04567-yShare this articleAnyone you share the following link with will be able to read this content:Get shareable linkSorry, a shareable link is not currently available for this article.Copy to clipboard

Provided by the Springer Nature SharedIt content-sharing initiative

Read full news in source page