f | { | f | { |
| "access_rights": "", | | "access_rights": "", |
n | "author": "Jianbo Jiao", | n | "author": "Mohamed A. KERKOURI", |
| "author_email": "", | | "author_email": "", |
n | "citation": [ | n | "citation": [], |
| "https://doi.org/10.48550/arXiv.1608.05203", | | |
| "https://doi.org/10.48550/arXiv.1903.02501", | | |
| "https://doi.org/10.48550/arXiv.1804.01793", | | |
| "https://doi.org/10.48550/arXiv.1808.09559" | | |
| ], | | |
| "creator_user_id": "17755db4-395a-4b3b-ac09-e8e3484ca700", | | "creator_user_id": "17755db4-395a-4b3b-ac09-e8e3484ca700", |
n | "defined_in": "https://doi.org/10.1007/978-3-030-58558-7_25", | n | "defined_in": "https://doi.org/10.48550/arXiv.2311.12860", |
| "doi": "10.57702/22u0fhma", | | "doi": "10.57702/22u0fhma", |
| "doi_date_published": "2024-12-16", | | "doi_date_published": "2024-12-16", |
| "doi_publisher": "TIB", | | "doi_publisher": "TIB", |
| "doi_status": true, | | "doi_status": true, |
| "domain": "https://service.tib.eu/ldmservice", | | "domain": "https://service.tib.eu/ldmservice", |
| "extra_authors": [ | | "extra_authors": [ |
| { | | { |
n | "extra_author": "J. Alison Noble", | n | "extra_author": "Marouane TLIBA", |
| | | "orcid": "" |
| | | }, |
| | | { |
| | | "extra_author": "Aladine CHETOUANI", |
| | | "orcid": "" |
| | | }, |
| | | { |
| | | "extra_author": "Rachid HARBA", |
| "orcid": "" | | "orcid": "" |
| } | | } |
| ], | | ], |
| "groups": [ | | "groups": [ |
| { | | { |
| "description": "", | | "description": "", |
n | "display_name": "Computer Vision", | n | "display_name": "Image Classification", |
| "id": "d09caf7c-26c7-4e4d-bb8e-49476a90ba25", | | "id": "18b77292-26aa-4caf-89ed-cbd35fa60474", |
| "image_display_url": "", | | "image_display_url": "", |
n | "name": "computer-vision", | n | "name": "image-classification", |
| "title": "Computer Vision" | | "title": "Image Classification" |
| }, | | }, |
| { | | { |
| "description": "", | | "description": "", |
n | "display_name": "Eye Tracking", | n | "display_name": "Scanpath Prediction", |
| "id": "ace34eff-ad3b-4126-a970-32c267180725", | | "id": "b64bbd5c-89c7-49e0-bf80-11435d0d9dda", |
| "image_display_url": "", | | "image_display_url": "", |
n | "name": "eye-tracking", | n | "name": "scanpath-prediction", |
| "title": "Eye Tracking" | | "title": "Scanpath Prediction" |
| | | }, |
| | | { |
| | | "description": "", |
| | | "display_name": "Visual Attention", |
| | | "id": "403b0e97-dd2b-48f8-9846-7f2e0affb8e4", |
| | | "image_display_url": "", |
| | | "name": "visual-attention", |
| | | "title": "Visual Attention" |
| } | | } |
| ], | | ], |
| "id": "29f3ec57-ed52-41b2-aece-54ebb1dc6563", | | "id": "29f3ec57-ed52-41b2-aece-54ebb1dc6563", |
| "isopen": false, | | "isopen": false, |
n | "landing_page": "https://github.com/rdroste/unisal", | n | "landing_page": |
| | | "https://www.sciencedirect.com/science/article/pii/S2352240615000465", |
| "license_title": null, | | "license_title": null, |
| "link_orkg": "", | | "link_orkg": "", |
| "metadata_created": "2024-12-16T19:02:30.592664", | | "metadata_created": "2024-12-16T19:02:30.592664", |
n | "metadata_modified": "2024-12-16T19:02:30.958527", | n | "metadata_modified": "2024-12-16T19:26:21.690757", |
| "name": "salicon", | | "name": "salicon", |
n | "notes": "SALICON is one of the largest saliency datasets available | n | "notes": "Human vision is naturally more attracted by some regions |
| in the public domain. It consists of eye-fixation information for | | within their field of view than others. This intrinsic selectivity |
| 20,000 images from the MS COCO dataset. These images contain diverse | | mechanism, so-called visual attention, is in\ufb02uenced by both high- |
| indoor and outdoor scenes and display a range of scene clutter.", | | and low-level factors; such as the global environment (illumination, |
| | | background texture, etc.), stimulus characteristics (color, intensity, |
| | | orientation, etc.), and some prior visual information.", |
| "num_resources": 1, | | "num_resources": 0, |
| "num_tags": 12, | | "num_tags": 7, |
| "organization": { | | "organization": { |
| "approval_status": "approved", | | "approval_status": "approved", |
| "created": "2024-11-25T12:11:38.292601", | | "created": "2024-11-25T12:11:38.292601", |
| "description": "", | | "description": "", |
| "id": "079d46db-32df-4b48-91f3-0a8bc8f69559", | | "id": "079d46db-32df-4b48-91f3-0a8bc8f69559", |
| "image_url": "", | | "image_url": "", |
| "is_organization": true, | | "is_organization": true, |
| "name": "no-organization", | | "name": "no-organization", |
| "state": "active", | | "state": "active", |
| "title": "No Organization", | | "title": "No Organization", |
| "type": "organization" | | "type": "organization" |
| }, | | }, |
| "owner_org": "079d46db-32df-4b48-91f3-0a8bc8f69559", | | "owner_org": "079d46db-32df-4b48-91f3-0a8bc8f69559", |
| "private": false, | | "private": false, |
| "relationships_as_object": [], | | "relationships_as_object": [], |
| "relationships_as_subject": [], | | "relationships_as_subject": [], |
n | "resources": [ | n | "resources": [], |
| { | | |
| "cache_last_updated": null, | | |
| "cache_url": null, | | |
| "created": "2024-12-16T18:25:38", | | |
| "data": [ | | |
| "dcterms:title", | | |
| "dcterms:accessRights", | | |
| "dcterms:creator", | | |
| "dcterms:description", | | |
| "dcterms:issued", | | |
| "dcterms:language", | | |
| "dcterms:identifier", | | |
| "dcat:theme", | | |
| "dcterms:type", | | |
| "dcat:keyword", | | |
| "dcat:landingPage", | | |
| "dcterms:hasVersion", | | |
| "dcterms:format", | | |
| "mls:task", | | |
| "datacite:isDescribedBy" | | |
| ], | | |
| "description": "The json representation of the dataset with its | | |
| distributions based on DCAT.", | | |
| "format": "JSON", | | |
| "hash": "", | | |
| "id": "b3b5a42a-4915-4161-8594-d0b786a51a7e", | | |
| "last_modified": "2024-12-16T19:02:30.949032", | | |
| "metadata_modified": "2024-12-16T19:02:30.961338", | | |
| "mimetype": "application/json", | | |
| "mimetype_inner": null, | | |
| "name": "Original Metadata", | | |
| "package_id": "29f3ec57-ed52-41b2-aece-54ebb1dc6563", | | |
| "position": 0, | | |
| "resource_type": null, | | |
| "size": 1249, | | |
| "state": "active", | | |
| "url": | | |
| resource/b3b5a42a-4915-4161-8594-d0b786a51a7e/download/metadata.json", | | |
| "url_type": "upload" | | |
| } | | |
| ], | | |
| "services_used_list": "", | | "services_used_list": "", |
| "state": "active", | | "state": "active", |
| "tags": [ | | "tags": [ |
| { | | { |
n | "display_name": "Image Saliency", | n | "display_name": "Gaze Fixation Density Maps", |
| "id": "004ff2f0-3fb4-4d47-b59c-4f3441f9b0c3", | | "id": "815392bb-68c9-4e96-a1bf-21db57cea222", |
| "name": "Image Saliency", | | "name": "Gaze Fixation Density Maps", |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| }, | | }, |
| { | | { |
n | "display_name": "MS COCO", | n | "display_name": "Image Classification", |
| "id": "b8dcb1ca-8488-4b9b-9c1b-568b6f956b50", | | "id": "418e2ddf-a1d3-42ac-ad05-156f79ca8e22", |
| "name": "MS COCO", | | "name": "Image Classification", |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| }, | | }, |
| { | | { |
n | "display_name": "SALICON", | n | "display_name": "Visual Attention", |
| "id": "497402d9-84ab-4e34-9177-186da45352f5", | | "id": "75d9febf-b2ed-45b2-8b35-90c67689b87b", |
| "name": "SALICON", | | "name": "Visual Attention", |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| }, | | }, |
| { | | { |
n | "display_name": "computer vision", | n | |
| "id": "f650b4e3-9955-49b0-ba7b-2d302a990978", | | |
| "name": "computer vision", | | |
| "state": "active", | | |
| "vocabulary_id": null | | |
| }, | | |
| { | | |
| "display_name": "eye tracking", | | "display_name": "deep learning", |
| "id": "16e5b240-6bcc-4821-905c-0acfcbbe6ea9", | | "id": "19e41883-3799-4184-9e0e-26c95795b119", |
| "name": "eye tracking", | | "name": "deep learning", |
| "state": "active", | | |
| "vocabulary_id": null | | |
| }, | | |
| { | | |
| "display_name": "eye-fixation", | | |
| "id": "be128c37-d897-4d7a-95d7-10e16d782634", | | |
| "name": "eye-fixation", | | |
| "state": "active", | | |
| "vocabulary_id": null | | |
| }, | | |
| { | | |
| "display_name": "gaze data", | | |
| "id": "3e2dbbd3-459d-4560-a880-3dd5231fc66c", | | |
| "name": "gaze data", | | |
| "state": "active", | | |
| "vocabulary_id": null | | |
| }, | | |
| { | | |
| "display_name": "image", | | |
| "id": "2750bf30-5ae4-4ae8-bfef-5a168733376b", | | |
| "name": "image", | | |
| "state": "active", | | |
| "vocabulary_id": null | | |
| }, | | |
| { | | |
| "display_name": "image saliency", | | |
| "id": "4b80a8c7-6da1-41d6-90ef-c5fd3484817d", | | |
| "name": "image saliency", | | |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| }, | | }, |
| { | | { |
| "display_name": "saliency", | | "display_name": "saliency", |
| "id": "e8c29766-f718-4bd8-87ca-f86bd741cba7", | | "id": "e8c29766-f718-4bd8-87ca-f86bd741cba7", |
| "name": "saliency", | | "name": "saliency", |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| }, | | }, |
| { | | { |
n | "display_name": "salient regions", | n | "display_name": "scanpath", |
| "id": "04077406-d8fb-4486-9b7a-95e5d769c72c", | | "id": "daeedfad-28a4-477b-b95b-4cfa815e51d9", |
| "name": "salient regions", | | "name": "scanpath", |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| }, | | }, |
| { | | { |
n | "display_name": "static saliency", | n | "display_name": "visual attention", |
| "id": "0202755f-53a2-4c69-9111-2420a9c53d33", | | "id": "bd5f6475-4fb3-4c8f-8a3b-c51bc58bc8b4", |
| "name": "static saliency", | | "name": "visual attention", |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| } | | } |
| ], | | ], |
t | "title": "SALICON", | t | "title": "Salicon", |
| "type": "dataset", | | "type": "dataset", |
| "version": "" | | "version": "" |
| } | | } |