f | { | f | { |
| "Author": [ | | "Author": [ |
| { | | { |
| "affiliation": "Centre for Earth Observation Science - | | "affiliation": "Centre for Earth Observation Science - |
| University of Manitoba", | | University of Manitoba", |
| "creatorName": "Harasyn, Madison", | | "creatorName": "Harasyn, Madison", |
| "email": "Madison.harasyn@umanitoba.ca", | | "email": "Madison.harasyn@umanitoba.ca", |
| "nameIdentifier": "https://orcid.org/0000-0002-5741-6766", | | "nameIdentifier": "https://orcid.org/0000-0002-5741-6766", |
| "nameIdentifierScheme": "ORCID", | | "nameIdentifierScheme": "ORCID", |
| "nameType": "Personal", | | "nameType": "Personal", |
| "schemeURI": "http://orcid.org/" | | "schemeURI": "http://orcid.org/" |
| }, | | }, |
| { | | { |
| "affiliation": "Centre for Earth Observation Science - | | "affiliation": "Centre for Earth Observation Science - |
| University of Manitoba", | | University of Manitoba", |
| "creatorName": "Chan, Wayne", | | "creatorName": "Chan, Wayne", |
| "email": "wayne.chan@umanitoba.ca", | | "email": "wayne.chan@umanitoba.ca", |
| "nameIdentifier": "", | | "nameIdentifier": "", |
| "nameType": "Personal" | | "nameType": "Personal" |
| }, | | }, |
| { | | { |
| "affiliation": "Centre for Earth Observation Science - | | "affiliation": "Centre for Earth Observation Science - |
| University of Manitoba", | | University of Manitoba", |
| "creatorName": "Ausen, Emma", | | "creatorName": "Ausen, Emma", |
| "email": "emma.ausen@umanitoba.ca", | | "email": "emma.ausen@umanitoba.ca", |
| "nameIdentifier": "", | | "nameIdentifier": "", |
| "nameType": "Personal" | | "nameType": "Personal" |
| }, | | }, |
| { | | { |
| "affiliation": "Centre for Earth Observation Science - | | "affiliation": "Centre for Earth Observation Science - |
| University of Manitoba", | | University of Manitoba", |
| "creatorName": "Barber, David", | | "creatorName": "Barber, David", |
| "email": "david.barber@umanitoba.ca", | | "email": "david.barber@umanitoba.ca", |
| "nameIdentifier": "0000-0001-9466-3291", | | "nameIdentifier": "0000-0001-9466-3291", |
| "nameIdentifierScheme": "ORCID", | | "nameIdentifierScheme": "ORCID", |
| "nameType": "Personal", | | "nameType": "Personal", |
| "schemeURI": "http://orcid.org/" | | "schemeURI": "http://orcid.org/" |
| } | | } |
| ], | | ], |
| "Identifier": "10.1139/juvs-2021-0024", | | "Identifier": "10.1139/juvs-2021-0024", |
| "PublicationYear": "2022", | | "PublicationYear": "2022", |
| "Publisher": "Drone Systems and Applications", | | "Publisher": "Drone Systems and Applications", |
| "ResourceType": "journal article", | | "ResourceType": "journal article", |
| "Rights": "Creative Commons Attribution 4.0 International", | | "Rights": "Creative Commons Attribution 4.0 International", |
| "Version": "1.0", | | "Version": "1.0", |
| "author": null, | | "author": null, |
| "author_email": null, | | "author_email": null, |
| "citation": "Madison L.Harasyn, Wayne S.Chan, Emma L.Ausen, and | | "citation": "Madison L.Harasyn, Wayne S.Chan, Emma L.Ausen, and |
| David G.Barber. Detection and tracking of belugas, kayaks and | | David G.Barber. Detection and tracking of belugas, kayaks and |
| motorized boats in drone video using deep learning. Drone Systems and | | motorized boats in drone video using deep learning. Drone Systems and |
| Applications. 10(1): 77-96. https://doi.org/10.1139/juvs-2021-0024", | | Applications. 10(1): 77-96. https://doi.org/10.1139/juvs-2021-0024", |
| "creator_user_id": "cde7b848-a882-4fc7-97c9-670417bd6b43", | | "creator_user_id": "cde7b848-a882-4fc7-97c9-670417bd6b43", |
| "descriptionType": "Abstract", | | "descriptionType": "Abstract", |
| "groups": [ | | "groups": [ |
| { | | { |
n | "description": "Inland water features, drainage systems and | n | "description": "Features and characteristics of salt water |
| their characteristics. Examples of data you can find here include | | bodies.\r\n\r\nIn CEOS, related research themes include |
| river and lake data, water quality data. \r\n\r\nIn CEOS, related | | biogeochemistry, modelling, marine mammals, oil spill response, |
| research themes include biogeochemistry, Inland lakes and waters, | | |
| modelling, remote sensing and technology, trace metals and | | physical oceanography, remote sensing and technology and trace metals |
| contaminants.", | | and contaminants", |
| "display_name": "Freshwater", | | "display_name": "Marine", |
| "id": "8f8cd877-b037-4b1a-b928-f86d9e093741", | | "id": "98238b1c-5be8-41ad-8c6e-74cdc4f5f369", |
| "image_display_url": | | "image_display_url": |
n | /data/uploads/group/2021-10-31-211937.658599hyinspirehydrography.svg", | n | ata/uploads/group/2021-10-31-211516.365746ofinspireoceanographic.svg", |
| "name": "freshwater", | | "name": "marine", |
| "title": "Freshwater" | | "title": "Marine" |
| | | }, |
| | | { |
| | | "description": "Image: \"Earth from Space\" by NASA Goddard |
| | | Photo and Video is licensed under CC BY 2.0", |
| | | "display_name": "Remote Sensing", |
| | | "id": "3ec49cbb-4da6-4fe8-8d54-5b6ce03b49d9", |
| | | "image_display_url": |
| | | anitoba.ca/data/uploads/group/2022-02-05-222621.346712earthimage.jpg", |
| | | "name": "remote-sensing", |
| | | "title": "Remote Sensing" |
| } | | } |
| ], | | ], |
| "id": "54b0d7a1-8536-4d40-b1bb-daad81805f43", | | "id": "54b0d7a1-8536-4d40-b1bb-daad81805f43", |
| "isopen": false, | | "isopen": false, |
| "language": "English", | | "language": "English", |
| "licenceType": "Open", | | "licenceType": "Open", |
| "license_id": null, | | "license_id": null, |
| "license_title": null, | | "license_title": null, |
| "maintainer": null, | | "maintainer": null, |
| "maintainer_email": null, | | "maintainer_email": null, |
| "metadata_created": "2022-04-07T19:45:13.021227", | | "metadata_created": "2022-04-07T19:45:13.021227", |
n | "metadata_modified": "2023-06-06T20:09:06.818440", | n | "metadata_modified": "2023-06-06T20:09:50.526039", |
| "name": "detect-video-deep-learning", | | "name": "detect-video-deep-learning", |
| "notes": "Aerial imagery surveys are commonly used in marine mammal | | "notes": "Aerial imagery surveys are commonly used in marine mammal |
| research to determine population size, distribution and habitat use. | | research to determine population size, distribution and habitat use. |
| Analysis of aerial photos involves hours of manually identifying | | Analysis of aerial photos involves hours of manually identifying |
| individuals present in each image and converting raw counts into | | individuals present in each image and converting raw counts into |
| useable biological statistics. Our research proposes the use of deep | | useable biological statistics. Our research proposes the use of deep |
| learning algorithms to increase the efficiency of the marine mammal | | learning algorithms to increase the efficiency of the marine mammal |
| research workflow. To test the feasibility of this proposal, the | | research workflow. To test the feasibility of this proposal, the |
| existing YOLOv4 convolutional neural network model was trained to | | existing YOLOv4 convolutional neural network model was trained to |
| detect belugas, kayaks and motorized boats in oblique drone imagery, | | detect belugas, kayaks and motorized boats in oblique drone imagery, |
| collected from a stationary tethered system. Automated computer-based | | collected from a stationary tethered system. Automated computer-based |
| object detection achieved the following precision and recall, | | object detection achieved the following precision and recall, |
| respectively, for each class: beluga = 74%/72%; boat = 97%/99%; and | | respectively, for each class: beluga = 74%/72%; boat = 97%/99%; and |
| kayak = 96%/96%. We then tested the performance of computer vision | | kayak = 96%/96%. We then tested the performance of computer vision |
| tracking of belugas and occupied watercraft in drone videos using the | | tracking of belugas and occupied watercraft in drone videos using the |
| DeepSORT tracking algorithm, which achieved a multiple-object tracking | | DeepSORT tracking algorithm, which achieved a multiple-object tracking |
| accuracy (MOTA) ranging from 37% to 88% and multiple object tracking | | accuracy (MOTA) ranging from 37% to 88% and multiple object tracking |
| precision (MOTP) between 63% and 86%. Results from this research | | precision (MOTP) between 63% and 86%. Results from this research |
| indicate that deep learning technology can detect and track features | | indicate that deep learning technology can detect and track features |
| more consistently than human annotators, allowing for larger datasets | | more consistently than human annotators, allowing for larger datasets |
| to be processed within a fraction of the time while avoiding | | to be processed within a fraction of the time while avoiding |
| discrepancies introduced by labeling fatigue or multiple human | | discrepancies introduced by labeling fatigue or multiple human |
t | annotators.\r\n\r\nR\u00e9sum\u00e9 Les relev\u00e9s par imagerie | t | annotators.\r\n\r\n**R\u00e9sum\u00e9** Les relev\u00e9s par imagerie |
| a\u00e9rienne sont couramment utilis\u00e9s dans la recherche sur les | | a\u00e9rienne sont couramment utilis\u00e9s dans la recherche sur les |
| mammif\u00e8res marins pour d\u00e9terminer la taille de la | | mammif\u00e8res marins pour d\u00e9terminer la taille de la |
| population, sa r\u00e9partition et l\u2019utilisation de | | population, sa r\u00e9partition et l\u2019utilisation de |
| l\u2019habitat. L\u2019analyse des photos a\u00e9riennes implique des | | l\u2019habitat. L\u2019analyse des photos a\u00e9riennes implique des |
| heures d\u2019identification manuelle des individus pr\u00e9sents dans | | heures d\u2019identification manuelle des individus pr\u00e9sents dans |
| chaque image et la conversion des chiffres bruts en statistiques | | chaque image et la conversion des chiffres bruts en statistiques |
| biologiques utilisables. Notre recherche propose l\u2019utilisation | | biologiques utilisables. Notre recherche propose l\u2019utilisation |
| d\u2019algorithmes d\u2019apprentissage en profondeur pour augmenter | | d\u2019algorithmes d\u2019apprentissage en profondeur pour augmenter |
| l\u2019efficacit\u00e9 du flux de recherche sur les mammif\u00e8res | | l\u2019efficacit\u00e9 du flux de recherche sur les mammif\u00e8res |
| marins. Pour mettre \u00e0 l\u2019essai la faisabilit\u00e9 de cette | | marins. Pour mettre \u00e0 l\u2019essai la faisabilit\u00e9 de cette |
| proposition, le mod\u00e8le de r\u00e9seau de neurones \u00e0 | | proposition, le mod\u00e8le de r\u00e9seau de neurones \u00e0 |
| convolution YOLOv4 existant a \u00e9t\u00e9 entra\u00een\u00e9 pour | | convolution YOLOv4 existant a \u00e9t\u00e9 entra\u00een\u00e9 pour |
| d\u00e9tecter les b\u00e9lugas, les kayaks et les embarcations | | d\u00e9tecter les b\u00e9lugas, les kayaks et les embarcations |
| motoris\u00e9es dans des images de drones obliques, recueillies \u00e0 | | motoris\u00e9es dans des images de drones obliques, recueillies \u00e0 |
| partir d\u2019un syst\u00e8me fixe reli\u00e9. La d\u00e9tection | | partir d\u2019un syst\u00e8me fixe reli\u00e9. La d\u00e9tection |
| automatis\u00e9e d\u2019objets par ordinateur a atteint la | | automatis\u00e9e d\u2019objets par ordinateur a atteint la |
| pr\u00e9cision et le rappel suivants, respectivement, pour chaque | | pr\u00e9cision et le rappel suivants, respectivement, pour chaque |
| classe : b\u00e9luga : 74 %/72 %; bateau : 97 %/99 %; kayak : 96 %/96 | | classe : b\u00e9luga : 74 %/72 %; bateau : 97 %/99 %; kayak : 96 %/96 |
| %. Les auteurs ont ensuite test\u00e9 la performance de poursuite au | | %. Les auteurs ont ensuite test\u00e9 la performance de poursuite au |
| moyen de la vision par ordinateur des b\u00e9lugas et des motomarines | | moyen de la vision par ordinateur des b\u00e9lugas et des motomarines |
| dans des vid\u00e9os de drones \u00e0 l\u2019aide de l\u2019algorithme | | dans des vid\u00e9os de drones \u00e0 l\u2019aide de l\u2019algorithme |
| de poursuite DeepSORT, qui a obtenu une exactitude de poursuite des | | de poursuite DeepSORT, qui a obtenu une exactitude de poursuite des |
| objets multiples (\u00ab MOTA \u00bb) allant de 37 \u00e0 88 % et une | | objets multiples (\u00ab MOTA \u00bb) allant de 37 \u00e0 88 % et une |
| pr\u00e9cision de poursuite des objets multiples (\u00ab MOTP \u00bb) | | pr\u00e9cision de poursuite des objets multiples (\u00ab MOTP \u00bb) |
| allant de 63 \u00e0 86 %. Les r\u00e9sultats de cette recherche | | allant de 63 \u00e0 86 %. Les r\u00e9sultats de cette recherche |
| indiquent que la technologie d\u2019apprentissage profond peut | | indiquent que la technologie d\u2019apprentissage profond peut |
| d\u00e9tecter et suivre les caract\u00e9ristiques plus | | d\u00e9tecter et suivre les caract\u00e9ristiques plus |
| r\u00e9guli\u00e8rement que les annotateurs humains, permettant de | | r\u00e9guli\u00e8rement que les annotateurs humains, permettant de |
| traiter des ensembles de donn\u00e9es plus volumineux en une fraction | | traiter des ensembles de donn\u00e9es plus volumineux en une fraction |
| de temps tout en \u00e9vitant les \u00e9carts introduits par la | | de temps tout en \u00e9vitant les \u00e9carts introduits par la |
| fatigue d\u2019\u00e9tiquetage ou de multiples annotateurs humains. | | fatigue d\u2019\u00e9tiquetage ou de multiples annotateurs humains. |
| [Traduit par la R\u00e9daction]", | | [Traduit par la R\u00e9daction]", |
| "num_resources": 2, | | "num_resources": 2, |
| "num_tags": 6, | | "num_tags": 6, |
| "organization": { | | "organization": { |
| "approval_status": "approved", | | "approval_status": "approved", |
| "created": "2017-07-21T13:15:49.935872", | | "created": "2017-07-21T13:15:49.935872", |
| "description": "The Centre for Earth Observation Science (CEOS) | | "description": "The Centre for Earth Observation Science (CEOS) |
| was established in 1994 with a mandate to research, preserve and | | was established in 1994 with a mandate to research, preserve and |
| communicate knowledge of Earth system processes using the technologies | | communicate knowledge of Earth system processes using the technologies |
| of Earth Observation Science. Research is multidisciplinary and | | of Earth Observation Science. Research is multidisciplinary and |
| collaborative seeking to understand the complex interrelationships | | collaborative seeking to understand the complex interrelationships |
| between elements of Earth systems, and how these systems will likely | | between elements of Earth systems, and how these systems will likely |
| respond to climate change. Although researchers have worked in many | | respond to climate change. Although researchers have worked in many |
| regions, the Arctic marine system has always been a unifying focus of | | regions, the Arctic marine system has always been a unifying focus of |
| activity.\r\n\r\nIn 2012, CEOS, along with the Greenland Climate | | activity.\r\n\r\nIn 2012, CEOS, along with the Greenland Climate |
| Research Centre (GCRC, Nuuk, Greenland) and the Arctic Research Centre | | Research Centre (GCRC, Nuuk, Greenland) and the Arctic Research Centre |
| (ARC, Aarhus, Denmark) established the Arctic Science Partnership, | | (ARC, Aarhus, Denmark) established the Arctic Science Partnership, |
| thereby integrating academic and research initiatives.\r\n\r\nAreas of | | thereby integrating academic and research initiatives.\r\n\r\nAreas of |
| existing research activity are divided among key themes:\r\n\r\nArctic | | existing research activity are divided among key themes:\r\n\r\nArctic |
| Anthropology/Paleoclimatology: LiDAR scanning and digital site | | Anthropology/Paleoclimatology: LiDAR scanning and digital site |
| preservation, archaeo-geophysics, permafrost degredation, lithic | | preservation, archaeo-geophysics, permafrost degredation, lithic |
| morphometrics, zooarchaeology, proxy studies, paleodistribution of sea | | morphometrics, zooarchaeology, proxy studies, paleodistribution of sea |
| ice, landscape learning, Paleo-Eskimo culture, Thule Inuit culture, | | ice, landscape learning, Paleo-Eskimo culture, Thule Inuit culture, |
| ethnographic analogy, traditional knowledge, climate change and | | ethnographic analogy, traditional knowledge, climate change and |
| northern heritage resource management.\r\n\r\nAtmospheric | | northern heritage resource management.\r\n\r\nAtmospheric |
| Studies/Meteorology: Boundary layer, precipitation, clouds, storms and | | Studies/Meteorology: Boundary layer, precipitation, clouds, storms and |
| extreme weather, circulation, eddy correlations, polar vortex, | | extreme weather, circulation, eddy correlations, polar vortex, |
| climate, teleconnections, geophysical fluid dynamics, flux and energy | | climate, teleconnections, geophysical fluid dynamics, flux and energy |
| budgets, ocean-sea ice-atmosphere interface, radiative transfer, ice | | budgets, ocean-sea ice-atmosphere interface, radiative transfer, ice |
| albedo feedback, cloud radiative forcing, pCO2. | | albedo feedback, cloud radiative forcing, pCO2. |
| \r\n\r\nBiogeochemistry: Organic carbon, greenhouse gases, bubbles, | | \r\n\r\nBiogeochemistry: Organic carbon, greenhouse gases, bubbles, |
| Ikaite, carbonate chemistry, CO2 fluxes, mercury and other trace | | Ikaite, carbonate chemistry, CO2 fluxes, mercury and other trace |
| metals, minerals, hydrocarbons, brine processes, otolith | | metals, minerals, hydrocarbons, brine processes, otolith |
| microchemistry, sediments, biomarkers. \r\n\r\nContaminants: Mercury, | | microchemistry, sediments, biomarkers. \r\n\r\nContaminants: Mercury, |
| trace metals, PAHs, source, transport, transformation, pathways, | | trace metals, PAHs, source, transport, transformation, pathways, |
| bioaccumulations, marine ecosystems, marine chemistry. \r\nEarth | | bioaccumulations, marine ecosystems, marine chemistry. \r\nEarth |
| Observation Science: Active and passive microwave, LiDAR, EM | | Observation Science: Active and passive microwave, LiDAR, EM |
| induction, spatial-temporal analysis, forward and inverse scattering | | induction, spatial-temporal analysis, forward and inverse scattering |
| models, complex permittivity, ocean colour, ocean surface roughness, | | models, complex permittivity, ocean colour, ocean surface roughness, |
| NIR, TIR, satellite telemetry, GPS. Ice-Associated Biology: | | NIR, TIR, satellite telemetry, GPS. Ice-Associated Biology: |
| Biophysical processes, primary production; ice algae, ice | | Biophysical processes, primary production; ice algae, ice |
| microbiology, bio-optics, under-ice phytoplankton. \r\n\r\nInland | | microbiology, bio-optics, under-ice phytoplankton. \r\n\r\nInland |
| Lakes and Waters: Hydrologic connectivity, watershed systems, sediment | | Lakes and Waters: Hydrologic connectivity, watershed systems, sediment |
| transport, nutrient transport, contaminants, landscape processes, | | transport, nutrient transport, contaminants, landscape processes, |
| remote sensing, freshwater-marine coupling. Marine Mammals: Seals, | | remote sensing, freshwater-marine coupling. Marine Mammals: Seals, |
| whales, habitat, conservation, satellite telemetry, distribution, | | whales, habitat, conservation, satellite telemetry, distribution, |
| population studies, prey behaviour, bioacoustics.\r\n\r\nModelling: | | population studies, prey behaviour, bioacoustics.\r\n\r\nModelling: |
| Simulation of sea ice and oceanic regional processes, Nucleus for | | Simulation of sea ice and oceanic regional processes, Nucleus for |
| European Modelling of the Ocean (NEMO), ice-ocean modelling and | | European Modelling of the Ocean (NEMO), ice-ocean modelling and |
| interactions, hind cast simulations and projections for sea ice state | | interactions, hind cast simulations and projections for sea ice state |
| and ocean variables based on CMIP5 scenarios and MIROC5 forcing, | | and ocean variables based on CMIP5 scenarios and MIROC5 forcing, |
| validation.\r\n\r\nOceanography: Circulation, temperature, in-flow and | | validation.\r\n\r\nOceanography: Circulation, temperature, in-flow and |
| out-flow shelves, water dynamics, microturbulence, Beaufort Gyre, eddy | | out-flow shelves, water dynamics, microturbulence, Beaufort Gyre, eddy |
| correlations.\r\n\r\nSea Ice Geophysics:Thermodynamic and dynamic | | correlations.\r\n\r\nSea Ice Geophysics:Thermodynamic and dynamic |
| processes, extreme ice features and hazards, snow, ridges, | | processes, extreme ice features and hazards, snow, ridges, |
| polynyas.\r\n\r\nTraditional and Local Knowledge: Indigenous cultures, | | polynyas.\r\n\r\nTraditional and Local Knowledge: Indigenous cultures, |
| Inuit, Inuvialuit, oral history, toponomy, mobility and settlement, | | Inuit, Inuvialuit, oral history, toponomy, mobility and settlement, |
| hunting, food security, sea ice use, community-based research, | | hunting, food security, sea ice use, community-based research, |
| community-based monitoring, two ways of knowing.", | | community-based monitoring, two ways of knowing.", |
| "id": "9e21f6b6-d13f-4ba2-a379-fd962f507071", | | "id": "9e21f6b6-d13f-4ba2-a379-fd962f507071", |
| "image_url": "2021-11-13-003953.952874UMLogoHORZ.jpg", | | "image_url": "2021-11-13-003953.952874UMLogoHORZ.jpg", |
| "is_organization": true, | | "is_organization": true, |
| "name": "ceos", | | "name": "ceos", |
| "state": "active", | | "state": "active", |
| "title": "Centre for Earth Observation Science", | | "title": "Centre for Earth Observation Science", |
| "type": "organization" | | "type": "organization" |
| }, | | }, |
| "owner_org": "9e21f6b6-d13f-4ba2-a379-fd962f507071", | | "owner_org": "9e21f6b6-d13f-4ba2-a379-fd962f507071", |
| "private": false, | | "private": false, |
| "related_datasets": [ | | "related_datasets": [ |
| "b5f259b4-3ace-4750-bfb0-47c4e794082f" | | "b5f259b4-3ace-4750-bfb0-47c4e794082f" |
| ], | | ], |
| "related_programs": [], | | "related_programs": [], |
| "relationships_as_object": [], | | "relationships_as_object": [], |
| "relationships_as_subject": [], | | "relationships_as_subject": [], |
| "resources": [ | | "resources": [ |
| { | | { |
| "cache_last_updated": null, | | "cache_last_updated": null, |
| "cache_url": null, | | "cache_url": null, |
| "created": "2022-04-07T19:49:13.974750", | | "created": "2022-04-07T19:49:13.974750", |
| "datastore_active": false, | | "datastore_active": false, |
| "datastore_contains_all_records_of_source_file": false, | | "datastore_contains_all_records_of_source_file": false, |
| "description": "Churchill Beluga Boat Drone Imagery related | | "description": "Churchill Beluga Boat Drone Imagery related |
| journal article published in Drone Systems and Applications.\r\nDOI: | | journal article published in Drone Systems and Applications.\r\nDOI: |
| https://doi.org/10.1139/juvs-2021-0024", | | https://doi.org/10.1139/juvs-2021-0024", |
| "format": "PDF", | | "format": "PDF", |
| "hash": "", | | "hash": "", |
| "id": "5bcbb0bc-425b-4fad-b7ff-4c8599043dcf", | | "id": "5bcbb0bc-425b-4fad-b7ff-4c8599043dcf", |
| "last_modified": "2022-04-07T20:02:20.594051", | | "last_modified": "2022-04-07T20:02:20.594051", |
| "metadata_modified": "2023-05-18T21:05:56.975275", | | "metadata_modified": "2023-05-18T21:05:56.975275", |
| "mimetype": "application/pdf", | | "mimetype": "application/pdf", |
| "mimetype_inner": null, | | "mimetype_inner": null, |
| "name": "Detection and tracking of belugas, kayaks and motorized | | "name": "Detection and tracking of belugas, kayaks and motorized |
| boats in drone video using deep learning", | | boats in drone video using deep learning", |
| "package_id": "54b0d7a1-8536-4d40-b1bb-daad81805f43", | | "package_id": "54b0d7a1-8536-4d40-b1bb-daad81805f43", |
| "position": 0, | | "position": 0, |
| "resCategory": "supplemental", | | "resCategory": "supplemental", |
| "resource_type": null, | | "resource_type": null, |
| "size": 4104522, | | "size": 4104522, |
| "state": "active", | | "state": "active", |
| "url": | | "url": |
| rce/5bcbb0bc-425b-4fad-b7ff-4c8599043dcf/download/juvs-2021-0024.pdf", | | rce/5bcbb0bc-425b-4fad-b7ff-4c8599043dcf/download/juvs-2021-0024.pdf", |
| "url_type": "upload" | | "url_type": "upload" |
| }, | | }, |
| { | | { |
| "cache_last_updated": null, | | "cache_last_updated": null, |
| "cache_url": null, | | "cache_url": null, |
| "created": "2023-05-18T21:05:57.006643", | | "created": "2023-05-18T21:05:57.006643", |
| "datastore_active": false, | | "datastore_active": false, |
| "datastore_contains_all_records_of_source_file": false, | | "datastore_contains_all_records_of_source_file": false, |
| "description": "Researchers at CEOS are often asked to write a | | "description": "Researchers at CEOS are often asked to write a |
| field story about their work, to make their research more accessible. | | field story about their work, to make their research more accessible. |
| We decided to do something a little different for our work on applying | | We decided to do something a little different for our work on applying |
| machine learning to detecting and tracking beluga whales: we are | | machine learning to detecting and tracking beluga whales: we are |
| presenting it as a comic-book style video!", | | presenting it as a comic-book style video!", |
| "format": "", | | "format": "", |
| "hash": "", | | "hash": "", |
| "id": "1cd6dbae-5c9d-440d-b29b-26c84fbc5a7c", | | "id": "1cd6dbae-5c9d-440d-b29b-26c84fbc5a7c", |
| "last_modified": null, | | "last_modified": null, |
| "metadata_modified": "2023-05-18T21:05:56.975546", | | "metadata_modified": "2023-05-18T21:05:56.975546", |
| "mimetype": null, | | "mimetype": null, |
| "mimetype_inner": null, | | "mimetype_inner": null, |
| "name": "One Beluga, Two Beluga, Three Beluga, Four: How to | | "name": "One Beluga, Two Beluga, Three Beluga, Four: How to |
| Count Belugas When You Run Out of Fingers and Toes", | | Count Belugas When You Run Out of Fingers and Toes", |
| "package_id": "54b0d7a1-8536-4d40-b1bb-daad81805f43", | | "package_id": "54b0d7a1-8536-4d40-b1bb-daad81805f43", |
| "position": 1, | | "position": 1, |
| "resCategory": "supplemental", | | "resCategory": "supplemental", |
| "resource_type": null, | | "resource_type": null, |
| "size": null, | | "size": null, |
| "state": "active", | | "state": "active", |
| "url": | | "url": |
| n/beluga-graphic-novel/resource/58aed159-4a62-4c2b-9978-967ad5f356a6", | | n/beluga-graphic-novel/resource/58aed159-4a62-4c2b-9978-967ad5f356a6", |
| "url_type": null | | "url_type": null |
| } | | } |
| ], | | ], |
| "rightsIdentifier": "CC-BY-4.0", | | "rightsIdentifier": "CC-BY-4.0", |
| "rightsIdentifierScheme": "SPDX", | | "rightsIdentifierScheme": "SPDX", |
| "rightsSchemeURI": "https://spdx.org/licenses", | | "rightsSchemeURI": "https://spdx.org/licenses", |
| "rightsURI": "https://spdx.org/licenses/CC-BY-4.0.html", | | "rightsURI": "https://spdx.org/licenses/CC-BY-4.0.html", |
| "schemeURI": "", | | "schemeURI": "", |
| "state": "active", | | "state": "active", |
| "subjectScheme": "", | | "subjectScheme": "", |
| "tags": [ | | "tags": [ |
| { | | { |
| "display_name": "Beluga", | | "display_name": "Beluga", |
| "id": "a9f25a89-b0ef-4d4d-993d-73f28e0d702a", | | "id": "a9f25a89-b0ef-4d4d-993d-73f28e0d702a", |
| "name": "Beluga", | | "name": "Beluga", |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| }, | | }, |
| { | | { |
| "display_name": "Unmanned Aerial Vehicle", | | "display_name": "Unmanned Aerial Vehicle", |
| "id": "a6dc9001-e6da-4a84-bfec-2941d3ebce78", | | "id": "a6dc9001-e6da-4a84-bfec-2941d3ebce78", |
| "name": "Unmanned Aerial Vehicle", | | "name": "Unmanned Aerial Vehicle", |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| }, | | }, |
| { | | { |
| "display_name": "computer vision", | | "display_name": "computer vision", |
| "id": "d7270905-c420-4d19-aa9c-c6f818ab5b67", | | "id": "d7270905-c420-4d19-aa9c-c6f818ab5b67", |
| "name": "computer vision", | | "name": "computer vision", |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| }, | | }, |
| { | | { |
| "display_name": "deep learning", | | "display_name": "deep learning", |
| "id": "87526358-2d8a-4c78-8375-38c132b53d5a", | | "id": "87526358-2d8a-4c78-8375-38c132b53d5a", |
| "name": "deep learning", | | "name": "deep learning", |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| }, | | }, |
| { | | { |
| "display_name": "object detection", | | "display_name": "object detection", |
| "id": "a3d44586-cba5-4685-b07a-2d2f16578353", | | "id": "a3d44586-cba5-4685-b07a-2d2f16578353", |
| "name": "object detection", | | "name": "object detection", |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| }, | | }, |
| { | | { |
| "display_name": "object tracking", | | "display_name": "object tracking", |
| "id": "28ce0864-2ed8-43d1-b80d-d79684cac63f", | | "id": "28ce0864-2ed8-43d1-b80d-d79684cac63f", |
| "name": "object tracking", | | "name": "object tracking", |
| "state": "active", | | "state": "active", |
| "vocabulary_id": null | | "vocabulary_id": null |
| } | | } |
| ], | | ], |
| "theme": [ | | "theme": [ |
| "8f8cd877-b037-4b1a-b928-f86d9e093741", | | "8f8cd877-b037-4b1a-b928-f86d9e093741", |
| "98238b1c-5be8-41ad-8c6e-74cdc4f5f369", | | "98238b1c-5be8-41ad-8c6e-74cdc4f5f369", |
| "3ec49cbb-4da6-4fe8-8d54-5b6ce03b49d9" | | "3ec49cbb-4da6-4fe8-8d54-5b6ce03b49d9" |
| ], | | ], |
| "title": "Detection and tracking of belugas, kayaks and motorized | | "title": "Detection and tracking of belugas, kayaks and motorized |
| boats in drone video using deep learning", | | boats in drone video using deep learning", |
| "type": "publication", | | "type": "publication", |
| "url": null, | | "url": null, |
| "version": null | | "version": null |
| } | | } |