SlideShare a Scribd company logo
1 of 86
FROM CROWD KNOWLEDGE
     TO MACHINE KNOWLEDGE
      use cases with
      semantics & user
      interaction in cultural
      heritage collections




      Lora Aroyo
VU University Amsterdam
       @laroyo
OBSERVATION I:
         VIDEO ANNOTATIONS

 Video: rich in meaning
 expressed with objects,
people, events & symbols
Annotation: tedious, time-
 consuming, incomplete

Professional annotations:
coarse-grained & refer to
 entire video and topics
OBSERVATION II:
            SEARCH BEHAVIOR

  People predominantly
request video fragments:
broadcast (33%), stories
 (17%), fragments (49%)

  Finding fragments takes
        much longer
  than finding broadcasts:
stories (2x), fragment (3x)
OBSERVATION III:
    VOCABULARY MISMATCH

experts use a specific
  vocabulary that is
 unknown to general
     audiences

35% of clicked results
are not found by title
  or thesaurus term
OBSERVATION IV:
      LIMITED ANNOTATION
           RESOURCES

the professionals can
no longer cope with
   the demand for
manual annotation on
  the ever growing
multimedia content in
   their collections
SO WE WANT TO ...
Make a massive set of videos accessible to end users

        Improve video search for end users

  Maintain a growing community of engaged users


          Support professional annotators
TO IMPROVE VIDEO SEARCH:

       fragment retrieval
     within video navigation
REQUIRES CHANGES IN
        ANNOTATIONS:

    Including time-based annotations

Bridging the vocabulary gap of searcher &.
                cataloguer
exploit crowdsourcing on the web
integrate in content management system
CULTURAL HERITAGE WITH
  INVOLVED CONSUMERS &
        PROVIDERS


data: open, shared & accessible
infrastructures: interoperable
smart services: relevant content
in right context
Rijksmuseum
        Amsterdam
http://chip-project.org


http://e-culture.multimedian.nl/
pk/annotate?uri=RP-P-
OB-77.320
Netherlands Institute
   for Sound & Vision

http://openimages.eu
http://academia.nl


MUNCH project
CHOICE project
Europeana
        Cultural Search
http://e-culture.multimedian.nl
http://europeana.eu/portal/
thoughtlab.html
http://europeana.eu
WAYS TO USE THE CROWD

    Tagging & classification
    Editing & transcribing
       Contextualising
         Acquisition
         Co-curation
        Crowdfunding
WHAT WE DID:
             THE WORKFLOW
collect-analyse-manage-integrate user-generated metadata
domain
         experts
gamers




                   collection
                   managers
Winner EuroITV Competition
Best Archives on the Web Award




  COLLECT
User Generated Metadata
TWO PILOTS
           Waisda? 1.0 (8 months)
focus group with lay users for motivation aspects
    cognitive walk through with professionals
 usability testing with lay users and professionals
          tag analysis - primarily WHAT & WHO



Waisda? 2.0 (starting this September)
              currently running
         initial user & tag analysis
Waisda? 1.0
(8 months)
Participation
 44,000 pageviews
2,000 different players
500 registered players
 thousands anonymous players

Time-based annotations
 612 videos tagged
 420,000 tags added
 46,000 unique tags

Vocabulary gap
User tags added the user perspective on the collection
                Community consensus as a measure of ‘validity’
     Players score points when tag exactly matches a tag entered by another within 10 secs
Waisda? 2.0
(4 weeks
not public)

Participation
 ~1500 users
 81 registered
 1435 anonymous
users
 2344 games




Time-based annotations                                   11,109 videos in the game
 32,200 tags --> 25,600 first time tags                  322 were played with
 19% are one time tags (81% appear more than once)
 12,000 validated tags (36%) / (4000 of them first time tag)
 1,900 match in vocabs - 257 GTAA people/83 validated; 1,661 GTAA geo/666 validated
 9,796 validated, but no match in GTAA
MANAGE
Analyze Waisda? Collected Tags; Quality Metrics
WHAT ARE THE RELATIONS
BETWEEN TAGS & PROFESSIONAL
      ANNOTATIONS?
           in terms of the vocabulary used?
         in terms of the topics they describe?


               Initially: manual analysis
   Output goal: to be expressed in a quality metric
WAISDA? TAG ANALYSIS

                       Study 1
Quantitative analysis of all 46,762 Waisda? tags
       Tag coverage w.r.t different vocabularies


                      Study 2
  Qualitative analysis of selected video fragments
 Manual classification of 1,343 “verified” tags
DURING FIRST 9 MONTHS
                        400000



                        300000                                                                                                                                                      > 600 videos
                                                                                                                                                                                    13 TV series
Number
of
tags




                        200000



                        100000                                                                                                                                                > 420,000 tags collected
                                  0
                                                                                                                                                                                46,792 unique tags
                                      1
                                           12
                                                23
                                                         34
                                                              45
                                                                   56
                                                                        67
                                                                             78
                                                                                  89
                                                                                        100
                                                                                              111
                                                                                                     122
                                                                                                           133
                                                                                                                     144
                                                                                                                           155
                                                                                                                                 166
                                                                                                                                       177
                                                                                                                                             188
                                                                                                                                                    199
                                                                                                                                                          210
                                                                                                                                                                221
                                                                                                                                                                       232
                            500
                                                                                                   Day
                                                                                                                                                                               each top 5 most-tagged
                                                                                                                                                                               video has > 23,000 tags
                            375
       Number
of
arrivals




                            250                                                                                                                                              avg. tag density > 8 tags/sec
                            125



                             0
                                      1
                                          11
                                               21
                                                    31
                                                         41
                                                              51
                                                                   61
                                                                        71
                                                                             81
                                                                                  91
                                                                                       101
                                                                                             111
                                                                                                   121
                                                                                                         131
                                                                                                               141
                                                                                                                     151
                                                                                                                           161
                                                                                                                                 171
                                                                                                                                       181
                                                                                                                                             191
                                                                                                                                                   201
                                                                                                                                                         211
                                                                                                                                                               221
                                                                                                                                                                     231




                                                                                               Day
1
    MAPPING TAGS TO
     VOCABULARIES
                                      Vocabulary of S&V
                          GTAA      160 000 terms in six
                         Thesaurus disjoint facets: keyword,
                    to

           ap
             pe
               d                    location, maker, genre,
       m
                                   person, named entities

       mapped                                Dutch lexical
                                              database
       m
           ap
             pe
               d
                   to
                                          collective
                                       vocabulary of
                                       internet users
RESULTS
Zero-hits
 (4404)
  11%




                      Google
                      (41388)
                        89%

 Corneto
                                (29815)
  (8723)
                                  63%
   18%
            (2216)
              5%
                     GTAA
                     (1634)
                       3%
RESULTS
Zero-hits
 (4404)
  11%

                                          Facet     Tags
                                         Subject    1199
                     Google
                     (41388)             Location   613
                       89%
                                          Genre      52
 Corneto
  (8723)
                               (29815)   Person     118
                                 63%
   18%
            (2216)
                                          Maker      4
              5%GTAA
                    GTAA                  Name      673
                (3850)
                    (1634)
                   8%3%
RESULTS
Zero-hits
 (4404)
  11%




                      Google
                      (41388)
                        89%

 Corneto
                                (29815)
  (8723)
                                  63%
   18%
            (2216)
              5%
                     GTAA
                     (1634)
                       3%
RESULTS
Zero-hits
 (4404)
  11%

                                       Types      Tags
                                       Noun       7222
                  Google
                  (41388)
                    89%                 Verb      2090

 Corneto                              Adjective   1693
  Corneto
  (8723)
                            (29815)
   (10939)                    63%
   18%
     23%(2216)
                                      Adverb      171
          5%
                 GTAA
                 (1634)
                   3%
RESULTS
Zero-hits
 (4404)
  11%




                      Google
                      (41388)
                        89%

 Corneto
                                (29815)
  (8723)
                                  63%
   18%
            (2216)
              5%
                     GTAA
                     (1634)
                       3%
RESULTS
 Zero-hits
Zero-hits
  (4404)
 (4404)
   11%
  11%
                                         Zero-hit tags
                     Google
                                         • garbled text
                     (41388)
                       89%
                                         • seriously mistyped words
 Corneto
                                         • grammatically incorrect
                               (29815)
  (8723)
   18%
                                 63%     sentences
           (2216)
             5%
                    GTAA
                    (1634)
                      3%
RESULTS
Zero-hits
 (4404)
  11%




                      Google
                      (41388)
                        89%

 Corneto
                                (29815)
  (8723)
                                  63%
   18%
            (2216)
              5%
                     GTAA
                     (1634)
                       3%
RESULTS
Zero-hits
 (4404)
  11%

                                           Google “Meaningful” Tags
                                           • multi-word tags
                      Google
                      (41388)              • slang
                        89%
                                           • names
 Corneto                        Pos-hits
  (8723)
                                (29815)
                                (29815)
                                  63%
                                           • morphological variations
   18%                            63%
            (2216)                         • minor typos
              5%
                     GTAA                  • etc…
                     (1634)
                       3%
RESULTS
Zero-hits
 (4404)
  11%




                      Google
                      (41388)
                        89%

 Corneto
                                (29815)
  (8723)
                                  63%
   18%
            (2216)
              5%
                     GTAA
                     (1634)
                       3%
2
             TAG CLASSIFICATION
 What video aspects are described by user tags?
        Manual classification of tag sample
                     5 videos
non-visual (0), perceptual (11), the rest conceptual
 Object (most of the tags) vs. Scene tags (only 30)

                  3 levels and 4 facets:
                abstract, general & specific
                 who, what, when, where
    Tag sample: all verified tags
      from random fragments
     (182 non-descriptive tags
         removed) #1,343
2
                                   RESULTS
                           Only 30 scene-level tags
                 Panofsky-Shatford matrix for object-level tags

                      Abstract           General            Specific             Total
     Who                 10                166                177                31%
     What                73                  5
                                           563                 12                57%
     Where                0                 68                  8                7%
     When                 4                 31                  6                5%
     Total              7%                 74%                9%
    195 tags (typically adverbs & adjectives) couldn’t be classified in
                            any of the facets

[1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
2
                                   RESULTS
                           Only 30 scene-level tags
                 Panofsky-Shatford matrix for object-level tags

                      Abstract           General            Specific             Total
     Who                 10                166                177                31%
     What                73                  5
                                           563                 12                57%
     Where                0                 68                  8                7%
     When                 4                 31                  6                5%
     Total              7%                 74%
                                          74%                 9%
    195 tags (typically adverbs & adjectives) couldn’t be classified in
                            any of the facets

[1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
2
                                   RESULTS
                           Only 30 scene-level tags
                 Panofsky-Shatford matrix for object-level tags

                      Abstract           General            Specific             Total
     Who                 10                166                177                31%
     What                73                  5
                                           563                 12                57%
     Where                0                 68                  8                7%
     When                 4                 31                  6                5%
     Total              7%                 74%                9%
    195 tags (typically adverbs & adjectives) couldn’t be classified in
                            any of the facets

[1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
2
                                   RESULTS
                           Only 30 scene-level tags
                 Panofsky-Shatford matrix for object-level tags

                      Abstract           General            Specific             Total
     Who                 10                166                177               31%
                                                                                 31%
     What                73                 5
                                           563                 12               57%
                                                                                 57%
     Where                0                 68                  8                7%
     When                 4                 31                  6                5%
     Total              7%                 74%                9%
    195 tags (typically adverbs & adjectives) couldn’t be classified in
                            any of the facets

[1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
EXPERIMENTAL SET

5 videos       Episodes        All tags   Verified    Genre
well tagged    Reality show     25,965     5837     amusement

well tagged    Reality show     22,792     6153     amusement

 medium       Missing people    1007        274     informative

low tagged      Reporter         403        73      informative

low tagged      The Walk         257        45       religious
genre might influence the usefulness of the user tags
users focus on what can be seen or heard (directly perceivable objects)
                because of the fast pace of the game




                  USEFULNESS
                  according to professionals
SUBTITLES USEFUL?
                    on the average
                   26% of all the tags
                 (35% of the verified)
                are matched in subtitles

             used to run bot-based games

otherwise, don’t really bring much new to the annotation
OBSERVATIONS FROM ANALYSIS:

taggers & professionals use different
vocabularies
  user tags complement professional annotations
  user tags could bridge the searcher vs. cataloger gap
taggers tend to use general concepts more
taggers focus mainly on the What and Who
taggers describe predominantly objects & rarely
scenes
OBSERVATIONS FROM ANALYSIS:

need for validation for quality & correctness
crowd-sourcing users need continuous support
& motivation to supply more and better
contributions  
THE QUESTION IS:

Can user tags be used for fragment search?
How to derive user tag quality metrics?
THE QUALITY OF TAGS?
MANAGE
Moderate Waisda? Analyzed Tags
Tag gardening

  annotations for video,
   fragments & scenes

 adding types & roles to
        user tags

mapping tags to concepts,
    disambiguation

   spelling corrections

single & batch tags/video
         editing
sorted by
frequency all frames per tag
Inspired by Google Refine
INITIAL (SMALL) EXPERIMENT
•1   video, 36 tags, 4 validators

• validatorsselects the most appropriate concept for each tag
 (if present)

• frames   within the video available - linked to the tags

• reconciliation   agains GTAA, Cornetto, Freebase

• (1)
    select a source, (2) start reconciliation (3) choose from
 suggested concepts
krippendorff
coverage 33 out of 36 tags                      reliability
                                               coefficient




Number of tags reconciled by the four participants

          more than 50% of concepts




           Ranks of the selected tags
Lessons
• Cornetto is good for subject terms
 • Selecting most suitable concept is time consuming as the
     differences sometimes are difficult to understand
• For all NE users could select a concept from Freebase & GTAA
 • Vocabularies like GTAA have less additional information, thus
     it’s more difficult to select the right concept
  • Freebase works well for English (Dutch might be a problem)
• Full coverage - prevented by spelling errors and variations
• Users adapt quickly, e.g. Freebase is good for people & locations
Recommendations
• reconcile agains multiple data sources at the same
  time for best coverage
  • merge duplicates into a single suggestion
  • organize results in categories, e.g. people, locations
     and subjects (to avoid suggestions from the wrong type)
• pre-processing to merge similar concepts (deal
  with spelling errors and variations)
• rank the suggested concepts by precision, then by recall
  (as alternative)
INTEGRATE
the improved (gardened) user annotations
FRAGMENT RETRIEVAL
FRAGMENT RETRIEVAL
FRAGMENT RETRIEVAL
WITHIN VIDEO NAVIGATION
WITHIN VIDEO NAVIGATION
group by type
  (faceted)
WITHIN VIDEO NAVIGATION
WITHIN VIDEO NAVIGATION

       concepts
       uniquely
      identified
WITHIN VIDEO NAVIGATION
WITHIN VIDEO NAVIGATION




            background
               info
THE TEAM


                   Riste Gligorov         @mbrinkerink
  @GuusSchreiber                    Marteen Brinkerink
Guus Schreiber                                            @johanoomen
                                                         Johan Oomen
                         @laroyo
                   Lora Aroyo
                          @McHildebrand
           @jrvosse           Michiel Hildebrand         @lottebelice
Jacco van Ossenbruggen                      Lotte Belice Baltussen
LINKS

      http://waisda.nl

http://www.prestoprime.org/

         @waisda



http://www.cs.vu.nl/~laroyo

http://slideshare.com/laroyo

More Related Content

More from Lora Aroyo

Data excellence: Better data for better AI
Data excellence: Better data for better AIData excellence: Better data for better AI
Data excellence: Better data for better AILora Aroyo
 
CHIP Demonstrator presentation @ CATCH Symposium
CHIP Demonstrator presentation @ CATCH SymposiumCHIP Demonstrator presentation @ CATCH Symposium
CHIP Demonstrator presentation @ CATCH SymposiumLora Aroyo
 
Semantic Web Challenge: CHIP Demonstrator
Semantic Web Challenge: CHIP DemonstratorSemantic Web Challenge: CHIP Demonstrator
Semantic Web Challenge: CHIP DemonstratorLora Aroyo
 
The Rijksmuseum Collection as Linked Data
The Rijksmuseum Collection as Linked DataThe Rijksmuseum Collection as Linked Data
The Rijksmuseum Collection as Linked DataLora Aroyo
 
Keynote at International Conference of Art Libraries 2018 @Rijksmuseum
Keynote at International Conference of Art Libraries 2018 @RijksmuseumKeynote at International Conference of Art Libraries 2018 @Rijksmuseum
Keynote at International Conference of Art Libraries 2018 @RijksmuseumLora Aroyo
 
FAIRview: Responsible Video Summarization @NYCML'18
FAIRview: Responsible Video Summarization @NYCML'18FAIRview: Responsible Video Summarization @NYCML'18
FAIRview: Responsible Video Summarization @NYCML'18Lora Aroyo
 
Understanding bias in video news & news filtering algorithms
Understanding bias in video news & news filtering algorithmsUnderstanding bias in video news & news filtering algorithms
Understanding bias in video news & news filtering algorithmsLora Aroyo
 
StorySourcing: Telling Stories with Humans & Machines
StorySourcing: Telling Stories with Humans & MachinesStorySourcing: Telling Stories with Humans & Machines
StorySourcing: Telling Stories with Humans & MachinesLora Aroyo
 
Data Science with Humans in the Loop
Data Science with Humans in the LoopData Science with Humans in the Loop
Data Science with Humans in the LoopLora Aroyo
 
Digital Humanities Benelux 2017: Keynote Lora Aroyo
Digital Humanities Benelux 2017: Keynote Lora AroyoDigital Humanities Benelux 2017: Keynote Lora Aroyo
Digital Humanities Benelux 2017: Keynote Lora AroyoLora Aroyo
 
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...Lora Aroyo
 
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017Crowdsourcing ambiguity aware ground truth - collective intelligence 2017
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017Lora Aroyo
 
My ESWC 2017 keynote: Disrupting the Semantic Comfort Zone
My ESWC 2017 keynote: Disrupting the Semantic Comfort ZoneMy ESWC 2017 keynote: Disrupting the Semantic Comfort Zone
My ESWC 2017 keynote: Disrupting the Semantic Comfort ZoneLora Aroyo
 
Data Science with Human in the Loop @Faculty of Science #Leiden University
Data Science with Human in the Loop @Faculty of Science #Leiden UniversityData Science with Human in the Loop @Faculty of Science #Leiden University
Data Science with Human in the Loop @Faculty of Science #Leiden UniversityLora Aroyo
 
SXSW2017 @NewDutchMedia Talk: Exploration is the New Search
SXSW2017 @NewDutchMedia Talk: Exploration is the New SearchSXSW2017 @NewDutchMedia Talk: Exploration is the New Search
SXSW2017 @NewDutchMedia Talk: Exploration is the New SearchLora Aroyo
 
Europeana GA 2016: Harnessing Crowds, Niches & Professionals in the Digital Age
Europeana GA 2016: Harnessing Crowds, Niches & Professionals  in the Digital AgeEuropeana GA 2016: Harnessing Crowds, Niches & Professionals  in the Digital Age
Europeana GA 2016: Harnessing Crowds, Niches & Professionals in the Digital AgeLora Aroyo
 
"Video Killed the Radio Star": From MTV to Snapchat
"Video Killed the Radio Star": From MTV to Snapchat"Video Killed the Radio Star": From MTV to Snapchat
"Video Killed the Radio Star": From MTV to SnapchatLora Aroyo
 
UMAP 2016 Opening Ceremony
UMAP 2016 Opening CeremonyUMAP 2016 Opening Ceremony
UMAP 2016 Opening CeremonyLora Aroyo
 
Crowdsourcing & Nichesourcing: Enriching Cultural Heritage with Experts & Cr...
Crowdsourcing & Nichesourcing: Enriching Cultural Heritagewith Experts & Cr...Crowdsourcing & Nichesourcing: Enriching Cultural Heritagewith Experts & Cr...
Crowdsourcing & Nichesourcing: Enriching Cultural Heritage with Experts & Cr...Lora Aroyo
 
Stitch by Stitch: Annotating Fashion at the Rijksmuseum
Stitch by Stitch: Annotating Fashion at the RijksmuseumStitch by Stitch: Annotating Fashion at the Rijksmuseum
Stitch by Stitch: Annotating Fashion at the RijksmuseumLora Aroyo
 

More from Lora Aroyo (20)

Data excellence: Better data for better AI
Data excellence: Better data for better AIData excellence: Better data for better AI
Data excellence: Better data for better AI
 
CHIP Demonstrator presentation @ CATCH Symposium
CHIP Demonstrator presentation @ CATCH SymposiumCHIP Demonstrator presentation @ CATCH Symposium
CHIP Demonstrator presentation @ CATCH Symposium
 
Semantic Web Challenge: CHIP Demonstrator
Semantic Web Challenge: CHIP DemonstratorSemantic Web Challenge: CHIP Demonstrator
Semantic Web Challenge: CHIP Demonstrator
 
The Rijksmuseum Collection as Linked Data
The Rijksmuseum Collection as Linked DataThe Rijksmuseum Collection as Linked Data
The Rijksmuseum Collection as Linked Data
 
Keynote at International Conference of Art Libraries 2018 @Rijksmuseum
Keynote at International Conference of Art Libraries 2018 @RijksmuseumKeynote at International Conference of Art Libraries 2018 @Rijksmuseum
Keynote at International Conference of Art Libraries 2018 @Rijksmuseum
 
FAIRview: Responsible Video Summarization @NYCML'18
FAIRview: Responsible Video Summarization @NYCML'18FAIRview: Responsible Video Summarization @NYCML'18
FAIRview: Responsible Video Summarization @NYCML'18
 
Understanding bias in video news & news filtering algorithms
Understanding bias in video news & news filtering algorithmsUnderstanding bias in video news & news filtering algorithms
Understanding bias in video news & news filtering algorithms
 
StorySourcing: Telling Stories with Humans & Machines
StorySourcing: Telling Stories with Humans & MachinesStorySourcing: Telling Stories with Humans & Machines
StorySourcing: Telling Stories with Humans & Machines
 
Data Science with Humans in the Loop
Data Science with Humans in the LoopData Science with Humans in the Loop
Data Science with Humans in the Loop
 
Digital Humanities Benelux 2017: Keynote Lora Aroyo
Digital Humanities Benelux 2017: Keynote Lora AroyoDigital Humanities Benelux 2017: Keynote Lora Aroyo
Digital Humanities Benelux 2017: Keynote Lora Aroyo
 
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...
 
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017Crowdsourcing ambiguity aware ground truth - collective intelligence 2017
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017
 
My ESWC 2017 keynote: Disrupting the Semantic Comfort Zone
My ESWC 2017 keynote: Disrupting the Semantic Comfort ZoneMy ESWC 2017 keynote: Disrupting the Semantic Comfort Zone
My ESWC 2017 keynote: Disrupting the Semantic Comfort Zone
 
Data Science with Human in the Loop @Faculty of Science #Leiden University
Data Science with Human in the Loop @Faculty of Science #Leiden UniversityData Science with Human in the Loop @Faculty of Science #Leiden University
Data Science with Human in the Loop @Faculty of Science #Leiden University
 
SXSW2017 @NewDutchMedia Talk: Exploration is the New Search
SXSW2017 @NewDutchMedia Talk: Exploration is the New SearchSXSW2017 @NewDutchMedia Talk: Exploration is the New Search
SXSW2017 @NewDutchMedia Talk: Exploration is the New Search
 
Europeana GA 2016: Harnessing Crowds, Niches & Professionals in the Digital Age
Europeana GA 2016: Harnessing Crowds, Niches & Professionals  in the Digital AgeEuropeana GA 2016: Harnessing Crowds, Niches & Professionals  in the Digital Age
Europeana GA 2016: Harnessing Crowds, Niches & Professionals in the Digital Age
 
"Video Killed the Radio Star": From MTV to Snapchat
"Video Killed the Radio Star": From MTV to Snapchat"Video Killed the Radio Star": From MTV to Snapchat
"Video Killed the Radio Star": From MTV to Snapchat
 
UMAP 2016 Opening Ceremony
UMAP 2016 Opening CeremonyUMAP 2016 Opening Ceremony
UMAP 2016 Opening Ceremony
 
Crowdsourcing & Nichesourcing: Enriching Cultural Heritage with Experts & Cr...
Crowdsourcing & Nichesourcing: Enriching Cultural Heritagewith Experts & Cr...Crowdsourcing & Nichesourcing: Enriching Cultural Heritagewith Experts & Cr...
Crowdsourcing & Nichesourcing: Enriching Cultural Heritage with Experts & Cr...
 
Stitch by Stitch: Annotating Fashion at the Rijksmuseum
Stitch by Stitch: Annotating Fashion at the RijksmuseumStitch by Stitch: Annotating Fashion at the Rijksmuseum
Stitch by Stitch: Annotating Fashion at the Rijksmuseum
 

Recently uploaded

Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxLoriGlavin3
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...Wes McKinney
 
2024 April Patch Tuesday
2024 April Patch Tuesday2024 April Patch Tuesday
2024 April Patch TuesdayIvanti
 
Testing tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesTesting tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesKari Kakkonen
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxLoriGlavin3
 
Generative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfGenerative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfIngrid Airi González
 
Scale your database traffic with Read & Write split using MySQL Router
Scale your database traffic with Read & Write split using MySQL RouterScale your database traffic with Read & Write split using MySQL Router
Scale your database traffic with Read & Write split using MySQL RouterMydbops
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxLoriGlavin3
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsNathaniel Shimoni
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxLoriGlavin3
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxLoriGlavin3
 
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...Alkin Tezuysal
 
Manual 508 Accessibility Compliance Audit
Manual 508 Accessibility Compliance AuditManual 508 Accessibility Compliance Audit
Manual 508 Accessibility Compliance AuditSkynet Technologies
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfMounikaPolabathina
 
A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI AgeCprime
 
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...panagenda
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024Lonnie McRorey
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteDianaGray10
 

Recently uploaded (20)

Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
 
2024 April Patch Tuesday
2024 April Patch Tuesday2024 April Patch Tuesday
2024 April Patch Tuesday
 
Testing tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examplesTesting tools and AI - ideas what to try with some tool examples
Testing tools and AI - ideas what to try with some tool examples
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
 
Generative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdfGenerative Artificial Intelligence: How generative AI works.pdf
Generative Artificial Intelligence: How generative AI works.pdf
 
Scale your database traffic with Read & Write split using MySQL Router
Scale your database traffic with Read & Write split using MySQL RouterScale your database traffic with Read & Write split using MySQL Router
Scale your database traffic with Read & Write split using MySQL Router
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directions
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
 
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
 
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
 
Manual 508 Accessibility Compliance Audit
Manual 508 Accessibility Compliance AuditManual 508 Accessibility Compliance Audit
Manual 508 Accessibility Compliance Audit
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdf
 
A Framework for Development in the AI Age
A Framework for Development in the AI AgeA Framework for Development in the AI Age
A Framework for Development in the AI Age
 
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test Suite
 

Crowdsourcing & Citizen Science

  • 1. FROM CROWD KNOWLEDGE TO MACHINE KNOWLEDGE use cases with semantics & user interaction in cultural heritage collections Lora Aroyo VU University Amsterdam @laroyo
  • 2. OBSERVATION I: VIDEO ANNOTATIONS Video: rich in meaning expressed with objects, people, events & symbols Annotation: tedious, time- consuming, incomplete Professional annotations: coarse-grained & refer to entire video and topics
  • 3. OBSERVATION II: SEARCH BEHAVIOR People predominantly request video fragments: broadcast (33%), stories (17%), fragments (49%) Finding fragments takes much longer than finding broadcasts: stories (2x), fragment (3x)
  • 4. OBSERVATION III: VOCABULARY MISMATCH experts use a specific vocabulary that is unknown to general audiences 35% of clicked results are not found by title or thesaurus term
  • 5. OBSERVATION IV: LIMITED ANNOTATION RESOURCES the professionals can no longer cope with the demand for manual annotation on the ever growing multimedia content in their collections
  • 6. SO WE WANT TO ... Make a massive set of videos accessible to end users Improve video search for end users Maintain a growing community of engaged users Support professional annotators
  • 7. TO IMPROVE VIDEO SEARCH: fragment retrieval within video navigation
  • 8. REQUIRES CHANGES IN ANNOTATIONS: Including time-based annotations Bridging the vocabulary gap of searcher &. cataloguer
  • 9. exploit crowdsourcing on the web integrate in content management system
  • 10. CULTURAL HERITAGE WITH INVOLVED CONSUMERS & PROVIDERS data: open, shared & accessible infrastructures: interoperable smart services: relevant content in right context
  • 11. Rijksmuseum Amsterdam http://chip-project.org http://e-culture.multimedian.nl/ pk/annotate?uri=RP-P- OB-77.320
  • 12. Netherlands Institute for Sound & Vision http://openimages.eu http://academia.nl MUNCH project CHOICE project
  • 13. Europeana Cultural Search http://e-culture.multimedian.nl http://europeana.eu/portal/ thoughtlab.html http://europeana.eu
  • 14. WAYS TO USE THE CROWD Tagging & classification Editing & transcribing Contextualising Acquisition Co-curation Crowdfunding
  • 15. WHAT WE DID: THE WORKFLOW collect-analyse-manage-integrate user-generated metadata
  • 16.
  • 17. domain experts gamers collection managers
  • 18. Winner EuroITV Competition Best Archives on the Web Award COLLECT User Generated Metadata
  • 19. TWO PILOTS Waisda? 1.0 (8 months) focus group with lay users for motivation aspects cognitive walk through with professionals usability testing with lay users and professionals tag analysis - primarily WHAT & WHO Waisda? 2.0 (starting this September) currently running initial user & tag analysis
  • 20.
  • 21. Waisda? 1.0 (8 months) Participation  44,000 pageviews 2,000 different players 500 registered players  thousands anonymous players Time-based annotations  612 videos tagged  420,000 tags added  46,000 unique tags Vocabulary gap User tags added the user perspective on the collection Community consensus as a measure of ‘validity’ Players score points when tag exactly matches a tag entered by another within 10 secs
  • 22.
  • 23. Waisda? 2.0 (4 weeks not public) Participation  ~1500 users  81 registered  1435 anonymous users  2344 games Time-based annotations 11,109 videos in the game  32,200 tags --> 25,600 first time tags 322 were played with  19% are one time tags (81% appear more than once)  12,000 validated tags (36%) / (4000 of them first time tag)  1,900 match in vocabs - 257 GTAA people/83 validated; 1,661 GTAA geo/666 validated  9,796 validated, but no match in GTAA
  • 24.
  • 25.
  • 26.
  • 27.
  • 28.
  • 29. MANAGE Analyze Waisda? Collected Tags; Quality Metrics
  • 30. WHAT ARE THE RELATIONS BETWEEN TAGS & PROFESSIONAL ANNOTATIONS? in terms of the vocabulary used? in terms of the topics they describe? Initially: manual analysis Output goal: to be expressed in a quality metric
  • 31. WAISDA? TAG ANALYSIS Study 1 Quantitative analysis of all 46,762 Waisda? tags Tag coverage w.r.t different vocabularies Study 2 Qualitative analysis of selected video fragments Manual classification of 1,343 “verified” tags
  • 32. DURING FIRST 9 MONTHS 400000 300000 > 600 videos 13 TV series Number
of
tags 200000 100000 > 420,000 tags collected 0 46,792 unique tags 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 500 Day each top 5 most-tagged video has > 23,000 tags 375 Number
of
arrivals 250 avg. tag density > 8 tags/sec 125 0 1 11 21 31 41 51 61 71 81 91 101 111 121 131 141 151 161 171 181 191 201 211 221 231 Day
  • 33. 1 MAPPING TAGS TO VOCABULARIES Vocabulary of S&V GTAA 160 000 terms in six Thesaurus disjoint facets: keyword, to ap pe d location, maker, genre, m person, named entities mapped Dutch lexical database m ap pe d to collective vocabulary of internet users
  • 34. RESULTS Zero-hits (4404) 11% Google (41388) 89% Corneto (29815) (8723) 63% 18% (2216) 5% GTAA (1634) 3%
  • 35. RESULTS Zero-hits (4404) 11% Facet Tags Subject 1199 Google (41388) Location 613 89% Genre 52 Corneto (8723) (29815) Person 118 63% 18% (2216) Maker 4 5%GTAA GTAA Name 673 (3850) (1634) 8%3%
  • 36. RESULTS Zero-hits (4404) 11% Google (41388) 89% Corneto (29815) (8723) 63% 18% (2216) 5% GTAA (1634) 3%
  • 37. RESULTS Zero-hits (4404) 11% Types Tags Noun 7222 Google (41388) 89% Verb 2090 Corneto Adjective 1693 Corneto (8723) (29815) (10939) 63% 18% 23%(2216) Adverb 171 5% GTAA (1634) 3%
  • 38. RESULTS Zero-hits (4404) 11% Google (41388) 89% Corneto (29815) (8723) 63% 18% (2216) 5% GTAA (1634) 3%
  • 39. RESULTS Zero-hits Zero-hits (4404) (4404) 11% 11% Zero-hit tags Google • garbled text (41388) 89% • seriously mistyped words Corneto • grammatically incorrect (29815) (8723) 18% 63% sentences (2216) 5% GTAA (1634) 3%
  • 40. RESULTS Zero-hits (4404) 11% Google (41388) 89% Corneto (29815) (8723) 63% 18% (2216) 5% GTAA (1634) 3%
  • 41. RESULTS Zero-hits (4404) 11% Google “Meaningful” Tags • multi-word tags Google (41388) • slang 89% • names Corneto Pos-hits (8723) (29815) (29815) 63% • morphological variations 18% 63% (2216) • minor typos 5% GTAA • etc… (1634) 3%
  • 42. RESULTS Zero-hits (4404) 11% Google (41388) 89% Corneto (29815) (8723) 63% 18% (2216) 5% GTAA (1634) 3%
  • 43. 2 TAG CLASSIFICATION What video aspects are described by user tags? Manual classification of tag sample 5 videos non-visual (0), perceptual (11), the rest conceptual Object (most of the tags) vs. Scene tags (only 30) 3 levels and 4 facets: abstract, general & specific who, what, when, where Tag sample: all verified tags from random fragments (182 non-descriptive tags removed) #1,343
  • 44. 2 RESULTS Only 30 scene-level tags Panofsky-Shatford matrix for object-level tags Abstract General Specific Total Who 10 166 177 31% What 73 5 563 12 57% Where 0 68 8 7% When 4 31 6 5% Total 7% 74% 9% 195 tags (typically adverbs & adjectives) couldn’t be classified in any of the facets [1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
  • 45. 2 RESULTS Only 30 scene-level tags Panofsky-Shatford matrix for object-level tags Abstract General Specific Total Who 10 166 177 31% What 73 5 563 12 57% Where 0 68 8 7% When 4 31 6 5% Total 7% 74% 74% 9% 195 tags (typically adverbs & adjectives) couldn’t be classified in any of the facets [1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
  • 46. 2 RESULTS Only 30 scene-level tags Panofsky-Shatford matrix for object-level tags Abstract General Specific Total Who 10 166 177 31% What 73 5 563 12 57% Where 0 68 8 7% When 4 31 6 5% Total 7% 74% 9% 195 tags (typically adverbs & adjectives) couldn’t be classified in any of the facets [1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
  • 47. 2 RESULTS Only 30 scene-level tags Panofsky-Shatford matrix for object-level tags Abstract General Specific Total Who 10 166 177 31% 31% What 73 5 563 12 57% 57% Where 0 68 8 7% When 4 31 6 5% Total 7% 74% 9% 195 tags (typically adverbs & adjectives) couldn’t be classified in any of the facets [1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
  • 48. EXPERIMENTAL SET 5 videos Episodes All tags Verified Genre well tagged Reality show 25,965 5837 amusement well tagged Reality show 22,792 6153 amusement medium Missing people 1007 274 informative low tagged Reporter 403 73 informative low tagged The Walk 257 45 religious
  • 49. genre might influence the usefulness of the user tags users focus on what can be seen or heard (directly perceivable objects) because of the fast pace of the game USEFULNESS according to professionals
  • 50. SUBTITLES USEFUL? on the average 26% of all the tags (35% of the verified) are matched in subtitles used to run bot-based games otherwise, don’t really bring much new to the annotation
  • 51. OBSERVATIONS FROM ANALYSIS: taggers & professionals use different vocabularies user tags complement professional annotations user tags could bridge the searcher vs. cataloger gap taggers tend to use general concepts more taggers focus mainly on the What and Who taggers describe predominantly objects & rarely scenes
  • 52. OBSERVATIONS FROM ANALYSIS: need for validation for quality & correctness crowd-sourcing users need continuous support & motivation to supply more and better contributions  
  • 53. THE QUESTION IS: Can user tags be used for fragment search? How to derive user tag quality metrics?
  • 54. THE QUALITY OF TAGS?
  • 56. Tag gardening annotations for video, fragments & scenes adding types & roles to user tags mapping tags to concepts, disambiguation spelling corrections single & batch tags/video editing
  • 57.
  • 58.
  • 59.
  • 60.
  • 61.
  • 62. sorted by frequency all frames per tag
  • 63.
  • 64.
  • 65.
  • 66.
  • 67.
  • 68.
  • 70. INITIAL (SMALL) EXPERIMENT •1 video, 36 tags, 4 validators • validatorsselects the most appropriate concept for each tag (if present) • frames within the video available - linked to the tags • reconciliation agains GTAA, Cornetto, Freebase • (1) select a source, (2) start reconciliation (3) choose from suggested concepts
  • 71.
  • 72. krippendorff coverage 33 out of 36 tags reliability coefficient Number of tags reconciled by the four participants more than 50% of concepts Ranks of the selected tags
  • 73. Lessons • Cornetto is good for subject terms • Selecting most suitable concept is time consuming as the differences sometimes are difficult to understand • For all NE users could select a concept from Freebase & GTAA • Vocabularies like GTAA have less additional information, thus it’s more difficult to select the right concept • Freebase works well for English (Dutch might be a problem) • Full coverage - prevented by spelling errors and variations • Users adapt quickly, e.g. Freebase is good for people & locations
  • 74. Recommendations • reconcile agains multiple data sources at the same time for best coverage • merge duplicates into a single suggestion • organize results in categories, e.g. people, locations and subjects (to avoid suggestions from the wrong type) • pre-processing to merge similar concepts (deal with spelling errors and variations) • rank the suggested concepts by precision, then by recall (as alternative)
  • 80. WITHIN VIDEO NAVIGATION group by type (faceted)
  • 82. WITHIN VIDEO NAVIGATION concepts uniquely identified
  • 84. WITHIN VIDEO NAVIGATION background info
  • 85. THE TEAM Riste Gligorov @mbrinkerink @GuusSchreiber Marteen Brinkerink Guus Schreiber @johanoomen Johan Oomen @laroyo Lora Aroyo @McHildebrand @jrvosse Michiel Hildebrand @lottebelice Jacco van Ossenbruggen Lotte Belice Baltussen
  • 86. LINKS http://waisda.nl http://www.prestoprime.org/ @waisda http://www.cs.vu.nl/~laroyo http://slideshare.com/laroyo

Editor's Notes

  1. From the different research topics I am involved in today I address some of the results from projects in which we use explicit semantics and user interaction to make crowd knowledge effectively processable for machines\n\n\n*****\nFrom Crowd Knowledge to Machine Knowledge: Use cases with semantics and user interaction in Dutch cultural heritage collections\n\nIn this talk I will discuss several projects, for example with the Dutch national archive for sound and vision or with the Rijksmuseum Amsterdam, where we have experimented with semantics-based technologies and user interaction paradigms to provide systems with additional support for users to turn their lay knowledge into machine-processable knowledge. Turning this crowd knowledge into machine knowledge makes the system more intelligent and thus makes them capitalize on the knowledge assets in the crowds of users.\n\nThe problem our system addresses is concerned with making the massive set of interesting multimedia content available in these cultural heritage institutions accessible for a large community of users.  On the one hand, this content is difficult to find for the average user as they have been indexed by experts (curators, art historians, etc) who use a very specific vocabulary that is unknown to the general audience.  On the other hand, these professionals can no longer  cope with the demand for annotation on the ever growing multimedia content in these collections.\n\nOur solution to both these problem is to exploit crowdsourcing on the web, where there is a lot of specific domain knowledge and processing power available that such archives and museums would gladly incorporate. However, turning this knowledge from the mass of lay users into additional intelligence in content management systems is not a simple challenge for several reasons: (1) the lay users use different vocabularies and are also interested in different aspects of the collection items; (2) the collected user metadata needs to be validated for quality and correctness; (3) crowd-sourcing users need to be continuously supported and motivated in order to supply more and better knowledge in such systems.  I will present our data analysis of the crowdsourced content, our techniques for quality control and incorporating domain semantics, and the results of our attempts to employ different interaction strategies and user interfaces (for example games and simplified interfaces for interactive annotation) to engage and stimulate the users.\n\nDemos:\n-----------\nVideo-tagging game: http://woordentikkertje.manbijthond.nl/\nArt recommender and personalized museum tours: http://www.chip-project.org/demo/ (3rd prize winner of ISWC Semantic Web challenge)\nHistorical events in museum collections: http://agora.cs.vu.nl/demo/\n\nRelevant Papers:\n------------------------\n- On the role of user-generated metadata in audio visual collections, at K-CAP2011\nhttp://portal.acm.org/citation.cfm?id=1999702\n\n- Digital Hermeneutics:  Agora and the Online Understanding of Cultural Heritage, at WebSci2011, (Best paper nominee)\nhttp://www.websci11.org/fileadmin/websci/Papers/116_paper.pdf\n\n- The effects of transparency on trust in and acceptance of a content-based art recommender, UMUAI journal, (Best journal paper for 2008)\nhttp://www.springerlink.com/content/81q34u73mpp58u75/\n\n- Recommendations based on semantically enriched museum collections, Journal of Web Semantics\nhttp://www.sciencedirect.com/science/article/pii/S1570826808000681\n\n- Enhancing Content-Based Recommendation with the Task Model of Classification, at EKAW2010\nhttp://www.springerlink.com/content/p78hl5r283x79r13/\n\nShort bio:\n-----------------\nLora Aroyo is an associate professor at the Web and Media group, at the Department of Computer Science, Free University Amsterdam, The Netherlands. Her research interests are in using semantic web technologies for modeling user interests and context, recommendation systems and personalized access in Web-based applications. Typical example domains are cultural heritage collections, multimedia archives and interactive TV. She has coordinated the research work in the CHIP project on Cultural Heritage Information Personalization (http://chip-project.org). Currently she is a scientific coordinator of the EU Integrated Project NoTube dealing with the integration of Web and TV data with the help of semantics (http://notube.tv), a project leader of VU INTERTAIN Experimental Research Lab initiative (http://www.cs.vu.nl/intertain), and involved in the research on motivational user interaction for video-tagging games in the PrestoPrime project (http://www.prestoprime.org/) and modeling\nhistoric events in the Agora project (http://agora.cs.vu.nl/). She has organized numerous workshops in the areas of personalized access to cultural heritage, e-learning, interactive television, as well as on visual interfaces to the social and semantic web (PATCH, FutureTV, PersWeb, VISSW and DeRIVE). Lora has been actively involved in both the Semantic Web (PC co-chair and conference chair for ESWC2009 and ESWC2010 and PC co-chair for ISWC2011) and the Personalization and User modeling communities (on the editorial board for the UMUAI journal and on the steering committee of UMAP conference).\n\nMore information can be found at:\n-----------------------------------------------\nWebpage: http://www.cs.vu.nl/~laroyo\nSlideshare: http://www.slideshare.net/laroyo\nTwitter: @laroyo\n\n
  2. Nowadays av collections are undergoing process of transformation from archives of analog material to large digital (online) data stores, as videos are very much wanted by different types of end users. \n\nFor example, the Netherlands Institute of Sound and Vision archives all radio and TV material broadcasted in the Netherlands (has appr. 700,000 hours radio and television programs available online. \n\nFacilitating a successfully access to av collection items demands quality metadata associated with them.\n\nTraditionally, in AV achives it is the task of professional catalogers to manually describe the videos. Usually, in the process they follow well-defined , well-established guidelines and rules. They also may make use of auxiliary materials like controlled vocabularies, thesauri, and such.\nHowever, as we all know video is medium that is extremely rich in meaning. Directors and screenwriters create entire universes with complex interplay between characters, objects and events. Sometimes they may employ rich and complex abstract symbolic language. This makes that task of describing the meaning of a video as complicated as describing the real worlds. Which is no trivial matter.\nAs a result the process of annotation is tedious, time-consuming and inevitably incomplete. According to some research, it takes approximately 5 times of the duration of the material to annotate it completely. So for example, if we are talking about a documentary that lasts one hour, it will take approximately 5 hours for a cataloger to fully describe it. Furthermore,\n\nConsequently, professional annotations are coarse-grained in a sense that they are referring to the entire video describing prevalent topics. It may happen that catalogers provide more fine-grained, shot-level descriptions for a video. But this is exception of the rule and it is reserved to the most important pieces of the AV collection.\n
  3. \n* Search Behavior of Media Professionals at an Audiovisual Archive: A Transaction Log Analysis\nB. Huurnink, L. Hollink, W. van den Heuvel, and M. de Rijke.\n\n
  4. \n
  5. The problem our system addresses is concerned with making the massive set of interesting multimedia content available in these cultural heritage institutions accessible for a large community of users.  On the one hand, this content is difficult to find for the average user as they have been indexed by experts (curators, art historians, etc) who use a very specific vocabulary that is unknown to the general audience.  On the other hand, these professionals can no longer  cope with the demand for annotation on the ever growing multimedia content in these collections.\n
  6. The problem our system addresses is concerned with making the massive set of interesting multimedia content available in these cultural heritage institutions accessible for a large community of users.  On the one hand, this content is difficult to find for the average user as they have been indexed by experts (curators, art historians, etc) who use a very specific vocabulary that is unknown to the general audience.  On the other hand, these professionals can no longer  cope with the demand for annotation on the ever growing multimedia content in these collections.\n\nRecommendation of video fragments\nLinks between video fragments\nLinks to background information\nintegrate the Goal with the general --> merge the two slides --> one slide for RMA and Europeana\n\n
  7. The problem our system addresses is concerned with making the massive set of interesting multimedia content available in these cultural heritage institutions accessible for a large community of users.  On the one hand, this content is difficult to find for the average user as they have been indexed by experts (curators, art historians, etc) who use a very specific vocabulary that is unknown to the general audience.  On the other hand, these professionals can no longer  cope with the demand for annotation on the ever growing multimedia content in these collections.\n\nRecommendation of video fragments\nLinks between video fragments\nLinks to background information\nintegrate the Goal with the general --> merge the two slides --> one slide for RMA and Europeana\n\n
  8. The problem our system addresses is concerned with making the massive set of interesting multimedia content available in these cultural heritage institutions accessible for a large community of users.  On the one hand, this content is difficult to find for the average user as they have been indexed by experts (curators, art historians, etc) who use a very specific vocabulary that is unknown to the general audience.  On the other hand, these professionals can no longer  cope with the demand for annotation on the ever growing multimedia content in these collections.\n\nRecommendation of video fragments\nLinks between video fragments\nLinks to background information\nintegrate the Goal with the general --> merge the two slides --> one slide for RMA and Europeana\n\n
  9. Our solution to both these problem is to exploit crowdsourcing on the web, where there is a lot of specific domain knowledge and processing power available that such archives and museums would gladly incorporate. However, turning this knowledge from the mass of lay users into additional intelligence in content management systems is not a simple challenge for several reasons: (1) the lay users use different vocabularies and are also interested in different aspects of the collection items; (2) the collected user metadata needs to be validated for quality and correctness; (3) crowd-sourcing users need to be continuously supported and motivated in order to supply more and better knowledge in such systems.  I will present our data analysis of the crowdsourced content, our techniques for quality control and incorporating domain semantics, and the results of our attempts to employ different interaction strategies and user interfaces (for example games and simplified interfaces for interactive annotation) to engage and stimulate the users.\n\nunderstanding the user-generated data\ncontextualize the user-generated metadata \n\n\n
  10. supplement video data with background knowledge\n introduce the context, rijksmuseum\n the general picture, examples, CH landscape\n \n \n
  11. \n
  12. \n
  13. \n
  14. We hebben naar uiteenlopende voorbeelden gekeken binnen de sector en deze geclusterd\n\ncriterium - voorname/leidende rol erfgoed. Uiteraard zijn er ook veel grassroots initiatieven. \n\nEen van mijn favoriete voorbeelden is de International Amateur Scanning league.\nDigitaliseren van historische video’s in het publieke domein.\n\n\nwe komen vervolgens op een indeling in zes categorieen. \n\nwe hebben al goede voorbeelden gezien vandaag - ik zal met name voorbeelden noemen uit categorieen die niet in andere presentaties aan bod komen.\n\n
  15. \n
  16. Wikipedia as a compelling example --> calculating that creating Wikipedia as it stands today has taken one hundred\nmillion hours of cumulative thought, he juxtaposes this to the astounding 200 billion hours people watch TV in the\nUS alone. 200 billion hours would amount to two thousand Wikipedia projects-worth of free time, annually. \n\n
  17. \n
  18. three types of users\ncollect: gamers\nmanage: domain experts\nintegrate: collection maintainer\n
  19. explain games with purpose\n\nplay on words\n\n
  20. \n
  21. multi-player game \n 4 channels each continuously streaming videos from a predefined category\n a player starts a game by selecting one of the channels\n plays agains all the players who selected the same channel\n\n
  22. multi-player game \n users score points by entering tags (based on the principle of Louis van Ahn’s ESP games\n points are scored when two or more players enter the same tag\n as in the Yahoo! video tag game --> this is done in a period of 10 sec\n each tag is shown together with the score it has earned\n\nThe relevance of the tags is ensured by two factors: independence of players and consensus among players. The players don’t know each other, they are grouped randomly, and the only thing they share at the moment when they are playing the game is the video. So if they enter the same tag within a time frame the changes are the tag is relevant for the video.\n
  23. the changes:\n- shorter videos (~3 mins), independent fragments vs. fragmented videos\n- improved scoring (including matches in Cornetto and GTAA; first time tag)\n- game recap with continuous scoring\n- profile page\n\nbots again\n- different handling of users to start playing the game\n\n
  24. total # tags 32,248 (25,685 first time tags)\n 19% are one time tags (81% appear more than once)\n \n 257 person names in GTAA, 83 also validated\n 1,661 geo names in GTAA, 666 also validated\n 9,796 validated, but no match in GTAA\n\n
  25. \n
  26. \n
  27. \n
  28. First step manual\n\nGoal is to automate it and make it an integral part of Waisda\n\nEventually the output can be expressed in a quality metric\n
  29. \n\n
  30. How/why did you select those fragments\nHow did you do the classifacation\n
  31. OK some statistics. These stats are computed on the dataset that was accumulated by waisda? in the first six months after it went live \n
  32. The goal of our first data analysis is to investigate user terminology and compare it to the terminology used by professionals. We de\n
  33. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  34. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  35. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  36. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  37. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  38. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  39. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  40. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  41. non-visual --> 0\nperceptual --> 11 tags (referring to color)\nconceptual --> the rest (describing objects, 30 about scenes only)\n
  42. 30% of scenes are conceptual (by Laura Hollink)\n\n
  43. 30% of scenes are conceptual (by Laura Hollink)\n\n
  44. 30% of scenes are conceptual (by Laura Hollink)\n\n
  45. \n
  46. genre is influencing the specificity and usefulness\nusers focus on what can be seen or heard\nusers focus on directly perceivable objects - because of the pace\n\n
  47. \n
  48. Our solution to both these problem is to exploit crowdsourcing on the web, where there is a lot of specific domain knowledge and processing power available that such archives and museums would gladly incorporate. However, turning this knowledge from the mass of lay users into additional intelligence in content management systems is not a simple challenge for several reasons: (1) the lay users use different vocabularies and are also interested in different aspects of the collection items; (2) the collected user metadata needs to be validated for quality and correctness; (3) crowd-sourcing users need to be continuously supported and motivated in order to supply more and better knowledge in such systems.  I will present our data analysis of the crowdsourced content, our techniques for quality control and incorporating domain semantics, and the results of our attempts to employ different interaction strategies and user interfaces (for example games and simplified interfaces for interactive annotation) to engage and stimulate the users.\n\n
  49. Our solution to both these problem is to exploit crowdsourcing on the web, where there is a lot of specific domain knowledge and processing power available that such archives and museums would gladly incorporate. However, turning this knowledge from the mass of lay users into additional intelligence in content management systems is not a simple challenge for several reasons: (1) the lay users use different vocabularies and are also interested in different aspects of the collection items; (2) the collected user metadata needs to be validated for quality and correctness; (3) crowd-sourcing users need to be continuously supported and motivated in order to supply more and better knowledge in such systems.  I will present our data analysis of the crowdsourced content, our techniques for quality control and incorporating domain semantics, and the results of our attempts to employ different interaction strategies and user interfaces (for example games and simplified interfaces for interactive annotation) to engage and stimulate the users.\n\n
  50. Our solution to both these problem is to exploit crowdsourcing on the web, where there is a lot of specific domain knowledge and processing power available that such archives and museums would gladly incorporate. However, turning this knowledge from the mass of lay users into additional intelligence in content management systems is not a simple challenge for several reasons: (1) the lay users use different vocabularies and are also interested in different aspects of the collection items; (2) the collected user metadata needs to be validated for quality and correctness; (3) crowd-sourcing users need to be continuously supported and motivated in order to supply more and better knowledge in such systems.  I will present our data analysis of the crowdsourced content, our techniques for quality control and incorporating domain semantics, and the results of our attempts to employ different interaction strategies and user interfaces (for example games and simplified interfaces for interactive annotation) to engage and stimulate the users.\n\n\nCan user tags be used to derive topical descriptions for scenes?\n\n
  51. \n
  52. \n
  53. Bridge communities \n crowd\n professionals\n Combine perspectives\n (subjective) user perspective\n (objective) professional perspective\n Link tags to concepts\n Annotations of video scenes \n Quality assurance by moderation\n
  54. \n
  55. \n
  56. \n
  57. frames per tag!\nview all frames for a tag\nsorted by frequency\nFrequently occurring tags for scene descriptions\n
  58. Selecting a role\n
  59. Selecting a role\n
  60. Link tags to concepts\n
  61. not always easy,\nwordnet: polysemi, interpretations are close\nsome vocabularies have no description\n\nhow do tags relate to a shot\n- the frame is often a bit later\n
  62. \n
  63. \n
  64. publicly available reconciliation APIs: e.g. Freebase, Europeana, Talis (kasabi.com)\n\n
  65. \n
  66. krippendorff alpha\n\nWe conclude that for this small experiment, with a coverage of 33 out of 36 tags and a total reliability coefficient of 0.91, users can effectively reconcile tags using the selected services.\n\nMore than half of the selected concepts were ranked first (32), while 23 concepts were ranked second or lower. Several (6) selected concepts were even ranked sixth or lower\nThus, often the first concept could be automatically selected, but not always\n\n\n
  67. \n
  68. \n
  69. \n
  70. \n
  71. \n
  72. \n
  73. Open Archives Initiative Object Reuse and Exchange (OAI-ORE) defines standards for the description and exchange of aggregations of Web resources. The goal of these standards is to expose the rich content in these aggregations to applications that support authoring, deposit, exchange, visualization, reuse, and preservation. \n
  74. \n
  75. \n
  76. \n
  77. - group by type (faceted)\n- concept is uniquely identified (see full name for Juliana and Bernhard)\n- background information\n
  78. - group by type (faceted)\n- concept is uniquely identified (see full name for Juliana and Bernhard)\n- background information\n
  79. - group by type (faceted)\n- concept is uniquely identified (see full name for Juliana and Bernhard)\n- background information\n
  80. - group by type (faceted)\n- concept is uniquely identified (see full name for Juliana and Bernhard)\n- background information\n
  81. - group by type (faceted)\n- concept is uniquely identified (see full name for Juliana and Bernhard)\n- background information\n
  82. \n
  83. \n