SlideShare a Scribd company logo
1 of 86
FROM CROWD KNOWLEDGE
     TO MACHINE KNOWLEDGE
      use cases with
      semantics & user
      interaction in cultural
      heritage collections




      Lora Aroyo
VU University Amsterdam
       @laroyo
OBSERVATION I:
         VIDEO ANNOTATIONS

 Video: rich in meaning
 expressed with objects,
people, events & symbols
Annotation: tedious, time-
 consuming, incomplete

Professional annotations:
coarse-grained & refer to
 entire video and topics
OBSERVATION II:
            SEARCH BEHAVIOR

  People predominantly
request video fragments:
broadcast (33%), stories
 (17%), fragments (49%)

  Finding fragments takes
        much longer
  than ļ¬nding broadcasts:
stories (2x), fragment (3x)
OBSERVATION III:
    VOCABULARY MISMATCH

experts use a speciļ¬c
  vocabulary that is
 unknown to general
     audiences

35% of clicked results
are not found by title
  or thesaurus term
OBSERVATION IV:
      LIMITED ANNOTATION
           RESOURCES

the professionals can
no longer cope with
   the demand for
manual annotation on
  the ever growing
multimedia content in
   their collections
SO WE WANT TO ...
Make a massive set of videos accessible to end users

        Improve video search for end users

  Maintain a growing community of engaged users


          Support professional annotators
TO IMPROVE VIDEO SEARCH:

       fragment retrieval
     within video navigation
REQUIRES CHANGES IN
        ANNOTATIONS:

    Including time-based annotations

Bridging the vocabulary gap of searcher &.
                cataloguer
exploit crowdsourcing on the web
integrate in content management system
CULTURAL HERITAGE WITH
  INVOLVED CONSUMERS &
        PROVIDERS


data: open, shared & accessible
infrastructures: interoperable
smart services: relevant content
in right context
Rijksmuseum
        Amsterdam
http://chip-project.org


http://e-culture.multimedian.nl/
pk/annotate?uri=RP-P-
OB-77.320
Netherlands Institute
   for Sound & Vision

http://openimages.eu
http://academia.nl


MUNCH project
CHOICE project
Europeana
        Cultural Search
http://e-culture.multimedian.nl
http://europeana.eu/portal/
thoughtlab.html
http://europeana.eu
WAYS TO USE THE CROWD

    Tagging & classiļ¬cation
    Editing & transcribing
       Contextualising
         Acquisition
         Co-curation
        Crowdfunding
WHAT WE DID:
             THE WORKFLOW
collect-analyse-manage-integrate user-generated metadata
domain
         experts
gamers




                   collection
                   managers
Winner EuroITV Competition
Best Archives on the Web Award




  COLLECT
User Generated Metadata
TWO PILOTS
           Waisda? 1.0 (8 months)
focus group with lay users for motivation aspects
    cognitive walk through with professionals
 usability testing with lay users and professionals
          tag analysis - primarily WHAT & WHO



Waisda? 2.0 (starting this September)
              currently running
         initial user & tag analysis
Waisda? 1.0
(8 months)
Participation
ļƒ˜ 44,000 pageviews
ļƒ˜2,000 different players
ļƒ˜500 registered players
ļƒ˜ thousands anonymous players

Time-based annotations
ļƒ˜ 612 videos tagged
ļƒ˜ 420,000 tags added
ļƒ˜ 46,000 unique tags

Vocabulary gap
User tags added the user perspective on the collection
                Community consensus as a measure of ā€˜validityā€™
     Players score points when tag exactly matches a tag entered by another within 10 secs
Waisda? 2.0
(4 weeks
not public)

Participation
ļƒ˜ ~1500 users
ļƒ˜ 81 registered
ļƒ˜ 1435 anonymous
users
ļƒ˜ 2344 games




Time-based annotations                                   11,109 videos in the game
ļƒ˜ 32,200 tags --> 25,600 ļ¬rst time tags                  322 were played with
ļƒ˜ 19% are one time tags (81% appear more than once)
ļƒ˜ 12,000 validated tags (36%) / (4000 of them ļ¬rst time tag)
ļƒ˜ 1,900 match in vocabs - 257 GTAA people/83 validated; 1,661 GTAA geo/666 validated
ļƒ˜ 9,796 validated, but no match in GTAA
MANAGE
Analyze Waisda? Collected Tags; Quality Metrics
WHAT ARE THE RELATIONS
BETWEEN TAGS & PROFESSIONAL
      ANNOTATIONS?
           in terms of the vocabulary used?
         in terms of the topics they describe?


               Initially: manual analysis
   Output goal: to be expressed in a quality metric
WAISDA? TAG ANALYSIS

                       Study 1
Quantitative analysis of all 46,762 Waisda? tags
       Tag coverage w.r.t different vocabularies


                      Study 2
  Qualitative analysis of selected video fragments
 Manual classiļ¬cation of 1,343 ā€œveriļ¬edā€ tags
DURING FIRST 9 MONTHS
                        400000



                        300000                                                                                                                                                      > 600 videos
                                                                                                                                                                                    13 TV series
Numberā€©ofā€©tags




                        200000



                        100000                                                                                                                                                > 420,000 tags collected
                                  0
                                                                                                                                                                                46,792 unique tags
                                      1
                                           12
                                                23
                                                         34
                                                              45
                                                                   56
                                                                        67
                                                                             78
                                                                                  89
                                                                                        100
                                                                                              111
                                                                                                     122
                                                                                                           133
                                                                                                                     144
                                                                                                                           155
                                                                                                                                 166
                                                                                                                                       177
                                                                                                                                             188
                                                                                                                                                    199
                                                                                                                                                          210
                                                                                                                                                                221
                                                                                                                                                                       232
                            500
                                                                                                   Day
                                                                                                                                                                               each top 5 most-tagged
                                                                                                                                                                               video has > 23,000 tags
                            375
       Numberā€©ofā€©arrivals




                            250                                                                                                                                              avg. tag density > 8 tags/sec
                            125



                             0
                                      1
                                          11
                                               21
                                                    31
                                                         41
                                                              51
                                                                   61
                                                                        71
                                                                             81
                                                                                  91
                                                                                       101
                                                                                             111
                                                                                                   121
                                                                                                         131
                                                                                                               141
                                                                                                                     151
                                                                                                                           161
                                                                                                                                 171
                                                                                                                                       181
                                                                                                                                             191
                                                                                                                                                   201
                                                                                                                                                         211
                                                                                                                                                               221
                                                                                                                                                                     231




                                                                                               Day
1
    MAPPING TAGS TO
     VOCABULARIES
                                      Vocabulary of S&V
                          GTAA      160 000 terms in six
                         Thesaurus disjoint facets: keyword,
                    to

           ap
             pe
               d                    location, maker, genre,
       m
                                   person, named entities

       mapped                                Dutch lexical
                                              database
       m
           ap
             pe
               d
                   to
                                          collective
                                       vocabulary of
                                       internet users
RESULTS
Zero-hits
 (4404)
  11%




                      Google
                      (41388)
                        89%

 Corneto
                                (29815)
  (8723)
                                  63%
   18%
            (2216)
              5%
                     GTAA
                     (1634)
                       3%
RESULTS
Zero-hits
 (4404)
  11%

                                          Facet     Tags
                                         Subject    1199
                     Google
                     (41388)             Location   613
                       89%
                                          Genre      52
 Corneto
  (8723)
                               (29815)   Person     118
                                 63%
   18%
            (2216)
                                          Maker      4
              5%GTAA
                    GTAA                  Name      673
                (3850)
                    (1634)
                   8%3%
RESULTS
Zero-hits
 (4404)
  11%




                      Google
                      (41388)
                        89%

 Corneto
                                (29815)
  (8723)
                                  63%
   18%
            (2216)
              5%
                     GTAA
                     (1634)
                       3%
RESULTS
Zero-hits
 (4404)
  11%

                                       Types      Tags
                                       Noun       7222
                  Google
                  (41388)
                    89%                 Verb      2090

 Corneto                              Adjective   1693
  Corneto
  (8723)
                            (29815)
   (10939)                    63%
   18%
     23%(2216)
                                      Adverb      171
          5%
                 GTAA
                 (1634)
                   3%
RESULTS
Zero-hits
 (4404)
  11%




                      Google
                      (41388)
                        89%

 Corneto
                                (29815)
  (8723)
                                  63%
   18%
            (2216)
              5%
                     GTAA
                     (1634)
                       3%
RESULTS
 Zero-hits
Zero-hits
  (4404)
 (4404)
   11%
  11%
                                         Zero-hit tags
                     Google
                                         ā€¢ā€ˆgarbled text
                     (41388)
                       89%
                                         ā€¢ā€ˆseriously mistyped words
 Corneto
                                         ā€¢ā€ˆgrammatically incorrect
                               (29815)
  (8723)
   18%
                                 63%     sentences
           (2216)
             5%
                    GTAA
                    (1634)
                      3%
RESULTS
Zero-hits
 (4404)
  11%




                      Google
                      (41388)
                        89%

 Corneto
                                (29815)
  (8723)
                                  63%
   18%
            (2216)
              5%
                     GTAA
                     (1634)
                       3%
RESULTS
Zero-hits
 (4404)
  11%

                                           Google ā€œMeaningfulā€ Tags
                                           ā€¢ā€ˆmulti-word tags
                      Google
                      (41388)              ā€¢ā€ˆslang
                        89%
                                           ā€¢ā€ˆnames
 Corneto                        Pos-hits
  (8723)
                                (29815)
                                (29815)
                                  63%
                                           ā€¢ā€ˆmorphological variations
   18%                            63%
            (2216)                         ā€¢ā€ˆminor typos
              5%
                     GTAA                  ā€¢ā€ˆetcā€¦
                     (1634)
                       3%
RESULTS
Zero-hits
 (4404)
  11%




                      Google
                      (41388)
                        89%

 Corneto
                                (29815)
  (8723)
                                  63%
   18%
            (2216)
              5%
                     GTAA
                     (1634)
                       3%
2
             TAG CLASSIFICATION
 What video aspects are described by user tags?
        Manual classiļ¬cation of tag sample
                     5 videos
non-visual (0), perceptual (11), the rest conceptual
 Object (most of the tags) vs. Scene tags (only 30)

                  3 levels and 4 facets:
                abstract, general & speciļ¬c
                 who, what, when, where
    Tag sample: all veriļ¬ed tags
      from random fragments
     (182 non-descriptive tags
         removed) #1,343
2
                                   RESULTS
                           Only 30 scene-level tags
                 Panofsky-Shatford matrix for object-level tags

                      Abstract           General            Speciļ¬c             Total
     Who                 10                166                177                31%
     What                73                  5
                                           563                 12                57%
     Where                0                 68                  8                7%
     When                 4                 31                  6                5%
     Total              7%                 74%                9%
    195 tags (typically adverbs & adjectives) couldnā€™t be classiļ¬ed in
                            any of the facets

[1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
2
                                   RESULTS
                           Only 30 scene-level tags
                 Panofsky-Shatford matrix for object-level tags

                      Abstract           General            Speciļ¬c             Total
     Who                 10                166                177                31%
     What                73                  5
                                           563                 12                57%
     Where                0                 68                  8                7%
     When                 4                 31                  6                5%
     Total              7%                 74%
                                          74%                 9%
    195 tags (typically adverbs & adjectives) couldnā€™t be classiļ¬ed in
                            any of the facets

[1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
2
                                   RESULTS
                           Only 30 scene-level tags
                 Panofsky-Shatford matrix for object-level tags

                      Abstract           General            Speciļ¬c             Total
     Who                 10                166                177                31%
     What                73                  5
                                           563                 12                57%
     Where                0                 68                  8                7%
     When                 4                 31                  6                5%
     Total              7%                 74%                9%
    195 tags (typically adverbs & adjectives) couldnā€™t be classiļ¬ed in
                            any of the facets

[1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
2
                                   RESULTS
                           Only 30 scene-level tags
                 Panofsky-Shatford matrix for object-level tags

                      Abstract           General            Speciļ¬c             Total
     Who                 10                166                177               31%
                                                                                 31%
     What                73                 5
                                           563                 12               57%
                                                                                 57%
     Where                0                 68                  8                7%
     When                 4                 31                  6                5%
     Total              7%                 74%                9%
    195 tags (typically adverbs & adjectives) couldnā€™t be classiļ¬ed in
                            any of the facets

[1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
EXPERIMENTAL SET

5 videos       Episodes        All tags   Veriļ¬ed    Genre
well tagged    Reality show     25,965     5837     amusement

well tagged    Reality show     22,792     6153     amusement

 medium       Missing people    1007        274     informative

low tagged      Reporter         403        73      informative

low tagged      The Walk         257        45       religious
genre might inļ¬‚uence the usefulness of the user tags
users focus on what can be seen or heard (directly perceivable objects)
                because of the fast pace of the game




                  USEFULNESS
                  according to professionals
SUBTITLES USEFUL?
                    on the average
                   26% of all the tags
                 (35% of the veriļ¬ed)
                are matched in subtitles

             used to run bot-based games

otherwise, donā€™t really bring much new to the annotation
OBSERVATIONS FROM ANALYSIS:

taggers & professionals use different
vocabularies
  user tags complement professional annotations
  user tags could bridge the searcher vs. cataloger gap
taggers tend to use general concepts more
taggers focus mainly on the What and Who
taggers describe predominantly objects & rarely
scenes
OBSERVATIONS FROM ANALYSIS:

need for validation for quality & correctness
crowd-sourcing users need continuous support
& motivation to supply more and better
contributions Ā 
THE QUESTION IS:

Can user tags be used for fragment search?
How to derive user tag quality metrics?
THE QUALITY OF TAGS?
MANAGE
Moderate Waisda? Analyzed Tags
Tag gardening

  annotations for video,
   fragments & scenes

 adding types & roles to
        user tags

mapping tags to concepts,
    disambiguation

   spelling corrections

single & batch tags/video
         editing
sorted by
frequency all frames per tag
Inspired by Google Reļ¬ne
INITIAL (SMALL) EXPERIMENT
ā€¢1   video, 36 tags, 4 validators

ā€¢ validatorsselects the most appropriate concept for each tag
 (if present)

ā€¢ frames   within the video available - linked to the tags

ā€¢ reconciliation   agains GTAA, Cornetto, Freebase

ā€¢ (1)
    select a source, (2) start reconciliation (3) choose from
 suggested concepts
krippendorff
coverage 33 out of 36 tags                      reliability
                                               coefļ¬cient




Number of tags reconciled by the four participants

          more than 50% of concepts




           Ranks of the selected tags
Lessons
ā€¢ Cornetto is good for subject terms
 ā€¢ Selecting most suitable concept is time consuming as the
     differences sometimes are difļ¬cult to understand
ā€¢ For all NE users could select a concept from Freebase & GTAA
 ā€¢ Vocabularies like GTAA have less additional information, thus
     itā€™s more difļ¬cult to select the right concept
  ā€¢ Freebase works well for English (Dutch might be a problem)
ā€¢ Full coverage - prevented by spelling errors and variations
ā€¢ Users adapt quickly, e.g. Freebase is good for people & locations
Recommendations
ā€¢ reconcile agains multiple data sources at the same
  time for best coverage
  ā€¢ merge duplicates into a single suggestion
  ā€¢ organize results in categories, e.g. people, locations
     and subjects (to avoid suggestions from the wrong type)
ā€¢ pre-processing to merge similar concepts (deal
  with spelling errors and variations)
ā€¢ rank the suggested concepts by precision, then by recall
  (as alternative)
INTEGRATE
the improved (gardened) user annotations
FRAGMENT RETRIEVAL
FRAGMENT RETRIEVAL
FRAGMENT RETRIEVAL
WITHIN VIDEO NAVIGATION
WITHIN VIDEO NAVIGATION
group by type
  (faceted)
WITHIN VIDEO NAVIGATION
WITHIN VIDEO NAVIGATION

       concepts
       uniquely
      identiļ¬ed
WITHIN VIDEO NAVIGATION
WITHIN VIDEO NAVIGATION




            background
               info
THE TEAM


                   Riste Gligorov         @mbrinkerink
  @GuusSchreiber                    Marteen Brinkerink
Guus Schreiber                                            @johanoomen
                                                         Johan Oomen
                         @laroyo
                   Lora Aroyo
                          @McHildebrand
           @jrvosse           Michiel Hildebrand         @lottebelice
Jacco van Ossenbruggen                      Lotte Belice Baltussen
LINKS

      http://waisda.nl

http://www.prestoprime.org/

         @waisda



http://www.cs.vu.nl/~laroyo

http://slideshare.com/laroyo

More Related Content

More from Lora Aroyo

Data excellence: Better data for better AI
Data excellence: Better data for better AIData excellence: Better data for better AI
Data excellence: Better data for better AILora Aroyo
Ā 
CHIP Demonstrator presentation @ CATCH Symposium
CHIP Demonstrator presentation @ CATCH SymposiumCHIP Demonstrator presentation @ CATCH Symposium
CHIP Demonstrator presentation @ CATCH SymposiumLora Aroyo
Ā 
Semantic Web Challenge: CHIP Demonstrator
Semantic Web Challenge: CHIP DemonstratorSemantic Web Challenge: CHIP Demonstrator
Semantic Web Challenge: CHIP DemonstratorLora Aroyo
Ā 
The Rijksmuseum Collection as Linked Data
The Rijksmuseum Collection as Linked DataThe Rijksmuseum Collection as Linked Data
The Rijksmuseum Collection as Linked DataLora Aroyo
Ā 
Keynote at International Conference of Art Libraries 2018 @Rijksmuseum
Keynote at International Conference of Art Libraries 2018 @RijksmuseumKeynote at International Conference of Art Libraries 2018 @Rijksmuseum
Keynote at International Conference of Art Libraries 2018 @RijksmuseumLora Aroyo
Ā 
FAIRview: Responsible Video Summarization @NYCML'18
FAIRview: Responsible Video Summarization @NYCML'18FAIRview: Responsible Video Summarization @NYCML'18
FAIRview: Responsible Video Summarization @NYCML'18Lora Aroyo
Ā 
Understanding bias in video news & news filtering algorithms
Understanding bias in video news & news filtering algorithmsUnderstanding bias in video news & news filtering algorithms
Understanding bias in video news & news filtering algorithmsLora Aroyo
Ā 
StorySourcing: Telling Stories with Humans & Machines
StorySourcing: Telling Stories with Humans & MachinesStorySourcing: Telling Stories with Humans & Machines
StorySourcing: Telling Stories with Humans & MachinesLora Aroyo
Ā 
Data Science with Humans in the Loop
Data Science with Humans in the LoopData Science with Humans in the Loop
Data Science with Humans in the LoopLora Aroyo
Ā 
Digital Humanities Benelux 2017: Keynote Lora Aroyo
Digital Humanities Benelux 2017: Keynote Lora AroyoDigital Humanities Benelux 2017: Keynote Lora Aroyo
Digital Humanities Benelux 2017: Keynote Lora AroyoLora Aroyo
Ā 
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...Lora Aroyo
Ā 
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017Crowdsourcing ambiguity aware ground truth - collective intelligence 2017
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017Lora Aroyo
Ā 
My ESWC 2017 keynote: Disrupting the Semantic Comfort Zone
My ESWC 2017 keynote: Disrupting the Semantic Comfort ZoneMy ESWC 2017 keynote: Disrupting the Semantic Comfort Zone
My ESWC 2017 keynote: Disrupting the Semantic Comfort ZoneLora Aroyo
Ā 
Data Science with Human in the Loop @Faculty of Science #Leiden University
Data Science with Human in the Loop @Faculty of Science #Leiden UniversityData Science with Human in the Loop @Faculty of Science #Leiden University
Data Science with Human in the Loop @Faculty of Science #Leiden UniversityLora Aroyo
Ā 
SXSW2017 @NewDutchMedia Talk: Exploration is the New Search
SXSW2017 @NewDutchMedia Talk: Exploration is the New SearchSXSW2017 @NewDutchMedia Talk: Exploration is the New Search
SXSW2017 @NewDutchMedia Talk: Exploration is the New SearchLora Aroyo
Ā 
Europeana GA 2016: Harnessing Crowds, Niches & Professionals in the Digital Age
Europeana GA 2016: Harnessing Crowds, Niches & Professionals  in the Digital AgeEuropeana GA 2016: Harnessing Crowds, Niches & Professionals  in the Digital Age
Europeana GA 2016: Harnessing Crowds, Niches & Professionals in the Digital AgeLora Aroyo
Ā 
"Video Killed the Radio Star": From MTV to Snapchat
"Video Killed the Radio Star": From MTV to Snapchat"Video Killed the Radio Star": From MTV to Snapchat
"Video Killed the Radio Star": From MTV to SnapchatLora Aroyo
Ā 
UMAP 2016 Opening Ceremony
UMAP 2016 Opening CeremonyUMAP 2016 Opening Ceremony
UMAP 2016 Opening CeremonyLora Aroyo
Ā 
Crowdsourcing & Nichesourcing: Enriching Cultural Heritage with Experts & Cr...
Crowdsourcing & Nichesourcing: Enriching Cultural Heritagewith Experts & Cr...Crowdsourcing & Nichesourcing: Enriching Cultural Heritagewith Experts & Cr...
Crowdsourcing & Nichesourcing: Enriching Cultural Heritage with Experts & Cr...Lora Aroyo
Ā 
Stitch by Stitch: Annotating Fashion at the Rijksmuseum
Stitch by Stitch: Annotating Fashion at the RijksmuseumStitch by Stitch: Annotating Fashion at the Rijksmuseum
Stitch by Stitch: Annotating Fashion at the RijksmuseumLora Aroyo
Ā 

More from Lora Aroyo (20)

Data excellence: Better data for better AI
Data excellence: Better data for better AIData excellence: Better data for better AI
Data excellence: Better data for better AI
Ā 
CHIP Demonstrator presentation @ CATCH Symposium
CHIP Demonstrator presentation @ CATCH SymposiumCHIP Demonstrator presentation @ CATCH Symposium
CHIP Demonstrator presentation @ CATCH Symposium
Ā 
Semantic Web Challenge: CHIP Demonstrator
Semantic Web Challenge: CHIP DemonstratorSemantic Web Challenge: CHIP Demonstrator
Semantic Web Challenge: CHIP Demonstrator
Ā 
The Rijksmuseum Collection as Linked Data
The Rijksmuseum Collection as Linked DataThe Rijksmuseum Collection as Linked Data
The Rijksmuseum Collection as Linked Data
Ā 
Keynote at International Conference of Art Libraries 2018 @Rijksmuseum
Keynote at International Conference of Art Libraries 2018 @RijksmuseumKeynote at International Conference of Art Libraries 2018 @Rijksmuseum
Keynote at International Conference of Art Libraries 2018 @Rijksmuseum
Ā 
FAIRview: Responsible Video Summarization @NYCML'18
FAIRview: Responsible Video Summarization @NYCML'18FAIRview: Responsible Video Summarization @NYCML'18
FAIRview: Responsible Video Summarization @NYCML'18
Ā 
Understanding bias in video news & news filtering algorithms
Understanding bias in video news & news filtering algorithmsUnderstanding bias in video news & news filtering algorithms
Understanding bias in video news & news filtering algorithms
Ā 
StorySourcing: Telling Stories with Humans & Machines
StorySourcing: Telling Stories with Humans & MachinesStorySourcing: Telling Stories with Humans & Machines
StorySourcing: Telling Stories with Humans & Machines
Ā 
Data Science with Humans in the Loop
Data Science with Humans in the LoopData Science with Humans in the Loop
Data Science with Humans in the Loop
Ā 
Digital Humanities Benelux 2017: Keynote Lora Aroyo
Digital Humanities Benelux 2017: Keynote Lora AroyoDigital Humanities Benelux 2017: Keynote Lora Aroyo
Digital Humanities Benelux 2017: Keynote Lora Aroyo
Ā 
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...
Ā 
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017Crowdsourcing ambiguity aware ground truth - collective intelligence 2017
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017
Ā 
My ESWC 2017 keynote: Disrupting the Semantic Comfort Zone
My ESWC 2017 keynote: Disrupting the Semantic Comfort ZoneMy ESWC 2017 keynote: Disrupting the Semantic Comfort Zone
My ESWC 2017 keynote: Disrupting the Semantic Comfort Zone
Ā 
Data Science with Human in the Loop @Faculty of Science #Leiden University
Data Science with Human in the Loop @Faculty of Science #Leiden UniversityData Science with Human in the Loop @Faculty of Science #Leiden University
Data Science with Human in the Loop @Faculty of Science #Leiden University
Ā 
SXSW2017 @NewDutchMedia Talk: Exploration is the New Search
SXSW2017 @NewDutchMedia Talk: Exploration is the New SearchSXSW2017 @NewDutchMedia Talk: Exploration is the New Search
SXSW2017 @NewDutchMedia Talk: Exploration is the New Search
Ā 
Europeana GA 2016: Harnessing Crowds, Niches & Professionals in the Digital Age
Europeana GA 2016: Harnessing Crowds, Niches & Professionals  in the Digital AgeEuropeana GA 2016: Harnessing Crowds, Niches & Professionals  in the Digital Age
Europeana GA 2016: Harnessing Crowds, Niches & Professionals in the Digital Age
Ā 
"Video Killed the Radio Star": From MTV to Snapchat
"Video Killed the Radio Star": From MTV to Snapchat"Video Killed the Radio Star": From MTV to Snapchat
"Video Killed the Radio Star": From MTV to Snapchat
Ā 
UMAP 2016 Opening Ceremony
UMAP 2016 Opening CeremonyUMAP 2016 Opening Ceremony
UMAP 2016 Opening Ceremony
Ā 
Crowdsourcing & Nichesourcing: Enriching Cultural Heritage with Experts & Cr...
Crowdsourcing & Nichesourcing: Enriching Cultural Heritagewith Experts & Cr...Crowdsourcing & Nichesourcing: Enriching Cultural Heritagewith Experts & Cr...
Crowdsourcing & Nichesourcing: Enriching Cultural Heritage with Experts & Cr...
Ā 
Stitch by Stitch: Annotating Fashion at the Rijksmuseum
Stitch by Stitch: Annotating Fashion at the RijksmuseumStitch by Stitch: Annotating Fashion at the Rijksmuseum
Stitch by Stitch: Annotating Fashion at the Rijksmuseum
Ā 

Recently uploaded

08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking MenDelhi Call girls
Ā 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
Ā 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
Ā 
šŸ¬ The future of MySQL is Postgres šŸ˜
šŸ¬  The future of MySQL is Postgres   šŸ˜šŸ¬  The future of MySQL is Postgres   šŸ˜
šŸ¬ The future of MySQL is Postgres šŸ˜RTylerCroy
Ā 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
Ā 
What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?Antenna Manufacturer Coco
Ā 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
Ā 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptxHampshireHUG
Ā 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
Ā 
A Call to Action for Generative AI in 2024
A Call to Action for Generative AI in 2024A Call to Action for Generative AI in 2024
A Call to Action for Generative AI in 2024Results
Ā 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...Martijn de Jong
Ā 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processorsdebabhi2
Ā 
Scaling API-first ā€“ The story of a global engineering organization
Scaling API-first ā€“ The story of a global engineering organizationScaling API-first ā€“ The story of a global engineering organization
Scaling API-first ā€“ The story of a global engineering organizationRadu Cotescu
Ā 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
Ā 
Finology Group ā€“ Insurtech Innovation Award 2024
Finology Group ā€“ Insurtech Innovation Award 2024Finology Group ā€“ Insurtech Innovation Award 2024
Finology Group ā€“ Insurtech Innovation Award 2024The Digital Insurer
Ā 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking MenDelhi Call girls
Ā 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreternaman860154
Ā 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Enterprise Knowledge
Ā 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...apidays
Ā 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxMalak Abu Hammad
Ā 

Recently uploaded (20)

08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
Ā 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
Ā 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Ā 
šŸ¬ The future of MySQL is Postgres šŸ˜
šŸ¬  The future of MySQL is Postgres   šŸ˜šŸ¬  The future of MySQL is Postgres   šŸ˜
šŸ¬ The future of MySQL is Postgres šŸ˜
Ā 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
Ā 
What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?
Ā 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
Ā 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
Ā 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
Ā 
A Call to Action for Generative AI in 2024
A Call to Action for Generative AI in 2024A Call to Action for Generative AI in 2024
A Call to Action for Generative AI in 2024
Ā 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
Ā 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
Ā 
Scaling API-first ā€“ The story of a global engineering organization
Scaling API-first ā€“ The story of a global engineering organizationScaling API-first ā€“ The story of a global engineering organization
Scaling API-first ā€“ The story of a global engineering organization
Ā 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
Ā 
Finology Group ā€“ Insurtech Innovation Award 2024
Finology Group ā€“ Insurtech Innovation Award 2024Finology Group ā€“ Insurtech Innovation Award 2024
Finology Group ā€“ Insurtech Innovation Award 2024
Ā 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men
Ā 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
Ā 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...
Ā 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Ā 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptx
Ā 

Crowdsourcing & Citizen Science

  • 1. FROM CROWD KNOWLEDGE TO MACHINE KNOWLEDGE use cases with semantics & user interaction in cultural heritage collections Lora Aroyo VU University Amsterdam @laroyo
  • 2. OBSERVATION I: VIDEO ANNOTATIONS Video: rich in meaning expressed with objects, people, events & symbols Annotation: tedious, time- consuming, incomplete Professional annotations: coarse-grained & refer to entire video and topics
  • 3. OBSERVATION II: SEARCH BEHAVIOR People predominantly request video fragments: broadcast (33%), stories (17%), fragments (49%) Finding fragments takes much longer than ļ¬nding broadcasts: stories (2x), fragment (3x)
  • 4. OBSERVATION III: VOCABULARY MISMATCH experts use a speciļ¬c vocabulary that is unknown to general audiences 35% of clicked results are not found by title or thesaurus term
  • 5. OBSERVATION IV: LIMITED ANNOTATION RESOURCES the professionals can no longer cope with the demand for manual annotation on the ever growing multimedia content in their collections
  • 6. SO WE WANT TO ... Make a massive set of videos accessible to end users Improve video search for end users Maintain a growing community of engaged users Support professional annotators
  • 7. TO IMPROVE VIDEO SEARCH: fragment retrieval within video navigation
  • 8. REQUIRES CHANGES IN ANNOTATIONS: Including time-based annotations Bridging the vocabulary gap of searcher &. cataloguer
  • 9. exploit crowdsourcing on the web integrate in content management system
  • 10. CULTURAL HERITAGE WITH INVOLVED CONSUMERS & PROVIDERS data: open, shared & accessible infrastructures: interoperable smart services: relevant content in right context
  • 11. Rijksmuseum Amsterdam http://chip-project.org http://e-culture.multimedian.nl/ pk/annotate?uri=RP-P- OB-77.320
  • 12. Netherlands Institute for Sound & Vision http://openimages.eu http://academia.nl MUNCH project CHOICE project
  • 13. Europeana Cultural Search http://e-culture.multimedian.nl http://europeana.eu/portal/ thoughtlab.html http://europeana.eu
  • 14. WAYS TO USE THE CROWD Tagging & classiļ¬cation Editing & transcribing Contextualising Acquisition Co-curation Crowdfunding
  • 15. WHAT WE DID: THE WORKFLOW collect-analyse-manage-integrate user-generated metadata
  • 16.
  • 17. domain experts gamers collection managers
  • 18. Winner EuroITV Competition Best Archives on the Web Award COLLECT User Generated Metadata
  • 19. TWO PILOTS Waisda? 1.0 (8 months) focus group with lay users for motivation aspects cognitive walk through with professionals usability testing with lay users and professionals tag analysis - primarily WHAT & WHO Waisda? 2.0 (starting this September) currently running initial user & tag analysis
  • 20.
  • 21. Waisda? 1.0 (8 months) Participation ļƒ˜ 44,000 pageviews ļƒ˜2,000 different players ļƒ˜500 registered players ļƒ˜ thousands anonymous players Time-based annotations ļƒ˜ 612 videos tagged ļƒ˜ 420,000 tags added ļƒ˜ 46,000 unique tags Vocabulary gap User tags added the user perspective on the collection Community consensus as a measure of ā€˜validityā€™ Players score points when tag exactly matches a tag entered by another within 10 secs
  • 22.
  • 23. Waisda? 2.0 (4 weeks not public) Participation ļƒ˜ ~1500 users ļƒ˜ 81 registered ļƒ˜ 1435 anonymous users ļƒ˜ 2344 games Time-based annotations 11,109 videos in the game ļƒ˜ 32,200 tags --> 25,600 ļ¬rst time tags 322 were played with ļƒ˜ 19% are one time tags (81% appear more than once) ļƒ˜ 12,000 validated tags (36%) / (4000 of them ļ¬rst time tag) ļƒ˜ 1,900 match in vocabs - 257 GTAA people/83 validated; 1,661 GTAA geo/666 validated ļƒ˜ 9,796 validated, but no match in GTAA
  • 24.
  • 25.
  • 26.
  • 27.
  • 28.
  • 29. MANAGE Analyze Waisda? Collected Tags; Quality Metrics
  • 30. WHAT ARE THE RELATIONS BETWEEN TAGS & PROFESSIONAL ANNOTATIONS? in terms of the vocabulary used? in terms of the topics they describe? Initially: manual analysis Output goal: to be expressed in a quality metric
  • 31. WAISDA? TAG ANALYSIS Study 1 Quantitative analysis of all 46,762 Waisda? tags Tag coverage w.r.t different vocabularies Study 2 Qualitative analysis of selected video fragments Manual classiļ¬cation of 1,343 ā€œveriļ¬edā€ tags
  • 32. DURING FIRST 9 MONTHS 400000 300000 > 600 videos 13 TV series Numberā€©ofā€©tags 200000 100000 > 420,000 tags collected 0 46,792 unique tags 1 12 23 34 45 56 67 78 89 100 111 122 133 144 155 166 177 188 199 210 221 232 500 Day each top 5 most-tagged video has > 23,000 tags 375 Numberā€©ofā€©arrivals 250 avg. tag density > 8 tags/sec 125 0 1 11 21 31 41 51 61 71 81 91 101 111 121 131 141 151 161 171 181 191 201 211 221 231 Day
  • 33. 1 MAPPING TAGS TO VOCABULARIES Vocabulary of S&V GTAA 160 000 terms in six Thesaurus disjoint facets: keyword, to ap pe d location, maker, genre, m person, named entities mapped Dutch lexical database m ap pe d to collective vocabulary of internet users
  • 34. RESULTS Zero-hits (4404) 11% Google (41388) 89% Corneto (29815) (8723) 63% 18% (2216) 5% GTAA (1634) 3%
  • 35. RESULTS Zero-hits (4404) 11% Facet Tags Subject 1199 Google (41388) Location 613 89% Genre 52 Corneto (8723) (29815) Person 118 63% 18% (2216) Maker 4 5%GTAA GTAA Name 673 (3850) (1634) 8%3%
  • 36. RESULTS Zero-hits (4404) 11% Google (41388) 89% Corneto (29815) (8723) 63% 18% (2216) 5% GTAA (1634) 3%
  • 37. RESULTS Zero-hits (4404) 11% Types Tags Noun 7222 Google (41388) 89% Verb 2090 Corneto Adjective 1693 Corneto (8723) (29815) (10939) 63% 18% 23%(2216) Adverb 171 5% GTAA (1634) 3%
  • 38. RESULTS Zero-hits (4404) 11% Google (41388) 89% Corneto (29815) (8723) 63% 18% (2216) 5% GTAA (1634) 3%
  • 39. RESULTS Zero-hits Zero-hits (4404) (4404) 11% 11% Zero-hit tags Google ā€¢ā€ˆgarbled text (41388) 89% ā€¢ā€ˆseriously mistyped words Corneto ā€¢ā€ˆgrammatically incorrect (29815) (8723) 18% 63% sentences (2216) 5% GTAA (1634) 3%
  • 40. RESULTS Zero-hits (4404) 11% Google (41388) 89% Corneto (29815) (8723) 63% 18% (2216) 5% GTAA (1634) 3%
  • 41. RESULTS Zero-hits (4404) 11% Google ā€œMeaningfulā€ Tags ā€¢ā€ˆmulti-word tags Google (41388) ā€¢ā€ˆslang 89% ā€¢ā€ˆnames Corneto Pos-hits (8723) (29815) (29815) 63% ā€¢ā€ˆmorphological variations 18% 63% (2216) ā€¢ā€ˆminor typos 5% GTAA ā€¢ā€ˆetcā€¦ (1634) 3%
  • 42. RESULTS Zero-hits (4404) 11% Google (41388) 89% Corneto (29815) (8723) 63% 18% (2216) 5% GTAA (1634) 3%
  • 43. 2 TAG CLASSIFICATION What video aspects are described by user tags? Manual classiļ¬cation of tag sample 5 videos non-visual (0), perceptual (11), the rest conceptual Object (most of the tags) vs. Scene tags (only 30) 3 levels and 4 facets: abstract, general & speciļ¬c who, what, when, where Tag sample: all veriļ¬ed tags from random fragments (182 non-descriptive tags removed) #1,343
  • 44. 2 RESULTS Only 30 scene-level tags Panofsky-Shatford matrix for object-level tags Abstract General Speciļ¬c Total Who 10 166 177 31% What 73 5 563 12 57% Where 0 68 8 7% When 4 31 6 5% Total 7% 74% 9% 195 tags (typically adverbs & adjectives) couldnā€™t be classiļ¬ed in any of the facets [1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
  • 45. 2 RESULTS Only 30 scene-level tags Panofsky-Shatford matrix for object-level tags Abstract General Speciļ¬c Total Who 10 166 177 31% What 73 5 563 12 57% Where 0 68 8 7% When 4 31 6 5% Total 7% 74% 74% 9% 195 tags (typically adverbs & adjectives) couldnā€™t be classiļ¬ed in any of the facets [1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
  • 46. 2 RESULTS Only 30 scene-level tags Panofsky-Shatford matrix for object-level tags Abstract General Speciļ¬c Total Who 10 166 177 31% What 73 5 563 12 57% Where 0 68 8 7% When 4 31 6 5% Total 7% 74% 9% 195 tags (typically adverbs & adjectives) couldnā€™t be classiļ¬ed in any of the facets [1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
  • 47. 2 RESULTS Only 30 scene-level tags Panofsky-Shatford matrix for object-level tags Abstract General Speciļ¬c Total Who 10 166 177 31% 31% What 73 5 563 12 57% 57% Where 0 68 8 7% When 4 31 6 5% Total 7% 74% 9% 195 tags (typically adverbs & adjectives) couldnā€™t be classiļ¬ed in any of the facets [1] Classification of user image descriptions. L. Hollink, G. Schreiber, B. Wielinga & M. Worring
  • 48. EXPERIMENTAL SET 5 videos Episodes All tags Veriļ¬ed Genre well tagged Reality show 25,965 5837 amusement well tagged Reality show 22,792 6153 amusement medium Missing people 1007 274 informative low tagged Reporter 403 73 informative low tagged The Walk 257 45 religious
  • 49. genre might inļ¬‚uence the usefulness of the user tags users focus on what can be seen or heard (directly perceivable objects) because of the fast pace of the game USEFULNESS according to professionals
  • 50. SUBTITLES USEFUL? on the average 26% of all the tags (35% of the veriļ¬ed) are matched in subtitles used to run bot-based games otherwise, donā€™t really bring much new to the annotation
  • 51. OBSERVATIONS FROM ANALYSIS: taggers & professionals use different vocabularies user tags complement professional annotations user tags could bridge the searcher vs. cataloger gap taggers tend to use general concepts more taggers focus mainly on the What and Who taggers describe predominantly objects & rarely scenes
  • 52. OBSERVATIONS FROM ANALYSIS: need for validation for quality & correctness crowd-sourcing users need continuous support & motivation to supply more and better contributions Ā 
  • 53. THE QUESTION IS: Can user tags be used for fragment search? How to derive user tag quality metrics?
  • 54. THE QUALITY OF TAGS?
  • 56. Tag gardening annotations for video, fragments & scenes adding types & roles to user tags mapping tags to concepts, disambiguation spelling corrections single & batch tags/video editing
  • 57.
  • 58.
  • 59.
  • 60.
  • 61.
  • 62. sorted by frequency all frames per tag
  • 63.
  • 64.
  • 65.
  • 66.
  • 67.
  • 68.
  • 69. Inspired by Google Reļ¬ne
  • 70. INITIAL (SMALL) EXPERIMENT ā€¢1 video, 36 tags, 4 validators ā€¢ validatorsselects the most appropriate concept for each tag (if present) ā€¢ frames within the video available - linked to the tags ā€¢ reconciliation agains GTAA, Cornetto, Freebase ā€¢ (1) select a source, (2) start reconciliation (3) choose from suggested concepts
  • 71.
  • 72. krippendorff coverage 33 out of 36 tags reliability coefļ¬cient Number of tags reconciled by the four participants more than 50% of concepts Ranks of the selected tags
  • 73. Lessons ā€¢ Cornetto is good for subject terms ā€¢ Selecting most suitable concept is time consuming as the differences sometimes are difļ¬cult to understand ā€¢ For all NE users could select a concept from Freebase & GTAA ā€¢ Vocabularies like GTAA have less additional information, thus itā€™s more difļ¬cult to select the right concept ā€¢ Freebase works well for English (Dutch might be a problem) ā€¢ Full coverage - prevented by spelling errors and variations ā€¢ Users adapt quickly, e.g. Freebase is good for people & locations
  • 74. Recommendations ā€¢ reconcile agains multiple data sources at the same time for best coverage ā€¢ merge duplicates into a single suggestion ā€¢ organize results in categories, e.g. people, locations and subjects (to avoid suggestions from the wrong type) ā€¢ pre-processing to merge similar concepts (deal with spelling errors and variations) ā€¢ rank the suggested concepts by precision, then by recall (as alternative)
  • 80. WITHIN VIDEO NAVIGATION group by type (faceted)
  • 82. WITHIN VIDEO NAVIGATION concepts uniquely identiļ¬ed
  • 84. WITHIN VIDEO NAVIGATION background info
  • 85. THE TEAM Riste Gligorov @mbrinkerink @GuusSchreiber Marteen Brinkerink Guus Schreiber @johanoomen Johan Oomen @laroyo Lora Aroyo @McHildebrand @jrvosse Michiel Hildebrand @lottebelice Jacco van Ossenbruggen Lotte Belice Baltussen
  • 86. LINKS http://waisda.nl http://www.prestoprime.org/ @waisda http://www.cs.vu.nl/~laroyo http://slideshare.com/laroyo

Editor's Notes

  1. From the different research topics I am involved in today I address some of the results from projects in which we use explicit semantics and user interaction to make crowd knowledge effectively processable for machines\n\n\n*****\nFrom Crowd Knowledge to Machine Knowledge: Use cases with semantics and user interaction in Dutch cultural heritage collections\n\nIn this talk I will discuss several projects, for example with the Dutch national archive for sound and vision or with the Rijksmuseum Amsterdam, where we have experimented with semantics-based technologies and user interaction paradigms to provide systems with additional support for users to turn their lay knowledge into machine-processable knowledge. Turning this crowd knowledge into machine knowledge makes the system more intelligent and thus makes them capitalize on the knowledge assets in the crowds of users.\n\nThe problem our system addresses is concerned with making the massive set of interesting multimedia content available in these cultural heritage institutions accessible for a large community of users.  On the one hand, this content is difficult to find for the average user as they have been indexed by experts (curators, art historians, etc) who use a very specific vocabulary that is unknown to the general audience.  On the other hand, these professionals can no longer  cope with the demand for annotation on the ever growing multimedia content in these collections.\n\nOur solution to both these problem is to exploit crowdsourcing on the web, where there is a lot of specific domain knowledge and processing power available that such archives and museums would gladly incorporate. However, turning this knowledge from the mass of lay users into additional intelligence in content management systems is not a simple challenge for several reasons: (1) the lay users use different vocabularies and are also interested in different aspects of the collection items; (2) the collected user metadata needs to be validated for quality and correctness; (3) crowd-sourcing users need to be continuously supported and motivated in order to supply more and better knowledge in such systems.  I will present our data analysis of the crowdsourced content, our techniques for quality control and incorporating domain semantics, and the results of our attempts to employ different interaction strategies and user interfaces (for example games and simplified interfaces for interactive annotation) to engage and stimulate the users.\n\nDemos:\n-----------\nVideo-tagging game: http://woordentikkertje.manbijthond.nl/\nArt recommender and personalized museum tours: http://www.chip-project.org/demo/ (3rd prize winner of ISWC Semantic Web challenge)\nHistorical events in museum collections: http://agora.cs.vu.nl/demo/\n\nRelevant Papers:\n------------------------\n- On the role of user-generated metadata in audio visual collections, at K-CAP2011\nhttp://portal.acm.org/citation.cfm?id=1999702\n\n- Digital Hermeneutics:  Agora and the Online Understanding of Cultural Heritage, at WebSci2011, (Best paper nominee)\nhttp://www.websci11.org/fileadmin/websci/Papers/116_paper.pdf\n\n- The effects of transparency on trust in and acceptance of a content-based art recommender, UMUAI journal, (Best journal paper for 2008)\nhttp://www.springerlink.com/content/81q34u73mpp58u75/\n\n- Recommendations based on semantically enriched museum collections, Journal of Web Semantics\nhttp://www.sciencedirect.com/science/article/pii/S1570826808000681\n\n- Enhancing Content-Based Recommendation with the Task Model of Classification, at EKAW2010\nhttp://www.springerlink.com/content/p78hl5r283x79r13/\n\nShort bio:\n-----------------\nLora Aroyo is an associate professor at the Web and Media group, at the Department of Computer Science, Free University Amsterdam, The Netherlands. Her research interests are in using semantic web technologies for modeling user interests and context, recommendation systems and personalized access in Web-based applications. Typical example domains are cultural heritage collections, multimedia archives and interactive TV. She has coordinated the research work in the CHIP project on Cultural Heritage Information Personalization (http://chip-project.org). Currently she is a scientific coordinator of the EU Integrated Project NoTube dealing with the integration of Web and TV data with the help of semantics (http://notube.tv), a project leader of VU INTERTAIN Experimental Research Lab initiative (http://www.cs.vu.nl/intertain), and involved in the research on motivational user interaction for video-tagging games in the PrestoPrime project (http://www.prestoprime.org/) and modeling\nhistoric events in the Agora project (http://agora.cs.vu.nl/). She has organized numerous workshops in the areas of personalized access to cultural heritage, e-learning, interactive television, as well as on visual interfaces to the social and semantic web (PATCH, FutureTV, PersWeb, VISSW and DeRIVE). Lora has been actively involved in both the Semantic Web (PC co-chair and conference chair for ESWC2009 and ESWC2010 and PC co-chair for ISWC2011) and the Personalization and User modeling communities (on the editorial board for the UMUAI journal and on the steering committee of UMAP conference).\n\nMore information can be found at:\n-----------------------------------------------\nWebpage: http://www.cs.vu.nl/~laroyo\nSlideshare: http://www.slideshare.net/laroyo\nTwitter: @laroyo\n\n
  2. Nowadays av collections are undergoing process of transformation from archives of analog material to large digital (online) data stores, as videos are very much wanted by different types of end users. \n\nFor example, the Netherlands Institute of Sound and Vision archives all radio and TV material broadcasted in the Netherlands (has appr. 700,000 hours radio and television programs available online. \n\nFacilitating a successfully access to av collection items demands quality metadata associated with them.\n\nTraditionally, in AV achives it is the task of professional catalogers to manually describe the videos. Usually, in the process they follow well-defined , well-established guidelines and rules. They also may make use of auxiliary materials like controlled vocabularies, thesauri, and such.\nHowever, as we all know video is medium that is extremely rich in meaning. Directors and screenwriters create entire universes with complex interplay between characters, objects and events. Sometimes they may employ rich and complex abstract symbolic language. This makes that task of describing the meaning of a video as complicated as describing the real worlds. Which is no trivial matter.\nAs a result the process of annotation is tedious, time-consuming and inevitably incomplete. According to some research, it takes approximately 5 times of the duration of the material to annotate it completely. So for example, if we are talking about a documentary that lasts one hour, it will take approximately 5 hours for a cataloger to fully describe it. Furthermore,\n\nConsequently, professional annotations are coarse-grained in a sense that they are referring to the entire video describing prevalent topics. It may happen that catalogers provide more fine-grained, shot-level descriptions for a video. But this is exception of the rule and it is reserved to the most important pieces of the AV collection.\n
  3. \n* Search Behavior of Media Professionals at an Audiovisual Archive: A Transaction Log Analysis\nB. Huurnink, L. Hollink, W. van den Heuvel, and M. de Rijke.\n\n
  4. \n
  5. The problem our system addresses is concerned with making the massive set of interesting multimedia content available in these cultural heritage institutions accessible for a large community of users.  On the one hand, this content is difficult to find for the average user as they have been indexed by experts (curators, art historians, etc) who use a very specific vocabulary that is unknown to the general audience.  On the other hand, these professionals can no longer  cope with the demand for annotation on the ever growing multimedia content in these collections.\n
  6. The problem our system addresses is concerned with making the massive set of interesting multimedia content available in these cultural heritage institutions accessible for a large community of users.  On the one hand, this content is difficult to find for the average user as they have been indexed by experts (curators, art historians, etc) who use a very specific vocabulary that is unknown to the general audience.  On the other hand, these professionals can no longer  cope with the demand for annotation on the ever growing multimedia content in these collections.\n\nRecommendation of video fragments\nLinks between video fragments\nLinks to background information\nintegrate the Goal with the general --> merge the two slides --> one slide for RMA and Europeana\n\n
  7. The problem our system addresses is concerned with making the massive set of interesting multimedia content available in these cultural heritage institutions accessible for a large community of users.  On the one hand, this content is difficult to find for the average user as they have been indexed by experts (curators, art historians, etc) who use a very specific vocabulary that is unknown to the general audience.  On the other hand, these professionals can no longer  cope with the demand for annotation on the ever growing multimedia content in these collections.\n\nRecommendation of video fragments\nLinks between video fragments\nLinks to background information\nintegrate the Goal with the general --> merge the two slides --> one slide for RMA and Europeana\n\n
  8. The problem our system addresses is concerned with making the massive set of interesting multimedia content available in these cultural heritage institutions accessible for a large community of users.  On the one hand, this content is difficult to find for the average user as they have been indexed by experts (curators, art historians, etc) who use a very specific vocabulary that is unknown to the general audience.  On the other hand, these professionals can no longer  cope with the demand for annotation on the ever growing multimedia content in these collections.\n\nRecommendation of video fragments\nLinks between video fragments\nLinks to background information\nintegrate the Goal with the general --> merge the two slides --> one slide for RMA and Europeana\n\n
  9. Our solution to both these problem is to exploit crowdsourcing on the web, where there is a lot of specific domain knowledge and processing power available that such archives and museums would gladly incorporate. However, turning this knowledge from the mass of lay users into additional intelligence in content management systems is not a simple challenge for several reasons: (1) the lay users use different vocabularies and are also interested in different aspects of the collection items; (2) the collected user metadata needs to be validated for quality and correctness; (3) crowd-sourcing users need to be continuously supported and motivated in order to supply more and better knowledge in such systems.  I will present our data analysis of the crowdsourced content, our techniques for quality control and incorporating domain semantics, and the results of our attempts to employ different interaction strategies and user interfaces (for example games and simplified interfaces for interactive annotation) to engage and stimulate the users.\n\nunderstanding the user-generated data\ncontextualize the user-generated metadata \n\n\n
  10. supplement video data with background knowledge\n introduce the context, rijksmuseum\n the general picture, examples, CH landscape\n \n \n
  11. \n
  12. \n
  13. \n
  14. We hebben naar uiteenlopende voorbeelden gekeken binnen de sector en deze geclusterd\n\ncriterium - voorname/leidende rol erfgoed. Uiteraard zijn er ook veel grassroots initiatieven. \n\nEen van mijn favoriete voorbeelden is de International Amateur Scanning league.\nDigitaliseren van historische video’s in het publieke domein.\n\n\nwe komen vervolgens op een indeling in zes categorieen. \n\nwe hebben al goede voorbeelden gezien vandaag - ik zal met name voorbeelden noemen uit categorieen die niet in andere presentaties aan bod komen.\n\n
  15. \n
  16. Wikipedia as a compelling example --> calculating that creating Wikipedia as it stands today has taken one hundred\nmillion hours of cumulative thought, he juxtaposes this to the astounding 200 billion hours people watch TV in the\nUS alone. 200 billion hours would amount to two thousand Wikipedia projects-worth of free time, annually. \n\n
  17. \n
  18. three types of users\ncollect: gamers\nmanage: domain experts\nintegrate: collection maintainer\n
  19. explain games with purpose\n\nplay on words\n\n
  20. \n
  21. multi-player game \n 4 channels each continuously streaming videos from a predefined category\n a player starts a game by selecting one of the channels\n plays agains all the players who selected the same channel\n\n
  22. multi-player game \n users score points by entering tags (based on the principle of Louis van Ahn’s ESP games\n points are scored when two or more players enter the same tag\n as in the Yahoo! video tag game --> this is done in a period of 10 sec\n each tag is shown together with the score it has earned\n\nThe relevance of the tags is ensured by two factors: independence of players and consensus among players. The players don’t know each other, they are grouped randomly, and the only thing they share at the moment when they are playing the game is the video. So if they enter the same tag within a time frame the changes are the tag is relevant for the video.\n
  23. the changes:\n- shorter videos (~3 mins), independent fragments vs. fragmented videos\n- improved scoring (including matches in Cornetto and GTAA; first time tag)\n- game recap with continuous scoring\n- profile page\n\nbots again\n- different handling of users to start playing the game\n\n
  24. total # tags 32,248 (25,685 first time tags)\n 19% are one time tags (81% appear more than once)\n \n 257 person names in GTAA, 83 also validated\n 1,661 geo names in GTAA, 666 also validated\n 9,796 validated, but no match in GTAA\n\n
  25. \n
  26. \n
  27. \n
  28. First step manual\n\nGoal is to automate it and make it an integral part of Waisda\n\nEventually the output can be expressed in a quality metric\n
  29. \n\n
  30. How/why did you select those fragments\nHow did you do the classifacation\n
  31. OK some statistics. These stats are computed on the dataset that was accumulated by waisda? in the first six months after it went live \n
  32. The goal of our first data analysis is to investigate user terminology and compare it to the terminology used by professionals. We de\n
  33. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  34. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  35. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  36. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  37. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  38. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  39. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  40. Google search - a tag is meaningful if the number of pages returned by Google is positive\nFor about 89% of the tags - not found in vocabs or not verified --> google returned positive # of hits\n200 Sample of the ‘no-hits’ (zero sample)\n200 Sample of the ‘hits’ (pos sample)\n\nThings heard or things seen on the screen --> subtitles (for heard)\n\nComparable with Steve.museum\n
  41. non-visual --> 0\nperceptual --> 11 tags (referring to color)\nconceptual --> the rest (describing objects, 30 about scenes only)\n
  42. 30% of scenes are conceptual (by Laura Hollink)\n\n
  43. 30% of scenes are conceptual (by Laura Hollink)\n\n
  44. 30% of scenes are conceptual (by Laura Hollink)\n\n
  45. \n
  46. genre is influencing the specificity and usefulness\nusers focus on what can be seen or heard\nusers focus on directly perceivable objects - because of the pace\n\n
  47. \n
  48. Our solution to both these problem is to exploit crowdsourcing on the web, where there is a lot of specific domain knowledge and processing power available that such archives and museums would gladly incorporate. However, turning this knowledge from the mass of lay users into additional intelligence in content management systems is not a simple challenge for several reasons: (1) the lay users use different vocabularies and are also interested in different aspects of the collection items; (2) the collected user metadata needs to be validated for quality and correctness; (3) crowd-sourcing users need to be continuously supported and motivated in order to supply more and better knowledge in such systems.  I will present our data analysis of the crowdsourced content, our techniques for quality control and incorporating domain semantics, and the results of our attempts to employ different interaction strategies and user interfaces (for example games and simplified interfaces for interactive annotation) to engage and stimulate the users.\n\n
  49. Our solution to both these problem is to exploit crowdsourcing on the web, where there is a lot of specific domain knowledge and processing power available that such archives and museums would gladly incorporate. However, turning this knowledge from the mass of lay users into additional intelligence in content management systems is not a simple challenge for several reasons: (1) the lay users use different vocabularies and are also interested in different aspects of the collection items; (2) the collected user metadata needs to be validated for quality and correctness; (3) crowd-sourcing users need to be continuously supported and motivated in order to supply more and better knowledge in such systems.  I will present our data analysis of the crowdsourced content, our techniques for quality control and incorporating domain semantics, and the results of our attempts to employ different interaction strategies and user interfaces (for example games and simplified interfaces for interactive annotation) to engage and stimulate the users.\n\n
  50. Our solution to both these problem is to exploit crowdsourcing on the web, where there is a lot of specific domain knowledge and processing power available that such archives and museums would gladly incorporate. However, turning this knowledge from the mass of lay users into additional intelligence in content management systems is not a simple challenge for several reasons: (1) the lay users use different vocabularies and are also interested in different aspects of the collection items; (2) the collected user metadata needs to be validated for quality and correctness; (3) crowd-sourcing users need to be continuously supported and motivated in order to supply more and better knowledge in such systems.  I will present our data analysis of the crowdsourced content, our techniques for quality control and incorporating domain semantics, and the results of our attempts to employ different interaction strategies and user interfaces (for example games and simplified interfaces for interactive annotation) to engage and stimulate the users.\n\n\nCan user tags be used to derive topical descriptions for scenes?\n\n
  51. \n
  52. \n
  53. Bridge communities \n crowd\n professionals\n Combine perspectives\n (subjective) user perspective\n (objective) professional perspective\n Link tags to concepts\n Annotations of video scenes \n Quality assurance by moderation\n
  54. \n
  55. \n
  56. \n
  57. frames per tag!\nview all frames for a tag\nsorted by frequency\nFrequently occurring tags for scene descriptions\n
  58. Selecting a role\n
  59. Selecting a role\n
  60. Link tags to concepts\n
  61. not always easy,\nwordnet: polysemi, interpretations are close\nsome vocabularies have no description\n\nhow do tags relate to a shot\n- the frame is often a bit later\n
  62. \n
  63. \n
  64. publicly available reconciliation APIs: e.g. Freebase, Europeana, Talis (kasabi.com)\n\n
  65. \n
  66. krippendorff alpha\n\nWe conclude that for this small experiment, with a coverage of 33 out of 36 tags and a total reliability coefficient of 0.91, users can effectively reconcile tags using the selected services.\n\nMore than half of the selected concepts were ranked first (32), while 23 concepts were ranked second or lower. Several (6) selected concepts were even ranked sixth or lower\nThus, often the first concept could be automatically selected, but not always\n\n\n
  67. \n
  68. \n
  69. \n
  70. \n
  71. \n
  72. \n
  73. Open Archives Initiative Object Reuse and Exchange (OAI-ORE) defines standards for the description and exchange of aggregations of Web resources. The goal of these standards is to expose the rich content in these aggregations to applications that support authoring, deposit, exchange, visualization, reuse, and preservation. \n
  74. \n
  75. \n
  76. \n
  77. - group by type (faceted)\n- concept is uniquely identified (see full name for Juliana and Bernhard)\n- background information\n
  78. - group by type (faceted)\n- concept is uniquely identified (see full name for Juliana and Bernhard)\n- background information\n
  79. - group by type (faceted)\n- concept is uniquely identified (see full name for Juliana and Bernhard)\n- background information\n
  80. - group by type (faceted)\n- concept is uniquely identified (see full name for Juliana and Bernhard)\n- background information\n
  81. - group by type (faceted)\n- concept is uniquely identified (see full name for Juliana and Bernhard)\n- background information\n
  82. \n
  83. \n