Скачать презентацию Lecture 10 Metadata for Media SIMS 202 Information Скачать презентацию Lecture 10 Metadata for Media SIMS 202 Information

1c5fc937536517595e058ee9d902cb9d.ppt

  • Количество слайдов: 62

Lecture 10: Metadata for Media SIMS 202: Information Organization and Retrieval Prof. Ray Larson Lecture 10: Metadata for Media SIMS 202: Information Organization and Retrieval Prof. Ray Larson & Prof. Marc Davis UC Berkeley SIMS Tuesday and Thursday 10: 30 am - 12: 00 pm Fall 2003 http: //www. sims. berkeley. edu/academics/courses/is 202/f 03/ IS 202 – FALL 2003. 09. 23 - SLIDE 1

Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Video – Current Approaches – Media Streams • Discussion Questions • Action Items for Next Time IS 202 – FALL 2003. 09. 23 - SLIDE 2

Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Video – Current Approaches – Media Streams • Discussion Questions • Action Items for Next Time IS 202 – FALL 2003. 09. 23 - SLIDE 3

The Media Opportunity • Vastly more media will be produced • Without ways to The Media Opportunity • Vastly more media will be produced • Without ways to manage it (metadata creation and use) we lose the advantages of digital media • Most current approaches are insufficient and perhaps misguided • Great opportunity for innovation and invention • Need interdisciplinary approaches to the problem IS 202 – FALL 2003. 09. 23 - SLIDE 4

What is the Problem? • Today people cannot easily find, edit, share, and reuse What is the Problem? • Today people cannot easily find, edit, share, and reuse media • Computers don’t understand media content – Media is opaque and data rich – We lack structured representations • Without content representation (metadata), manipulating digital media will remain like wordprocessing with bitmaps IS 202 – FALL 2003. 09. 23 - SLIDE 5

Traditional Media Production Chain Metadata-Centric Production Chain METADATA PRE-PRODUCTION IS 202 – FALL 2003 Traditional Media Production Chain Metadata-Centric Production Chain METADATA PRE-PRODUCTION IS 202 – FALL 2003 PRODUCTION POST-PRODUCTION METADATA DISTRIBUTION 2003. 09. 23 - SLIDE 6

Automated Media Production Process 1 Active Capture 3 Automatic Editing Adaptive Media Engine 4 Automated Media Production Process 1 Active Capture 3 Automatic Editing Adaptive Media Engine 4 Personalized/ Customized Delivery Web Integration and Streaming Media Services Flash Generator HTML Email Annotation of Media Assets WAP 2 Annotation and Retrieval Asset Retrieval and Reuse Print/Physical Media Reusable Online Asset Database IS 202 – FALL 2003. 09. 23 - SLIDE 7

Technology Summary • Media Streams provides a framework for creating metadata throughout the media Technology Summary • Media Streams provides a framework for creating metadata throughout the media production cycle to make media assets searchable and reusable • Active Capture automates direction and cinematography using real-time audio-video analysis in an interactive control loop to create reusable media assets • Adaptive Media uses adaptive media templates and automatic editing functions to mass customize and personalize media and thereby eliminate the need for editing on the part of end users • Together, these technologies will automate, personalize, and speed up media production, distribution, and reuse IS 202 – FALL 2003. 09. 23 - SLIDE 8

Active Capture IS 202 – FALL 2003. 09. 23 - SLIDE 9 Active Capture IS 202 – FALL 2003. 09. 23 - SLIDE 9

Active Capture: Reusable Shots IS 202 – FALL 2003. 09. 23 - SLIDE 10 Active Capture: Reusable Shots IS 202 – FALL 2003. 09. 23 - SLIDE 10

Marc Davis in Godzilla Scene IS 202 – FALL 2003. 09. 23 - SLIDE Marc Davis in Godzilla Scene IS 202 – FALL 2003. 09. 23 - SLIDE 11

Evolution of Media Production • Customized production – Skilled creation of one media product Evolution of Media Production • Customized production – Skilled creation of one media product • Mass production – Automatic replication of one media product • Mass customization – Skilled creation of adaptive media templates – Automatic production of customized media IS 202 – FALL 2003. 09. 23 - SLIDE 12

Central Idea: Movies as Programs • Movies change from being static data to programs Central Idea: Movies as Programs • Movies change from being static data to programs Media Parser Content Representation Producer Media Parser Media Content Representation • Shots are inputs to a program that computes new media based on content representation and functional dependency (US Patents 6, 243, 087 & 5, 969, 716) IS 202 – FALL 2003. 09. 23 - SLIDE 13

Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Video – Current Approaches – Media Streams • Discussion Questions • Action Items for Next Time IS 202 – FALL 2003. 09. 23 - SLIDE 14

Representing Video • Streams vs. Clips • Video syntax and semantics • Ontological issues Representing Video • Streams vs. Clips • Video syntax and semantics • Ontological issues in video representation IS 202 – FALL 2003. 09. 23 - SLIDE 15

Video is Temporal IS 202 – FALL 2003. 09. 23 - SLIDE 16 Video is Temporal IS 202 – FALL 2003. 09. 23 - SLIDE 16

Streams vs. Clips IS 202 – FALL 2003. 09. 23 - SLIDE 17 Streams vs. Clips IS 202 – FALL 2003. 09. 23 - SLIDE 17

Stream-Based Representation • Makes annotation pay off – The richer the annotation, the more Stream-Based Representation • Makes annotation pay off – The richer the annotation, the more numerous the possible segmentations of the video stream • Clips – Change from being fixed segmentations of the video stream, to being the results of retrieval queries based on annotations of the video stream • Annotations – Create representations which make clips, not representations of clips IS 202 – FALL 2003. 09. 23 - SLIDE 18

Video Syntax and Semantics • The Kuleshov Effect • Video has a dual semantics Video Syntax and Semantics • The Kuleshov Effect • Video has a dual semantics – Sequence-independent invariant semantics of shots – Sequence-dependent variable semantics of shots IS 202 – FALL 2003. 09. 23 - SLIDE 19

Ontological Issues for Video • Video plays with rules for identity and continuity – Ontological Issues for Video • Video plays with rules for identity and continuity – Space – Time – Person – Action IS 202 – FALL 2003. 09. 23 - SLIDE 20

Space and Time: Actual vs. Inferable • Actual Recorded Space and Time – GPS Space and Time: Actual vs. Inferable • Actual Recorded Space and Time – GPS – Studio space and time • Inferable Space and Time – Establishing shots – Cues and clues IS 202 – FALL 2003. 09. 23 - SLIDE 21

Time: Temporal Durations • Story (Fabula) Duration – Example: Brushing teeth in story world Time: Temporal Durations • Story (Fabula) Duration – Example: Brushing teeth in story world (5 minutes) • Plot (Syuzhet) Duration – Example: Brushing teeth in plot world (1 minute: 6 steps of 10 seconds each) • Screen Duration – Example: Brushing teeth (10 seconds: 2 shots of 5 seconds each) IS 202 – FALL 2003. 09. 23 - SLIDE 22

Character and Continuity • Identity of character is constructed through – Continuity of actor Character and Continuity • Identity of character is constructed through – Continuity of actor – Continuity of role • Alternative continuities – Continuity of actor only – Continuity of role only IS 202 – FALL 2003. 09. 23 - SLIDE 23

Representing Action • Physically-based description for sequenceindependent action semantics – Abstract vs. conventionalized descriptions Representing Action • Physically-based description for sequenceindependent action semantics – Abstract vs. conventionalized descriptions – Temporally and spatially decomposable actions and subactions • Issues in describing sequence-dependent action semantics – Mental states (emotions vs. expressions) – Cultural differences (e. g. , bowing vs. greeting) IS 202 – FALL 2003. 09. 23 - SLIDE 24

“Cinematic” Actions • Cinematic actions support the basic narrative structure of cinema – Reactions/Proactions “Cinematic” Actions • Cinematic actions support the basic narrative structure of cinema – Reactions/Proactions • Nodding, screaming, laughing, etc. – Focus of Attention • Gazing, headturning, pointing, etc. – Locomotion • Walking, running, etc. • Cinematic actions can occur • Within the frame/shot boundary • Across the frame boundary • Across shot boundaries IS 202 – FALL 2003. 09. 23 - SLIDE 25

Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Video – Current Approaches – Media Streams • Discussion Questions • Action Items for Next Time IS 202 – FALL 2003. 09. 23 - SLIDE 26

The Search for Solutions • Current approaches to creating metadata don’t work – Signal-based The Search for Solutions • Current approaches to creating metadata don’t work – Signal-based analysis – Keywords – Natural language • Need standardized metadata framework – – Designed for video and rich media data Human and machine readable and writable Standardized and scaleable Integrated into media capture, archiving, editing, distribution, and reuse IS 202 – FALL 2003. 09. 23 - SLIDE 27

Signal-Based Parsing • Practical problem – Parsing unstructured, unknown video is very, very hard Signal-Based Parsing • Practical problem – Parsing unstructured, unknown video is very, very hard • Theoretical problem – Mismatch between percepts and concepts IS 202 – FALL 2003. 09. 23 - SLIDE 28

Perceptual/Conceptual Issue Similar Percepts / Dissimilar Concepts Clown Nose IS 202 – FALL 2003 Perceptual/Conceptual Issue Similar Percepts / Dissimilar Concepts Clown Nose IS 202 – FALL 2003 Red Sun 2003. 09. 23 - SLIDE 29

Perceptual/Conceptual Issue Dissimilar Percepts / Similar Concepts John Dillinger’s Car IS 202 – FALL Perceptual/Conceptual Issue Dissimilar Percepts / Similar Concepts John Dillinger’s Car IS 202 – FALL 2003 Timothy Mc. Veigh’s Car 2003. 09. 23 - SLIDE 30

Signal-Based Parsing • Effective and useful automatic parsing – Video • • • Shot Signal-Based Parsing • Effective and useful automatic parsing – Video • • • Shot boundary detection Camera motion analysis Low level visual similarity Feature tracking Face detection – Audio • • Pause detection Audio pattern matching Simple speech recognition Speech vs. music detection IS 202 – FALL 2003 • Approaches to automated parsing – At the point of capture, integrate the recording device, the environment, and agents in the environment into an interactive system – After capture, use “humanin-the-loop” algorithms to leverage human and machine intelligence 2003. 09. 23 - SLIDE 31

Keywords vs. Semantic Descriptors dog, biting, Steve IS 202 – FALL 2003. 09. 23 Keywords vs. Semantic Descriptors dog, biting, Steve IS 202 – FALL 2003. 09. 23 - SLIDE 32

Keywords vs. Semantic Descriptors dog, biting, Steve IS 202 – FALL 2003. 09. 23 Keywords vs. Semantic Descriptors dog, biting, Steve IS 202 – FALL 2003. 09. 23 - SLIDE 33

Why Keywords Don’t Work • Are not a semantic representation • Do not describe Why Keywords Don’t Work • Are not a semantic representation • Do not describe relations between descriptors • Do not describe temporal structure • Do not converge • Do not scale IS 202 – FALL 2003. 09. 23 - SLIDE 34

Natural Language vs. Visual Language Jack, an adult male police officer, while walking to Natural Language vs. Visual Language Jack, an adult male police officer, while walking to the left, starts waving with his left arm, and then has a puzzled look on his face as he turns his head to the right; he then drops his facial expression and stops turning his head, immediately looks up, and then stops looking up after he stops waving but before he stops walking. IS 202 – FALL 2003. 09. 23 - SLIDE 35

Natural Language vs. Visual Language Jack, an adult male police officer, while walking to Natural Language vs. Visual Language Jack, an adult male police officer, while walking to the left, starts waving with his left arm, and then has a puzzled look on his face as he turns his head to the right; he then drops his facial expression and stops turning his head, immediately looks up, and then stops looking up after he stops waving but before he stops walking. IS 202 – FALL 2003. 09. 23 - SLIDE 36

Notation for Time-Based Media: Music IS 202 – FALL 2003. 09. 23 - SLIDE Notation for Time-Based Media: Music IS 202 – FALL 2003. 09. 23 - SLIDE 37

Visual Language Advantages • A language designed as an accurate and readable representation of Visual Language Advantages • A language designed as an accurate and readable representation of time-based media – For video, especially important for actions, expressions, and spatial relations • Enables Gestalt view and quick recognition of descriptors due to designed visual similarities • Supports global use of annotations IS 202 – FALL 2003. 09. 23 - SLIDE 38

Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Video – Current Approaches – Media Streams • Discussion Questions • Action Items for Next Time IS 202 – FALL 2003. 09. 23 - SLIDE 39

After Capture: Media Streams IS 202 – FALL 2003. 09. 23 - SLIDE 40 After Capture: Media Streams IS 202 – FALL 2003. 09. 23 - SLIDE 40

Media Streams Features • Key features – – – Stream-based representation (better segmentation) Semantic Media Streams Features • Key features – – – Stream-based representation (better segmentation) Semantic indexing (what things are similar to) Relational indexing (who is doing what to whom) Temporal indexing (when things happen) Iconic interface (designed visual language) Universal annotation (standardized markup schema) • Key benefits – More accurate annotation and retrieval – Global usability and standardization – Reuse of rich media according to content and structure IS 202 – FALL 2003. 09. 23 - SLIDE 41

Media Streams GUI Components • Media Time Line • Icon Space – Icon Workshop Media Streams GUI Components • Media Time Line • Icon Space – Icon Workshop – Icon Palette IS 202 – FALL 2003. 09. 23 - SLIDE 42

Media Time Line • Visualize video at multiple time scales • Write and read Media Time Line • Visualize video at multiple time scales • Write and read multi-layered iconic annotations • One interface for annotation, query, and composition IS 202 – FALL 2003. 09. 23 - SLIDE 43

Media Time Line IS 202 – FALL 2003. 09. 23 - SLIDE 44 Media Time Line IS 202 – FALL 2003. 09. 23 - SLIDE 44

Icon Space • Icon Workshop – Utilize categories of video representation – Create iconic Icon Space • Icon Workshop – Utilize categories of video representation – Create iconic descriptors by compounding iconic primitives – Extend set of iconic descriptors • Icon Palette – Dynamically group related sets of iconic descriptors – Reuse descriptive effort of others – View and use query results IS 202 – FALL 2003. 09. 23 - SLIDE 45

Icon Space IS 202 – FALL 2003. 09. 23 - SLIDE 46 Icon Space IS 202 – FALL 2003. 09. 23 - SLIDE 46

Icon Space: Icon Workshop • General to specific (horizontal) – Cascading hierarchy of icons Icon Space: Icon Workshop • General to specific (horizontal) – Cascading hierarchy of icons with increasing specificity on subordinate levels • Combinatorial (vertical) – Compounding of hierarchically organized icons across multiple axes of description IS 202 – FALL 2003. 09. 23 - SLIDE 47

Icon Space: Icon Workshop Detail IS 202 – FALL 2003. 09. 23 - SLIDE Icon Space: Icon Workshop Detail IS 202 – FALL 2003. 09. 23 - SLIDE 48

Icon Space: Icon Palette • Dynamically group related sets of iconic descriptors • Collect Icon Space: Icon Palette • Dynamically group related sets of iconic descriptors • Collect icon sentences • Reuse descriptive effort of others IS 202 – FALL 2003. 09. 23 - SLIDE 49

Icon Space: Icon Palette Detail IS 202 – FALL 2003. 09. 23 - SLIDE Icon Space: Icon Palette Detail IS 202 – FALL 2003. 09. 23 - SLIDE 50

Video Retrieval In Media Streams • Same interface for annotation and retrieval • Assembles Video Retrieval In Media Streams • Same interface for annotation and retrieval • Assembles responses to queries as well as finds them • Query responses use semantics to degrade gracefully IS 202 – FALL 2003. 09. 23 - SLIDE 51

Media Streams Technologies • Minimal video representation distinguishing syntax and semantics • Iconic visual Media Streams Technologies • Minimal video representation distinguishing syntax and semantics • Iconic visual language for annotating and retrieving video content • Retrieval-by-composition methods for repurposing video IS 202 – FALL 2003. 09. 23 - SLIDE 52

Non-Technical Challenges • Standardization of media metadata (MPEG-7) • Broadband infrastructure and deployment • Non-Technical Challenges • Standardization of media metadata (MPEG-7) • Broadband infrastructure and deployment • Intellectual property and economic models for sharing and reuse of media assets IS 202 – FALL 2003. 09. 23 - SLIDE 53

Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Video – Current Approaches – Media Streams • Discussion Questions • Action Items for Next Time IS 202 – FALL 2003. 09. 23 - SLIDE 54

Discussion Questions (Davis) • John Snydal on Media Streams – What is the target Discussion Questions (Davis) • John Snydal on Media Streams – What is the target audience of users (annotators/retrievers) for Media Streams? In the article the following groups are mentioned: • • • Content providers Video editors News teams Documentary film makers Film archives Stock photo houses Video archivists Video producers (international audience) (illiterate and preliterate people) – Is it possible that Media Streams could satisfy the needs, goals and requirements of all of these groups, or would it be more appropriate to develop separate, tailored applications for the unique needs of each group? IS 202 – FALL 2003. 09. 23 - SLIDE 55

Discussion Questions (Davis) • danah boyd on Media Streams – Icons require visual literacy. Discussion Questions (Davis) • danah boyd on Media Streams – Icons require visual literacy. Icons are also culturally constructed. Thus, for them to work as an information access bit, people must learn the visual language; it is not inherent. What are the social consequences of a system dependent on unfamiliar cues? IS 202 – FALL 2003. 09. 23 - SLIDE 56

Discussion Questions (Davis) • danah boyd on Media Streams – Films are constructed narratives. Discussion Questions (Davis) • danah boyd on Media Streams – Films are constructed narratives. But most commonplace storytelling is not. Even in a creative form, people often piece together found objects instead of finding objects to fit their story. (Think teenage girls making collages out of the latest YM. ) Storytelling also happens around media far more than through media (i. e. telling a story about a picture rather than using a collection of pictures to tell a story). My guess is that this social phenomenon goes beyond the retrieval issues. Do you think that Media Streams would encourage new behavior regarding storytelling or will it only be useful for those with a constructed narrative in mind? Why (not)? IS 202 – FALL 2003. 09. 23 - SLIDE 57

Discussion Questions (Davis) • Jesse Mendelsohn on Media Streams – Media Streams does not Discussion Questions (Davis) • Jesse Mendelsohn on Media Streams – Media Streams does not allow iconic descriptions of emotion or scene-interpretation. How would someone searching stock footage for a “suspenseful scene of two men beating each other” go about doing it? The actual sense of “suspense” and the act of “beating” cannot be iconified. Does this limit Media Streams' ability or is there a way around it within its capabilities as described? IS 202 – FALL 2003. 09. 23 - SLIDE 58

Discussion Questions (Davis) • Jesse Mendelsohn on Media Streams – In order for Media Discussion Questions (Davis) • Jesse Mendelsohn on Media Streams – In order for Media Streams to work well it relies on a the availability of a very large and extensive resource of well-annotated video. Is the current annotation process too primitive and/or time consuming to allow Media Streams to work to its full potential? Will changing how Media Streams can be used to annotate video or changing video annotation methods in general make Media Streams more effective? IS 202 – FALL 2003. 09. 23 - SLIDE 59

Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Today’s Agenda • Review of Last Time • Metadata for Motion Pictures – Representing Video – Current Approaches – Media Streams • Discussion Questions • Action Items for Next Time IS 202 – FALL 2003. 09. 23 - SLIDE 60

Assignment 4. 1 • Phone Metadata Design - Part 1 – Due Oct 2 Assignment 4. 1 • Phone Metadata Design - Part 1 – Due Oct 2 IS 202 – FALL 2003. 09. 23 - SLIDE 61

Next Time • Database Design (RRL) • Readings – Handouts in Class • Database Next Time • Database Design (RRL) • Readings – Handouts in Class • Database Modeling and Design -- Ch. 2 The ER Model - Basic Concepts (Teorey, T. J. ) • Logical Database Design and the Relational Model (F. R. Mc. Fadden, J. A. Hoffer) IS 202 – FALL 2003. 09. 23 - SLIDE 62