US20060090123A1 - System and method for acquisition and storage of presentations - Google Patents
System and method for acquisition and storage of presentations Download PDFInfo
- Publication number
- US20060090123A1 US20060090123A1 US10/973,698 US97369804A US2006090123A1 US 20060090123 A1 US20060090123 A1 US 20060090123A1 US 97369804 A US97369804 A US 97369804A US 2006090123 A1 US2006090123 A1 US 2006090123A1
- Authority
- US
- United States
- Prior art keywords
- image
- presentation
- signal
- video
- association
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/166—Editing, e.g. inserting or deleting
- G06F40/169—Annotation, e.g. comment data or footnotes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/43—Querying
- G06F16/438—Presentation of query results
- G06F16/4387—Presentation of query results by the use of playlists
- G06F16/4393—Multimedia presentations, e.g. slide shows, multimedia albums
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/48—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/53—Querying
- G06F16/538—Presentation of query results
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/24—Querying
- G06F16/245—Query processing
- G06F16/2457—Query processing with adaptation to user needs
- G06F16/24573—Query processing with adaptation to user needs using data annotations, e.g. user-defined metadata
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/44—Browsing; Visualisation therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/44—Browsing; Visualisation therefor
- G06F16/447—Temporal browsing, e.g. timeline
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/71—Indexing; Data structures therefor; Storage structures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/93—Document management systems
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/912—Applications of a database
- Y10S707/913—Multimedia
- Y10S707/914—Video
Definitions
- the present invention relates generally to processing and storing images. More particularly it relates to extracting information from video presentations and storing the video presentations for later use.
- the media e.g slides, video, audio
- the media e.g slides, video, audio
- This limitation is especially troublesome as these presentations may be the only broadly accessible documents through which certain types of gathered information are available.
- presentation software is usually unable to solicit input from the viewers of the presentation, limiting the presentation to a passive experience.
- Embodiments of the present invention enable the extraction, classification, storage, and supplementation of presentation video.
- a media system receives a signal carrying presentation video.
- the media system processes the signal and generates images for slides of the presentation.
- the media system then extracts text from the images and uses the text and other characteristics to classify the images and store them in a database. Additionally, the system enables viewers of the presentation to provide feedback on the presentation, which can be used to supplement the presentation.
- FIG. 1 is a block diagram illustrating one embodiment of interaction among a computer system, a media system, and a display device;
- FIG. 2 is a block diagram illustrating an alternate embodiment of interaction among a computer system, a media system, and a display device;
- FIG. 3 is a block diagram illustrating a closer view of a media system in accordance with one embodiment of the present invention
- FIG. 4 is a block diagram illustrating one embodiment of categorization information for stored video
- FIG. 5 is a flow chart illustrating a process for handling presentation video input from a computer system
- FIG. 6 is a flow chart illustrating a process for categorizing and storing video input
- FIG. 7 is a flow chart illustrating a process for supplementing presentations with user input
- FIG. 8 is a flow chart illustrating a process for utilizing stored content in new presentations.
- Embodiments of the present invention enable the extraction, classification, storage, and supplementation of presentation video.
- a media system receives a signal carrying presentation video.
- the media system processes the signal and generates images for slides of the presentation.
- the media system then extracts text from the images and uses the text and other characteristics to classify the images and store them in a database.
- the present system automates the process of detecting, capturing, interpreting, and storing presentations.
- the system can detect when a presentation is beginning and initiate a process that detects whether content is static or dynamic and stores and classifies it accordingly.
- the system can also modify the content for ease of organization and distribution, distribute the content to viewers in an original or modified format, and end operations when a presentation is no longer detected.
- the steps above can be performed without any direct user commands to start and stop operations or any user sorting/separation/organization of the media.
- FIG. 1 is a block diagram illustrating one embodiment of interaction among a computer system, a media system, and a display device.
- a computer system 105 such as a laptop computer, desktop computer, tablet system, or any other type of computer, is connected to a video splitter 110 .
- the computer system 105 transmits an output video signal to the splitter 110 , which splits the video signal and outputs it to the media system 115 and the display device 120 .
- the video signal can be digital or analog and can comprise any number of signal formats.
- the video signal can also be a data signal containing video information, such as a Virtual Network Computing (VNC) signal.
- VNC Virtual Network Computing
- the splitter can also perform conversion of a data signal to a video signal.
- the display device 120 is a device used to display the video output to viewers of the presentation.
- the display device can be a Liquid Crystal Display (LCD) projector, analog projector, a Cathode Ray Tube(CRT) display, an LCD display or any other type of display.
- LCD Liquid Crystal Display
- CRT Cathode Ray Tube
- the media system 115 receives the video output from the splitter 1 10 , uses it to generate audio and video media for the presentation, and extract relevant information from the media.
- the media system 115 is a conventional computer using specialized software, in alternate embodiments, the media system 115 is a computer specially configured to function as a media system.
- the media system is also configured to collect audio through a microphone or other input. The audio can be stored in association with the presentation images and video.
- FIG. 2 is a block diagram illustrating an alternate embodiment of interaction among a computer system, a media system, and a display device.
- the media system 115 sits between the computer system and the display device 120 .
- the media system processes the video signal, generates slide images and displays the generated slide images on the display device 120 .
- the media system 115 may also accept image overlays and supplements, or other modifications, and output them to the display device 120 .
- the overlays can be generated by an automatic agent such as a translator program that automatically translates the text of the presentation.
- the media system 115 can also include a “pass-through” mode where the input video signal is passed directly, without modification, to the output device 120 .
- FIG. 3 is a block diagram illustrating a closer view of a media system 115 in accordance with one embodiment of the present invention.
- the media system 115 includes a video capture module 305 , an image sampling module 310 , an image converter 315 , an update module 320 , a text extraction module 325 , a database 330 , and an input/output module 335 . These components may be implemented through any combination of hardware, software, and firmware.
- the video capture module 305 receives the video signal from the splitter 110 or computer system 105 .
- the image sampling module 310 generates slide images from the video captured by the video capture module. In one embodiment, the image sampling module detects if a particular image has been broadcast steadily for a predetermined amount of time and treats it as a single slide. Alternately, continuous video is recorded in full. If the sampling module 310 determines that the image is a slide it generates a bitmap for the image. If it determines that the media is video, a video recording of either the whole capture or a segment of the window that contains video, is captured.
- the image converter 315 may optionally convert the bitmap to a more size efficient format such as JPEG or another format.
- An update module 320 is configured to generate categorization information for media and to store the media, with the categorization information, in the database 330 .
- the update module 320 first utilizes the text extraction module 325 , which detects text in the image and provides the text to the update module.
- the categorization information can include date/time information for the presentation, an identifier for the particular presentation being shown, characteristics of the image, supplemental information received from either the presenter or the viewers, and text within the image. Some categorization information is generated after the presentation has been recorded while some categorization information is generated in real time.
- the input/output module 335 is used to generate an interface for configuring the media system 115 .
- the interface can be a console interface on the media system 115 itself, graphical user interface that is accessed through input/output devices such as a keyboard and monitor that are connected to the media system, or a web interface that is accessed over a network.
- the input/output module 335 can also be used to transmit overlays and video supplements to the media system 115 , which uses the overlays to modify the image.
- the input/output module comprises a web server running on the media system 115 . By viewing an interface page on the web server, viewers of the presentation can submit questions and comments as overlays for the presentation.
- the web server can also be used as an interface for submitting search queries for images stored in the database 330 .
- FIG. 4 is a block diagram illustrating one embodiment of categorization information 400 for stored media.
- the categorization information 400 includes metadata 402 and classification information 418 .
- the metadata 402 is preferably stored in association with the media and is generated when the media is first captured.
- the classification information 418 can be stored in association with the media or centrally.
- the classification information 418 is often generated after a presentation rather than in real-time.
- the metadata includes content information 405 .
- the content information indicates whether the stored video comprises a single slide image, a video clip containing continuous video, audio, or some other type of media.
- the metadata additionally includes text information 410 .
- the text information 410 includes text that has been extracted from the slide image by the text extraction module 325 .
- the information can include all of the text or particular key words that were designated as representative words for searches.
- the text information 410 can include weights or other information indicating the importance of particular text in the slides.
- the text extraction module 325 can be programmed to recognize title text or section headings and give that text greater importance in classifying the slide image.
- the metadata additionally includes video characteristics 415 .
- the video characteristics include image characteristics that are extracted from the slide image. These can include colors or distinctive shapes or other image qualities.
- the metadata additionally includes supplemented information 425 .
- the supplemented information includes overlays and other information that is provided by a presenter, automatic agent, or the audience during a presentation.
- the classification information 418 can include an identifier for the presentation from which the image is extracted. It may also include time and date information for the presentation. For example, all of the video or slides for a single presentation would include the same identifier within the classification information 418 . Presentation data can also be grouped by meeting or day with all of the presentation data for a single meeting or day classified associatively. Artificial categorizations that associate presentations that are related in other ways can also be added.
- the categorization information 400 can be used by an associated search utility to retrieve presentation content in response to submitted search requests.
- Users of the search utility can search according to content or organizational data (i.e. when a presentation was shown, content shown at a meeting or presentation) and the search utility will return media, complete presentations, or sections of presentations matching the search request.
- FIG. 5 is a flow chart illustrating a process for handling presentation video input from a computer system.
- the media system 115 accepts presentation video and/or audio, preferably through the video capture module 305 .
- the system can detect when a presentation has begun by analyzing an incoming video stream and detecting characteristics indicative of a presentation. This process can also be used to stop recording when the detected video characteristics indicate that a presentation is no longer being transmitted.
- the media system extracts the presentation information. This step includes the determination of what type of media is being presented, the extraction of slide images or video streams from the video, the conversion of the slide images to JPEGs, and the extraction of text from the image. This step is described in greater detail with respect to FIG. 6 .
- This step may also include the extraction of video streams and audio streams.
- This step can also include analysis of audio content, for changes in volume, detection of words through speech to text extraction, and any other useful or relevant characteristics of the audio. Audio content can be classified according to characteristics of the audio, characteristics of video detected at the same time, or both.
- the system can use a variety of methods for categorizing the input received from the video signal and categorizing it accordingly. Usually the system will analyze a predetermined number of consecutive frames and categorize it appropriately. In one embodiment, the system detects a slide or other stable presentation by detecting unchanging video frames for more than a predetermined amount of time.
- Video can be detected in a similar manner.
- the system computes the difference between a series of consecutive frames. The system checks for a region in the series of frames in which the frames are always changing (the difference between successive frames is not null). If it finds a region that changes continually it determines that a video clip is playing. In some embodiments, the system can crop the sections of the frames that are not changing. In alternate embodiments, the entire frame is cropped.
- the media which can include video, slides, or audio, is stored in association with the presentation information of FIG. 4 .
- the presentation information is supplemented with overlays. These overlays can be received from the presenter, an automatic agent, or the audience through the web server generated by the input/output module.
- the presenter can accept questions from audience members through a network interface. The questions can be overlayed on the slide image.
- the supplemented image is output to the display device 120 .
- FIG. 6 is a flow chart illustrating a process for categorizing and storing video input.
- the system accepts media input.
- the media input is received through the video capture module 305 .
- the image sampling module 310 extracts content from the video stream.
- the image sampling module 612 determines a type for the content. For example, video clips can be identified if a section of the image changes continuously and stored as continuous segments.
- the image sampling module 310 checks for images that are displayed continuously for a predetermined amount of time, designates those images as static images, and generates bitmaps for the images. The system can apply other criteria as well.
- the font size of any text in the image is used, with larger text indicating a greater likelihood that the image is a slide.
- the image sampling module can also extract audio from the media stream, to be stored in association with video or images captured concurrently.
- step 615 the image converter 315 converts the content to more compact format, such as GIF or JPEG for images, or MPEG for video. This step is optional, and in some embodiments, the image is stored in an uncompressed form.
- step 620 the update module 320 generates a new entry in the database 330 . The entry is created with initial categorization information such as the content type 405 for the media and video characteristics 415 .
- the update module 320 utilizes the text extraction module 325 to extract text from the image or video.
- the text can include weights or other information indicating the importance of particular text in the slides.
- the text extraction module 325 can be programmed to recognize title text or section headings and give that text greater importance in classifying the content.
- the content is stored in the database 330 . This step also entails adding the extracted text and any other supplemental information.
- FIG. 7 is a flow chart illustrating a process for supplementing presentations with user input.
- the video signal is received by the video capture module 305 .
- the presentation video is transmitted to the viewers of the presentation.
- the media system 115 transmits special presentation display information over a network connection, which is received by the viewers at their terminals or computers, and is processed and displayed by an application on the recipients' computers.
- the media system through the input/output module 335 accepts annotations from either the viewers or the presenter.
- the annotations can be comments or supplemental overlays (drawings added to the slides through a mouse or writing tool). Alternately, the annotations can be questions or comments transmitted from the viewers. In some embodiments, the questions or comments are displayed in a preset section of the image.
- step 725 the annotated presentation is displayed.
- the annotations are displayed in real time.
- the annotations are collected during the presentation and displayed when the presenter returns to an earlier stage of the presentation.
- step 730 the slide image is stored in the database 330 with the annotations stored in the supplemented information 425 .
- FIG. 8 is a flow chart illustrating a process for utilizing stored content in new presentations.
- a user starts creation of a new slide presentation. In some embodiments, this presentation is generated on the computer system 105 .
- An application module on the computer system is configured to detect the creation of a new presentation.
- the application module is configured to access the database 330 on the media server 115 .
- the application module according to the text input in the presentation determines search terms for the presentation.
- the application uses the search terms, the application searches the database 330 for related content, cross-referencing the search terms with the identification information described in FIG. 4 .
- the system provides images matching the search term and prompts a user to include them.
- step 825 responsive to user acceptance, the search terms are included in the presentation.
- the present invention may be conveniently implemented using a conventional general purpose or a specialized digital computer or microprocessor programmed according to the teachings of the present disclosure, as will be apparent to those skilled in the computer art.
- the present invention includes a computer program product which is a storage medium (media) having instructions stored thereon/in which can be used to program a computer to perform any of the processes of the present invention.
- the storage medium can include, but is not limited to, any type of disk including floppy disks, optical discs, DVD, CD-ROMs, microdrive, and magneto-optical disks, ROMs, RAMs, EPROMs, EEPROMs, DRAMs, VRAMs, flash memory devices, magnetic or optical cards, nanosystems (including molecular memory ICs), or any type of media or device suitable for storing instructions and/or data.
- the present invention includes software for controlling both the hardware of the general purpose/specialized computer or microprocessor, and for enabling the computer or microprocessor to interact with a human user or other mechanism utilizing the results of the present invention.
- software may include, but is not limited to, device drivers, operating systems, and user applications.
Abstract
Embodiments of the present invention enable the extraction, classification, storage, and supplementation of presentation video. A media system receives a video signal carrying presentation video. The media system processes the video signal and generates images for slides of the presentation. The media system then extracts text from the images and uses the text and other characteristics to classify the images and store them in a database. Additionally, the system enables viewers of the presentation to provide feedback on the presentation, which can be used to supplement the presentation.
Description
- 1. Field of the Invention
- The present invention relates generally to processing and storing images. More particularly it relates to extracting information from video presentations and storing the video presentations for later use.
- 2. Description of the Related Art
- In modern business environments a greater emphasis has been placed on the transfer and exchange of information. During this time, slide-based presentations using computer presentation software such as Microsoft Power Point, web based presentations, and video presentations, have become a staple of modern business environments. However, such presentation software, while often superficially useful for presenting information to others, possesses a number of severe limitations.
- Firstly, the media (e.g slides, video, audio) used in the presentation are seldom stored in a format that is easily searchable or accessible. Thus, it is often difficult for presenters and recipients of these presentations to search the content. This limitation is especially troublesome as these presentations may be the only broadly accessible documents through which certain types of gathered information are available. Additionally, such presentation software is usually unable to solicit input from the viewers of the presentation, limiting the presentation to a passive experience.
- Attempts to address these problems have usually centered around additions or modifications to the presentation software. However, such modifications must be performed on a per-application basis, and in the case of soliciting input, usually require configuration on the systems of the viewers.
- What is needed is an improved system for storing, organizing, and modifying presentations.
- Embodiments of the present invention enable the extraction, classification, storage, and supplementation of presentation video. A media system receives a signal carrying presentation video. The media system processes the signal and generates images for slides of the presentation. The media system then extracts text from the images and uses the text and other characteristics to classify the images and store them in a database. Additionally, the system enables viewers of the presentation to provide feedback on the presentation, which can be used to supplement the presentation.
- Preferred embodiments of the present invention will be described in detail based on the following figures, wherein:
-
FIG. 1 is a block diagram illustrating one embodiment of interaction among a computer system, a media system, and a display device; -
FIG. 2 is a block diagram illustrating an alternate embodiment of interaction among a computer system, a media system, and a display device; -
FIG. 3 is a block diagram illustrating a closer view of a media system in accordance with one embodiment of the present invention; -
FIG. 4 is a block diagram illustrating one embodiment of categorization information for stored video; -
FIG. 5 is a flow chart illustrating a process for handling presentation video input from a computer system; -
FIG. 6 is a flow chart illustrating a process for categorizing and storing video input; -
FIG. 7 is a flow chart illustrating a process for supplementing presentations with user input; -
FIG. 8 is a flow chart illustrating a process for utilizing stored content in new presentations. - Embodiments of the present invention enable the extraction, classification, storage, and supplementation of presentation video. A media system receives a signal carrying presentation video. The media system processes the signal and generates images for slides of the presentation. The media system then extracts text from the images and uses the text and other characteristics to classify the images and store them in a database.
- The present system automates the process of detecting, capturing, interpreting, and storing presentations. The system can detect when a presentation is beginning and initiate a process that detects whether content is static or dynamic and stores and classifies it accordingly. The system can also modify the content for ease of organization and distribution, distribute the content to viewers in an original or modified format, and end operations when a presentation is no longer detected. The steps above can be performed without any direct user commands to start and stop operations or any user sorting/separation/organization of the media.
-
FIG. 1 is a block diagram illustrating one embodiment of interaction among a computer system, a media system, and a display device. Acomputer system 105, such as a laptop computer, desktop computer, tablet system, or any other type of computer, is connected to avideo splitter 110. Thecomputer system 105 transmits an output video signal to thesplitter 110, which splits the video signal and outputs it to themedia system 115 and the display device 120.The video signal can be digital or analog and can comprise any number of signal formats. The video signal can also be a data signal containing video information, such as a Virtual Network Computing (VNC) signal. The splitter can also perform conversion of a data signal to a video signal. - The
display device 120 is a device used to display the video output to viewers of the presentation. The display device can be a Liquid Crystal Display (LCD) projector, analog projector, a Cathode Ray Tube(CRT) display, an LCD display or any other type of display. - The
media system 115 receives the video output from the splitter 1 10, uses it to generate audio and video media for the presentation, and extract relevant information from the media. In some embodiments, themedia system 115 is a conventional computer using specialized software, in alternate embodiments, themedia system 115 is a computer specially configured to function as a media system. In some embodiments, the media system is also configured to collect audio through a microphone or other input. The audio can be stored in association with the presentation images and video. -
FIG. 2 is a block diagram illustrating an alternate embodiment of interaction among a computer system, a media system, and a display device. In the present embodiment, themedia system 115 sits between the computer system and thedisplay device 120. In this embodiment, the media system processes the video signal, generates slide images and displays the generated slide images on thedisplay device 120. Themedia system 115 may also accept image overlays and supplements, or other modifications, and output them to thedisplay device 120. Alternately, the overlays can be generated by an automatic agent such as a translator program that automatically translates the text of the presentation. Themedia system 115 can also include a “pass-through” mode where the input video signal is passed directly, without modification, to theoutput device 120. -
FIG. 3 is a block diagram illustrating a closer view of amedia system 115 in accordance with one embodiment of the present invention. Themedia system 115 includes avideo capture module 305, animage sampling module 310, animage converter 315, anupdate module 320, atext extraction module 325, adatabase 330, and an input/output module 335. These components may be implemented through any combination of hardware, software, and firmware. - The
video capture module 305 receives the video signal from thesplitter 110 orcomputer system 105. Theimage sampling module 310 generates slide images from the video captured by the video capture module. In one embodiment, the image sampling module detects if a particular image has been broadcast steadily for a predetermined amount of time and treats it as a single slide. Alternately, continuous video is recorded in full. If thesampling module 310 determines that the image is a slide it generates a bitmap for the image. If it determines that the media is video, a video recording of either the whole capture or a segment of the window that contains video, is captured. - The
image converter 315 may optionally convert the bitmap to a more size efficient format such as JPEG or another format. Anupdate module 320 is configured to generate categorization information for media and to store the media, with the categorization information, in thedatabase 330. In some embodiments, theupdate module 320 first utilizes thetext extraction module 325, which detects text in the image and provides the text to the update module. - The categorization information can include date/time information for the presentation, an identifier for the particular presentation being shown, characteristics of the image, supplemental information received from either the presenter or the viewers, and text within the image. Some categorization information is generated after the presentation has been recorded while some categorization information is generated in real time.
- The input/
output module 335 is used to generate an interface for configuring themedia system 115. The interface can be a console interface on themedia system 115 itself, graphical user interface that is accessed through input/output devices such as a keyboard and monitor that are connected to the media system, or a web interface that is accessed over a network. The input/output module 335 can also be used to transmit overlays and video supplements to themedia system 115, which uses the overlays to modify the image. In one embodiment, the input/output module comprises a web server running on themedia system 115. By viewing an interface page on the web server, viewers of the presentation can submit questions and comments as overlays for the presentation. The web server can also be used as an interface for submitting search queries for images stored in thedatabase 330. -
FIG. 4 is a block diagram illustrating one embodiment ofcategorization information 400 for stored media. Thecategorization information 400 includesmetadata 402 andclassification information 418. Themetadata 402 is preferably stored in association with the media and is generated when the media is first captured. Theclassification information 418 can be stored in association with the media or centrally. Theclassification information 418 is often generated after a presentation rather than in real-time. The metadata includescontent information 405. The content information indicates whether the stored video comprises a single slide image, a video clip containing continuous video, audio, or some other type of media. - The metadata additionally includes
text information 410. Thetext information 410 includes text that has been extracted from the slide image by thetext extraction module 325. The information can include all of the text or particular key words that were designated as representative words for searches. Thetext information 410 can include weights or other information indicating the importance of particular text in the slides. For example, thetext extraction module 325 can be programmed to recognize title text or section headings and give that text greater importance in classifying the slide image. - The metadata additionally includes
video characteristics 415. The video characteristics include image characteristics that are extracted from the slide image. These can include colors or distinctive shapes or other image qualities. The metadata additionally includes supplementedinformation 425. The supplemented information includes overlays and other information that is provided by a presenter, automatic agent, or the audience during a presentation. - The
classification information 418 can include an identifier for the presentation from which the image is extracted. It may also include time and date information for the presentation. For example, all of the video or slides for a single presentation would include the same identifier within theclassification information 418. Presentation data can also be grouped by meeting or day with all of the presentation data for a single meeting or day classified associatively. Artificial categorizations that associate presentations that are related in other ways can also be added. - The
categorization information 400 can be used by an associated search utility to retrieve presentation content in response to submitted search requests. Users of the search utility can search according to content or organizational data (i.e. when a presentation was shown, content shown at a meeting or presentation) and the search utility will return media, complete presentations, or sections of presentations matching the search request. -
FIG. 5 is a flow chart illustrating a process for handling presentation video input from a computer system. Instep 505, themedia system 115 accepts presentation video and/or audio, preferably through thevideo capture module 305. In some embodiments, the system can detect when a presentation has begun by analyzing an incoming video stream and detecting characteristics indicative of a presentation. This process can also be used to stop recording when the detected video characteristics indicate that a presentation is no longer being transmitted. Instep 510, the media system extracts the presentation information. This step includes the determination of what type of media is being presented, the extraction of slide images or video streams from the video, the conversion of the slide images to JPEGs, and the extraction of text from the image. This step is described in greater detail with respect toFIG. 6 . This step may also include the extraction of video streams and audio streams. This step can also include analysis of audio content, for changes in volume, detection of words through speech to text extraction, and any other useful or relevant characteristics of the audio. Audio content can be classified according to characteristics of the audio, characteristics of video detected at the same time, or both. - The system can use a variety of methods for categorizing the input received from the video signal and categorizing it accordingly. Usually the system will analyze a predetermined number of consecutive frames and categorize it appropriately. In one embodiment, the system detects a slide or other stable presentation by detecting unchanging video frames for more than a predetermined amount of time.
- Video can be detected in a similar manner. In one embodiment, the system computes the difference between a series of consecutive frames. The system checks for a region in the series of frames in which the frames are always changing (the difference between successive frames is not null). If it finds a region that changes continually it determines that a video clip is playing. In some embodiments, the system can crop the sections of the frames that are not changing. In alternate embodiments, the entire frame is cropped.
- In
step 515, the media, which can include video, slides, or audio, is stored in association with the presentation information ofFIG. 4 . Instep 520 the presentation information is supplemented with overlays. These overlays can be received from the presenter, an automatic agent, or the audience through the web server generated by the input/output module. In one embodiment, the presenter can accept questions from audience members through a network interface. The questions can be overlayed on the slide image. Instep 520, the supplemented image is output to thedisplay device 120. -
FIG. 6 is a flow chart illustrating a process for categorizing and storing video input. Instep 605 the system accepts media input. In one embodiment, the media input is received through thevideo capture module 305. Instep 610, theimage sampling module 310 extracts content from the video stream. Instep 612, theimage sampling module 612 determines a type for the content. For example, video clips can be identified if a section of the image changes continuously and stored as continuous segments. In one embodiment, theimage sampling module 310 checks for images that are displayed continuously for a predetermined amount of time, designates those images as static images, and generates bitmaps for the images. The system can apply other criteria as well. In one embodiment, the font size of any text in the image is used, with larger text indicating a greater likelihood that the image is a slide. During this step the image sampling module can also extract audio from the media stream, to be stored in association with video or images captured concurrently. - In
step 615 theimage converter 315 converts the content to more compact format, such as GIF or JPEG for images, or MPEG for video. This step is optional, and in some embodiments, the image is stored in an uncompressed form. Instep 620, theupdate module 320 generates a new entry in thedatabase 330. The entry is created with initial categorization information such as thecontent type 405 for the media andvideo characteristics 415. - In
step 625, theupdate module 320 utilizes thetext extraction module 325 to extract text from the image or video. The text can include weights or other information indicating the importance of particular text in the slides. For example, thetext extraction module 325 can be programmed to recognize title text or section headings and give that text greater importance in classifying the content. Instep 630, the content is stored in thedatabase 330. This step also entails adding the extracted text and any other supplemental information. -
FIG. 7 is a flow chart illustrating a process for supplementing presentations with user input. Instep 705, the video signal is received by thevideo capture module 305. Instep 710, the presentation video is transmitted to the viewers of the presentation. In some embodiments, themedia system 115 transmits special presentation display information over a network connection, which is received by the viewers at their terminals or computers, and is processed and displayed by an application on the recipients' computers. Instep 720 the media system, through the input/output module 335 accepts annotations from either the viewers or the presenter. The annotations can be comments or supplemental overlays (drawings added to the slides through a mouse or writing tool). Alternately, the annotations can be questions or comments transmitted from the viewers. In some embodiments, the questions or comments are displayed in a preset section of the image. - In
step 725 the annotated presentation is displayed. In some embodiments, the annotations are displayed in real time. In alternate embodiments, the annotations are collected during the presentation and displayed when the presenter returns to an earlier stage of the presentation. - In
step 730, the slide image is stored in thedatabase 330 with the annotations stored in the supplementedinformation 425. -
FIG. 8 is a flow chart illustrating a process for utilizing stored content in new presentations. In step 805 a user starts creation of a new slide presentation. In some embodiments, this presentation is generated on thecomputer system 105. An application module on the computer system, either as part of the presentation generation program, or independently, is configured to detect the creation of a new presentation. The application module is configured to access thedatabase 330 on themedia server 115. Instep 810, the application module, according to the text input in the presentation determines search terms for the presentation. Instep 815, using the search terms, the application searches thedatabase 330 for related content, cross-referencing the search terms with the identification information described inFIG. 4 . In step 820, the system provides images matching the search term and prompts a user to include them. - In
step 825, responsive to user acceptance, the search terms are included in the presentation. - Other features, aspects and objects of the invention can be obtained from a review of the figures and the claims. It is to be understood that other embodiments of the invention can be developed and fall within the spirit and scope of the invention and claims.
- The foregoing description of preferred embodiments of the present invention has been provided for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Obviously, many modifications and variations will be apparent to the practitioner skilled in the art. The embodiments were chosen and described in order to best explain the principles of the invention and its practical application, thereby enabling others skilled in the art to understand the invention for various embodiments and with various modifications that are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the following claims and their equivalence.
- In addition to an embodiment consisting of specifically designed integrated circuits or other electronics, the present invention may be conveniently implemented using a conventional general purpose or a specialized digital computer or microprocessor programmed according to the teachings of the present disclosure, as will be apparent to those skilled in the computer art.
- Appropriate software coding can readily be prepared by skilled programmers based on the teachings of the present disclosure, as will be apparent to those skilled in the software art. The invention may also be implemented by the preparation of application specific integrated circuits or by interconnecting an appropriate network of conventional component circuits, as will be readily apparent to those skilled in the art.
- The present invention includes a computer program product which is a storage medium (media) having instructions stored thereon/in which can be used to program a computer to perform any of the processes of the present invention. The storage medium can include, but is not limited to, any type of disk including floppy disks, optical discs, DVD, CD-ROMs, microdrive, and magneto-optical disks, ROMs, RAMs, EPROMs, EEPROMs, DRAMs, VRAMs, flash memory devices, magnetic or optical cards, nanosystems (including molecular memory ICs), or any type of media or device suitable for storing instructions and/or data.
- Stored on any one of the computer readable medium (media), the present invention includes software for controlling both the hardware of the general purpose/specialized computer or microprocessor, and for enabling the computer or microprocessor to interact with a human user or other mechanism utilizing the results of the present invention. Such software may include, but is not limited to, device drivers, operating systems, and user applications.
- Included in the programming (software) of the general/specialized computer or microprocessor are software modules for implementing the teachings of the present invention.
Claims (89)
1. A method for capturing video presentations, the method comprising:
collecting a signal comprising video information, the signal associated with a presentation;
generating at least one image from the signal;
determining one or more categorization criteria for the image; and
storing the image in association with the categorization criteria.
2. The method of claim 1 , wherein the image comprises a slide in the presentation.
3. The method of claim 1 , wherein the categorization criteria comprise text within the image.
4. The method of claim 1 , wherein the categorization criteria comprise visual characteristics of the image.
5. The method of claim 1 , wherein the categorization criteria comprise a time in which the presentation was shown.
6. The method of claim 1 , wherein the categorization criteria comprise a meeting in which the presentation was shown.
7. The method of claim 1 , further comprising:
receiving a search query; and
returning presentation content according to a similarity between the search query and categorization criteria for the presentation content.
8. The method of claim 1 , wherein the image is stored in JPEG format.
9. The method of claim 1 , further comprising transmitting the image to a display.
10. The method of claim 1 , wherein the image is stored in a Structured Query Language database.
11. The method of claim 1 , further comprising compressing the image.
12. The method of claim 1 , further comprising modifying the image.
13. The method of claim 1 , further comprising:
transmitting the image to a viewer of the presentation; and
accepting an annotation for the image.
14. The method of claim 13 , further comprising storing the annotation in association with the image.
15. The method of claim 1 , wherein the categorization information comprises metadata.
16. The method of claim 1 , further comprising:
capturing an audio signal; and
storing the audio signal in association with categorization criteria of an image captured at approximately the time in which the audio signal was captured.
17. The method of claim 1 , further comprising:
capturing an audio signal;
extracting features of the audio signal; and
storing the audio signal in association with the extracted features.
18. The method of claim 1 , wherein generating an image from the signal comprises determining a content type for the signal.
19. The method of claim 1 , further comprising:
accepting an overlay for the image; and
storing the overlay in association with the image.
20. A machine readable medium having instructions stored thereon that when executed by a processor cause a system to:
collect a signal comprising video information, the signal associated with a presentation;
generate at least one image from the signal;
determine one or more categorization criteria for the image; and
store the image in association with the categorization criteria.
21. The machine readable medium of claim 20 , wherein the image comprises a slide in the presentation.
22. The machine readable medium of claim 20 , wherein the categorization criteria comprise text within the image.
23. The machine readable medium of claim 20 , wherein the categorization criteria comprise visual characteristics of the image.
24. The machine readable medium of claim 20 , wherein the categorization criteria comprise a time in which the presentation was shown.
25. The machine readable medium of claim 20 , further comprising instructions that when executed by a processor cause the system to:
receive a search query; and
return presentation content according to a similarity between the search query and categorization criteria for the presentation content.
26. The machine readable medium of claim 20 , wherein the image is stored in JPEG format.
27. The machine readable medium of claim 20 , wherein the image is stored in a Structured Query Language database.
28. The machine readable medium of claim 20 , further comprising instructions that when executed by a processor cause the system to:
transmit the image to a viewer of the presentation; and
accept an annotation for the image.
29. The machine readable medium of claim 28 , further comprising instructions that when executed by a processor cause the system to store the annotation in association with the image.
30. The machine readable medium of claim 20 , wherein the categorization information comprises metadata.
31. The machine readable medium of claim 20 , further comprising instructions that when executed by the processor cause the system to:
capture an audio signal; and
store the audio signal in association with categorization criteria of an image captured at approximately the time in which the audio signal was captured.
32. The machine readable medium of claim 20 , further comprising instructions that when executed by the processor cause the system to:
accept an overlay for the image; and
store the overlay in association with the image.
33. The machine readable medium of claim 20 , wherein the instructions for generating an image from the signal comprise instructions for determining a content type for the signal.
34. The machine readable medium of claim 20 , wherein the categorization criteria comprise a meeting in which a presentation was shown.
35. The machine readable medium of claim 20 , further comprising instructions that when executed by the processor cause the system to transmit the image to a display.
36. The machine readable medium of claim 20 , further comprising instructions that when executed by the processor cause the system to modify the image.
37. The machine readable medium of claim 20 , further comprising instructions that when executed by the processor cause the system to:
capture an audio signal;
extract features of the audio signal; and
store the audio signal in association with the extracted features.
38. A system for storing video presentations, the system comprising:
a database for storing images;
an image capture module configured to convert a data signal associated with a presentation into at least one image;
an update module configured to:
determine one or more categorization criteria for the image; and
store the image in the database in association with the categorization criteria.
39. The system of claim 38 , wherein the image comprises a slide in the presentation.
40. The system of claim 38 , wherein the categorization criteria comprise text within the image.
41. The system of claim 38 , wherein the categorization criteria comprise visual characteristics of the image.
42. The system of claim 38 , wherein the categorization criteria comprise a time in which the presentation was shown.
43. The system of claim 38 , wherein the update module is further configured to:
receive a search query; and
return presentation content according to a similarity between the search query and categorization criteria for the presentation content.
44. The system of claim 38 , wherein the image is stored in JPEG format.
45. The system of claim 38 , wherein the image is stored in a Structured Query Language database.
46. The system of claim 38 , wherein the update module is further configured to:
transmit the image to a viewer of the presentation; and
accept an annotation for the image.
47. The system of claim 46 , wherein the update module is further configured to store the annotation in association with the image.
48. The system of claim 38 , wherein the categorization information comprises metadata.
49. The system of claim 38 , wherein the image capture module is further configured to:
capture an audio signal; and
store the audio signal in association with categorization criteria of an image captured at approximately the time in which the audio signal was captured.
50. The system of claim 38 , wherein the update module is further configured to:
accept an overlay for the image; and
store the overlay in association with the image.
51. The system of claim 38 , wherein image capture module, when converting the signal, determines a content type for the signal.
52. The system of claim 38 , wherein the categorization criteria comprise a meeting in which a presentation was shown.
53. The system of claim 38 , wherein the image capture module is further configured to transmit the image to a display.
54. The system of claim 38 , wherein the update module is further configured to modify the image.
55. The system of claim 38 , wherein the image capture module is further configured to:
capture an audio signal;
extract features of the audio signal; and
store the audio signal in association with the extracted features.
56. A method for capturing video presentations, the method comprising:
collecting a signal comprising video information, the signal associated with a presentation;
determining a media type from the signal;
responsive to determining the media type, generating an image from the signal;
extracting one or more items of metadata from the image; and
storing the image in association with the metadata.
57. The method of claim 56 , wherein extracting the one or more items of metadata comprises extracting text from the image.
58. The method of claim 56 , wherein determining the media type comprises determining that the media type comprises a slide.
59. The method of claim 56 , wherein determining the media type comprises determining that the media type comprises a video stream.
60. A method for capturing video presentations, the method comprising:
collecting a signal comprising video information, the signal associated with a presentation;
determining whether a presentation element in the signal is static or dynamic;
generating an image from the presentation element when the presentation element is static;
generating a video clip from the presentation element when the presentation element is dynamic; and
determining one or more characteristics associated with the presentation element.
61. The method of claim 60 , further comprising storing the video clip in association with the characteristics when the presentation element is dynamic.
62. The method of claim 60 , further comprising storing the image in association with the characteristics when the presentation element is static.
63. The method of claim 60 , wherein the image corresponds to a slide in the presentation.
64. The method of claim 60 , wherein the video clip corresponds to video shown within the presentation.
65. The method of claim 60 , wherein the video clip corresponds to interactions with a software application shown within the presentation.
66. The method of claim 60 , wherein the characteristics comprise text within the image or video clip.
67. The method of claim 60 , wherein the characteristics comprise visual characteristics of the image or video clip.
68. The method of claim 60 , wherein the characteristics comprise a time in which the image or video clip was shown.
69. The method of claim 60 , wherein the characteristics are determined after the image or video clip is stored.
70. The method of claim 60 , further comprising:
receiving a search query for presentation content; and
returning the presentation content according to a similarity between the search query and characteristics of the presentation element.
71. The method of claim 60 , further comprising:
generating a classification for the presentation element according to whether the presentation element is static or dynamic.
72. The method of claim 71 , further comprising:
receiving a search query for presentation content; and
returning the presentation content according to a similarity between the search query and the classification of the presentation content.
73. The method of claim 60 , wherein the image is stored in JPEG format.
74. The method of claim 60 , wherein the video clip is stored in MPEG format.
75. The method of claim 60 , wherein the image and video clip is stored in a Structured Query Language database.
76. The method of claim 60 , further comprising transmitting the signal for display.
77. The method of claim 60 , further comprising:
transmitting the image or video clip for display.
78. The method of claim 60 , further comprising:
accepting supplemental information for the presentation element; and
transmitting a signal including the presentation element and the supplemental information.
79. The method of claim 78 , further comprising storing the supplemental information in association with the image or video clip.
80. The method of claim 60 , further comprising:
capturing an audio signal associated with the presentation element; and
storing the audio signal.
81. The method of claim 80 , further comprising:
receiving a search query for a presentation element;
returning an image or video clip according to a similarity between the search query and characteristics of the image or video clip; and
returning an audio signal associated with the image or video clip.
82. The method of claim 60 , wherein the characteristics comprise text within the image or video clip.
83. The method of claim 60 , wherein the characteristics comprise visual characteristics of the image or video clip.
84. The method of claim 60 , wherein the characteristics comprise a time in which the presentation was shown.
85. The method of claim 60 , further comprising:
reducing a size of the presentation element; and
transmitting the presentation element.
86. The method of claim 60 , further comprising:
accepting supplemental information for the presentation element;
retrieving a search request for which the presentation element is returned; and
returning the supplemental information with the presentation element.
87. The method of claim 60 , further comprising:
accepting supplemental information for the presentation element;
receiving a search query; and
returning the supplemental information in response to a similarity between the search query and the supplemental information.
88. The method of claim 60 , wherein collecting a signal comprising video information comprises:
determining that the signal contains presentation elements; and
in response to a determination that the signal contains presentation elements, initiating the step of determining whether a presentation element is static or dynamic.
89. The method of claim 60 , further comprising:
determining that the signal no longer contains presentation elements; and
in response to a determination that the signal no longer contains presentation elements, halting the step of determining whether a presentation element is static or dynamic.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/973,698 US20060090123A1 (en) | 2004-10-26 | 2004-10-26 | System and method for acquisition and storage of presentations |
JP2005311276A JP2006127518A (en) | 2004-10-26 | 2005-10-26 | System and method for acquisition and storage of presentation |
US12/482,315 US9875222B2 (en) | 2004-10-26 | 2009-06-10 | Capturing and storing elements from a video presentation for later retrieval in response to queries |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/973,698 US20060090123A1 (en) | 2004-10-26 | 2004-10-26 | System and method for acquisition and storage of presentations |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/482,315 Division US9875222B2 (en) | 2004-10-26 | 2009-06-10 | Capturing and storing elements from a video presentation for later retrieval in response to queries |
Publications (1)
Publication Number | Publication Date |
---|---|
US20060090123A1 true US20060090123A1 (en) | 2006-04-27 |
Family
ID=36207385
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/973,698 Abandoned US20060090123A1 (en) | 2004-10-26 | 2004-10-26 | System and method for acquisition and storage of presentations |
US12/482,315 Active 2029-09-07 US9875222B2 (en) | 2004-10-26 | 2009-06-10 | Capturing and storing elements from a video presentation for later retrieval in response to queries |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/482,315 Active 2029-09-07 US9875222B2 (en) | 2004-10-26 | 2009-06-10 | Capturing and storing elements from a video presentation for later retrieval in response to queries |
Country Status (2)
Country | Link |
---|---|
US (2) | US20060090123A1 (en) |
JP (1) | JP2006127518A (en) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060250508A1 (en) * | 2005-05-09 | 2006-11-09 | Chia-Kai Chang | Method for inserting a picture into a video frame |
US20080007567A1 (en) * | 2005-12-18 | 2008-01-10 | Paul Clatworthy | System and Method for Generating Advertising in 2D or 3D Frames and Scenes |
US20080042923A1 (en) * | 2006-08-16 | 2008-02-21 | Rick De Laet | Systems, methods, and apparatus for recording of graphical display |
US20080183820A1 (en) * | 2007-01-29 | 2008-07-31 | Fuji Xerox Co., Ltd. | Device Orchestration |
US20080263010A1 (en) * | 2006-12-12 | 2008-10-23 | Microsoft Corporation | Techniques to selectively access meeting content |
US20080276159A1 (en) * | 2007-05-01 | 2008-11-06 | International Business Machines Corporation | Creating Annotated Recordings and Transcripts of Presentations Using a Mobile Device |
US20130036235A1 (en) * | 2010-04-16 | 2013-02-07 | Pablo Lopez Garcia | Method of transmission of visual content |
US8626487B2 (en) | 2010-09-08 | 2014-01-07 | Sharp Kabushiki Kaisha | Translation apparatus, translation method, computer program, and recording medium |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9116989B1 (en) | 2005-08-19 | 2015-08-25 | At&T Intellectual Property Ii, L.P. | System and method for using speech for data searching during presentations |
US8977965B1 (en) | 2005-08-19 | 2015-03-10 | At&T Intellectual Property Ii, L.P. | System and method for controlling presentations using a multimodal interface |
US9026915B1 (en) * | 2005-10-31 | 2015-05-05 | At&T Intellectual Property Ii, L.P. | System and method for creating a presentation using natural language |
US7904825B2 (en) * | 2007-03-14 | 2011-03-08 | Xerox Corporation | Graphical user interface for gathering image evaluation information |
JP4967950B2 (en) * | 2007-09-18 | 2012-07-04 | 富士ゼロックス株式会社 | Information recording apparatus and program |
US9177285B2 (en) * | 2008-05-06 | 2015-11-03 | Fuji Xerox Co., Ltd. | Method and system for controlling a space based on media content |
US9684736B2 (en) | 2009-01-29 | 2017-06-20 | Oracle International Corporation | Communication handler for flex integration with a secure application |
US9659335B2 (en) | 2009-01-29 | 2017-05-23 | Oracle International Corporation | Sample management for a sales call |
US9063806B2 (en) * | 2009-01-29 | 2015-06-23 | Oracle International Corporation | Flex integration with a secure application |
US8762883B2 (en) | 2009-01-30 | 2014-06-24 | Oracle International Corporation | Manipulation of window controls in a popup window |
US9760381B2 (en) | 2009-01-30 | 2017-09-12 | Oracle International Corporation | Configurable toolbar |
US8452640B2 (en) * | 2009-01-30 | 2013-05-28 | Oracle International Corporation | Personalized content delivery and analytics |
US8762448B2 (en) | 2009-01-30 | 2014-06-24 | Oracle International Corporation | Implementing asynchronous processes on a mobile client |
US8380487B2 (en) | 2010-06-21 | 2013-02-19 | International Business Machines Corporation | Language translation of selected content in a web conference |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5848409A (en) * | 1993-11-19 | 1998-12-08 | Smartpatents, Inc. | System, method and computer program product for maintaining group hits tables and document index tables for the purpose of searching through individual documents and groups of documents |
US20010020954A1 (en) * | 1999-11-17 | 2001-09-13 | Ricoh Company, Ltd. | Techniques for capturing information during multimedia presentations |
US6332144B1 (en) * | 1998-03-11 | 2001-12-18 | Altavista Company | Technique for annotating media |
US6646655B1 (en) * | 1999-03-09 | 2003-11-11 | Webex Communications, Inc. | Extracting a time-sequence of slides from video |
US20040202349A1 (en) * | 2003-04-11 | 2004-10-14 | Ricoh Company, Ltd. | Automated techniques for comparing contents of images |
US20040247206A1 (en) * | 2003-02-21 | 2004-12-09 | Canon Kabushiki Kaisha | Image processing method and image processing system |
US20050078868A1 (en) * | 2003-09-26 | 2005-04-14 | William Chen | Method and apparatus for summarizing and indexing the contents of an audio-visual presentation |
US7010751B2 (en) * | 2000-02-18 | 2006-03-07 | University Of Maryland, College Park | Methods for the electronic annotation, retrieval, and use of electronic images |
US20060074895A1 (en) * | 2004-09-29 | 2006-04-06 | International Business Machines Corporation | Method and system for extracting and utilizing metadata to improve accuracy in speech to text conversions |
US7149755B2 (en) * | 2002-07-29 | 2006-12-12 | Hewlett-Packard Development Company, Lp. | Presenting a collection of media objects |
US7392469B1 (en) * | 2003-05-19 | 2008-06-24 | Sidney Bailin | Non-intrusive commentary capture for document authors |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2965010B2 (en) * | 1997-08-30 | 1999-10-18 | 日本電気株式会社 | Related information search method and apparatus, and machine-readable recording medium recording program |
US6404856B1 (en) * | 1998-04-10 | 2002-06-11 | Fuji Xerox Co., Ltd. | System for recording, annotating and indexing audio data |
US6654735B1 (en) * | 1999-01-08 | 2003-11-25 | International Business Machines Corporation | Outbound information analysis for generating user interest profiles and improving user productivity |
KR100374040B1 (en) * | 2001-03-09 | 2003-03-03 | 엘지전자 주식회사 | Method for detecting caption synthetic key frame in video stream |
US20030123841A1 (en) * | 2001-12-27 | 2003-07-03 | Sylvie Jeannin | Commercial detection in audio-visual content based on scene change distances on separator boundaries |
JP2004112379A (en) * | 2002-09-19 | 2004-04-08 | Fuji Xerox Co Ltd | Image retrieving system |
US20040113933A1 (en) * | 2002-10-08 | 2004-06-17 | Northrop Grumman Corporation | Split and merge behavior analysis and understanding using Hidden Markov Models |
US7324736B2 (en) * | 2002-10-09 | 2008-01-29 | Lsi Logic Corporation | Intelligent recording control system |
JP2004185424A (en) * | 2002-12-04 | 2004-07-02 | Fuji Xerox Co Ltd | Presentation recording device |
JP4112968B2 (en) * | 2002-12-26 | 2008-07-02 | 富士通株式会社 | Video text processing device |
US7412453B2 (en) * | 2002-12-30 | 2008-08-12 | International Business Machines Corporation | Document analysis and retrieval |
US7904823B2 (en) * | 2003-03-17 | 2011-03-08 | Oracle International Corporation | Transparent windows methods and apparatus therefor |
US8904267B2 (en) * | 2003-10-14 | 2014-12-02 | International Business Machines Corporation | Retrieving slide show content from presentation documents |
JPWO2005043915A1 (en) * | 2003-10-31 | 2007-05-10 | 株式会社Kddiメディアウィル | Video analysis device and video failure detection device |
US20060047637A1 (en) * | 2004-09-02 | 2006-03-02 | Microsoft Corporation | System and method for managing information by answering a predetermined number of predefined questions |
US7554576B2 (en) * | 2005-06-20 | 2009-06-30 | Ricoh Company, Ltd. | Information capture and recording system for controlling capture devices |
-
2004
- 2004-10-26 US US10/973,698 patent/US20060090123A1/en not_active Abandoned
-
2005
- 2005-10-26 JP JP2005311276A patent/JP2006127518A/en active Pending
-
2009
- 2009-06-10 US US12/482,315 patent/US9875222B2/en active Active
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5848409A (en) * | 1993-11-19 | 1998-12-08 | Smartpatents, Inc. | System, method and computer program product for maintaining group hits tables and document index tables for the purpose of searching through individual documents and groups of documents |
US6332144B1 (en) * | 1998-03-11 | 2001-12-18 | Altavista Company | Technique for annotating media |
US6646655B1 (en) * | 1999-03-09 | 2003-11-11 | Webex Communications, Inc. | Extracting a time-sequence of slides from video |
US20010020954A1 (en) * | 1999-11-17 | 2001-09-13 | Ricoh Company, Ltd. | Techniques for capturing information during multimedia presentations |
US7167191B2 (en) * | 1999-11-17 | 2007-01-23 | Ricoh Company, Ltd. | Techniques for capturing information during multimedia presentations |
US7010751B2 (en) * | 2000-02-18 | 2006-03-07 | University Of Maryland, College Park | Methods for the electronic annotation, retrieval, and use of electronic images |
US7149755B2 (en) * | 2002-07-29 | 2006-12-12 | Hewlett-Packard Development Company, Lp. | Presenting a collection of media objects |
US20040247206A1 (en) * | 2003-02-21 | 2004-12-09 | Canon Kabushiki Kaisha | Image processing method and image processing system |
US20040202349A1 (en) * | 2003-04-11 | 2004-10-14 | Ricoh Company, Ltd. | Automated techniques for comparing contents of images |
US7392469B1 (en) * | 2003-05-19 | 2008-06-24 | Sidney Bailin | Non-intrusive commentary capture for document authors |
US20050078868A1 (en) * | 2003-09-26 | 2005-04-14 | William Chen | Method and apparatus for summarizing and indexing the contents of an audio-visual presentation |
US20060074895A1 (en) * | 2004-09-29 | 2006-04-06 | International Business Machines Corporation | Method and system for extracting and utilizing metadata to improve accuracy in speech to text conversions |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060250508A1 (en) * | 2005-05-09 | 2006-11-09 | Chia-Kai Chang | Method for inserting a picture into a video frame |
US20080007567A1 (en) * | 2005-12-18 | 2008-01-10 | Paul Clatworthy | System and Method for Generating Advertising in 2D or 3D Frames and Scenes |
US20080042923A1 (en) * | 2006-08-16 | 2008-02-21 | Rick De Laet | Systems, methods, and apparatus for recording of graphical display |
US8878833B2 (en) | 2006-08-16 | 2014-11-04 | Barco, Inc. | Systems, methods, and apparatus for recording of graphical display |
US20080263010A1 (en) * | 2006-12-12 | 2008-10-23 | Microsoft Corporation | Techniques to selectively access meeting content |
US20080183820A1 (en) * | 2007-01-29 | 2008-07-31 | Fuji Xerox Co., Ltd. | Device Orchestration |
US20080276159A1 (en) * | 2007-05-01 | 2008-11-06 | International Business Machines Corporation | Creating Annotated Recordings and Transcripts of Presentations Using a Mobile Device |
US20130036235A1 (en) * | 2010-04-16 | 2013-02-07 | Pablo Lopez Garcia | Method of transmission of visual content |
US8626487B2 (en) | 2010-09-08 | 2014-01-07 | Sharp Kabushiki Kaisha | Translation apparatus, translation method, computer program, and recording medium |
Also Published As
Publication number | Publication date |
---|---|
US9875222B2 (en) | 2018-01-23 |
US20090254828A1 (en) | 2009-10-08 |
JP2006127518A (en) | 2006-05-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9875222B2 (en) | Capturing and storing elements from a video presentation for later retrieval in response to queries | |
US7466858B2 (en) | Methods for slide image classification | |
US9244923B2 (en) | Hypervideo browsing using links generated based on user-specified content features | |
US7131059B2 (en) | Scalably presenting a collection of media objects | |
US7616840B2 (en) | Techniques for using an image for the retrieval of television program information | |
US7149755B2 (en) | Presenting a collection of media objects | |
US7793212B2 (en) | System and method for annotating multi-modal characteristics in multimedia documents | |
US10438050B2 (en) | Image analysis device, image analysis system, and image analysis method | |
US7167191B2 (en) | Techniques for capturing information during multimedia presentations | |
US9569428B2 (en) | Providing an electronic summary of source content | |
US20160014482A1 (en) | Systems and Methods for Generating Video Summary Sequences From One or More Video Segments | |
US20020051077A1 (en) | Videoabstracts: a system for generating video summaries | |
US7908141B2 (en) | Extracting and utilizing metadata to improve accuracy in speech to text conversions | |
CN105786969B (en) | Information display method and device | |
JPH07219971A (en) | Indexing device of table of contents | |
US8931002B2 (en) | Explanatory-description adding apparatus, computer program product, and explanatory-description adding method | |
JP5894149B2 (en) | Enhancement of meaning using TOP-K processing | |
CN112911324B (en) | Content display method and device for live broadcast room, server and storage medium | |
US10691879B2 (en) | Smart multimedia processing | |
US20180151178A1 (en) | Interactive question-answering apparatus and method thereof | |
CN104185041A (en) | Video interaction advertisement automatic generation method and system | |
CN111723235B (en) | Music content identification method, device and equipment | |
CN114625918A (en) | Video recommendation method, device, equipment, storage medium and program product | |
US20150006497A1 (en) | Slideshow Builder and Method Associated Thereto | |
CN113778717A (en) | Content sharing method, device, equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: FUJI XEROX CO., LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:DENOUE, LAURENT;TREVOR, JONATHAN J.;HILBERT, DAVID M.;AND OTHERS;REEL/FRAME:016241/0244 Effective date: 20050131 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |