US6580438B1 - Systems and methods for maintaining uniformity in a presentation environment - Google Patents
Systems and methods for maintaining uniformity in a presentation environment Download PDFInfo
- Publication number
- US6580438B1 US6580438B1 US09/444,204 US44420499A US6580438B1 US 6580438 B1 US6580438 B1 US 6580438B1 US 44420499 A US44420499 A US 44420499A US 6580438 B1 US6580438 B1 US 6580438B1
- Authority
- US
- United States
- Prior art keywords
- presentation
- style
- presentation element
- information
- font
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/103—Formatting, i.e. changing of presentation of documents
Definitions
- This invention is related to presentation systems.
- presentation authoring software allows a user to input data which is then translated into “slides.” These slides are then electronically displayed to an audience using a presentation system during a presentation to facilitate and supplement a speaker's presentation.
- presentation systems allow a pre-assembled presentation to be replayed. Specifically, these presentation systems allow a user to navigate, via a graphical user interface or keyboard control, to a previous slide, a next slide or another slide within the pre-assembled presentation, or cycle through unrelated slides, i.e., skip slides, to present a more relevant slide of the presentation.
- presentation authoring software and presentation systems work well for sequential presentations drawing their content from a common source, this technology fails to manipulate the presentations to create a uniform appearance when “slides” are drawn from varying sources.
- numerous file types are available from which a presenter can compile information to generate “slides.”
- context elements may be background color or texture, orientation of the document, transition of the elements into and out of the presentation, font, font size, bullet style, or the like.
- the systems and methods of this invention are more attuned to a presenter's environment and allow a majority of the user's attention, as well as the presenter's attention, to be dedicated to tasks other than manipulation of the presentation system, especially when dealing with electronic media of varying formats.
- Selecting material for viewing and subsequent display to an audience should be as simple as retrieving a transparency, yet sufficiently flexible, functional and dynamic to deal with the dynamic content afforded by available and anticipated presentation and multimedia systems and formats.
- This invention provides presentation control systems and methods that provide support for manipulating the context of elements within a presentation.
- the systems and methods of this invention receive a request to display a presentation element, e.g., a slide.
- a presentation element e.g., a slide.
- the context of the slide is then manipulated in accordance with a style profile, or template, that can ensure a uniform display characteristic between presentation elements of varying formats.
- an identification-carrying device is presented to a sensor of a presentation control system.
- the presentation control system associates an identification on the identification-carrying device with a presentation element.
- the presenter then has a variety of options for modifying the style, or context of the presentation element.
- the stylized presentation element is then presented for display.
- This invention separately provides systems and methods for assisting users in presenting electronic media.
- This invention separately provides systems and methods that allow a user to interact with a presentation system via a tangible identification-carrying device.
- This invention separately provides systems and methods that allow users to create tangible identification-carrying devices containing style identifiers to be applied to the context of presentation elements.
- the presentation systems and methods of this invention provide a user with a mechanism that provides ease of use and better support to the user during, for example, a presentation.
- the presentation systems and methods of this invention can use physical objects, such as note cards, that contain, or are provided with, identifiers that represent styles to be applied to presentation elements.
- Mnemonics which clearly express the style, e.g., background color, are represented by an object affixed to the physical object.
- users can incorporate additional notes or graphics which may supplement the mnemonic, such as notes describing which background is better for a particular presentation element, which further aid in the display characteristics of the presentation.
- a presenter may distribute a set of cards on a table or podium which contain mnemonics representing the material for the presentation.
- the presenter may also distribute a set of cards that contain identifications of styles in which to present those elements.
- the presenter can then select a presentation element card, introduce it to the sensor device of the presentation system, and have the context of current or subsequent presentation elements modified to create a custom appearance by introducing a “style card” to the sensor device.
- the presentation control systems and methods according to this invention off-load activities from the presenter's over-taxed cognitive system to under-utilized areas, such as the peripheral ranges of the presenter's senses.
- the presentation control systems and methods according to this invention enable a user to manipulate the context of a presentation while very little attention is dedicated to manipulating a particular user interface.
- the presentation systems and methods according to this invention engage a wider range of human perception and enable a larger degree of low-intentional interaction than is found in current presentation systems, methods and interfaces.
- the presentation systems and methods of this invention recognize a direct physical action can be more efficient and reliable than interactions with a graphical user interface. For example, using the knob on a computer's speaker to adjust its loudness is more easily accomplished than finding and adjusting a slider in a control window.
- FIG. 1 is a functional block diagram showing a first embodiment of the presentation system according to this invention
- FIG. 2 shows a first exemplary context identification-carrying device
- FIG. 3 illustrates a second exemplary context identification-carrying device
- FIG. 4 illustrates a third exemplary context identification-carrying device
- FIG. 5 is a flowchart outlining one exemplary embodiment of a method for controlling the context of presentation elements during a presentation
- FIG. 6 is a flowchart outlining one exemplary embodiment of a method for controlling the context of a stored presentation according to this invention.
- FIG. 7 is a flowchart outlining one exemplary embodiment of a method for automatically controlling the context of a presentation according to this invention.
- FIG. 8 is a flowchart outlining in greater detail the applied style steps of FIGS. 5-7.
- FIG. 9 is a flowchart outlining one exemplary embodiment of a method for controlling font attributes according to this invention.
- Operation of the presentation control systems and methods according to this invention reduce the amount of attention required from a presenter to maintain an electronic system supporting the presentation, while allowing immediate access to an array of presentation aids with differing formats.
- a presenter generates a set of presentation aids, or elements, such as a slide show, to supplement a presentation to be given to an audience.
- the presenter can then associate at least one tangible sensible object, such as a card carrying a sensible code or other identification element, with at least one presentation aid element of the presentation as described in co-pending U.S. patent applications Ser. No. 09/152,677 filed Sep. 14, 1998now U.S. Pat. No. 6,195,093, and Ser. No. 09/393,962, filed Sep. 10, 1999, each incorporated herein by reference in their entirety.
- the presenter thus possesses physical objects that correspond to the presentation elements.
- the presenter presents an object, such as one of the cards, to a sensor of the presentation system which causes the presentation system to retrieve and display the slide corresponding to the selected card to, for example, an audience.
- the presenter can also generate, or input, context control identification devices for the presentation.
- This context control information can then be applied, similarly to the physical objects corresponding to the presentation elements, to manipulate the display characteristics of the presentation elements, or slides, of the presentation.
- the systems and methods of this invention modify the context of the presentation element based on a style chosen by the presenter.
- Each modified presentation element is then presented, via a presentation display device, while maintaining a specific appearance or display characteristic.
- the user introduces one of the tangible, sensible identification-carrying devices into the sensible area of a sensor of the presentation control system of this invention.
- the user has a variety of options for manipulating the style of the sensed presentation element associated with the tangible, sensible identification-carrying device.
- the presentation control system manipulates the presentation element in accordance with the style selected by the presenter and presents the updated presentation element with a modified display characteristic on a presentation display device. In this manner, the presenter's attention can be redirected from the presentation system and especially away from the user interface and to an audience.
- FIG. 1 illustrates one embodiment of the components of a presentation control system 100 used during a presentation.
- the presentation control system 100 includes an I/O interface 110 , a controller 120 , an identification controller 130 , a presentation controller 140 , a history memory 150 , a style memory 160 , a stylizer 170 , a presentation element type identifier 180 , a style memory portion 220 with style identifications 222 - 226 , a presentation element memory portion 200 with presentation elements 202 - 206 , and a presentation element identification memory portion 210 with presentation element identifiers 212 - 216 , all stored in memory 190 , an identification sensing device 300 , a presentation display device 310 , a user input device 320 , one or more tangible sensible identification-carrying devices 330 , one or more tangible sensible style identification-carrying devices 340 and links 105 and 305 .
- the systems and methods of this invention will be described in relation to an embodiment in which tangible sensible identification-carrying devices are presented to the presentation control system 100 .
- the presentation control system 100 then retrieves presentation elements corresponding to a presentation element identifier that is sensed from the tangible sensible identification-carrying device.
- the systems and methods of this invention will work equally well on any known, or later developed, presentation control system.
- a user need not select a presentation element for presentation with the use of a tangible, sensible identification-carrying device.
- the user can use an input device, such as a keyboard, a mouse, or a voice recognition system, to control the retrieval, and subsequent display, of a presentation element.
- tangible sensible identification-carrying devices 330 have been produced, for example by the systems and methods described in co-pending U.S. patent applications Ser. No. 09/156,677 or Ser. No. 09/393,962, a user can use the identification-carrying devices 330 during a presentation.
- the particular tangible sensible identification-carrying device 330 for a desired presentation element is identified by the user based on the mnemonic for that presentation element that is attached to, or otherwise associated with, one of the tangible sensible identification-carrying devices 330 .
- That tangible sensible identification-carrying device 330 is placed, at least momentarily, in the sensible area of the identification sensing device 300 .
- the identification sensing device 300 outputs, via the I/O interface 110 and the link 305 , the presentation element identifier associated with that particular tangible sensible identification-carrying device 330 .
- the identification controller 130 associates the presentation element identifier information received from the identification-carrying device 330 with a particular presentation element based on the presentation element identifier stored in the presentation element identification memory portion 210 .
- the identification controller 130 then outputs a signal to the presentation controller 140 indicating the selected presentation element.
- the presentation controller 140 retrieves, via the link 105 , a particular presentation element indicated by the received signal.
- a style or certain display characteristic, should be applied to the particular presentation element. If a style is not desired, a signal representing the selected presentation element is output via the link 105 , the I/O interface 110 and link 305 to the presentation display device 310 and displayed.
- the presentation element type identifier 180 determines the format of the particular presentation element.
- presentation element types may be in a POWERPOINT® format by Microsoft, a PERSUASION® format by Adobe, a CORELSHOW® format by Corel, or an image element such as a GIF, JPEG, bitmap, or the like, or any other known or later developed format that can be used in conjunction with a presentation.
- the presentation element type identifier 180 queries the style memory 160 to determine if attribute tags, i.e., information that identifies display characteristics of the presentation, such as background color or the like, for that particular presentation format are available.
- the presentation system 100 assembles a list of predetermined available styles for selection by the user.
- available styles that can be applied to a particular presentation element are of background color, background texture, background image, presentation element transition, font, font size, font spacing, font color, font family, font style, word spacing, letter spacing, text alignment, text indent, line spacing, bullet style, bullet size, bullet color, list style, presentation element orientation, presentation element display time, presentation element size, border color, border width, visual effect, audio effect, design selection, template selection or any other style which affects the context or display characteristics of the presentation element.
- the stylizer 170 Upon selecting a style, via, for example, the user input device 320 , such as a mouse, keyboard, voice control system, or the like, or through the use of a style identification-carrying device 340 , the stylizer 170 applies the selected style to the presentation element(s).
- the user input device 320 such as a mouse, keyboard, voice control system, or the like
- a style identification-carrying device 340 the stylizer 170 applies the selected style to the presentation element(s).
- the user can enter, via user input device 320 , or retrieve, for example, from a database or data source collocated or remote from the presentation control system 100 , a custom style.
- a custom style As previously noted, the examples discussed above are illustrative in nature and are not intended to limit the scope or combination of styles that can be applied to any particular presentation element type.
- the stylizer 170 applies the selected or custom style to the presentation element with the cooperation of controller 120 .
- the stylizer 170 compares the detected presentation element type to a presentation element attribute table, which can be stored in style memory 160 . If the presentation element attribute table for the particular presentation element type contains attribute tags which correspond to the attribute tags affected by the style selection of the presenter, the stylizer 170 replaces the attributes of the tags with new attributes that correspond to the selected style. For example, if the presentation element type is a product of a well known presentation authoring system, the presentation element will contain tags, or identifiers, e.g., function calls, followed by attributes that describe various features of the presentation element. Table 1 illustrates some of the common tags used to identify attributes within a presentation element.
- the background color will be identified by the ActiveWindow.Selection.SlideRange.Background.Fill.ForeColor.RGB value tag.
- the stylizer 170 then replaces the attributes of tags with the new attributes that correspond to the selected style.
- the stylizer 170 updates the global presentation context style attributes. This allows all presentation elements within the presentation to be aligned with the currently selected style attribute.
- the stylizer 170 in cooperation with the presentation controller 140 , outputs a modified presentation element to the presentation display device 340 .
- the presentation display device will show the bitmap image superimposed on, for example, a blue background.
- a signal representing the stylized presentation element is then output via the link 305 and the I/O interface 110 to the presentation display device 310 and displayed.
- the stylizer 170 in cooperation with the presentation controller 140 , outputs a modified presentation element to the presentation display device 340 .
- the presentation display device 340 will show the updated presentation element containing the new font size and style.
- a user can, for example, select a style that indicates all presentation elements should be displayed in, for example, New Times Roman font with a font size of 14 point.
- the system Upon selection of a presentation element, the system will determine a first text box within the presentation element and retrieve its identification. Then, in order to maintain the spatial relationship existing in the presentation element, a determination will be made as to whether the text box has a border, or frame. If a border is present, the system retrieves the dimensions for the text box and records them in association with the text box identifier. The stylizer 170 then applies the new font size and shape to the text within the selected text box.
- the stylizer 170 will retrieve the saved text box identifier and dimensions previously stored. If after the new font size and shape attributes have been applied to the selected text box, the borders are larger than the stored borders, the stylizer 170 will decrease the font size accordingly until the original boundary size is restored. Alternatively, if the text box boundaries are smaller than the original stored text box boundaries, the stylizer 170 will proportionally increase the font size to achieve the same size text box that appeared in the original presentation. Thus, the spatial relationships between the text boxes are maintained. This process can continue until each text box has been subjected to the new style attribute.
- the spatial relationships with a presentation element may not be critical. Therefore, if, for example, the spatial relationships within a presentation element are irrelevant, a user could alternatively force all text to a particular font or font size in accordance with the user's selected, or customized style attribute.
- the stylized presentation element along with any associated style identification, is stored in, or in a combination of, the memory 190 and the history memory 150 . This allows the presentation system 100 to track at least one previous style that has been applied to at least one particular presentation element.
- the presentation control system 100 works equally well for presentations that have been stored with an associated style identification. For example, if an exemplary presentation has one or more presentation elements 202 - 206 , those presentation elements are identified by corresponding presentation element identifiers 212 - 216 , respectively. Furthermore, each of those presentation elements 202 - 206 has an associated style identification 222 - 226 , respectively. These style identifications 222 - 226 correspond to previous styles that were associated with the respective presentation element 202 - 206 . Therefore, when subsequently displaying a particular presentation element, the presentation controller 140 can query the memory 190 to determine if a style identifier is associated with the selected presentation element. If a style identifier is associated with the selected presentation element, the presentation controller 140 can retrieve the previously stored style information and forward a signal representing the selected presentation element and associated style to the presentation display device 310 .
- the presentation control system 100 can query the user whether a style is to be applied to the selected presentation element. If no style is desired for the selected presentation element, a signal representing the selected, unmodified presentation element is output via the link 305 and I/O interface 10 to the presentation display device 310 and displayed.
- the presentation element type identifier 180 identifies the presentation element type. Alternatively, the identification controller 130 determines whether any style information is currently associated with that particular presentation element. Therefore, the presentation control system 100 can determine if styles have been previously associated with the presentation element. Furthermore, the presentation control system 100 can determine if the user has selected a style to be applied generically to one or more particular presentation element types. If a particular style is associated with the selected presentation element, or the presentation element type matches a flag set by the user indicating all elements of this type are to have a particular style, the associated style attributes can be displayed to the user. The user then has the option of applying a new style to the selected presentation element, or, alternatively, displaying the presentation element using the previously associated style information.
- the presentation control system 100 is also capable of operating in an “auto stylize” mode. Upon being placed in this mode, the presentation control system 100 receives a request to display a particular presentation element. Upon receiving this request, the presentation element type identifier 180 , in conjunction with the controller 120 , determines if a style has previously been associated with the selected presentation element. If an associated style for the selected presentation element is not available, the stylizer 170 , in conjunction with the presentation element type identifier 180 , analyzes the presentation element to create a style based on that presentation element's display characteristics.
- the stylizer 170 creates a style based on the display characteristics of the requested presentation element. For example, the stylizer 170 can query the presentation element and identify the attributes associated with the “Background” and “Transition” tags. The attributes of these tags are then associated with a new style having, for example, a white background and transition of fade in/out. Therefore, this created style can be used, and applied to, subsequently requested presentation elements, with or without interaction on the user's part.
- the user can establish a half-automatic, half-manual mode of operation. For example, the user can identify that only certain display characteristics be queried by the stylizer 170 . If the stylizer 170 finds attributes associated with the limited set of display characteristics, the stylizer 170 creates a custom style based on the attributes associated with those display characteristics. However, if none of the limited set of display characteristics are found in the requested presentation element, the presentation control system 100 could default to, for example, a predefined background color with no transition parameters.
- a style has been previously associated with a presentation element, the user is not required to use that style for display of the presentation element.
- the user can select a style from the available styles based on the detected element type, or create a custom style, as described above.
- tangible sensible identification-carrying device 330 can be any device capable of carrying the mnemonic relating to the presentation element associated with the presentation element identifier contained in that tangible sensible identification-carrying device 330 .
- tangible sensible identification-carrying device 340 can be any device capable of carrying a mnemonic relating to a style identification that can be applied to a particular presentation element.
- the tangible sensible identification-carrying device 330 and the style identification-carrying device 340 need not be limited to printed media.
- the tangible sensible identification-carrying device and the style identification-carrying device can be a physical device such as a microchip or other device capable of carrying information based on its mechanical, electrical, or magnetic properties.
- a tangible sensible identification-carrying device and/or style identification-carrying device can be an ultrasonic transmitter and the identification sensing device an ultrasonic receiver.
- the identification sensing device Upon placing the ultrasound emitting identification-carrying device in the sensible area of the identification sensing device 300 , the identification sensing device reads the presentation element identifier, or style identification, respectively, from the identification-carrying device.
- the tangible sensible identification-carrying device 330 and style identification-carrying device 340 can maintain, transmit or respond to an electric or magnetic field.
- the identification sensing device 300 would then produce an electromagnetic field into which the identification-carrying device 330 can be placed.
- the identification-carrying device When the identification-carrying device is placed in a sensible area of the identification sensing device, the identification-carrying device becomes capacitively, inductively or otherwise electromechanically coupled to the identification sensing device, and the identification can be extracted from the identification-carrying device and correlated to the particular presentation element identifier or style identifier.
- one of the identification-carrying devices can be a chip and the identification sensing device a chip reader.
- the chip when placed in the sensible area of the identification sensing device could then be read and interpreted by the presentation control system 100 .
- the systems and methods of this invention work equally well in conjunction with known, or later developed, presentation authoring and displaying systems.
- a user can implement traditional techniques, for example selecting entries with a keyboard and/or mouse, to select a presentation element and/or a desired style.
- the identification-carrying device 330 , the style identification-carrying device 340 , the identification sensing device 300 and the user input device 320 can be any device, or combination of devices, which is capable of receiving a user's selection of a presentation element, and/or style to be applied thereto.
- style identification-carrying device 340 and identification-carrying device 330 can be incorporated into one tangible sensible identification-carrying device. Additionally, more than one style or presentation element can also be associated with a single tangible sensible identification-carrying device.
- the presentation display device 310 can be any device capable of displaying electronic media or multimedia to an audience, such as an overhead projector, a large screen TV, a video projector, a slide projector, or the like.
- the links 105 and 305 can be a wired or wireless link or any known or later developed element that is capable of supplying electronic data to and from the connected elements.
- a tangible sensible identification-carrying device is a card containing a thumbnail image forming the mnemonic where the thumbnail image represents a particular presentation element, or style, and at least one barcode forming the presentation element identifier, or style, respectively. While the presentation system 100 of this invention is by no means limited to this particular exemplary embodiment, this example is chosen for its simplicity.
- This exemplary embodiment uses a barcode scanner as the identification sensing device 300 and a card, such as a 3 ⁇ 5 card as the identification-carrying, or style identification-carrying device.
- a printer for example at the direction of the user and/or the presentation control system, produces a printed card or page, i.e., the tangible sensible identification-carrying device or style identification-carrying device, as described in co-pending U.S. patent application Ser. No. 09/152,677 now U.S. Pat. No. 6,195,093.
- the tangible sensible identification-carrying devices contain a barcode representing the identifiers printed on the card or page.
- a thumbnail image representing the identifiers as the mnemonic is also printed on the card.
- FIGS. 2-4 illustrate exemplary printed cards that the inventors have developed as exemplary style identification-carrying devices 340 .
- FIGS. 2-4 illustrate how exemplary style identification-carrying devices 340 may appear.
- FIG. 2 illustrates an exemplary style identification-carrying device 500 for setting the background in the presentation to blue.
- the style identification-carrying device 500 comprises a mnemonic 502 corresponding to the style identified in the barcode 504 .
- the style identification-carrying device 500 contains a graphical mnemonic 506 which shows an exemplary presentation element 508 superimposed on a blue background 509 .
- at least the current presentation element is associated with a style element that directs the presentation control system to display a blue background.
- FIG. 3 illustrates an exemplary style identification-carrying device 510 that controls the display characteristics for the introduction of the presentation element on the presentation display device 310 , and the transition of that presentation element to another presentation element, or close of the presentation.
- the style identification-carrying device 510 comprises a textual mnemonic 512 , a graphical mnemonic 514 and a barcode 516 corresponding to the identification of the particular style stored in style memory 160 .
- the style identification for this style identification carrying device directs the presentation control system to fade-in a requested presentation element, and fade-out the requested presentation element when a new presentation element is requested or the presentation closed.
- FIG. 4 illustrates another exemplary style identification-carrying device 520 that modifies the presentation element to a predetermined size.
- the style identification-carrying device comprises a textual mnemonic 522 , a graphical mnemonic 524 and a barcode representing the style identification 526 .
- the stylizer 170 reduces or enlarges, as appropriate, the presentation element to the size defined by the style attributes.
- the barcode as a style identifier, is associated with one or more mnemonics relating to a particular style identification-stored in style memory 160 .
- the style identifier need not be linked to a particular style resident on the presentation control system 100 .
- the style identifier identifies any style, or set of styles, in any presentation.
- the style identifier can be, for example, a hyperlink, a file name, a page number, a uniform resource locator (URL), a full text search keyword or other query syntax, a database address and/or a path to a particular memory location inside or outside of the presentation control system.
- the style identifier is in no way limited to identifying a particular style in one particular loaded presentation, or present on the presentation control system, but can also be a path to at least one memory location containing at least one style to be applied to at least one presentation element.
- the only requirement of the style identifier is that it contain sufficient information that the presentation control system can, possibly along with static information, identify the desired style.
- static information can be, for example, the current presentation of the current presentation element being displayed.
- FIG. 5 is a flowchart outlining one embodiment of a method for applying a style to a presentation element according to this invention.
- Control begins in step S 100 , and continues to step S 110 , where a presentation is opened.
- step S 120 a request by the user for a presentation element is detected.
- step S 130 the presentation element identifier is associated with a particular presentation element.
- step S 140 the presentation element identifier is associated with a particular presentation element.
- step S 140 a determination is made whether a style should be applied to the selected presentation element. If a style is to be applied to the selected presentation element, control continues to step S 150 . Otherwise, control jumps to step S 210 .
- step S 150 the type of the presentation element is detected. Then, in step S 160 , the style available for the detected element type are retrieved. Next, in step S 170 , the user is queried whether a custom style is desired, as opposed to the styles retrieved that correspond to the detected element type. If a custom style is desired, control continues to step S 180 . Otherwise, control jumps to step S 190 .
- step S 180 the user enters, or retrieves, the style to be applied to the particular presentation element. Control then jumps to step S 200 .
- step S 190 a style is selected by the user from the available style retrieved and corresponding to the detected presentation element type. Control then continues to step S 200 .
- step S 200 the style information is applied to the particular presentation element.
- step S 210 the presentation element is displayed.
- step S 220 the history information, as well as any corresponding style identification information, is stored.
- step S 230 the user is queried whether another presentation element is to be presented. If another presentation element is desired for presentation, control jumps back to step S 120 . Otherwise, control continues to step S 240 , where the control sequence ends.
- FIG. 6 is a flowchart outlining one exemplary embodiment of a presentation method used in conjunction with stored presentation elements that may, or may not, already have associated styles.
- Control begins in step S 1000 and continues to step S 1100 , where a presentation is opened.
- step S 1200 a presentation element identifier is received and detected.
- step S 1300 the presentation element identifier is associated with a particular presentation element.
- step S 1400 the presentation element identifier is associated with a particular presentation element.
- step S 1400 a determination is made whether any style information should be applied to the requested presentation element. If style information should be applied to the element, control continues to step S 1500 . Otherwise, control jumps to step S 2400 .
- step S 1500 the presentation element type is detected.
- step S 1600 a determination is made whether any style identification information is currently associated with that particular presentation element type. If a style is currently associated with the presentation element, control continues to step S 1700 . Otherwise, control jumps to step S 1900 .
- step S 1700 the style information associated with the presentation element is displayed.
- step S 1800 a determination is made whether to apply a new style. If a new style is to be applied, control continues to step S 1900 . Otherwise, if the currently associated style is to be used, control jumps to step S 2400 .
- step S 1900 a list of available styles is retrieved based on the detected element type.
- step S 2000 a determination is made whether a custom style is desired. If a custom style is desired, control continues to step S 2100 . Otherwise control jumps to step S 1450 .
- step S 1422 the custom style is entered by a user, or retrieved from a database. Control then continues to step S 2300 .
- step S 2200 a style is selected from the available styles list. Then, in step S 2200 , this style is applied to the selected presentation element. Control then continues to step S 2400 .
- step S 2400 the presentation element, modified by the selected style, if appropriate, is displayed.
- step S 2500 the history and/or style information associated with the displayed presentation element is stored.
- step S 2600 a determination is made whether another presentation element is desired for presentation. If another presentation element is desired for presentation, control jumps back to step S 1200 . Otherwise, control continues to step S 2700 , where the control sequence ends.
- step S 1500 it is not necessary to detect the element type in step S 1500 .
- step S 1500 can be eliminated.
- FIG. 7 is a flowchart outlining one exemplary embodiment of an “auto stylize” method for controlling style information in accordance with this invention.
- Control begins in step S 3000 , and continues to step S 3100 where a presentation is opened. Then, in step S 3200 , an input presentation element identifier is detected. Next, in step S 3300 , a presentation element is associated with the input presentation element identifier. Control then continues to step S 3400 .
- step S 3400 a determination is made whether a style is associated with the selected presentation element. If a style is currently associated with the selected presentation element, control jumps to step S 3700 . Otherwise, control continues to step S 3500 .
- step S 3500 when style information is not currently associated with the selected presentation element, the presentation element is analyzed to create a style based on that particular element's display characteristics.
- step S 3600 a determination is made whether the created style is acceptable. If the created style is not acceptable, control jumps to step S 3800 where a new style can be selected or input. Otherwise, control jumps to step S 4400 .
- step S 3700 a determination is made whether the associated style should be applied to the selected presentation element. If the associated style is to be applied to the selected presentation element, control jumps to step S 4300 . Otherwise, control continues to step S 3800 .
- the style may already have been applied to or associated with the selected presentation element. In this case, control would jump directly to step S 4400 to display the presentation element.
- step S 3800 the presentation element type is detected. Then, in step S 3900 , a list of available styles available for the detected presentation element are displayed, or summarized for a user.
- step S 4000 a determination is made whether a custom style, other than the listed available styles, is desired. If a custom style is desired, control continues to step S 4100 . Otherwise, control jumps to step S 4200 .
- step S 4100 the user enters, or retrieves, custom style identification information. Control then continues to step S 4300 .
- step S 4200 a style is selected from the available styles list, or, for example, from a menu. Control then continues to step S 4300 .
- step S 4300 the style is applied to the presentation element.
- the method need not reapply the style to the presentation element, but can simply retrieve, and display, the already applied style. Control then continues to step S 4400 .
- step S 4400 the presentation element is displayed. Then, in step S 4500 , the history, and associated style identification information, are stored in relation to the selected presentation element. Next, in step S 4600 , a determination is made whether another presentation element is desired for presentation. If another presentation element is desired, control jumps back to step S 3200 . Otherwise, control continues to step S 4700 where the control sequence ends.
- FIG. 8 is a flowchart outlining one exemplary embodiment of the applying style to element step shown in FIGS. 5-7 in greater detail.
- Control commences in step S 5000 , and continues to step S 5100 , where the detected presentation element type is compared to an element attribute table. Then, in step S 5200 , a determination is made whether tags or identifiers corresponding to the desired style attributes are present. If the tags are present, control continues to step S 5300 . Otherwise, control jumps to step S 5400 .
- step S 5300 the current attributes of the tags are replaced with new attributes which correspond to the selected style. Control then jumps to step S 5500 where the control sequence returns to the main control sequence.
- a presentation context style attribute is updated.
- This presentation context style attribute governs global display characteristics of the presentation element.
- the context style attribute can be, for example, a background color, image, texture, or the like, or matting, that surrounds the selected presentation element to fill an entire presentation display device screen.
- the global context style attribute can be a presentation element transition as previously described. Control then continues to step S 5500 .
- FIG. 9 illustrates step S 5300 of FIG. 8 in greater detail.
- FIG. 9 illustrates an embodiment in which portions of the presentation element contain borders.
- portions of the presentation element contain borders.
- text boxes can have borders. This flowchart outlines how the spatial relationships are maintained between portions of a presentation element while still allowing for the attributes to be adjusted in accordance with a selected style.
- Control begins in step S 6000 and continues to step S 6100 .
- step S 6100 a portion, e.g., a text box within a presentation element is identified.
- step S 6200 a determination is made whether the identified portion has a border. If the identified portion has a border, control continues to S 6300 . Otherwise, control jumps to step S 7100 .
- step S 6300 the border dimensions for the identified portion are retrieved.
- step S 6400 the dimensions are associated with the identified portion.
- step S 6500 the attributes of tags with new attributes corresponding to style for the selected portion are replaced. Control then continues to step S 6600 .
- step S 6600 the current dimensions associated with the selected portion are retrieved.
- step S 6700 a determination is made whether the borders are larger than the stored original dimensions. If the borders are larger than the stored original dimensions, control continues to step S 6800 , otherwise control jumps to step S 6900 .
- step S 6800 the font size is decremented and control returns to step S 6700 .
- step S 6900 a determination is made whether the borders are smaller than the stored dimensions. If the borders are smaller than the stored dimensions, control continues to step S 7000 , otherwise control jumps to step S 7100 . In step S 7000 , the font size is incremented and control returns to step S 6900 . Control then continues to step S 7100 .
- step S 7100 a determination is made whether all portions within the presentation element have been identified. If all portions have been identified, control continues to step S 7200 where the control sequence ends. Otherwise, control returns to step S 6100 .
- the presentation system is preferably implemented either on a single program general purpose computer or separate program general purpose computers, with associated tangible sensible identification-carrying devices.
- the presentation system can also be implemented on a special purpose computer, a programmed microprocessor or microcontroller and peripheral integrated circuit elements, an ASIC, or other integrated circuit, a digital signal processor, hard wired electronic or logic circuit such as a discrete element circuit, a programmable logic device such as a PLD, PLA, FPGA, PAL, or the like.
- any device capable of implementing a finite state machine that is in turn capable of implementing the flowcharts shown in FIGS. 5-9 can be used to implement the presentation system.
- the disclosed method may be readily implemented in software using object or object-oriented software development environments that provide portable source code that can be used on a variety of computer or work station hardware platforms.
- the disclosed presentation system may be implemented partially or fully in hardware using standard logic circuits or a VLSI design. Whether software or hardware is used to implement the systems in accordance with this invention is dependent on the speed and/or efficiency requirements of the system, the particular function, and the particular software or hardware systems or microprocessor or microcomputer systems being utilized.
- the presentation system and methods described above can be readily implemented in hardware or software using any known or later developed systems or structures, devices and/or software by those skilled in the applicable art without undue experimentation from the functional description provided herein together with a general knowledge of the computer arts.
- the disclosed methods may be readily implemented as software executed on a programmed general purpose computer, a special purpose computer, a microprocessor, or the like.
- the methods and systems of this invention can be implemented as a routine embedded in a personal computer, as a resource residing on a server or graphics workstation, as a routine embedded in a dedicated presentation control system, or the like.
- the presentation system can also be implemented by physically incorporating the system and method into a software and/or hardware system, such as the hardware and software systems of a graphics workstation or dedicated presentation control system.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Controls And Circuits For Display Device (AREA)
- User Interface Of Digital Computer (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
TABLE 1 | ||
PRESENTATION ELEMENT TYPE |
HTML | ||
Tag and Cascading Style | ||
PowerPoint ® | Sheet Access to Attribute | |
API Access to Attribute | HTML Tag(s) listed first; | |
ATTRIBUTE | Visual Basic Function Call(s): | CSS properties listed next |
background color | ActiveWindow.Selection.SlideRange | <body bgcolor=#XXXXXX> |
.Background.Fill.ForeColor.RGB | CSS | |
value | background-color: <value> | |
background image | ActiveWindow.Selection.SlideRange | <body |
.Background.Fill.UserPicture | background = “filename.gif”> | |
“C:\pic.bmp” | CSS | |
background-image: <value> | ||
background texture | ActiveWindow.Selection.SlideRange | <body |
.Background.Fill.Patterned value | background=“filename.gif”> | |
ActiveWindow.Selection.SlideRange | CSS | |
.Background.Fill.PresetTextured | background-image: <value> | |
value | ||
font family | ActiveWindow.Selection.TextRange. | <FONT FACE=“facename1, |
Font.Name = “font-name” | facename2 . . . ”> | |
CSS | ||
font-family:<name> | ||
font style | ActiveWindow.Selection.TextRange | <B></B> |
.Font.Bold = msoTrue | <BLINK></BLINK> | |
ActiveWindow.Selection.TextRange | <CITE> </CITE> | |
.Font.Emboss = msoTrue | <CODE> </CODE> | |
ActiveWindow.Selection.TextRange | <EM> </EM> | |
.Font.Italic = msoTrue | <I></I> | |
ActiveWindow.Selection.TextRange | <KBD> </KBD> | |
.Font.Shadow = msoTrue | <LISTING> </LISTING> | |
ActiveWindow.Selection.TextRange | <PLAINTEXT></PLAINTEXT> | |
.Font.Subscript = msoTrue | <SAMP> </SAMP> | |
ActiveWindow.Selection.TextRange | <STRIKE> </STRIKE> | |
.Font.Superscript = msoTrue | <STRONG> </STRONG> | |
ActiveWindow.Selection.TextRange | <SUB> </SUB> | |
.Font.Underline = msoTrue | <SUP> </SUP> | |
<TT> </TT> | ||
<U> </U> | ||
<VAR> </VAR> | ||
CSS | ||
font-style: <value> | ||
font-variant: <value> | ||
font-weight: <value> | ||
text-transform: <value> | ||
text-decoration: <value> | ||
vertical-align: <value> | ||
font size | ActiveWindow.Selection.TextRange. | <BASEFONT SIZE=number> |
Font.Size = value | <BIG> </BIG> | |
<FONT SIZE=+1></FONT> | ||
<SMALL> </SMALL> | ||
CSS | ||
font-size: <value> | ||
font color | ActiveWindow.Selection.TextRange | <FONT |
.Font.Color.RGB = value | COLOR=FFFFFF></FONT> | |
CSS | ||
color: <color> | ||
word spacing | Adjust by font change | CSS |
word-spacing: <value> | ||
letter spacing | Adjust by font change | CSS |
letter-spacing: <value> | ||
line spacing | With ActiveWindow.Selection.TextRange | CSS |
.ParagraphFormat | line-height: 200% | |
.LineRuleWithin = msoTrue | ||
.SpaceWithin = value | ||
End With | ||
text align | ActiveWindow.Selection.TextRange | <CENTER></CENTER> |
.ParagraphFormat.Alignment = value | <P ALIGN=alignment> | |
CSS | ||
text-align: <value> | ||
text indent | ActiveWindow.Selection.ShapeRange | <DFN></DFN> |
.TextFrame.MarginLeft = value | <FRAME | |
MARGINWIDTH=number> | ||
ActiveWindow.Selection.ShapeRange | <FRAME | |
.TextFrame.MarginRight = value | MARGINHEIGHT=number> | |
ActiveWindow.Selection.ShapeRange | CSS | |
.TextFrame.MarginTop = value | text-indent: <value> | |
ActiveWindow.Selection.ShapeRange | border-bottom-width: <value> | |
.TextFrame.MarginBottom = value | border-bottom: <value> | |
border-color: <value> | ||
border-left-width: <value> | ||
border-left: <value> | ||
border-right-width: <value> | ||
border-right: <value> | ||
border-style: <value> | ||
border-top-width: <value> | ||
border-top: <value> | ||
border-width: <value> | ||
border: <value> | ||
clear: <value> | ||
float: <value> | ||
height: <value> | ||
margin-bottom: <value> | ||
margin-left: <value> | ||
margin-right: <value> | ||
margin-top: <value> | ||
margin: <value> | ||
padding-bottom: <value> | ||
padding-left: <value> | ||
padding-right: <value> | ||
padding-top: <value> | ||
padding: <value> | ||
width: <value> | ||
list style | ActivePresentation.ApplyTemplate | CSS |
FileName:=“C:\file.dot” | list-style-type: <value> | |
list-style-position: <value> | ||
bullet for list | ActiveWindow.Selection.TextRange | <LI DINGBAT=“ name”></LI> |
.ParagraphFormat.Bullet.Character = value | <LI TYPE=type> </LI> | |
<LI VALUE=number> </LI> | ||
same for <OL> and <UL> tags | ||
CSS | ||
list-style-image: <value> | ||
visual effect | ActiveWindow.Selection.SlideRange | <MARQUEE behavior=“slide” |
.SlideShowTransition.EntryEffect = value | direction = “left, right, up, or | |
down” LOOP=1> | ||
audio effect | ActiveWindow.Selection.SlideRange | <BGSOUND SRC=“URL”> |
.SlideShowTransition.SoundEffect | ||
.Name= value | ||
design/template | ActivePresentation.ApplyTemplate | Apply a CSS using |
selection | FileName:=“C:\file.dot” | CLASS=‘name’ tag to different |
HTML elements | ||
Or <LINK REL=StyleSheet | ||
HREF=“style.css” | ||
TYPE=“text/css”> | ||
Claims (47)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/444,204 US6580438B1 (en) | 1999-11-22 | 1999-11-22 | Systems and methods for maintaining uniformity in a presentation environment |
JP2000355253A JP4277443B2 (en) | 1999-11-22 | 2000-11-22 | Presentation system and program |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/444,204 US6580438B1 (en) | 1999-11-22 | 1999-11-22 | Systems and methods for maintaining uniformity in a presentation environment |
Publications (1)
Publication Number | Publication Date |
---|---|
US6580438B1 true US6580438B1 (en) | 2003-06-17 |
Family
ID=23763926
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/444,204 Expired - Lifetime US6580438B1 (en) | 1999-11-22 | 1999-11-22 | Systems and methods for maintaining uniformity in a presentation environment |
Country Status (2)
Country | Link |
---|---|
US (1) | US6580438B1 (en) |
JP (1) | JP4277443B2 (en) |
Cited By (43)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20010054176A1 (en) * | 2000-02-11 | 2001-12-20 | Vin Harrick M. | Run-time personalization architecture |
US20030208534A1 (en) * | 2002-05-02 | 2003-11-06 | Dennis Carmichael | Enhanced productivity electronic meeting system |
US20030229851A1 (en) * | 2001-07-05 | 2003-12-11 | Heinz Willumeit | Defining layout files by markup language documents |
US20040056897A1 (en) * | 2002-08-30 | 2004-03-25 | Fujitsu Limited | Mobile terminal |
US6760884B1 (en) * | 1999-08-09 | 2004-07-06 | Internal Research Corporation | Interactive memory archive |
US20040133855A1 (en) * | 2002-09-27 | 2004-07-08 | Blair Robert Bruce | Providing a presentation engine adapted for use by a constrained resource client device |
US20040196293A1 (en) * | 2000-04-06 | 2004-10-07 | Microsoft Corporation | Application programming interface for changing the visual style |
US20050012949A1 (en) * | 2003-05-22 | 2005-01-20 | Katsuhito Kitahara | Barcode printing system, and method and program of setting a virtual barcode font |
US20050154982A1 (en) * | 2004-01-13 | 2005-07-14 | International Business Machines Corporation | Apparatus, system and method of importing cascading style sheets to macromedia flash |
US20050183022A1 (en) * | 2004-02-12 | 2005-08-18 | International Business Machines Corporation | System, method and program product for control of a presentation |
US6938210B1 (en) * | 2000-04-11 | 2005-08-30 | Liztech Co., Ltd. | Computer-Based lecture recording and reproducing method |
US20050223314A1 (en) * | 2004-03-31 | 2005-10-06 | Satyam Computer Services Inc. | System and method for automatic generation of presentations based on agenda |
US20050240407A1 (en) * | 2004-04-22 | 2005-10-27 | Simske Steven J | Method and system for presenting content to an audience |
EP1672518A2 (en) | 2004-12-20 | 2006-06-21 | Microsoft Corporation | Method, system, and computer-readable medium for applying a global formatting scheme to a chart in an electronic document |
US7131065B1 (en) * | 2000-06-30 | 2006-10-31 | International Business Machines Corporation | Dynamically altering web page content based on a page's title properties |
US20060271874A1 (en) * | 2000-04-06 | 2006-11-30 | Microsoft Corporation | Focus state themeing |
US20070277106A1 (en) * | 2006-05-26 | 2007-11-29 | International Business Machines Corporation | Method and structure for managing electronic slides using a slide-reading program |
US20080040340A1 (en) * | 2004-03-31 | 2008-02-14 | Satyam Computer Services Ltd | System and method for automatic generation of presentations based on agenda |
US7370274B1 (en) * | 2003-09-18 | 2008-05-06 | Microsoft Corporation | System and method for formatting objects on a page of an electronic document by reference |
US7376897B1 (en) * | 2000-09-30 | 2008-05-20 | Intel Corporation | Method, apparatus, and system for determining information representations and modalities based on user preferences and resource consumption |
US20080126943A1 (en) * | 1999-06-15 | 2008-05-29 | Microsoft Corporation | System and method for recording a presentation for on-demand viewing over a computer network |
US20080288864A1 (en) * | 2007-05-15 | 2008-11-20 | International Business Machines Corporation | Method and system to enable prioritized presentation content delivery and display |
US20090119578A1 (en) * | 2000-04-06 | 2009-05-07 | Microsoft Corporation | Programming Interface for a Computer Platform |
US20090157609A1 (en) * | 2007-12-12 | 2009-06-18 | Yahoo! Inc. | Analyzing images to derive supplemental web page layout characteristics |
US20090210796A1 (en) * | 2008-02-15 | 2009-08-20 | Bhogal Kulvir S | System and Method for Dynamically Modifying a Sequence of Slides in a Slideshow Set During a Presentation of the Slideshow |
WO2009101252A1 (en) * | 2008-02-12 | 2009-08-20 | Progressiivinen Informatio Ja Kommunikaatio | Data structure and method for its construction |
US20100064222A1 (en) * | 2008-09-08 | 2010-03-11 | Apple Inc. | Object-aware transitions |
US20100064223A1 (en) * | 2008-09-08 | 2010-03-11 | Apple Inc. | Object-aware transitions |
US20100118037A1 (en) * | 2008-09-08 | 2010-05-13 | Apple Inc. | Object-aware transitions |
US20100168881A1 (en) * | 2008-12-30 | 2010-07-01 | Apple Inc. | Multimedia Display Based on Audio and Visual Complexity |
US20100169777A1 (en) * | 2008-12-30 | 2010-07-01 | Apple Inc. | Light Table for Editing Digital Media |
US20100169784A1 (en) * | 2008-12-30 | 2010-07-01 | Apple Inc. | Slide Show Effects Style |
US20100169783A1 (en) * | 2008-12-30 | 2010-07-01 | Apple, Inc. | Framework for Slideshow Object |
US20100318916A1 (en) * | 2009-06-11 | 2010-12-16 | David Wilkins | System and method for generating multimedia presentations |
US20130086468A1 (en) * | 2005-01-09 | 2013-04-04 | Roger Rosner | Efficient creation of documents |
US8676813B2 (en) | 2011-09-06 | 2014-03-18 | Denis J. Alarie | Method and system for selecting a subset of information to communicate to others from a set of information |
US9032300B2 (en) | 2010-08-24 | 2015-05-12 | Apple Inc. | Visual presentation composition |
US9619128B2 (en) | 2013-07-01 | 2017-04-11 | Microsoft Technology Licensing, Llc | Dynamic presentation prototyping and generation |
US10242672B2 (en) | 2016-10-28 | 2019-03-26 | Microsoft Technology Licensing, Llc | Intelligent assistance in presentations |
US10496251B2 (en) | 2015-10-16 | 2019-12-03 | Fanuc Corporation | Screen creation apparatus having application screen creation support function |
US10572128B2 (en) | 2013-09-29 | 2020-02-25 | Microsoft Technology Licensing, Llc | Media presentation effects |
US10846336B2 (en) | 2014-10-30 | 2020-11-24 | Microsoft Technology Licensing, Llc | Authoring tools for synthesizing hybrid slide-canvas presentations |
US11042694B2 (en) * | 2017-09-01 | 2021-06-22 | Adobe Inc. | Document beautification using smart feature suggestions based on textual analysis |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7102643B2 (en) * | 2001-11-09 | 2006-09-05 | Vibe Solutions Group, Inc. | Method and apparatus for controlling the visual presentation of data |
JP2003271892A (en) * | 2002-03-18 | 2003-09-26 | Nippon Telegraph & Telephone East Corp | Automatic scenario production system, automatically scenario producing method, computer program, and recording medium with the program recorded thereon |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4994987A (en) | 1987-11-20 | 1991-02-19 | Minnesota Mining And Manufacturing Company | Image access system providing easier access to images |
US5001696A (en) | 1987-11-20 | 1991-03-19 | Minnesota Mining And Manufacturing Company | Method of preparing an image access card |
US5331547A (en) | 1993-01-29 | 1994-07-19 | Clinical Multiphase Research, Inc. | Process and computer system for control of interface software and data files |
US5473744A (en) * | 1992-09-28 | 1995-12-05 | Optical Magnetic Imaging Corporation | Computer-assisted interactive method and apparatus for making a multi-media presentation |
US5539658A (en) * | 1995-05-16 | 1996-07-23 | Minnesota Mining And Manufacturing Company | Electronic presentation system using portable storage media |
US5859623A (en) * | 1996-05-14 | 1999-01-12 | Proxima Corporation | Intelligent display system presentation projection arrangement and method of using same |
US5874724A (en) | 1997-01-10 | 1999-02-23 | International Business Machines Corporation | Light selectable radio frequency identification tag and method therefor |
US5965860A (en) * | 1996-05-28 | 1999-10-12 | Fujitsu Limited | Management system for using IC card with registered personal information |
US6008807A (en) * | 1997-07-14 | 1999-12-28 | Microsoft Corporation | Method and system for controlling the display of objects in a slide show presentation |
US6191783B1 (en) * | 1996-12-06 | 2001-02-20 | U.S. Philips Corp. | Method and device for configuring a multimedia message for presentation |
US6362838B1 (en) * | 1995-05-16 | 2002-03-26 | Inventions, Inc. | Method for consolidation of multiple data sources |
-
1999
- 1999-11-22 US US09/444,204 patent/US6580438B1/en not_active Expired - Lifetime
-
2000
- 2000-11-22 JP JP2000355253A patent/JP4277443B2/en not_active Expired - Fee Related
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4994987A (en) | 1987-11-20 | 1991-02-19 | Minnesota Mining And Manufacturing Company | Image access system providing easier access to images |
US5001696A (en) | 1987-11-20 | 1991-03-19 | Minnesota Mining And Manufacturing Company | Method of preparing an image access card |
US5473744A (en) * | 1992-09-28 | 1995-12-05 | Optical Magnetic Imaging Corporation | Computer-assisted interactive method and apparatus for making a multi-media presentation |
US5331547A (en) | 1993-01-29 | 1994-07-19 | Clinical Multiphase Research, Inc. | Process and computer system for control of interface software and data files |
US5539658A (en) * | 1995-05-16 | 1996-07-23 | Minnesota Mining And Manufacturing Company | Electronic presentation system using portable storage media |
US6362838B1 (en) * | 1995-05-16 | 2002-03-26 | Inventions, Inc. | Method for consolidation of multiple data sources |
US5859623A (en) * | 1996-05-14 | 1999-01-12 | Proxima Corporation | Intelligent display system presentation projection arrangement and method of using same |
US5965860A (en) * | 1996-05-28 | 1999-10-12 | Fujitsu Limited | Management system for using IC card with registered personal information |
US6191783B1 (en) * | 1996-12-06 | 2001-02-20 | U.S. Philips Corp. | Method and device for configuring a multimedia message for presentation |
US5874724A (en) | 1997-01-10 | 1999-02-23 | International Business Machines Corporation | Light selectable radio frequency identification tag and method therefor |
US6008807A (en) * | 1997-07-14 | 1999-12-28 | Microsoft Corporation | Method and system for controlling the display of objects in a slide show presentation |
Non-Patent Citations (29)
Title |
---|
"Alternative Human-Web Interaction Systems", R. Schlosser et al., http://ils.unc.edu/alternative/alternative.html, Jun. 1999-abstract only. |
"An Overview of the Parc Tab Ubiquitous Computing Experiment", R. Want et al., IEEE Personal Communications, Dec. 1995. |
"BiStatix Whitepaper", BiStatix Technology, http://www.mot.com/LMPS/Indala/bistatix.htm, pp. 1-9, Aug. 31, 1999. |
"Bricks: Laying the Foundations for Graspable User Interfaces", G. Fitzmaurice et al., Proceedings of CHI '95, pp. 422-449, 1995. |
"Bridging Physical and Virtual Worlds with Electronic Tags", Roy Want et al., INTERCHI '99, pp. 370-377, May 15-20, 1999. |
"Bridging the Paper and Electronic Worlds: The Paper User Interface", W. Johnson et al., INTERCHI '93, Apr. 24-29, 1993. |
"Collaboration Using Multiple PDAs Connected to a PC", Brad A. Myers et al., CSCW 98, pp. 285-294, 1998. |
"Computer-Augmented Environments: Back to the Real World", P. Weliner et al., Communications of the ACM, vol. 36, No. 7, Jul. 1993. |
"Design of Spatially Aware Graspable Displays", D. Small et al., Extended Abstracts of CHI '97, Mar. 22-27, 1997. |
"Dual Device User Interface Design: PDAs and Interactive Television", Scott Robertson et al., INTERCHI '96, pp. 79-86, Apr. 13-18, 1996. |
"Illuminating Light: An Optical Design Tool with a Luminous-Tangible Interface", John Underkoffer et al., Proceedings of CHI '98, pp. 1-8, Apr. 18-23, 1998. |
"Insight Lab: An Immersive Team Environment Linking Paper, Displays, and Data", B. Lange et al., CHI'98, Apr. 18-23, 1998. |
"Liveboard: A Large Interactive Display Supporting Group Meetings, Presentations and Remote Collaboration", S. Elrod et al., CHI '92, Human Factors in Computing Systems, ACM Press, pp. 599-607, Monterey. |
"Palette: A Paper Interface for Giving Presentations", Les Nelson et al., Proceeding of the CHI 99 Conference of Human Factors in Computing Systems, pp. 1-8, May 1999. |
"PaperClipX-Summer Intern", Tomas Sokoler, http://intern99-1.pal.xerox.com:80/PaperClipX, pp. 1-2, Jun. 24, 1999. |
"PaperLink: A Technique of Hyperlinking from Real Paper to Electronic Content", Toshifumi Arai et al., INTERCHI '97, pp. 327-344, Mar. 22-27, 1997. |
"Passive Real-World Interface Props for Neurosurgical Visualization", K. Hinckley et al., CHI '94-abstract only. |
"Situated Information Spaces and Spatially Aware Palmtop Computers", G. Fitzmaurice, CACM, 36(7), pp. 38-49, Jul. 1993. |
"Squeeze Me, Hold Me, Tilt Me! An Exploration of Manipulative User Interfaces", B. Harrison et al., CHI '98, Apr. 18-23, 1993. |
"Supporting Dynamic Downloadable Appearances in an Extensible User Interface Toolkit", Scott E. Hudson et al., UIST 97 Banff, pp. 159-168, 1997. |
"Tangible Bits: Towards Seamless Interfaces Between People, Bits, and Atoms", H. Ishii et al., Chi'97, Mar. 22-27, 1997. |
"Telerobotic Control Using Augmented Reality", P. Milgram et al., IEEE International Workshop on Robot and Human Communication, 1995. |
"The Computer for the 21st Century", M. Weiser, Scientific American, Sep. 1991. |
"Tilting Operations for Small Screen Interfaces (Tech Note)", J. Rekimoto, UIST '96. |
"Triangles: Design of a Physical/Digital Construction Kit", M. Gorbet et al., Proceedings of DIS '97, Mar. 1997. |
"Triangles: Tangible Interface for Manipulation and Exploration of Digital Information Topography", M. Gorbet et al., Proceedings of CHI'98, ACM Press, Apr. 1998. |
"Video Mosaic: Laying Out Time in a Physical Space", Wendy E. Mackay et al., Multimedia 94, pp. 165-172, Oct. 1994. |
A Model of the Acquisition of Menu Knowledge by Exploration, A. Howes, CHI '94-abstract only. |
Indala's Proximity Access Field Programming, BiStatix Technology, http://www.mot.com/LMPS/Indala/motrelease. html#bistatix, pp. 1-4, Aug. 31, 1999. |
Cited By (62)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080126943A1 (en) * | 1999-06-15 | 2008-05-29 | Microsoft Corporation | System and method for recording a presentation for on-demand viewing over a computer network |
US7552228B2 (en) * | 1999-06-15 | 2009-06-23 | Microsoft Corporation | System and method for recording a presentation for on-demand viewing over a computer network |
US6760884B1 (en) * | 1999-08-09 | 2004-07-06 | Internal Research Corporation | Interactive memory archive |
US20010054176A1 (en) * | 2000-02-11 | 2001-12-20 | Vin Harrick M. | Run-time personalization architecture |
US20040196293A1 (en) * | 2000-04-06 | 2004-10-07 | Microsoft Corporation | Application programming interface for changing the visual style |
US20090119578A1 (en) * | 2000-04-06 | 2009-05-07 | Microsoft Corporation | Programming Interface for a Computer Platform |
US20060271874A1 (en) * | 2000-04-06 | 2006-11-30 | Microsoft Corporation | Focus state themeing |
US8458608B2 (en) | 2000-04-06 | 2013-06-04 | Microsoft Corporation | Focus state themeing |
US6938210B1 (en) * | 2000-04-11 | 2005-08-30 | Liztech Co., Ltd. | Computer-Based lecture recording and reproducing method |
US7131065B1 (en) * | 2000-06-30 | 2006-10-31 | International Business Machines Corporation | Dynamically altering web page content based on a page's title properties |
US7376897B1 (en) * | 2000-09-30 | 2008-05-20 | Intel Corporation | Method, apparatus, and system for determining information representations and modalities based on user preferences and resource consumption |
US20030229851A1 (en) * | 2001-07-05 | 2003-12-11 | Heinz Willumeit | Defining layout files by markup language documents |
US20030208534A1 (en) * | 2002-05-02 | 2003-11-06 | Dennis Carmichael | Enhanced productivity electronic meeting system |
US7660604B2 (en) | 2002-08-30 | 2010-02-09 | Fujitsu Limited | Mobile terminal |
US20040056897A1 (en) * | 2002-08-30 | 2004-03-25 | Fujitsu Limited | Mobile terminal |
US20040133855A1 (en) * | 2002-09-27 | 2004-07-08 | Blair Robert Bruce | Providing a presentation engine adapted for use by a constrained resource client device |
US7299411B2 (en) * | 2002-09-27 | 2007-11-20 | Liberate Technologies | Providing a presentation engine adapted for use by a constrained resource client device |
US20050012949A1 (en) * | 2003-05-22 | 2005-01-20 | Katsuhito Kitahara | Barcode printing system, and method and program of setting a virtual barcode font |
US7561283B2 (en) * | 2003-05-22 | 2009-07-14 | Seiko Epson Corporation | Barcode printing system, and method and program of setting a virtual barcode font |
US7370274B1 (en) * | 2003-09-18 | 2008-05-06 | Microsoft Corporation | System and method for formatting objects on a page of an electronic document by reference |
US20050154982A1 (en) * | 2004-01-13 | 2005-07-14 | International Business Machines Corporation | Apparatus, system and method of importing cascading style sheets to macromedia flash |
US20050183022A1 (en) * | 2004-02-12 | 2005-08-18 | International Business Machines Corporation | System, method and program product for control of a presentation |
US8689097B2 (en) | 2004-03-31 | 2014-04-01 | Satyam Computer Services Ltd. | System and method for automatic generation of presentations based on agenda |
US20050223314A1 (en) * | 2004-03-31 | 2005-10-06 | Satyam Computer Services Inc. | System and method for automatic generation of presentations based on agenda |
US20080040340A1 (en) * | 2004-03-31 | 2008-02-14 | Satyam Computer Services Ltd | System and method for automatic generation of presentations based on agenda |
US20050240407A1 (en) * | 2004-04-22 | 2005-10-27 | Simske Steven J | Method and system for presenting content to an audience |
KR101169087B1 (en) | 2004-12-20 | 2012-07-26 | 마이크로소프트 코포레이션 | Method, system, and computer-readable medium for applying a global formatting scheme to a chart in an electronic document |
US7546522B2 (en) | 2004-12-20 | 2009-06-09 | Microsoft Corporation | Method, system, and computer-readable medium for applying a global formatting scheme to a chart in an electronic document |
EP1672518A3 (en) * | 2004-12-20 | 2009-01-21 | Microsoft Corporation | Method, system, and computer-readable medium for applying a global formatting scheme to a chart in an electronic document |
EP1672518A2 (en) | 2004-12-20 | 2006-06-21 | Microsoft Corporation | Method, system, and computer-readable medium for applying a global formatting scheme to a chart in an electronic document |
US20060136819A1 (en) * | 2004-12-20 | 2006-06-22 | Microsoft Corporation | Method, system, ane computer-readable medium for applying a global formatting scheme to a chart in an electronic document |
US9639504B2 (en) * | 2005-01-09 | 2017-05-02 | Apple Inc. | Efficient creation of documents |
US20130086468A1 (en) * | 2005-01-09 | 2013-04-04 | Roger Rosner | Efficient creation of documents |
US20070277106A1 (en) * | 2006-05-26 | 2007-11-29 | International Business Machines Corporation | Method and structure for managing electronic slides using a slide-reading program |
US20080288864A1 (en) * | 2007-05-15 | 2008-11-20 | International Business Machines Corporation | Method and system to enable prioritized presentation content delivery and display |
US20090157609A1 (en) * | 2007-12-12 | 2009-06-18 | Yahoo! Inc. | Analyzing images to derive supplemental web page layout characteristics |
WO2009101252A1 (en) * | 2008-02-12 | 2009-08-20 | Progressiivinen Informatio Ja Kommunikaatio | Data structure and method for its construction |
US8041724B2 (en) | 2008-02-15 | 2011-10-18 | International Business Machines Corporation | Dynamically modifying a sequence of slides in a slideshow set during a presentation of the slideshow |
US20090210796A1 (en) * | 2008-02-15 | 2009-08-20 | Bhogal Kulvir S | System and Method for Dynamically Modifying a Sequence of Slides in a Slideshow Set During a Presentation of the Slideshow |
US20100118037A1 (en) * | 2008-09-08 | 2010-05-13 | Apple Inc. | Object-aware transitions |
US20100064222A1 (en) * | 2008-09-08 | 2010-03-11 | Apple Inc. | Object-aware transitions |
US20100223554A1 (en) * | 2008-09-08 | 2010-09-02 | Apple Inc. | Object-aware transitions |
US10984577B2 (en) | 2008-09-08 | 2021-04-20 | Apple Inc. | Object-aware transitions |
US8694889B2 (en) | 2008-09-08 | 2014-04-08 | Appl Inc. | Object-aware transitions |
US20100064223A1 (en) * | 2008-09-08 | 2010-03-11 | Apple Inc. | Object-aware transitions |
US7721209B2 (en) | 2008-09-08 | 2010-05-18 | Apple Inc. | Object-aware transitions |
US8621357B2 (en) | 2008-12-30 | 2013-12-31 | Apple Inc. | Light table for editing digital media |
US8626322B2 (en) | 2008-12-30 | 2014-01-07 | Apple Inc. | Multimedia display based on audio and visual complexity |
US20100169777A1 (en) * | 2008-12-30 | 2010-07-01 | Apple Inc. | Light Table for Editing Digital Media |
US20100169784A1 (en) * | 2008-12-30 | 2010-07-01 | Apple Inc. | Slide Show Effects Style |
US8832555B2 (en) | 2008-12-30 | 2014-09-09 | Apple Inc. | Framework for slideshow object |
US20100168881A1 (en) * | 2008-12-30 | 2010-07-01 | Apple Inc. | Multimedia Display Based on Audio and Visual Complexity |
US20100169783A1 (en) * | 2008-12-30 | 2010-07-01 | Apple, Inc. | Framework for Slideshow Object |
US20100318916A1 (en) * | 2009-06-11 | 2010-12-16 | David Wilkins | System and method for generating multimedia presentations |
US9032300B2 (en) | 2010-08-24 | 2015-05-12 | Apple Inc. | Visual presentation composition |
US8676813B2 (en) | 2011-09-06 | 2014-03-18 | Denis J. Alarie | Method and system for selecting a subset of information to communicate to others from a set of information |
US9619128B2 (en) | 2013-07-01 | 2017-04-11 | Microsoft Technology Licensing, Llc | Dynamic presentation prototyping and generation |
US10572128B2 (en) | 2013-09-29 | 2020-02-25 | Microsoft Technology Licensing, Llc | Media presentation effects |
US10846336B2 (en) | 2014-10-30 | 2020-11-24 | Microsoft Technology Licensing, Llc | Authoring tools for synthesizing hybrid slide-canvas presentations |
US10496251B2 (en) | 2015-10-16 | 2019-12-03 | Fanuc Corporation | Screen creation apparatus having application screen creation support function |
US10242672B2 (en) | 2016-10-28 | 2019-03-26 | Microsoft Technology Licensing, Llc | Intelligent assistance in presentations |
US11042694B2 (en) * | 2017-09-01 | 2021-06-22 | Adobe Inc. | Document beautification using smart feature suggestions based on textual analysis |
Also Published As
Publication number | Publication date |
---|---|
JP2001228840A (en) | 2001-08-24 |
JP4277443B2 (en) | 2009-06-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6580438B1 (en) | Systems and methods for maintaining uniformity in a presentation environment | |
US8671351B2 (en) | Application modification based on feed content | |
US6732915B1 (en) | Systems and methods for controlling a presentation using physical objects | |
US7418656B1 (en) | Dynamic annotations for electronics documents | |
US7287018B2 (en) | Browsing electronically-accessible resources | |
US10417316B2 (en) | Emphasizing a portion of the visible content elements of a markup language document | |
JP4344693B2 (en) | System and method for browser document editing | |
US20050060046A1 (en) | Information processing apparatus, its control method, and program | |
US7240281B2 (en) | System, method and program for printing an electronic document | |
CN102414682A (en) | Method and apparatus for providing scraped web pages in a graphical user interface | |
US20110072390A1 (en) | System and method for selective control of zoom adjustments in an electronic publication | |
US20110082749A1 (en) | System And Method For Template-Based Assembly Of Publications | |
US20110154186A1 (en) | Method, program and terminal device for rendering web page | |
KR102350540B1 (en) | digital component background rendering | |
JP2000348040A (en) | Information processor and information processing system | |
US20040205643A1 (en) | Reproduction of documents using intent information | |
US20070136348A1 (en) | Screen-wise presentation of search results | |
US20080155389A1 (en) | Mobile terminal and method for providing hyperlink information | |
US6964017B1 (en) | System and method of creating interactive visual content | |
KR101160973B1 (en) | Effective Graphic Format image file forming method and device therefor | |
US20070206022A1 (en) | Method and apparatus for associating text with animated graphics | |
US20040237026A1 (en) | System and method for creating reminders in electronic documents | |
US8276089B1 (en) | System and method for producing, displaying and printing documents | |
JP2002251338A (en) | Document display device having bookmark presenting function | |
KR101645487B1 (en) | Apparatus and method of offering electronic manual having user editing function |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: FUJI XEROX CO., LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ICHIMURA, SATOSHI;NELSON, LESTER D.;PEDERSEN, ELIN R.;REEL/FRAME:010433/0779 Effective date: 19991119 |
|
AS | Assignment |
Owner name: BANK ONE, NA, AS ADMINISTRATIVE AGENT, ILLINOIS Free format text: SECURITY AGREEMENT;ASSIGNOR:XEROX CORPORATION;REEL/FRAME:013111/0001 Effective date: 20020621 Owner name: BANK ONE, NA, AS ADMINISTRATIVE AGENT,ILLINOIS Free format text: SECURITY AGREEMENT;ASSIGNOR:XEROX CORPORATION;REEL/FRAME:013111/0001 Effective date: 20020621 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: XEROX CORPORATION, CONNECTICUT Free format text: RELEASE OF LIEN IN PATENTS;ASSIGNOR:BANK ONE, NA;REEL/FRAME:013707/0924 Effective date: 20030530 |
|
FEPP | Fee payment procedure |
Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
FPAY | Fee payment |
Year of fee payment: 8 |
|
FPAY | Fee payment |
Year of fee payment: 12 |
|
AS | Assignment |
Owner name: XEROX CORPORATION, CONNECTICUT Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:JPMORGAN CHASE BANK, N.A. AS SUCCESSOR-IN-INTEREST ADMINISTRATIVE AGENT AND COLLATERAL AGENT TO BANK ONE, N.A.;REEL/FRAME:061388/0388 Effective date: 20220822 |