US20020138264A1 - Apparatus to convey depth information in graphical images and method therefor - Google Patents
Apparatus to convey depth information in graphical images and method therefor Download PDFInfo
- Publication number
- US20020138264A1 US20020138264A1 US09/814,397 US81439701A US2002138264A1 US 20020138264 A1 US20020138264 A1 US 20020138264A1 US 81439701 A US81439701 A US 81439701A US 2002138264 A1 US2002138264 A1 US 2002138264A1
- Authority
- US
- United States
- Prior art keywords
- image
- depth
- depth map
- response
- cue
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims abstract description 31
- 230000004044 response Effects 0.000 claims abstract description 18
- 238000012545 processing Methods 0.000 claims abstract description 10
- 238000004590 computer program Methods 0.000 claims abstract description 6
- 238000005516 engineering process Methods 0.000 description 5
- 230000001771 impaired effect Effects 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 3
- 238000004891 communication Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 230000008859 change Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000003909 pattern recognition Methods 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 229920006395 saturated elastomer Polymers 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 208000029257 vision disease Diseases 0.000 description 1
- 230000004393 visual impairment Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/016—Input arrangements with force or tactile feedback as computer generated output to the user
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B21/00—Teaching, or communicating with, the blind, deaf or mute
- G09B21/001—Teaching or communicating with blind persons
- G09B21/003—Teaching or communicating with blind persons using tactile presentation of the information, e.g. Braille displays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
- H04N21/440236—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by media transcoding, e.g. video is transformed into a slideshow of still pictures, audio is converted into text
Definitions
- the present invention is related to the following U.S. Patent Applications which are hereby incorporated herein by reference: Ser. No. 09/______, “Apparatus for Outputting Textual Renditions of Graphical Data and Method Therefor” (Attorney Docket No. AUS9-2001-0095US1); Ser. No. 09/______, “Scanning and Outputting Textual Information in Web Page Images” (Attorney Docket No. AUS9-2001-0096US1); and Ser. No. 09/______, “Extracting Textual Equivalents of Multimedia Content Stored in Multimedia Files” (Attorney Docket No. AUS9-2001-0097US1).
- the present invention relates to the field of assisting individuals with disabilities through technology, and more particularly to providing depth cues in graphical information contained in web pages to promote accessibility to individuals with disabilities.
- Congress passed the “Assistive Technology Act of 1998” to promote the assistance of individuals with disabilities through technology such as encouraging the promotion of technology that will allow individuals with disabilities to partake in the information technology, e.g., Internet.
- Web server 102 may store web pages for transmission to a web client 104 , via Internet 106 .
- a computer user may “browse”, i.e., navigate around, the WWW by utilizing a suitable web browser, e.g., Netscape NavigatorTM, Internet ExploderTM, or a talking browser such as, Home Page ReaderTM (HPR) available from International Business Machines Corporation, and a network gateway, e.g., Internet Service Provider (ISP).
- a web browser allows the user to specify or search for a web page on the WWW and subsequently retrieve and display web pages on the user's computer screen.
- Such web browsers are typically installed on personal computers or workstations to provide web client services, such as web client 104 , but increasingly may be found on wireless devices such as cell phones.
- the Internet is based upon a suite of communication protocols known as Transmission Control Protocol/Internet Protocol (TCP/IP) which sends packets of data between a host machine, e.g., server computer on the Internet commonly referred to as a web server, and a client machine, e.g., a user's computer connected to the Internet.
- TCP/IP Transmission Control Protocol/Internet Protocol
- the WWW is a network of computers that use an Internet interface protocol which is supported by the same TCP/IP transmission protocol.
- a web page may typically include content in a multiplicity of media. In addition to text, these may include images, audio and video. Examples of images may include charts and graphs. Images audio and video may be specified in a HyperText Markup Language (HTML) file that is sent from the web server, such as web server 102 , to the client machine, such as web client 104 . HTML files may be exchanged on the Internet in accordance with the HyperText Transfer Protocol (HTTP).
- HTML source code images, video and audio may be specified in various files of different formats. For example, an image may be represented in a Graphics Interchange Format (GIF), Joint Photographic Experts Group (JPEG) and Portable Network Graphics (PNG) file format.
- Video may be represented in a Moving Pictures Expert Group (MPEG) file format.
- Audio maybe represented in a MPEG Audio Layer 3 (MP3) file format.
- the HTML file may then be parsed by the web browser in order to display the images and graphics on the client machine.
- a depth cue method includes scanning a depth map corresponding to an image, in response to user input. A nonvisual cue corresponding to a depth value in the depth map is output, for each pixel scanned.
- the program product includes a program of instructions for performing a scan of a depth map corresponding to an image, in which the scan is performed in response to user input.
- a nonvisual cue is output corresponding to a depth value in the depth map, for each pixel scanned.
- a data processing system includes circuitry operable for scanning a depth map.
- the depth map is scanned in response to user input.
- circuitry operable for outputting a nonvisual cue corresponding to a depth value in the depth map.
- the novisual cue is output for each pixel scanned.
- FIG. 1 illustrates a network system which may be used with the present invention
- FIG. 2 illustrates, in block diagram form, a data processing system implemented in accordance with the present invention
- FIG. 3 illustrates, in flow chart form, an image depth representation methodology in accordance with an embodiment of the present invention
- FIGS. 3. 1 - 3 . 3 schematically illustrate an image and corresponding pixel intensity and depth maps in conjunction with the embodiment of the present invention in FIG. 3;
- FIG. 4 illustrates, in flow chart form a depth map generation methodology which may be used with the embodiment in FIG. 2.
- the present invention provides a system and method for providing depth cues drawn from images appearing in a web page.
- the depth cues may be output in a format accessible to those users with visual impairments.
- the depth cues may be output in an audio form.
- a tactile format may be used.
- the depth information may be incorporated in the web page itself, via, for example, an “ALT” tag. Additionally, the depth information may be generated from the images themselves.
- FIG. 2 an example is shown of a data processing system 200 which may be used for the invention.
- the system has a central processing unit (CPU) 210 , which is coupled to various other components by system bus 212 .
- Read only memory (“ROM”) 216 is coupled to the system bus 212 and includes a basic input/output system (“BIOS”) that controls certain basic functions of the data processing system 200 .
- RAM random access memory
- I/O adapter 218 may be a small computer system interface (“SCSI”) adapter that communicates with a disk storage device 220 .
- SCSI small computer system interface
- Communications adapter 234 interconnects bus 212 with an outside network enabling the data processing system to communicate with other such systems.
- Input/Output devices are also connected to system bus 212 via user interface adapter 222 and display adapter 236 .
- Keyboard 224 , track ball 232 , mouse 226 , speaker 228 , microphone 250 and tactile display 242 are all interconnected to bus 212 via user interface adapter 222 .
- Display monitor 238 is connected to system bus 212 by display adapter 236 . In this manner, a user is capable of inputting to the system throughout the keyboard 224 , trackball 232 or mouse 226 and receiving output from the system via speaker 228 , display 238 and tactile display 242 .
- Preferred implementations of the invention include implementations as a computer system programmed to execute the method or methods described herein, and as a computer program product.
- sets of instructions for executing the method or methods are resident in the random access memory 214 of one or more computer systems configured generally as described above.
- the set of instructions may be stored as a computer program product in another computer memory, for example, in disk drive 220 (which may include a removable memory such as an optical disk or floppy disk for eventual use in the disk drive 220 ).
- the computer program product can also be stored at another computer and transmitted when desired to the user's work station by a network or by an external network such as the Internet.
- step 302 a web page is received.
- step 304 images incorporated in the web page are extracted. That is, in step 304 the image information is identified, and the associated image files are retrieved for further processing in accordance with the principles discussed hereinbelow.
- an image file may be represented in a multiplicity of formats, for example, in a GIF, JPEG, or PNG file format, for example.
- a sequence of images operable for displaying motion such as images forming a “motion picture,” which may be represented in an MPEG file format.
- step 306 it is determined if a depth map is associated with an image in the web page.
- a depth map may be associated with an image in the HTML of the web page by the following exemplary code snippet:
- a depth map is a data structure, for example, a two-dimensional array, in which each member thereof corresponds to a pixel in the image associated with the depth map.
- FIGS. 3. 1 - 3 . 3 schematically depict an image and associated intensity and depth maps, for illustrative purposes.
- FIG. 3. 1 illustrates an elevation view image of a “white” cylindrical object 350 against a “black” background 352 .
- FIG. 3. 2 illustrates an intensity pixel map 354 , corresponding to the image of FIG. 3. 1 .
- the value “255” represents saturated “white” pixels of cylindrical object 350 and the value “0” represents pixels of the “black” background 352 .
- intensity values are represented by an eight-bit gray scale, however, it would be recognized by artisans of ordinary skill that this is illustrative, and other numbers of bits may be used to represent intensity values.
- FIG. 3. 3 illustrates a depth map 356 , corresponding to the image of FIG. 3. 1 .
- the value “30” represents pixels of background 352 , the portion of the image “furthest” from a viewer, and the value “6” represents pixels of the object 350 that are “nearest” the viewer.
- Intermediate values, “8” and “10”, represent pixels of object 350 corresponding to portions of the curved cylindrical surface that are receding from the viewer toward background 352 .
- depth values are represented by an five-bit value, however, it would be recognized by artisans of ordinary skill that this is illustrative, and other numbers of bits may be used to represent depth values. It would be further understood that the one-hundred entries in depth map 356 is also illustrative, and an image may be represented by other numbers of pixels. It would be appreciated that depth maps having depth values represented by other numbers of bits and containing other numbers of entries would fall within the spirit and scope of the present invention.
- step 205 the depth map is fetched from the web server, such as web server 102 , FIG. 1.
- Methodology 300 proceeds to step 307 , and while receiving user input scans the image depth map, step 308 and outputs a representation thereof, step 310 .
- User input maybe in the form, for example, of keystrokes on the keyboard, such as keyboard 124 , FIG. 1, in which keyboard arrows are used to scan through the depth map as methodology 300 loops over steps 307 , 308 and 310 .
- step 308 in response to the user input, a scan through the image depth map is performed.
- a representation of the depth value associated with the pixel is output.
- the output may be in an audio format, wherein a pitch or tone of the audio signal represents the depth value.
- a pitch or tone of the audio signal represents the depth value.
- a “low” pitch may represent a foreground, or “near” element of the image corresponding to the pixel, and, conversely, a “high” pitch may represent a “distant” element.
- Gradations in tone between a predetermined lowest pitch (corresponding, for example, to the smallest depth value in the predetermined range) and a predetermined highest pitch (corresponding to the largest depth value) may, thus, represent to the visually impaired user a range of depths from the “foreground” to the “background” of the image.
- amplitude, rather than pitch may similarly be used to provide depth cues to the visually impaired user.
- a tactile representation may be used via a tactile display, such as tactile display 142 , FIG. 1.
- an array of mechanical elements for example “pins” or similar elastic members (for example, springs) may be excited with an amplitude corresponding to the depth value as the image depth map is scanned.
- an elastic member is capable of returning to an undeformed state on removal of a deforming stress, and is not meant to be limited to members in which the stress-strain relationship is linear.
- methodology 300 breaks out of the loop via the “False” path in step 307 , and process 300 terminates, step 318 .
- step 306 if a depth map has not been associated with an image in the web page, methodology 300 determines if image information is available from which a depth map may be generated.
- step 314 it is determined if either a stereographic image pair has been provided in the web page, or a motion picture image file is included in the page. If so, in step 316 , discussed further hereinbelow in conjunction with FIG. 4, a depth map is generated and process 300 proceeds to step 307 to perform the image depth map scan as previously described. If, however, in step 314 , image information from which a depth map may be generated has not been included in the web page, process 300 terminates, step 318 .
- step 416 of FIG. 3 for generating a depth map is described in additional detail.
- an image set is input. This may, for example, be a sequence of images constituting a portion of a motion picture file. Additionally, the image set maybe a pair of stereographic images.
- the image depth is analyzed, and a depth value assigned to each pixel of the image. Techniques for analyzing depths in an image from stereographic views and motion are described in R. Dutta and C. C.
- step 406 the depth map is filled by setting the data values in a data structure, such as a two-dimensional array, and the depth map containing the depth values generated in analysis step 404 , is output. This depth map may then be scanned in accordance with the principles of methodology 300 , FIG. 3, as previously described.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Processing Or Creating Images (AREA)
Abstract
The aforementioned needs are addressed by the present invention. Accordingly, there is provided, in a first form, a depth cue method. The method includes scanning a depth map corresponding to an image, in response to user input. A nonvisual cue corresponding to a depth value in the depth map is output, for each pixel scanned.
There is also provided, in a second form, a computer program product. The program product includes a program of instructions for performing a scan of a depth map corresponding to an image, in which the scan is performed in response to user input. In response, a nonvisual cue is output corresponding to a depth value in the depth map, for each pixel scanned.
Additionally, there is provided, in a third form, a data processing system. The system includes circuitry operable for scanning a depth map. The depth map is scanned in response to user input. Also included is circuitry operable for outputting a nonvisual cue corresponding to a depth value in the depth map. The novisual cue is output for each pixel scanned.
Description
- The present invention is related to the following U.S. Patent Applications which are hereby incorporated herein by reference: Ser. No. 09/______, “Apparatus for Outputting Textual Renditions of Graphical Data and Method Therefor” (Attorney Docket No. AUS9-2001-0095US1); Ser. No. 09/______, “Scanning and Outputting Textual Information in Web Page Images” (Attorney Docket No. AUS9-2001-0096US1); and Ser. No. 09/______, “Extracting Textual Equivalents of Multimedia Content Stored in Multimedia Files” (Attorney Docket No. AUS9-2001-0097US1).
- The present invention relates to the field of assisting individuals with disabilities through technology, and more particularly to providing depth cues in graphical information contained in web pages to promote accessibility to individuals with disabilities.
- Congress passed the “Assistive Technology Act of 1998” to promote the assistance of individuals with disabilities through technology such as encouraging the promotion of technology that will allow individuals with disabilities to partake in the information technology, e.g., Internet.
- The development of computerized distribution information systems, such as the Internet, allows users to link with servers and networks, and thus retrieve vast amounts of electronic information that was previously unavailable using conventional electronic mediums. Such electronic information increasingly is replacing the more conventional means of information distribution such as newspapers, magazines and television.
- Users may be linked to the Internet through a hypertext system of servers commonly referred to as the World Wide Web (WWW). With the World Wide Web, an entity having a domain name may create a “web page” or “page” that can provide information and, to some degree, some interactivity. Referring to FIG. 1, systematically illustrating
network system 100.Web server 102 may store web pages for transmission to aweb client 104, via Internet 106. - A computer user may “browse”, i.e., navigate around, the WWW by utilizing a suitable web browser, e.g., Netscape Navigator™, Internet Exploder™, or a talking browser such as, Home Page Reader™ (HPR) available from International Business Machines Corporation, and a network gateway, e.g., Internet Service Provider (ISP). A web browser allows the user to specify or search for a web page on the WWW and subsequently retrieve and display web pages on the user's computer screen. Such web browsers are typically installed on personal computers or workstations to provide web client services, such as
web client 104, but increasingly may be found on wireless devices such as cell phones. - The Internet is based upon a suite of communication protocols known as Transmission Control Protocol/Internet Protocol (TCP/IP) which sends packets of data between a host machine, e.g., server computer on the Internet commonly referred to as a web server, and a client machine, e.g., a user's computer connected to the Internet. The WWW is a network of computers that use an Internet interface protocol which is supported by the same TCP/IP transmission protocol.
- A web page may typically include content in a multiplicity of media. In addition to text, these may include images, audio and video. Examples of images may include charts and graphs. Images audio and video may be specified in a HyperText Markup Language (HTML) file that is sent from the web server, such as
web server 102, to the client machine, such asweb client 104. HTML files may be exchanged on the Internet in accordance with the HyperText Transfer Protocol (HTTP). In the HTML source code, images, video and audio may be specified in various files of different formats. For example, an image may be represented in a Graphics Interchange Format (GIF), Joint Photographic Experts Group (JPEG) and Portable Network Graphics (PNG) file format. Video may be represented in a Moving Pictures Expert Group (MPEG) file format. Audio maybe represented in a MPEG Audio Layer 3 (MP3) file format. The HTML file may then be parsed by the web browser in order to display the images and graphics on the client machine. - Images in a web page are inaccessible to the visually impaired user. Consequently, there is a need in the art, generally, to improve the accessibility of this information to such users. In particular, there is a need in the art to convey to the visually impaired user depth cues contained in the images in a web page.
- The aforementioned needs are addressed by the present invention. Accordingly, there is provided, in a first form, a depth cue method. The method includes scanning a depth map corresponding to an image, in response to user input. A nonvisual cue corresponding to a depth value in the depth map is output, for each pixel scanned.
- There is also provided, in a second form, a computer program product. The program product includes a program of instructions for performing a scan of a depth map corresponding to an image, in which the scan is performed in response to user input. In response, a nonvisual cue is output corresponding to a depth value in the depth map, for each pixel scanned.
- Additionally, there is provided, in a third form, a data processing system. The system includes circuitry operable for scanning a depth map. The depth map is scanned in response to user input. Also included is circuitry operable for outputting a nonvisual cue corresponding to a depth value in the depth map. The novisual cue is output for each pixel scanned.
- The foregoing has outlined rather broadly the features and technical advantages of the present invention in order that the detailed description of the invention that follows may be better understood. Additional features and advantages of the invention will be described hereinafter which form the subject of the claims of the invention.
- For a more complete understanding of the present invention, and the advantages thereof, reference is now made to the following descriptions taken in conjunction with the accompanying drawings, in which:
- FIG. 1 illustrates a network system which may be used with the present invention;
- FIG. 2 illustrates, in block diagram form, a data processing system implemented in accordance with the present invention;
- FIG. 3 illustrates, in flow chart form, an image depth representation methodology in accordance with an embodiment of the present invention;
- FIGS. 3.1-3.3 schematically illustrate an image and corresponding pixel intensity and depth maps in conjunction with the embodiment of the present invention in FIG. 3; and
- FIG. 4 illustrates, in flow chart form a depth map generation methodology which may be used with the embodiment in FIG. 2.
- The present invention provides a system and method for providing depth cues drawn from images appearing in a web page. The depth cues may be output in a format accessible to those users with visual impairments. For example, the depth cues may be output in an audio form. Alternatively, a tactile format may be used. The depth information may be incorporated in the web page itself, via, for example, an “ALT” tag. Additionally, the depth information may be generated from the images themselves.
- In the following description, numerous specific details are set forth to provide a thorough understanding of the present invention. However, it will be obvious to those skilled in the art that the present invention may be practiced without such specific details. In other instances, well-known circuits have been shown in block diagram form in order not to obscure the present invention in unnecessary detail. For the most part, details concerning timing considerations and the like have been omitted in as much as such details are not necessary to obtain a complete understanding of the present invention and are within the skills of persons of ordinary skill in the relevant art.
- Refer now to the drawings wherein depicted elements are not necessarily shown to scale and wherein like or similar elements are designated by the same reference numeral through the several views.
- Referring first to FIG. 2, an example is shown of a
data processing system 200 which may be used for the invention. The system has a central processing unit (CPU) 210, which is coupled to various other components bysystem bus 212. Read only memory (“ROM”) 216 is coupled to thesystem bus 212 and includes a basic input/output system (“BIOS”) that controls certain basic functions of thedata processing system 200. Random access memory (“RAM”) 214, I/O adapter 218, andcommunications adapter 234 are also coupled to thesystem bus 212. I/O adapter 218 may be a small computer system interface (“SCSI”) adapter that communicates with adisk storage device 220.Communications adapter 234interconnects bus 212 with an outside network enabling the data processing system to communicate with other such systems. Input/Output devices are also connected tosystem bus 212 viauser interface adapter 222 anddisplay adapter 236.Keyboard 224,track ball 232,mouse 226, speaker 228, microphone 250 andtactile display 242 are all interconnected tobus 212 viauser interface adapter 222.Display monitor 238 is connected tosystem bus 212 bydisplay adapter 236. In this manner, a user is capable of inputting to the system throughout thekeyboard 224,trackball 232 ormouse 226 and receiving output from the system via speaker 228,display 238 andtactile display 242. - Preferred implementations of the invention include implementations as a computer system programmed to execute the method or methods described herein, and as a computer program product. According to the computer system implementation, sets of instructions for executing the method or methods are resident in the random access memory214 of one or more computer systems configured generally as described above. Until required by the computer system, the set of instructions may be stored as a computer program product in another computer memory, for example, in disk drive 220 (which may include a removable memory such as an optical disk or floppy disk for eventual use in the disk drive 220). Further, the computer program product can also be stored at another computer and transmitted when desired to the user's work station by a network or by an external network such as the Internet. One skilled in the art would appreciate that the physical storage of the sets of instructions physically changes the medium upon which it is stored so that the medium carries computer readable information. The change maybe electrical, magnetic, chemical, biological, or some other physical change. While it is convenient to describe the invention in terms of instructions, symbols, characters, or the like, the reader should remember that all of these and similar terms should be associated with the appropriate physical elements.
- Refer now to FIG. 3 illustrating, in flow chart form, image
depth representation methodology 300 in accordance with the principles of the present invention. In step 302 a web page is received. Instep 304 images incorporated in the web page are extracted. That is, instep 304 the image information is identified, and the associated image files are retrieved for further processing in accordance with the principles discussed hereinbelow. As would be recognized by persons of ordinary skill in the data processing art, an image file may be represented in a multiplicity of formats, for example, in a GIF, JPEG, or PNG file format, for example. Additionally, a sequence of images operable for displaying motion, such as images forming a “motion picture,” which may be represented in an MPEG file format. - In step306, it is determined if a depth map is associated with an image in the web page.
- (A depth map may be associated with an image in the HTML of the web page by the following exemplary code snippet:
- <HTML>
- <IMG SRC=“cyl_img.gif” LONGDESC=cyl_ingdepth.txt>
- </HTML>
- where, the image file, in GIF format, is called cyl_img.gif and the associated depth map is called cyl_imgdepthmap.txt. An artisan of ordinary skill would appreciate that the file names are illustrative, and that the code snippet is exemplary, and other coding may be used to associate a depth image with an image in a web page, and would fall within the spirit and scope of the present invention. A depth map is a data structure, for example, a two-dimensional array, in which each member thereof corresponds to a pixel in the image associated with the depth map. (Note that a depth map may be associated with the image by use of an HTML “ALT” tag.) Each element of the data structure, such as a two-dimensional array, has a value in a predetermined range, in which the value represents a depth of the image element represented by the corresponding pixel. (R. Dutta and C. C. Weems,Parallel Dense Depth from Motion on the Image Understanding Architecture; 1993 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION 154 (1993), which is hereby incorporated herein by reference.) FIGS. 3.1-3.3 schematically depict an image and associated intensity and depth maps, for illustrative purposes. FIG. 3.1 illustrates an elevation view image of a “white”
cylindrical object 350 against a “black”background 352. (Note that, for ease of illustration, “black”background 352 is rendered as a mottled pattern.) FIG. 3.2 illustrates anintensity pixel map 354, corresponding to the image of FIG. 3.1. In FIG. 3.2 the value “255” represents saturated “white” pixels ofcylindrical object 350 and the value “0” represents pixels of the “black”background 352. For the purposes of FIG. 3.2, it is assumed that intensity values are represented by an eight-bit gray scale, however, it would be recognized by artisans of ordinary skill that this is illustrative, and other numbers of bits may be used to represent intensity values. It would be further understood that the one-hundred entries inintensity map 354 is also illustrative, and an image may be represented by other numbers of pixels. FIG. 3.3 illustrates adepth map 356, corresponding to the image of FIG. 3.1. In FIG. 3.3 the value “30” represents pixels ofbackground 352, the portion of the image “furthest” from a viewer, and the value “6” represents pixels of theobject 350 that are “nearest” the viewer. Intermediate values, “8” and “10”, represent pixels ofobject 350 corresponding to portions of the curved cylindrical surface that are receding from the viewer towardbackground 352. For the purposes of FIG. 3.3, it is assumed that depth values are represented by an five-bit value, however, it would be recognized by artisans of ordinary skill that this is illustrative, and other numbers of bits may be used to represent depth values. It would be further understood that the one-hundred entries indepth map 356 is also illustrative, and an image may be represented by other numbers of pixels. It would be appreciated that depth maps having depth values represented by other numbers of bits and containing other numbers of entries would fall within the spirit and scope of the present invention. - If a depth map corresponding to an image in the web page is associated therewith, in step205, the depth map is fetched from the web server, such as
web server 102, FIG. 1.Methodology 300 proceeds to step 307, and while receiving user input scans the image depth map,step 308 and outputs a representation thereof,step 310. User input maybe in the form, for example, of keystrokes on the keyboard, such as keyboard 124, FIG. 1, in which keyboard arrows are used to scan through the depth map asmethodology 300 loops oversteps step 308, in response to the user input, a scan through the image depth map is performed. At each pixel, instep 310, a representation of the depth value associated with the pixel is output. The output may be in an audio format, wherein a pitch or tone of the audio signal represents the depth value. For example, a “low” pitch may represent a foreground, or “near” element of the image corresponding to the pixel, and, conversely, a “high” pitch may represent a “distant” element. Gradations in tone between a predetermined lowest pitch (corresponding, for example, to the smallest depth value in the predetermined range) and a predetermined highest pitch (corresponding to the largest depth value) may, thus, represent to the visually impaired user a range of depths from the “foreground” to the “background” of the image. Alternatively, amplitude, rather than pitch may similarly be used to provide depth cues to the visually impaired user. In yet another embodiment, a tactile representation may be used via a tactile display, such as tactile display 142, FIG. 1. In such a display, an array of mechanical elements, for example “pins” or similar elastic members (for example, springs) may be excited with an amplitude corresponding to the depth value as the image depth map is scanned. As used herein, an elastic member is capable of returning to an undeformed state on removal of a deforming stress, and is not meant to be limited to members in which the stress-strain relationship is linear.) - After the user input terminates,
methodology 300 breaks out of the loop via the “False” path instep 307, andprocess 300 terminates,step 318. - Returning to step306, if a depth map has not been associated with an image in the web page,
methodology 300 determines if image information is available from which a depth map may be generated. Instep 314, it is determined if either a stereographic image pair has been provided in the web page, or a motion picture image file is included in the page. If so, instep 316, discussed further hereinbelow in conjunction with FIG. 4, a depth map is generated andprocess 300 proceeds to step 307 to perform the image depth map scan as previously described. If, however, instep 314, image information from which a depth map may be generated has not been included in the web page,process 300 terminates,step 318. - Referring now to FIG. 4, step416 of FIG. 3 for generating a depth map is described in additional detail. In
step 402, an image set is input. This may, for example, be a sequence of images constituting a portion of a motion picture file. Additionally, the image set maybe a pair of stereographic images. Instep 404, the image depth is analyzed, and a depth value assigned to each pixel of the image. Techniques for analyzing depths in an image from stereographic views and motion are described in R. Dutta and C. C. Weems, Parallel Dense Depth from Motion on the Image Understanding Architecture; 1993 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION 154 (1993), incorporated herein by reference. Alternatively, image depth may be analyzed using commercially available software, for example, KBVision™ from Amerinex Applied Imaging, Inc., Amherst MA, or Khoros Pro™ from Khoral Research, Inc., Albuquerque, N.M., maybe used. Instep 406, the depth map is filled by setting the data values in a data structure, such as a two-dimensional array, and the depth map containing the depth values generated inanalysis step 404, is output. This depth map may then be scanned in accordance with the principles ofmethodology 300, FIG. 3, as previously described. - Although the present invention and its advantages have been described in detail, it should be understood that various changes, substitutions and alterations can be made herein without departing from the spirit and scope of the invention as defined by the appended claims.
Claims (21)
1. A depth cue method comprising the steps of:
scanning a depth map corresponding to an image, in response to user input; and
outputting a nonvisual cue corresponding to a depth value in said depth map, for each pixel scanned.
2. The method of claim 1 wherein said nonvisual cue is selected from the group consisting of auditory cues and tactile cues.
3. The method of claim 1 wherein said depth map is received in response to a web page containing said image.
4. The method of claim 3 further comprising the step of, if no depth map is received in response to said web page containing said image, generating said depth map.
5. The method of claim 4 wherein said step of generating said depth map comprises:
performing a depth analysis of a set of images associated with said image, said set of images operable for extracting depth information therefrom.; and
assigning a depth value corresponding to said depth information for each pixel corresponding to said image.
6. The method of claim 5 wherein said set of images associated with said image is selected from the group consisting of a stereographic pair including said image and a plurality of images operable for displaying motion.
7. The method of claim 5 wherein said step of generating said depth map further comprises the steps of:
setting each depth value in a data structure to form said depth map; and
outputting said data structure.
8. A computer program product embodied in a tangible storage medium, the program product for accessing graphical data, the program product including a program of instructions for performing the steps of:
scanning a depth map corresponding to an image, in response to user input; and
outputting a nonvisual cue corresponding to a depth value in said depth map, for each pixel scanned.
9. The program product of claim 8 wherein said nonvisual cue is selected from the group consisting of auditory cues and tactile cues.
10. The program product of claim 8 wherein said depth map is received in response to a web page containing said image.
11. The program product ofclaim 10 further comprising programming for performing the step of, if no depth map is received in response to said web page containing said image, generating said depth map.
12. The method of claim 11 wherein said programming for performing step of generating said depth map comprises programming for performing the steps of:
performing a depth analysis of a set of images associated with said image, said set of images operable for extracting depth information therefrom.; and
assigning a depth value corresponding to said depth information for each pixel corresponding to said image.
13. The program product of claim 12 wherein said set of images associated with said image is selected from the group consisting of a stereographic pair including said image and a plurality of images operable for displaying motion.
14. The program product of claim 12 wherein said programming for performing step of generating said depth map further comprises programming for performing the steps of:
setting each depth value in a data structure to form said depth map; and
outputting said data structure.
15. A data processing system comprising:
circuitry operable for scanning a depth map corresponding to an image, in response to user input; and
outputting a nonvisual cue corresponding to a depth value in said depth map, for each pixel scanned.
16. The system of claim 15 wherein said nonvisual cue is selected from the group consisting of auditory cues and tactile cues.
17. The system of claim 15 wherein said depth map is received in response to a web page containing said image.
18. The system of claim 17 further comprising circuitry operable for, if no depth map is received in response to said web page containing said image, generating said depth map.
19. The system of claim 18 wherein said circuitry operable for generating said depth map comprises:
circuitry operable for performing a depth analysis of a set of images associated with said image, said set of images operable for extracting depth information therefrom.; and
circuitry operable for assigning a depth value corresponding to said depth information for each pixel corresponding to said image.
20. The system of claim 19 wherein said set of images associated with said image is selected from the group consisting of a stereographic pair including said image and a plurality of images operable for displaying motion.
21. The system of claim 17 wherein said circuitry operable for generating said depth map further comprises:
circuitry operable for setting each depth value in a data structure to form said depth map; and
circuitry operable for outputting said data structure.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/814,397 US20020138264A1 (en) | 2001-03-21 | 2001-03-21 | Apparatus to convey depth information in graphical images and method therefor |
TW091105071A TW552509B (en) | 2001-03-21 | 2002-03-18 | Apparatus to convey depth information in graphical images and method therefor |
JP2002073546A JP2002358540A (en) | 2001-03-21 | 2002-03-18 | Device and method for transmitting depth information in graphical image |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/814,397 US20020138264A1 (en) | 2001-03-21 | 2001-03-21 | Apparatus to convey depth information in graphical images and method therefor |
Publications (1)
Publication Number | Publication Date |
---|---|
US20020138264A1 true US20020138264A1 (en) | 2002-09-26 |
Family
ID=25214940
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/814,397 Abandoned US20020138264A1 (en) | 2001-03-21 | 2001-03-21 | Apparatus to convey depth information in graphical images and method therefor |
Country Status (3)
Country | Link |
---|---|
US (1) | US20020138264A1 (en) |
JP (1) | JP2002358540A (en) |
TW (1) | TW552509B (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080021866A1 (en) * | 2006-07-20 | 2008-01-24 | Heather M Hinton | Method and system for implementing a floating identity provider model across data centers |
US20090103616A1 (en) * | 2007-10-19 | 2009-04-23 | Gwangju Institute Of Science And Technology | Method and device for generating depth image using reference image, method for encoding/decoding depth image, encoder or decoder for the same, and recording medium recording image generated using the method |
US20120050284A1 (en) * | 2010-08-27 | 2012-03-01 | Samsung Electronics Co., Ltd. | Method and apparatus for implementing three-dimensional image |
EP2642370A1 (en) * | 2012-03-21 | 2013-09-25 | Samsung Electronics Co., Ltd | Display method and apparatus for an electronic device |
US9162061B2 (en) | 2009-10-09 | 2015-10-20 | National Ict Australia Limited | Vision enhancement for a vision impaired user |
CN105892971A (en) * | 2016-03-30 | 2016-08-24 | 联想(北京)有限公司 | Image display method and electronic equipment |
US10402159B1 (en) * | 2015-03-13 | 2019-09-03 | Amazon Technologies, Inc. | Audible user interface system |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150379720A1 (en) * | 2013-01-31 | 2015-12-31 | Threevolution Llc | Methods for converting two-dimensional images into three-dimensional images |
Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5237647A (en) * | 1989-09-15 | 1993-08-17 | Massachusetts Institute Of Technology | Computer aided drawing in three dimensions |
US5255211A (en) * | 1990-02-22 | 1993-10-19 | Redmond Productions, Inc. | Methods and apparatus for generating and processing synthetic and absolute real time environments |
US5371627A (en) * | 1992-10-23 | 1994-12-06 | N.E. Thing Enterprises, Inc. | Random dot stereogram and method for making the same |
US5416510A (en) * | 1991-08-28 | 1995-05-16 | Stereographics Corporation | Camera controller for stereoscopic video system |
US5495576A (en) * | 1993-01-11 | 1996-02-27 | Ritchey; Kurtis J. | Panoramic image based virtual reality/telepresence audio-visual system and method |
US5687331A (en) * | 1995-08-03 | 1997-11-11 | Microsoft Corporation | Method and system for displaying an animated focus item |
US5784052A (en) * | 1995-03-13 | 1998-07-21 | U.S. Philips Corporation | Vertical translation of mouse or trackball enables truly 3D input |
US5984475A (en) * | 1997-12-05 | 1999-11-16 | Mcgill University | Stereoscopic gaze controller |
US6164973A (en) * | 1995-01-20 | 2000-12-26 | Vincent J. Macri | Processing system method to provide users with user controllable image for use in interactive simulated physical movements |
US6320496B1 (en) * | 1999-04-29 | 2001-11-20 | Fuji Xerox Co., Ltd | Systems and methods providing tactile guidance using sensory supplementation |
US6366281B1 (en) * | 1996-12-06 | 2002-04-02 | Stereographics Corporation | Synthetic panoramagram |
US6466185B2 (en) * | 1998-04-20 | 2002-10-15 | Alan Sullivan | Multi-planar volumetric display system and method of operation using psychological vision cues |
US6536553B1 (en) * | 2000-04-25 | 2003-03-25 | The United States Of America As Represented By The Secretary Of The Army | Method and apparatus using acoustic sensor for sub-surface object detection and visualization |
US6563105B2 (en) * | 1999-06-08 | 2003-05-13 | University Of Washington | Image acquisition with depth enhancement |
US6664962B1 (en) * | 2000-08-23 | 2003-12-16 | Nintendo Co., Ltd. | Shadow mapping in a low cost graphics system |
US6674877B1 (en) * | 2000-02-03 | 2004-01-06 | Microsoft Corporation | System and method for visually tracking occluded objects in real time |
-
2001
- 2001-03-21 US US09/814,397 patent/US20020138264A1/en not_active Abandoned
-
2002
- 2002-03-18 JP JP2002073546A patent/JP2002358540A/en active Pending
- 2002-03-18 TW TW091105071A patent/TW552509B/en not_active IP Right Cessation
Patent Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5237647A (en) * | 1989-09-15 | 1993-08-17 | Massachusetts Institute Of Technology | Computer aided drawing in three dimensions |
US5255211A (en) * | 1990-02-22 | 1993-10-19 | Redmond Productions, Inc. | Methods and apparatus for generating and processing synthetic and absolute real time environments |
US5416510A (en) * | 1991-08-28 | 1995-05-16 | Stereographics Corporation | Camera controller for stereoscopic video system |
US5371627A (en) * | 1992-10-23 | 1994-12-06 | N.E. Thing Enterprises, Inc. | Random dot stereogram and method for making the same |
US5495576A (en) * | 1993-01-11 | 1996-02-27 | Ritchey; Kurtis J. | Panoramic image based virtual reality/telepresence audio-visual system and method |
US6164973A (en) * | 1995-01-20 | 2000-12-26 | Vincent J. Macri | Processing system method to provide users with user controllable image for use in interactive simulated physical movements |
US5784052A (en) * | 1995-03-13 | 1998-07-21 | U.S. Philips Corporation | Vertical translation of mouse or trackball enables truly 3D input |
US5687331A (en) * | 1995-08-03 | 1997-11-11 | Microsoft Corporation | Method and system for displaying an animated focus item |
US6366281B1 (en) * | 1996-12-06 | 2002-04-02 | Stereographics Corporation | Synthetic panoramagram |
US5984475A (en) * | 1997-12-05 | 1999-11-16 | Mcgill University | Stereoscopic gaze controller |
US6466185B2 (en) * | 1998-04-20 | 2002-10-15 | Alan Sullivan | Multi-planar volumetric display system and method of operation using psychological vision cues |
US6320496B1 (en) * | 1999-04-29 | 2001-11-20 | Fuji Xerox Co., Ltd | Systems and methods providing tactile guidance using sensory supplementation |
US6563105B2 (en) * | 1999-06-08 | 2003-05-13 | University Of Washington | Image acquisition with depth enhancement |
US6674877B1 (en) * | 2000-02-03 | 2004-01-06 | Microsoft Corporation | System and method for visually tracking occluded objects in real time |
US6536553B1 (en) * | 2000-04-25 | 2003-03-25 | The United States Of America As Represented By The Secretary Of The Army | Method and apparatus using acoustic sensor for sub-surface object detection and visualization |
US6664962B1 (en) * | 2000-08-23 | 2003-12-16 | Nintendo Co., Ltd. | Shadow mapping in a low cost graphics system |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080021866A1 (en) * | 2006-07-20 | 2008-01-24 | Heather M Hinton | Method and system for implementing a floating identity provider model across data centers |
US20090103616A1 (en) * | 2007-10-19 | 2009-04-23 | Gwangju Institute Of Science And Technology | Method and device for generating depth image using reference image, method for encoding/decoding depth image, encoder or decoder for the same, and recording medium recording image generated using the method |
US9162061B2 (en) | 2009-10-09 | 2015-10-20 | National Ict Australia Limited | Vision enhancement for a vision impaired user |
US20120050284A1 (en) * | 2010-08-27 | 2012-03-01 | Samsung Electronics Co., Ltd. | Method and apparatus for implementing three-dimensional image |
EP2642370A1 (en) * | 2012-03-21 | 2013-09-25 | Samsung Electronics Co., Ltd | Display method and apparatus for an electronic device |
US20130249975A1 (en) * | 2012-03-21 | 2013-09-26 | Samsung Electronics Co., Ltd | Method and apparatus for displaying on electronic device |
US10402159B1 (en) * | 2015-03-13 | 2019-09-03 | Amazon Technologies, Inc. | Audible user interface system |
CN105892971A (en) * | 2016-03-30 | 2016-08-24 | 联想(北京)有限公司 | Image display method and electronic equipment |
Also Published As
Publication number | Publication date |
---|---|
TW552509B (en) | 2003-09-11 |
JP2002358540A (en) | 2002-12-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
TW552527B (en) | Method and system for the international support of Internet web pages | |
CN1152320C (en) | Web page adaption system relative to size of display screen and window | |
CN1142513C (en) | Dynamic content supplied processor | |
Lei et al. | Context-based media adaptation in pervasive computing | |
US7228495B2 (en) | Method and system for providing an index to linked sites on a web page for individuals with visual disabilities | |
JP4851763B2 (en) | Document retrieval technology using image capture device | |
US9779066B2 (en) | Method and system for converting punycode text to ASCII/unicode text | |
JP2000090001A (en) | Method and system for conversion of electronic data using conversion setting | |
EP0844573A2 (en) | Method and system for rendering hyper-link information in a printable medium | |
US20020097264A1 (en) | Apparatus and methods for management of temporal parameters to provide enhanced accessibility to computer programs | |
US6961458B2 (en) | Method and apparatus for presenting 3-dimensional objects to visually impaired users | |
CN1168506A (en) | Method and apparatus for controlling peripheral equipment | |
JP2002108870A (en) | System and method for processing information | |
CN101582083A (en) | Web-based detection in image, extraction and recognition | |
CN106383875A (en) | Artificial intelligence-based man-machine interaction method and device | |
CN1602482A (en) | System and method to facilitate translation of communications between entities over a network | |
US20020138264A1 (en) | Apparatus to convey depth information in graphical images and method therefor | |
JPH10177613A (en) | Method and device for generating and inputting url | |
US20020111974A1 (en) | Method and apparatus for early presentation of emphasized regions in a web page | |
US6636235B1 (en) | Lettering adjustments for display resolution | |
US20020158903A1 (en) | Apparatus for outputting textual renditions of graphical data and method therefor | |
US20050193018A1 (en) | Utilizing a scannable URL (Universal Resource Locator) | |
WO2002099697A1 (en) | Method and system for displaying the searching result with specified image | |
KR100351478B1 (en) | Method for setting up and using preference in web browser | |
JPH10301944A (en) | Www browser device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JANANIKIRAMAN, JANANI;DUTTA, RABINDRANATH;REEL/FRAME:011686/0137;SIGNING DATES FROM 20010305 TO 20010307 |
|
STCB | Information on status: application discontinuation |
Free format text: EXPRESSLY ABANDONED -- DURING EXAMINATION |