US20060098864A1 - Method and apparatus to display 3d rendered ultrasound data on an ultrasound cart in stereovision - Google Patents

Method and apparatus to display 3d rendered ultrasound data on an ultrasound cart in stereovision Download PDF

Info

Publication number
US20060098864A1
US20060098864A1 US10/536,643 US53664305A US2006098864A1 US 20060098864 A1 US20060098864 A1 US 20060098864A1 US 53664305 A US53664305 A US 53664305A US 2006098864 A1 US2006098864 A1 US 2006098864A1
Authority
US
United States
Prior art keywords
ultrasound
stereovision
image
data volumes
signals
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/536,643
Inventor
Jonathan Ziel
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Koninklijke Philips NV
Original Assignee
Koninklijke Philips Electronics NV
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips Electronics NV filed Critical Koninklijke Philips Electronics NV
Priority to US10/536,643 priority Critical patent/US20060098864A1/en
Assigned to KONINKLIJKE PHILIPS ELECTRONICS, N.V. reassignment KONINKLIJKE PHILIPS ELECTRONICS, N.V. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ZIEL, J. MARK
Publication of US20060098864A1 publication Critical patent/US20060098864A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/52Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
    • G01S7/52017Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
    • G01S7/52053Display arrangements
    • G01S7/52057Cathode ray tube displays
    • G01S7/52068Stereoscopic displays; Three-dimensional displays; Pseudo 3D displays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/275Image signal generators from 3D object models, e.g. computer-generated stereoscopic image signals
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S15/00Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
    • G01S15/88Sonar systems specially adapted for specific applications
    • G01S15/89Sonar systems specially adapted for specific applications for mapping or imaging
    • G01S15/8906Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
    • G01S15/8993Three dimensional imaging systems

Definitions

  • the present invention relates to a portable ultrasound device that is housed on an ultrasound cart, and more particularly, to an ultrasound device that produces a real time stereovision image while still on the ultrasound cart.
  • FIG. 1 is a diagram illustrating a conventional system 1 of generating the ultrasound image.
  • An ultrasound machine 10 is small enough to fit on a cart 14 , thereby allowing the ultrasound machine 10 to be transported to bed-ridden patients or from one operating room to another.
  • the ultrasound machine 10 emits and receives the ultrasound signals and generates a volume of data from the received ultrasound signals.
  • This data is used to generate the ultrasound image on a screen 12 of the ultrasound machine 10 , or may be downloaded onto a disc (not shown) which is physically carried to an off-line workstation 20 , or exported electronically over a network 30 and stored at an external location such as an internet or intranet website or a network storage server, where the data can be accessed by a workstation to generate and analyze images.
  • a disadvantage of the conventional system 1 is that it cannot generate a stereovision image in real time while on the cart 14 .
  • “Real time” means that it appears from the point of view of a user that the image generated on the screen 12 represents the actual condition of a patient at a particular instant in time, even though it may take a very small but finite amount of time for the system to process the information and display the same. Thus, as far as the user can detect, the ultrasound image is contemporaneously displaying the object being analyzed.
  • “Real Time” can be achieved with a frame update rate greater than or equal to 5 Hz (frames per second) and a latency from the start of dataset acquisition to display of less than or equal to 0.5 seconds.
  • a stereovision image has true depth, as opposed to two-dimensional (2D) images that try to achieve a three-dimensional (3D) effect by shading or other methods of providing depth dependent visual cues such as 1) Perspective Projection which makes objects farther away from the viewer look smaller, for example, railroad tracks appear to converge in the distance; 2) Depth Dependent Shading which makes objects farther away from the viewer look darker, 3) Lighting which indicates depth with shadows.
  • the stereovision effect is achieved by reproducing an image from two slightly different angles, for example, a left eye angle and a right eye angle, and alternately displaying the left and right images, thereby fooling the brain into seeing an image with true depth.
  • the user feels as if he can put his hand behind the image or even inside it.
  • Images generated on the cart 14 have previously been limited to images requiring only small amounts of data.
  • the conventional system 1 can generate a 2D image in real time on the cart 14 , if there is no 3D effect.
  • the conventional system 1 can generate a 2D image with a 3D effect on the cart 14 , but this is not a stereovision image, and this image is not in real time. Instead, there is a delay between acquiring the image data and generating the image. Historically, it has taken 3-5 minutes to acquire the data, and then upwards of 10-20 minutes before the image is displayed. Often, the image is rotated slightly, typically between 5 and 20 degrees, back and forth, in order to achieve a perception of depth.
  • This system is disadvantageous because it does not display an image having true depth. Furthermore, due to the delay between acquiring the data and generating the image, it is extremely time consuming to achieve an acceptable image.
  • the initial image is not satisfactory, and the sonographer (such as a doctor or technician) must adjust scanning parameters such as gain and angle of insonation with image controls 15 .
  • Color doppler imaging mode requires additional user controls that are interdependent and must be optimized together such as wall filter settings, scale settings, threshold, etc.
  • the sonographer may want to view an entirely different portion of the object. Each time an adjustment is made, the user must wait up to 20 minutes to see whether the adjustment achieved the desired result, instead of receiving immediate feedback. Multiple adjustments are often required, resulting in a significant waiting period before a useful image is achieved.
  • the conventional ultrasound machine 10 collects 3D data from the reflected ultrasound signals, and this data must be rendered into a 2D representation.
  • the conventional ultrasound machine 10 cannot render the data “on-the-fly” by streaming the data, but instead rendering is accomplished in “batch” mode after the acquisition is complete. Streaming indicates that the acquisition does not stop when the rendering starts: as one dataset is being rendered the next dataset is being acquired. Batch mode means that first, the entire data set is acquired. Then, it is saved as a file. Then a rendering program opens that file and renders the data. Often, “batch” processing includes writing the file to a compact disc (CD) or sending the data over a network to a storage server. The data is then analyzed with an off-line workstation 20 .
  • CD compact disc
  • Stereovision is an example of an image type that previously required off-cart analysis. Since stereovision requires left and right angle views of each dataset, twice as much processing must occur on each dataset. Stereovision also requires a rapidly alternating display of the left and right views. To successfully achieve the 3D effect, the two images must be displayed alternately at 120 Hz. That is, the left angle image must be displayed at 60 Hz interleaved with the right angle image at 60 Hz. However, in previous designs, a maximum of only 30 image frames per second could be displayed on the cart, which is too slow to achieve the stereovision effect. This limitation of the conventional system 1 is not limited to stereovision images, but applies to other 3D representations as well.
  • Generating the image off of the cart 14 is disadvantageous because it is time consuming. Furthermore, if the ultrasound procedure was not performed properly, this fact would not be discovered until a later time. Thus, the patient would have to have the ultrasound device applied a second time, either after waiting for the off-line results, or by coming back at a later date.
  • the present invention relates to an ultrasound apparatus comprising an emitter to emit ultrasound signals, a receiver to receive reflected ultrasound signals, and a display unit to display a stereovision ultrasound image in real time from the reflected ultrasound signals.
  • the apparatus also includes an acquisition subsystem to acquire the 3D ultrasound volume data from the reflected ultrasound signals, and a rendering processor to render the 3D ultrasound data volumes into left and right angle 2D images in streaming mode.
  • a generator to generate 3D ultrasound data volumes from the reflected ultrasound signals and a rendering processor to render the 3D ultrasound data volumes into first and second 2D images by streaming may also be included.
  • a transport unit such as a cart houses said emitter, receiver, display unit, acquisition subsystem, and rendering processor.
  • FIG. 1 is a diagram illustrating a conventional system of generating an ultrasound image
  • FIG. 2 is a block diagram of an ultrasound apparatus in accordance with the present invention.
  • FIG. 3 is a diagram illustrating different methods of viewing the image displayed by the ultrasound apparatus of FIG. 2 to achieve a stereovision image
  • FIG. 4 is a block diagram of an ultrasound apparatus using a time-interleaved, or serial rendering.
  • FIG. 2 is a block diagram of an ultrasound apparatus 100 in accordance with the present invention, which is similar in appearance to the conventional ultrasound machine 10 of FIG. 4 .
  • the ultrasound apparatus 100 includes a transducer 110 to emit a plurality of ultrasound signals to an object (not shown).
  • the transducer 110 emits and receives ultrasound signals to create a volume dataset.
  • the ultrasound signals are reflected by the object and received by the transducer 110 .
  • the transducer 110 may be a two-dimensional phased array transducer.
  • the two-dimensional phased array transducer includes a probe (not shown) including a plurality of elements to generate the emitted ultrasound signals. These elements are arranged in a two-dimensional array, for example, in a rectangular or circular shape.
  • a mechanical transducer having a one-dimensional column of elements that is swept mechanically to interrogate a volume.
  • the transducer 110 then receives the reflected ultrasound signals and passes them on to an ultrasound scanner 120 , which performs beamforming and generates a stream of detected 3D ultrasound volume datasets from the reflected ultrasound signals.
  • the ultrasound scanner 120 may be of a basic “front end” type.
  • the stream of detected ultrasound data volumes is then received by left and right eye rendering processors 130 , 132 which simultaneously render the left eye angle and the right eye angle detected ultrasound data volumes to generate left and right 2D rendered images.
  • the 2D rendered images are then received by left and right 2D frame buffers 140 , 142 , respectively, which hold the 2D rendered images as frames.
  • a multiplexor 150 then toggles between frame buffers 140 , 142 , alternately selecting the left 2D rendered image and the right 2D rendered image and transmitting the selected image to a display monitor 160 . This may be done at a rate of 120 Hertz, or 60 left images and 60 right images being displayed each second.
  • a toggle 152 changes the rate at which the multiplexor 150 selects between the frame buffers 140 , 142 .
  • a personal computer may be used as the rendering processors 130 , 132 and the frame buffers 140 , 142 , as indicated by dotted box 190 a .
  • the PC may also include the multiplexor 150 , as indicated by dotted box 190 b .
  • Display monitor 160 alternately displays the left 2D rendered image and the right 2D rendered image.
  • the transducer 110 , ultrasound scanner 120 , rendering processors 130 , 132 , frame buffers 140 , 142 , multiplexor 150 , and display monitor 160 are all housed on a single cart (not shown), which is similar to the cart 14 of FIG. 4 .
  • FIG. 3 is a diagram illustrating different methods of viewing the image displayed by the display monitor 160 to achieve the stereovision image.
  • the top portion of FIG. 3 illustrates a user wearing shuttered glasses 170 .
  • the shuttered glasses 170 include shutters 172 , which alternately open and close on the left and right sides in synch with the display of the left and right 2D rendered images.
  • FIG. 3 further illustrates that the display monitor 160 may track a movement of the eye 180 of the user, to thereby generate the stereovision image.
  • the present invention may also use recently developed monitors that do not need to track the eye movement.
  • FIG. 3 illustrates that the user may wear a virtual reality headset 190 .
  • the user may view an image having true depth, but as the user changes his view, the image changes. For example, the user could be virtually positioned in the middle of the heart. If the user turns his head to the right, he would be looking at that section of the heart.
  • the analysis of the heart or other object is facilitated by eliminating the need to manually select the view area.
  • FIG. 4 is a block diagram of an ultrasound apparatus 200 using a time-interleaved, or serial rendering.
  • the ultrasound apparatus 200 of FIG. 4 is essentially the same as the ultrasound apparatus 100 of FIG. 2 .
  • a single rendering processor is provided 234 that alternately renders both the left eye angle and right eye angle detected ultrasound volumes, and alternately generates the left and right 2D rendered images to the frame buffers 140 , 142 , respectively. This differs from the parallel rendering performed by the rendering processors 130 , 132 of FIG. 1 .
  • the present invention is able to collect and process the large amount of data necessary to generate a stereovision image while still on the cart 14 .
  • the rendering processors 130 , 132 , 234 render these volumes by streaming.
  • the rendering processors 130 , 132 , 234 may use any of several rendering algorithms to stream the ultrasound volumes. For example, the Shear-warp rendering algorithm published by Lacroute, P. & Levoy, M. [1994]. Fast volume rendering using a shear-warp factorization of the viewing transformation, Computer Graphics Proceedings, Annual Conference Series (SIGGRAPH '94), Orlando, pp. 451-458, may be used. Ray Casting, published by LeVoy, M. [1990]. Efficient ray tracing of volume data, ACM Transactions on Graphics 9(3): 245-261 is another algorithm that may be used.
  • the present invention eliminates the need for the time consuming process of downloading data and displaying the image at a remote location. Thus, errors in the sonography are detected immediately. Furthermore, a volume update rate in the order of 20-30 frames per second is possible. Thus, 20-30 new images are generated per second as opposed to the previous rate of one frame per second. Due to the improved display rate, when image parameters are adjusted, the updated image is displayed within a fraction of a second, as opposed to the previous 10-20 minute delay. The user perceives the feedback as contemporaneous with the adjustments because only a small fraction of a second is required to process the data Thus, unlike the previous designs, the present invention can provide a truly “real time” stereovision image on the cart 14 .
  • the image can be updated at a rate of greater than or equal to 20 frames per second, and a latency of less than or equal to 100 milliseconds from the start of acquisition to display is achieved.
  • the present invention may be used in CFM (Color Flow Mode) displays, which display blood velocity in color mode rather than as a black and white part of the image.
  • CFM Color Flow Mode
  • Other applications include Power Doppler, or “Angio,” in which the amplitude of the blood flow signal is displayed, and AQ (Acoustic Quantification).
  • the present invention may also be used when displaying fetal images. In recent years, the demand for three-dimensional fetal photos has increased, as proud parents are more willing than ever to pay extra for a photo of their unborn “bundle of joy.”

Abstract

An ultrasound apparatus and method including an emitter to emit ultrasound signals, a receiver to receive reflected ultrasound signals, and a display unit to display a stereovision ultrasound image in real time from the reflected ultrasound signals. The apparatus also includes a generator to generate 3D ultrasound data volumes from the reflected ultrasound signals and a rendering processor to render the 3D ultrasound data volumes into first and second 2D images by streaming. A transport unit such as a cart houses the emitter, the receiver, the display unit, the rendering processor and the generator.

Description

  • The present invention relates to a portable ultrasound device that is housed on an ultrasound cart, and more particularly, to an ultrasound device that produces a real time stereovision image while still on the ultrasound cart.
  • Ultrasound generally operates by transmitting ultrasound signals into the body, and then receiving the echoes of the ultrasound signals off of internal objects to generate an image. These internal objects may be a fetus or internal organs such as a heart or kidney. FIG. 1 is a diagram illustrating a conventional system 1 of generating the ultrasound image. An ultrasound machine 10 is small enough to fit on a cart 14, thereby allowing the ultrasound machine 10 to be transported to bed-ridden patients or from one operating room to another. The ultrasound machine 10 emits and receives the ultrasound signals and generates a volume of data from the received ultrasound signals. This data is used to generate the ultrasound image on a screen 12 of the ultrasound machine 10, or may be downloaded onto a disc (not shown) which is physically carried to an off-line workstation 20, or exported electronically over a network 30 and stored at an external location such as an internet or intranet website or a network storage server, where the data can be accessed by a workstation to generate and analyze images.
  • A disadvantage of the conventional system 1 is that it cannot generate a stereovision image in real time while on the cart 14. “Real time” means that it appears from the point of view of a user that the image generated on the screen 12 represents the actual condition of a patient at a particular instant in time, even though it may take a very small but finite amount of time for the system to process the information and display the same. Thus, as far as the user can detect, the ultrasound image is contemporaneously displaying the object being analyzed. “Real Time” can be achieved with a frame update rate greater than or equal to 5 Hz (frames per second) and a latency from the start of dataset acquisition to display of less than or equal to 0.5 seconds.
  • A stereovision image has true depth, as opposed to two-dimensional (2D) images that try to achieve a three-dimensional (3D) effect by shading or other methods of providing depth dependent visual cues such as 1) Perspective Projection which makes objects farther away from the viewer look smaller, for example, railroad tracks appear to converge in the distance; 2) Depth Dependent Shading which makes objects farther away from the viewer look darker, 3) Lighting which indicates depth with shadows. The stereovision effect is achieved by reproducing an image from two slightly different angles, for example, a left eye angle and a right eye angle, and alternately displaying the left and right images, thereby fooling the brain into seeing an image with true depth. Thus, the user feels as if he can put his hand behind the image or even inside it.
  • Images generated on the cart 14 have previously been limited to images requiring only small amounts of data. For example, the conventional system 1 can generate a 2D image in real time on the cart 14, if there is no 3D effect. The conventional system 1 can generate a 2D image with a 3D effect on the cart 14, but this is not a stereovision image, and this image is not in real time. Instead, there is a delay between acquiring the image data and generating the image. Historically, it has taken 3-5 minutes to acquire the data, and then upwards of 10-20 minutes before the image is displayed. Often, the image is rotated slightly, typically between 5 and 20 degrees, back and forth, in order to achieve a perception of depth.
  • This system is disadvantageous because it does not display an image having true depth. Furthermore, due to the delay between acquiring the data and generating the image, it is extremely time consuming to achieve an acceptable image. Typically, the initial image is not satisfactory, and the sonographer (such as a doctor or technician) must adjust scanning parameters such as gain and angle of insonation with image controls 15. Color doppler imaging mode requires additional user controls that are interdependent and must be optimized together such as wall filter settings, scale settings, threshold, etc. Furthermore, the sonographer may want to view an entirely different portion of the object. Each time an adjustment is made, the user must wait up to 20 minutes to see whether the adjustment achieved the desired result, instead of receiving immediate feedback. Multiple adjustments are often required, resulting in a significant waiting period before a useful image is achieved.
  • Another disadvantage is that the resulting image jumps around, as opposed to being smooth and lifelike. This is particularly disadvantageous when viewing a continuously moving body, such as a beating heart Thus, although vendors of these previous designs may market their images as “real time,” this depiction is erroneous, insofar as there is a significant delay between adjusting the parameters/acquiring the data and displaying the image.
  • Using previous designs, if the desired image requires large amounts of data, then the image must be generated off of the cart 14. This limitation is due to the fact that the data rendering processing of the conventional ultrasound machine 10 is too slow. The conventional ultrasound machine 10 collects 3D data from the reflected ultrasound signals, and this data must be rendered into a 2D representation. The conventional ultrasound machine 10 cannot render the data “on-the-fly” by streaming the data, but instead rendering is accomplished in “batch” mode after the acquisition is complete. Streaming indicates that the acquisition does not stop when the rendering starts: as one dataset is being rendered the next dataset is being acquired. Batch mode means that first, the entire data set is acquired. Then, it is saved as a file. Then a rendering program opens that file and renders the data. Often, “batch” processing includes writing the file to a compact disc (CD) or sending the data over a network to a storage server. The data is then analyzed with an off-line workstation 20.
  • Stereovision is an example of an image type that previously required off-cart analysis. Since stereovision requires left and right angle views of each dataset, twice as much processing must occur on each dataset. Stereovision also requires a rapidly alternating display of the left and right views. To successfully achieve the 3D effect, the two images must be displayed alternately at 120 Hz. That is, the left angle image must be displayed at 60 Hz interleaved with the right angle image at 60 Hz. However, in previous designs, a maximum of only 30 image frames per second could be displayed on the cart, which is too slow to achieve the stereovision effect. This limitation of the conventional system 1 is not limited to stereovision images, but applies to other 3D representations as well.
  • Generating the image off of the cart 14 is disadvantageous because it is time consuming. Furthermore, if the ultrasound procedure was not performed properly, this fact would not be discovered until a later time. Thus, the patient would have to have the ultrasound device applied a second time, either after waiting for the off-line results, or by coming back at a later date.
  • The present invention relates to an ultrasound apparatus comprising an emitter to emit ultrasound signals, a receiver to receive reflected ultrasound signals, and a display unit to display a stereovision ultrasound image in real time from the reflected ultrasound signals. The apparatus also includes an acquisition subsystem to acquire the 3D ultrasound volume data from the reflected ultrasound signals, and a rendering processor to render the 3D ultrasound data volumes into left and right angle 2D images in streaming mode. A generator to generate 3D ultrasound data volumes from the reflected ultrasound signals and a rendering processor to render the 3D ultrasound data volumes into first and second 2D images by streaming may also be included. A transport unit such as a cart houses said emitter, receiver, display unit, acquisition subsystem, and rendering processor.
  • These and other objects and advantages of the invention will become apparent and more readily appreciated from the following description of the preferred embodiments, taken in conjunction with the accompanying drawings of which:
  • FIG. 1 is a diagram illustrating a conventional system of generating an ultrasound image;
  • FIG. 2 is a block diagram of an ultrasound apparatus in accordance with the present invention;
  • FIG. 3 is a diagram illustrating different methods of viewing the image displayed by the ultrasound apparatus of FIG. 2 to achieve a stereovision image;
  • FIG. 4 is a block diagram of an ultrasound apparatus using a time-interleaved, or serial rendering.
  • Reference will now be made in detail to the present preferred embodiments of the present invention, examples of which are illustrated in the accompanying drawings. The embodiments are described below in order to explain the present invention by referring to the figures.
  • FIG. 2 is a block diagram of an ultrasound apparatus 100 in accordance with the present invention, which is similar in appearance to the conventional ultrasound machine 10 of FIG. 4. The ultrasound apparatus 100 includes a transducer 110 to emit a plurality of ultrasound signals to an object (not shown). The transducer 110 emits and receives ultrasound signals to create a volume dataset. The ultrasound signals are reflected by the object and received by the transducer 110.
  • The transducer 110 may be a two-dimensional phased array transducer. The two-dimensional phased array transducer includes a probe (not shown) including a plurality of elements to generate the emitted ultrasound signals. These elements are arranged in a two-dimensional array, for example, in a rectangular or circular shape. When generating an image of a fetus, it is sufficient to use a mechanical transducer, having a one-dimensional column of elements that is swept mechanically to interrogate a volume.
  • The transducer 110 then receives the reflected ultrasound signals and passes them on to an ultrasound scanner 120, which performs beamforming and generates a stream of detected 3D ultrasound volume datasets from the reflected ultrasound signals. The ultrasound scanner 120 may be of a basic “front end” type. The stream of detected ultrasound data volumes is then received by left and right eye rendering processors 130, 132 which simultaneously render the left eye angle and the right eye angle detected ultrasound data volumes to generate left and right 2D rendered images. The 2D rendered images are then received by left and right 2D frame buffers 140, 142, respectively, which hold the 2D rendered images as frames. A multiplexor 150 then toggles between frame buffers 140, 142, alternately selecting the left 2D rendered image and the right 2D rendered image and transmitting the selected image to a display monitor 160. This may be done at a rate of 120 Hertz, or 60 left images and 60 right images being displayed each second. A toggle 152 changes the rate at which the multiplexor 150 selects between the frame buffers 140, 142. A personal computer (PC) may be used as the rendering processors 130, 132 and the frame buffers 140, 142, as indicated by dotted box 190 a. The PC may also include the multiplexor 150, as indicated by dotted box 190 b. Display monitor 160 alternately displays the left 2D rendered image and the right 2D rendered image. The transducer 110, ultrasound scanner 120, rendering processors 130, 132, frame buffers 140, 142, multiplexor 150, and display monitor 160 are all housed on a single cart (not shown), which is similar to the cart 14 of FIG. 4.
  • FIG. 3 is a diagram illustrating different methods of viewing the image displayed by the display monitor 160 to achieve the stereovision image. The top portion of FIG. 3 illustrates a user wearing shuttered glasses 170. The shuttered glasses 170 include shutters 172, which alternately open and close on the left and right sides in synch with the display of the left and right 2D rendered images.
  • FIG. 3 further illustrates that the display monitor 160 may track a movement of the eye 180 of the user, to thereby generate the stereovision image. The present invention may also use recently developed monitors that do not need to track the eye movement. Finally, FIG. 3 illustrates that the user may wear a virtual reality headset 190. Thus, not only does the user view an image having true depth, but as the user changes his view, the image changes. For example, the user could be virtually positioned in the middle of the heart. If the user turns his head to the right, he would be looking at that section of the heart. Thus, the analysis of the heart or other object is facilitated by eliminating the need to manually select the view area.
  • FIG. 4 is a block diagram of an ultrasound apparatus 200 using a time-interleaved, or serial rendering. The ultrasound apparatus 200 of FIG. 4 is essentially the same as the ultrasound apparatus 100 of FIG. 2. However, a single rendering processor is provided 234 that alternately renders both the left eye angle and right eye angle detected ultrasound volumes, and alternately generates the left and right 2D rendered images to the frame buffers 140, 142, respectively. This differs from the parallel rendering performed by the rendering processors 130,132 of FIG. 1.
  • Unlike previous designs, the present invention is able to collect and process the large amount of data necessary to generate a stereovision image while still on the cart 14. Instead of processing the detected ultrasound volumes in batch mode, the rendering processors 130, 132, 234 render these volumes by streaming. Thus, the rendering is more continuous than in previous designs, resulting in a smoother image that can be more quickly adjusted. The rendering processors 130, 132, 234 may use any of several rendering algorithms to stream the ultrasound volumes. For example, the Shear-warp rendering algorithm published by Lacroute, P. & Levoy, M. [1994]. Fast volume rendering using a shear-warp factorization of the viewing transformation, Computer Graphics Proceedings, Annual Conference Series (SIGGRAPH '94), Orlando, pp. 451-458, may be used. Ray Casting, published by LeVoy, M. [1990]. Efficient ray tracing of volume data, ACM Transactions on Graphics 9(3): 245-261 is another algorithm that may be used.
  • The present invention eliminates the need for the time consuming process of downloading data and displaying the image at a remote location. Thus, errors in the sonography are detected immediately. Furthermore, a volume update rate in the order of 20-30 frames per second is possible. Thus, 20-30 new images are generated per second as opposed to the previous rate of one frame per second. Due to the improved display rate, when image parameters are adjusted, the updated image is displayed within a fraction of a second, as opposed to the previous 10-20 minute delay. The user perceives the feedback as contemporaneous with the adjustments because only a small fraction of a second is required to process the data Thus, unlike the previous designs, the present invention can provide a truly “real time” stereovision image on the cart 14.
  • With the present invention, the image can be updated at a rate of greater than or equal to 20 frames per second, and a latency of less than or equal to 100 milliseconds from the start of acquisition to display is achieved.
  • This image is smooth, life-like, and has true depth, and therefore improves over the previous choppy images which rely upon 2D representations of 3D objects. This feature is especially advantageous when displaying moving objects such as beating hearts or blood flow. Specifically, the present invention may be used in CFM (Color Flow Mode) displays, which display blood velocity in color mode rather than as a black and white part of the image. Other applications include Power Doppler, or “Angio,” in which the amplitude of the blood flow signal is displayed, and AQ (Acoustic Quantification). The present invention may also be used when displaying fetal images. In recent years, the demand for three-dimensional fetal photos has increased, as proud parents are more willing than ever to pay extra for a photo of their unborn “bundle of joy.”
  • Although a few preferred embodiments of the present invention have been shown and described, it would be appreciated by those skilled in the art that changes may be made in this embodiment without departing from the principles and spirit of the invention, the scope of which is defined in the claims and their equivalents.

Claims (31)

1. A method of generating an image, comprising:
emitting ultrasound signals;
receiving reflected ultrasound signals;
converting the reflected ultrasound signals to a stereovision ultrasound image in real time; and
displaying the stereovision ultrasound image in real time.
2. The method of claim 1, wherein the converting of the reflected ultrasound signals comprises:
generating 3D ultrasound data volumes from the reflected ultrasound signals; and
rendering the 3D ultrasound data volumes into first and second 2D images by streaming,
the first and second 2D images comprising the stereovision ultrasound image.
3. The method of claim 2, further comprising adjusting the stereovision ultrasound image in real time.
4. The method of claim 3, further comprising updating the stereovision ultrasound image at a rate of greater than or equal to 10 frames per second.
5. An ultrasound apparatus, comprising:
an emitter to emit ultrasound signals;
a receiver to receive reflected ultrasound signals;
a signal processor to convert the reflected ultrasound signals to a stereovision ultrasound image in real time; and
a display unit to display the stereovision ultrasound image in real time.
6. The ultrasound apparatus of claim 5, wherein the signal processor comprises:
a generator to generate 3D ultrasound data volumes from the reflected ultrasound signals; and
a rendering processor to render the 3D ultrasound data volumes into first and second 2D images by streaming,
the first and second 2D images comprising the stereovision ultrasound image.
7. The ultrasound apparatus of claim 6, further comprising a transport unit to house said emitter, receiver, display unit, rendering processor and said generator.
8. The ultrasound apparatus of claim 7, wherein said transport unit is a cart.
9. The ultrasound apparatus of claim 7, wherein said display unit further comprises a control unit to control the stereovision ultrasound image in real time.
10. The ultrasound apparatus of claim 9, wherein the stereovision ultrasound image is updated at a rate of greater than or equal to 20 frames per second.
11. The ultrasound apparatus of claim 10, wherein the 3D ultrasound data volumes comprise first and second 3D data volumes, and said rendering processor renders the first and second 3D data volumes into the first and second 2D images, respectively.
12. The ultrasound apparatus of claim 11, further comprising a select unit to alternately transmit the first and second 2D images to said display unit to display the stereovision ultrasound image.
13. The ultrasound apparatus of claim 12, wherein a user views the stereovision ultrasound image through shuttered glasses.
14. The ultrasound apparatus of claim 12, wherein said display unit tracks an eye movement of a user to create the stereovision ultrasound image.
15. The ultrasound apparatus of claim 6, wherein said rendering processor continuously streams the 3D ultrasound data volumes.
16. The ultrasound apparatus of claim 12, wherein a user views the stereovision ultrasound image, and the stereovision ultrasound image changes corresponding to a movement of the user.
17. The ultrasound apparatus of claim 16, wherein the user views the stereovision ultrasound image through a virtual reality viewing unit connectible to the display unit to change the stereovision ultrasound image in accordance with the movement of the user.
18. The ultrasound apparatus of claim 12, wherein said rendering processor renders the first and second 3D data volumes in series.
19. The ultrasound apparatus of claim 12, wherein said rendering processor comprises left and right rendering processors to render the first and second 3D data volumes, respectively, in parallel.
20. The ultrasound apparatus of claim 12, wherein said select unit is a multiplexor.
21. The ultrasound apparatus of claim 5, wherein said emitter and said receiver comprise a two-dimensional phased array transducer.
22. An ultrasound apparatus, comprising:
an emitter to emit ultrasound signals;
a receiver to receive reflected ultrasound signals;
a signal processor to convert the reflected ultrasound signals to a stereovision ultrasound image in real time;
a display unit to display the stereovision ultrasound image; and
a transport unit to house said emitter, receiver, signal processor and said display unit.
23. The ultrasound apparatus of claim 22, wherein the signal processor comprises:
a generator to generate 3D ultrasound data volumes from the reflected ultrasound signals; and
a rendering processor to render the 3D ultrasound data volumes into first and second 2D images,
the first and second 2D images comprising the stereovision ultrasound image.
24. The ultrasound apparatus of claim 23, wherein said emitter and said receiver comprise a two-dimensional phased array transducer.
25. An ultrasound apparatus, comprising:
a transducer to emit ultrasound signals and to receive reflected ultrasound signals;
a scanner to generate a stream of detected ultrasound data volumes from the reflected ultrasound signals;
a rendering processor to render the stream of detected ultrasound data volumes into first and second 2D rendered images;
first and second buffers to hold the first and second 2D rendered images, respectively;
a display unit; and
a multiplexor to alternately transmit the first and second 2D rendered images to the display unit to generate a stereovision ultrasound image in real time.
26. The ultrasound apparatus of claim 25, further comprising a cart to house said transducer, scanner, rendering processor, first and second buffers, multiplexor and said display unit.
27. The ultrasound apparatus of claim 25, wherein said rendering processor renders the stream of detected ultrasound data volumes by streaming.
28. The ultrasound apparatus of claim 27, wherein the stereovision ultrasound image is a Color Flow Mode (CFM) image.
29. The ultrasound apparatus of claim 27, wherein the stereovision ultrasound image is a Power Doppler image.
30. The ultrasound apparatus of claim 27, wherein the stereovision ultrasound image is an Acoustic Quantification (AQ) image.
31. The method of claim 3, further comprising updating the stereovision ultrasound image at a latency of less than or equal to 200 milliseconds from start of acquisition to display.
US10/536,643 2002-12-03 2003-11-13 Method and apparatus to display 3d rendered ultrasound data on an ultrasound cart in stereovision Abandoned US20060098864A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US10/536,643 US20060098864A1 (en) 2002-12-03 2003-11-13 Method and apparatus to display 3d rendered ultrasound data on an ultrasound cart in stereovision

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US43039602P 2002-12-03 2002-12-03
PCT/IB2003/005438 WO2004051307A2 (en) 2002-12-03 2003-11-13 Method and apparatus to display 3d rendered ultrasound data on an ultrasound cart in stereovision
US10/536,643 US20060098864A1 (en) 2002-12-03 2003-11-13 Method and apparatus to display 3d rendered ultrasound data on an ultrasound cart in stereovision

Publications (1)

Publication Number Publication Date
US20060098864A1 true US20060098864A1 (en) 2006-05-11

Family

ID=32469462

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/536,643 Abandoned US20060098864A1 (en) 2002-12-03 2003-11-13 Method and apparatus to display 3d rendered ultrasound data on an ultrasound cart in stereovision

Country Status (4)

Country Link
US (1) US20060098864A1 (en)
EP (1) EP1570294A2 (en)
AU (1) AU2003280180A1 (en)
WO (1) WO2004051307A2 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040150963A1 (en) * 2003-01-31 2004-08-05 Sonosite, Inc. System for use with a modular apparatus for diagnostic ultrasound
US20040152982A1 (en) * 2002-03-29 2004-08-05 Sonosite, Inc. Modular apparatus for diagnostic ultrasound
US20080266301A1 (en) * 2007-04-25 2008-10-30 Atmel Corporation Display controller operating mode using multiple data buffers
US7849250B2 (en) 2006-10-31 2010-12-07 Sonosite, Inc. Docking station with hierarchal battery management for use with portable medical equipment
US8398408B1 (en) 2009-02-25 2013-03-19 Sonosite, Inc. Charging station for cordless ultrasound cart
US20130150721A1 (en) * 2010-12-24 2013-06-13 Panasonic Corporation Ultrasound diagnostic apparatus and ultrasound diagnostic apparatus control method
US20140303491A1 (en) * 2013-04-04 2014-10-09 Children's National Medical Center Device and method for generating composite images for endoscopic surgery of moving and deformable anatomy
US20180152694A1 (en) * 2015-06-19 2018-05-31 Maxell, Ltd. Head mounted display device and method for providing visual aid using same
US10743844B2 (en) 2014-07-29 2020-08-18 Koninklijke Philips N.V. Ultrasound imaging apparatus
US10970921B2 (en) 2016-09-30 2021-04-06 University Hospitals Cleveland Medical Center Apparatus and method for constructing a virtual 3D model from a 2D ultrasound video

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100932977B1 (en) 2005-07-05 2009-12-21 삼성모바일디스플레이주식회사 Stereoscopic video display
KR100913173B1 (en) * 2005-07-05 2009-08-19 삼성모바일디스플레이주식회사 3 dimension graphic processor and autostereoscopic display device using the same
EP1750460A1 (en) 2005-08-05 2007-02-07 Samsung SDI Co., Ltd. 3D graphics processor and autostereoscopic display device using the same

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5917937A (en) * 1997-04-15 1999-06-29 Microsoft Corporation Method for performing stereo matching to recover depths, colors and opacities of surface elements
US5993391A (en) * 1997-09-25 1999-11-30 Kabushiki Kaisha Toshiba Ultrasound diagnostic apparatus
US6210334B1 (en) * 1999-03-31 2001-04-03 Acuson Corporation Medical diagnostic ultrasound method and apparatus for harmonic detection using doppler processing
US6245017B1 (en) * 1998-10-30 2001-06-12 Kabushiki Kaisha Toshiba 3D ultrasonic diagnostic apparatus
US20010007919A1 (en) * 1996-06-28 2001-07-12 Ramin Shahidi Method and apparatus for volumetric image navigation
US20020041327A1 (en) * 2000-07-24 2002-04-11 Evan Hildreth Video-based image control system
US6436049B1 (en) * 1999-05-31 2002-08-20 Kabushiki Kaisha Toshiba Three-dimensional ultrasound diagnosis based on contrast echo technique
US6450961B1 (en) * 1999-06-03 2002-09-17 Kabushiki Kaisha Toshiba Ultrasound imaging using flash echo imaging technique
US20030013955A1 (en) * 2001-06-22 2003-01-16 Poland Mckee D. Ultrasound system for the production of 3-D images
US6980210B1 (en) * 1997-11-24 2005-12-27 3-D Image Processing Gmbh 3D stereo real-time sensor system, method and computer program therefor

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5899861A (en) * 1995-03-31 1999-05-04 Siemens Medical Systems, Inc. 3-dimensional volume by aggregating ultrasound fields of view
JP4298016B2 (en) * 1997-09-25 2009-07-15 株式会社東芝 Ultrasonic diagnostic equipment
US6048312A (en) * 1998-04-23 2000-04-11 Ishrak; Syed Omar Method and apparatus for three-dimensional ultrasound imaging of biopsy needle
US6413219B1 (en) * 1999-03-31 2002-07-02 General Electric Company Three-dimensional ultrasound data display using multiple cut planes

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20010007919A1 (en) * 1996-06-28 2001-07-12 Ramin Shahidi Method and apparatus for volumetric image navigation
US5917937A (en) * 1997-04-15 1999-06-29 Microsoft Corporation Method for performing stereo matching to recover depths, colors and opacities of surface elements
US5993391A (en) * 1997-09-25 1999-11-30 Kabushiki Kaisha Toshiba Ultrasound diagnostic apparatus
US6980210B1 (en) * 1997-11-24 2005-12-27 3-D Image Processing Gmbh 3D stereo real-time sensor system, method and computer program therefor
US6245017B1 (en) * 1998-10-30 2001-06-12 Kabushiki Kaisha Toshiba 3D ultrasonic diagnostic apparatus
US6210334B1 (en) * 1999-03-31 2001-04-03 Acuson Corporation Medical diagnostic ultrasound method and apparatus for harmonic detection using doppler processing
US6436049B1 (en) * 1999-05-31 2002-08-20 Kabushiki Kaisha Toshiba Three-dimensional ultrasound diagnosis based on contrast echo technique
US6450961B1 (en) * 1999-06-03 2002-09-17 Kabushiki Kaisha Toshiba Ultrasound imaging using flash echo imaging technique
US20020041327A1 (en) * 2000-07-24 2002-04-11 Evan Hildreth Video-based image control system
US20030013955A1 (en) * 2001-06-22 2003-01-16 Poland Mckee D. Ultrasound system for the production of 3-D images

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040152982A1 (en) * 2002-03-29 2004-08-05 Sonosite, Inc. Modular apparatus for diagnostic ultrasound
US7534211B2 (en) 2002-03-29 2009-05-19 Sonosite, Inc. Modular apparatus for diagnostic ultrasound
US8088071B2 (en) 2002-03-29 2012-01-03 Sonosite, Inc. Modular apparatus for diagnostic ultrasound
US20040150963A1 (en) * 2003-01-31 2004-08-05 Sonosite, Inc. System for use with a modular apparatus for diagnostic ultrasound
US7591786B2 (en) 2003-01-31 2009-09-22 Sonosite, Inc. Dock for connecting peripheral devices to a modular diagnostic ultrasound apparatus
US7849250B2 (en) 2006-10-31 2010-12-07 Sonosite, Inc. Docking station with hierarchal battery management for use with portable medical equipment
US20080266301A1 (en) * 2007-04-25 2008-10-30 Atmel Corporation Display controller operating mode using multiple data buffers
US8102401B2 (en) * 2007-04-25 2012-01-24 Atmel Corporation Display controller operating mode using multiple data buffers
US8398408B1 (en) 2009-02-25 2013-03-19 Sonosite, Inc. Charging station for cordless ultrasound cart
US20130150721A1 (en) * 2010-12-24 2013-06-13 Panasonic Corporation Ultrasound diagnostic apparatus and ultrasound diagnostic apparatus control method
US20140303491A1 (en) * 2013-04-04 2014-10-09 Children's National Medical Center Device and method for generating composite images for endoscopic surgery of moving and deformable anatomy
US10426345B2 (en) * 2013-04-04 2019-10-01 Children's National Medical Center System for generating composite images for endoscopic surgery of moving and deformable anatomy
US10743844B2 (en) 2014-07-29 2020-08-18 Koninklijke Philips N.V. Ultrasound imaging apparatus
US20180152694A1 (en) * 2015-06-19 2018-05-31 Maxell, Ltd. Head mounted display device and method for providing visual aid using same
US10951882B2 (en) * 2015-06-19 2021-03-16 Maxell, Ltd. Head mounted display device and method for providing visual aid using same
US11350081B2 (en) 2015-06-19 2022-05-31 Maxell, Ltd. Head mounted display device and method for providing visual aid using same
US11606548B2 (en) 2015-06-19 2023-03-14 Maxell, Ltd. Head mounted display device and method for providing visual aid using same
US10970921B2 (en) 2016-09-30 2021-04-06 University Hospitals Cleveland Medical Center Apparatus and method for constructing a virtual 3D model from a 2D ultrasound video

Also Published As

Publication number Publication date
WO2004051307A2 (en) 2004-06-17
EP1570294A2 (en) 2005-09-07
WO2004051307A3 (en) 2004-07-29
AU2003280180A1 (en) 2004-06-23

Similar Documents

Publication Publication Date Title
US5488952A (en) Stereoscopically display three dimensional ultrasound imaging
US5993391A (en) Ultrasound diagnostic apparatus
US7356178B2 (en) System and method for improved multiple-dimension image displays
JP6147489B2 (en) Ultrasonic imaging system
US5493595A (en) Stereoscopically displayed three dimensional medical imaging
JP5688197B2 (en) 3D display of 2D ultrasound images
US10157500B2 (en) Utilizing depth from ultrasound volume rendering for 3D printing
US20150173715A1 (en) Apparatus and method for distributed ultrasound diagnostics
JP6058283B2 (en) Ultrasonic diagnostic equipment
US20060098864A1 (en) Method and apparatus to display 3d rendered ultrasound data on an ultrasound cart in stereovision
CN106456112B (en) Imaging system and method for positioning a 3D ultrasound volume in a desired orientation
US20150065877A1 (en) Method and system for generating a composite ultrasound image
JP2012252697A (en) Method and system for indicating depth of 3d cursor in volume-rendered image
JP6058282B2 (en) Medical image diagnostic apparatus and image processing apparatus
JP4855926B2 (en) Synchronizing swivel 3D ultrasonic display with vibration target
KR20140063993A (en) Apparatus and method for generating medical image
US20080024488A1 (en) Real Time Stereoscopic Imaging Apparatus and Method
KR102218308B1 (en) ultrasonic image processing apparatus and method
JPH0622964A (en) Three-dimensional display method of ultrasonic image
US20140063208A1 (en) Medical image diagnostic apparatus, image processing apparatus, and ultrasonic diagnostic apparatus
JP2009519085A (en) Fast rate for real-time 3D volume rendering images
JP4298016B2 (en) Ultrasonic diagnostic equipment
JP2004209247A (en) Method to stream three-dimensional ultrasonic volume to complete checking place for off-cart
CN117795560A (en) Visual data transmission system, display system and operation method thereof
JPH11113903A (en) Ultrasonograph

Legal Events

Date Code Title Description
AS Assignment

Owner name: KONINKLIJKE PHILIPS ELECTRONICS, N.V., NETHERLANDS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ZIEL, J. MARK;REEL/FRAME:017397/0607

Effective date: 20021119

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION