WO2011146671A1 - Computing device notes - Google Patents

Computing device notes Download PDF

Info

Publication number
WO2011146671A1
WO2011146671A1 PCT/US2011/037073 US2011037073W WO2011146671A1 WO 2011146671 A1 WO2011146671 A1 WO 2011146671A1 US 2011037073 W US2011037073 W US 2011037073W WO 2011146671 A1 WO2011146671 A1 WO 2011146671A1
Authority
WO
WIPO (PCT)
Prior art keywords
note
computing device
characters
indication
input
Prior art date
Application number
PCT/US2011/037073
Other languages
French (fr)
Inventor
Jonathan R. Harris
Andrew S. Allen
Original Assignee
Microsoft Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Microsoft Corporation filed Critical Microsoft Corporation
Priority to EP11784210.4A priority Critical patent/EP2572270A4/en
Publication of WO2011146671A1 publication Critical patent/WO2011146671A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • G06F40/169Annotation, e.g. comment data or footnotes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • G06F40/171Editing, e.g. inserting or deleting by use of digital ink

Definitions

  • inclusion of additional functions in a menu may add additional levels to the menu as well as additional choices at each of the levels.
  • inclusion of these features using traditional techniques may force the user to navigate through menus to access the features "away" from the current user interface. Consequently, the addition of these functions in the menu may frustrate users by the sheer number of choices of functions and thereby result in decreased utilization of both the additional functions as well as the device itself that employs the functions.
  • traditional techniques that were used to access the functions may limit the usefulness of the functions and the device as a whole to a user of the computing device.
  • an input is recognized as selecting at least one object displayed in a user interface by a display device of a computing device. Responsive to the recognition, an indication is displayed on the display device that is selectable. Responsive to the selection of the indication, a portion is displayed by the computing device that is configured to accept one or more inputs as a note to be associated with the at least one object.
  • an input is recognized as movement over characters displayed in a user interface by a display device of a computing device.
  • a line is displayed in the user interface that corresponds to the movement. Responsive to the recognition, selection of the characters is indicated by changing display characteristics of the characters.
  • a portion is displayed in the user interface by the computing device that is configured to accept one or more inputs as a note to be associated with at least the characters associated with the movement.
  • a computing device includes a first housing having a first display device disposed therein and a second housing physically coupled to the first housing, the second housing having a second display device disposed therein.
  • One or more modules disposed within at least one of the first or second housings, the one or more modules configured to responsive to recognition of an input selecting an objected displayed on the first display device, display an indication that is selectable proximal to the object and responsive to selection of the indication, display a portion on the second display device that is configured to accept one or more inputs as a note to be associated with the at least one object.
  • FIG. 1 is an illustration of an environment in an example implementation that is operable to employ computing device note techniques described herein.
  • FIG. 2 depicts a system in an example implementation showing selection of an object and output of an indication responsive to the selection in relation to a computing device of FIG. 1, the indication selectable to output a portion to accept one or more inputs as a note.
  • FIG. 3 depicts a system in an example implementation in which a portion is output that is configured to receive one or more inputs as a note in response to selection of the indication of FIG. 2.
  • FIG. 4 depicts a system in an example implementation in which an input is recognized as selecting multiple rows of characters displayed on a display device of the computing device of FIG. 1.
  • FIG. 5 is a flow diagram depicting a procedure in an example implementation in which computing device note techniques are performed that include output of an indication responsive to selection of an object, the indication selectable to output a portion that is configured to receive one or more inputs as a note to be associated with the object.
  • FIG. 6 illustrates various components of an example device that can be implemented as any type of portable and/or computer device as described with reference to FIGS. 1-4 to implement embodiments of the computing device note techniques described herein.
  • a user may select an object displayed by a display device of a computing device, such as by highlighting a section of text or other object using a stylus. Responsive to this selection, an indication is output that is selectable to cause a portion to be displayed that is configured to work as a note. A user may then enter text (e.g., handwritten using the stylus) or other data into the portion, which is associated with the text, such as for display with the text, display when the text is selected, and so on. Thus, through automatic display both the indication and the note may be accessed without navigating through one or more menus to locate the function.
  • text e.g., handwritten using the stylus
  • the indication may "time out” so as not to obstruct the user's view of an underlying user interface.
  • Example illustrations of the techniques and procedures are then described, which may be employed in the example environment as well as in other environments. Accordingly, the example environment is not limited to performing the example techniques and procedures. Likewise, the example techniques and procedures are not limited to implementation in the example environment.
  • FIG. 1 is an illustration of an environment 100 in an example implementation that is operable to employ computing device note techniques.
  • the illustrated environment 100 includes an example of a computing device 102 that may be configured in a variety of ways.
  • the computing device 102 may be configured as a traditional computer (e.g., a desktop personal computer, laptop computer, and so on), a mobile station, an entertainment appliance, a set-top box communicatively coupled to a television, a wireless phone, a netbook, a game console, and so forth.
  • the computing device 102 may range from full resource devices with substantial memory and processor resources (e.g., personal computers, game consoles) to a low-resource device with limited memory and/or processing resources (e.g., traditional set-top boxes, hand-held game consoles).
  • the computing device 102 may also relate to software that causes the computing device 102 to perform one or more operations.
  • the computing device 102 is illustrated as including an input module 104.
  • the input module 104 is representative of functionality relating to inputs of the computing device 102.
  • the input module 104 may be configured to receive inputs from a keyboard, mouse, to identify gestures and cause operations to be performed that correspond to the gestures, and so on.
  • the inputs may be identified by the input module 104 in a variety of different ways.
  • the input module 104 may be configured to recognize an input received via touchscreen functionality of a display device 106, such as a finger of a user's hand 108 as proximal to the display device 106 of the computing device 102, from a stylus 110, and so on.
  • the input may take a variety of different forms, such as to recognize movement of the stylus 110 and/or a finger of the user's hand 108 across the display device 106, such as a tap, drawing of a line, and so on. In implementations, these inputs may be recognized as gestures.
  • gestures may be recognized, such a gestures that are recognized from a single type of input (e.g., touch gestures) as well as gestures involving multiple types of inputs.
  • the computing device 102 may be configured to detect and differentiate between a touch input (e.g., provided by one or more fingers of the user's hand 108) and a stylus input (e.g., provided by a stylus 110).
  • the differentiation may be performed in a variety of ways, such as by detecting an amount of the display device 108 that is contacted by the finger of the user's hand 108 versus an amount of the display device 106 that is contacted by the stylus 110.
  • Differentiation may also be performed through use of a camera to distinguish a touch input (e.g., holding up one or more fingers) from a stylus input (e.g., holding two fingers together to indicate a point) in a natural user interface (NUI).
  • NUI natural user interface
  • the input module 104 may support a variety of different gesture techniques by recognizing and leveraging a division between stylus and touch inputs.
  • the input module 104 may be configured to recognize the stylus as a writing tool, whereas touch is employed to manipulate objects displayed by the display device 108. Consequently, the combination of touch and stylus inputs may serve as a basis to indicate a variety of different gestures. For instance, primitives of touch (e.g., tap, hold, two-fmger hold, grab, cross, pinch, hand or finger postures, and so on) and stylus (e.g., tap, hold-and-drag-off, drag-into, cross, stroke) may be composed to create a space involving a plurality of gestures.
  • primitives of touch e.g., tap, hold, two-fmger hold, grab, cross, pinch, hand or finger postures, and so on
  • stylus e.g., tap, hold-and-drag-off, drag-into, cross, stroke
  • the computing device 102 in this example is further illustrated as assuming a dual screen configuration.
  • the computing device 102 includes the first display device 106, which is disposed in a first housing 112.
  • the computing device 102 also has a second display device 114 disposed in a second housing 116.
  • a binding system 1 18 is also included (a rotatable mechanism such as a hinge, pivot, and so on), which movably connects the first housing 112 to the second housing 114.
  • the binding system 118 can be implemented as a multi-axis hinge that movably connects the first housing 112 to rotate in at least two different directions relative to the second housing 114.
  • the binding system 118 can include an internal drive mechanism to position the first housing 112 and/or the second housing 114 in one or more various display postures, the example of which is a "flat" posture such that the first and second housings 112, 1 14 may be positioned when laid on a surface.
  • the binding system 118 may also include passive mechanical positioned detents to retain the two halves in various relative postures as further described below.
  • the computing device 102 is further illustrated as including a note taking module 120.
  • the note taking module 120 is representative of functionality of the computing device 102 relating to computing device note techniques.
  • the note taking module 120 may be configured to output a portion that is configured to accept one or more notes (e.g., comments or other data) responsive to selection of an object that is displayed in a display device of the computing device 102.
  • the note functionality may be accessed without navigating through menus, input of one or more key combinations, and so on, further discussion of which may be found in relation to the following computing device note implementation example.
  • any of the functions described herein can be implemented using software, firmware, hardware (e.g., fixed logic circuitry), or a combination of these implementations.
  • the terms “module,” “functionality,” and “logic” as used herein generally represent software, firmware, hardware, or a combination thereof.
  • the module, functionality, or logic represents program code that performs specified tasks when executed on a processor (e.g., CPU or CPUs).
  • the program code can be stored in one or more computer readable memory devices.
  • FIG. 2 depicts a system 200 in an example implementation showing selection of an object and output of an indication responsive to the selection, the indication selectable to output a portion to accepts one or more inputs as a note.
  • the system 200 of FIG. 2 is illustrated as including first and second stages 202, 204.
  • highlighter functionality has been selected for the stylus 110 to highlight objects, although other functionality is also contemplated, e.g., such as to underline objects, circle objects, and so on.
  • the first display device 106 is illustrated as outputting a Gaelic book.
  • the stylus 110 is illustrated as having moved across a surface of the first display device 106 to select characters of the book. This movement is also illustrated as causing display of a line 206 by the display device to show the movement, e.g., the line 206 may be drawn in real time to follow the movement of the stylus 110.
  • a stylus 110 is shown, it should be readily apparent that a variety of other inputs may be recognized (e.g., by touchscreen functionality of the computing device 102), such as a touch input from a user's hand 108 as previously described.
  • Text 208 that corresponds to the movement is also indicated as selected through holding of the text.
  • the computing device 102 may employ functionality to determine what is to be included in the object selected. For example, the line 206 illustrated in the first step 202 is shown as selecting a letter "e” in the Gaelic word “caillte.” However, responsive to this selection the computing device (through the note taking module 120) may determine that the word "caillte” is to be selected as part of the text 208, even though other letters of the word are not disposed "beneath" the line 206 drawn by the stylus 110.
  • an indication 210 is illustrated as being disposed proximal to the selected object, e.g., the text 208 in this instance.
  • the indication 210 is configured by the note taking module 120 to "time out" after a predetermined period of time if the indication 210 is not selected during that time. In this way, the indication 210 may provide an option to input a note without cluttering a user interface for an extended period of time. Further, this indication 210 may be output automatically and without further user intervention (e.g., past selection of the object) such that a user is not forced to navigate away from the current user experience. Selection of the indication 210 may then cause the user interface to be configured to provide a portion for a note, further discussion of which may be found in relation to the following figure.
  • FIG. 3 depicts a system 300 in an example implementation in which a portion is output in response to selection of the indication 210 of FIG. 2.
  • the system 300 of FIG. 3 is also illustrated as including first and second stages 302, 304.
  • a portion 302 is output on the second display device 114 of the second housing 116 responsive to selection of the indication 210 displayed in the first display device 106 of the first housing 112.
  • the indication 210 may be selected in a variety of ways, such as by the stylus 110, a finger of the user's hand 108, and so on.
  • selection of the indication 210 causes further display of the portion to cease, as shown in the second stage 304 of FIG. 3.
  • continued display of the indication 210 is also contemplated, such as to select subsequent output of the portion 302.
  • the portion 302 is configured to receive one or more inputs. As illustrated in the second stage 304 of FIG. 3, for instance, the stylus 110 is shown as providing handwritten freeform lines 304 that provide a translation for the selected Gaelic words.
  • a variety of other data may also be captured by the portion 302, such as an image, link, keyboard inputs, drawings, and so on.
  • the portion 302 also includes a section 306 (e.g., illustrated as a button saying "Done” in the figure) that is selectable to indicate that entry of inputs is complete.
  • the portion 302 may then be saved as a note that is associated with the selected object.
  • the note may be displayed in response to a variety of different situations. For example, display of the text 208 may automatically cause display of the note. In another example, the text 208 and/or the indication 210 may be selectable to cause display of the note. A variety of other examples are also contemplated.
  • the portion 302 is also illustrated as having focus. This is illustrated in FIG. 3 by displaying the portion 302 over a background having phantom lines (which is readily identifiable in comparison with FIG. 2). However, it should also be apparent that focus may be provided in a variety of different ways, such as by darkening the background "beneath" the portion 302, using different levels of opacity, and so on.
  • FIG. 4 depicts a system 400 in an example implementation in which an input is recognized as selecting multiple rows of characters displayed on a display device of the computing device of FIG. 1.
  • the system 400 of FIG. 4 is illustrated as including first and second stages 402, 404.
  • the stylus 110 is shown at a first position 406 in relation to a user interface output by the computing device 102.
  • the stylus 110 is illustrated as having been moved from the first position 406 to a second position 408 in the user interface and as having a line 410 draw to indicate the movement of the stylus 1 10.
  • the movement from the first position 406 to the second position 408 is recognized by the note taking module 120 as movement that intersects three rows of characters.
  • the note taking module 120 may ascertain the movement of the stylus 1 10 as selecting those rows. As before, this selection may be indicated in a variety of ways, such as through holding the text as illustrated in the second stage 404. Thus, as previously described in relation to FIG.
  • selection of the object may be performed using logic of the note taking module 120 to include portions that are not disposed directly "beneath" the movement of an input device, e.g., the stylus 110, a finger of the user's hand 108, and so on.
  • an input device e.g., the stylus 110, a finger of the user's hand 108, and so on.
  • a dual screen computing device 102 was described in which a document to be reviewed was displayed on a first display device 102 and notes input from the user were displayed on the second display device 114. It should be apparent, however, that a wide variety of configurations (e.g., a "slate,” laptop, and so on) are also contemplated without departing from the spirit and scope thereof.
  • FIG. 5 depicts a procedure 500 in an example implementation in which computing device note techniques are performed that include output of an indication responsive to selection of an object, the indication selectable to output a portion that is configured to receive one or more inputs as a note to be associated with the object.
  • An input is recognized as selecting at least one object displayed in user interface by a display device of a computing device (block 502).
  • the object may be selected in a variety of ways. For example, highlighter functionality may be selected such that inputs received from the stylus 110 are highlighted in a user interface to mimic an ink highlighter. In another example, the object may be selected by circling the object, underlining the object, "clicking" on the object using a cursor control device, and so on.
  • an indication is displayed on the display device that is selectable (block 504).
  • the indication 210 may be displayed proximal to the selected object. Further, the indication 210 may be output automatically and without user intervention responsive to recognition of the selection. In this way, the indication may be output without having the user navigate away from the current user experience in the user interface.
  • a portion is displayed by the computing device that is configured to accept one or more inputs as a note to be associated with the at least one object (block 506).
  • the portion for instance, may be displayed proximally to the indication, and consequently the selected object. In another instance, the portion may be displayed on another screen, such as on the second display device 114 as illustrated in FIG. 3.
  • the one or more inputs are accepted in the portion (block 508).
  • the one or more inputs may take a variety of forms, such as handwritten freeform lines 304, inputs from a keyboard, images, lines, media files, and so on.
  • the one or more inputs in the portion are associated with the object (block 510).
  • the note may be displayed responsive to subsequent selection of the object (block 512).
  • the note may be displayed responsive to display of the object (block 514).
  • a variety of other examples are also contemplated.
  • FIG. 6 illustrates various components of an example device 600 that can be implemented as any type of portable and/or computer device as described with reference to FIGS. 1 and 2 to implement embodiments of the gesture techniques described herein.
  • Device 600 includes communication devices 602 that enable wired and/or wireless communication of device data 604 (e.g., received data, data that is being received, data scheduled for broadcast, data packets of the data, etc.).
  • the device data 604 or other device content can include configuration settings of the device, media content stored on the device, and/or information associated with a user of the device.
  • Media content stored on device 600 can include any type of audio, video, and/or image data.
  • Device 600 includes one or more data inputs 606 via which any type of data, media content, and/or inputs can be received, such as user-selectable inputs, messages, music, television media content, recorded video content, and any other type of audio, video, and/or image data received from any content and/or data source.
  • any type of data, media content, and/or inputs can be received, such as user-selectable inputs, messages, music, television media content, recorded video content, and any other type of audio, video, and/or image data received from any content and/or data source.
  • Device 600 also includes communication interfaces 608 that can be implemented as any one or more o ⁇ f a serial and/or parallel interface, a wireless interface, any type of network interface, a modem, and as any other type of communication interface.
  • the communication interfaces 608 provide a connection and/or communication links between device 600 and a communication network by which other electronic, computing, and communication devices communicate data with device 600.
  • Device 600 includes one or more processors 610 (e.g., any of microprocessors, controllers, and the like) which process various computer- executable instructions to control the operation of device 600 and to implement embodiments of a touch pull-in gesture.
  • processors 610 e.g., any of microprocessors, controllers, and the like
  • device 600 can be implemented with any one or combination of hardware, firmware, or fixed logic circuitry that is implemented in connection with processing and control circuits which are generally identified at 612.
  • device 600 can include a system bus or data transfer system that couples the various components within the device.
  • a system bus can include any one or combination of different bus structures, such as a memory bus or memory controller, a peripheral bus, a universal serial bus, and/or a processor or local bus that utilizes any of a variety of bus architectures.
  • Device 600 also includes computer-readable media 614, such as one or more memory components, examples of which include random access memory (RAM), non-volatile memory (e.g., any one or more of a read-only memory (ROM), flash memory, EPROM, EEPROM, etc.), and a disk storage device.
  • RAM random access memory
  • non-volatile memory e.g., any one or more of a read-only memory (ROM), flash memory, EPROM, EEPROM, etc.
  • a disk storage device may be implemented as any type of magnetic or optical storage device, such as a hard disk drive, a recordable and/or rewriteable compact disc (CD), any type of a digital versatile disc (DVD), and the like.
  • Device 600 can also include a mass storage media device 616.
  • Computer-readable media 614 provides data storage mechanisms to store the device data 604, as well as various device applications 618 and any other types of information and/or data related to operational aspects of device 600.
  • an operating system 620 can be maintained as a computer application with the computer-readable media 614 and executed on processors 610.
  • the device applications 618 can include a device manager (e.g., a control application, software application, signal processing and control module, code that is native to a particular device, a hardware abstraction layer for a particular device, etc.).
  • the device applications 618 also include any system components or modules to implement embodiments of the gesture techniques described herein.
  • the device applications 618 include an interface application 622 and an input module 624 (which may be the same or different as input module 114) that are shown as software modules and/or computer applications.
  • the input module 624 is representative of software that is used to provide an interface with a device configured to capture inputs, such as a touchscreen, track pad, camera, and so on.
  • the interface application 622 and the input module 624 can be implemented as hardware, software, firmware, or any combination thereof.
  • the input module 624 may be configured to support multiple input devices, such as separate devices to capture touch and stylus inputs, respectively.
  • the device may be configured to include dual display devices, in which one of the display device is configured to capture touch inputs while the other stylus inputs.
  • Device 600 also includes an audio and/or video input-output system 626 that provides audio data to an audio system 628 and/or provides video data to a display system 630.
  • the audio system 628 and/or the display system 630 can include any devices that process, display, and/or otherwise render audio, video, and image data.
  • Video signals and audio signals can be communicated from device 600 to an audio device and/or to a display device via an RF (radio frequency) link, S- video link, composite video link, component video link, DVI (digital video interface), analog audio connection, or other similar communication link.
  • the audio system 628 and/or the display system 630 are implemented as external components to device 600.
  • the audio system 628 and/or the display system 630 are implemented as integrated components of example device 600.

Abstract

Computing device note techniques are described. In implementations, an input is recognized as selecting at least one object displayed in a user interface by a display device of a computing device. Responsive to the recognition, an indication is displayed on the display device that is selectable. Responsive to selection of the indication, a portion is displayed by the computing device that is configured to accept one or more inputs as a note to be associated with the at least one object.

Description

Computing Device Notes
BACKGROUND
[0001] The amount of functionality that is available from computing devices is ever increasing, such as from mobile devices, game consoles, televisions, set-top boxes, personal computers, and so on. However, traditional techniques that were employed to interact with the computing devices may become less efficient as the amount of functionality increases.
[0002] For example, inclusion of additional functions in a menu may add additional levels to the menu as well as additional choices at each of the levels. Additionally, inclusion of these features using traditional techniques may force the user to navigate through menus to access the features "away" from the current user interface. Consequently, the addition of these functions in the menu may frustrate users by the sheer number of choices of functions and thereby result in decreased utilization of both the additional functions as well as the device itself that employs the functions. Thus, traditional techniques that were used to access the functions may limit the usefulness of the functions and the device as a whole to a user of the computing device.
SUMMARY
[0003] Computing device note techniques are described. In implementations, an input is recognized as selecting at least one object displayed in a user interface by a display device of a computing device. Responsive to the recognition, an indication is displayed on the display device that is selectable. Responsive to the selection of the indication, a portion is displayed by the computing device that is configured to accept one or more inputs as a note to be associated with the at least one object.
[0004] In implementations, an input is recognized as movement over characters displayed in a user interface by a display device of a computing device. A line is displayed in the user interface that corresponds to the movement. Responsive to the recognition, selection of the characters is indicated by changing display characteristics of the characters. A portion is displayed in the user interface by the computing device that is configured to accept one or more inputs as a note to be associated with at least the characters associated with the movement. [0005] In implementations, a computing device includes a first housing having a first display device disposed therein and a second housing physically coupled to the first housing, the second housing having a second display device disposed therein. One or more modules disposed within at least one of the first or second housings, the one or more modules configured to responsive to recognition of an input selecting an objected displayed on the first display device, display an indication that is selectable proximal to the object and responsive to selection of the indication, display a portion on the second display device that is configured to accept one or more inputs as a note to be associated with the at least one object.
[0006] This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
BRIEF DESCRIPTION OF THE DRAWINGS
[0007] The detailed description is described with reference to the accompanying figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The use of the same reference numbers in different instances in the description and the figures may indicate similar or identical items.
[0008] FIG. 1 is an illustration of an environment in an example implementation that is operable to employ computing device note techniques described herein.
[0009] FIG. 2 depicts a system in an example implementation showing selection of an object and output of an indication responsive to the selection in relation to a computing device of FIG. 1, the indication selectable to output a portion to accept one or more inputs as a note.
[0010] FIG. 3 depicts a system in an example implementation in which a portion is output that is configured to receive one or more inputs as a note in response to selection of the indication of FIG. 2. [ooii] FIG. 4 depicts a system in an example implementation in which an input is recognized as selecting multiple rows of characters displayed on a display device of the computing device of FIG. 1.
[0012] FIG. 5 is a flow diagram depicting a procedure in an example implementation in which computing device note techniques are performed that include output of an indication responsive to selection of an object, the indication selectable to output a portion that is configured to receive one or more inputs as a note to be associated with the object.
[0013] FIG. 6 illustrates various components of an example device that can be implemented as any type of portable and/or computer device as described with reference to FIGS. 1-4 to implement embodiments of the computing device note techniques described herein.
DETAILED DESCRIPTION
Overview
[0014] Traditional functionality of a computing device was often accessed using a menu, which may contain a variety of hierarchical levels as well as a variety of choices at each of the levels. Accordingly, access to these functions may interrupt a user's interaction with the computing device.
[0015] Computing device note techniques are described. In implementations, a user may select an object displayed by a display device of a computing device, such as by highlighting a section of text or other object using a stylus. Responsive to this selection, an indication is output that is selectable to cause a portion to be displayed that is configured to work as a note. A user may then enter text (e.g., handwritten using the stylus) or other data into the portion, which is associated with the text, such as for display with the text, display when the text is selected, and so on. Thus, through automatic display both the indication and the note may be accessed without navigating through one or more menus to locate the function. Further, the indication may "time out" so as not to obstruct the user's view of an underlying user interface. Further discussion of this and other examples of computing device notes may be found in relation to the following sections. [0016] In the following discussion, an example environment is first described that is operable to employ the computing device note techniques described herein. Example illustrations of the techniques and procedures are then described, which may be employed in the example environment as well as in other environments. Accordingly, the example environment is not limited to performing the example techniques and procedures. Likewise, the example techniques and procedures are not limited to implementation in the example environment.
Example Environment
[0017] FIG. 1 is an illustration of an environment 100 in an example implementation that is operable to employ computing device note techniques. The illustrated environment 100 includes an example of a computing device 102 that may be configured in a variety of ways. For example, the computing device 102 may be configured as a traditional computer (e.g., a desktop personal computer, laptop computer, and so on), a mobile station, an entertainment appliance, a set-top box communicatively coupled to a television, a wireless phone, a netbook, a game console, and so forth. Thus, the computing device 102 may range from full resource devices with substantial memory and processor resources (e.g., personal computers, game consoles) to a low-resource device with limited memory and/or processing resources (e.g., traditional set-top boxes, hand-held game consoles). The computing device 102 may also relate to software that causes the computing device 102 to perform one or more operations.
[0018] The computing device 102 is illustrated as including an input module 104. The input module 104 is representative of functionality relating to inputs of the computing device 102. For example, the input module 104 may be configured to receive inputs from a keyboard, mouse, to identify gestures and cause operations to be performed that correspond to the gestures, and so on. The inputs may be identified by the input module 104 in a variety of different ways.
[0019] For example, the input module 104 may be configured to recognize an input received via touchscreen functionality of a display device 106, such as a finger of a user's hand 108 as proximal to the display device 106 of the computing device 102, from a stylus 110, and so on. The input may take a variety of different forms, such as to recognize movement of the stylus 110 and/or a finger of the user's hand 108 across the display device 106, such as a tap, drawing of a line, and so on. In implementations, these inputs may be recognized as gestures.
[0020] A variety of different types of gestures may be recognized, such a gestures that are recognized from a single type of input (e.g., touch gestures) as well as gestures involving multiple types of inputs. For example, the computing device 102 may be configured to detect and differentiate between a touch input (e.g., provided by one or more fingers of the user's hand 108) and a stylus input (e.g., provided by a stylus 110). The differentiation may be performed in a variety of ways, such as by detecting an amount of the display device 108 that is contacted by the finger of the user's hand 108 versus an amount of the display device 106 that is contacted by the stylus 110. Differentiation may also be performed through use of a camera to distinguish a touch input (e.g., holding up one or more fingers) from a stylus input (e.g., holding two fingers together to indicate a point) in a natural user interface (NUI). A variety of other example techniques for distinguishing touch and stylus inputs are contemplated, further discussion of which may be found in relation to FIG. 6.
[0021] Thus, the input module 104 may support a variety of different gesture techniques by recognizing and leveraging a division between stylus and touch inputs. For instance, the input module 104 may be configured to recognize the stylus as a writing tool, whereas touch is employed to manipulate objects displayed by the display device 108. Consequently, the combination of touch and stylus inputs may serve as a basis to indicate a variety of different gestures. For instance, primitives of touch (e.g., tap, hold, two-fmger hold, grab, cross, pinch, hand or finger postures, and so on) and stylus (e.g., tap, hold-and-drag-off, drag-into, cross, stroke) may be composed to create a space involving a plurality of gestures. It should be noted that by differentiating between stylus and touch inputs, the number of gestures that are made possible by each of these inputs alone is also increased. For example, although the movements may be the same, different gestures (or different parameters to analogous commands) may be indicated using touch inputs versus stylus inputs. [0022] Additionally, although the following discussion may describe specific examples of touch and stylus inputs, in instances the types of inputs may be switched (e.g., touch may be used to replace stylus and vice versa) and even removed (e.g., both inputs may be provided using touch or a stylus) without departing from the spirit and scope thereof. Further, although in instances in the following discussion the gestures are illustrated as being input using touchscreen functionality, the gestures may be input using a variety of different techniques by a variety of different devices.
[0023] The computing device 102 in this example is further illustrated as assuming a dual screen configuration. In the illustrated example, the computing device 102 includes the first display device 106, which is disposed in a first housing 112. The computing device 102 also has a second display device 114 disposed in a second housing 116. A binding system 1 18 is also included (a rotatable mechanism such as a hinge, pivot, and so on), which movably connects the first housing 112 to the second housing 114. In an embodiment, the binding system 118 can be implemented as a multi-axis hinge that movably connects the first housing 112 to rotate in at least two different directions relative to the second housing 114. Alternatively or in addition, the binding system 118 can include an internal drive mechanism to position the first housing 112 and/or the second housing 114 in one or more various display postures, the example of which is a "flat" posture such that the first and second housings 112, 1 14 may be positioned when laid on a surface. Alternatively or in addition, the binding system 118 may also include passive mechanical positioned detents to retain the two halves in various relative postures as further described below.
[0024] The computing device 102 is further illustrated as including a note taking module 120. The note taking module 120 is representative of functionality of the computing device 102 relating to computing device note techniques. For example, the note taking module 120 may be configured to output a portion that is configured to accept one or more notes (e.g., comments or other data) responsive to selection of an object that is displayed in a display device of the computing device 102. In this way, the note functionality may be accessed without navigating through menus, input of one or more key combinations, and so on, further discussion of which may be found in relation to the following computing device note implementation example.
[0025] Generally, any of the functions described herein can be implemented using software, firmware, hardware (e.g., fixed logic circuitry), or a combination of these implementations. The terms "module," "functionality," and "logic" as used herein generally represent software, firmware, hardware, or a combination thereof. In the case of a software implementation, the module, functionality, or logic represents program code that performs specified tasks when executed on a processor (e.g., CPU or CPUs). The program code can be stored in one or more computer readable memory devices. The features of the computing device note techniques described below are platform-independent, meaning that the techniques may be implemented on a variety of commercial computing platforms having a variety of processors.
Computing Device Note Implementation Example
[0026] FIG. 2 depicts a system 200 in an example implementation showing selection of an object and output of an indication responsive to the selection, the indication selectable to output a portion to accepts one or more inputs as a note. The system 200 of FIG. 2 is illustrated as including first and second stages 202, 204. In these examples, highlighter functionality has been selected for the stylus 110 to highlight objects, although other functionality is also contemplated, e.g., such as to underline objects, circle objects, and so on.
[0027] At the first stage 202, the first display device 106 is illustrated as outputting a Gaelic book. The stylus 110 is illustrated as having moved across a surface of the first display device 106 to select characters of the book. This movement is also illustrated as causing display of a line 206 by the display device to show the movement, e.g., the line 206 may be drawn in real time to follow the movement of the stylus 110. Although a stylus 110 is shown, it should be readily apparent that a variety of other inputs may be recognized (e.g., by touchscreen functionality of the computing device 102), such as a touch input from a user's hand 108 as previously described. [0028] Text 208 that corresponds to the movement is also indicated as selected through holding of the text. In implementations, the computing device 102 may employ functionality to determine what is to be included in the object selected. For example, the line 206 illustrated in the first step 202 is shown as selecting a letter "e" in the Gaelic word "caillte." However, responsive to this selection the computing device (through the note taking module 120) may determine that the word "caillte" is to be selected as part of the text 208, even though other letters of the word are not disposed "beneath" the line 206 drawn by the stylus 110.
[0029] At the second stage 204, an indication 210 is illustrated as being disposed proximal to the selected object, e.g., the text 208 in this instance. In implementations, the indication 210 is configured by the note taking module 120 to "time out" after a predetermined period of time if the indication 210 is not selected during that time. In this way, the indication 210 may provide an option to input a note without cluttering a user interface for an extended period of time. Further, this indication 210 may be output automatically and without further user intervention (e.g., past selection of the object) such that a user is not forced to navigate away from the current user experience. Selection of the indication 210 may then cause the user interface to be configured to provide a portion for a note, further discussion of which may be found in relation to the following figure.
[0030] FIG. 3 depicts a system 300 in an example implementation in which a portion is output in response to selection of the indication 210 of FIG. 2. The system 300 of FIG. 3 is also illustrated as including first and second stages 302, 304. At the first stage 302, a portion 302 is output on the second display device 114 of the second housing 116 responsive to selection of the indication 210 displayed in the first display device 106 of the first housing 112. The indication 210 may be selected in a variety of ways, such as by the stylus 110, a finger of the user's hand 108, and so on.
[0031] In implementations, selection of the indication 210 causes further display of the portion to cease, as shown in the second stage 304 of FIG. 3. However, it should be readily apparent that continued display of the indication 210 is also contemplated, such as to select subsequent output of the portion 302. [0032] The portion 302 is configured to receive one or more inputs. As illustrated in the second stage 304 of FIG. 3, for instance, the stylus 110 is shown as providing handwritten freeform lines 304 that provide a translation for the selected Gaelic words. A variety of other data may also be captured by the portion 302, such as an image, link, keyboard inputs, drawings, and so on. The portion 302 also includes a section 306 (e.g., illustrated as a button saying "Done" in the figure) that is selectable to indicate that entry of inputs is complete. The portion 302 may then be saved as a note that is associated with the selected object.
[0033] Once saved, the note may be displayed in response to a variety of different situations. For example, display of the text 208 may automatically cause display of the note. In another example, the text 208 and/or the indication 210 may be selectable to cause display of the note. A variety of other examples are also contemplated.
[0034] The portion 302 is also illustrated as having focus. This is illustrated in FIG. 3 by displaying the portion 302 over a background having phantom lines (which is readily identifiable in comparison with FIG. 2). However, it should also be apparent that focus may be provided in a variety of different ways, such as by darkening the background "beneath" the portion 302, using different levels of opacity, and so on.
[0035] FIG. 4 depicts a system 400 in an example implementation in which an input is recognized as selecting multiple rows of characters displayed on a display device of the computing device of FIG. 1. The system 400 of FIG. 4 is illustrated as including first and second stages 402, 404. At the first stage 402, the stylus 110 is shown at a first position 406 in relation to a user interface output by the computing device 102.
[0036] At the second stage 404, the stylus 110 is illustrated as having been moved from the first position 406 to a second position 408 in the user interface and as having a line 410 draw to indicate the movement of the stylus 1 10. In this implementation, the movement from the first position 406 to the second position 408 is recognized by the note taking module 120 as movement that intersects three rows of characters. [0037] Accordingly, the note taking module 120 may ascertain the movement of the stylus 1 10 as selecting those rows. As before, this selection may be indicated in a variety of ways, such as through holding the text as illustrated in the second stage 404. Thus, as previously described in relation to FIG. 2 selection of the object may be performed using logic of the note taking module 120 to include portions that are not disposed directly "beneath" the movement of an input device, e.g., the stylus 110, a finger of the user's hand 108, and so on.
[0038] In the examples systems 200-400 of FIGS. 2-4, a dual screen computing device 102 was described in which a document to be reviewed was displayed on a first display device 102 and notes input from the user were displayed on the second display device 114. It should be apparent, however, that a wide variety of configurations (e.g., a "slate," laptop, and so on) are also contemplated without departing from the spirit and scope thereof.
Example Procedures
[0039] The following discussion describes computing device note techniques that may be implemented utilizing the previously described systems and devices. Aspects of each of the procedures may be implemented in hardware, firmware, software, or a combination thereof. The procedures are shown as a set of blocks that specify operations performed by one or more devices and are not necessarily limited to the orders shown for performing the operations by the respective blocks. In portions of the following discussion, reference will be made to the environment 100 of FIG. 1 and the systems 200-400 of FIGS. 2-4.
[0040] FIG. 5 depicts a procedure 500 in an example implementation in which computing device note techniques are performed that include output of an indication responsive to selection of an object, the indication selectable to output a portion that is configured to receive one or more inputs as a note to be associated with the object. An input is recognized as selecting at least one object displayed in user interface by a display device of a computing device (block 502). The object may be selected in a variety of ways. For example, highlighter functionality may be selected such that inputs received from the stylus 110 are highlighted in a user interface to mimic an ink highlighter. In another example, the object may be selected by circling the object, underlining the object, "clicking" on the object using a cursor control device, and so on.
[0041] Responsive to the recognition, an indication is displayed on the display device that is selectable (block 504). The indication 210, for instance, may be displayed proximal to the selected object. Further, the indication 210 may be output automatically and without user intervention responsive to recognition of the selection. In this way, the indication may be output without having the user navigate away from the current user experience in the user interface.
[0042] Responsive to selection of the indication, a portion is displayed by the computing device that is configured to accept one or more inputs as a note to be associated with the at least one object (block 506). The portion, for instance, may be displayed proximally to the indication, and consequently the selected object. In another instance, the portion may be displayed on another screen, such as on the second display device 114 as illustrated in FIG. 3.
[0043] The one or more inputs are accepted in the portion (block 508). As previously stated, the one or more inputs may take a variety of forms, such as handwritten freeform lines 304, inputs from a keyboard, images, lines, media files, and so on.
[0044] The one or more inputs in the portion are associated with the object (block 510). For example, the note may be displayed responsive to subsequent selection of the object (block 512). In another example the note may be displayed responsive to display of the object (block 514). A variety of other examples are also contemplated.
Example Device
[0045] FIG. 6 illustrates various components of an example device 600 that can be implemented as any type of portable and/or computer device as described with reference to FIGS. 1 and 2 to implement embodiments of the gesture techniques described herein. Device 600 includes communication devices 602 that enable wired and/or wireless communication of device data 604 (e.g., received data, data that is being received, data scheduled for broadcast, data packets of the data, etc.). The device data 604 or other device content can include configuration settings of the device, media content stored on the device, and/or information associated with a user of the device. Media content stored on device 600 can include any type of audio, video, and/or image data. Device 600 includes one or more data inputs 606 via which any type of data, media content, and/or inputs can be received, such as user-selectable inputs, messages, music, television media content, recorded video content, and any other type of audio, video, and/or image data received from any content and/or data source.
[0046] Device 600 also includes communication interfaces 608 that can be implemented as any one or more o\f a serial and/or parallel interface, a wireless interface, any type of network interface, a modem, and as any other type of communication interface. The communication interfaces 608 provide a connection and/or communication links between device 600 and a communication network by which other electronic, computing, and communication devices communicate data with device 600.
[0047] Device 600 includes one or more processors 610 (e.g., any of microprocessors, controllers, and the like) which process various computer- executable instructions to control the operation of device 600 and to implement embodiments of a touch pull-in gesture. Alternatively or in addition, device 600 can be implemented with any one or combination of hardware, firmware, or fixed logic circuitry that is implemented in connection with processing and control circuits which are generally identified at 612. Although not shown, device 600 can include a system bus or data transfer system that couples the various components within the device. A system bus can include any one or combination of different bus structures, such as a memory bus or memory controller, a peripheral bus, a universal serial bus, and/or a processor or local bus that utilizes any of a variety of bus architectures.
[0048] Device 600 also includes computer-readable media 614, such as one or more memory components, examples of which include random access memory (RAM), non-volatile memory (e.g., any one or more of a read-only memory (ROM), flash memory, EPROM, EEPROM, etc.), and a disk storage device. A disk storage device may be implemented as any type of magnetic or optical storage device, such as a hard disk drive, a recordable and/or rewriteable compact disc (CD), any type of a digital versatile disc (DVD), and the like. Device 600 can also include a mass storage media device 616.
[0049] Computer-readable media 614 provides data storage mechanisms to store the device data 604, as well as various device applications 618 and any other types of information and/or data related to operational aspects of device 600. For example, an operating system 620 can be maintained as a computer application with the computer-readable media 614 and executed on processors 610. The device applications 618 can include a device manager (e.g., a control application, software application, signal processing and control module, code that is native to a particular device, a hardware abstraction layer for a particular device, etc.). The device applications 618 also include any system components or modules to implement embodiments of the gesture techniques described herein. In this example, the device applications 618 include an interface application 622 and an input module 624 (which may be the same or different as input module 114) that are shown as software modules and/or computer applications. The input module 624 is representative of software that is used to provide an interface with a device configured to capture inputs, such as a touchscreen, track pad, camera, and so on. Alternatively or in addition, the interface application 622 and the input module 624 can be implemented as hardware, software, firmware, or any combination thereof. Additionally, the input module 624 may be configured to support multiple input devices, such as separate devices to capture touch and stylus inputs, respectively. For example, the device may be configured to include dual display devices, in which one of the display device is configured to capture touch inputs while the other stylus inputs.
[0050] Device 600 also includes an audio and/or video input-output system 626 that provides audio data to an audio system 628 and/or provides video data to a display system 630. The audio system 628 and/or the display system 630 can include any devices that process, display, and/or otherwise render audio, video, and image data. Video signals and audio signals can be communicated from device 600 to an audio device and/or to a display device via an RF (radio frequency) link, S- video link, composite video link, component video link, DVI (digital video interface), analog audio connection, or other similar communication link. In an embodiment, the audio system 628 and/or the display system 630 are implemented as external components to device 600. Alternatively, the audio system 628 and/or the display system 630 are implemented as integrated components of example device 600.
Conclusion
[0051] Although the invention has been described in language specific to structural features and/or methodological acts, it is to be understood that the invention defined in the appended claims is not necessarily limited to the specific features or acts described. Rather, the specific features and acts are disclosed as example forms of implementing the claimed invention.

Claims

CLAIMS What is claimed
1. A method comprising:
recognizing an input as selecting at least one object displayed in a user interface by a display device of a computing device;
responsive to the recognizing, displaying an indication on the display device that is selectable; and
responsive to selection of the indication, displaying a portion by the computing device that is configured to accept one or more inputs as a note to be associated with the at least one object.
2. A method as described in claim 1, wherein the input is recognized as highlighting the object.
3. A method as described in claim 2, wherein the object includes one or more characters.
4. A method as described in claim 1, wherein the input is recognized as a touch input or a stylus input detected using touchscreen functionality.
5. A method as described in claim 1, wherein the object includes a plurality of characters arranged as a word and the recognizing includes:
detecting that the input describes a line drawn over one of the characters of the word but not another one of the characters of the word; and
ascertaining the selection of the word from the detecting.
6. A method as described in claim 1, wherein the object includes a plurality of rows of characters and the recognizing includes:
detecting that the input describes a line drawn at least one of the characters in a particular said row but not another one of the characters of the of the particular said row; and
ascertaining the selection of the particular said row from the detecting.
7. A method as described in claim 1, wherein the at least one object is displayed in on the display device disposed in a first housing of the display device and the portion is displayed in a second display device disposed in a second housing of the computing device.
8. A method as described in claim 1, wherein the displaying of the indication ceases after a predetermined period of time if the indication is not selected during the predetermined period of time.
9. A method as described in claim 1, wherein the note is associated with the at least one object such that subsequent selection of the at least one object causes display of the note.
10. A method as described in claim 1, wherein the note is associated with the at least one object such that display of the at least object causes display of the note.
11. A method as described in claim 1 , wherein the one or more inputs of the note are handwritten.
12. A method comprising:
recognizing an input as movement over characters displayed in a user interface by a display device of a computing device;
displaying a line in the user interface that corresponds to the movement; responsive to the recognizing, indicating selection of the characters by changing display characteristics of the characters; and
displaying a portion in the user interface by the computing device that is configured to accept one or more inputs as a note to be associated with at least the characters associated with the movement.
13. A method as described in claim 12, wherein the indicating is performed for at least one additional character of a word that includes at least one said character associated with the movement.
14. A method as described in claim 12, further comprising responsive to the recognizing, displaying an indication on the display device that is selectable to cause the displaying of the portion.
15. A method as described in claim 12, wherein the indicating is performed for at least one additional character of a row that includes at least one said character associated with the movement.
PCT/US2011/037073 2010-05-21 2011-05-19 Computing device notes WO2011146671A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP11784210.4A EP2572270A4 (en) 2010-05-21 2011-05-19 Computing device notes

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US12/784,803 US8791900B2 (en) 2010-05-21 2010-05-21 Computing device notes
US12/784,803 2010-05-21

Publications (1)

Publication Number Publication Date
WO2011146671A1 true WO2011146671A1 (en) 2011-11-24

Family

ID=44972111

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2011/037073 WO2011146671A1 (en) 2010-05-21 2011-05-19 Computing device notes

Country Status (3)

Country Link
US (1) US8791900B2 (en)
EP (1) EP2572270A4 (en)
WO (1) WO2011146671A1 (en)

Families Citing this family (61)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9400390B2 (en) 2014-01-24 2016-07-26 Osterhout Group, Inc. Peripheral lighting for head worn computing
US9229233B2 (en) 2014-02-11 2016-01-05 Osterhout Group, Inc. Micro Doppler presentations in head worn computing
US9952664B2 (en) 2014-01-21 2018-04-24 Osterhout Group, Inc. Eye imaging in head worn computing
US9965681B2 (en) 2008-12-16 2018-05-08 Osterhout Group, Inc. Eye imaging in head worn computing
US20150205111A1 (en) 2014-01-21 2015-07-23 Osterhout Group, Inc. Optical configurations for head worn computing
US9298007B2 (en) 2014-01-21 2016-03-29 Osterhout Group, Inc. Eye imaging in head worn computing
US9715112B2 (en) 2014-01-21 2017-07-25 Osterhout Group, Inc. Suppression of stray light in head worn computing
JP5284524B1 (en) * 2012-09-07 2013-09-11 株式会社東芝 Electronic device and handwritten document processing method
US9575321B2 (en) 2014-06-09 2017-02-21 Osterhout Group, Inc. Content presentation in head worn computing
US9829707B2 (en) 2014-08-12 2017-11-28 Osterhout Group, Inc. Measuring content brightness in head worn computing
US10254856B2 (en) * 2014-01-17 2019-04-09 Osterhout Group, Inc. External user interface for head worn computing
US9841599B2 (en) 2014-06-05 2017-12-12 Osterhout Group, Inc. Optical configurations for head-worn see-through displays
US9529195B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. See-through computer display systems
US20150277118A1 (en) 2014-03-28 2015-10-01 Osterhout Group, Inc. Sensor dependent content position in head worn computing
US9299194B2 (en) 2014-02-14 2016-03-29 Osterhout Group, Inc. Secure sharing in head worn computing
US9810906B2 (en) 2014-06-17 2017-11-07 Osterhout Group, Inc. External user interface for head worn computing
US11103122B2 (en) 2014-07-15 2021-08-31 Mentor Acquisition One, Llc Content presentation in head worn computing
US9671613B2 (en) 2014-09-26 2017-06-06 Osterhout Group, Inc. See-through computer display systems
US9594246B2 (en) 2014-01-21 2017-03-14 Osterhout Group, Inc. See-through computer display systems
US20160019715A1 (en) 2014-07-15 2016-01-21 Osterhout Group, Inc. Content presentation in head worn computing
US10649220B2 (en) 2014-06-09 2020-05-12 Mentor Acquisition One, Llc Content presentation in head worn computing
US9939934B2 (en) 2014-01-17 2018-04-10 Osterhout Group, Inc. External user interface for head worn computing
US9746686B2 (en) 2014-05-19 2017-08-29 Osterhout Group, Inc. Content position calibration in head worn computing
US9448409B2 (en) 2014-11-26 2016-09-20 Osterhout Group, Inc. See-through computer display systems
US10684687B2 (en) 2014-12-03 2020-06-16 Mentor Acquisition One, Llc See-through computer display systems
US11227294B2 (en) 2014-04-03 2022-01-18 Mentor Acquisition One, Llc Sight information collection in head worn computing
US10191279B2 (en) 2014-03-17 2019-01-29 Osterhout Group, Inc. Eye imaging in head worn computing
US9651788B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9766463B2 (en) 2014-01-21 2017-09-19 Osterhout Group, Inc. See-through computer display systems
US9811159B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US11737666B2 (en) 2014-01-21 2023-08-29 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11669163B2 (en) 2014-01-21 2023-06-06 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US9494800B2 (en) 2014-01-21 2016-11-15 Osterhout Group, Inc. See-through computer display systems
US9532714B2 (en) 2014-01-21 2017-01-03 Osterhout Group, Inc. Eye imaging in head worn computing
US9836122B2 (en) 2014-01-21 2017-12-05 Osterhout Group, Inc. Eye glint imaging in see-through computer display systems
US11487110B2 (en) 2014-01-21 2022-11-01 Mentor Acquisition One, Llc Eye imaging in head worn computing
US9651784B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9753288B2 (en) 2014-01-21 2017-09-05 Osterhout Group, Inc. See-through computer display systems
US20150205135A1 (en) 2014-01-21 2015-07-23 Osterhout Group, Inc. See-through computer display systems
US11892644B2 (en) 2014-01-21 2024-02-06 Mentor Acquisition One, Llc See-through computer display systems
US9846308B2 (en) 2014-01-24 2017-12-19 Osterhout Group, Inc. Haptic systems for head-worn computers
US9401540B2 (en) 2014-02-11 2016-07-26 Osterhout Group, Inc. Spatial location presentation in head worn computing
US20150241963A1 (en) 2014-02-11 2015-08-27 Osterhout Group, Inc. Eye imaging in head worn computing
US20160187651A1 (en) 2014-03-28 2016-06-30 Osterhout Group, Inc. Safety for a vehicle operator with an hmd
US10853589B2 (en) 2014-04-25 2020-12-01 Mentor Acquisition One, Llc Language translation with head-worn computing
US9672210B2 (en) 2014-04-25 2017-06-06 Osterhout Group, Inc. Language translation with head-worn computing
US9423842B2 (en) 2014-09-18 2016-08-23 Osterhout Group, Inc. Thermal management for head-worn computer
US9651787B2 (en) 2014-04-25 2017-05-16 Osterhout Group, Inc. Speaker assembly for headworn computer
JP2015215680A (en) * 2014-05-08 2015-12-03 ブラザー工業株式会社 Electronic writing device
US10663740B2 (en) 2014-06-09 2020-05-26 Mentor Acquisition One, Llc Content presentation in head worn computing
US9684172B2 (en) 2014-12-03 2017-06-20 Osterhout Group, Inc. Head worn computer display systems
USD751552S1 (en) 2014-12-31 2016-03-15 Osterhout Group, Inc. Computer glasses
USD753114S1 (en) 2015-01-05 2016-04-05 Osterhout Group, Inc. Air mouse
US20160239985A1 (en) 2015-02-17 2016-08-18 Osterhout Group, Inc. See-through computer display systems
US10139966B2 (en) 2015-07-22 2018-11-27 Osterhout Group, Inc. External user interface for head worn computing
US11003246B2 (en) 2015-07-22 2021-05-11 Mentor Acquisition One, Llc External user interface for head worn computing
US10684478B2 (en) 2016-05-09 2020-06-16 Mentor Acquisition One, Llc User interface systems for head-worn computers
US10466491B2 (en) 2016-06-01 2019-11-05 Mentor Acquisition One, Llc Modular systems for head-worn computers
US10824253B2 (en) 2016-05-09 2020-11-03 Mentor Acquisition One, Llc User interface systems for head-worn computers
US10152141B1 (en) 2017-08-18 2018-12-11 Osterhout Group, Inc. Controller movement tracking with light emitters
US10423321B2 (en) * 2017-11-02 2019-09-24 Dell Products L. P. Defining a zone to perform an action in a dual-screen tablet

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050076300A1 (en) * 2003-10-02 2005-04-07 International Business Machines Corporation Block marker system
KR20090042740A (en) * 2007-10-26 2009-04-30 리서치 인 모션 리미티드 Text selection using a touch sensitive screen of a handheld mobile communication device
US20090228842A1 (en) * 2008-03-04 2009-09-10 Apple Inc. Selecting of text using gestures
KR20090102135A (en) * 2008-03-25 2009-09-30 건국대학교 산학협력단 Installtion for management of dictionary information in electric device and method thereof

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5398310A (en) 1992-04-13 1995-03-14 Apple Computer, Incorporated Pointing gesture based computer note pad paging and scrolling interface
US5592607A (en) 1993-10-15 1997-01-07 Xerox Corporation Interactive method and system for producing address-correlated information using user-specified address zones
US5852436A (en) 1994-06-30 1998-12-22 Microsoft Corporation Notes facility for receiving notes while the computer system is in a screen mode
US6970266B2 (en) 2001-01-23 2005-11-29 Xerox Corporation Fault notes user interface for a printing system
US7823072B2 (en) 2001-12-11 2010-10-26 International Business Machines Corporation Method and system for controlling multiple electronic mail messages in a data processing system
US20030204567A1 (en) 2002-04-26 2003-10-30 International Business Machines Corporation System and method for creating e-mail object attributes
US20040135814A1 (en) * 2003-01-15 2004-07-15 Vendelin George David Reading tool and method
US7793233B1 (en) * 2003-03-12 2010-09-07 Microsoft Corporation System and method for customizing note flags
US20050091578A1 (en) * 2003-10-24 2005-04-28 Microsoft Corporation Electronic sticky notes
US20070239831A1 (en) 2006-04-06 2007-10-11 Yahoo! Inc. Interface for editing, binding, and displaying an annotation for a message
US20090327890A1 (en) 2008-06-26 2009-12-31 Raytheon Company Graphical user interface (gui), display module and methods for displaying and comparing skin features
US20110102354A1 (en) * 2008-06-30 2011-05-05 Tetsuya Fuyuno Information processing device, display control method, and recording medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050076300A1 (en) * 2003-10-02 2005-04-07 International Business Machines Corporation Block marker system
KR20090042740A (en) * 2007-10-26 2009-04-30 리서치 인 모션 리미티드 Text selection using a touch sensitive screen of a handheld mobile communication device
US20090228842A1 (en) * 2008-03-04 2009-09-10 Apple Inc. Selecting of text using gestures
KR20090102135A (en) * 2008-03-25 2009-09-30 건국대학교 산학협력단 Installtion for management of dictionary information in electric device and method thereof

Also Published As

Publication number Publication date
US8791900B2 (en) 2014-07-29
US20110285638A1 (en) 2011-11-24
EP2572270A1 (en) 2013-03-27
EP2572270A4 (en) 2018-01-03

Similar Documents

Publication Publication Date Title
US8791900B2 (en) Computing device notes
US9727149B2 (en) Stylus settings
US9857970B2 (en) Copy and staple gestures
US10282086B2 (en) Brush, carbon-copy, and fill gestures
EP2580643B1 (en) Jump, checkmark, and strikethrough gestures
EP2529288B1 (en) Edge gestures
US20170038965A1 (en) Link Gestures
US8751550B2 (en) Freeform mathematical computations
US8799827B2 (en) Page manipulations using on and off-screen gestures
US20170300221A1 (en) Erase, Circle, Prioritize and Application Tray Gestures
US20110304556A1 (en) Activate, fill, and level gestures
US20110289462A1 (en) Computing Device Magnification Gesture
US20090100383A1 (en) Predictive gesturing in graphical user interface
US8786547B2 (en) Effects of gravity on gestures
US20110191719A1 (en) Cut, Punch-Out, and Rip Gestures
US20110185299A1 (en) Stamp Gestures
US20110185320A1 (en) Cross-reference Gestures
US20110191704A1 (en) Contextual multiplexing gestures
US20110285639A1 (en) Computing Device Writing Implement Techniques
JP6232694B2 (en) Information processing apparatus, control method thereof, and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11784210

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2011784210

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE