WO2015175634A1 - Electronic device and method for controlling access to same - Google Patents

Electronic device and method for controlling access to same Download PDF

Info

Publication number
WO2015175634A1
WO2015175634A1 PCT/US2015/030527 US2015030527W WO2015175634A1 WO 2015175634 A1 WO2015175634 A1 WO 2015175634A1 US 2015030527 W US2015030527 W US 2015030527W WO 2015175634 A1 WO2015175634 A1 WO 2015175634A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
electronic device
image data
access
level
Prior art date
Application number
PCT/US2015/030527
Other languages
French (fr)
Inventor
Rachid M. Alameh
Jiri Slaby
Original Assignee
Google Technology Holdings LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Google Technology Holdings LLC filed Critical Google Technology Holdings LLC
Priority to EP15726444.1A priority Critical patent/EP3143545B1/en
Priority to KR1020167034208A priority patent/KR102253098B1/en
Priority to CN201580024496.9A priority patent/CN106464679B/en
Priority to AU2015259277A priority patent/AU2015259277B2/en
Priority to CN202110023106.3A priority patent/CN112887269B/en
Publication of WO2015175634A1 publication Critical patent/WO2015175634A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/30Authentication, i.e. establishing the identity or authorisation of security principals
    • G06F21/31User authentication
    • G06F21/32User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/08Network architectures or network communication protocols for network security for authentication of entities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/08Network architectures or network communication protocols for network security for authentication of entities
    • H04L63/0861Network architectures or network communication protocols for network security for authentication of entities using biometrical features, e.g. fingerprint, retina-scan
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W12/00Security arrangements; Authentication; Protecting privacy or anonymity
    • H04W12/06Authentication
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2221/00Indexing scheme relating to security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F2221/21Indexing scheme relating to G06F21/00 and subgroups addressing additional information or applications relating to security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F2221/2111Location-sensitive, e.g. geographical location, GPS
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W12/00Security arrangements; Authentication; Protecting privacy or anonymity
    • H04W12/60Context-dependent security
    • H04W12/69Identity-dependent
    • H04W12/77Graphical identity

Definitions

  • the present disclosure is related generally to user authentication techniques on electronic devices.
  • biometric authentication Although the potential advantages of using biometric authentication over traditional personal identification number (“PIN”) authentication have long been understood, its use in consumer electronic devices has only recently become popular. With biometric authentication, a user does not need to enter a PIN and, under the right conditions, does not even need to touch the device in order to unlock it.
  • PIN personal identification number
  • FIG. 1 A is a front view of an electronic device according to an embodiment
  • FIG. IB is a rear view of an electronic device according to an embodiment
  • FIG. 2 is a block diagram of the electronic device according to an embodiment
  • FIG. 3 is a diagrammatic view of a scenario in which the electronic device may be used
  • FIG. 4 is a process flow diagram of a method that may be carried out in an embodiment
  • FIG. 5 is a diagrammatic view of another scenario in which the electronic device may be used.
  • FIG. 6 is a process flow diagram of a method that may be carried out in another embodiment.
  • an electronic device (also referred to as “the device") is able to alter one or more settings of its imager (e.g., its camera) based on the motion of a user that the device is attempting to authenticate.
  • the device captures a first set of image data of the user (e.g., a moving video or still image of the user), alters a setting of the imager based on the motion, captures a second set of image data of the user, and authenticates the user based on the second set of image data.
  • the device grants the user a first level of access to the device based on the first set of image data and grants the user second level of access to the device based on the second set of image data.
  • the number of possible access levels is not limited, and the example of two levels discussed herein is only meant to be illustrative.
  • the electronic device may capture the two sets of image data with two different imagers, stitch the sets of image data together, and carry out authentication on the stitched sets of image data.
  • the number of imagers that may be used is not limited to two, however.
  • an embodiment of the electronic device (“the device"), generally labeled 100, includes a housing 102 having a front side 104 and a rear side 106.
  • a first imager 11 OA Set along the perimeter of the housing are a first imager 11 OA, a second imager HOB, a third imager HOC, and a fourth imager HOD.
  • Each of the first through fourth imagers has a field of view that extends outwardly from the perimeter of the device 100.
  • Also set along the perimeter of the device 100 are a first motion sensor 116A, a second motion sensor 116B, a third motion sensor 116C, and a fourth motion sensor 116D.
  • Each motion sensor is configured to sense motion external to device 100.
  • Each motion sensor may be implemented as a passive infrared detector, such as a digital thermopile sensor, or as an active sensor that uses reflected light of a light source of the device 100.
  • a display 108 e.g., an organic light-emitting diode display
  • a fifth imager 110E e.g., a front facing camera
  • a sixth imager 11 OF e.g., a rear facing camera
  • the electronic device 100 may be implemented as other types of devices, including a tablet computer, portable gaming device, and a wearable device (e.g., a smart watch).
  • an embodiment of the electronic device 100 includes a processor 202, network communication hardware 204 (e.g., WiFI chip or a cellular baseband chipset), an audio output 206 (e.g., a speaker), a memory 208 (which can be implemented as volatile memory or non-volatile memory), and a light source 212 (e.g., an infrared light-emitting diode).
  • the processor 202 retrieves instructions and data from the memory 208 and, using the instructions and data, carries out the methods described herein.
  • FIG. 2 including the elements of FIGS. 1A and IB that appear in FIG.
  • data pathways 226 are communicatively linked to one or more other elements via one or more data pathways 226.
  • Possible implementations of the data pathways 226 include wires, conductive pathways on a microchip, and wireless connections.
  • Possible implementations of the processor 202 include a microprocessor and a controller.
  • FIG. 3 a procedure that the device 100 carries out to authenticate a user in an embodiment will now be described.
  • the electronic device 100 is lying on a table in a room 304.
  • a user 302 of the device enters the room 104 at position A and is moving.
  • the first motion sensor 116A detects the user 302 and provides data regarding the user to the processor 202 (FIG. 2), including data regarding the user's position, motion (including the user's gait), speed, and context.
  • the processor 202 turns on the first imager 110A and controls the first imager 110A to capture a first set of image data (i.e., a still image, multiple still images, or multiple images organized as a moving image) of the user 302 (block 402) and provides the first set of image data to the processor 202.
  • the processor 202 attempts to authenticate the user 302 using the first set of image data.
  • the processor 202 may attempt to authenticate the user 302 based on biometric data, such as the user's body geometry (e.g., the user's body shape, gender, height, girth, and gait).
  • the processor 202 determines that it is possible that the user 302 is the authorized user. Conversely, if the image data indicates that the user 302 is a short female, then the authentication will fail.
  • the processor 202 determines, with at least a 50% confidence level (based on its authentication attempt with the first set of image data) that the user 302 is an authorized user. Based on this determination, the processor 202 grants the user 302 a first level of access to the device 100.
  • the first level of access may involve granting the user 302 access to telephone functions or lower security applications of the device 100.
  • the processor 202 may control the audio output 206 to inform that user 302 that "You missed two phone calls and have one voicemail.”
  • the processor 202 may also control the display 108 to display the user's access level (e.g., "You are now able to access the phone functions").
  • the processor 202 continues to receive data (position, motion, speed, and context) from the first motion sensor 116A.
  • the processor 202 analyzes the data from the first motion sensor 116A.
  • the processor 202 alters a setting of the first imager 116A based on the detected motion. For example, the processor 202 may determine, based on the detected motion, that the user 302 is moving at or above a certain speed threshold (e.g., 3 feet per second), and, based on this fact, may increase the frame rate of the first imager 11 OA (e.g., from 20 frames per second ("fps") to 50 fps).
  • a certain speed threshold e.g., 3 feet per second
  • This increase in frame rate allows the first imager 110A to obtain more detail about the user in order to compensate for the fact that the user 302 is now in motion or is now moving faster.
  • Other ways that the processor 202 can alter a setting of the first imager 11 OA include controlling the first imager 11 OA to change one or more of its shutter speed, shutter timing, illumination setting, resolution, aperture, and zoom setting. In various embodiments, any or all of these changes may be triggered by the same motion sensor that prompted the processor 202 to turn on the first imager 202.
  • the processor 202 controls the first imager 11 OA to capture a second set of image data of the user 302 (block 406) and provide the second set of image data to the processor 202.
  • the processor 202 may receive a second moving video of the user from the first imager 11 OA, this time at the higher frame rate.
  • the processor 202 is able to use the second set of image data (e.g., the second, higher-frame-rate moving video) to authenticate the user 302 (block 408).
  • the processor 202 may authenticate the user 302 with a high enough confidence level to grant the user 302 a second level of access.
  • the processor 214 grants the user 302 the second level of access to the device 100 based on the second set of image data. Granting the second level of access may involve the processor 202 granting the user 302 access to one or more of pictures, files, emails, or higher security applications on the device 100.
  • the processor 202 may also control the display 108 to display the user's access level (e.g., "You are now able to access email").
  • the device 100 uses multiple imagers to gradually authenticate a user. Referring to FIGS. 5 and 6, a procedure for doing so will now be described.
  • the electronic device 100 is lying on a table in a room 504.
  • a user 502 of the device enters the room 504 at position A and is moving.
  • the first motion sensor 116A detects the user 502 and provides data regarding the user, such as the user's position, motion (including the user's gait), speed, and context, to the processor 202 (FIG. 2).
  • the processor 202 turns on the first imager 110A and controls the first imager 110A to capture a first set of image data of the user 302 (block 602) and provides the first set of image data to the processor 202.
  • the processor 202 attempts to authenticate the user 502 using the first set of image data.
  • the processor 202 determines, with at least a 50% confidence level based on its authentication attempt with the image data that the user 502 is an authorized user. Based on this determination, the processor 202 grants the user 502 a first level of access to the device 100 (block 604).
  • the processor 202 then receives data regarding the user, including the user's position, motion (including the user's gait), speed, and context, from the second motion sensor 116B.
  • the processor 202 analyzes the data from the second motion sensor 116B and, based on this motion data (and possibly based on further data from the first motion sensor 116A) determines that the user 502 has moved within viewing range of the second imager HOB.
  • the processor 202 reacts by turning on the second imager HOB and controlling the second imager 110B to capture a second set of image data of the user 502 (block 606).
  • the controller 202 then stitches the first set of image data and the second set of image data together (block 608).
  • the processor 202 grants the user 502 a second level of access to the electronic device 100 based on the stitched first and second sets of image data.
  • the processor 202 may also use the stitched images to assess the environment surrounding the device 100—such as the walls, ceiling, room settings, and table— and grant a level access to the user if the processor 202 determines that the environment is specific to the user (the user's house, office, car, etc.)
  • the processor 202 can also use the surrounding environment to reinforce the biometric data (i.e., the user's gait, etc.) collected regarding the user.
  • the combination of the environmental authentication and the biometric authentication is enough for the processor 202 to raise the level of access from a first level to a second level at block 610.
  • the process described in conjunction with FIG. 6 is not limited to two imagers.
  • the third and fourth motion sensors 116C and 116D could detect the motion and signal the processor 202.
  • the processor 202 could react by activating the third imager 1 IOC and the fourth imager HOD, respectively, control the imagers to capture third and fourth sets of video data, and perform stitching (and possibly environmental analysis) in order to grant the second level of access, or even to grant further levels of access.
  • the process described in conjunction with FIG. 6 may also be carried out with sensors of the device 100, such as the motion sensors 116A-116D.
  • the processor 202 may stitch the data from the first motion sensor 116A and the second motion sensor 116B together.
  • the stitched data can be used, for example, to map the XY position of the user, and may be part of the basis upon which the processor 202 grants the first or second level of access.

Abstract

An electronic device is able to alter one or more settings of its imager based on the motion of a user that the device is attempting to authenticate. The electronic device, in one implementation, captures a first set of image data of the user (e.g., a video or still photo of the user), detects motion of the user, alters a setting of the imager based on the motion, captures a second set of image data of the user, and authenticates the user based on the second set of image data. In some implementations, the electronic device has multiple imagers, and activates one or more additional imagers based on the detected motion of the user.

Description

ELECTRONIC DEVICE AND METHOD FOR CONTROLLING ACCESS TO SAME
TECHNICAL FIELD
[0001] The present disclosure is related generally to user authentication techniques on electronic devices.
BACKGROUND
[0002] Although the potential advantages of using biometric authentication over traditional personal identification number ("PIN") authentication have long been understood, its use in consumer electronic devices has only recently become popular. With biometric authentication, a user does not need to enter a PIN and, under the right conditions, does not even need to touch the device in order to unlock it.
[0003] Most existing biometric authentication schemes use the same basic access logic that traditional PIN-based systems use. That is, a user is either authenticated or is not. The user either gains full access or no access. Furthermore, they generally do not adjust in real-time for dynamic conditions such as the movement and position of the user.
DRAWINGS
[0004] While the appended claims set forth the features of the present techniques with particularity, these techniques, together with their objects and advantages, may be best understood from the following detailed description taken in conjunction with the accompanying drawings of which:
[0005] FIG. 1 A is a front view of an electronic device according to an embodiment;
[0006] FIG. IB is a rear view of an electronic device according to an embodiment;
[0007] FIG. 2 is a block diagram of the electronic device according to an embodiment;
[0008] FIG. 3 is a diagrammatic view of a scenario in which the electronic device may be used;
[0009] FIG. 4 is a process flow diagram of a method that may be carried out in an embodiment; [0010] FIG. 5 is a diagrammatic view of another scenario in which the electronic device may be used; and
[0011] FIG. 6 is a process flow diagram of a method that may be carried out in another embodiment.
DESCRIPTION
[0012] According to various embodiments of the disclosure, an electronic device (also referred to as "the device") is able to alter one or more settings of its imager (e.g., its camera) based on the motion of a user that the device is attempting to authenticate. In an embodiment, the device captures a first set of image data of the user (e.g., a moving video or still image of the user), alters a setting of the imager based on the motion, captures a second set of image data of the user, and authenticates the user based on the second set of image data.
[0013] According to an embodiment of the disclosure, the device grants the user a first level of access to the device based on the first set of image data and grants the user second level of access to the device based on the second set of image data. The number of possible access levels is not limited, and the example of two levels discussed herein is only meant to be illustrative. Additionally, the electronic device may capture the two sets of image data with two different imagers, stitch the sets of image data together, and carry out authentication on the stitched sets of image data. The number of imagers that may be used is not limited to two, however.
[0014] Turning to FIG. 1A and FIG. IB, an embodiment of the electronic device ("the device"), generally labeled 100, includes a housing 102 having a front side 104 and a rear side 106. Set along the perimeter of the housing are a first imager 11 OA, a second imager HOB, a third imager HOC, and a fourth imager HOD. Each of the first through fourth imagers has a field of view that extends outwardly from the perimeter of the device 100. Also set along the perimeter of the device 100 are a first motion sensor 116A, a second motion sensor 116B, a third motion sensor 116C, and a fourth motion sensor 116D. Each motion sensor is configured to sense motion external to device 100. Each motion sensor may be implemented as a passive infrared detector, such as a digital thermopile sensor, or as an active sensor that uses reflected light of a light source of the device 100. [0015] Set within the front side 104 of the housing 102 is a display 108 (e.g., an organic light-emitting diode display) and a fifth imager 110E (e.g., a front facing camera). Set within the rear side 106 of the housing 102 is a sixth imager 11 OF (e.g., a rear facing camera). Although depicted in FIGS. 1A and IB as a smartphone, the electronic device 100 may be implemented as other types of devices, including a tablet computer, portable gaming device, and a wearable device (e.g., a smart watch).
[0016] Turning to FIG. 2, an embodiment of the electronic device 100 includes a processor 202, network communication hardware 204 (e.g., WiFI chip or a cellular baseband chipset), an audio output 206 (e.g., a speaker), a memory 208 (which can be implemented as volatile memory or non-volatile memory), and a light source 212 (e.g., an infrared light-emitting diode). In various embodiments, the processor 202 retrieves instructions and data from the memory 208 and, using the instructions and data, carries out the methods described herein. Each of the elements of FIG. 2 (including the elements of FIGS. 1A and IB that appear in FIG. 2) is communicatively linked to one or more other elements via one or more data pathways 226. Possible implementations of the data pathways 226 include wires, conductive pathways on a microchip, and wireless connections. Possible implementations of the processor 202 include a microprocessor and a controller.
[0017] Turning to FIG. 3 and to the flowchart of FIG. 4, a procedure that the device 100 carries out to authenticate a user in an embodiment will now be described. As shown in FIG. 3, the electronic device 100 is lying on a table in a room 304. A user 302 of the device enters the room 104 at position A and is moving. When the user is at position A, the first motion sensor 116A detects the user 302 and provides data regarding the user to the processor 202 (FIG. 2), including data regarding the user's position, motion (including the user's gait), speed, and context. In response to receiving the data, the processor 202 turns on the first imager 110A and controls the first imager 110A to capture a first set of image data (i.e., a still image, multiple still images, or multiple images organized as a moving image) of the user 302 (block 402) and provides the first set of image data to the processor 202. The processor 202 attempts to authenticate the user 302 using the first set of image data. For example, the processor 202 may attempt to authenticate the user 302 based on biometric data, such as the user's body geometry (e.g., the user's body shape, gender, height, girth, and gait). Thus, if the processor 202 knows that an authorized user is a tall male, and the image data indicates that the user 302 is a tall male, then the processor 202 will determine that it is possible that the user 302 is the authorized user. Conversely, if the image data indicates that the user 302 is a short female, then the authentication will fail.
[0018] In this scenario, the processor 202 determines, with at least a 50% confidence level (based on its authentication attempt with the first set of image data) that the user 302 is an authorized user. Based on this determination, the processor 202 grants the user 302 a first level of access to the device 100. The first level of access may involve granting the user 302 access to telephone functions or lower security applications of the device 100. For example, the processor 202 may control the audio output 206 to inform that user 302 that "You missed two phone calls and have one voicemail." The processor 202 may also control the display 108 to display the user's access level (e.g., "You are now able to access the phone functions").
[0019] The processor 202 continues to receive data (position, motion, speed, and context) from the first motion sensor 116A. The processor 202 analyzes the data from the first motion sensor 116A. At block 404, the processor 202 alters a setting of the first imager 116A based on the detected motion. For example, the processor 202 may determine, based on the detected motion, that the user 302 is moving at or above a certain speed threshold (e.g., 3 feet per second), and, based on this fact, may increase the frame rate of the first imager 11 OA (e.g., from 20 frames per second ("fps") to 50 fps). This increase in frame rate allows the first imager 110A to obtain more detail about the user in order to compensate for the fact that the user 302 is now in motion or is now moving faster. Other ways that the processor 202 can alter a setting of the first imager 11 OA include controlling the first imager 11 OA to change one or more of its shutter speed, shutter timing, illumination setting, resolution, aperture, and zoom setting. In various embodiments, any or all of these changes may be triggered by the same motion sensor that prompted the processor 202 to turn on the first imager 202.
[0020] After the processor 202 alters the setting, the processor 202 controls the first imager 11 OA to capture a second set of image data of the user 302 (block 406) and provide the second set of image data to the processor 202. For example, the processor 202 may receive a second moving video of the user from the first imager 11 OA, this time at the higher frame rate.
[0021] In this example, it is assumed that the processor 202 is able to use the second set of image data (e.g., the second, higher-frame-rate moving video) to authenticate the user 302 (block 408). For example, the processor 202 may authenticate the user 302 with a high enough confidence level to grant the user 302 a second level of access. The processor 214 grants the user 302 the second level of access to the device 100 based on the second set of image data. Granting the second level of access may involve the processor 202 granting the user 302 access to one or more of pictures, files, emails, or higher security applications on the device 100. The processor 202 may also control the display 108 to display the user's access level (e.g., "You are now able to access email").
[0022] In another embodiment, the device 100 uses multiple imagers to gradually authenticate a user. Referring to FIGS. 5 and 6, a procedure for doing so will now be described.
[0023] As shown in FIG. 5, the electronic device 100 is lying on a table in a room 504. A user 502 of the device enters the room 504 at position A and is moving. When the user is at position A, the first motion sensor 116A detects the user 502 and provides data regarding the user, such as the user's position, motion (including the user's gait), speed, and context, to the processor 202 (FIG. 2). In response to receiving the data, the processor 202 turns on the first imager 110A and controls the first imager 110A to capture a first set of image data of the user 302 (block 602) and provides the first set of image data to the processor 202. The processor 202 attempts to authenticate the user 502 using the first set of image data. In this scenario, the processor 202 determines, with at least a 50% confidence level based on its authentication attempt with the image data that the user 502 is an authorized user. Based on this determination, the processor 202 grants the user 502 a first level of access to the device 100 (block 604).
[0024] The processor 202 then receives data regarding the user, including the user's position, motion (including the user's gait), speed, and context, from the second motion sensor 116B. The processor 202 analyzes the data from the second motion sensor 116B and, based on this motion data (and possibly based on further data from the first motion sensor 116A) determines that the user 502 has moved within viewing range of the second imager HOB. The processor 202 reacts by turning on the second imager HOB and controlling the second imager 110B to capture a second set of image data of the user 502 (block 606). The controller 202 then stitches the first set of image data and the second set of image data together (block 608). This stitching process allows the processor 202 to get a more comprehensive view of the user 502 and attempt to authenticate the user 502 on that basis. At block 610, the processor 202 grants the user 502 a second level of access to the electronic device 100 based on the stitched first and second sets of image data. In doing so, the processor 202 may also use the stitched images to assess the environment surrounding the device 100— such as the walls, ceiling, room settings, and table— and grant a level access to the user if the processor 202 determines that the environment is specific to the user (the user's house, office, car, etc.) The processor 202 can also use the surrounding environment to reinforce the biometric data (i.e., the user's gait, etc.) collected regarding the user. In this scenario, the combination of the environmental authentication and the biometric authentication is enough for the processor 202 to raise the level of access from a first level to a second level at block 610.
[0025] The process described in conjunction with FIG. 6 is not limited to two imagers. For example, if the user 502 continued to walk around the device 100, the third and fourth motion sensors 116C and 116D could detect the motion and signal the processor 202. The processor 202 could react by activating the third imager 1 IOC and the fourth imager HOD, respectively, control the imagers to capture third and fourth sets of video data, and perform stitching (and possibly environmental analysis) in order to grant the second level of access, or even to grant further levels of access.
[0026] Furthermore, the process described in conjunction with FIG. 6 may also be carried out with sensors of the device 100, such as the motion sensors 116A-116D. For example, as the user walks around the device 100, the processor 202 may stitch the data from the first motion sensor 116A and the second motion sensor 116B together. The stitched data can be used, for example, to map the XY position of the user, and may be part of the basis upon which the processor 202 grants the first or second level of access.
[0027] It should be understood that the embodiments described herein should be considered in a descriptive sense only and not for purposes of limitation. Descriptions of features or aspects within each embodiment should typically be considered as available for other similar features or aspects in other embodiments.
[0028] While one or more embodiments of the have been described with reference to the figures, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from their spirit and scope of as defined by the following claims. For example, the steps of the flow diagrams of FIGS. 4 and 6 can be reordered in way that will be apparent to those of skill in the art. Steps may also be added to the flow diagrams of FIGS. 4 and 6 without departing from the spirit of the disclosure.

Claims

CLAIMS We claim:
1. A method for controlling access to an electronic device, the method comprising: capturing a first set of image data of a user with an imager of the electronic device; altering a setting of the imager based on detected motion of the user;
capturing a second set of image data of the user with the imager after altering the setting; and
authenticating the user a second level of access to the electronic device based on the second set of image data.
2. The method of claim 1, further comprising:
granting the user a first level of access to the electronic device based on the first set of image data; and
granting the user a second level of access to the electronic device based on the second set of image data.
3. The method of claim 2, wherein
granting the first level of access comprises granting the user access to telephone functions or lower security applications of the electronic device,
granting the second level of access comprises granting the user access to one or more of pictures, files, emails, and higher security applications on the electronic device.
4. The method of claim 1, further comprising
carrying out a first authentication procedure on the user based on the user's body geometry with the first set of image data;
carrying out a second authentication procedure on the user based on the user's body geometry with the second set of image data.
5. The method of claim 4, wherein the user's body geometry is one or more of the user's body shape, body size, and gait.
6. The method of claim 1, wherein the setting is one or more of the imager's frame rate, shutter speed, shutter timing, illumination, resolution, aperture, and zoom.
7. The method of claim 1, further comprising changing the illumination intensity of a light source on the electronic device based on the detected motion of the user.
8. A method for controlling access to an electronic device, the method comprising:
capturing a first set of image data of a user with a first imager of the electronic device; granting the user a first level of access to the electronic device based on the first set of image data;
based on detected motion of the user, capturing a second set of image data of the user with a second imager of the electronic device;
stitching the first and second sets of image data together; and
granting the user a second level of access to the electronic device based on the stitched first and second sets of image data.
9. The method of claim 8, further comprising:
using the stitched first and second sets of image data to assess the environment surrounding the electronic device; and
granting the second level access based on the assessment.
10. The method of claim 9, further comprising:
using one or more of the first and second sets of image data to make a biometric assessment of the user; and
granting the second level access based on the combination of the environmental assessment and the biometric assessment.
11. An electronic device comprising:
an imager configured to capture a first set of image data of a user; and
a processor configured to
grant the user a first level of access to the electronic device based on the first set of image data;
alter a setting of the imager based on detected motion of the user, wherein the imager is further configured to capture a second set of image data of the user based on the altered setting, and wherein the processor is further configured to grant the user a second level of access to the electronic device based on the second set of image data.
12. The electronic device of claim 11, wherein the processor is further configured to
carry out a first authentication procedure on the user based on the user's body geometry with the first set of image data; and
carry out a second authentication procedure on the user based on the user's body geometry with the second set of image data.
13. The electronic device of claim 1 1, wherein
the processor is further configured to grant the first level of access by granting the user access to telephone functions or lower security applications of the electronic device, the processor is further configured to grant the second level of access by granting the user access to one or more of pictures, files, emails, and higher security applications on the electronic device.
14. A method for controlling access to an electronic device, the method comprising:
capturing a first set of image data of a user with a first imager of the electronic device; granting the user a first level of access to the electronic device based on the first set of image data;
based on detected motion of the user, capturing a second set of image data with a second imager of the electronic device;
stitching the first and second sets of data together; and
granting the user a second level of access to the electronic device based on the stitched first and second sets of image data.
15. The method of claim 14, further comprising activating the second imager based on the detected motion.
16. The method of claim 14, further comprising:
receiving data regarding the user from a first motion sensor and a second motion sensor;
stitching the data from the first motion sensor and the second motion sensor together, wherein granting the user the second level of access comprises granting the user a second level of access based, at least in part, on the stitched data from the first motion sensor and the second motion sensor.
PCT/US2015/030527 2014-05-13 2015-05-13 Electronic device and method for controlling access to same WO2015175634A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
EP15726444.1A EP3143545B1 (en) 2014-05-13 2015-05-13 Electronic device with method for controlling access to the same
KR1020167034208A KR102253098B1 (en) 2014-05-13 2015-05-13 Electronic device and method for controlling access to same
CN201580024496.9A CN106464679B (en) 2014-05-13 2015-05-13 Electronic device and method for controlling access thereto
AU2015259277A AU2015259277B2 (en) 2014-05-13 2015-05-13 Electronic device and method for controlling access to same
CN202110023106.3A CN112887269B (en) 2014-05-13 2015-05-13 Electronic device and method for controlling access thereto

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US14/276,107 US9710629B2 (en) 2014-05-13 2014-05-13 Electronic device with method for controlling access to same
US14/276,107 2014-05-13

Publications (1)

Publication Number Publication Date
WO2015175634A1 true WO2015175634A1 (en) 2015-11-19

Family

ID=53276281

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2015/030527 WO2015175634A1 (en) 2014-05-13 2015-05-13 Electronic device and method for controlling access to same

Country Status (6)

Country Link
US (2) US9710629B2 (en)
EP (1) EP3143545B1 (en)
KR (1) KR102253098B1 (en)
CN (2) CN112887269B (en)
AU (1) AU2015259277B2 (en)
WO (1) WO2015175634A1 (en)

Families Citing this family (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9229233B2 (en) 2014-02-11 2016-01-05 Osterhout Group, Inc. Micro Doppler presentations in head worn computing
US9298007B2 (en) 2014-01-21 2016-03-29 Osterhout Group, Inc. Eye imaging in head worn computing
US9965681B2 (en) 2008-12-16 2018-05-08 Osterhout Group, Inc. Eye imaging in head worn computing
US11103122B2 (en) 2014-07-15 2021-08-31 Mentor Acquisition One, Llc Content presentation in head worn computing
US9594246B2 (en) 2014-01-21 2017-03-14 Osterhout Group, Inc. See-through computer display systems
US20160147063A1 (en) * 2014-11-26 2016-05-26 Osterhout Group, Inc. See-through computer display systems
US20160019715A1 (en) 2014-07-15 2016-01-21 Osterhout Group, Inc. Content presentation in head worn computing
US20150228119A1 (en) 2014-02-11 2015-08-13 Osterhout Group, Inc. Spatial location presentation in head worn computing
US10684687B2 (en) 2014-12-03 2020-06-16 Mentor Acquisition One, Llc See-through computer display systems
US9746686B2 (en) 2014-05-19 2017-08-29 Osterhout Group, Inc. Content position calibration in head worn computing
US9829707B2 (en) 2014-08-12 2017-11-28 Osterhout Group, Inc. Measuring content brightness in head worn computing
US10649220B2 (en) 2014-06-09 2020-05-12 Mentor Acquisition One, Llc Content presentation in head worn computing
US9841599B2 (en) 2014-06-05 2017-12-12 Osterhout Group, Inc. Optical configurations for head-worn see-through displays
US10191279B2 (en) 2014-03-17 2019-01-29 Osterhout Group, Inc. Eye imaging in head worn computing
US9939934B2 (en) 2014-01-17 2018-04-10 Osterhout Group, Inc. External user interface for head worn computing
US9575321B2 (en) 2014-06-09 2017-02-21 Osterhout Group, Inc. Content presentation in head worn computing
US10254856B2 (en) 2014-01-17 2019-04-09 Osterhout Group, Inc. External user interface for head worn computing
US9299194B2 (en) 2014-02-14 2016-03-29 Osterhout Group, Inc. Secure sharing in head worn computing
US9836122B2 (en) 2014-01-21 2017-12-05 Osterhout Group, Inc. Eye glint imaging in see-through computer display systems
US11892644B2 (en) 2014-01-21 2024-02-06 Mentor Acquisition One, Llc See-through computer display systems
US9766463B2 (en) 2014-01-21 2017-09-19 Osterhout Group, Inc. See-through computer display systems
US11487110B2 (en) 2014-01-21 2022-11-01 Mentor Acquisition One, Llc Eye imaging in head worn computing
US9494800B2 (en) 2014-01-21 2016-11-15 Osterhout Group, Inc. See-through computer display systems
US11669163B2 (en) 2014-01-21 2023-06-06 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US9740280B2 (en) 2014-01-21 2017-08-22 Osterhout Group, Inc. Eye imaging in head worn computing
US20150205135A1 (en) 2014-01-21 2015-07-23 Osterhout Group, Inc. See-through computer display systems
US9529199B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. See-through computer display systems
US9753288B2 (en) 2014-01-21 2017-09-05 Osterhout Group, Inc. See-through computer display systems
US9852545B2 (en) 2014-02-11 2017-12-26 Osterhout Group, Inc. Spatial location presentation in head worn computing
US9401540B2 (en) 2014-02-11 2016-07-26 Osterhout Group, Inc. Spatial location presentation in head worn computing
US20150241963A1 (en) 2014-02-11 2015-08-27 Osterhout Group, Inc. Eye imaging in head worn computing
US20160187651A1 (en) 2014-03-28 2016-06-30 Osterhout Group, Inc. Safety for a vehicle operator with an hmd
US9651787B2 (en) 2014-04-25 2017-05-16 Osterhout Group, Inc. Speaker assembly for headworn computer
US10853589B2 (en) 2014-04-25 2020-12-01 Mentor Acquisition One, Llc Language translation with head-worn computing
US9710629B2 (en) 2014-05-13 2017-07-18 Google Technology Holdings LLC Electronic device with method for controlling access to same
US10663740B2 (en) 2014-06-09 2020-05-26 Mentor Acquisition One, Llc Content presentation in head worn computing
US9684172B2 (en) 2014-12-03 2017-06-20 Osterhout Group, Inc. Head worn computer display systems
USD751552S1 (en) 2014-12-31 2016-03-15 Osterhout Group, Inc. Computer glasses
US20160239985A1 (en) 2015-02-17 2016-08-18 Osterhout Group, Inc. See-through computer display systems
US10878775B2 (en) 2015-02-17 2020-12-29 Mentor Acquisition One, Llc See-through computer display systems
KR20160109586A (en) * 2015-03-12 2016-09-21 삼성전자주식회사 Image processing system and mobile computing device including the same
WO2016149208A1 (en) * 2015-03-17 2016-09-22 Microsoft Technology Licensing, Llc Selectively providing personal information and access to functionality on lock screen based on biometric user authentication
US10133304B2 (en) * 2015-05-26 2018-11-20 Motorola Mobility Llc Portable electronic device proximity sensors and mode switching functionality
KR102423830B1 (en) * 2015-11-10 2022-07-22 삼성전자주식회사 Wearable device and control method thereof
US10591728B2 (en) 2016-03-02 2020-03-17 Mentor Acquisition One, Llc Optical systems for head-worn computers
US10667981B2 (en) 2016-02-29 2020-06-02 Mentor Acquisition One, Llc Reading assistance system for visually impaired
WO2019182562A1 (en) * 2018-03-20 2019-09-26 Hewlett-Packard Development Company, L.P. Sensors access control
US10909225B2 (en) 2018-09-17 2021-02-02 Motorola Mobility Llc Electronic devices and corresponding methods for precluding entry of authentication codes in multi-person environments
WO2022066712A1 (en) * 2020-09-23 2022-03-31 Gentex Corporation Mirror assembly with passenger camera

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1990769A1 (en) * 2006-02-21 2008-11-12 Oki Electric Industry Company, Limited Personal identification device and personal identification method
GB2474536A (en) * 2009-10-13 2011-04-20 Pointgrab Ltd Computer vision gesture based control by hand shape recognition and object tracking
US20120038796A1 (en) * 2010-08-12 2012-02-16 Posa John G Apparatus and method providing auto zoom in response to relative movement of target subject matter

Family Cites Families (98)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR200172315Y1 (en) * 1997-03-26 2000-04-01 김기일 Cellular phone with the functions of alarming emergency and acquiring speech and image
AU2223999A (en) * 1998-01-12 1999-07-26 David Monroe Apparatus for capturing, converting and transmitting a visual image signal via adigital transmission system
US6707487B1 (en) * 1998-11-20 2004-03-16 In The Play, Inc. Method for representing real-time motion
US7193652B2 (en) * 1999-08-17 2007-03-20 Applied Vision Systems, Inc. Dynamic range video recording and playback system and method
JP2002197437A (en) 2000-12-27 2002-07-12 Sony Corp Walking detection system, walking detector, device and walking detecting method
US20030053662A1 (en) * 2001-09-19 2003-03-20 Koninklijke Philips Electronics N.V. Method and apparatus for digital encoding and operator identification using stored user image
SE525304C2 (en) * 2002-04-22 2005-01-25 Snalle Ab Method and apparatus for controlling access between a computer and a communication network
US20050232470A1 (en) 2004-03-31 2005-10-20 Ibm Corporation Method and apparatus for determining the identity of a user by narrowing down from user groups
US7840308B2 (en) 2004-09-10 2010-11-23 Honda Motor Co., Ltd. Robot device control based on environment and position of a movable robot
US7498951B2 (en) * 2004-10-18 2009-03-03 Ixi Mobile (R &D), Ltd. Motion sensitive illumination system and method for a mobile computing device
US8370639B2 (en) * 2005-06-16 2013-02-05 Sensible Vision, Inc. System and method for providing secure access to an electronic device using continuous facial biometrics
US8079079B2 (en) * 2005-06-29 2011-12-13 Microsoft Corporation Multimodal authentication
AU2006312012B2 (en) * 2005-11-03 2011-11-17 Barry Stuecker Security system
CN1992596A (en) * 2005-12-27 2007-07-04 国际商业机器公司 User authentication device and method
US7903168B2 (en) * 2006-04-06 2011-03-08 Eastman Kodak Company Camera and method with additional evaluation image capture based on scene brightness changes
US20070286596A1 (en) * 2006-06-08 2007-12-13 Lonn Fredrik A Method and system for adjusting camera settings in a camera equipped mobile radio terminal
US20080220809A1 (en) * 2007-03-07 2008-09-11 Sony Ericsson Mobile Communications Ab Method and system for a self timer function for a camera and ...
US8953849B2 (en) * 2007-04-19 2015-02-10 Eyelock, Inc. Method and system for biometric recognition
EP3882797A1 (en) * 2007-09-24 2021-09-22 Apple Inc. Embedded authentication systems in an electronic device
US8942764B2 (en) * 2007-10-01 2015-01-27 Apple Inc. Personal media device controlled via user initiated movements utilizing movement based interfaces
JP5326527B2 (en) * 2008-11-28 2013-10-30 富士通株式会社 Authentication apparatus and authentication method
WO2010101697A2 (en) * 2009-02-06 2010-09-10 Oculis Labs, Inc. Video-based privacy supporting system
US8364971B2 (en) * 2009-02-26 2013-01-29 Kynen Llc User authentication system and method
US8477175B2 (en) * 2009-03-09 2013-07-02 Cisco Technology, Inc. System and method for providing three dimensional imaging in a network environment
US8111247B2 (en) * 2009-03-27 2012-02-07 Sony Ericsson Mobile Communications Ab System and method for changing touch screen functionality
SG177156A1 (en) * 2009-06-16 2012-01-30 Intel Corp Camera applications in a handheld device
JP5397081B2 (en) * 2009-08-12 2014-01-22 富士通モバイルコミュニケーションズ株式会社 Mobile device
FR2961621B1 (en) * 2010-06-22 2014-09-05 Arjowiggins Security METHOD OF AUTHENTICATING AND / OR IDENTIFYING A SECURITY ARTICLE
KR20120010764A (en) * 2010-07-27 2012-02-06 엘지전자 주식회사 MOBILE TERMINAL AND METHOD FOR CONTROLLING A THREE DIMENSION IMAGE in thereof
US8412158B2 (en) * 2010-08-17 2013-04-02 Qualcomm Incorporated Mobile device having increased security that is less obtrusive
US10395018B2 (en) * 2010-11-29 2019-08-27 Biocatch Ltd. System, method, and device of detecting identity of a user and authenticating a user
CN102137005B (en) 2010-12-31 2014-04-02 华为技术有限公司 Method, device and system for forwarding date in communication system
US8994499B2 (en) * 2011-03-16 2015-03-31 Apple Inc. Locking and unlocking a mobile device using facial recognition
US8594374B1 (en) * 2011-03-30 2013-11-26 Amazon Technologies, Inc. Secure device unlock with gaze calibration
US20120287031A1 (en) * 2011-05-12 2012-11-15 Apple Inc. Presence sensing
EP2712431A1 (en) * 2011-05-12 2014-04-02 Apple Inc. Presence sensing
US20120315016A1 (en) * 2011-06-12 2012-12-13 Hei Tao Fung Multi-Purpose Image and Video Capturing Device
KR101279576B1 (en) * 2011-06-15 2013-06-27 삼성테크윈 주식회사 Method for generating panorama image within digital image processing apparatus
US8693726B2 (en) * 2011-06-29 2014-04-08 Amazon Technologies, Inc. User identification by gesture recognition
US9082235B2 (en) * 2011-07-12 2015-07-14 Microsoft Technology Licensing, Llc Using facial data for device authentication or subject identification
US8839358B2 (en) * 2011-08-31 2014-09-16 Microsoft Corporation Progressive authentication
US8966613B2 (en) * 2011-09-30 2015-02-24 Microsoft Technology Licensing, Llc Multi-frame depth image information identification
EP3697110A1 (en) * 2012-01-08 2020-08-19 ImagiStar LLC Portable item containers which identify missing and/or extra container contents
CN102662554B (en) 2012-01-09 2015-06-24 联想(北京)有限公司 Information processing apparatus and code input mode switching method thereof
US9519769B2 (en) * 2012-01-09 2016-12-13 Sensible Vision, Inc. System and method for disabling secure access to an electronic device using detection of a predetermined device orientation
US9202105B1 (en) * 2012-01-13 2015-12-01 Amazon Technologies, Inc. Image analysis for user authentication
US8984622B1 (en) * 2012-01-17 2015-03-17 Amazon Technologies, Inc. User authentication through video analysis
US9066125B2 (en) 2012-02-10 2015-06-23 Advanced Biometric Controls, Llc Secure display
US9100825B2 (en) * 2012-02-28 2015-08-04 Verizon Patent And Licensing Inc. Method and system for multi-factor biometric authentication based on different device capture modalities
US9323912B2 (en) * 2012-02-28 2016-04-26 Verizon Patent And Licensing Inc. Method and system for multi-factor biometric authentication
US9137246B2 (en) * 2012-04-09 2015-09-15 Brivas Llc Systems, methods and apparatus for multivariate authentication
US9208492B2 (en) * 2013-05-13 2015-12-08 Hoyos Labs Corp. Systems and methods for biometric authentication of transactions
US20130322705A1 (en) * 2012-05-30 2013-12-05 Google Inc. Facial and fingerprint authentication
US20130326613A1 (en) * 2012-05-31 2013-12-05 Gregory Peter Kochanski Dynamic control of device unlocking security level
US8396265B1 (en) * 2012-06-26 2013-03-12 Google Inc. Facial recognition
US20140013422A1 (en) * 2012-07-03 2014-01-09 Scott Janus Continuous Multi-factor Authentication
US8983201B2 (en) 2012-07-30 2015-03-17 Microsoft Technology Licensing, Llc Three-dimensional visual phrases for object recognition
US10567376B2 (en) * 2012-08-24 2020-02-18 Sensible Vision, Inc. System and method for providing secure access to an electronic device using multifactor authentication
US9032510B2 (en) * 2012-09-11 2015-05-12 Sony Corporation Gesture- and expression-based authentication
US8886953B1 (en) * 2012-09-14 2014-11-11 Google Inc. Image processing
US20140250523A1 (en) * 2012-10-11 2014-09-04 Carnegie Mellon University Continuous Authentication, and Methods, Systems, and Software Therefor
US20140118520A1 (en) * 2012-10-29 2014-05-01 Motorola Mobility Llc Seamless authorized access to an electronic device
US8584219B1 (en) * 2012-11-07 2013-11-12 Fmr Llc Risk adjusted, multifactor authentication
US9166962B2 (en) * 2012-11-14 2015-10-20 Blackberry Limited Mobile communications device providing heuristic security authentication features and related methods
WO2014089576A1 (en) * 2012-12-07 2014-06-12 Chamtech Technologies Incorporated Techniques for biometric authentication of user of mobile device
US8983207B1 (en) * 2013-01-10 2015-03-17 Intuit Inc. Mitigating replay attacks using multiple-image authentication
US8856541B1 (en) * 2013-01-10 2014-10-07 Google Inc. Liveness detection
US20140197922A1 (en) * 2013-01-16 2014-07-17 Cygnus Broadband, Inc. System and method for positive identification on a mobile device
US10152135B2 (en) 2013-03-15 2018-12-11 Intel Corporation User interface responsive to operator position and gestures
US10706132B2 (en) * 2013-03-22 2020-07-07 Nok Nok Labs, Inc. System and method for adaptive user authentication
US20140310801A1 (en) 2013-04-11 2014-10-16 Nokia Corporation Method and Apparatus for Performing Authentication
US9384751B2 (en) 2013-05-06 2016-07-05 Honeywell International Inc. User authentication of voice controlled devices
US9760785B2 (en) * 2013-05-08 2017-09-12 Jpmorgan Chase Bank, N.A. Systems and methods for high fidelity multi-modal out-of-band biometric authentication
US9721175B2 (en) * 2013-05-08 2017-08-01 Jpmorgan Chase Bank, N.A. Systems and methods for high fidelity multi-modal out-of-band biometric authentication through vector-based multi-profile storage
EP3005607B1 (en) * 2013-05-30 2020-01-08 Intel Corporation Adaptive authentication systems and methods
BR112015030439A2 (en) * 2013-06-06 2017-07-25 Kustom Signals Inc Time-tracking traffic enforcement system with integrated video capture
US9626493B2 (en) * 2013-06-08 2017-04-18 Microsoft Technology Licensing, Llc Continuous digital content protection
CN104298909B (en) * 2013-07-19 2018-11-30 富泰华工业(深圳)有限公司 Electronic device, authentication system and method
US8646060B1 (en) * 2013-07-30 2014-02-04 Mourad Ben Ayed Method for adaptive authentication using a mobile device
US9213817B2 (en) * 2013-08-28 2015-12-15 Paypal, Inc. Motion-based credentials using magnified motion
US9928355B2 (en) * 2013-09-09 2018-03-27 Apple Inc. Background enrollment and authentication of a user
US20150071508A1 (en) * 2013-09-09 2015-03-12 Apple Inc. Background Enrollment and Authentication of a User
US9147061B1 (en) * 2013-11-22 2015-09-29 Google Inc. Multi-level authentication
US20150177842A1 (en) * 2013-12-23 2015-06-25 Yuliya Rudenko 3D Gesture Based User Authorization and Device Control Methods
CN103744506A (en) * 2013-12-26 2014-04-23 乐视致新电子科技(天津)有限公司 Electronic device and gesture unlocking method
US20150186628A1 (en) * 2013-12-27 2015-07-02 Isabel F. Bush Authentication with an electronic device
CN103761463B (en) 2014-01-13 2017-09-01 联想(北京)有限公司 A kind of information processing method and electronic equipment
US9710691B1 (en) * 2014-01-23 2017-07-18 Diamond Fortress Technologies, Inc. Touchless fingerprint matching systems and methods
US9600304B2 (en) * 2014-01-23 2017-03-21 Apple Inc. Device configuration for multiple users using remote user biometrics
US9760383B2 (en) * 2014-01-23 2017-09-12 Apple Inc. Device configuration with multiple profiles for a single user using remote user biometrics
US9747428B2 (en) * 2014-01-30 2017-08-29 Qualcomm Incorporated Dynamic keyboard and touchscreen biometrics
US20150220931A1 (en) * 2014-01-31 2015-08-06 Apple Inc. Use of a Biometric Image for Authorization
US9552684B2 (en) * 2014-02-04 2017-01-24 Secure Gravity Inc. Methods and systems configured to detect and guarantee identity for the purpose of data protection and access control
US9510196B2 (en) * 2014-03-17 2016-11-29 Qualcomm Incorporated Method and apparatus for authenticating a user on a mobile device
US9710629B2 (en) 2014-05-13 2017-07-18 Google Technology Holdings LLC Electronic device with method for controlling access to same
US9408076B2 (en) * 2014-05-14 2016-08-02 The Regents Of The University Of California Sensor-assisted biometric authentication for smartphones
US20150347732A1 (en) * 2014-05-29 2015-12-03 Google Technology Holdings LLC Electronic Device and Method for Controlling Access to Same
WO2016119696A1 (en) * 2015-01-29 2016-08-04 艾尔希格科技股份有限公司 Action based identity identification system and method

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1990769A1 (en) * 2006-02-21 2008-11-12 Oki Electric Industry Company, Limited Personal identification device and personal identification method
GB2474536A (en) * 2009-10-13 2011-04-20 Pointgrab Ltd Computer vision gesture based control by hand shape recognition and object tracking
US20120038796A1 (en) * 2010-08-12 2012-02-16 Posa John G Apparatus and method providing auto zoom in response to relative movement of target subject matter

Also Published As

Publication number Publication date
US10255417B2 (en) 2019-04-09
EP3143545A1 (en) 2017-03-22
EP3143545B1 (en) 2020-07-29
US20150332032A1 (en) 2015-11-19
US9710629B2 (en) 2017-07-18
CN106464679A (en) 2017-02-22
AU2015259277B2 (en) 2020-03-19
CN112887269A (en) 2021-06-01
CN112887269B (en) 2022-12-27
KR20170007772A (en) 2017-01-20
CN106464679B (en) 2021-01-26
US20170277876A1 (en) 2017-09-28
AU2015259277A1 (en) 2016-11-10
KR102253098B1 (en) 2021-05-14

Similar Documents

Publication Publication Date Title
US10255417B2 (en) Electronic device with method for controlling access to same
US11330012B2 (en) System, method, and device of authenticating a user based on selfie image or selfie video
WO2019109767A1 (en) Task execution method, terminal device and computer readable storage medium
JP6613740B2 (en) Display control apparatus, display control method, and display control program
US9652663B2 (en) Using facial data for device authentication or subject identification
US20150347732A1 (en) Electronic Device and Method for Controlling Access to Same
KR102230691B1 (en) Method and device for recognizing biometric information
TWI700607B (en) Unlocking system and method
US9924090B2 (en) Method and device for acquiring iris image
TWI752105B (en) Feature image acquisition method, acquisition device, and user authentication method
KR102371211B1 (en) Device and method for authenticating biometric data using a plural camera
KR102532746B1 (en) Electronic device and control method thereof
KR102423364B1 (en) Method for providing image and electronic device supporting the same
US11687635B2 (en) Automatic exposure and gain control for face authentication
EP3495999A1 (en) Image recognition method, apparatus and computer readable storage medium
CN112417998A (en) Method and device for acquiring living body face image, medium and equipment
JP6492483B2 (en) Detection apparatus, detection method, and program
KR20180076779A (en) Terminal and iris recognition method using the same

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15726444

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2015259277

Country of ref document: AU

Date of ref document: 20150513

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

REEP Request for entry into the european phase

Ref document number: 2015726444

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2015726444

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 20167034208

Country of ref document: KR

Kind code of ref document: A