US20140306986A1 - Identifying Content in Electronic Images - Google Patents

Identifying Content in Electronic Images Download PDF

Info

Publication number
US20140306986A1
US20140306986A1 US13/862,212 US201313862212A US2014306986A1 US 20140306986 A1 US20140306986 A1 US 20140306986A1 US 201313862212 A US201313862212 A US 201313862212A US 2014306986 A1 US2014306986 A1 US 2014306986A1
Authority
US
United States
Prior art keywords
graphics
type
text
percentage
surface area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US13/862,212
Other versions
US10296933B2 (en
Inventor
Frederick Gottesman
David Clune
James Andrews
Igor Gevka
Satwik Shukla
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Meta Platforms Inc
Original Assignee
Facebook Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US13/862,212 priority Critical patent/US10296933B2/en
Application filed by Facebook Inc filed Critical Facebook Inc
Priority to BR112015025970A priority patent/BR112015025970A2/en
Priority to EP14783229.9A priority patent/EP2984556A4/en
Priority to CN201480033674.XA priority patent/CN105408855B/en
Priority to KR1020157032308A priority patent/KR102025615B1/en
Priority to AU2014251221A priority patent/AU2014251221B2/en
Priority to MX2015014302A priority patent/MX364910B/en
Priority to JP2016507563A priority patent/JP6473135B2/en
Priority to CA2908554A priority patent/CA2908554A1/en
Priority to PCT/US2014/032669 priority patent/WO2014168795A1/en
Assigned to FACEBOOK, INC. reassignment FACEBOOK, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CLUNE, David, ANDREWS, JAMES, GEVKA, Igor, GOTTESMAN, Frederick, SHUKLA, SATWIK
Publication of US20140306986A1 publication Critical patent/US20140306986A1/en
Priority to JP2019010276A priority patent/JP6744932B2/en
Application granted granted Critical
Publication of US10296933B2 publication Critical patent/US10296933B2/en
Assigned to META PLATFORMS, INC. reassignment META PLATFORMS, INC. CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: FACEBOOK, INC.
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T1/00General purpose image data processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0241Advertisements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/40Document-oriented image-based pattern recognition
    • G06V30/41Analysis of document content
    • G06V30/413Classification of content, e.g. text, photographs or tables

Definitions

  • This disclosure generally relates to electronic image analysis.
  • Electronic images are presented as surfaces having graphical content.
  • Electronic images may be stored in any suitable electronic storage medium, such as a harddrive, flash drive, etc and may be stored according to any suitable format, such as JPEG, TIFF, GIF, etc..
  • Electronic images may be presented to a viewer using any suitable display, such as on a display on a personal computer, mobile device, touchscreen, kiosk, television screen, etc.
  • Electronic images may have variable characteristics such as size, shape, resolution, color schemes used, etc, and those characteristics may be varied for presentation on different types of displays or in different contexts, windows, etc. on a single display.
  • An electronic image such as an advertisement, may contain a type of graphics such as text, shapes, etc.
  • An example method may access the electronic image and divide the image into a number of surfaces.
  • the surfaces may have any suitable shape or size.
  • a determination is made about whether the type of graphics is present on one or more of the surfaces. For example, the determination may be made by any suitable image processing technique or based on input received from a user.
  • a percentage of the image's total surface area occupied by the surfaces determined to contain the type of graphics is calculated. In particular embodiments, if that percentage exceeds a predetermined percentage, the type of graphics may be resized, rearranged, or removed such that the total surface area occupied by the surfaces containing the type of graphics is less than the predetermined percentage.
  • FIG. 1 illustrates an example network environment associated with a social-networking system.
  • FIG. 2 illustrates an example method for determining the percentage of an electronic image occupied by a type of graphics.
  • FIG. 3 illustrates an example of determining the percentage of an electronic advertisement occupied by text.
  • FIG. 4 illustrates an example computer system.
  • FIG. 1 illustrates an example network environment 100 associated with a social-networking system.
  • Network environment 100 includes a user 101 , a client system 130 , a social-networking system 160 , and a third-party system 170 connected to each other by a network 110 .
  • FIG. 1 illustrates a particular arrangement of user 101 , client system 130 , social-networking system 160 , third-party system 170 , and network 110 , this disclosure contemplates any suitable arrangement of user 101 , client system 130 , social-networking system 160 , third-party system 170 , and network 110 .
  • two or more of client system 130 , social-networking system 160 , and third-party system 170 may be connected to each other directly, bypassing network 110 .
  • two or more of client system 130 , social-networking system 160 , and third-party system 170 may be physically or logically co-located with each other in whole or in part.
  • FIG. 1 illustrates a particular number of users 101 , client systems 130 , social-networking systems 160 , third-party systems 170 , and networks 110
  • this disclosure contemplates any suitable number of users 101 , client systems 130 , social-networking systems 160 , third-party systems 170 , and networks 110 .
  • network environment 100 may include multiple users 101 , client system 130 , social-networking systems 160 , third-party systems 170 , and networks 110 .
  • user 101 may be an individual (human user), an entity (e.g., an enterprise, business, or third-party application), or a group (e.g., of individuals or entities) that interacts or communicates with or over social-networking system 160 .
  • social-networking system 160 may be a network-addressable computing system hosting an online social network. Social-networking system 160 may generate, store, receive, and send social-networking data, such as, for example, user-profile data, concept-profile data, social-graph information, or other suitable data related to the online social network. Social-networking system 160 may be accessed by the other components of network environment 100 either directly or via network 110 .
  • social-networking system 160 may include an authorization server (or other suitable component(s)) that allows users 101 to opt in to or opt out of having their actions logged by social-networking system 160 or shared with other systems (e.g., third-party systems 170 ), for example, by setting appropriate privacy settings.
  • a privacy setting of a user may determine what information associated with the user may be logged, how information associated with the user may be logged, when information associated with the user may be logged, who may log information associated with the user, whom information associated with the user may be shared with, and for what purposes information associated with the user may be logged or shared.
  • Authorization servers may be used to enforce one or more privacy settings of the users of social-networking system 30 through blocking, data hashing, anonymization, or other suitable techniques as appropriate.
  • Third-party system 170 may generate, store, receive, and send data, such as, for example, an advertisement. Third-party system 170 may be accessed by the other components of network environment 100 either directly or via network 110 .
  • one or more users 101 may use one or more client systems 130 to access, send data to, and receive data from social-networking system 160 or third-party system 170 .
  • Client system 130 may access social-networking system 160 or third-party system 170 directly, via network 110 , or via a third-party system.
  • client system 130 may access third-party system 170 via social-networking system 160 .
  • Client system 130 may be any suitable computing device, such as, for example, a personal computer, a laptop computer, a cellular telephone, a smartphone, or a tablet computer.
  • network 110 may include any suitable network 110 .
  • one or more portions of network 110 may include an ad hoc network, an intranet, an extranet, a virtual private network (VPN), a local area network (LAN), a wireless LAN (WLAN), a wide area network (WAN), a wireless WAN (WWAN), a metropolitan area network (MAN), a portion of the Internet, a portion of the Public Switched Telephone Network (PSTN), a cellular telephone network, or a combination of two or more of these.
  • Network 110 may include one or more networks 110 .
  • Links 150 may connect client system 130 , social-networking system 160 , and third-party system 170 to communication network 110 or to each other.
  • This disclosure contemplates any suitable links 150 .
  • one or more links 150 include one or more wireline (such as for example Digital Subscriber Line (DSL) or Data Over Cable Service Interface Specification (DOCSIS)), wireless (such as for example Wi-Fi or Worldwide Interoperability for Microwave Access (WiMAX)), or optical (such as for example Synchronous Optical Network (SONET) or Synchronous Digital Hierarchy (SDH)) links.
  • wireline such as for example Digital Subscriber Line (DSL) or Data Over Cable Service Interface Specification (DOCSIS)
  • wireless such as for example Wi-Fi or Worldwide Interoperability for Microwave Access (WiMAX)
  • optical such as for example Synchronous Optical Network (SONET) or Synchronous Digital Hierarchy (SDH) links.
  • SONET Synchronous Optical Network
  • SDH Synchronous Digital Hierarchy
  • one or more links 150 each include an ad hoc network, an intranet, an extranet, a VPN, a LAN, a WLAN, a WAN, a WWAN, a MAN, a portion of the Internet, a portion of the PSTN, a cellular technology-based network, a satellite communications technology-based network, another link 150 , or a combination of two or more such links 150 .
  • Links 150 need not necessarily be the same throughout network environment 100 .
  • One or more first links 150 may differ in one or more respects from one or more second links 150 .
  • an advertisement may be text (which may be HTML-linked), one or more images (which may be HTML-linked), one or more videos, audio, one or more ADOBE FLASH files, a suitable combination of these, or any other suitable advertisement in any suitable digital format presented on one or more webpages, in one or more e-mails, or in connection with search results requested by a user.
  • an advertisement may be one or more sponsored stories (e.g., a news-feed or ticker item on social-networking system 160 ).
  • a sponsored story may be a social action by a user (such as “liking” a page, “liking” or commenting on a post on a page, RSVPing to an event associated with a page, voting on a question posted on a page, checking in to a place, using an application or playing a game, or “liking” or sharing a website) that an advertiser promotes, for example, by having the social action presented within a pre-determined area of a profile page of a user or other page, presented with additional information associated with the advertiser, bumped up or otherwise highlighted within news feeds or tickers of other users, or otherwise promoted.
  • the advertiser may pay to have the social action promoted.
  • advertisements may be included among the search results of a search-results page, where sponsored content is promoted over non-sponsored content.
  • an advertisement may be requested for display within social-networking-system webpages, third-party webpages, or other pages.
  • An advertisement may be displayed in a dedicated portion of a page, such as in a banner area at the top of the page, in a column at the side of the page, in a GUI of the page, in a pop-up window, in a drop-down menu, in an input field of the page, over the top of content of the page, or elsewhere with respect to the page.
  • an advertisement may be displayed within an application.
  • An advertisement may be displayed within dedicated pages, requiring the user to interact with or watch the advertisement before the user may access a page or utilize an application. The user may, for example view the advertisement through a web browser.
  • FIG. 2 illustrates an example method 200 for determining the percentage of an electronic image occupied by a type of graphics.
  • the method may begin at step 210 , where an electronic image is accessed.
  • Electronic images may be accessed from any suitable non-volatile storage medium, such as a harddrive, flash drive, ROM, etc.
  • Electronic images may accessed from any suitable volatile storage media, such as RAM, caches, buffers, etc.
  • surfaces may be represented by identifying the memory elements describing the portion of an image the surface comprises.
  • Electronic images may be stored according to any suitable format, such as JPEG, TIFF, GIF, PNG, etc.
  • Electronic images may be of any suitable type, such as photos or advertisements, and may include any suitable graphical content such as text, colors, shapes, etc.
  • the electronic image is divided into a plurality of surfaces.
  • the surfaces may take any suitable shape and be of any suitable size.
  • the surfaces may be non-overlapping and/or may collectively span the surface area of the image.
  • the surfaces may be squares, rectangles, hexagons, polygons, ellipses, or any suitable combination thereof.
  • the surfaces and corresponding divisions may be presented as divided to a user.
  • the divisions may be represented as a grid (e.g. by lines or other suitable markings), and the grid and surfaces may be presented to the user.
  • one surface of an image may have a different size or shape than another surface of the image.
  • divisions or the size, shape, or locations of one or more surfaces may be determined by identifying or estimating the presence of particular content or types of graphics in the image, as explained more fully below.
  • This disclosure contemplates dividing an image into any suitable type, number, or size of surfaces, and representing those surfaces and/or divisions in any suitable way.
  • This disclosure contemplates any suitable type of graphics, such as text, colors, shapes, etc.
  • the determination may be made based on input received from a user.
  • the surfaces may be presented to the user and the user may indicate (e.g. by clicking using a cursor, dragging using a cursor, touching on a touch-sensitive portion of the display, etc.) which surfaces contain a type of graphics.
  • the number and size of surfaces may be such that the user can make provide input quickly or within a predetermined amount of time, for example by using a uniform grid of rectangles each having a height and width of at least one inch or by using a uniform grid of a predetermined number of squares, such as a 5 ⁇ 5 grid.
  • the determination may be made by a computing device, for example by using one or more image-processing techniques, such as optical character recognition (OCR) or line recognition for text, or techniques (such as Fourier transforms) for identifying shapes, patterns, etc, or characteristics of the same.
  • OCR optical character recognition
  • line recognition for text
  • a determination may be made by a combination of user input and a computing device.
  • a user's determination of which surfaces contain a type of graphics may be checked by image processing techniques, and vice versa.
  • the divisions described above may be altered by or based on estimates of which portions of an image contain a type of graphics.
  • a user may identify where a type of graphics is present or identify regions containing the type of graphics, and that identification may be used to determine the size, shape, or location of one or more surfaces.
  • a surface may be enlarged or made smaller to more-accurately represent the shape or location of at least part of the particular type of graphics, such as by closely matching a surface to the area occupied by the type of graphics.
  • a computing device may estimate where a type of graphics is present, for example by using image-processing techniques as described above, and those estimates may be used to determine the size, shape, or location of one or more surfaces.
  • portions of an image where a type of graphics is likely to occur e.g. based on analysis of similar images
  • This disclosure contemplates determining that one or more surfaces comprise a type of graphics by any suitable technique.
  • the percentage of the image's surface area occupied by surfaces determined to include the type of graphics is calculated.
  • This disclosure contemplates any suitable method of calculating the percentage, such as adding the surface areas of the surfaces and dividing by the surface area of the image, by adding each surface area of a surface divided by the surface area of the image, dividing the number of surfaces containing the type of graphics by the total number of surfaces the image is divided into (if the surfaces are of uniform surface area), etc.
  • the percentage may be compared to a predetermined percentage, and some action may be taken as a result.
  • the image may be an advertisement and the type of graphics may be text.
  • the advertisement may be prohibited from being displayed, and the advertiser or publisher may be notified that the advertisement does not comply with rules governing the area of an advertisement text is permitted to cover.
  • the surface area of surfaces containing text is a greater percentage of the advertisement's surface area than a pre-determined percentage, at least some of the text may be rearranged, resized, and/or removed.
  • the text may iteratively rearranged, resized, and/or removed until the surface area of surfaces containing text is not a greater percentage of the advertisement's surface area than a pre-determined percentage, or a calculation may be made regarding one or more configurations of text that would result in the surface area of surfaces containing text to not be a greater percentage of the advertisement's surface area than a predetermined percentage, and the text may be configured to one of the calculated configurations.
  • a user may select all or a portion of the text and indicate a size, shape, or location of the text or a size, shape, or location of one or more surfaces to move the text onto. While this disclosure describes particular embodiments in the context of text on advertisements, this disclosure applies to any suitable type of graphics on any suitable type of electronic image.
  • Particular embodiments may repeat one or more steps of the method of FIG. 2 , where appropriate.
  • this disclosure describes and illustrates particular steps of the method of FIG. 2 as occurring in a particular order, this disclosure contemplates any suitable steps of the method of FIG. 2 occurring in any suitable order.
  • this disclosure describes and illustrates particular components, devices, or systems carrying out particular steps of the method of FIG. 2 , this disclosure contemplates any suitable combination of any suitable components, devices, or systems carrying out any suitable steps of the method of FIG. 2 .
  • a determination (whether automated, based on manual input, or some combination thereof) about whether a surface contains a type of graphics may be used to train automated methods used to determine or aid in a determination of whether a surface contains a type of graphics.
  • a determination may be used to train one or more image-processing tools such as OCR, line detection, etc.
  • machine-learning techniques such as decision trees, neural networks, clustering, reinforcement learning, etc. may be used to facilitate the training process.
  • FIG. 3 illustrates a specific example for determining the percentage of an electronic advertisement occupied by text (as described more fully in step 230 , above).
  • device 300 displays an advertisement 310 on a display.
  • the advertisement is divided into a 5 ⁇ 5 grid of rectangular surfaces 320 , and the divisions are indicated by gridlines 330 .
  • the user selects the rectangular surfaces where text appears, such as rectangular surface 340 .
  • the user may select the rectangular surfaces by any suitable method, such as by using a cursor, tapping the squares on a touch-sensitive display, swiping the squares on a touch-sensitive display, etc.
  • selected surfaces may be presented differently than unselected surfaces, for example by shading the selected surfaces, coloring the selected surfaces, etc. As illustrated by the example of FIG.
  • a type of graphics may be determined not to be a on a surface if only a small amount of the type of graphics is on the surface or the type of graphics occupies only a small amount of surface area of the surface.
  • surface 350 contains a small portion of text, and the user may not indicate that the surface contains text.
  • FIG. 4 illustrates an example computer system 400 .
  • one or more computer systems 400 perform one or more steps of one or more methods described or illustrated herein.
  • one or more computer systems 400 provide functionality described or illustrated herein.
  • software running on one or more computer systems 400 performs one or more steps of one or more methods described or illustrated herein or provides functionality described or illustrated herein.
  • Particular embodiments include one or more portions of one or more computer systems 400 .
  • reference to a computer system may encompass a computing device, and vice versa, where appropriate.
  • reference to a computer system may encompass one or more computer systems, where appropriate.
  • computer system 400 may be an embedded computer system, a system-on-chip (SOC), a single-board computer system (SB) (such as, for example, a computer-on-module (COM) or system-on-module (SOM)), a desktop computer system, a laptop or notebook computer system, an interactive kiosk, a mainframe, a mesh of computer systems, a mobile telephone, a personal digital assistant (PDA), a server, a tablet computer system, or a combination of two or more of these.
  • SOC system-on-chip
  • SB single-board computer system
  • COM computer-on-module
  • SOM system-on-module
  • desktop computer system such as, for example, a computer-on-module (COM) or system-on-module (SOM)
  • laptop or notebook computer system such as, for example, a computer-on-module (COM) or system-on-module (SOM)
  • desktop computer system such as, for example, a computer-on-module (COM
  • computer system 400 may include one or more computer systems 400 ; be unitary or distributed; span multiple locations; span multiple machines; span multiple data centers; or reside in a cloud, which may include one or more cloud components in one or more networks.
  • one or more computer systems 400 may perform without substantial spatial or temporal limitation one or more steps of one or more methods described or illustrated herein.
  • one or more computer systems 400 may perform in real time or in batch mode one or more steps of one or more methods described or illustrated herein.
  • One or more computer systems 400 may perform at different times or at different locations one or more steps of one or more methods described or illustrated herein, where appropriate.
  • computer system 400 includes a processor 402 , memory 404 , storage 406 , an input/output (I/O) interface 408 , a communication interface 410 , and a bus 412 .
  • I/O input/output
  • this disclosure describes and illustrates a particular computer system having a particular number of particular components in a particular arrangement, this disclosure contemplates any suitable computer system having any suitable number of any suitable components in any suitable arrangement.
  • processor 402 includes hardware for executing instructions, such as those making up a computer program.
  • processor 402 may retrieve (or fetch) the instructions from an internal register, an internal cache, memory 404 , or storage 406 ; decode and execute them; and then write one or more results to an internal register, an internal cache, memory 404 , or storage 406 .
  • processor 402 may include one or more internal caches for data, instructions, or addresses. This disclosure contemplates processor 402 including any suitable number of any suitable internal caches, where appropriate.
  • processor 402 may include one or more instruction caches, one or more data caches, and one or more translation lookaside buffers (TLBs). Instructions in the instruction caches may be copies of instructions in memory 404 or storage 406 , and the instruction caches may speed up retrieval of those instructions by processor 402 . Data in the data caches may be copies of data in memory 404 or storage 406 for instructions executing at processor 402 to operate on; the results of previous instructions executed at processor 402 for access by subsequent instructions executing at processor 402 or for writing to memory 404 or storage 406 ; or other suitable data. The data caches may speed up read or write operations by processor 402 . The TLBs may speed up virtual-address translation for processor 402 .
  • TLBs translation lookaside buffers
  • processor 402 may include one or more internal registers for data, instructions, or addresses. This disclosure contemplates processor 402 including any suitable number of any suitable internal registers, where appropriate. Where appropriate, processor 402 may include one or more arithmetic logic units (ALUs); be a multi-core processor; or include one or more processors 402 . Although this disclosure describes and illustrates a particular processor, this disclosure contemplates any suitable processor.
  • ALUs arithmetic logic units
  • memory 404 includes main memory for storing instructions for processor 402 to execute or data for processor 402 to operate on.
  • computer system 400 may load instructions from storage 406 or another source (such as, for example, another computer system 400 ) to memory 404 .
  • Processor 402 may then load the instructions from memory 404 to an internal register or internal cache.
  • processor 402 may retrieve the instructions from the internal register or internal cache and decode them.
  • processor 402 may write one or more results (which may be intermediate or final results) to the internal register or internal cache.
  • Processor 402 may then write one or more of those results to memory 404 .
  • processor 402 executes only instructions in one or more internal registers or internal caches or in memory 404 (as opposed to storage 406 or elsewhere) and operates only on data in one or more internal registers or internal caches or in memory 404 (as opposed to storage 406 or elsewhere).
  • One or more memory buses (which may each include an address bus and a data bus) may couple processor 402 to memory 404 .
  • Bus 412 may include one or more memory buses, as described below.
  • one or more memory management units reside between processor 402 and memory 404 and facilitate accesses to memory 404 requested by processor 402 .
  • memory 404 includes random access memory (RAM).
  • This RAM may be volatile memory, where appropriate Where appropriate, this RAM may be dynamic RAM (DRAM) or static RAM (SRAM). Moreover, where appropriate, this RAM may be single-ported or multi-ported RAM. This disclosure contemplates any suitable RAM.
  • Memory 404 may include one or more memories 404 , where appropriate. Although this disclosure describes and illustrates particular memory, this disclosure contemplates any suitable memory.
  • storage 406 includes mass storage for data or instructions.
  • storage 406 may include a hard disk drive (HDD), a floppy disk drive, flash memory, an optical disc, a magneto-optical disc, magnetic tape, or a Universal Serial Bus (USB) drive or a combination of two or more of these.
  • Storage 406 may include removable or non-removable (or fixed) media, where appropriate.
  • Storage 406 may be internal or external to computer system 400 , where appropriate.
  • storage 406 is non-volatile, solid-state memory.
  • storage 406 includes read-only memory (ROM).
  • this ROM may be mask-programmed ROM, programmable ROM (PROM), erasable PROM (EPROM), electrically erasable PROM (EEPROM), electrically alterable ROM (EAROM), or flash memory or a combination of two or more of these.
  • This disclosure contemplates mass storage 406 taking any suitable physical form.
  • Storage 406 may include one or more storage control units facilitating communication between processor 402 and storage 406 , where appropriate. Where appropriate, storage 406 may include one or more storages 406 . Although this disclosure describes and illustrates particular storage, this disclosure contemplates any suitable storage.
  • I/O interface 408 includes hardware, software, or both, providing one or more interfaces for communication between computer system 400 and one or more I/O devices.
  • Computer system 400 may include one or more of these I/O devices, where appropriate.
  • One or more of these I/O devices may enable communication between a person and computer system 400 .
  • an I/O device may include a keyboard, keypad, microphone, monitor, mouse, printer, scanner, speaker, still camera, stylus, tablet, touch screen, trackball, video camera, another suitable I/O device or a combination of two or more of these.
  • An I/O device may include one or more sensors. This disclosure contemplates any suitable I/O devices and any suitable I/O interfaces 408 for them.
  • I/O interface 408 may include one or more device or software drivers enabling processor 402 to drive one or more of these I/O devices.
  • I/O interface 408 may include one or more I/O interfaces 408 , where appropriate. Although this disclosure describes and illustrates a particular I/O interface, this disclosure contemplates any suitable I/O interface.
  • communication interface 410 includes hardware, software, or both providing one or more interfaces for communication (such as, for example, packet-based communication) between computer system 400 and one or more other computer systems 400 or one or more networks.
  • communication interface 410 may include a network interface controller (NIC) or network adapter for communicating with an Ethernet or other wire-based network or a wireless NIC (WNIC) or wireless adapter for communicating with a wireless network, such as a WI-FI network.
  • NIC network interface controller
  • WNIC wireless NIC
  • WI-FI network wireless network
  • computer system 400 may communicate with an ad hoc network, a personal area network (PAN), a local area network (LAN), a wide area network (WAN), a metropolitan area network (MAN), or one or more portions of the Internet or a combination of two or more of these.
  • PAN personal area network
  • LAN local area network
  • WAN wide area network
  • MAN metropolitan area network
  • computer system 400 may communicate with a wireless PAN (WPAN) (such as, for example, a BLUETOOTH WPAN), a WI-FI network, a WI-MAX network, a cellular telephone network (such as, for example, a Global System for Mobile Communications (GSM) network), or other suitable wireless network or a combination of two or more of these.
  • WPAN wireless PAN
  • WI-FI wireless personal area network
  • WI-MAX wireless personal area network
  • WI-MAX wireless personal area network
  • cellular telephone network such as, for example, a Global System for Mobile Communications (GSM) network
  • GSM Global System
  • bus 412 includes hardware, software, or both coupling components of computer system 400 to each other.
  • bus 412 may include an Accelerated Graphics Port (AGP) or other graphics bus, an Enhanced Industry Standard Architecture (EISA) bus, a front-side bus (FSB), a HYPERTRANSPORT (HT) interconnect, an Industry Standard Architecture (ISA) bus, an INFINIBAND interconnect, a low-pin-count (LPC) bus, a memory bus, a Micro Channel Architecture (MCA) bus, a Peripheral Component Interconnect (PCI) bus, a PCI-Express (PCIe) bus, a serial advanced technology attachment (SATA) bus, a Video Electronics Standards Association local (VLB) bus, or another suitable bus or a combination of two or more of these.
  • Bus 412 may include one or more buses 412 , where appropriate.
  • a computer-readable non-transitory storage medium or media may include one or more semiconductor-based or other integrated circuits (ICs) (such, as for example, field-programmable gate arrays (FPGAs) or application-specific ICs (ASICs)), hard disk drives (HDDs), hybrid hard drives (HHDs), optical discs, optical disc drives (ODDs), magneto-optical discs, magneto-optical drives, floppy diskettes, floppy disk drives (FDDs), magnetic tapes, solid-state drives (SSDs), RAM-drives, SECURE DIGITAL cards or drives, any other suitable computer-readable non-transitory storage media, or any suitable combination of two or more of these, where appropriate.
  • ICs such, as for example, field-programmable gate arrays (FPGAs) or application-specific ICs (ASICs)
  • HDDs hard disk drives
  • HHDs hybrid hard drives
  • ODDs optical disc drives
  • magneto-optical discs magneto-optical drives
  • an apparatus or system or a component of an apparatus or system being adapted to, arranged to, capable of, configured to, enabled to, operable to, or operative to perform a particular function encompasses that apparatus, system, component, whether or not it or that particular function is activated, turned on, or unlocked, as long as that apparatus, system, or component is so adapted, arranged, capable, configured, enabled, operable, or operative.

Abstract

In one embodiment, a method includes accessing an electronic image comprising a surface area and dividing the electronic image into a plurality of surfaces. The method further includes determining that one or more of the surfaces comprise a type of graphics, and determining a percentage of the surface area of the image that is occupied by the one or more surfaces determined to comprise the type of graphics.

Description

    TECHNICAL FIELD
  • This disclosure generally relates to electronic image analysis.
  • BACKGROUND
  • Electronic images are presented as surfaces having graphical content. Electronic images may be stored in any suitable electronic storage medium, such as a harddrive, flash drive, etc and may be stored according to any suitable format, such as JPEG, TIFF, GIF, etc.. Electronic images may be presented to a viewer using any suitable display, such as on a display on a personal computer, mobile device, touchscreen, kiosk, television screen, etc. Electronic images may have variable characteristics such as size, shape, resolution, color schemes used, etc, and those characteristics may be varied for presentation on different types of displays or in different contexts, windows, etc. on a single display.
  • SUMMARY OF PARTICULAR EMBODIMENTS
  • An electronic image, such as an advertisement, may contain a type of graphics such as text, shapes, etc. An example method may access the electronic image and divide the image into a number of surfaces. The surfaces may have any suitable shape or size. A determination is made about whether the type of graphics is present on one or more of the surfaces. For example, the determination may be made by any suitable image processing technique or based on input received from a user. A percentage of the image's total surface area occupied by the surfaces determined to contain the type of graphics is calculated. In particular embodiments, if that percentage exceeds a predetermined percentage, the type of graphics may be resized, rearranged, or removed such that the total surface area occupied by the surfaces containing the type of graphics is less than the predetermined percentage.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 illustrates an example network environment associated with a social-networking system.
  • FIG. 2 illustrates an example method for determining the percentage of an electronic image occupied by a type of graphics.
  • FIG. 3 illustrates an example of determining the percentage of an electronic advertisement occupied by text.
  • FIG. 4 illustrates an example computer system.
  • DESCRIPTION OF EXAMPLE EMBODIMENTS
  • FIG. 1 illustrates an example network environment 100 associated with a social-networking system. Network environment 100 includes a user 101, a client system 130, a social-networking system 160, and a third-party system 170 connected to each other by a network 110. Although FIG. 1 illustrates a particular arrangement of user 101, client system 130, social-networking system 160, third-party system 170, and network 110, this disclosure contemplates any suitable arrangement of user 101, client system 130, social-networking system 160, third-party system 170, and network 110. As an example and not by way of limitation, two or more of client system 130, social-networking system 160, and third-party system 170 may be connected to each other directly, bypassing network 110. As another example, two or more of client system 130, social-networking system 160, and third-party system 170 may be physically or logically co-located with each other in whole or in part. Moreover, although FIG. 1 illustrates a particular number of users 101, client systems 130, social-networking systems 160, third-party systems 170, and networks 110, this disclosure contemplates any suitable number of users 101, client systems 130, social-networking systems 160, third-party systems 170, and networks 110. As an example and not by way of limitation, network environment 100 may include multiple users 101, client system 130, social-networking systems 160, third-party systems 170, and networks 110.
  • In particular embodiments, user 101 may be an individual (human user), an entity (e.g., an enterprise, business, or third-party application), or a group (e.g., of individuals or entities) that interacts or communicates with or over social-networking system 160. In particular embodiments, social-networking system 160 may be a network-addressable computing system hosting an online social network. Social-networking system 160 may generate, store, receive, and send social-networking data, such as, for example, user-profile data, concept-profile data, social-graph information, or other suitable data related to the online social network. Social-networking system 160 may be accessed by the other components of network environment 100 either directly or via network 110. In particular embodiments, social-networking system 160 may include an authorization server (or other suitable component(s)) that allows users 101 to opt in to or opt out of having their actions logged by social-networking system 160 or shared with other systems (e.g., third-party systems 170), for example, by setting appropriate privacy settings. A privacy setting of a user may determine what information associated with the user may be logged, how information associated with the user may be logged, when information associated with the user may be logged, who may log information associated with the user, whom information associated with the user may be shared with, and for what purposes information associated with the user may be logged or shared. Authorization servers may be used to enforce one or more privacy settings of the users of social-networking system 30 through blocking, data hashing, anonymization, or other suitable techniques as appropriate. Third-party system 170 may generate, store, receive, and send data, such as, for example, an advertisement. Third-party system 170 may be accessed by the other components of network environment 100 either directly or via network 110. In particular embodiments, one or more users 101 may use one or more client systems 130 to access, send data to, and receive data from social-networking system 160 or third-party system 170. Client system 130 may access social-networking system 160 or third-party system 170 directly, via network 110, or via a third-party system. As an example and not by way of limitation, client system 130 may access third-party system 170 via social-networking system 160. Client system 130 may be any suitable computing device, such as, for example, a personal computer, a laptop computer, a cellular telephone, a smartphone, or a tablet computer.
  • This disclosure contemplates any suitable network 110. As an example and not by way of limitation, one or more portions of network 110 may include an ad hoc network, an intranet, an extranet, a virtual private network (VPN), a local area network (LAN), a wireless LAN (WLAN), a wide area network (WAN), a wireless WAN (WWAN), a metropolitan area network (MAN), a portion of the Internet, a portion of the Public Switched Telephone Network (PSTN), a cellular telephone network, or a combination of two or more of these. Network 110 may include one or more networks 110.
  • Links 150 may connect client system 130, social-networking system 160, and third-party system 170 to communication network 110 or to each other. This disclosure contemplates any suitable links 150. In particular embodiments, one or more links 150 include one or more wireline (such as for example Digital Subscriber Line (DSL) or Data Over Cable Service Interface Specification (DOCSIS)), wireless (such as for example Wi-Fi or Worldwide Interoperability for Microwave Access (WiMAX)), or optical (such as for example Synchronous Optical Network (SONET) or Synchronous Digital Hierarchy (SDH)) links. In particular embodiments, one or more links 150 each include an ad hoc network, an intranet, an extranet, a VPN, a LAN, a WLAN, a WAN, a WWAN, a MAN, a portion of the Internet, a portion of the PSTN, a cellular technology-based network, a satellite communications technology-based network, another link 150, or a combination of two or more such links 150. Links 150 need not necessarily be the same throughout network environment 100. One or more first links 150 may differ in one or more respects from one or more second links 150.
  • In particular embodiments, an advertisement may be text (which may be HTML-linked), one or more images (which may be HTML-linked), one or more videos, audio, one or more ADOBE FLASH files, a suitable combination of these, or any other suitable advertisement in any suitable digital format presented on one or more webpages, in one or more e-mails, or in connection with search results requested by a user. In addition or as an alternative, an advertisement may be one or more sponsored stories (e.g., a news-feed or ticker item on social-networking system 160). A sponsored story may be a social action by a user (such as “liking” a page, “liking” or commenting on a post on a page, RSVPing to an event associated with a page, voting on a question posted on a page, checking in to a place, using an application or playing a game, or “liking” or sharing a website) that an advertiser promotes, for example, by having the social action presented within a pre-determined area of a profile page of a user or other page, presented with additional information associated with the advertiser, bumped up or otherwise highlighted within news feeds or tickers of other users, or otherwise promoted. The advertiser may pay to have the social action promoted. As an example and not by way of limitation, advertisements may be included among the search results of a search-results page, where sponsored content is promoted over non-sponsored content.
  • In particular embodiments, an advertisement may be requested for display within social-networking-system webpages, third-party webpages, or other pages. An advertisement may be displayed in a dedicated portion of a page, such as in a banner area at the top of the page, in a column at the side of the page, in a GUI of the page, in a pop-up window, in a drop-down menu, in an input field of the page, over the top of content of the page, or elsewhere with respect to the page. In addition or as an alternative, an advertisement may be displayed within an application. An advertisement may be displayed within dedicated pages, requiring the user to interact with or watch the advertisement before the user may access a page or utilize an application. The user may, for example view the advertisement through a web browser.
  • FIG. 2 illustrates an example method 200 for determining the percentage of an electronic image occupied by a type of graphics. The method may begin at step 210, where an electronic image is accessed. Electronic images may be accessed from any suitable non-volatile storage medium, such as a harddrive, flash drive, ROM, etc. Electronic images may accessed from any suitable volatile storage media, such as RAM, caches, buffers, etc. In particular embodiments, surfaces may be represented by identifying the memory elements describing the portion of an image the surface comprises. Electronic images may be stored according to any suitable format, such as JPEG, TIFF, GIF, PNG, etc. Electronic images may be of any suitable type, such as photos or advertisements, and may include any suitable graphical content such as text, colors, shapes, etc.
  • At step 220, the electronic image is divided into a plurality of surfaces. The surfaces (and thus, the divisions) may take any suitable shape and be of any suitable size. In particular embodiments, the surfaces may be non-overlapping and/or may collectively span the surface area of the image. For example, the surfaces may be squares, rectangles, hexagons, polygons, ellipses, or any suitable combination thereof. In particular embodiments, the surfaces and corresponding divisions may be presented as divided to a user. For example, the divisions may be represented as a grid (e.g. by lines or other suitable markings), and the grid and surfaces may be presented to the user. In particular embodiments, one surface of an image may have a different size or shape than another surface of the image. In particular embodiments, divisions or the size, shape, or locations of one or more surfaces may be determined by identifying or estimating the presence of particular content or types of graphics in the image, as explained more fully below. This disclosure contemplates dividing an image into any suitable type, number, or size of surfaces, and representing those surfaces and/or divisions in any suitable way.
  • At step 230, a determination is made whether a type of graphics exists on at least one of the surfaces. This disclosure contemplates any suitable type of graphics, such as text, colors, shapes, etc. In particular embodiments, the determination may be made based on input received from a user. For example, the surfaces may be presented to the user and the user may indicate (e.g. by clicking using a cursor, dragging using a cursor, touching on a touch-sensitive portion of the display, etc.) which surfaces contain a type of graphics. In particular embodiments, the number and size of surfaces may be such that the user can make provide input quickly or within a predetermined amount of time, for example by using a uniform grid of rectangles each having a height and width of at least one inch or by using a uniform grid of a predetermined number of squares, such as a 5×5 grid. In particular embodiments, the determination may be made by a computing device, for example by using one or more image-processing techniques, such as optical character recognition (OCR) or line recognition for text, or techniques (such as Fourier transforms) for identifying shapes, patterns, etc, or characteristics of the same. In particular embodiments, a determination may be made by a combination of user input and a computing device. For example, a user's determination of which surfaces contain a type of graphics may be checked by image processing techniques, and vice versa. In particular embodiments, the divisions described above may be altered by or based on estimates of which portions of an image contain a type of graphics. As an example, a user may identify where a type of graphics is present or identify regions containing the type of graphics, and that identification may be used to determine the size, shape, or location of one or more surfaces. For example, a surface may be enlarged or made smaller to more-accurately represent the shape or location of at least part of the particular type of graphics, such as by closely matching a surface to the area occupied by the type of graphics. As another example, a computing device may estimate where a type of graphics is present, for example by using image-processing techniques as described above, and those estimates may be used to determine the size, shape, or location of one or more surfaces. As another example, portions of an image where a type of graphics is likely to occur (e.g. based on analysis of similar images) may have pre-sized and/or pre-shaped surfaces. This disclosure contemplates determining that one or more surfaces comprise a type of graphics by any suitable technique.
  • At step 240, the percentage of the image's surface area occupied by surfaces determined to include the type of graphics is calculated. This disclosure contemplates any suitable method of calculating the percentage, such as adding the surface areas of the surfaces and dividing by the surface area of the image, by adding each surface area of a surface divided by the surface area of the image, dividing the number of surfaces containing the type of graphics by the total number of surfaces the image is divided into (if the surfaces are of uniform surface area), etc. In particular embodiments, the percentage may be compared to a predetermined percentage, and some action may be taken as a result. As an example, the image may be an advertisement and the type of graphics may be text. If the surface area of surfaces containing text is a greater percentage of the advertisement's surface area than a pre-determined percentage, the advertisement may be prohibited from being displayed, and the advertiser or publisher may be notified that the advertisement does not comply with rules governing the area of an advertisement text is permitted to cover. In addition or the alternative, if the surface area of surfaces containing text is a greater percentage of the advertisement's surface area than a pre-determined percentage, at least some of the text may be rearranged, resized, and/or removed. For example, the text may iteratively rearranged, resized, and/or removed until the surface area of surfaces containing text is not a greater percentage of the advertisement's surface area than a pre-determined percentage, or a calculation may be made regarding one or more configurations of text that would result in the surface area of surfaces containing text to not be a greater percentage of the advertisement's surface area than a predetermined percentage, and the text may be configured to one of the calculated configurations. As another example, a user may select all or a portion of the text and indicate a size, shape, or location of the text or a size, shape, or location of one or more surfaces to move the text onto. While this disclosure describes particular embodiments in the context of text on advertisements, this disclosure applies to any suitable type of graphics on any suitable type of electronic image.
  • Particular embodiments may repeat one or more steps of the method of FIG. 2, where appropriate. Although this disclosure describes and illustrates particular steps of the method of FIG. 2 as occurring in a particular order, this disclosure contemplates any suitable steps of the method of FIG. 2 occurring in any suitable order. Moreover, although this disclosure describes and illustrates particular components, devices, or systems carrying out particular steps of the method of FIG. 2, this disclosure contemplates any suitable combination of any suitable components, devices, or systems carrying out any suitable steps of the method of FIG. 2.
  • In particular embodiments, a determination (whether automated, based on manual input, or some combination thereof) about whether a surface contains a type of graphics may be used to train automated methods used to determine or aid in a determination of whether a surface contains a type of graphics. For example, a determination may be used to train one or more image-processing tools such as OCR, line detection, etc. In particular embodiments, machine-learning techniques such as decision trees, neural networks, clustering, reinforcement learning, etc. may be used to facilitate the training process.
  • FIG. 3 illustrates a specific example for determining the percentage of an electronic advertisement occupied by text (as described more fully in step 230, above). As illustrated, device 300 displays an advertisement 310 on a display. The advertisement is divided into a 5×5 grid of rectangular surfaces 320, and the divisions are indicated by gridlines 330. The user selects the rectangular surfaces where text appears, such as rectangular surface 340. The user may select the rectangular surfaces by any suitable method, such as by using a cursor, tapping the squares on a touch-sensitive display, swiping the squares on a touch-sensitive display, etc. In particular embodiments, selected surfaces may be presented differently than unselected surfaces, for example by shading the selected surfaces, coloring the selected surfaces, etc. As illustrated by the example of FIG. 3, in particular embodiments a type of graphics may be determined not to be a on a surface if only a small amount of the type of graphics is on the surface or the type of graphics occupies only a small amount of surface area of the surface. For example, surface 350 contains a small portion of text, and the user may not indicate that the surface contains text.
  • FIG. 4 illustrates an example computer system 400. In particular embodiments, one or more computer systems 400 perform one or more steps of one or more methods described or illustrated herein. In particular embodiments, one or more computer systems 400 provide functionality described or illustrated herein. In particular embodiments, software running on one or more computer systems 400 performs one or more steps of one or more methods described or illustrated herein or provides functionality described or illustrated herein. Particular embodiments include one or more portions of one or more computer systems 400. Herein, reference to a computer system may encompass a computing device, and vice versa, where appropriate. Moreover, reference to a computer system may encompass one or more computer systems, where appropriate.
  • This disclosure contemplates any suitable number of computer systems 400. This disclosure contemplates computer system 400 taking any suitable physical form. As example and not by way of limitation, computer system 400 may be an embedded computer system, a system-on-chip (SOC), a single-board computer system (SB) (such as, for example, a computer-on-module (COM) or system-on-module (SOM)), a desktop computer system, a laptop or notebook computer system, an interactive kiosk, a mainframe, a mesh of computer systems, a mobile telephone, a personal digital assistant (PDA), a server, a tablet computer system, or a combination of two or more of these. Where appropriate, computer system 400 may include one or more computer systems 400; be unitary or distributed; span multiple locations; span multiple machines; span multiple data centers; or reside in a cloud, which may include one or more cloud components in one or more networks. Where appropriate, one or more computer systems 400 may perform without substantial spatial or temporal limitation one or more steps of one or more methods described or illustrated herein. As an example and not by way of limitation, one or more computer systems 400 may perform in real time or in batch mode one or more steps of one or more methods described or illustrated herein. One or more computer systems 400 may perform at different times or at different locations one or more steps of one or more methods described or illustrated herein, where appropriate.
  • In particular embodiments, computer system 400 includes a processor 402, memory 404, storage 406, an input/output (I/O) interface 408, a communication interface 410, and a bus 412. Although this disclosure describes and illustrates a particular computer system having a particular number of particular components in a particular arrangement, this disclosure contemplates any suitable computer system having any suitable number of any suitable components in any suitable arrangement.
  • In particular embodiments, processor 402 includes hardware for executing instructions, such as those making up a computer program. As an example and not by way of limitation, to execute instructions, processor 402 may retrieve (or fetch) the instructions from an internal register, an internal cache, memory 404, or storage 406; decode and execute them; and then write one or more results to an internal register, an internal cache, memory 404, or storage 406. In particular embodiments, processor 402 may include one or more internal caches for data, instructions, or addresses. This disclosure contemplates processor 402 including any suitable number of any suitable internal caches, where appropriate. As an example and not by way of limitation, processor 402 may include one or more instruction caches, one or more data caches, and one or more translation lookaside buffers (TLBs). Instructions in the instruction caches may be copies of instructions in memory 404 or storage 406, and the instruction caches may speed up retrieval of those instructions by processor 402. Data in the data caches may be copies of data in memory 404 or storage 406 for instructions executing at processor 402 to operate on; the results of previous instructions executed at processor 402 for access by subsequent instructions executing at processor 402 or for writing to memory 404 or storage 406; or other suitable data. The data caches may speed up read or write operations by processor 402. The TLBs may speed up virtual-address translation for processor 402. In particular embodiments, processor 402 may include one or more internal registers for data, instructions, or addresses. This disclosure contemplates processor 402 including any suitable number of any suitable internal registers, where appropriate. Where appropriate, processor 402 may include one or more arithmetic logic units (ALUs); be a multi-core processor; or include one or more processors 402. Although this disclosure describes and illustrates a particular processor, this disclosure contemplates any suitable processor.
  • In particular embodiments, memory 404 includes main memory for storing instructions for processor 402 to execute or data for processor 402 to operate on. As an example and not by way of limitation, computer system 400 may load instructions from storage 406 or another source (such as, for example, another computer system 400) to memory 404. Processor 402 may then load the instructions from memory 404 to an internal register or internal cache. To execute the instructions, processor 402 may retrieve the instructions from the internal register or internal cache and decode them. During or after execution of the instructions, processor 402 may write one or more results (which may be intermediate or final results) to the internal register or internal cache. Processor 402 may then write one or more of those results to memory 404. In particular embodiments, processor 402 executes only instructions in one or more internal registers or internal caches or in memory 404 (as opposed to storage 406 or elsewhere) and operates only on data in one or more internal registers or internal caches or in memory 404 (as opposed to storage 406 or elsewhere). One or more memory buses (which may each include an address bus and a data bus) may couple processor 402 to memory 404. Bus 412 may include one or more memory buses, as described below. In particular embodiments, one or more memory management units (MMUs) reside between processor 402 and memory 404 and facilitate accesses to memory 404 requested by processor 402. In particular embodiments, memory 404 includes random access memory (RAM). This RAM may be volatile memory, where appropriate Where appropriate, this RAM may be dynamic RAM (DRAM) or static RAM (SRAM). Moreover, where appropriate, this RAM may be single-ported or multi-ported RAM. This disclosure contemplates any suitable RAM. Memory 404 may include one or more memories 404, where appropriate. Although this disclosure describes and illustrates particular memory, this disclosure contemplates any suitable memory.
  • In particular embodiments, storage 406 includes mass storage for data or instructions. As an example and not by way of limitation, storage 406 may include a hard disk drive (HDD), a floppy disk drive, flash memory, an optical disc, a magneto-optical disc, magnetic tape, or a Universal Serial Bus (USB) drive or a combination of two or more of these. Storage 406 may include removable or non-removable (or fixed) media, where appropriate. Storage 406 may be internal or external to computer system 400, where appropriate. In particular embodiments, storage 406 is non-volatile, solid-state memory. In particular embodiments, storage 406 includes read-only memory (ROM). Where appropriate, this ROM may be mask-programmed ROM, programmable ROM (PROM), erasable PROM (EPROM), electrically erasable PROM (EEPROM), electrically alterable ROM (EAROM), or flash memory or a combination of two or more of these. This disclosure contemplates mass storage 406 taking any suitable physical form. Storage 406 may include one or more storage control units facilitating communication between processor 402 and storage 406, where appropriate. Where appropriate, storage 406 may include one or more storages 406. Although this disclosure describes and illustrates particular storage, this disclosure contemplates any suitable storage.
  • In particular embodiments, I/O interface 408 includes hardware, software, or both, providing one or more interfaces for communication between computer system 400 and one or more I/O devices. Computer system 400 may include one or more of these I/O devices, where appropriate. One or more of these I/O devices may enable communication between a person and computer system 400. As an example and not by way of limitation, an I/O device may include a keyboard, keypad, microphone, monitor, mouse, printer, scanner, speaker, still camera, stylus, tablet, touch screen, trackball, video camera, another suitable I/O device or a combination of two or more of these. An I/O device may include one or more sensors. This disclosure contemplates any suitable I/O devices and any suitable I/O interfaces 408 for them. Where appropriate, I/O interface 408 may include one or more device or software drivers enabling processor 402 to drive one or more of these I/O devices. I/O interface 408 may include one or more I/O interfaces 408, where appropriate. Although this disclosure describes and illustrates a particular I/O interface, this disclosure contemplates any suitable I/O interface.
  • In particular embodiments, communication interface 410 includes hardware, software, or both providing one or more interfaces for communication (such as, for example, packet-based communication) between computer system 400 and one or more other computer systems 400 or one or more networks. As an example and not by way of limitation, communication interface 410 may include a network interface controller (NIC) or network adapter for communicating with an Ethernet or other wire-based network or a wireless NIC (WNIC) or wireless adapter for communicating with a wireless network, such as a WI-FI network. This disclosure contemplates any suitable network and any suitable communication interface 410 for it. As an example and not by way of limitation, computer system 400 may communicate with an ad hoc network, a personal area network (PAN), a local area network (LAN), a wide area network (WAN), a metropolitan area network (MAN), or one or more portions of the Internet or a combination of two or more of these. One or more portions of one or more of these networks may be wired or wireless. As an example, computer system 400 may communicate with a wireless PAN (WPAN) (such as, for example, a BLUETOOTH WPAN), a WI-FI network, a WI-MAX network, a cellular telephone network (such as, for example, a Global System for Mobile Communications (GSM) network), or other suitable wireless network or a combination of two or more of these. Computer system 400 may include any suitable communication interface 410 for any of these networks, where appropriate. Communication interface 410 may include one or more communication interfaces 410, where appropriate. Although this disclosure describes and illustrates a particular communication interface, this disclosure contemplates any suitable communication interface.
  • In particular embodiments, bus 412 includes hardware, software, or both coupling components of computer system 400 to each other. As an example and not by way of limitation, bus 412 may include an Accelerated Graphics Port (AGP) or other graphics bus, an Enhanced Industry Standard Architecture (EISA) bus, a front-side bus (FSB), a HYPERTRANSPORT (HT) interconnect, an Industry Standard Architecture (ISA) bus, an INFINIBAND interconnect, a low-pin-count (LPC) bus, a memory bus, a Micro Channel Architecture (MCA) bus, a Peripheral Component Interconnect (PCI) bus, a PCI-Express (PCIe) bus, a serial advanced technology attachment (SATA) bus, a Video Electronics Standards Association local (VLB) bus, or another suitable bus or a combination of two or more of these. Bus 412 may include one or more buses 412, where appropriate. Although this disclosure describes and illustrates a particular bus, this disclosure contemplates any suitable bus or interconnect.
  • Herein, a computer-readable non-transitory storage medium or media may include one or more semiconductor-based or other integrated circuits (ICs) (such, as for example, field-programmable gate arrays (FPGAs) or application-specific ICs (ASICs)), hard disk drives (HDDs), hybrid hard drives (HHDs), optical discs, optical disc drives (ODDs), magneto-optical discs, magneto-optical drives, floppy diskettes, floppy disk drives (FDDs), magnetic tapes, solid-state drives (SSDs), RAM-drives, SECURE DIGITAL cards or drives, any other suitable computer-readable non-transitory storage media, or any suitable combination of two or more of these, where appropriate. A computer-readable non-transitory storage medium may be volatile, non-volatile, or a combination of volatile and non-volatile, where appropriate.
  • Herein, “or” is inclusive and not exclusive, unless expressly indicated otherwise or indicated otherwise by context. Therefore, herein, “A or B” means “A, B, or both,” unless expressly indicated otherwise or indicated otherwise by context. Moreover, “and” is both joint and several, unless expressly indicated otherwise or indicated otherwise by context. Therefore, herein, “A and B” means “A and B, jointly or severally,” unless expressly indicated otherwise or indicated otherwise by context.
  • The scope of this disclosure encompasses all changes, substitutions, variations, alterations, and modifications to the example embodiments described or illustrated herein that a person having ordinary skill in the art would comprehend. The scope of this disclosure is not limited to the example embodiments described or illustrated herein. Moreover, although this disclosure describes and illustrates respective embodiments herein as including particular components, elements, functions, operations, or steps, any of these embodiments may include any combination or permutation of any of the components, elements, functions, operations, or steps described or illustrated anywhere herein that a person having ordinary skill in the art would comprehend. Furthermore, reference in the appended claims to an apparatus or system or a component of an apparatus or system being adapted to, arranged to, capable of, configured to, enabled to, operable to, or operative to perform a particular function encompasses that apparatus, system, component, whether or not it or that particular function is activated, turned on, or unlocked, as long as that apparatus, system, or component is so adapted, arranged, capable, configured, enabled, operable, or operative.

Claims (20)

What is claimed is:
1. A method comprising:
by one or more computing devices, accessing an electronic image comprising a surface area;
by one or more computing devices, dividing the electronic image into a plurality of surfaces;
by one or more computing devices, determining that one or more of the surfaces comprise a type of graphics; and
by one or more computing devices, determining a percentage of the surface area of the image that is occupied by the one or more surfaces determined to comprise the type of graphics.
2. The method of claim 1, wherein determining that one or more of the surfaces comprise a type of graphics comprises:
displaying the divided electronic image to a user; and
receiving input from the user indicating that the one or more of the surfaces comprise the type of graphics.
3. The method of claim 1, further comprising:
by one or more computing devices, providing the determination that one or more of the surfaces comprise the type of graphics as training input to an image-processing algorithm.
4. The method of claim 1, wherein the type of graphics comprises text.
5. The method of claim 4, wherein determining that one or more of the surfaces comprises text comprises:
determining whether the amount of text comprising at least one of the one or more surfaces is greater than a predetermined amount; and
when the amount of text is greater than the predetermined amount, determining that the one of the one or more surfaces comprises text.
6. The method of claim 1, wherein the image comprises an advertisement.
7. The method of claim 6, further comprising determining whether the percentage of the surface area of the advertisement that is occupied by the one or more surfaces determined to comprise the type of graphics exceeds a predetermined percentage, wherein the type of graphics comprises text.
8. The method of claim 7, further comprising:
by one or more computing devices, prohibiting the advertisement from display to a user when the percentage of the surface area of the advertisement that is occupied by the one or more surfaces determined to comprise text exceeds the predetermined percentage.
9. The method of claim 7, further comprising:
by one or more computing devices, adjusting the size or location of at least some of the text when the percentage of the surface area of the advertisement that is occupied by the one or more surfaces determined to comprise text exceeds the predetermined percentage.
10. The method of claim 1, wherein at least one of the plurality of surfaces has a different size or shape than at least one other one of the plurality of surfaces.
11. The method of claim 1, wherein a size or shape of at least one of the plurality of surfaces is determined based on an estimate of a size or shape of at least part of the type of graphics the surface comprises.
12. One or more computer-readable non-transitory storage media embodying software that is operable when executed to:
access an electronic image comprising a surface area;
divide the electronic image into a plurality of surfaces;
determine that one or more of the surfaces comprise a type of graphics; and
determine a percentage of the surface area of the image that is occupied by the one or more surfaces determined to comprise the type of graphics.
13. The media of claim 12, wherein the software that is operable when executed to determine that one or more of the surfaces comprise a type of graphics comprises software that is operable when executed to:
display the divided electronic image to a user; and
receive input from the user indicating that the one or more of the surfaces comprise the type of graphics.
14. The media of claim 12, wherein the software is further operable when executed to:
provide the determination that one or more of the surfaces comprise the type of graphics as training input to an image-processing algorithm.
15. The media of claim 12, wherein the type of graphics comprises text.
16. The media of claim 15, wherein the software that is operable when executed to determine that one or more of the surfaces comprise text comprises software that is operable when executed to:
determine whether the amount of text comprising at least one of the one or more surfaces is greater than a predetermined amount; and
when the amount of text is greater than the predetermined amount, determine that the one of the one or more surfaces comprises text.
17. The media of claim 12, wherein the image comprises an advertisement.
18. The media of claim 17, wherein the software is further operable when executed to:
determine whether the percentage of the surface area of the advertisement that is occupied by the one or more surfaces determined to comprise the type of graphics exceeds a predetermined percentage, wherein the type of graphics comprises text.
19. The media of claim 18, wherein the software is further operable when executed to:
prohibit the advertisement from display to a user when the percentage of the surface area of the advertisement that is occupied by the one or more surfaces determined to comprise text exceeds the predetermined percentage.
20. A system comprising:
one or more processors; and
a memory coupled to the processors comprising instructions executable by the processors, the processors being operable when executing the instructions to:
access an electronic image comprising a surface area;
divide the electronic image into a plurality of surfaces;
determine that one or more of the surfaces comprise a type of graphics; and
determine a percentage of the surface area of the image that is occupied by the one or more surfaces determined to comprise the type of graphics.
US13/862,212 2013-04-12 2013-04-12 Identifying content in electronic images Active US10296933B2 (en)

Priority Applications (11)

Application Number Priority Date Filing Date Title
US13/862,212 US10296933B2 (en) 2013-04-12 2013-04-12 Identifying content in electronic images
JP2016507563A JP6473135B2 (en) 2013-04-12 2014-04-02 Content identification in electronic images
CN201480033674.XA CN105408855B (en) 2013-04-12 2014-04-02 Identify the content in electronic image
KR1020157032308A KR102025615B1 (en) 2013-04-12 2014-04-02 Identifying content in electronic images
AU2014251221A AU2014251221B2 (en) 2013-04-12 2014-04-02 Identifying content in electronic images
MX2015014302A MX364910B (en) 2013-04-12 2014-04-02 Identifying content in electronic images.
BR112015025970A BR112015025970A2 (en) 2013-04-12 2014-04-02 content identification in electronic images
CA2908554A CA2908554A1 (en) 2013-04-12 2014-04-02 Identifying content in electronic images
PCT/US2014/032669 WO2014168795A1 (en) 2013-04-12 2014-04-02 Identifying content in electronic images
EP14783229.9A EP2984556A4 (en) 2013-04-12 2014-04-02 Identifying content in electronic images
JP2019010276A JP6744932B2 (en) 2013-04-12 2019-01-24 Identifying content in electronic images

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/862,212 US10296933B2 (en) 2013-04-12 2013-04-12 Identifying content in electronic images

Publications (2)

Publication Number Publication Date
US20140306986A1 true US20140306986A1 (en) 2014-10-16
US10296933B2 US10296933B2 (en) 2019-05-21

Family

ID=51686488

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/862,212 Active US10296933B2 (en) 2013-04-12 2013-04-12 Identifying content in electronic images

Country Status (10)

Country Link
US (1) US10296933B2 (en)
EP (1) EP2984556A4 (en)
JP (2) JP6473135B2 (en)
KR (1) KR102025615B1 (en)
CN (1) CN105408855B (en)
AU (1) AU2014251221B2 (en)
BR (1) BR112015025970A2 (en)
CA (1) CA2908554A1 (en)
MX (1) MX364910B (en)
WO (1) WO2014168795A1 (en)

Cited By (119)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170374003A1 (en) 2014-10-02 2017-12-28 Snapchat, Inc. Ephemeral gallery of ephemeral messages
US9866999B1 (en) 2014-01-12 2018-01-09 Investment Asset Holdings Llc Location-based messaging
US10123166B2 (en) 2015-01-26 2018-11-06 Snap Inc. Content request by location
US10154192B1 (en) 2014-07-07 2018-12-11 Snap Inc. Apparatus and method for supplying content aware photo filters
US10157449B1 (en) 2015-01-09 2018-12-18 Snap Inc. Geo-location-based image filters
US10165402B1 (en) 2016-06-28 2018-12-25 Snap Inc. System to track engagement of media items
US10182311B2 (en) 2014-06-13 2019-01-15 Snap Inc. Prioritization of messages within a message collection
US10203855B2 (en) 2016-12-09 2019-02-12 Snap Inc. Customized user-controlled media overlays
US10219111B1 (en) 2018-04-18 2019-02-26 Snap Inc. Visitation tracking system
US10223397B1 (en) 2015-03-13 2019-03-05 Snap Inc. Social graph based co-location of network users
US10319149B1 (en) 2017-02-17 2019-06-11 Snap Inc. Augmented reality anamorphosis system
US10327096B1 (en) 2018-03-06 2019-06-18 Snap Inc. Geo-fence selection system
US10334307B2 (en) 2011-07-12 2019-06-25 Snap Inc. Methods and systems of providing visual content editing functions
US10348662B2 (en) 2016-07-19 2019-07-09 Snap Inc. Generating customized electronic messaging graphics
US10354425B2 (en) 2015-12-18 2019-07-16 Snap Inc. Method and system for providing context relevant media augmentation
US10366543B1 (en) 2015-10-30 2019-07-30 Snap Inc. Image based tracking in augmented reality systems
US10387730B1 (en) 2017-04-20 2019-08-20 Snap Inc. Augmented reality typography personalization system
US10387514B1 (en) 2016-06-30 2019-08-20 Snap Inc. Automated content curation and communication
US10423983B2 (en) 2014-09-16 2019-09-24 Snap Inc. Determining targeting information based on a predictive targeting model
US10430838B1 (en) 2016-06-28 2019-10-01 Snap Inc. Methods and systems for generation, curation, and presentation of media collections with automated advertising
US10474321B2 (en) 2015-11-30 2019-11-12 Snap Inc. Network resource location linking and visual content sharing
US10499191B1 (en) 2017-10-09 2019-12-03 Snap Inc. Context sensitive presentation of content
US10523625B1 (en) 2017-03-09 2019-12-31 Snap Inc. Restricted group content collection
US10572681B1 (en) 2014-05-28 2020-02-25 Snap Inc. Apparatus and method for automated privacy protection in distributed images
US10580458B2 (en) 2014-12-19 2020-03-03 Snap Inc. Gallery of videos set to an audio time line
US10592574B2 (en) 2015-05-05 2020-03-17 Snap Inc. Systems and methods for automated local story generation and curation
US10616476B1 (en) 2014-11-12 2020-04-07 Snap Inc. User interface for accessing media at a geographic location
US10616239B2 (en) 2015-03-18 2020-04-07 Snap Inc. Geo-fence authorization provisioning
US10623666B2 (en) 2016-11-07 2020-04-14 Snap Inc. Selective identification and order of image modifiers
US10679393B2 (en) 2018-07-24 2020-06-09 Snap Inc. Conditional modification of augmented reality object
US10678818B2 (en) 2018-01-03 2020-06-09 Snap Inc. Tag distribution visualization system
US10679389B2 (en) 2016-02-26 2020-06-09 Snap Inc. Methods and systems for generation, curation, and presentation of media collections
US10740974B1 (en) 2017-09-15 2020-08-11 Snap Inc. Augmented reality system
US10817898B2 (en) 2015-08-13 2020-10-27 Placed, Llc Determining exposures to content presented by physical objects
US10824654B2 (en) * 2014-09-18 2020-11-03 Snap Inc. Geolocation-based pictographs
US10834525B2 (en) 2016-02-26 2020-11-10 Snap Inc. Generation, curation, and presentation of media collections
US10862951B1 (en) 2007-01-05 2020-12-08 Snap Inc. Real-time display of multiple images
US10885136B1 (en) 2018-02-28 2021-01-05 Snap Inc. Audience filtering system
US10911575B1 (en) 2015-05-05 2021-02-02 Snap Inc. Systems and methods for story and sub-story navigation
US10915911B2 (en) 2017-02-03 2021-02-09 Snap Inc. System to determine a price-schedule to distribute media content
US10933311B2 (en) 2018-03-14 2021-03-02 Snap Inc. Generating collectible items based on location information
US10948717B1 (en) 2015-03-23 2021-03-16 Snap Inc. Reducing boot time and power consumption in wearable display systems
US10952013B1 (en) 2017-04-27 2021-03-16 Snap Inc. Selective location-based identity communication
US10963529B1 (en) 2017-04-27 2021-03-30 Snap Inc. Location-based search mechanism in a graphical user interface
US10979752B1 (en) 2018-02-28 2021-04-13 Snap Inc. Generating media content items based on location information
US10993069B2 (en) 2015-07-16 2021-04-27 Snap Inc. Dynamically adaptive media content delivery
US10997783B2 (en) 2015-11-30 2021-05-04 Snap Inc. Image and point cloud based tracking and in augmented reality systems
US10997760B2 (en) 2018-08-31 2021-05-04 Snap Inc. Augmented reality anthropomorphization system
US11017173B1 (en) 2017-12-22 2021-05-25 Snap Inc. Named entity recognition visual context and caption data
US11023514B2 (en) 2016-02-26 2021-06-01 Snap Inc. Methods and systems for generation, curation, and presentation of media collections
US11030787B2 (en) 2017-10-30 2021-06-08 Snap Inc. Mobile-based cartographic control of display content
US11038829B1 (en) 2014-10-02 2021-06-15 Snap Inc. Ephemeral gallery of ephemeral messages with opt-in permanence
US11037372B2 (en) 2017-03-06 2021-06-15 Snap Inc. Virtual vision system
US11128715B1 (en) 2019-12-30 2021-09-21 Snap Inc. Physical friend proximity in chat
US11163941B1 (en) 2018-03-30 2021-11-02 Snap Inc. Annotating a collection of media content items
US11170393B1 (en) 2017-04-11 2021-11-09 Snap Inc. System to calculate an engagement score of location based media content
US11182383B1 (en) 2012-02-24 2021-11-23 Placed, Llc System and method for data collection to validate location data
US11189299B1 (en) 2017-02-20 2021-11-30 Snap Inc. Augmented reality speech balloon system
US11199957B1 (en) 2018-11-30 2021-12-14 Snap Inc. Generating customized avatars based on location information
US11206615B2 (en) 2019-05-30 2021-12-21 Snap Inc. Wearable device location systems
US11216869B2 (en) 2014-09-23 2022-01-04 Snap Inc. User interface to augment an image using geolocation
US11218838B2 (en) 2019-10-31 2022-01-04 Snap Inc. Focused map-based context information surfacing
US11228551B1 (en) 2020-02-12 2022-01-18 Snap Inc. Multiple gateway message exchange
US11232040B1 (en) 2017-04-28 2022-01-25 Snap Inc. Precaching unlockable data elements
US11249614B2 (en) 2019-03-28 2022-02-15 Snap Inc. Generating personalized map interface with enhanced icons
US11250075B1 (en) 2017-02-17 2022-02-15 Snap Inc. Searching social media content
US11249617B1 (en) 2015-01-19 2022-02-15 Snap Inc. Multichannel system
US11265273B1 (en) 2017-12-01 2022-03-01 Snap, Inc. Dynamic media overlay with smart widget
US11290851B2 (en) 2020-06-15 2022-03-29 Snap Inc. Location sharing using offline and online objects
US11297399B1 (en) 2017-03-27 2022-04-05 Snap Inc. Generating a stitched data stream
US11294936B1 (en) 2019-01-30 2022-04-05 Snap Inc. Adaptive spatial density based clustering
US11301117B2 (en) 2019-03-08 2022-04-12 Snap Inc. Contextual information in chat
US11314776B2 (en) 2020-06-15 2022-04-26 Snap Inc. Location sharing using friend list versions
US11343323B2 (en) 2019-12-31 2022-05-24 Snap Inc. Augmented reality objects registry
US11349796B2 (en) 2017-03-27 2022-05-31 Snap Inc. Generating a stitched data stream
US11361493B2 (en) 2019-04-01 2022-06-14 Snap Inc. Semantic texture mapping system
US11372608B2 (en) 2014-12-19 2022-06-28 Snap Inc. Gallery of messages from individuals with a shared interest
US11388226B1 (en) 2015-01-13 2022-07-12 Snap Inc. Guided personal identity based actions
US11430091B2 (en) 2020-03-27 2022-08-30 Snap Inc. Location mapping for large scale augmented-reality
US11429618B2 (en) 2019-12-30 2022-08-30 Snap Inc. Surfacing augmented reality objects
US11455082B2 (en) 2018-09-28 2022-09-27 Snap Inc. Collaborative achievement interface
US11475254B1 (en) 2017-09-08 2022-10-18 Snap Inc. Multimodal entity identification
US11483267B2 (en) 2020-06-15 2022-10-25 Snap Inc. Location sharing using different rate-limited links
US11500525B2 (en) 2019-02-25 2022-11-15 Snap Inc. Custom media overlay system
US11503432B2 (en) 2020-06-15 2022-11-15 Snap Inc. Scalable real-time location sharing framework
US11507614B1 (en) 2018-02-13 2022-11-22 Snap Inc. Icon based tagging
US11516167B2 (en) 2020-03-05 2022-11-29 Snap Inc. Storing data based on device location
US11558709B2 (en) 2018-11-30 2023-01-17 Snap Inc. Position service to determine relative position to map features
US11574431B2 (en) 2019-02-26 2023-02-07 Snap Inc. Avatar based on weather
US11601888B2 (en) 2021-03-29 2023-03-07 Snap Inc. Determining location using multi-source geolocation data
US11601783B2 (en) 2019-06-07 2023-03-07 Snap Inc. Detection of a physical collision between two client devices in a location sharing system
US11606755B2 (en) 2019-05-30 2023-03-14 Snap Inc. Wearable device location systems architecture
US11616745B2 (en) 2017-01-09 2023-03-28 Snap Inc. Contextual generation and selection of customized media content
US11619501B2 (en) 2020-03-11 2023-04-04 Snap Inc. Avatar based on trip
US11625443B2 (en) 2014-06-05 2023-04-11 Snap Inc. Web document enhancement
US11631276B2 (en) 2016-03-31 2023-04-18 Snap Inc. Automated avatar generation
US11645324B2 (en) 2021-03-31 2023-05-09 Snap Inc. Location-based timeline media content system
US11675831B2 (en) 2017-05-31 2023-06-13 Snap Inc. Geolocation based playlists
US11676378B2 (en) 2020-06-29 2023-06-13 Snap Inc. Providing travel-based augmented reality content with a captured image
US11714535B2 (en) 2019-07-11 2023-08-01 Snap Inc. Edge gesture interface with smart interactions
US11734712B2 (en) 2012-02-24 2023-08-22 Foursquare Labs, Inc. Attributing in-store visits to media consumption based on data collected from user devices
US11751015B2 (en) 2019-01-16 2023-09-05 Snap Inc. Location-based context information sharing in a messaging system
US11776256B2 (en) 2020-03-27 2023-10-03 Snap Inc. Shared augmented reality system
US11799811B2 (en) 2018-10-31 2023-10-24 Snap Inc. Messaging and gaming applications communication platform
US11809624B2 (en) 2019-02-13 2023-11-07 Snap Inc. Sleep detection in a location sharing system
US11816853B2 (en) 2016-08-30 2023-11-14 Snap Inc. Systems and methods for simultaneous localization and mapping
US11821742B2 (en) 2019-09-26 2023-11-21 Snap Inc. Travel based notifications
US11829834B2 (en) 2021-10-29 2023-11-28 Snap Inc. Extended QR code
US11843456B2 (en) 2016-10-24 2023-12-12 Snap Inc. Generating and displaying customized avatars in media overlays
US11842411B2 (en) 2017-04-27 2023-12-12 Snap Inc. Location-based virtual avatars
US11852554B1 (en) 2019-03-21 2023-12-26 Snap Inc. Barometer calibration in a location sharing system
US11860888B2 (en) 2018-05-22 2024-01-02 Snap Inc. Event detection system
US11868414B1 (en) 2019-03-14 2024-01-09 Snap Inc. Graph-based prediction for contact suggestion in a location sharing system
US11870743B1 (en) 2017-01-23 2024-01-09 Snap Inc. Customized digital avatar accessories
US11877211B2 (en) 2019-01-14 2024-01-16 Snap Inc. Destination sharing in location sharing system
US11893208B2 (en) 2019-12-31 2024-02-06 Snap Inc. Combined map icon with action indicator
US11925869B2 (en) 2012-05-08 2024-03-12 Snap Inc. System and method for generating and displaying avatars
US11943192B2 (en) 2020-08-31 2024-03-26 Snap Inc. Co-location connection service
US11954314B2 (en) 2022-09-09 2024-04-09 Snap Inc. Custom media overlay system

Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070027749A1 (en) * 2005-07-27 2007-02-01 Hewlett-Packard Development Company, L.P. Advertisement detection
US7203360B2 (en) * 2003-04-09 2007-04-10 Lee Shih-Jong J Learnable object segmentation
US20080231751A1 (en) * 2007-03-22 2008-09-25 Sony Computer Entertainment America Inc. Scheme for determining the locations and timing of advertisements and other insertions in media
US20080260248A1 (en) * 2007-04-23 2008-10-23 Sony Corporation Image processing apparatus, image processing method, and program
US20090150210A1 (en) * 2007-12-10 2009-06-11 Athellina Athsani Advertising associated with multimedia content
US20090171920A1 (en) * 2007-12-28 2009-07-02 Microsoft Corporation Adaptive page layout utilizing block-level elements
US20100100811A1 (en) * 2008-10-20 2010-04-22 Canon Kabushiki Kaisha Information processing apparatus and layout processing method
US20100157340A1 (en) * 2008-12-18 2010-06-24 Canon Kabushiki Kaisha Object extraction in colour compound documents
US7930647B2 (en) * 2005-12-11 2011-04-19 Topix Llc System and method for selecting pictures for presentation with text content
US20110091098A1 (en) * 2005-09-02 2011-04-21 Alan Yuille System and Method for Detecting Text in Real-World Color Images
US20110302524A1 (en) * 2010-06-04 2011-12-08 Scott Forstall Progress indicators for loading content
US20110302510A1 (en) * 2010-06-04 2011-12-08 David Frank Harrison Reader mode presentation of web content
US20110321084A1 (en) * 2010-06-25 2011-12-29 Kddi Corporation Apparatus and method for optimizing on-screen location of additional content overlay on video content
US20120159307A1 (en) * 2010-12-17 2012-06-21 Microsoft Corporation Rendering source regions into target regions of web pages
US20130058575A1 (en) * 2011-09-06 2013-03-07 Qualcomm Incorporated Text detection using image regions
US20130069924A1 (en) * 2011-09-01 2013-03-21 Research In Motion Limited Data display adapted for bright ambient light
US8655716B1 (en) * 2007-01-27 2014-02-18 Sprint Communications Company L.P. Integrated advertisement delivery
US8818913B1 (en) * 2004-01-14 2014-08-26 Junkin Holdings Llc Wireless access using preexisting data connection
US20140289614A1 (en) * 2005-05-16 2014-09-25 Adobe Systems Incorporated Object resizing with content reflow

Family Cites Families (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH06208625A (en) 1993-01-11 1994-07-26 Canon Inc Method and device for processing image
US7278117B2 (en) 2002-07-02 2007-10-02 Hewlett-Packard Development Company, L.P. Image segmentation and warping for specialized display viewing
AU2002952711A0 (en) * 2002-11-18 2002-11-28 Typefi Systems Pty Ltd A method of formatting documents
US20060103667A1 (en) * 2004-10-28 2006-05-18 Universal-Ad. Ltd. Method, system and computer readable code for automatic reize of product oriented advertisements
JP4661474B2 (en) 2005-09-13 2011-03-30 富士ゼロックス株式会社 Image processing apparatus, abnormality detection method, program, and recording medium
US9110903B2 (en) * 2006-11-22 2015-08-18 Yahoo! Inc. Method, system and apparatus for using user profile electronic device data in media delivery
CN101419661B (en) 2007-10-26 2011-08-24 国际商业机器公司 Method for displaying image based on text in image and system
US8799068B2 (en) * 2007-11-05 2014-08-05 Facebook, Inc. Social advertisements and other informational messages on a social networking website, and advertising model for same
JP5006263B2 (en) 2008-06-03 2012-08-22 株式会社リコー Image processing apparatus, program, and image processing method
US8189917B2 (en) 2008-09-25 2012-05-29 Sharp Laboratories Of America, Inc. Methods and systems for locating text in a digital image
US8914736B2 (en) * 2010-03-30 2014-12-16 International Business Machines Corporation On-page manipulation and real-time replacement of content
US8670597B2 (en) * 2009-08-07 2014-03-11 Google Inc. Facial recognition with social network aiding
US8352856B2 (en) 2009-11-11 2013-01-08 Xerox Corporation Systems and methods to resize document content
KR20130009754A (en) * 2010-02-01 2013-01-23 점프탭, 인크. Integrated advertising system
US20110313859A1 (en) 2010-06-22 2011-12-22 Microsoft Corporation Techniques for advertiser geotargeting using map coordinates
KR101577508B1 (en) * 2010-06-29 2015-12-15 라쿠텐 인코포레이티드 Information processing device, information processing method, and recording medium in which information processing program is recorded
KR101223784B1 (en) 2011-05-27 2013-01-21 송정욱 Apparatus, system and method for transforming image

Patent Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7203360B2 (en) * 2003-04-09 2007-04-10 Lee Shih-Jong J Learnable object segmentation
US8818913B1 (en) * 2004-01-14 2014-08-26 Junkin Holdings Llc Wireless access using preexisting data connection
US20140289614A1 (en) * 2005-05-16 2014-09-25 Adobe Systems Incorporated Object resizing with content reflow
US20070027749A1 (en) * 2005-07-27 2007-02-01 Hewlett-Packard Development Company, L.P. Advertisement detection
US20110091098A1 (en) * 2005-09-02 2011-04-21 Alan Yuille System and Method for Detecting Text in Real-World Color Images
US7930647B2 (en) * 2005-12-11 2011-04-19 Topix Llc System and method for selecting pictures for presentation with text content
US8655716B1 (en) * 2007-01-27 2014-02-18 Sprint Communications Company L.P. Integrated advertisement delivery
US20080231751A1 (en) * 2007-03-22 2008-09-25 Sony Computer Entertainment America Inc. Scheme for determining the locations and timing of advertisements and other insertions in media
US20080260248A1 (en) * 2007-04-23 2008-10-23 Sony Corporation Image processing apparatus, image processing method, and program
US20090150210A1 (en) * 2007-12-10 2009-06-11 Athellina Athsani Advertising associated with multimedia content
US20090171920A1 (en) * 2007-12-28 2009-07-02 Microsoft Corporation Adaptive page layout utilizing block-level elements
US20110010661A1 (en) * 2007-12-28 2011-01-13 Microsoft Corporation Adaptive page layout utilizing block-level elements
US20100100811A1 (en) * 2008-10-20 2010-04-22 Canon Kabushiki Kaisha Information processing apparatus and layout processing method
US20100157340A1 (en) * 2008-12-18 2010-06-24 Canon Kabushiki Kaisha Object extraction in colour compound documents
US20110302524A1 (en) * 2010-06-04 2011-12-08 Scott Forstall Progress indicators for loading content
US20110302510A1 (en) * 2010-06-04 2011-12-08 David Frank Harrison Reader mode presentation of web content
US20110321084A1 (en) * 2010-06-25 2011-12-29 Kddi Corporation Apparatus and method for optimizing on-screen location of additional content overlay on video content
US20120159307A1 (en) * 2010-12-17 2012-06-21 Microsoft Corporation Rendering source regions into target regions of web pages
US20130069924A1 (en) * 2011-09-01 2013-03-21 Research In Motion Limited Data display adapted for bright ambient light
US20130058575A1 (en) * 2011-09-06 2013-03-07 Qualcomm Incorporated Text detection using image regions

Cited By (255)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10862951B1 (en) 2007-01-05 2020-12-08 Snap Inc. Real-time display of multiple images
US11588770B2 (en) 2007-01-05 2023-02-21 Snap Inc. Real-time display of multiple images
US11750875B2 (en) 2011-07-12 2023-09-05 Snap Inc. Providing visual content editing functions
US10999623B2 (en) 2011-07-12 2021-05-04 Snap Inc. Providing visual content editing functions
US10334307B2 (en) 2011-07-12 2019-06-25 Snap Inc. Methods and systems of providing visual content editing functions
US11451856B2 (en) 2011-07-12 2022-09-20 Snap Inc. Providing visual content editing functions
US11182383B1 (en) 2012-02-24 2021-11-23 Placed, Llc System and method for data collection to validate location data
US11734712B2 (en) 2012-02-24 2023-08-22 Foursquare Labs, Inc. Attributing in-store visits to media consumption based on data collected from user devices
US11925869B2 (en) 2012-05-08 2024-03-12 Snap Inc. System and method for generating and displaying avatars
US10349209B1 (en) 2014-01-12 2019-07-09 Investment Asset Holdings Llc Location-based messaging
US10080102B1 (en) 2014-01-12 2018-09-18 Investment Asset Holdings Llc Location-based messaging
US9866999B1 (en) 2014-01-12 2018-01-09 Investment Asset Holdings Llc Location-based messaging
US10990697B2 (en) 2014-05-28 2021-04-27 Snap Inc. Apparatus and method for automated privacy protection in distributed images
US10572681B1 (en) 2014-05-28 2020-02-25 Snap Inc. Apparatus and method for automated privacy protection in distributed images
US11625443B2 (en) 2014-06-05 2023-04-11 Snap Inc. Web document enhancement
US11921805B2 (en) 2014-06-05 2024-03-05 Snap Inc. Web document enhancement
US10200813B1 (en) 2014-06-13 2019-02-05 Snap Inc. Geo-location based event gallery
US10659914B1 (en) 2014-06-13 2020-05-19 Snap Inc. Geo-location based event gallery
US11166121B2 (en) 2014-06-13 2021-11-02 Snap Inc. Prioritization of messages within a message collection
US10524087B1 (en) 2014-06-13 2019-12-31 Snap Inc. Message destination list mechanism
US10182311B2 (en) 2014-06-13 2019-01-15 Snap Inc. Prioritization of messages within a message collection
US10779113B2 (en) 2014-06-13 2020-09-15 Snap Inc. Prioritization of messages within a message collection
US11317240B2 (en) 2014-06-13 2022-04-26 Snap Inc. Geo-location based event gallery
US10448201B1 (en) 2014-06-13 2019-10-15 Snap Inc. Prioritization of messages within a message collection
US10623891B2 (en) 2014-06-13 2020-04-14 Snap Inc. Prioritization of messages within a message collection
US11595569B2 (en) 2014-07-07 2023-02-28 Snap Inc. Supplying content aware photo filters
US10432850B1 (en) 2014-07-07 2019-10-01 Snap Inc. Apparatus and method for supplying content aware photo filters
US11849214B2 (en) 2014-07-07 2023-12-19 Snap Inc. Apparatus and method for supplying content aware photo filters
US11122200B2 (en) 2014-07-07 2021-09-14 Snap Inc. Supplying content aware photo filters
US10602057B1 (en) 2014-07-07 2020-03-24 Snap Inc. Supplying content aware photo filters
US10154192B1 (en) 2014-07-07 2018-12-11 Snap Inc. Apparatus and method for supplying content aware photo filters
US10423983B2 (en) 2014-09-16 2019-09-24 Snap Inc. Determining targeting information based on a predictive targeting model
US11625755B1 (en) 2014-09-16 2023-04-11 Foursquare Labs, Inc. Determining targeting information based on a predictive targeting model
US11741136B2 (en) * 2014-09-18 2023-08-29 Snap Inc. Geolocation-based pictographs
US10824654B2 (en) * 2014-09-18 2020-11-03 Snap Inc. Geolocation-based pictographs
US20220318281A1 (en) * 2014-09-18 2022-10-06 Snap Inc. Geolocation-based pictographs
US11281701B2 (en) * 2014-09-18 2022-03-22 Snap Inc. Geolocation-based pictographs
US11216869B2 (en) 2014-09-23 2022-01-04 Snap Inc. User interface to augment an image using geolocation
US20170374003A1 (en) 2014-10-02 2017-12-28 Snapchat, Inc. Ephemeral gallery of ephemeral messages
US11522822B1 (en) 2014-10-02 2022-12-06 Snap Inc. Ephemeral gallery elimination based on gallery and message timers
US11038829B1 (en) 2014-10-02 2021-06-15 Snap Inc. Ephemeral gallery of ephemeral messages with opt-in permanence
US10476830B2 (en) 2014-10-02 2019-11-12 Snap Inc. Ephemeral gallery of ephemeral messages
US11411908B1 (en) 2014-10-02 2022-08-09 Snap Inc. Ephemeral message gallery user interface with online viewing history indicia
US11190679B2 (en) 2014-11-12 2021-11-30 Snap Inc. Accessing media at a geographic location
US10616476B1 (en) 2014-11-12 2020-04-07 Snap Inc. User interface for accessing media at a geographic location
US11783862B2 (en) 2014-12-19 2023-10-10 Snap Inc. Routing messages by message parameter
US11803345B2 (en) 2014-12-19 2023-10-31 Snap Inc. Gallery of messages from individuals with a shared interest
US11250887B2 (en) 2014-12-19 2022-02-15 Snap Inc. Routing messages by message parameter
US11372608B2 (en) 2014-12-19 2022-06-28 Snap Inc. Gallery of messages from individuals with a shared interest
US10580458B2 (en) 2014-12-19 2020-03-03 Snap Inc. Gallery of videos set to an audio time line
US10811053B2 (en) 2014-12-19 2020-10-20 Snap Inc. Routing messages by message parameter
US10157449B1 (en) 2015-01-09 2018-12-18 Snap Inc. Geo-location-based image filters
US11301960B2 (en) 2015-01-09 2022-04-12 Snap Inc. Object recognition based image filters
US11734342B2 (en) 2015-01-09 2023-08-22 Snap Inc. Object recognition based image overlays
US10380720B1 (en) 2015-01-09 2019-08-13 Snap Inc. Location-based image filters
US11388226B1 (en) 2015-01-13 2022-07-12 Snap Inc. Guided personal identity based actions
US11249617B1 (en) 2015-01-19 2022-02-15 Snap Inc. Multichannel system
US10932085B1 (en) 2015-01-26 2021-02-23 Snap Inc. Content request by location
US10123166B2 (en) 2015-01-26 2018-11-06 Snap Inc. Content request by location
US11910267B2 (en) 2015-01-26 2024-02-20 Snap Inc. Content request by location
US11528579B2 (en) 2015-01-26 2022-12-13 Snap Inc. Content request by location
US10536800B1 (en) 2015-01-26 2020-01-14 Snap Inc. Content request by location
US10223397B1 (en) 2015-03-13 2019-03-05 Snap Inc. Social graph based co-location of network users
US10893055B2 (en) 2015-03-18 2021-01-12 Snap Inc. Geo-fence authorization provisioning
US10616239B2 (en) 2015-03-18 2020-04-07 Snap Inc. Geo-fence authorization provisioning
US11902287B2 (en) 2015-03-18 2024-02-13 Snap Inc. Geo-fence authorization provisioning
US11320651B2 (en) 2015-03-23 2022-05-03 Snap Inc. Reducing boot time and power consumption in displaying data content
US11662576B2 (en) 2015-03-23 2023-05-30 Snap Inc. Reducing boot time and power consumption in displaying data content
US10948717B1 (en) 2015-03-23 2021-03-16 Snap Inc. Reducing boot time and power consumption in wearable display systems
US10911575B1 (en) 2015-05-05 2021-02-02 Snap Inc. Systems and methods for story and sub-story navigation
US11496544B2 (en) 2015-05-05 2022-11-08 Snap Inc. Story and sub-story navigation
US10592574B2 (en) 2015-05-05 2020-03-17 Snap Inc. Systems and methods for automated local story generation and curation
US11449539B2 (en) 2015-05-05 2022-09-20 Snap Inc. Automated local story generation and curation
US11392633B2 (en) 2015-05-05 2022-07-19 Snap Inc. Systems and methods for automated local story generation and curation
US10993069B2 (en) 2015-07-16 2021-04-27 Snap Inc. Dynamically adaptive media content delivery
US10817898B2 (en) 2015-08-13 2020-10-27 Placed, Llc Determining exposures to content presented by physical objects
US11315331B2 (en) 2015-10-30 2022-04-26 Snap Inc. Image based tracking in augmented reality systems
US11769307B2 (en) 2015-10-30 2023-09-26 Snap Inc. Image based tracking in augmented reality systems
US10733802B2 (en) 2015-10-30 2020-08-04 Snap Inc. Image based tracking in augmented reality systems
US10366543B1 (en) 2015-10-30 2019-07-30 Snap Inc. Image based tracking in augmented reality systems
US10997783B2 (en) 2015-11-30 2021-05-04 Snap Inc. Image and point cloud based tracking and in augmented reality systems
US11380051B2 (en) 2015-11-30 2022-07-05 Snap Inc. Image and point cloud based tracking and in augmented reality systems
US11599241B2 (en) 2015-11-30 2023-03-07 Snap Inc. Network resource location linking and visual content sharing
US10474321B2 (en) 2015-11-30 2019-11-12 Snap Inc. Network resource location linking and visual content sharing
US10354425B2 (en) 2015-12-18 2019-07-16 Snap Inc. Method and system for providing context relevant media augmentation
US11830117B2 (en) 2015-12-18 2023-11-28 Snap Inc Media overlay publication system
US11468615B2 (en) 2015-12-18 2022-10-11 Snap Inc. Media overlay publication system
US11197123B2 (en) 2016-02-26 2021-12-07 Snap Inc. Generation, curation, and presentation of media collections
US10679389B2 (en) 2016-02-26 2020-06-09 Snap Inc. Methods and systems for generation, curation, and presentation of media collections
US11023514B2 (en) 2016-02-26 2021-06-01 Snap Inc. Methods and systems for generation, curation, and presentation of media collections
US11889381B2 (en) 2016-02-26 2024-01-30 Snap Inc. Generation, curation, and presentation of media collections
US10834525B2 (en) 2016-02-26 2020-11-10 Snap Inc. Generation, curation, and presentation of media collections
US11611846B2 (en) 2016-02-26 2023-03-21 Snap Inc. Generation, curation, and presentation of media collections
US11631276B2 (en) 2016-03-31 2023-04-18 Snap Inc. Automated avatar generation
US11445326B2 (en) 2016-06-28 2022-09-13 Snap Inc. Track engagement of media items
US10885559B1 (en) 2016-06-28 2021-01-05 Snap Inc. Generation, curation, and presentation of media collections with automated advertising
US10219110B2 (en) 2016-06-28 2019-02-26 Snap Inc. System to track engagement of media items
US10430838B1 (en) 2016-06-28 2019-10-01 Snap Inc. Methods and systems for generation, curation, and presentation of media collections with automated advertising
US10327100B1 (en) 2016-06-28 2019-06-18 Snap Inc. System to track engagement of media items
US10506371B2 (en) 2016-06-28 2019-12-10 Snap Inc. System to track engagement of media items
US10785597B2 (en) 2016-06-28 2020-09-22 Snap Inc. System to track engagement of media items
US10165402B1 (en) 2016-06-28 2018-12-25 Snap Inc. System to track engagement of media items
US10735892B2 (en) 2016-06-28 2020-08-04 Snap Inc. System to track engagement of media items
US11640625B2 (en) 2016-06-28 2023-05-02 Snap Inc. Generation, curation, and presentation of media collections with automated advertising
US11080351B1 (en) 2016-06-30 2021-08-03 Snap Inc. Automated content curation and communication
US11895068B2 (en) 2016-06-30 2024-02-06 Snap Inc. Automated content curation and communication
US10387514B1 (en) 2016-06-30 2019-08-20 Snap Inc. Automated content curation and communication
US11509615B2 (en) 2016-07-19 2022-11-22 Snap Inc. Generating customized electronic messaging graphics
US10348662B2 (en) 2016-07-19 2019-07-09 Snap Inc. Generating customized electronic messaging graphics
US11816853B2 (en) 2016-08-30 2023-11-14 Snap Inc. Systems and methods for simultaneous localization and mapping
US11843456B2 (en) 2016-10-24 2023-12-12 Snap Inc. Generating and displaying customized avatars in media overlays
US11876762B1 (en) 2016-10-24 2024-01-16 Snap Inc. Generating and displaying customized avatars in media overlays
US10623666B2 (en) 2016-11-07 2020-04-14 Snap Inc. Selective identification and order of image modifiers
US11750767B2 (en) 2016-11-07 2023-09-05 Snap Inc. Selective identification and order of image modifiers
US11233952B2 (en) 2016-11-07 2022-01-25 Snap Inc. Selective identification and order of image modifiers
US10754525B1 (en) 2016-12-09 2020-08-25 Snap Inc. Customized media overlays
US10203855B2 (en) 2016-12-09 2019-02-12 Snap Inc. Customized user-controlled media overlays
US11397517B2 (en) 2016-12-09 2022-07-26 Snap Inc. Customized media overlays
US11616745B2 (en) 2017-01-09 2023-03-28 Snap Inc. Contextual generation and selection of customized media content
US11870743B1 (en) 2017-01-23 2024-01-09 Snap Inc. Customized digital avatar accessories
US10915911B2 (en) 2017-02-03 2021-02-09 Snap Inc. System to determine a price-schedule to distribute media content
US11861795B1 (en) 2017-02-17 2024-01-02 Snap Inc. Augmented reality anamorphosis system
US11720640B2 (en) 2017-02-17 2023-08-08 Snap Inc. Searching social media content
US11250075B1 (en) 2017-02-17 2022-02-15 Snap Inc. Searching social media content
US10319149B1 (en) 2017-02-17 2019-06-11 Snap Inc. Augmented reality anamorphosis system
US11189299B1 (en) 2017-02-20 2021-11-30 Snap Inc. Augmented reality speech balloon system
US11748579B2 (en) 2017-02-20 2023-09-05 Snap Inc. Augmented reality speech balloon system
US11670057B2 (en) 2017-03-06 2023-06-06 Snap Inc. Virtual vision system
US11037372B2 (en) 2017-03-06 2021-06-15 Snap Inc. Virtual vision system
US10887269B1 (en) 2017-03-09 2021-01-05 Snap Inc. Restricted group content collection
US10523625B1 (en) 2017-03-09 2019-12-31 Snap Inc. Restricted group content collection
US11258749B2 (en) 2017-03-09 2022-02-22 Snap Inc. Restricted group content collection
US11349796B2 (en) 2017-03-27 2022-05-31 Snap Inc. Generating a stitched data stream
US11558678B2 (en) 2017-03-27 2023-01-17 Snap Inc. Generating a stitched data stream
US11297399B1 (en) 2017-03-27 2022-04-05 Snap Inc. Generating a stitched data stream
US11170393B1 (en) 2017-04-11 2021-11-09 Snap Inc. System to calculate an engagement score of location based media content
US11195018B1 (en) 2017-04-20 2021-12-07 Snap Inc. Augmented reality typography personalization system
US10387730B1 (en) 2017-04-20 2019-08-20 Snap Inc. Augmented reality typography personalization system
US11842411B2 (en) 2017-04-27 2023-12-12 Snap Inc. Location-based virtual avatars
US11385763B2 (en) 2017-04-27 2022-07-12 Snap Inc. Map-based graphical user interface indicating geospatial activity metrics
US11418906B2 (en) 2017-04-27 2022-08-16 Snap Inc. Selective location-based identity communication
US11409407B2 (en) 2017-04-27 2022-08-09 Snap Inc. Map-based graphical user interface indicating geospatial activity metrics
US10963529B1 (en) 2017-04-27 2021-03-30 Snap Inc. Location-based search mechanism in a graphical user interface
US11392264B1 (en) 2017-04-27 2022-07-19 Snap Inc. Map-based graphical user interface for multi-type social media galleries
US11451956B1 (en) 2017-04-27 2022-09-20 Snap Inc. Location privacy management on map-based social media platforms
US10952013B1 (en) 2017-04-27 2021-03-16 Snap Inc. Selective location-based identity communication
US11556221B2 (en) 2017-04-27 2023-01-17 Snap Inc. Friend location sharing mechanism for social media platforms
US11893647B2 (en) 2017-04-27 2024-02-06 Snap Inc. Location-based virtual avatars
US11474663B2 (en) 2017-04-27 2022-10-18 Snap Inc. Location-based search mechanism in a graphical user interface
US11782574B2 (en) 2017-04-27 2023-10-10 Snap Inc. Map-based graphical user interface indicating geospatial activity metrics
US11232040B1 (en) 2017-04-28 2022-01-25 Snap Inc. Precaching unlockable data elements
US11675831B2 (en) 2017-05-31 2023-06-13 Snap Inc. Geolocation based playlists
US11475254B1 (en) 2017-09-08 2022-10-18 Snap Inc. Multimodal entity identification
US10740974B1 (en) 2017-09-15 2020-08-11 Snap Inc. Augmented reality system
US11721080B2 (en) 2017-09-15 2023-08-08 Snap Inc. Augmented reality system
US11335067B2 (en) 2017-09-15 2022-05-17 Snap Inc. Augmented reality system
US10499191B1 (en) 2017-10-09 2019-12-03 Snap Inc. Context sensitive presentation of content
US11006242B1 (en) 2017-10-09 2021-05-11 Snap Inc. Context sensitive presentation of content
US11617056B2 (en) 2017-10-09 2023-03-28 Snap Inc. Context sensitive presentation of content
US11670025B2 (en) 2017-10-30 2023-06-06 Snap Inc. Mobile-based cartographic control of display content
US11030787B2 (en) 2017-10-30 2021-06-08 Snap Inc. Mobile-based cartographic control of display content
US11265273B1 (en) 2017-12-01 2022-03-01 Snap, Inc. Dynamic media overlay with smart widget
US11943185B2 (en) 2017-12-01 2024-03-26 Snap Inc. Dynamic media overlay with smart widget
US11558327B2 (en) 2017-12-01 2023-01-17 Snap Inc. Dynamic media overlay with smart widget
US11017173B1 (en) 2017-12-22 2021-05-25 Snap Inc. Named entity recognition visual context and caption data
US11687720B2 (en) 2017-12-22 2023-06-27 Snap Inc. Named entity recognition visual context and caption data
US11487794B2 (en) 2018-01-03 2022-11-01 Snap Inc. Tag distribution visualization system
US10678818B2 (en) 2018-01-03 2020-06-09 Snap Inc. Tag distribution visualization system
US11507614B1 (en) 2018-02-13 2022-11-22 Snap Inc. Icon based tagging
US11841896B2 (en) 2018-02-13 2023-12-12 Snap Inc. Icon based tagging
US10885136B1 (en) 2018-02-28 2021-01-05 Snap Inc. Audience filtering system
US11523159B2 (en) 2018-02-28 2022-12-06 Snap Inc. Generating media content items based on location information
US10979752B1 (en) 2018-02-28 2021-04-13 Snap Inc. Generating media content items based on location information
US11044574B2 (en) 2018-03-06 2021-06-22 Snap Inc. Geo-fence selection system
US10524088B2 (en) 2018-03-06 2019-12-31 Snap Inc. Geo-fence selection system
US11722837B2 (en) 2018-03-06 2023-08-08 Snap Inc. Geo-fence selection system
US11570572B2 (en) 2018-03-06 2023-01-31 Snap Inc. Geo-fence selection system
US10327096B1 (en) 2018-03-06 2019-06-18 Snap Inc. Geo-fence selection system
US11491393B2 (en) 2018-03-14 2022-11-08 Snap Inc. Generating collectible items based on location information
US10933311B2 (en) 2018-03-14 2021-03-02 Snap Inc. Generating collectible items based on location information
US11163941B1 (en) 2018-03-30 2021-11-02 Snap Inc. Annotating a collection of media content items
US10779114B2 (en) 2018-04-18 2020-09-15 Snap Inc. Visitation tracking system
US10924886B2 (en) 2018-04-18 2021-02-16 Snap Inc. Visitation tracking system
US11297463B2 (en) 2018-04-18 2022-04-05 Snap Inc. Visitation tracking system
US10219111B1 (en) 2018-04-18 2019-02-26 Snap Inc. Visitation tracking system
US10681491B1 (en) 2018-04-18 2020-06-09 Snap Inc. Visitation tracking system
US11683657B2 (en) 2018-04-18 2023-06-20 Snap Inc. Visitation tracking system
US10448199B1 (en) 2018-04-18 2019-10-15 Snap Inc. Visitation tracking system
US11860888B2 (en) 2018-05-22 2024-01-02 Snap Inc. Event detection system
US11367234B2 (en) 2018-07-24 2022-06-21 Snap Inc. Conditional modification of augmented reality object
US10789749B2 (en) 2018-07-24 2020-09-29 Snap Inc. Conditional modification of augmented reality object
US10679393B2 (en) 2018-07-24 2020-06-09 Snap Inc. Conditional modification of augmented reality object
US11670026B2 (en) 2018-07-24 2023-06-06 Snap Inc. Conditional modification of augmented reality object
US10943381B2 (en) 2018-07-24 2021-03-09 Snap Inc. Conditional modification of augmented reality object
US10997760B2 (en) 2018-08-31 2021-05-04 Snap Inc. Augmented reality anthropomorphization system
US11450050B2 (en) 2018-08-31 2022-09-20 Snap Inc. Augmented reality anthropomorphization system
US11676319B2 (en) 2018-08-31 2023-06-13 Snap Inc. Augmented reality anthropomorphtzation system
US11455082B2 (en) 2018-09-28 2022-09-27 Snap Inc. Collaborative achievement interface
US11704005B2 (en) 2018-09-28 2023-07-18 Snap Inc. Collaborative achievement interface
US11799811B2 (en) 2018-10-31 2023-10-24 Snap Inc. Messaging and gaming applications communication platform
US11698722B2 (en) 2018-11-30 2023-07-11 Snap Inc. Generating customized avatars based on location information
US11558709B2 (en) 2018-11-30 2023-01-17 Snap Inc. Position service to determine relative position to map features
US11812335B2 (en) 2018-11-30 2023-11-07 Snap Inc. Position service to determine relative position to map features
US11199957B1 (en) 2018-11-30 2021-12-14 Snap Inc. Generating customized avatars based on location information
US11877211B2 (en) 2019-01-14 2024-01-16 Snap Inc. Destination sharing in location sharing system
US11751015B2 (en) 2019-01-16 2023-09-05 Snap Inc. Location-based context information sharing in a messaging system
US11294936B1 (en) 2019-01-30 2022-04-05 Snap Inc. Adaptive spatial density based clustering
US11693887B2 (en) 2019-01-30 2023-07-04 Snap Inc. Adaptive spatial density based clustering
US11809624B2 (en) 2019-02-13 2023-11-07 Snap Inc. Sleep detection in a location sharing system
US11500525B2 (en) 2019-02-25 2022-11-15 Snap Inc. Custom media overlay system
US11574431B2 (en) 2019-02-26 2023-02-07 Snap Inc. Avatar based on weather
US11301117B2 (en) 2019-03-08 2022-04-12 Snap Inc. Contextual information in chat
US11868414B1 (en) 2019-03-14 2024-01-09 Snap Inc. Graph-based prediction for contact suggestion in a location sharing system
US11852554B1 (en) 2019-03-21 2023-12-26 Snap Inc. Barometer calibration in a location sharing system
US11249614B2 (en) 2019-03-28 2022-02-15 Snap Inc. Generating personalized map interface with enhanced icons
US11740760B2 (en) 2019-03-28 2023-08-29 Snap Inc. Generating personalized map interface with enhanced icons
US11361493B2 (en) 2019-04-01 2022-06-14 Snap Inc. Semantic texture mapping system
US11785549B2 (en) 2019-05-30 2023-10-10 Snap Inc. Wearable device location systems
US11606755B2 (en) 2019-05-30 2023-03-14 Snap Inc. Wearable device location systems architecture
US11206615B2 (en) 2019-05-30 2021-12-21 Snap Inc. Wearable device location systems
US11917495B2 (en) 2019-06-07 2024-02-27 Snap Inc. Detection of a physical collision between two client devices in a location sharing system
US11601783B2 (en) 2019-06-07 2023-03-07 Snap Inc. Detection of a physical collision between two client devices in a location sharing system
US11714535B2 (en) 2019-07-11 2023-08-01 Snap Inc. Edge gesture interface with smart interactions
US11821742B2 (en) 2019-09-26 2023-11-21 Snap Inc. Travel based notifications
US11218838B2 (en) 2019-10-31 2022-01-04 Snap Inc. Focused map-based context information surfacing
US11128715B1 (en) 2019-12-30 2021-09-21 Snap Inc. Physical friend proximity in chat
US11429618B2 (en) 2019-12-30 2022-08-30 Snap Inc. Surfacing augmented reality objects
US11943303B2 (en) 2019-12-31 2024-03-26 Snap Inc. Augmented reality objects registry
US11343323B2 (en) 2019-12-31 2022-05-24 Snap Inc. Augmented reality objects registry
US11893208B2 (en) 2019-12-31 2024-02-06 Snap Inc. Combined map icon with action indicator
US11228551B1 (en) 2020-02-12 2022-01-18 Snap Inc. Multiple gateway message exchange
US11888803B2 (en) 2020-02-12 2024-01-30 Snap Inc. Multiple gateway message exchange
US11765117B2 (en) 2020-03-05 2023-09-19 Snap Inc. Storing data based on device location
US11516167B2 (en) 2020-03-05 2022-11-29 Snap Inc. Storing data based on device location
US11619501B2 (en) 2020-03-11 2023-04-04 Snap Inc. Avatar based on trip
US11915400B2 (en) 2020-03-27 2024-02-27 Snap Inc. Location mapping for large scale augmented-reality
US11430091B2 (en) 2020-03-27 2022-08-30 Snap Inc. Location mapping for large scale augmented-reality
US11776256B2 (en) 2020-03-27 2023-10-03 Snap Inc. Shared augmented reality system
US11503432B2 (en) 2020-06-15 2022-11-15 Snap Inc. Scalable real-time location sharing framework
US11314776B2 (en) 2020-06-15 2022-04-26 Snap Inc. Location sharing using friend list versions
US11290851B2 (en) 2020-06-15 2022-03-29 Snap Inc. Location sharing using offline and online objects
US11483267B2 (en) 2020-06-15 2022-10-25 Snap Inc. Location sharing using different rate-limited links
US11676378B2 (en) 2020-06-29 2023-06-13 Snap Inc. Providing travel-based augmented reality content with a captured image
US11943192B2 (en) 2020-08-31 2024-03-26 Snap Inc. Co-location connection service
US11961116B2 (en) 2020-10-26 2024-04-16 Foursquare Labs, Inc. Determining exposures to content presented by physical objects
US11902902B2 (en) 2021-03-29 2024-02-13 Snap Inc. Scheduling requests for location data
US11601888B2 (en) 2021-03-29 2023-03-07 Snap Inc. Determining location using multi-source geolocation data
US11606756B2 (en) 2021-03-29 2023-03-14 Snap Inc. Scheduling requests for location data
US11645324B2 (en) 2021-03-31 2023-05-09 Snap Inc. Location-based timeline media content system
US11829834B2 (en) 2021-10-29 2023-11-28 Snap Inc. Extended QR code
US11956533B2 (en) 2021-11-29 2024-04-09 Snap Inc. Accessing media at a geographic location
US11962645B2 (en) 2022-06-02 2024-04-16 Snap Inc. Guided personal identity based actions
US11954314B2 (en) 2022-09-09 2024-04-09 Snap Inc. Custom media overlay system
US11963105B2 (en) 2023-02-10 2024-04-16 Snap Inc. Wearable device location systems architecture
US11961196B2 (en) 2023-03-17 2024-04-16 Snap Inc. Virtual vision system

Also Published As

Publication number Publication date
JP2016529570A (en) 2016-09-23
US10296933B2 (en) 2019-05-21
MX364910B (en) 2019-05-13
JP6744932B2 (en) 2020-08-19
CN105408855B (en) 2019-06-11
EP2984556A1 (en) 2016-02-17
JP6473135B2 (en) 2019-02-20
KR102025615B1 (en) 2019-09-27
JP2019075172A (en) 2019-05-16
AU2014251221A1 (en) 2015-11-26
BR112015025970A2 (en) 2017-07-25
MX2015014302A (en) 2016-04-20
EP2984556A4 (en) 2016-10-12
CA2908554A1 (en) 2014-10-16
CN105408855A (en) 2016-03-16
AU2014251221B2 (en) 2019-09-12
KR20150142030A (en) 2015-12-21
WO2014168795A1 (en) 2014-10-16

Similar Documents

Publication Publication Date Title
US10296933B2 (en) Identifying content in electronic images
CA2917976C (en) Optimizing electronic layouts for media content
US10261982B2 (en) Asynchronous execution of animation tasks for a GUI
US10896484B2 (en) Method and system of display object pre-generation on windows of idle time available after each frame buffer fill tasks
US10628516B2 (en) Progressive rendering of data sets
US20150339006A1 (en) Asynchronous Preparation of Displayable Sections of a Graphical User Interface
US9607235B2 (en) Image cropping according to points of interest
US20150339033A1 (en) Asynchronous Execution of Tasks for a GUI
US20180189471A1 (en) Visual CAPTCHA Based On Image Segmentation
EP2987074B1 (en) Application-tailored object pre-inflation
AU2014235402B2 (en) Defer heavy operations while scrolling

Legal Events

Date Code Title Description
AS Assignment

Owner name: FACEBOOK, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GOTTESMAN, FREDERICK;CLUNE, DAVID;ANDREWS, JAMES;AND OTHERS;SIGNING DATES FROM 20130509 TO 20140416;REEL/FRAME:032832/0496

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: META PLATFORMS, INC., CALIFORNIA

Free format text: CHANGE OF NAME;ASSIGNOR:FACEBOOK, INC.;REEL/FRAME:058553/0802

Effective date: 20211028

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4