US20100214436A1 - Method of adjusting white balance of image, recording medium having program for performing the method, and apparatus applying the method - Google Patents
Method of adjusting white balance of image, recording medium having program for performing the method, and apparatus applying the method Download PDFInfo
- Publication number
- US20100214436A1 US20100214436A1 US12/708,642 US70864210A US2010214436A1 US 20100214436 A1 US20100214436 A1 US 20100214436A1 US 70864210 A US70864210 A US 70864210A US 2010214436 A1 US2010214436 A1 US 2010214436A1
- Authority
- US
- United States
- Prior art keywords
- display image
- image
- white balance
- sample images
- area
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/64—Circuits for processing colour signals
- H04N9/73—Colour balance circuits, e.g. white balance circuits or colour temperature control
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
- H04N23/84—Camera processing pipelines; Components thereof for processing colour signals
- H04N23/88—Camera processing pipelines; Components thereof for processing colour signals for colour balance, e.g. white-balance circuits or colour temperature control
Definitions
- methods of adjusting white balance may be classified as the following three modes: automatic white balance control mode, preset white balance control mode, and manual white balance control mode.
- the method may further comprise applying the correction gain to a third input image to generate a third display image.
- FIGS. 6 through 8 are schematic views illustrating display images displayed in the digital camera according to the method described with reference to FIG. 5 ;
- FIG. 9 is a flowchart illustrating a method of adjusting white balance in the digital camera illustrated in FIG. 1 , according to another embodiment of the present invention.
- a white balance adjusting apparatus that is used, for example, in a digital camera will be described.
- the white balance adjusting apparatus is not limited to use in digital cameras and may also be used in phones with cameras, personal digital assistants (PDAs), portable multimedia players (PMPs), and so forth, which include a white balance adjusting function.
- PDAs personal digital assistants
- PMPs portable multimedia players
- FIG. 1 is a block diagram illustrating a digital camera including a white balance adjusting apparatus according to an embodiment of the present invention.
- the digital camera includes an optical unit 11 , a motor 12 that drives the optical unit 11 , an image sensor 13 that converts an optical signal that is input through the optical unit 11 to an electrical signal, and an input signal processing unit 14 that performs signal processing such as noise reduction, digital signal conversion, etc., on the electrical signal supplied from the image sensor 13 .
- the digital camera includes a manipulation unit 20 that inputs a manipulation signal of a user and a memory 30 that stores data or algorithms for calculations.
- a liquid crystal display (LCD) 40 is mounted in the digital camera as a display device.
- LCD liquid crystal display
- a flash 50 that provides an amount of light and a light emitting diode (LED) 60 that displays the status of the digital camera are included in the digital camera.
- the digital camera may further include an audio signal processing unit 71 that converts sound into a digital signal or a digital signal of a sound source into an analog signal, a speaker 72 for outputting sound, and a microphone 73 for inputting sound.
- the digital camera includes a digital signal processor (DSP) 80 that controls the overall operations of the digital camera and performs image signal processing on an input image.
- DSP digital signal processor
- the optical unit 11 may include a lens which focuses an optical signal, an aperture which controls an amount of the optical signal (light amount) which reaches the image sensor 13 , and a shutter controlling input of the optical signal.
- the lens may include a zoom lens that controls an angle of view to be narrow or wide according to a focal length, and/or a focus lens that adjusts a focus of a subject.
- the lens may be a single lens or a group of lenses.
- a mechanical shutter over which a cover moves up and down may be used as the shutter.
- the function of a shutter may be conducted by controlling the supply of an electrical signal to the image sensor 13 .
- the motor 12 which drives the optical unit 11 may adjust the position of the lens, may open or close the aperture, or may operate the shutter in order to perform auto-focusing, automatic exposure control, aperture control, zooming, focusing, and so forth.
- the motor 12 receives a control signal from the DSP 80 to control the operation of the optical unit 11 .
- the image sensor 13 receives an optical signal from the optical unit 11 to form an image of a subject.
- Examples of the image sensor 13 include a complementary metal-oxide semiconductor (CMOS) sensor array, a charge-coupled device (CCD) sensor array, etc.
- CMOS complementary metal-oxide semiconductor
- CCD charge-coupled device
- the input signal processing unit 14 may further include an analog-to-digital (A/D) converter that digitalizes an electrical signal, which is an analog signal and supplied from a CCD. Also, the input signal processing unit 14 may include a circuit that adjusts a gain or standardizes a waveform of the electrical signal supplied from the image sensor 13 .
- A/D analog-to-digital
- the MCP temporarily stores RAW data (RGB data) of an image provided from the input signal processing unit 14 .
- a predetermined image signal processing may be performed to the temporarily stored RAW data according to a calculation of the DSP 80 or the RAW data may be transmitted to other elements such as the memory.
- the MCP may convert data constituting the algorithms stored in the E2PROM to executable data and store the same temporarily. By calculating in the DSP 80 using the data stored in the MCP, an operation according to the algorithms may be performed.
- image data which is converted by decompressing an image file stored in the E2PROM may be temporarily stored in the MCP.
- the temporarily stored image data is transmitted to the LCD 40 to display a predetermined image.
- the MCP may be a volatile memory of various types such as a dynamic random access memory (DRAM) or synchronous DRAM (SDRAM), which temporarily stores data while power is being supplied.
- DRAM dynamic random access memory
- SDRAM synchronous DRAM
- the E2PROM may store operation systems needed to operate the digital camera, application programs, data executing algorithms of methods of adjusting white balance according to the present invention, and so forth.
- Examples of the E2PROM include various non-volatile memories such as a flash memory, read-only memory (ROM), etc.
- the white balance adjusting unit 85 a selects one of the first and second sample images, and applies a correction gain corresponding to the selected sample image to the whole input image. For example, if the first sample image is selected, the first correction gain corresponding to the first sample image is applied to the whole input image to adjust white balance.
- the representative color output unit 82 b outputs a representative color of the area.
- the representative color may be a color having a largest surface area among the area.
- the representative color may not be necessarily white.
- the representative color may be output from the area based on preset standards set by the user or the manufacturer using various methods.
- a white balance adjusting apparatus may be in a white balance set up mode among the setup menus.
- a white balance set up mode is set in operation S 11 .
- one of an automatic white balance control mode (operation S 12 ), a preset white balance control mode (operation S 13 ), and a manual white balance control mode (operation S 14 ) may be selected by the user through a manipulation unit.
- the preset white balance control mode specific examples of white balancing under various lighting conditions such as a fluorescent lamp, natural light (sunlight), cloudy (shade), a tungsten bulb, etc., are stored in advance, and the user selects a lighting condition to adjust white balance.
- the method of adjusting white balance according to the present invention may be used in the manual white balance control mode among the above modes. The method of adjusting white balance used in the manual white balance control mode will be described with reference to embodiments below.
- FIG. 5 is a flowchart illustrating a method of adjusting white balance in the digital camera illustrated in FIG. 1 , according to an embodiment of the present invention. Operations of the method of adjusting white balance performed in each of the elements of the digital camera such as an image sensor, a DSP, a LCD, and a manipulation unit will be described individually, but the present invention is not limited thereto.
- a first input image, a second input image, and a third input image may be sequentially generated in operations S 21 , S 22 , and S 23 , respectively.
- the DSP may perform display image signal processing on a second input image and the output sample images and the second input image to generate a second display image and supply the second display image to the LCD. Then the second display image may be displayed on the LCD in operation S 42 .
- the user may check the sample images of the second display image through the LCD. Then the user may select one of the sample images. For example, the user may select a first sample image through the manipulation unit to output the first sample image from the DSP in operation S 36 and output a correction gain corresponding to the output first sample image in operation S 37 . Then the correction gain output may be applied with respect to the third input image in operation S 38 and a third display image including an image to which display image signal processing is performed may be generated in operation S 39 . Then the third display image may be displayed on the LCD in operation S 43 . Accordingly, the user may obtain the third display image including the image having desired color sensitivity.
- the user may proceed photographing by adjusting the focus on a subject and pressing a shutter button.
- a shutter button an image having color sensitivity desired by the user can be obtained.
- FIGS. 6 through 8 are schematic views illustrating a rear side of the digital camera on which a first display image P 1 , a second display image P 2 , and a third display image P 3 are displayed according to the method described with reference to FIG. 5 .
- the second display image P 2 including sample images M showing various color sensitivities of the representative color of the selected area X and an input image are displayed.
- the sample images M may be displayed in a matrix manner, and a selection window S for specifying each of the sample images M may also be displayed together with the sample images M. The user may select one of the sample images M by moving the selection window S.
- the third display image P 3 including an image obtained by applying a correction gain corresponding to the selected sample image to the third input image, may be displayed.
- the third display image P 3 also includes the sample image M arranged in the matrix manner.
- the image sensor sequentially generates a first input image, a second input image, and a third input image in operations S 51 , S 52 , and S 53 , respectively.
- the user may input a control signal through the manipulation unit, and in the DSP, an area of the first display image may be specified according to the control signal.
- a plurality of sample images having various color sensitivities with respect to the area of the first display image may be output using the DSP.
- display image signal processing is performed to the second input image and the output sample images to generate a second display image, and in operation S 72 , the second display image may be displayed on the LCD.
- FIGS. 10 through 12 are schematic views illustrating a rear side of the digital camera on which a first display image Q 1 , a second display image Q 2 , and a third display image Q 3 are displayed according to the method described with reference to FIG. 9 .
- FIG. 10 illustrates a rear side of the digital camera displaying the first display image Q 1 .
- an area X of the first display image Q 1 may be specified.
- FIG. 11 illustrates the second display image Q 2 including a plurality of sample images M resulted by applying various correction gains to the specified area X and an input image.
- FIG. 12 illustrates the third display image Q 3 .
- a sample image in a right upper section among the sample images M arranged in a matrix manner is selected using the selection window S. Then a correction gain applied to the selected sample image is applied to a third input image to perform white balance.
- the third display image Q 3 includes an image to which white balance is adjusted.
- the third display image Q 3 may also include the sample images M.
- white balance can be manually adjusted by using a subject to be photographed, thereby easily performing a manual white balance control mode.
- an apparatus for performing the method of adjusting white balance may be implemented using any general purpose computing device or devices.
- Any of the computing devices may comprise a processor, a memory for storing program data and executing the program data, a permanent storage such as a disk drive, a communications port for handling communications with external devices, and user interface devices, including a display, keyboard, mouse, etc.
- these software modules may be stored as program instructions executable on the processor on a computer-readable storage medium, where the program instructions stored on this medium can be read by the computing device, stored in the memory, and executed by the processor.
- Examples of the storage medium include magnetic storage media (e.g., floppy disks, hard disks, or magnetic tape), optical recording media (e.g., CD-ROMs or digital versatile disks (DVDs)), and electronic storage media (e.g., integrated circuits (IC's), ROM, RAM, EEPROM, or flash memory).
- the storage medium may also be distributed over network-coupled computing devices so that the program instructions are stored and executed in a distributed fashion.
- the present invention may be described in terms of functional block components and various processing steps. Such functional blocks may be realized by any number of hardware and/or software components configured to perform the specified functions.
- the present invention may employ various integrated circuit components, e.g., memory elements, processing elements, logic elements, look-up tables, and the like, which may carry out a variety of functions under the control of one or more microprocessors or other control devices.
- the elements of the present invention are implemented using software programming or software elements the invention may be implemented with any programming or scripting language such as C, C++, Java, assembler, or the like, with the various algorithms being implemented with any combination of data structures, objects, processes, routines or other programming elements.
- the present invention could employ any number of conventional techniques for electronics configuration, signal processing and/or control, data processing and the like.
- the word mechanism is used broadly and is not limited to mechanical or physical embodiments, but can include software routines in conjunction with processors, etc.
Abstract
Description
- This application claims the priority benefit of Korean Patent Application No. 10-2009-0014410, filed on Feb. 20, 2009, in the Korean Intellectual Property Office, the disclosure of which is incorporated herein in its entirety by reference.
- 1. Field of the Invention
- The present invention relates to a method of adjusting white balance of an image, a computer readable storage medium having stored thereon a program executable by a processor for performing the method, and a digital apparatus for executing the method.
- 2. Description of the Related Art
- In conventional digital cameras, which is one kind of digital apparatus, methods of adjusting white balance may be classified as the following three modes: automatic white balance control mode, preset white balance control mode, and manual white balance control mode.
- In the manual white balance control mode of conventional digital cameras, a nearby white subject such as white paper is searched for to adjust white balance that is appropriate for a particular circumstance.
- Accordingly, the manual white balance control mode always requires a white subject. However, the user may not obtain desired color sensitivity in other colors than white when the manual white balance control mode is used.
- Embodiments of the present invention include a method of adjusting white balance in which a white subject does not need to be searched for in order to adjust white balance, but images having desired color sensitivity can be obtained by adjusting white balance of colors other than white.
- According to an aspect of the present invention, a method of adjusting white balance comprises generating a display image by performing display image signal processing to an input image, specifying an area of the display image, outputting at least two sample images by applying at least two types of correction gains to the area of the display image, and applying the correction gain that is applied to one of the at least two sample images to the input image.
- The method may further comprise performing display image signal processing to a first input image to generate and display a first display image, specifying an area of the first display image, outputting at least two sample images by applying at least two types of correction gains to the area of the first display image, and performing display image signal processing to a second input image and the at least two sample images to generate and display a second display image.
- The method may further comprise applying the correction gain to a third input image to generate a third display image.
- The method may further comprise selecting one of the at least two sample images.
- The method may further comprise outputting a representative color having a largest surface area of the specified area of the display image, and outputting at least two sample images to which at least two types of correction gains are applied with respect to an image of the specified area of the display image that displays the output representative color.
- The at least two sample images may be output from a database in which a plurality of sample images to which different correction gains are applied with respect to an image having a plurality of colors are stored.
- According to another aspect of the present invention, a computer readable storage medium may have stored thereon a computer program executable by a processor for performing a method of adjusting white balance as described above.
- According to another aspect of the present invention, a white balance adjusting apparatus comprises a display image processing unit that performs display image signal processing to an input image to generate a display image, an area set up unit that specifies an area of the display image, a sample image output unit that outputs at least two sample images to which at least two types of correction gains are applied with respect to the area of the display image, and a white balance adjusting unit that applies the correction gain that is applied to one of the at least two sample images to the input image.
- The white balance adjusting apparatus may further comprise an imaging device that inputs the input image.
- The imaging device may input a first input image, and the display image processing unit may perform display image signal processing to the first input image to generate a first display image, and the area set up unit may specify an area of the first display image.
- The imaging device may input a second input image, and the display image processing unit may perform display image signal processing to the second input image and the sample images to generate a second display image.
- The imaging device may input a third input image, and the white balance adjusting unit may apply the correction gain to the third input image, and the display image processing unit may perform display image signal processing to the third input image, to which the correction gain is applied, to generate a third display image.
- The white balance adjusting apparatus may further comprise a display unit that displays the display image.
- The white balance adjusting apparatus may further comprise a representative color outputting unit that outputs a representative color having a largest surface area of the display image.
- The white balance adjusting apparatus may further comprise a database in which a plurality of sample images, to which different correction gains are applied with respect to an image having a plurality of colors, are stored.
- The sample image output unit may output the at least two sample images from the database.
- The above and other features and advantages of the present invention will become more apparent by describing in detail exemplary embodiments thereof with reference to the attached drawings in which:
-
FIG. 1 is a block diagram illustrating a digital camera including a white balance adjusting apparatus according to an embodiment of the present invention; -
FIG. 2 is a block diagram illustrating a digital signal processor (DSP) of the digital camera ofFIG. 1 , according to an embodiment of the present invention; -
FIG. 3 is a block diagram illustrating a DSP of the digital camera ofFIG. 1 , according to another embodiment of the present invention; -
FIG. 4 is a flowchart illustrating a method of selecting a manual white balance control mode in order to apply a method of adjusting white balance according to an embodiment of the present invention; -
FIG. 5 is a flowchart illustrating a method of adjusting white balance in the digital camera illustrated inFIG. 1 , according to an embodiment of the present invention; -
FIGS. 6 through 8 are schematic views illustrating display images displayed in the digital camera according to the method described with reference toFIG. 5 ; -
FIG. 9 is a flowchart illustrating a method of adjusting white balance in the digital camera illustrated inFIG. 1 , according to another embodiment of the present invention; and -
FIGS. 10 through 12 are schematic views illustrating display images displayed in the digital camera according to the method described with reference toFIG. 9 . - A white balance adjusting apparatus, according to an embodiment of the present invention, that is used, for example, in a digital camera will be described. However, the white balance adjusting apparatus is not limited to use in digital cameras and may also be used in phones with cameras, personal digital assistants (PDAs), portable multimedia players (PMPs), and so forth, which include a white balance adjusting function.
-
FIG. 1 is a block diagram illustrating a digital camera including a white balance adjusting apparatus according to an embodiment of the present invention. Referring toFIG. 1 , the digital camera includes anoptical unit 11, amotor 12 that drives theoptical unit 11, animage sensor 13 that converts an optical signal that is input through theoptical unit 11 to an electrical signal, and an inputsignal processing unit 14 that performs signal processing such as noise reduction, digital signal conversion, etc., on the electrical signal supplied from theimage sensor 13. Also, the digital camera includes amanipulation unit 20 that inputs a manipulation signal of a user and amemory 30 that stores data or algorithms for calculations. In addition, a liquid crystal display (LCD) 40 is mounted in the digital camera as a display device. Also, aflash 50 that provides an amount of light and a light emitting diode (LED) 60 that displays the status of the digital camera are included in the digital camera. The digital camera may further include an audiosignal processing unit 71 that converts sound into a digital signal or a digital signal of a sound source into an analog signal, aspeaker 72 for outputting sound, and amicrophone 73 for inputting sound. Also, the digital camera includes a digital signal processor (DSP) 80 that controls the overall operations of the digital camera and performs image signal processing on an input image. - The elements of the digital camera will hereinafter be described in detail.
- The
optical unit 11 may include a lens which focuses an optical signal, an aperture which controls an amount of the optical signal (light amount) which reaches theimage sensor 13, and a shutter controlling input of the optical signal. The lens may include a zoom lens that controls an angle of view to be narrow or wide according to a focal length, and/or a focus lens that adjusts a focus of a subject. The lens may be a single lens or a group of lenses. A mechanical shutter over which a cover moves up and down may be used as the shutter. Alternatively, instead of installing a shutter device, the function of a shutter may be conducted by controlling the supply of an electrical signal to theimage sensor 13. - The
motor 12 which drives theoptical unit 11 may adjust the position of the lens, may open or close the aperture, or may operate the shutter in order to perform auto-focusing, automatic exposure control, aperture control, zooming, focusing, and so forth. Themotor 12 receives a control signal from the DSP 80 to control the operation of theoptical unit 11. - The
image sensor 13 receives an optical signal from theoptical unit 11 to form an image of a subject. Examples of theimage sensor 13 include a complementary metal-oxide semiconductor (CMOS) sensor array, a charge-coupled device (CCD) sensor array, etc. - The input
signal processing unit 14 may further include an analog-to-digital (A/D) converter that digitalizes an electrical signal, which is an analog signal and supplied from a CCD. Also, the inputsignal processing unit 14 may include a circuit that adjusts a gain or standardizes a waveform of the electrical signal supplied from theimage sensor 13. - The
manipulation unit 20 may include elements for the user to manipulate the digital camera or may set various functions during photographing. For example, themanipulation unit 20 may include elements such as buttons, keys, touch panels, touch screens, dials, etc. User control signals such as power on/off, photographing start/stop, replay start/stop/search, operation of an optical system, mode conversion, menu manipulation, selection manipulation, and so forth may be input using the elements of themanipulation unit 20. - The
memory 30 may include a multi-chip package (MCP) for temporarily storing data of an input image, data for calculations, processing results, etc., an electrically erasable programmable read-only memory (E2PROM) for storing algorithms required for operations of the digital camera or set up data, and a secure digital (SD) card/MultiMediaCard (MMC) as a recording device for storing image files. - The MCP temporarily stores RAW data (RGB data) of an image provided from the input
signal processing unit 14. A predetermined image signal processing may be performed to the temporarily stored RAW data according to a calculation of theDSP 80 or the RAW data may be transmitted to other elements such as the memory. In addition, the MCP may convert data constituting the algorithms stored in the E2PROM to executable data and store the same temporarily. By calculating in theDSP 80 using the data stored in the MCP, an operation according to the algorithms may be performed. Also, image data which is converted by decompressing an image file stored in the E2PROM may be temporarily stored in the MCP. The temporarily stored image data is transmitted to theLCD 40 to display a predetermined image. The MCP may be a volatile memory of various types such as a dynamic random access memory (DRAM) or synchronous DRAM (SDRAM), which temporarily stores data while power is being supplied. - The E2PROM may store operation systems needed to operate the digital camera, application programs, data executing algorithms of methods of adjusting white balance according to the present invention, and so forth. Examples of the E2PROM include various non-volatile memories such as a flash memory, read-only memory (ROM), etc.
- In the SD card/MMC, image files that are generated by compressing the image data supplied from the input
signal processing unit 14 may be stored. Examples of the SD card/MMC include a flash memory, hard disk drive (HDD), an optical disk, an optical magnetic disk, a hologram memory, etc. - The
LCD 40 may realize an image corresponding to the image data provided from the inputsignal processing unit 14, or may display an image corresponding to image data restored from an image file stored in the SD card/MMC. However, the display device according to the current embodiment of the present invention is not limited to theLCD 40, and an organic electroluminescent display apparatus, an electrophoresis display apparatus, etc., may also be used. - The audio
signal processing unit 71 converts a digital signal of a sound source from theDSP 80 into sound and amplifies and transmits the sound to thespeaker 72 to be output. Alternatively, sound may be input through themicrophone 73, and the sound may be converted into a digital signal and compressed to an audio file. Then the audio file may be transmitted to theDSP 80 to be calculated. - The
DSP 80 may reduce noise of input image data, and may perform image signal processing for improving image quality, such as gamma correction, color filter array interpolation, color matrix, color correction, color enhancement, etc. Also, the generated image data may be compressed to an image file by image signal processing for improving image quality, or the image data may be restored from the image file. The compression type of an image may be reversible or irreversible. Examples of appropriate compression type may be a joint photographic experts group (JPEG) or JPEG 2000. Also, theDSP 80 may perform functions of obscure processing, coloring, blurring, edge emphasis, image interpreting, image recognition, image effect, etc. Examples of the image recognition include face recognition or scene recognition. Also, theDSP 80 may perform display image signal processing to display the image on theLCD 40. For example, luminance level control, color correction, contrast control, edge emphasis control, screen split, character image, and so forth, or synthesis of images may be performed. TheDSP 80 may be connected to an external monitor and perform predetermined display image signal processing so that image data is displayed on the external monitor. That is, theDSP 80 may control the image data to be transmitted to the external monitor and be displayed on the external monitor. - The
DSP 80 may perform the above-described display image signal processing, and may control the elements of the digital camera according to the results of the display image signal processing. Also, theDSP 80 may control each of the elements according to a control signal of the user, which is input through themanipulation unit 20. Algorithms for performing the display image signal processing are stored in the E2PROM, and they are converted into executable data for calculations and stored in thememory 30. Thus, theDSP 80 may perform calculations for display image signal processing according to the algorithms. - The controlling operation of the
DSP 80 will be described in detail with reference toFIG. 2 or 3. -
FIG. 2 is a block diagram illustrating aDSP 80 a of the digital camera ofFIG. 1 , according to an embodiment of the present invention. - Referring to
FIG. 2 , theDSP 80 a includes an area set upunit 81 a, a sampleimage output unit 83 a, a whitebalance adjusting unit 85 a, and a displayimage processing unit 87 a. - The area set up
unit 81 a sets up an area of an input image displayed on theLCD 40 ofFIG. 1 , by displaying a selection frame on the input image. An area of the input image may be selected by placing the selection frame on the area. Alternatively, the area may be automatically set up as an area including a predetermined object by the above-described image recognition such as face recognition and scene recognition. - The sample
image output unit 83 a generates sample images by applying at least two correction gains to the selected area of the input image. For example, a first sample image may be generated by applying a first correction gain to the area of the input image, and a second sample image may be generated by applying a second correction gain, which is different from the first correction gain, to the area of the input image. - Then the white
balance adjusting unit 85 a selects one of the first and second sample images, and applies a correction gain corresponding to the selected sample image to the whole input image. For example, if the first sample image is selected, the first correction gain corresponding to the first sample image is applied to the whole input image to adjust white balance. - The display
image processing unit 87 a performs the above-described display image signal processing on the input image to generate a display image, and provides the display image to theLCD 40. The displayimage processing unit 87 a performs predetermined display image signal processing to the input image in order to display the input image which is input in real-time. A first display image may be generated by performing display image signal processing to a first input image. By displaying the first display image on theLCD 40, the user may set up an area for adjusting white balance. That is, the user may check the first display image displayed on theLCD 40 to set up an area as a reference for white balance control. The area may also be set up automatically by image recognition. Also, the displayimage processing unit 87 a may perform display image signal processing on the sample images of the selected area and a second input image, thereby generating a second display image that is also displayed on theLCD 40. The second input image may be input after the first input image or may be the same image as the first input image. Also, the user may select one sample image through the second display image and apply a correction gain which is applied to the selected sample image, to the second input image, and perform display image signal processing to the second input image to which the correction gain is applied in the displayimage processing unit 87 a, to generate a third display image. The second input image to which the correction gain is applied may be a third input image that is input after the second input image, and may also be the same as the first input image or the second input image. White balance is performed to the third display image, and the third display image may be displayed on theLCD 40. -
FIG. 3 is a block diagram illustrating aDSP 80 b of the digital camera ofFIG. 1 , according to another embodiment of the present invention. TheDSP 80 b is similar to theDSP 80 a illustrated inFIG. 2 in that sample images are generated and white balance is performed using a correction gain applied to the selected sample images. However, the type of the sample images is different. Hereinafter, the description will focus on the differences of theDSP 80 b from theDSP 80 a ofFIG. 2 . - Referring to
FIGS. 1 and 3 , theDSP 80 b includes an area set upunit 81 b, a representativecolor output unit 82 b, a sampleimage output unit 83 b, a whitebalance adjusting unit 85 b, and a displayimage processing unit 87 b. - The area set up
unit 81 b sets up an area of an input image displayed on theLCD 40 ofFIG. 1 . - The representative
color output unit 82 b outputs a representative color of the area. The representative color may be a color having a largest surface area among the area. The representative color may not be necessarily white. The representative color may be output from the area based on preset standards set by the user or the manufacturer using various methods. - The sample
image output unit 83 b may generate at least two sample images to which at least two types of correction gains are applied, with respect to the input image displaying the representative color. For example, if red is set as the representative color, red sample images having various color sensitivities may be output by applying different correction gains to the sample images. - White balance is performed by applying the correction gain, which is applied to the selected sample image among the sample images output from the white
balance adjusting unit 85 b, to the input image. - The display
image processing unit 87 b performs display image signal processing on the input image to generate a display image, and provides the display image to theLCD 40. In detail, display image signal processing is performed to a first input image to generate a first display image. Also, display image signal processing is performed on sample images generated from an area of the first display image and to a second input image, to generate a second display image. Furthermore, the correction gain applied to the selected sample image may be applied to a third input image to perform white balance to the third input image. Next, by performing display image signal processing to the third input image, a third display image may be generated. The first input image, the second input image, and the third input image may be the same images or different images that are input sequentially. The first input image, the second input image, and the third input image may be displayed on theLCD 40. The third display image has color sensitivity desired by the user since white balance is performed thereto. - Also, the
DSP 80 b may further include adatabase 84 b that includes sample images displaying different color sensitivities of different colors. In thedatabase 84 b, a plurality of sample images, to which different correction gains with respect to an image having a plurality of colors are applied, may be stored. For example, a plurality of sample images having color sensitivities with respect to red and blue and correction gains corresponding to the sample images may be stored in thedatabase 84 b. Accordingly, if red is set as the representative color, red sample images displaying various color sensitivities of the red color may be output from thedatabase 84 b. - By using the
DSPs -
FIG. 4 is a flowchart illustrating a method of selecting a manual white balance control mode in order to apply a method of adjusting white balance according to an embodiment of the present invention. - For example, in a digital camera, a white balance adjusting apparatus may be in a white balance set up mode among the setup menus. Thus, first, a white balance set up mode is set in operation S11. Then one of an automatic white balance control mode (operation S12), a preset white balance control mode (operation S13), and a manual white balance control mode (operation S14) may be selected by the user through a manipulation unit. In the preset white balance control mode, specific examples of white balancing under various lighting conditions such as a fluorescent lamp, natural light (sunlight), cloudy (shade), a tungsten bulb, etc., are stored in advance, and the user selects a lighting condition to adjust white balance. The method of adjusting white balance according to the present invention may be used in the manual white balance control mode among the above modes. The method of adjusting white balance used in the manual white balance control mode will be described with reference to embodiments below.
-
FIG. 5 is a flowchart illustrating a method of adjusting white balance in the digital camera illustrated inFIG. 1 , according to an embodiment of the present invention. Operations of the method of adjusting white balance performed in each of the elements of the digital camera such as an image sensor, a DSP, a LCD, and a manipulation unit will be described individually, but the present invention is not limited thereto. - Referring to
FIG. 5 , light reflected from a subject is received in the image sensor, thereby generating an input image. A first input image, a second input image, and a third input image may be sequentially generated in operations S21, S22, and S23, respectively. - The DSP may generate a first display image by performing display image signal processing on the first input image in operation S31 and display the first display image on the LCD in operation S41.
- The user may check the first display image through the LCD and select an area of the first display image via the manipulation unit. The DSP may specify the area of the first display image selected by the user in operation S32 and output a representative color that occupies a largest surface area of the area in operation S33. Then the DSP may output sample images displaying various color sensitivities of the representative color in operation S34. That is, various correction gains may be applied to the area of the first display image displaying the representative color to output sample images.
- In operation S35, the DSP may perform display image signal processing on a second input image and the output sample images and the second input image to generate a second display image and supply the second display image to the LCD. Then the second display image may be displayed on the LCD in operation S42.
- The user may check the sample images of the second display image through the LCD. Then the user may select one of the sample images. For example, the user may select a first sample image through the manipulation unit to output the first sample image from the DSP in operation S36 and output a correction gain corresponding to the output first sample image in operation S37. Then the correction gain output may be applied with respect to the third input image in operation S38 and a third display image including an image to which display image signal processing is performed may be generated in operation S39. Then the third display image may be displayed on the LCD in operation S43. Accordingly, the user may obtain the third display image including the image having desired color sensitivity.
- Next, after adjusting white balance with desired color sensitivity, the user may proceed photographing by adjusting the focus on a subject and pressing a shutter button. Thus, an image having color sensitivity desired by the user can be obtained.
-
FIGS. 6 through 8 are schematic views illustrating a rear side of the digital camera on which a first display image P1, a second display image P2, and a third display image P3 are displayed according to the method described with reference toFIG. 5 . - Referring to
FIG. 6 , the first display image, P1 is displayed on the LCD. An area X of the first display image P1 may be specified. - Referring to
FIG. 7 , the second display image P2 including sample images M showing various color sensitivities of the representative color of the selected area X and an input image are displayed. The sample images M may be displayed in a matrix manner, and a selection window S for specifying each of the sample images M may also be displayed together with the sample images M. The user may select one of the sample images M by moving the selection window S. - Referring to
FIG. 8 , by moving the selection window S to the right, a sample image in an upper right section is selected, and the third display image P3, including an image obtained by applying a correction gain corresponding to the selected sample image to the third input image, may be displayed. The third display image P3 also includes the sample image M arranged in the matrix manner. -
FIG. 9 is a flowchart illustrating a method of adjusting white balance in the digital camera illustrated inFIG. 1 , according to another embodiment of the present invention. Operations of the method of adjusting white balance performed in each of the elements of the digital camera such as an image sensor, a DSP, a LCD, and a manipulation unit will be described individually, but the present invention is not limited thereto. - Referring to
FIG. 9 , the image sensor sequentially generates a first input image, a second input image, and a third input image in operations S51, S52, and S53, respectively. - Then in operation S61, display image signal processing is performed to the first input image in the DSP to generate a first display image. In operation S71, the first display image may be displayed on the LCD.
- In operation S62, the user may input a control signal through the manipulation unit, and in the DSP, an area of the first display image may be specified according to the control signal. In operation S63, a plurality of sample images having various color sensitivities with respect to the area of the first display image may be output using the DSP. In operation S64, display image signal processing is performed to the second input image and the output sample images to generate a second display image, and in operation S72, the second display image may be displayed on the LCD.
- In operation S65, the user inputs a control signal for selecting one of the sample images of the second display image through the manipulation unit, and outputs a first sample image in the DSP accordingly. In operation S66, a correction gain corresponding to the first sample image is output, and in operation S67, the output correction gain is applied to the third input image to adjust white balance. In operation S68, display image signal processing is performed to the third input image to which white balance is adjusted to generate a third display image, and the third display image is provided to the LCD. In operation S73, the third display image may be displayed on the LCD.
-
FIGS. 10 through 12 are schematic views illustrating a rear side of the digital camera on which a first display image Q1, a second display image Q2, and a third display image Q3 are displayed according to the method described with reference toFIG. 9 . In detail,FIG. 10 illustrates a rear side of the digital camera displaying the first display image Q1. Referring toFIG. 10 , an area X of the first display image Q1 may be specified. -
FIG. 11 illustrates the second display image Q2 including a plurality of sample images M resulted by applying various correction gains to the specified area X and an input image. -
FIG. 12 illustrates the third display image Q3. A sample image in a right upper section among the sample images M arranged in a matrix manner is selected using the selection window S. Then a correction gain applied to the selected sample image is applied to a third input image to perform white balance. Thus the third display image Q3 includes an image to which white balance is adjusted. The third display image Q3 may also include the sample images M. - According to the present invention, white balance can be manually adjusted by using a subject to be photographed, thereby easily performing a manual white balance control mode.
- Also, according to the present invention, white balance is adjusted for other colors than white, and thus the user can obtain images having desired color sensitivity.
- In general, an apparatus for performing the method of adjusting white balance according to the embodiments of the present invention may be implemented using any general purpose computing device or devices. Any of the computing devices may comprise a processor, a memory for storing program data and executing the program data, a permanent storage such as a disk drive, a communications port for handling communications with external devices, and user interface devices, including a display, keyboard, mouse, etc. When software modules are involved, these software modules may be stored as program instructions executable on the processor on a computer-readable storage medium, where the program instructions stored on this medium can be read by the computing device, stored in the memory, and executed by the processor. Examples of the storage medium include magnetic storage media (e.g., floppy disks, hard disks, or magnetic tape), optical recording media (e.g., CD-ROMs or digital versatile disks (DVDs)), and electronic storage media (e.g., integrated circuits (IC's), ROM, RAM, EEPROM, or flash memory). The storage medium may also be distributed over network-coupled computing devices so that the program instructions are stored and executed in a distributed fashion.
- The present invention may be described in terms of functional block components and various processing steps. Such functional blocks may be realized by any number of hardware and/or software components configured to perform the specified functions. For example, the present invention may employ various integrated circuit components, e.g., memory elements, processing elements, logic elements, look-up tables, and the like, which may carry out a variety of functions under the control of one or more microprocessors or other control devices. Similarly, where the elements of the present invention are implemented using software programming or software elements the invention may be implemented with any programming or scripting language such as C, C++, Java, assembler, or the like, with the various algorithms being implemented with any combination of data structures, objects, processes, routines or other programming elements. Furthermore, the present invention could employ any number of conventional techniques for electronics configuration, signal processing and/or control, data processing and the like. The word mechanism is used broadly and is not limited to mechanical or physical embodiments, but can include software routines in conjunction with processors, etc.
- The particular implementations shown and described herein are illustrative examples of the invention and are not intended to otherwise limit the scope of the invention in any way. For the sake of brevity, conventional electronics, control systems, software development and other functional aspects of the systems (and components of the individual operating components of the systems) may not be described in detail. Furthermore, the connecting lines, or connectors shown in the various figures presented are intended to represent exemplary functional relationships and/or physical or logical couplings between the various elements. It should be noted that many alternative or additional functional relationships, physical connections or logical connections may be present in a practical device. Moreover, no item or component is essential to the practice of the invention unless the element is specifically described as “essential” or “critical”.
- As these embodiments of the present invention are described with reference to illustrations, various modifications or adaptations of the methods and or specific structures described may become apparent to those skilled in the art. All such modifications, adaptations, or variations that rely upon the teachings of the present invention, and through which these teachings have advanced the art, are considered to be within the spirit and scope of the present invention. Hence, these descriptions and drawings should not be considered in a limiting sense, as it is understood that the present invention is in no way limited to only the embodiments illustrated.
- It will be recognized that the terms “comprising,” “including,” and “having,” as used herein, are specifically intended to be read as open-ended terms of art. The use of the terms “a” and “and” and “the” and similar referents in the context of describing the invention (especially in the context of the following claims) are to be construed to cover both the singular and the plural. Furthermore, recitation of ranges of values herein are merely intended to serve as a shorthand method of referring individually to each separate value falling within the range, unless otherwise indicated herein, and each separate value is incorporated into the specification as if it were individually recited herein. Finally, the steps of all methods described herein can be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context.
Claims (20)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020090014410A KR101595254B1 (en) | 2009-02-20 | 2009-02-20 | Method for controlling white balance of an image medium of recording the method and apparatus applying the method |
KR10-2009-0014410 | 2009-02-20 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20100214436A1 true US20100214436A1 (en) | 2010-08-26 |
US8754958B2 US8754958B2 (en) | 2014-06-17 |
Family
ID=42622322
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/708,642 Active 2032-03-23 US8754958B2 (en) | 2009-02-20 | 2010-02-19 | Method of adjusting white balance of image, recording medium having program for performing the method, and apparatus applying the method |
Country Status (3)
Country | Link |
---|---|
US (1) | US8754958B2 (en) |
KR (1) | KR101595254B1 (en) |
CN (2) | CN105657393B (en) |
Cited By (148)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102209246A (en) * | 2011-05-23 | 2011-10-05 | 北京工业大学 | Real-time video white balance processing system |
US20140267867A1 (en) * | 2013-03-14 | 2014-09-18 | Samsung Electronics Co., Ltd. | Electronic device and method for image processing |
US20140301638A1 (en) * | 2013-04-08 | 2014-10-09 | Samsung Electronics Co., Ltd. | Color extraction-based image processing method, computer-readable storage medium storing the same, and digital image apparatus |
US20150002693A1 (en) * | 2013-06-26 | 2015-01-01 | Nvidia Corporation | Method and system for performing white balancing operations on captured images |
CN104618645A (en) * | 2015-01-20 | 2015-05-13 | 广东欧珀移动通信有限公司 | Method and device for shooting through two cameras |
US9083770B1 (en) | 2013-11-26 | 2015-07-14 | Snapchat, Inc. | Method and system for integrating real time communication features in applications |
US9094137B1 (en) | 2014-06-13 | 2015-07-28 | Snapchat, Inc. | Priority based placement of messages in a geo-location based event gallery |
US9225897B1 (en) * | 2014-07-07 | 2015-12-29 | Snapchat, Inc. | Apparatus and method for supplying content aware photo filters |
US9237202B1 (en) | 2014-03-07 | 2016-01-12 | Snapchat, Inc. | Content delivery network for ephemeral objects |
US9276886B1 (en) | 2014-05-09 | 2016-03-01 | Snapchat, Inc. | Apparatus and method for dynamically configuring application component tiles |
US9385983B1 (en) | 2014-12-19 | 2016-07-05 | Snapchat, Inc. | Gallery of messages from individuals with a shared interest |
US9396354B1 (en) | 2014-05-28 | 2016-07-19 | Snapchat, Inc. | Apparatus and method for automated privacy protection in distributed images |
US9537811B2 (en) | 2014-10-02 | 2017-01-03 | Snap Inc. | Ephemeral gallery of ephemeral messages |
US9705831B2 (en) | 2013-05-30 | 2017-07-11 | Snap Inc. | Apparatus and method for maintaining a message thread with opt-in permanence for entries |
US9721394B2 (en) | 2012-08-22 | 2017-08-01 | Snaps Media, Inc. | Augmented reality virtual content platform apparatuses, methods and systems |
US9742713B2 (en) | 2013-05-30 | 2017-08-22 | Snap Inc. | Apparatus and method for maintaining a message thread with opt-in permanence for entries |
US9826208B2 (en) | 2013-06-26 | 2017-11-21 | Nvidia Corporation | Method and system for generating weights for use in white balancing an image |
US9843720B1 (en) | 2014-11-12 | 2017-12-12 | Snap Inc. | User interface for accessing media at a geographic location |
US9854219B2 (en) | 2014-12-19 | 2017-12-26 | Snap Inc. | Gallery of videos set to an audio time line |
US9866999B1 (en) | 2014-01-12 | 2018-01-09 | Investment Asset Holdings Llc | Location-based messaging |
US9882907B1 (en) | 2012-11-08 | 2018-01-30 | Snap Inc. | Apparatus and method for single action control of social network profile access |
US9936030B2 (en) | 2014-01-03 | 2018-04-03 | Investel Capital Corporation | User content sharing system and method with location-based external content integration |
US10055717B1 (en) | 2014-08-22 | 2018-08-21 | Snap Inc. | Message processor with application prompts |
US10084735B1 (en) | 2014-02-21 | 2018-09-25 | Snap Inc. | Apparatus and method for alternate channel communication initiated through a common message thread |
US10123166B2 (en) | 2015-01-26 | 2018-11-06 | Snap Inc. | Content request by location |
US10133705B1 (en) | 2015-01-19 | 2018-11-20 | Snap Inc. | Multichannel system |
US10135949B1 (en) | 2015-05-05 | 2018-11-20 | Snap Inc. | Systems and methods for story and sub-story navigation |
US10157449B1 (en) | 2015-01-09 | 2018-12-18 | Snap Inc. | Geo-location-based image filters |
US10165402B1 (en) | 2016-06-28 | 2018-12-25 | Snap Inc. | System to track engagement of media items |
US10203855B2 (en) | 2016-12-09 | 2019-02-12 | Snap Inc. | Customized user-controlled media overlays |
US10219111B1 (en) | 2018-04-18 | 2019-02-26 | Snap Inc. | Visitation tracking system |
US10223397B1 (en) | 2015-03-13 | 2019-03-05 | Snap Inc. | Social graph based co-location of network users |
US10284508B1 (en) | 2014-10-02 | 2019-05-07 | Snap Inc. | Ephemeral gallery of ephemeral messages with opt-in permanence |
US10311916B2 (en) | 2014-12-19 | 2019-06-04 | Snap Inc. | Gallery of videos set to an audio time line |
US10319149B1 (en) | 2017-02-17 | 2019-06-11 | Snap Inc. | Augmented reality anamorphosis system |
US10327096B1 (en) | 2018-03-06 | 2019-06-18 | Snap Inc. | Geo-fence selection system |
US10334307B2 (en) | 2011-07-12 | 2019-06-25 | Snap Inc. | Methods and systems of providing visual content editing functions |
US10348662B2 (en) | 2016-07-19 | 2019-07-09 | Snap Inc. | Generating customized electronic messaging graphics |
US10354425B2 (en) | 2015-12-18 | 2019-07-16 | Snap Inc. | Method and system for providing context relevant media augmentation |
US10366543B1 (en) | 2015-10-30 | 2019-07-30 | Snap Inc. | Image based tracking in augmented reality systems |
US10387730B1 (en) | 2017-04-20 | 2019-08-20 | Snap Inc. | Augmented reality typography personalization system |
US10387514B1 (en) | 2016-06-30 | 2019-08-20 | Snap Inc. | Automated content curation and communication |
US10423983B2 (en) | 2014-09-16 | 2019-09-24 | Snap Inc. | Determining targeting information based on a predictive targeting model |
US10430838B1 (en) | 2016-06-28 | 2019-10-01 | Snap Inc. | Methods and systems for generation, curation, and presentation of media collections with automated advertising |
US10439972B1 (en) | 2013-05-30 | 2019-10-08 | Snap Inc. | Apparatus and method for maintaining a message thread with opt-in permanence for entries |
US10474321B2 (en) | 2015-11-30 | 2019-11-12 | Snap Inc. | Network resource location linking and visual content sharing |
US10499191B1 (en) | 2017-10-09 | 2019-12-03 | Snap Inc. | Context sensitive presentation of content |
US10523625B1 (en) | 2017-03-09 | 2019-12-31 | Snap Inc. | Restricted group content collection |
US10582277B2 (en) | 2017-03-27 | 2020-03-03 | Snap Inc. | Generating a stitched data stream |
US10581782B2 (en) | 2017-03-27 | 2020-03-03 | Snap Inc. | Generating a stitched data stream |
US10592574B2 (en) | 2015-05-05 | 2020-03-17 | Snap Inc. | Systems and methods for automated local story generation and curation |
US10616239B2 (en) | 2015-03-18 | 2020-04-07 | Snap Inc. | Geo-fence authorization provisioning |
US10623666B2 (en) | 2016-11-07 | 2020-04-14 | Snap Inc. | Selective identification and order of image modifiers |
US10638256B1 (en) | 2016-06-20 | 2020-04-28 | Pipbin, Inc. | System for distribution and display of mobile targeted augmented reality content |
US10679389B2 (en) | 2016-02-26 | 2020-06-09 | Snap Inc. | Methods and systems for generation, curation, and presentation of media collections |
US10678818B2 (en) | 2018-01-03 | 2020-06-09 | Snap Inc. | Tag distribution visualization system |
US10679393B2 (en) | 2018-07-24 | 2020-06-09 | Snap Inc. | Conditional modification of augmented reality object |
US10740974B1 (en) | 2017-09-15 | 2020-08-11 | Snap Inc. | Augmented reality system |
US10805696B1 (en) | 2016-06-20 | 2020-10-13 | Pipbin, Inc. | System for recording and targeting tagged content of user interest |
US10817898B2 (en) | 2015-08-13 | 2020-10-27 | Placed, Llc | Determining exposures to content presented by physical objects |
US10824654B2 (en) | 2014-09-18 | 2020-11-03 | Snap Inc. | Geolocation-based pictographs |
US10834525B2 (en) | 2016-02-26 | 2020-11-10 | Snap Inc. | Generation, curation, and presentation of media collections |
US10839219B1 (en) | 2016-06-20 | 2020-11-17 | Pipbin, Inc. | System for curation, distribution and display of location-dependent augmented reality content |
US10862951B1 (en) | 2007-01-05 | 2020-12-08 | Snap Inc. | Real-time display of multiple images |
US10885136B1 (en) | 2018-02-28 | 2021-01-05 | Snap Inc. | Audience filtering system |
US10915911B2 (en) | 2017-02-03 | 2021-02-09 | Snap Inc. | System to determine a price-schedule to distribute media content |
US10933311B2 (en) | 2018-03-14 | 2021-03-02 | Snap Inc. | Generating collectible items based on location information |
US10948717B1 (en) | 2015-03-23 | 2021-03-16 | Snap Inc. | Reducing boot time and power consumption in wearable display systems |
US10952013B1 (en) | 2017-04-27 | 2021-03-16 | Snap Inc. | Selective location-based identity communication |
US10963529B1 (en) | 2017-04-27 | 2021-03-30 | Snap Inc. | Location-based search mechanism in a graphical user interface |
US10979752B1 (en) | 2018-02-28 | 2021-04-13 | Snap Inc. | Generating media content items based on location information |
US10993069B2 (en) | 2015-07-16 | 2021-04-27 | Snap Inc. | Dynamically adaptive media content delivery |
US10997760B2 (en) | 2018-08-31 | 2021-05-04 | Snap Inc. | Augmented reality anthropomorphization system |
US10997783B2 (en) | 2015-11-30 | 2021-05-04 | Snap Inc. | Image and point cloud based tracking and in augmented reality systems |
US11017173B1 (en) | 2017-12-22 | 2021-05-25 | Snap Inc. | Named entity recognition visual context and caption data |
US11023514B2 (en) | 2016-02-26 | 2021-06-01 | Snap Inc. | Methods and systems for generation, curation, and presentation of media collections |
US11030787B2 (en) | 2017-10-30 | 2021-06-08 | Snap Inc. | Mobile-based cartographic control of display content |
US11037372B2 (en) | 2017-03-06 | 2021-06-15 | Snap Inc. | Virtual vision system |
US11044393B1 (en) | 2016-06-20 | 2021-06-22 | Pipbin, Inc. | System for curation and display of location-dependent augmented reality content in an augmented estate system |
US11128715B1 (en) | 2019-12-30 | 2021-09-21 | Snap Inc. | Physical friend proximity in chat |
US11140331B2 (en) * | 2017-06-21 | 2021-10-05 | Fujifilm Corporation | Image capturing apparatus, control method for image capturing apparatus, and control program for image capturing apparatus |
US11163941B1 (en) | 2018-03-30 | 2021-11-02 | Snap Inc. | Annotating a collection of media content items |
US11170393B1 (en) | 2017-04-11 | 2021-11-09 | Snap Inc. | System to calculate an engagement score of location based media content |
US11182383B1 (en) | 2012-02-24 | 2021-11-23 | Placed, Llc | System and method for data collection to validate location data |
US11189299B1 (en) | 2017-02-20 | 2021-11-30 | Snap Inc. | Augmented reality speech balloon system |
US11199957B1 (en) | 2018-11-30 | 2021-12-14 | Snap Inc. | Generating customized avatars based on location information |
US11201981B1 (en) | 2016-06-20 | 2021-12-14 | Pipbin, Inc. | System for notification of user accessibility of curated location-dependent content in an augmented estate |
US11206615B2 (en) | 2019-05-30 | 2021-12-21 | Snap Inc. | Wearable device location systems |
US11216869B2 (en) | 2014-09-23 | 2022-01-04 | Snap Inc. | User interface to augment an image using geolocation |
US11218838B2 (en) | 2019-10-31 | 2022-01-04 | Snap Inc. | Focused map-based context information surfacing |
US11228551B1 (en) | 2020-02-12 | 2022-01-18 | Snap Inc. | Multiple gateway message exchange |
US11232040B1 (en) | 2017-04-28 | 2022-01-25 | Snap Inc. | Precaching unlockable data elements |
US11250075B1 (en) | 2017-02-17 | 2022-02-15 | Snap Inc. | Searching social media content |
US11249614B2 (en) | 2019-03-28 | 2022-02-15 | Snap Inc. | Generating personalized map interface with enhanced icons |
US11265273B1 (en) | 2017-12-01 | 2022-03-01 | Snap, Inc. | Dynamic media overlay with smart widget |
US11290851B2 (en) | 2020-06-15 | 2022-03-29 | Snap Inc. | Location sharing using offline and online objects |
US11294936B1 (en) | 2019-01-30 | 2022-04-05 | Snap Inc. | Adaptive spatial density based clustering |
US11301117B2 (en) | 2019-03-08 | 2022-04-12 | Snap Inc. | Contextual information in chat |
US11314776B2 (en) | 2020-06-15 | 2022-04-26 | Snap Inc. | Location sharing using friend list versions |
US11343323B2 (en) | 2019-12-31 | 2022-05-24 | Snap Inc. | Augmented reality objects registry |
US11361493B2 (en) | 2019-04-01 | 2022-06-14 | Snap Inc. | Semantic texture mapping system |
US11388226B1 (en) | 2015-01-13 | 2022-07-12 | Snap Inc. | Guided personal identity based actions |
US11430091B2 (en) | 2020-03-27 | 2022-08-30 | Snap Inc. | Location mapping for large scale augmented-reality |
US11429618B2 (en) | 2019-12-30 | 2022-08-30 | Snap Inc. | Surfacing augmented reality objects |
US11455082B2 (en) | 2018-09-28 | 2022-09-27 | Snap Inc. | Collaborative achievement interface |
US11475254B1 (en) | 2017-09-08 | 2022-10-18 | Snap Inc. | Multimodal entity identification |
US11483267B2 (en) | 2020-06-15 | 2022-10-25 | Snap Inc. | Location sharing using different rate-limited links |
US11500525B2 (en) | 2019-02-25 | 2022-11-15 | Snap Inc. | Custom media overlay system |
US11503432B2 (en) | 2020-06-15 | 2022-11-15 | Snap Inc. | Scalable real-time location sharing framework |
US11507614B1 (en) | 2018-02-13 | 2022-11-22 | Snap Inc. | Icon based tagging |
US11516167B2 (en) | 2020-03-05 | 2022-11-29 | Snap Inc. | Storing data based on device location |
US11558709B2 (en) | 2018-11-30 | 2023-01-17 | Snap Inc. | Position service to determine relative position to map features |
US11574431B2 (en) | 2019-02-26 | 2023-02-07 | Snap Inc. | Avatar based on weather |
US11601888B2 (en) | 2021-03-29 | 2023-03-07 | Snap Inc. | Determining location using multi-source geolocation data |
US11601783B2 (en) | 2019-06-07 | 2023-03-07 | Snap Inc. | Detection of a physical collision between two client devices in a location sharing system |
US11606755B2 (en) | 2019-05-30 | 2023-03-14 | Snap Inc. | Wearable device location systems architecture |
US11616745B2 (en) | 2017-01-09 | 2023-03-28 | Snap Inc. | Contextual generation and selection of customized media content |
US11619501B2 (en) | 2020-03-11 | 2023-04-04 | Snap Inc. | Avatar based on trip |
US11625443B2 (en) | 2014-06-05 | 2023-04-11 | Snap Inc. | Web document enhancement |
US11631276B2 (en) | 2016-03-31 | 2023-04-18 | Snap Inc. | Automated avatar generation |
US11645324B2 (en) | 2021-03-31 | 2023-05-09 | Snap Inc. | Location-based timeline media content system |
US11676378B2 (en) | 2020-06-29 | 2023-06-13 | Snap Inc. | Providing travel-based augmented reality content with a captured image |
US11675831B2 (en) | 2017-05-31 | 2023-06-13 | Snap Inc. | Geolocation based playlists |
US11714535B2 (en) | 2019-07-11 | 2023-08-01 | Snap Inc. | Edge gesture interface with smart interactions |
US11729343B2 (en) | 2019-12-30 | 2023-08-15 | Snap Inc. | Including video feed in message thread |
US11734712B2 (en) | 2012-02-24 | 2023-08-22 | Foursquare Labs, Inc. | Attributing in-store visits to media consumption based on data collected from user devices |
US11751015B2 (en) | 2019-01-16 | 2023-09-05 | Snap Inc. | Location-based context information sharing in a messaging system |
US11776256B2 (en) | 2020-03-27 | 2023-10-03 | Snap Inc. | Shared augmented reality system |
US11785161B1 (en) | 2016-06-20 | 2023-10-10 | Pipbin, Inc. | System for user accessibility of tagged curated augmented reality content |
US11799811B2 (en) | 2018-10-31 | 2023-10-24 | Snap Inc. | Messaging and gaming applications communication platform |
US11800074B2 (en) * | 2015-03-30 | 2023-10-24 | Nikon Corporation | Electronic device and computer program product |
US11809624B2 (en) | 2019-02-13 | 2023-11-07 | Snap Inc. | Sleep detection in a location sharing system |
US11816853B2 (en) | 2016-08-30 | 2023-11-14 | Snap Inc. | Systems and methods for simultaneous localization and mapping |
US11821742B2 (en) | 2019-09-26 | 2023-11-21 | Snap Inc. | Travel based notifications |
US11829834B2 (en) | 2021-10-29 | 2023-11-28 | Snap Inc. | Extended QR code |
US11843456B2 (en) | 2016-10-24 | 2023-12-12 | Snap Inc. | Generating and displaying customized avatars in media overlays |
US11842411B2 (en) | 2017-04-27 | 2023-12-12 | Snap Inc. | Location-based virtual avatars |
US11852554B1 (en) | 2019-03-21 | 2023-12-26 | Snap Inc. | Barometer calibration in a location sharing system |
US11860888B2 (en) | 2018-05-22 | 2024-01-02 | Snap Inc. | Event detection system |
US11868414B1 (en) | 2019-03-14 | 2024-01-09 | Snap Inc. | Graph-based prediction for contact suggestion in a location sharing system |
US11870743B1 (en) | 2017-01-23 | 2024-01-09 | Snap Inc. | Customized digital avatar accessories |
US11877211B2 (en) | 2019-01-14 | 2024-01-16 | Snap Inc. | Destination sharing in location sharing system |
US11876941B1 (en) | 2016-06-20 | 2024-01-16 | Pipbin, Inc. | Clickable augmented reality content manager, system, and network |
US11893208B2 (en) | 2019-12-31 | 2024-02-06 | Snap Inc. | Combined map icon with action indicator |
US11900418B2 (en) | 2016-04-04 | 2024-02-13 | Snap Inc. | Mutable geo-fencing system |
US11925869B2 (en) | 2012-05-08 | 2024-03-12 | Snap Inc. | System and method for generating and displaying avatars |
US11943192B2 (en) | 2020-08-31 | 2024-03-26 | Snap Inc. | Co-location connection service |
US11961116B2 (en) | 2020-10-26 | 2024-04-16 | Foursquare Labs, Inc. | Determining exposures to content presented by physical objects |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101795602B1 (en) * | 2011-08-12 | 2017-11-08 | 삼성전자주식회사 | Digital imaging apparatus and controlling method of thereof |
KR101805005B1 (en) * | 2011-11-07 | 2018-01-10 | 삼성전자주식회사 | Digital photographing apparatus |
US9049382B2 (en) * | 2012-04-05 | 2015-06-02 | Canon Kabushiki Kaisha | Image processing apparatus and image processing method |
KR101680446B1 (en) * | 2015-02-09 | 2016-11-29 | 주식회사 넥서스칩스 | Creation device for color table, correction and control device for camera image and method thereof |
CN105163099A (en) * | 2015-10-30 | 2015-12-16 | 努比亚技术有限公司 | While balance adjustment method and device and mobile terminal |
CN106385573A (en) * | 2016-09-06 | 2017-02-08 | 努比亚技术有限公司 | Picture processing method and terminal |
CN107071274B (en) * | 2017-03-13 | 2020-08-21 | 麒和科技(南京)有限公司 | Distortion processing method and terminal |
CN111064938A (en) * | 2019-12-30 | 2020-04-24 | 西安易朴通讯技术有限公司 | Multi-skin-color-weighted white balance correction method, system and storage medium thereof |
Citations (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6344907B1 (en) * | 1997-05-30 | 2002-02-05 | Fuji Photo Film Co., Ltd. | Image modification apparatus and method |
US20020130959A1 (en) * | 2001-01-12 | 2002-09-19 | Mcgarvey James E. | Venue customizable white balance digital camera system |
US20030122951A1 (en) * | 2001-12-27 | 2003-07-03 | Manabu Hara | Image processing apparatus and image processing method |
US20030146984A1 (en) * | 2002-02-05 | 2003-08-07 | Norihiro Sakaguchi | Image device having a selectively adapted white balance control |
US20050128316A1 (en) * | 2003-12-12 | 2005-06-16 | Canon Kabushiki Kaisha | Image processing system |
US20050212819A1 (en) * | 2004-03-29 | 2005-09-29 | Fuji Photo Film Co., Ltd. | Image display method, image display apparatus and image display program |
US6958773B1 (en) * | 1999-07-15 | 2005-10-25 | Pentax Corporation | White balance control amount calculating device |
US20060081760A1 (en) * | 2004-10-15 | 2006-04-20 | Tetsuo Kikuchi | Camera having focus detection device |
US20060087566A1 (en) * | 2004-10-12 | 2006-04-27 | Pentax Corporation | White balance adjustment device |
US20070058064A1 (en) * | 2005-09-14 | 2007-03-15 | Sony Corporation | Image processing apparatus and method, and program therefor |
US20070065137A1 (en) * | 2005-09-21 | 2007-03-22 | Sony Corporation | Photographic device, method of processing information, and program |
US7324749B2 (en) * | 2004-03-25 | 2008-01-29 | Fujifilm Corporation | Image display method, image display apparatus and image display program |
US20090027515A1 (en) * | 2007-07-26 | 2009-01-29 | Atsushi Maruyama | Image pickup apparatus |
US20090180002A1 (en) * | 2002-07-15 | 2009-07-16 | Olympus Corporation | White balance processing apparatus and processing method thereof |
US7656451B2 (en) * | 2006-01-18 | 2010-02-02 | Casio Computer Co., Ltd. | Camera apparatus and imaging method |
US7720258B1 (en) * | 2006-01-26 | 2010-05-18 | Adobe Systems Incorporated | Structured comparison of objects from similar images |
US7777784B2 (en) * | 2002-06-25 | 2010-08-17 | Hewlett-Packard Development Company, L.P. | Apparatus and method for generating multiple images from a single image |
US7839436B2 (en) * | 2006-09-13 | 2010-11-23 | Hoya Corporation | Photographing apparatus with setting for white balance adjustment |
US7978248B2 (en) * | 2007-09-14 | 2011-07-12 | Sony Corporation | Data processing apparatus and data processing method for displaying image capture mode candidates |
US8040412B2 (en) * | 2005-12-26 | 2011-10-18 | Panasonic Corporation | Imaging apparatus and image data correcting method |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPWO2005076637A1 (en) * | 2004-02-10 | 2007-08-02 | 松下電器産業株式会社 | White balance adjusting device and video display device |
JP2006074710A (en) | 2004-08-04 | 2006-03-16 | Konica Minolta Photo Imaging Inc | Digital camera and camera system |
KR20070053543A (en) * | 2005-11-21 | 2007-05-25 | 삼성전자주식회사 | White balance adjustment apparatus and method |
JP4708192B2 (en) * | 2006-01-10 | 2011-06-22 | パナソニック株式会社 | Dynamic camera color correction device and video search device using the same |
US7804525B2 (en) * | 2006-08-24 | 2010-09-28 | Micron Technology, Inc. | Method, apparatus, and system for selecting pixels for automatic white balance processing |
US7679786B2 (en) * | 2006-09-06 | 2010-03-16 | Eastman Kodak Company | Color correction method |
JP4254841B2 (en) * | 2006-10-20 | 2009-04-15 | ソニー株式会社 | Imaging apparatus, imaging method, image processing apparatus, image processing method, and image processing program |
JP4702635B2 (en) * | 2007-07-17 | 2011-06-15 | 富士フイルム株式会社 | AUTO WHITE BALANCE CORRECTION VALUE CALCULATION DEVICE, METHOD, PROGRAM, AND IMAGING DEVICE |
-
2009
- 2009-02-20 KR KR1020090014410A patent/KR101595254B1/en active IP Right Grant
-
2010
- 2010-02-19 US US12/708,642 patent/US8754958B2/en active Active
- 2010-02-20 CN CN201610195693.3A patent/CN105657393B/en not_active Expired - Fee Related
- 2010-02-20 CN CN201010119512A patent/CN101815223A/en active Pending
Patent Citations (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6344907B1 (en) * | 1997-05-30 | 2002-02-05 | Fuji Photo Film Co., Ltd. | Image modification apparatus and method |
US6958773B1 (en) * | 1999-07-15 | 2005-10-25 | Pentax Corporation | White balance control amount calculating device |
US20020130959A1 (en) * | 2001-01-12 | 2002-09-19 | Mcgarvey James E. | Venue customizable white balance digital camera system |
US20030122951A1 (en) * | 2001-12-27 | 2003-07-03 | Manabu Hara | Image processing apparatus and image processing method |
US20030146984A1 (en) * | 2002-02-05 | 2003-08-07 | Norihiro Sakaguchi | Image device having a selectively adapted white balance control |
US7777784B2 (en) * | 2002-06-25 | 2010-08-17 | Hewlett-Packard Development Company, L.P. | Apparatus and method for generating multiple images from a single image |
US20090180002A1 (en) * | 2002-07-15 | 2009-07-16 | Olympus Corporation | White balance processing apparatus and processing method thereof |
US20050128316A1 (en) * | 2003-12-12 | 2005-06-16 | Canon Kabushiki Kaisha | Image processing system |
US7324749B2 (en) * | 2004-03-25 | 2008-01-29 | Fujifilm Corporation | Image display method, image display apparatus and image display program |
US20050212819A1 (en) * | 2004-03-29 | 2005-09-29 | Fuji Photo Film Co., Ltd. | Image display method, image display apparatus and image display program |
US20060087566A1 (en) * | 2004-10-12 | 2006-04-27 | Pentax Corporation | White balance adjustment device |
US7551208B2 (en) * | 2004-10-12 | 2009-06-23 | Hoya Corporation | Device for adjusting white balance based on a selected part field of a displayed image |
US20060081760A1 (en) * | 2004-10-15 | 2006-04-20 | Tetsuo Kikuchi | Camera having focus detection device |
US20070058064A1 (en) * | 2005-09-14 | 2007-03-15 | Sony Corporation | Image processing apparatus and method, and program therefor |
US20070065137A1 (en) * | 2005-09-21 | 2007-03-22 | Sony Corporation | Photographic device, method of processing information, and program |
US8040412B2 (en) * | 2005-12-26 | 2011-10-18 | Panasonic Corporation | Imaging apparatus and image data correcting method |
US7656451B2 (en) * | 2006-01-18 | 2010-02-02 | Casio Computer Co., Ltd. | Camera apparatus and imaging method |
US7720258B1 (en) * | 2006-01-26 | 2010-05-18 | Adobe Systems Incorporated | Structured comparison of objects from similar images |
US7839436B2 (en) * | 2006-09-13 | 2010-11-23 | Hoya Corporation | Photographing apparatus with setting for white balance adjustment |
US20090027515A1 (en) * | 2007-07-26 | 2009-01-29 | Atsushi Maruyama | Image pickup apparatus |
US7978248B2 (en) * | 2007-09-14 | 2011-07-12 | Sony Corporation | Data processing apparatus and data processing method for displaying image capture mode candidates |
Cited By (346)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10862951B1 (en) | 2007-01-05 | 2020-12-08 | Snap Inc. | Real-time display of multiple images |
US11588770B2 (en) | 2007-01-05 | 2023-02-21 | Snap Inc. | Real-time display of multiple images |
CN102209246A (en) * | 2011-05-23 | 2011-10-05 | 北京工业大学 | Real-time video white balance processing system |
US11750875B2 (en) | 2011-07-12 | 2023-09-05 | Snap Inc. | Providing visual content editing functions |
US11451856B2 (en) | 2011-07-12 | 2022-09-20 | Snap Inc. | Providing visual content editing functions |
US10999623B2 (en) | 2011-07-12 | 2021-05-04 | Snap Inc. | Providing visual content editing functions |
US10334307B2 (en) | 2011-07-12 | 2019-06-25 | Snap Inc. | Methods and systems of providing visual content editing functions |
US11182383B1 (en) | 2012-02-24 | 2021-11-23 | Placed, Llc | System and method for data collection to validate location data |
US11734712B2 (en) | 2012-02-24 | 2023-08-22 | Foursquare Labs, Inc. | Attributing in-store visits to media consumption based on data collected from user devices |
US11925869B2 (en) | 2012-05-08 | 2024-03-12 | Snap Inc. | System and method for generating and displaying avatars |
US10169924B2 (en) | 2012-08-22 | 2019-01-01 | Snaps Media Inc. | Augmented reality virtual content platform apparatuses, methods and systems |
US9792733B2 (en) | 2012-08-22 | 2017-10-17 | Snaps Media, Inc. | Augmented reality virtual content platform apparatuses, methods and systems |
US9721394B2 (en) | 2012-08-22 | 2017-08-01 | Snaps Media, Inc. | Augmented reality virtual content platform apparatuses, methods and systems |
US9882907B1 (en) | 2012-11-08 | 2018-01-30 | Snap Inc. | Apparatus and method for single action control of social network profile access |
US11252158B2 (en) | 2012-11-08 | 2022-02-15 | Snap Inc. | Interactive user-interface to adjust access privileges |
US10887308B1 (en) | 2012-11-08 | 2021-01-05 | Snap Inc. | Interactive user-interface to adjust access privileges |
US10841511B1 (en) | 2013-03-14 | 2020-11-17 | Samsung Electronics Co., Ltd. | Electronic device and method for image processing |
US10506176B2 (en) | 2013-03-14 | 2019-12-10 | Samsung Electronics Co., Ltd. | Electronic device and method for image processing |
US9571736B2 (en) * | 2013-03-14 | 2017-02-14 | Samsung Electronics Co., Ltd. | Electronic device and method for image processing |
US9674462B2 (en) | 2013-03-14 | 2017-06-06 | Samsung Electronics Co., Ltd. | Electronic device and method for image processing |
US20140267867A1 (en) * | 2013-03-14 | 2014-09-18 | Samsung Electronics Co., Ltd. | Electronic device and method for image processing |
US10841510B2 (en) | 2013-03-14 | 2020-11-17 | Samsung Electronics Co., Ltd. | Electronic device and method for image processing |
US10284788B2 (en) | 2013-03-14 | 2019-05-07 | Samsung Electronics Co., Ltd. | Electronic device and method for image processing |
US20140301638A1 (en) * | 2013-04-08 | 2014-10-09 | Samsung Electronics Co., Ltd. | Color extraction-based image processing method, computer-readable storage medium storing the same, and digital image apparatus |
US10439972B1 (en) | 2013-05-30 | 2019-10-08 | Snap Inc. | Apparatus and method for maintaining a message thread with opt-in permanence for entries |
US11509618B2 (en) | 2013-05-30 | 2022-11-22 | Snap Inc. | Maintaining a message thread with opt-in permanence for entries |
US10587552B1 (en) | 2013-05-30 | 2020-03-10 | Snap Inc. | Apparatus and method for maintaining a message thread with opt-in permanence for entries |
US11134046B2 (en) | 2013-05-30 | 2021-09-28 | Snap Inc. | Apparatus and method for maintaining a message thread with opt-in permanence for entries |
US9742713B2 (en) | 2013-05-30 | 2017-08-22 | Snap Inc. | Apparatus and method for maintaining a message thread with opt-in permanence for entries |
US9705831B2 (en) | 2013-05-30 | 2017-07-11 | Snap Inc. | Apparatus and method for maintaining a message thread with opt-in permanence for entries |
US11115361B2 (en) | 2013-05-30 | 2021-09-07 | Snap Inc. | Apparatus and method for maintaining a message thread with opt-in permanence for entries |
US20150002693A1 (en) * | 2013-06-26 | 2015-01-01 | Nvidia Corporation | Method and system for performing white balancing operations on captured images |
US9756222B2 (en) * | 2013-06-26 | 2017-09-05 | Nvidia Corporation | Method and system for performing white balancing operations on captured images |
US9826208B2 (en) | 2013-06-26 | 2017-11-21 | Nvidia Corporation | Method and system for generating weights for use in white balancing an image |
US11546388B2 (en) | 2013-11-26 | 2023-01-03 | Snap Inc. | Method and system for integrating real time communication features in applications |
US9794303B1 (en) | 2013-11-26 | 2017-10-17 | Snap Inc. | Method and system for integrating real time communication features in applications |
US10069876B1 (en) | 2013-11-26 | 2018-09-04 | Snap Inc. | Method and system for integrating real time communication features in applications |
US11102253B2 (en) | 2013-11-26 | 2021-08-24 | Snap Inc. | Method and system for integrating real time communication features in applications |
US9083770B1 (en) | 2013-11-26 | 2015-07-14 | Snapchat, Inc. | Method and system for integrating real time communication features in applications |
US10681092B1 (en) | 2013-11-26 | 2020-06-09 | Snap Inc. | Method and system for integrating real time communication features in applications |
US9936030B2 (en) | 2014-01-03 | 2018-04-03 | Investel Capital Corporation | User content sharing system and method with location-based external content integration |
US9866999B1 (en) | 2014-01-12 | 2018-01-09 | Investment Asset Holdings Llc | Location-based messaging |
US10080102B1 (en) | 2014-01-12 | 2018-09-18 | Investment Asset Holdings Llc | Location-based messaging |
US10349209B1 (en) | 2014-01-12 | 2019-07-09 | Investment Asset Holdings Llc | Location-based messaging |
US10082926B1 (en) | 2014-02-21 | 2018-09-25 | Snap Inc. | Apparatus and method for alternate channel communication initiated through a common message thread |
US10949049B1 (en) | 2014-02-21 | 2021-03-16 | Snap Inc. | Apparatus and method for alternate channel communication initiated through a common message thread |
US10958605B1 (en) | 2014-02-21 | 2021-03-23 | Snap Inc. | Apparatus and method for alternate channel communication initiated through a common message thread |
US11463393B2 (en) | 2014-02-21 | 2022-10-04 | Snap Inc. | Apparatus and method for alternate channel communication initiated through a common message thread |
US11463394B2 (en) | 2014-02-21 | 2022-10-04 | Snap Inc. | Apparatus and method for alternate channel communication initiated through a common message thread |
US10084735B1 (en) | 2014-02-21 | 2018-09-25 | Snap Inc. | Apparatus and method for alternate channel communication initiated through a common message thread |
US11902235B2 (en) | 2014-02-21 | 2024-02-13 | Snap Inc. | Apparatus and method for alternate channel communication initiated through a common message thread |
US9407712B1 (en) | 2014-03-07 | 2016-08-02 | Snapchat, Inc. | Content delivery network for ephemeral objects |
US9237202B1 (en) | 2014-03-07 | 2016-01-12 | Snapchat, Inc. | Content delivery network for ephemeral objects |
US10817156B1 (en) | 2014-05-09 | 2020-10-27 | Snap Inc. | Dynamic configuration of application component tiles |
US11310183B2 (en) | 2014-05-09 | 2022-04-19 | Snap Inc. | Dynamic configuration of application component tiles |
US11743219B2 (en) | 2014-05-09 | 2023-08-29 | Snap Inc. | Dynamic configuration of application component tiles |
US9276886B1 (en) | 2014-05-09 | 2016-03-01 | Snapchat, Inc. | Apparatus and method for dynamically configuring application component tiles |
US10990697B2 (en) | 2014-05-28 | 2021-04-27 | Snap Inc. | Apparatus and method for automated privacy protection in distributed images |
US10572681B1 (en) | 2014-05-28 | 2020-02-25 | Snap Inc. | Apparatus and method for automated privacy protection in distributed images |
US9396354B1 (en) | 2014-05-28 | 2016-07-19 | Snapchat, Inc. | Apparatus and method for automated privacy protection in distributed images |
US9785796B1 (en) | 2014-05-28 | 2017-10-10 | Snap Inc. | Apparatus and method for automated privacy protection in distributed images |
US11625443B2 (en) | 2014-06-05 | 2023-04-11 | Snap Inc. | Web document enhancement |
US11921805B2 (en) | 2014-06-05 | 2024-03-05 | Snap Inc. | Web document enhancement |
US10524087B1 (en) | 2014-06-13 | 2019-12-31 | Snap Inc. | Message destination list mechanism |
US10779113B2 (en) | 2014-06-13 | 2020-09-15 | Snap Inc. | Prioritization of messages within a message collection |
US9094137B1 (en) | 2014-06-13 | 2015-07-28 | Snapchat, Inc. | Priority based placement of messages in a geo-location based event gallery |
US10623891B2 (en) | 2014-06-13 | 2020-04-14 | Snap Inc. | Prioritization of messages within a message collection |
US9113301B1 (en) | 2014-06-13 | 2015-08-18 | Snapchat, Inc. | Geo-location based event gallery |
US9825898B2 (en) | 2014-06-13 | 2017-11-21 | Snap Inc. | Prioritization of messages within a message collection |
US10200813B1 (en) | 2014-06-13 | 2019-02-05 | Snap Inc. | Geo-location based event gallery |
US9693191B2 (en) | 2014-06-13 | 2017-06-27 | Snap Inc. | Prioritization of messages within gallery |
US10182311B2 (en) | 2014-06-13 | 2019-01-15 | Snap Inc. | Prioritization of messages within a message collection |
US10659914B1 (en) | 2014-06-13 | 2020-05-19 | Snap Inc. | Geo-location based event gallery |
US11317240B2 (en) | 2014-06-13 | 2022-04-26 | Snap Inc. | Geo-location based event gallery |
US9532171B2 (en) | 2014-06-13 | 2016-12-27 | Snap Inc. | Geo-location based event gallery |
US10448201B1 (en) | 2014-06-13 | 2019-10-15 | Snap Inc. | Prioritization of messages within a message collection |
US9430783B1 (en) | 2014-06-13 | 2016-08-30 | Snapchat, Inc. | Prioritization of messages within gallery |
US11166121B2 (en) | 2014-06-13 | 2021-11-02 | Snap Inc. | Prioritization of messages within a message collection |
US20230020575A1 (en) * | 2014-07-07 | 2023-01-19 | Snap Inc. | Apparatus and method for supplying content aware photo filters |
US10432850B1 (en) | 2014-07-07 | 2019-10-01 | Snap Inc. | Apparatus and method for supplying content aware photo filters |
US11122200B2 (en) | 2014-07-07 | 2021-09-14 | Snap Inc. | Supplying content aware photo filters |
US11595569B2 (en) | 2014-07-07 | 2023-02-28 | Snap Inc. | Supplying content aware photo filters |
US9225897B1 (en) * | 2014-07-07 | 2015-12-29 | Snapchat, Inc. | Apparatus and method for supplying content aware photo filters |
US10348960B1 (en) * | 2014-07-07 | 2019-07-09 | Snap Inc. | Apparatus and method for supplying content aware photo filters |
US10154192B1 (en) | 2014-07-07 | 2018-12-11 | Snap Inc. | Apparatus and method for supplying content aware photo filters |
US11496673B1 (en) | 2014-07-07 | 2022-11-08 | Snap Inc. | Apparatus and method for supplying content aware photo filters |
US9407816B1 (en) | 2014-07-07 | 2016-08-02 | Snapchat, Inc. | Apparatus and method for supplying content aware photo filters |
US11849214B2 (en) * | 2014-07-07 | 2023-12-19 | Snap Inc. | Apparatus and method for supplying content aware photo filters |
US10602057B1 (en) * | 2014-07-07 | 2020-03-24 | Snap Inc. | Supplying content aware photo filters |
US10701262B1 (en) | 2014-07-07 | 2020-06-30 | Snap Inc. | Apparatus and method for supplying content aware photo filters |
US10055717B1 (en) | 2014-08-22 | 2018-08-21 | Snap Inc. | Message processor with application prompts |
US11017363B1 (en) | 2014-08-22 | 2021-05-25 | Snap Inc. | Message processor with application prompts |
US10423983B2 (en) | 2014-09-16 | 2019-09-24 | Snap Inc. | Determining targeting information based on a predictive targeting model |
US11625755B1 (en) | 2014-09-16 | 2023-04-11 | Foursquare Labs, Inc. | Determining targeting information based on a predictive targeting model |
US11741136B2 (en) | 2014-09-18 | 2023-08-29 | Snap Inc. | Geolocation-based pictographs |
US10824654B2 (en) | 2014-09-18 | 2020-11-03 | Snap Inc. | Geolocation-based pictographs |
US11281701B2 (en) | 2014-09-18 | 2022-03-22 | Snap Inc. | Geolocation-based pictographs |
US11216869B2 (en) | 2014-09-23 | 2022-01-04 | Snap Inc. | User interface to augment an image using geolocation |
US10476830B2 (en) | 2014-10-02 | 2019-11-12 | Snap Inc. | Ephemeral gallery of ephemeral messages |
US11855947B1 (en) | 2014-10-02 | 2023-12-26 | Snap Inc. | Gallery of ephemeral messages |
US11038829B1 (en) | 2014-10-02 | 2021-06-15 | Snap Inc. | Ephemeral gallery of ephemeral messages with opt-in permanence |
US10944710B1 (en) | 2014-10-02 | 2021-03-09 | Snap Inc. | Ephemeral gallery user interface with remaining gallery time indication |
US11411908B1 (en) | 2014-10-02 | 2022-08-09 | Snap Inc. | Ephemeral message gallery user interface with online viewing history indicia |
US10958608B1 (en) | 2014-10-02 | 2021-03-23 | Snap Inc. | Ephemeral gallery of visual media messages |
US9537811B2 (en) | 2014-10-02 | 2017-01-03 | Snap Inc. | Ephemeral gallery of ephemeral messages |
US10708210B1 (en) | 2014-10-02 | 2020-07-07 | Snap Inc. | Multi-user ephemeral message gallery |
US20170374003A1 (en) | 2014-10-02 | 2017-12-28 | Snapchat, Inc. | Ephemeral gallery of ephemeral messages |
US10284508B1 (en) | 2014-10-02 | 2019-05-07 | Snap Inc. | Ephemeral gallery of ephemeral messages with opt-in permanence |
US11522822B1 (en) | 2014-10-02 | 2022-12-06 | Snap Inc. | Ephemeral gallery elimination based on gallery and message timers |
US11012398B1 (en) | 2014-10-02 | 2021-05-18 | Snap Inc. | Ephemeral message gallery user interface with screenshot messages |
US10616476B1 (en) | 2014-11-12 | 2020-04-07 | Snap Inc. | User interface for accessing media at a geographic location |
US11956533B2 (en) | 2014-11-12 | 2024-04-09 | Snap Inc. | Accessing media at a geographic location |
US9843720B1 (en) | 2014-11-12 | 2017-12-12 | Snap Inc. | User interface for accessing media at a geographic location |
US11190679B2 (en) | 2014-11-12 | 2021-11-30 | Snap Inc. | Accessing media at a geographic location |
US10811053B2 (en) | 2014-12-19 | 2020-10-20 | Snap Inc. | Routing messages by message parameter |
US9385983B1 (en) | 2014-12-19 | 2016-07-05 | Snapchat, Inc. | Gallery of messages from individuals with a shared interest |
US11250887B2 (en) | 2014-12-19 | 2022-02-15 | Snap Inc. | Routing messages by message parameter |
US11783862B2 (en) | 2014-12-19 | 2023-10-10 | Snap Inc. | Routing messages by message parameter |
US9854219B2 (en) | 2014-12-19 | 2017-12-26 | Snap Inc. | Gallery of videos set to an audio time line |
US11803345B2 (en) | 2014-12-19 | 2023-10-31 | Snap Inc. | Gallery of messages from individuals with a shared interest |
US11372608B2 (en) | 2014-12-19 | 2022-06-28 | Snap Inc. | Gallery of messages from individuals with a shared interest |
US10580458B2 (en) | 2014-12-19 | 2020-03-03 | Snap Inc. | Gallery of videos set to an audio time line |
US10311916B2 (en) | 2014-12-19 | 2019-06-04 | Snap Inc. | Gallery of videos set to an audio time line |
US10514876B2 (en) | 2014-12-19 | 2019-12-24 | Snap Inc. | Gallery of messages from individuals with a shared interest |
US10157449B1 (en) | 2015-01-09 | 2018-12-18 | Snap Inc. | Geo-location-based image filters |
US10380720B1 (en) | 2015-01-09 | 2019-08-13 | Snap Inc. | Location-based image filters |
US11301960B2 (en) | 2015-01-09 | 2022-04-12 | Snap Inc. | Object recognition based image filters |
US11734342B2 (en) | 2015-01-09 | 2023-08-22 | Snap Inc. | Object recognition based image overlays |
US11388226B1 (en) | 2015-01-13 | 2022-07-12 | Snap Inc. | Guided personal identity based actions |
US10416845B1 (en) | 2015-01-19 | 2019-09-17 | Snap Inc. | Multichannel system |
US11249617B1 (en) | 2015-01-19 | 2022-02-15 | Snap Inc. | Multichannel system |
US10133705B1 (en) | 2015-01-19 | 2018-11-20 | Snap Inc. | Multichannel system |
CN104618645A (en) * | 2015-01-20 | 2015-05-13 | 广东欧珀移动通信有限公司 | Method and device for shooting through two cameras |
US11910267B2 (en) | 2015-01-26 | 2024-02-20 | Snap Inc. | Content request by location |
US10932085B1 (en) | 2015-01-26 | 2021-02-23 | Snap Inc. | Content request by location |
US11528579B2 (en) | 2015-01-26 | 2022-12-13 | Snap Inc. | Content request by location |
US10536800B1 (en) | 2015-01-26 | 2020-01-14 | Snap Inc. | Content request by location |
US10123166B2 (en) | 2015-01-26 | 2018-11-06 | Snap Inc. | Content request by location |
US10223397B1 (en) | 2015-03-13 | 2019-03-05 | Snap Inc. | Social graph based co-location of network users |
US11902287B2 (en) | 2015-03-18 | 2024-02-13 | Snap Inc. | Geo-fence authorization provisioning |
US10616239B2 (en) | 2015-03-18 | 2020-04-07 | Snap Inc. | Geo-fence authorization provisioning |
US10893055B2 (en) | 2015-03-18 | 2021-01-12 | Snap Inc. | Geo-fence authorization provisioning |
US11320651B2 (en) | 2015-03-23 | 2022-05-03 | Snap Inc. | Reducing boot time and power consumption in displaying data content |
US10948717B1 (en) | 2015-03-23 | 2021-03-16 | Snap Inc. | Reducing boot time and power consumption in wearable display systems |
US11662576B2 (en) | 2015-03-23 | 2023-05-30 | Snap Inc. | Reducing boot time and power consumption in displaying data content |
US11800074B2 (en) * | 2015-03-30 | 2023-10-24 | Nikon Corporation | Electronic device and computer program product |
US10911575B1 (en) | 2015-05-05 | 2021-02-02 | Snap Inc. | Systems and methods for story and sub-story navigation |
US11496544B2 (en) | 2015-05-05 | 2022-11-08 | Snap Inc. | Story and sub-story navigation |
US11392633B2 (en) | 2015-05-05 | 2022-07-19 | Snap Inc. | Systems and methods for automated local story generation and curation |
US10135949B1 (en) | 2015-05-05 | 2018-11-20 | Snap Inc. | Systems and methods for story and sub-story navigation |
US10592574B2 (en) | 2015-05-05 | 2020-03-17 | Snap Inc. | Systems and methods for automated local story generation and curation |
US11449539B2 (en) | 2015-05-05 | 2022-09-20 | Snap Inc. | Automated local story generation and curation |
US10993069B2 (en) | 2015-07-16 | 2021-04-27 | Snap Inc. | Dynamically adaptive media content delivery |
US10817898B2 (en) | 2015-08-13 | 2020-10-27 | Placed, Llc | Determining exposures to content presented by physical objects |
US10733802B2 (en) | 2015-10-30 | 2020-08-04 | Snap Inc. | Image based tracking in augmented reality systems |
US10366543B1 (en) | 2015-10-30 | 2019-07-30 | Snap Inc. | Image based tracking in augmented reality systems |
US11315331B2 (en) | 2015-10-30 | 2022-04-26 | Snap Inc. | Image based tracking in augmented reality systems |
US11769307B2 (en) | 2015-10-30 | 2023-09-26 | Snap Inc. | Image based tracking in augmented reality systems |
US10474321B2 (en) | 2015-11-30 | 2019-11-12 | Snap Inc. | Network resource location linking and visual content sharing |
US11599241B2 (en) | 2015-11-30 | 2023-03-07 | Snap Inc. | Network resource location linking and visual content sharing |
US11380051B2 (en) | 2015-11-30 | 2022-07-05 | Snap Inc. | Image and point cloud based tracking and in augmented reality systems |
US10997783B2 (en) | 2015-11-30 | 2021-05-04 | Snap Inc. | Image and point cloud based tracking and in augmented reality systems |
US10354425B2 (en) | 2015-12-18 | 2019-07-16 | Snap Inc. | Method and system for providing context relevant media augmentation |
US11468615B2 (en) | 2015-12-18 | 2022-10-11 | Snap Inc. | Media overlay publication system |
US11830117B2 (en) | 2015-12-18 | 2023-11-28 | Snap Inc | Media overlay publication system |
US10997758B1 (en) | 2015-12-18 | 2021-05-04 | Snap Inc. | Media overlay publication system |
US11197123B2 (en) | 2016-02-26 | 2021-12-07 | Snap Inc. | Generation, curation, and presentation of media collections |
US11611846B2 (en) | 2016-02-26 | 2023-03-21 | Snap Inc. | Generation, curation, and presentation of media collections |
US11889381B2 (en) | 2016-02-26 | 2024-01-30 | Snap Inc. | Generation, curation, and presentation of media collections |
US11023514B2 (en) | 2016-02-26 | 2021-06-01 | Snap Inc. | Methods and systems for generation, curation, and presentation of media collections |
US10834525B2 (en) | 2016-02-26 | 2020-11-10 | Snap Inc. | Generation, curation, and presentation of media collections |
US10679389B2 (en) | 2016-02-26 | 2020-06-09 | Snap Inc. | Methods and systems for generation, curation, and presentation of media collections |
US11631276B2 (en) | 2016-03-31 | 2023-04-18 | Snap Inc. | Automated avatar generation |
US11900418B2 (en) | 2016-04-04 | 2024-02-13 | Snap Inc. | Mutable geo-fencing system |
US10805696B1 (en) | 2016-06-20 | 2020-10-13 | Pipbin, Inc. | System for recording and targeting tagged content of user interest |
US10638256B1 (en) | 2016-06-20 | 2020-04-28 | Pipbin, Inc. | System for distribution and display of mobile targeted augmented reality content |
US11201981B1 (en) | 2016-06-20 | 2021-12-14 | Pipbin, Inc. | System for notification of user accessibility of curated location-dependent content in an augmented estate |
US11876941B1 (en) | 2016-06-20 | 2024-01-16 | Pipbin, Inc. | Clickable augmented reality content manager, system, and network |
US11044393B1 (en) | 2016-06-20 | 2021-06-22 | Pipbin, Inc. | System for curation and display of location-dependent augmented reality content in an augmented estate system |
US10839219B1 (en) | 2016-06-20 | 2020-11-17 | Pipbin, Inc. | System for curation, distribution and display of location-dependent augmented reality content |
US11785161B1 (en) | 2016-06-20 | 2023-10-10 | Pipbin, Inc. | System for user accessibility of tagged curated augmented reality content |
US10992836B2 (en) | 2016-06-20 | 2021-04-27 | Pipbin, Inc. | Augmented property system of curated augmented reality media elements |
US10219110B2 (en) | 2016-06-28 | 2019-02-26 | Snap Inc. | System to track engagement of media items |
US10785597B2 (en) | 2016-06-28 | 2020-09-22 | Snap Inc. | System to track engagement of media items |
US10327100B1 (en) | 2016-06-28 | 2019-06-18 | Snap Inc. | System to track engagement of media items |
US10506371B2 (en) | 2016-06-28 | 2019-12-10 | Snap Inc. | System to track engagement of media items |
US11640625B2 (en) | 2016-06-28 | 2023-05-02 | Snap Inc. | Generation, curation, and presentation of media collections with automated advertising |
US10430838B1 (en) | 2016-06-28 | 2019-10-01 | Snap Inc. | Methods and systems for generation, curation, and presentation of media collections with automated advertising |
US10165402B1 (en) | 2016-06-28 | 2018-12-25 | Snap Inc. | System to track engagement of media items |
US10885559B1 (en) | 2016-06-28 | 2021-01-05 | Snap Inc. | Generation, curation, and presentation of media collections with automated advertising |
US10735892B2 (en) | 2016-06-28 | 2020-08-04 | Snap Inc. | System to track engagement of media items |
US11445326B2 (en) | 2016-06-28 | 2022-09-13 | Snap Inc. | Track engagement of media items |
US11895068B2 (en) | 2016-06-30 | 2024-02-06 | Snap Inc. | Automated content curation and communication |
US11080351B1 (en) | 2016-06-30 | 2021-08-03 | Snap Inc. | Automated content curation and communication |
US10387514B1 (en) | 2016-06-30 | 2019-08-20 | Snap Inc. | Automated content curation and communication |
US11509615B2 (en) | 2016-07-19 | 2022-11-22 | Snap Inc. | Generating customized electronic messaging graphics |
US10348662B2 (en) | 2016-07-19 | 2019-07-09 | Snap Inc. | Generating customized electronic messaging graphics |
US11816853B2 (en) | 2016-08-30 | 2023-11-14 | Snap Inc. | Systems and methods for simultaneous localization and mapping |
US11876762B1 (en) | 2016-10-24 | 2024-01-16 | Snap Inc. | Generating and displaying customized avatars in media overlays |
US11843456B2 (en) | 2016-10-24 | 2023-12-12 | Snap Inc. | Generating and displaying customized avatars in media overlays |
US11750767B2 (en) | 2016-11-07 | 2023-09-05 | Snap Inc. | Selective identification and order of image modifiers |
US10623666B2 (en) | 2016-11-07 | 2020-04-14 | Snap Inc. | Selective identification and order of image modifiers |
US11233952B2 (en) | 2016-11-07 | 2022-01-25 | Snap Inc. | Selective identification and order of image modifiers |
US10754525B1 (en) | 2016-12-09 | 2020-08-25 | Snap Inc. | Customized media overlays |
US11397517B2 (en) | 2016-12-09 | 2022-07-26 | Snap Inc. | Customized media overlays |
US10203855B2 (en) | 2016-12-09 | 2019-02-12 | Snap Inc. | Customized user-controlled media overlays |
US11616745B2 (en) | 2017-01-09 | 2023-03-28 | Snap Inc. | Contextual generation and selection of customized media content |
US11870743B1 (en) | 2017-01-23 | 2024-01-09 | Snap Inc. | Customized digital avatar accessories |
US10915911B2 (en) | 2017-02-03 | 2021-02-09 | Snap Inc. | System to determine a price-schedule to distribute media content |
US10319149B1 (en) | 2017-02-17 | 2019-06-11 | Snap Inc. | Augmented reality anamorphosis system |
US11720640B2 (en) | 2017-02-17 | 2023-08-08 | Snap Inc. | Searching social media content |
US11861795B1 (en) | 2017-02-17 | 2024-01-02 | Snap Inc. | Augmented reality anamorphosis system |
US11250075B1 (en) | 2017-02-17 | 2022-02-15 | Snap Inc. | Searching social media content |
US11189299B1 (en) | 2017-02-20 | 2021-11-30 | Snap Inc. | Augmented reality speech balloon system |
US11748579B2 (en) | 2017-02-20 | 2023-09-05 | Snap Inc. | Augmented reality speech balloon system |
US11037372B2 (en) | 2017-03-06 | 2021-06-15 | Snap Inc. | Virtual vision system |
US11670057B2 (en) | 2017-03-06 | 2023-06-06 | Snap Inc. | Virtual vision system |
US11258749B2 (en) | 2017-03-09 | 2022-02-22 | Snap Inc. | Restricted group content collection |
US10523625B1 (en) | 2017-03-09 | 2019-12-31 | Snap Inc. | Restricted group content collection |
US10887269B1 (en) | 2017-03-09 | 2021-01-05 | Snap Inc. | Restricted group content collection |
US11297399B1 (en) | 2017-03-27 | 2022-04-05 | Snap Inc. | Generating a stitched data stream |
US10581782B2 (en) | 2017-03-27 | 2020-03-03 | Snap Inc. | Generating a stitched data stream |
US11558678B2 (en) | 2017-03-27 | 2023-01-17 | Snap Inc. | Generating a stitched data stream |
US11349796B2 (en) | 2017-03-27 | 2022-05-31 | Snap Inc. | Generating a stitched data stream |
US10582277B2 (en) | 2017-03-27 | 2020-03-03 | Snap Inc. | Generating a stitched data stream |
US11170393B1 (en) | 2017-04-11 | 2021-11-09 | Snap Inc. | System to calculate an engagement score of location based media content |
US11195018B1 (en) | 2017-04-20 | 2021-12-07 | Snap Inc. | Augmented reality typography personalization system |
US10387730B1 (en) | 2017-04-20 | 2019-08-20 | Snap Inc. | Augmented reality typography personalization system |
US10952013B1 (en) | 2017-04-27 | 2021-03-16 | Snap Inc. | Selective location-based identity communication |
US11385763B2 (en) | 2017-04-27 | 2022-07-12 | Snap Inc. | Map-based graphical user interface indicating geospatial activity metrics |
US11842411B2 (en) | 2017-04-27 | 2023-12-12 | Snap Inc. | Location-based virtual avatars |
US11474663B2 (en) | 2017-04-27 | 2022-10-18 | Snap Inc. | Location-based search mechanism in a graphical user interface |
US11418906B2 (en) | 2017-04-27 | 2022-08-16 | Snap Inc. | Selective location-based identity communication |
US11409407B2 (en) | 2017-04-27 | 2022-08-09 | Snap Inc. | Map-based graphical user interface indicating geospatial activity metrics |
US10963529B1 (en) | 2017-04-27 | 2021-03-30 | Snap Inc. | Location-based search mechanism in a graphical user interface |
US11392264B1 (en) | 2017-04-27 | 2022-07-19 | Snap Inc. | Map-based graphical user interface for multi-type social media galleries |
US11556221B2 (en) | 2017-04-27 | 2023-01-17 | Snap Inc. | Friend location sharing mechanism for social media platforms |
US11451956B1 (en) | 2017-04-27 | 2022-09-20 | Snap Inc. | Location privacy management on map-based social media platforms |
US11893647B2 (en) | 2017-04-27 | 2024-02-06 | Snap Inc. | Location-based virtual avatars |
US11782574B2 (en) | 2017-04-27 | 2023-10-10 | Snap Inc. | Map-based graphical user interface indicating geospatial activity metrics |
US11232040B1 (en) | 2017-04-28 | 2022-01-25 | Snap Inc. | Precaching unlockable data elements |
US11675831B2 (en) | 2017-05-31 | 2023-06-13 | Snap Inc. | Geolocation based playlists |
US11140331B2 (en) * | 2017-06-21 | 2021-10-05 | Fujifilm Corporation | Image capturing apparatus, control method for image capturing apparatus, and control program for image capturing apparatus |
US11475254B1 (en) | 2017-09-08 | 2022-10-18 | Snap Inc. | Multimodal entity identification |
US10740974B1 (en) | 2017-09-15 | 2020-08-11 | Snap Inc. | Augmented reality system |
US11335067B2 (en) | 2017-09-15 | 2022-05-17 | Snap Inc. | Augmented reality system |
US11721080B2 (en) | 2017-09-15 | 2023-08-08 | Snap Inc. | Augmented reality system |
US10499191B1 (en) | 2017-10-09 | 2019-12-03 | Snap Inc. | Context sensitive presentation of content |
US11617056B2 (en) | 2017-10-09 | 2023-03-28 | Snap Inc. | Context sensitive presentation of content |
US11006242B1 (en) | 2017-10-09 | 2021-05-11 | Snap Inc. | Context sensitive presentation of content |
US11030787B2 (en) | 2017-10-30 | 2021-06-08 | Snap Inc. | Mobile-based cartographic control of display content |
US11670025B2 (en) | 2017-10-30 | 2023-06-06 | Snap Inc. | Mobile-based cartographic control of display content |
US11265273B1 (en) | 2017-12-01 | 2022-03-01 | Snap, Inc. | Dynamic media overlay with smart widget |
US11558327B2 (en) | 2017-12-01 | 2023-01-17 | Snap Inc. | Dynamic media overlay with smart widget |
US11943185B2 (en) | 2017-12-01 | 2024-03-26 | Snap Inc. | Dynamic media overlay with smart widget |
US11687720B2 (en) | 2017-12-22 | 2023-06-27 | Snap Inc. | Named entity recognition visual context and caption data |
US11017173B1 (en) | 2017-12-22 | 2021-05-25 | Snap Inc. | Named entity recognition visual context and caption data |
US10678818B2 (en) | 2018-01-03 | 2020-06-09 | Snap Inc. | Tag distribution visualization system |
US11487794B2 (en) | 2018-01-03 | 2022-11-01 | Snap Inc. | Tag distribution visualization system |
US11507614B1 (en) | 2018-02-13 | 2022-11-22 | Snap Inc. | Icon based tagging |
US11841896B2 (en) | 2018-02-13 | 2023-12-12 | Snap Inc. | Icon based tagging |
US10885136B1 (en) | 2018-02-28 | 2021-01-05 | Snap Inc. | Audience filtering system |
US11523159B2 (en) | 2018-02-28 | 2022-12-06 | Snap Inc. | Generating media content items based on location information |
US10979752B1 (en) | 2018-02-28 | 2021-04-13 | Snap Inc. | Generating media content items based on location information |
US10327096B1 (en) | 2018-03-06 | 2019-06-18 | Snap Inc. | Geo-fence selection system |
US11722837B2 (en) | 2018-03-06 | 2023-08-08 | Snap Inc. | Geo-fence selection system |
US11044574B2 (en) | 2018-03-06 | 2021-06-22 | Snap Inc. | Geo-fence selection system |
US11570572B2 (en) | 2018-03-06 | 2023-01-31 | Snap Inc. | Geo-fence selection system |
US10524088B2 (en) | 2018-03-06 | 2019-12-31 | Snap Inc. | Geo-fence selection system |
US11491393B2 (en) | 2018-03-14 | 2022-11-08 | Snap Inc. | Generating collectible items based on location information |
US10933311B2 (en) | 2018-03-14 | 2021-03-02 | Snap Inc. | Generating collectible items based on location information |
US11163941B1 (en) | 2018-03-30 | 2021-11-02 | Snap Inc. | Annotating a collection of media content items |
US10448199B1 (en) | 2018-04-18 | 2019-10-15 | Snap Inc. | Visitation tracking system |
US10779114B2 (en) | 2018-04-18 | 2020-09-15 | Snap Inc. | Visitation tracking system |
US10924886B2 (en) | 2018-04-18 | 2021-02-16 | Snap Inc. | Visitation tracking system |
US11683657B2 (en) | 2018-04-18 | 2023-06-20 | Snap Inc. | Visitation tracking system |
US11297463B2 (en) | 2018-04-18 | 2022-04-05 | Snap Inc. | Visitation tracking system |
US10681491B1 (en) | 2018-04-18 | 2020-06-09 | Snap Inc. | Visitation tracking system |
US10219111B1 (en) | 2018-04-18 | 2019-02-26 | Snap Inc. | Visitation tracking system |
US11860888B2 (en) | 2018-05-22 | 2024-01-02 | Snap Inc. | Event detection system |
US10679393B2 (en) | 2018-07-24 | 2020-06-09 | Snap Inc. | Conditional modification of augmented reality object |
US11367234B2 (en) | 2018-07-24 | 2022-06-21 | Snap Inc. | Conditional modification of augmented reality object |
US10943381B2 (en) | 2018-07-24 | 2021-03-09 | Snap Inc. | Conditional modification of augmented reality object |
US10789749B2 (en) | 2018-07-24 | 2020-09-29 | Snap Inc. | Conditional modification of augmented reality object |
US11670026B2 (en) | 2018-07-24 | 2023-06-06 | Snap Inc. | Conditional modification of augmented reality object |
US11676319B2 (en) | 2018-08-31 | 2023-06-13 | Snap Inc. | Augmented reality anthropomorphtzation system |
US11450050B2 (en) | 2018-08-31 | 2022-09-20 | Snap Inc. | Augmented reality anthropomorphization system |
US10997760B2 (en) | 2018-08-31 | 2021-05-04 | Snap Inc. | Augmented reality anthropomorphization system |
US11455082B2 (en) | 2018-09-28 | 2022-09-27 | Snap Inc. | Collaborative achievement interface |
US11704005B2 (en) | 2018-09-28 | 2023-07-18 | Snap Inc. | Collaborative achievement interface |
US11799811B2 (en) | 2018-10-31 | 2023-10-24 | Snap Inc. | Messaging and gaming applications communication platform |
US11812335B2 (en) | 2018-11-30 | 2023-11-07 | Snap Inc. | Position service to determine relative position to map features |
US11558709B2 (en) | 2018-11-30 | 2023-01-17 | Snap Inc. | Position service to determine relative position to map features |
US11199957B1 (en) | 2018-11-30 | 2021-12-14 | Snap Inc. | Generating customized avatars based on location information |
US11698722B2 (en) | 2018-11-30 | 2023-07-11 | Snap Inc. | Generating customized avatars based on location information |
US11877211B2 (en) | 2019-01-14 | 2024-01-16 | Snap Inc. | Destination sharing in location sharing system |
US11751015B2 (en) | 2019-01-16 | 2023-09-05 | Snap Inc. | Location-based context information sharing in a messaging system |
US11693887B2 (en) | 2019-01-30 | 2023-07-04 | Snap Inc. | Adaptive spatial density based clustering |
US11294936B1 (en) | 2019-01-30 | 2022-04-05 | Snap Inc. | Adaptive spatial density based clustering |
US11809624B2 (en) | 2019-02-13 | 2023-11-07 | Snap Inc. | Sleep detection in a location sharing system |
US11500525B2 (en) | 2019-02-25 | 2022-11-15 | Snap Inc. | Custom media overlay system |
US11954314B2 (en) | 2019-02-25 | 2024-04-09 | Snap Inc. | Custom media overlay system |
US11574431B2 (en) | 2019-02-26 | 2023-02-07 | Snap Inc. | Avatar based on weather |
US11301117B2 (en) | 2019-03-08 | 2022-04-12 | Snap Inc. | Contextual information in chat |
US11868414B1 (en) | 2019-03-14 | 2024-01-09 | Snap Inc. | Graph-based prediction for contact suggestion in a location sharing system |
US11852554B1 (en) | 2019-03-21 | 2023-12-26 | Snap Inc. | Barometer calibration in a location sharing system |
US11249614B2 (en) | 2019-03-28 | 2022-02-15 | Snap Inc. | Generating personalized map interface with enhanced icons |
US11740760B2 (en) | 2019-03-28 | 2023-08-29 | Snap Inc. | Generating personalized map interface with enhanced icons |
US11361493B2 (en) | 2019-04-01 | 2022-06-14 | Snap Inc. | Semantic texture mapping system |
US11785549B2 (en) | 2019-05-30 | 2023-10-10 | Snap Inc. | Wearable device location systems |
US11606755B2 (en) | 2019-05-30 | 2023-03-14 | Snap Inc. | Wearable device location systems architecture |
US11206615B2 (en) | 2019-05-30 | 2021-12-21 | Snap Inc. | Wearable device location systems |
US11917495B2 (en) | 2019-06-07 | 2024-02-27 | Snap Inc. | Detection of a physical collision between two client devices in a location sharing system |
US11601783B2 (en) | 2019-06-07 | 2023-03-07 | Snap Inc. | Detection of a physical collision between two client devices in a location sharing system |
US11714535B2 (en) | 2019-07-11 | 2023-08-01 | Snap Inc. | Edge gesture interface with smart interactions |
US11821742B2 (en) | 2019-09-26 | 2023-11-21 | Snap Inc. | Travel based notifications |
US11218838B2 (en) | 2019-10-31 | 2022-01-04 | Snap Inc. | Focused map-based context information surfacing |
US11429618B2 (en) | 2019-12-30 | 2022-08-30 | Snap Inc. | Surfacing augmented reality objects |
US11128715B1 (en) | 2019-12-30 | 2021-09-21 | Snap Inc. | Physical friend proximity in chat |
US11729343B2 (en) | 2019-12-30 | 2023-08-15 | Snap Inc. | Including video feed in message thread |
US11893208B2 (en) | 2019-12-31 | 2024-02-06 | Snap Inc. | Combined map icon with action indicator |
US11943303B2 (en) | 2019-12-31 | 2024-03-26 | Snap Inc. | Augmented reality objects registry |
US11343323B2 (en) | 2019-12-31 | 2022-05-24 | Snap Inc. | Augmented reality objects registry |
US11888803B2 (en) | 2020-02-12 | 2024-01-30 | Snap Inc. | Multiple gateway message exchange |
US11228551B1 (en) | 2020-02-12 | 2022-01-18 | Snap Inc. | Multiple gateway message exchange |
US11516167B2 (en) | 2020-03-05 | 2022-11-29 | Snap Inc. | Storing data based on device location |
US11765117B2 (en) | 2020-03-05 | 2023-09-19 | Snap Inc. | Storing data based on device location |
US11619501B2 (en) | 2020-03-11 | 2023-04-04 | Snap Inc. | Avatar based on trip |
US11430091B2 (en) | 2020-03-27 | 2022-08-30 | Snap Inc. | Location mapping for large scale augmented-reality |
US11915400B2 (en) | 2020-03-27 | 2024-02-27 | Snap Inc. | Location mapping for large scale augmented-reality |
US11776256B2 (en) | 2020-03-27 | 2023-10-03 | Snap Inc. | Shared augmented reality system |
US11503432B2 (en) | 2020-06-15 | 2022-11-15 | Snap Inc. | Scalable real-time location sharing framework |
US11314776B2 (en) | 2020-06-15 | 2022-04-26 | Snap Inc. | Location sharing using friend list versions |
US11483267B2 (en) | 2020-06-15 | 2022-10-25 | Snap Inc. | Location sharing using different rate-limited links |
US11290851B2 (en) | 2020-06-15 | 2022-03-29 | Snap Inc. | Location sharing using offline and online objects |
US11676378B2 (en) | 2020-06-29 | 2023-06-13 | Snap Inc. | Providing travel-based augmented reality content with a captured image |
US11943192B2 (en) | 2020-08-31 | 2024-03-26 | Snap Inc. | Co-location connection service |
US11961116B2 (en) | 2020-10-26 | 2024-04-16 | Foursquare Labs, Inc. | Determining exposures to content presented by physical objects |
US11601888B2 (en) | 2021-03-29 | 2023-03-07 | Snap Inc. | Determining location using multi-source geolocation data |
US11606756B2 (en) | 2021-03-29 | 2023-03-14 | Snap Inc. | Scheduling requests for location data |
US11902902B2 (en) | 2021-03-29 | 2024-02-13 | Snap Inc. | Scheduling requests for location data |
US11645324B2 (en) | 2021-03-31 | 2023-05-09 | Snap Inc. | Location-based timeline media content system |
US11829834B2 (en) | 2021-10-29 | 2023-11-28 | Snap Inc. | Extended QR code |
US11962645B2 (en) | 2022-06-02 | 2024-04-16 | Snap Inc. | Guided personal identity based actions |
US11963105B2 (en) | 2023-02-10 | 2024-04-16 | Snap Inc. | Wearable device location systems architecture |
US11961196B2 (en) | 2023-03-17 | 2024-04-16 | Snap Inc. | Virtual vision system |
Also Published As
Publication number | Publication date |
---|---|
CN101815223A (en) | 2010-08-25 |
US8754958B2 (en) | 2014-06-17 |
CN105657393A (en) | 2016-06-08 |
CN105657393B (en) | 2018-06-22 |
KR101595254B1 (en) | 2016-02-18 |
KR20100095232A (en) | 2010-08-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8754958B2 (en) | Method of adjusting white balance of image, recording medium having program for performing the method, and apparatus applying the method | |
US9426359B2 (en) | Digital image signal processing method, medium for recording the method, and digital image signal processing apparatus | |
US8823829B2 (en) | Image capture with adjustment of imaging properties at transitions between regions | |
US8625021B2 (en) | Image capture with region-based adjustment of imaging properties | |
KR101643321B1 (en) | A digital photographing apparatus, a method for controlling the same, and a computer-readable medium | |
JP2014120844A (en) | Image processing apparatus and imaging apparatus | |
KR101626004B1 (en) | Method and apparatus for selective support of the RAW format in digital imaging processor | |
TW201515466A (en) | Control device, control method, and electronic device | |
US20100329552A1 (en) | Method and apparatus for guiding user with suitable composition, and digital photographing apparatus | |
KR102158844B1 (en) | Apparatus and method for processing image, and computer-readable recording medium | |
US8773546B2 (en) | Photographing control method and apparatus for capturing image data at a moment in time when booting | |
US8665355B2 (en) | Image capture with region-based adjustment of contrast | |
JPWO2019064757A1 (en) | Imaging equipment, imaging methods, and programs | |
US8872959B2 (en) | Digital photographing apparatus, method of controlling the same, and recording medium having recorded thereon program for executing the method | |
JP2015211233A (en) | Image processing apparatus and control method for image processing apparatus | |
US9413974B2 (en) | Information processing apparatus, image sensing apparatus, control method, and recording medium for conversion processing | |
KR20090076394A (en) | Digital photographing apparatus and method for controlling the same | |
JP5247900B2 (en) | Image sensor compensation | |
US8547447B2 (en) | Image sensor compensation | |
JP2012050121A (en) | Imaging device, imaging method and imaging program | |
KR102090273B1 (en) | Photographing apparatus and method | |
US11770511B2 (en) | Image processing apparatus, image capture apparatus, control method, and computer-readable storage medium | |
JP2006197478A (en) | Photographing device | |
JP2021078051A (en) | Image processing device and image control method, program, and storage medium | |
JP2009225252A (en) | Imaging apparatus and imaging method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SAMSUNG DIGITAL IMAGING CO., LTD., KOREA, REPUBLIC Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KIM, KYOUNG-JAE;KIM, YONG-GI;REEL/FRAME:024099/0383 Effective date: 20091229 |
|
AS | Assignment |
Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF Free format text: MERGER;ASSIGNOR:SAMSUNG DIGITAL IMAGING CO., LTD.;REEL/FRAME:026128/0759 Effective date: 20100402 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551) Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |