US20110049245A1 - Optical reader having reduced specular reflection read failures - Google Patents
Optical reader having reduced specular reflection read failures Download PDFInfo
- Publication number
- US20110049245A1 US20110049245A1 US12/853,090 US85309010A US2011049245A1 US 20110049245 A1 US20110049245 A1 US 20110049245A1 US 85309010 A US85309010 A US 85309010A US 2011049245 A1 US2011049245 A1 US 2011049245A1
- Authority
- US
- United States
- Prior art keywords
- pixels
- image sensor
- image data
- sensor array
- color
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000003287 optical effect Effects 0.000 title claims abstract description 308
- 238000003384 imaging method Methods 0.000 claims description 78
- 238000005286 illumination Methods 0.000 claims description 54
- 238000004458 analytical method Methods 0.000 claims description 23
- 239000007787 solid Substances 0.000 claims description 17
- 239000000284 extract Substances 0.000 claims description 4
- 230000000007 visual effect Effects 0.000 description 86
- 238000000034 method Methods 0.000 description 85
- 230000008569 process Effects 0.000 description 60
- 238000010586 diagram Methods 0.000 description 51
- 238000012545 processing Methods 0.000 description 50
- 230000000875 corresponding effect Effects 0.000 description 41
- 238000004891 communication Methods 0.000 description 27
- 238000012546 transfer Methods 0.000 description 26
- 238000005096 rolling process Methods 0.000 description 18
- 238000010276 construction Methods 0.000 description 16
- 238000003491 array Methods 0.000 description 12
- 238000009792 diffusion process Methods 0.000 description 12
- 230000004927 fusion Effects 0.000 description 12
- 230000001965 increasing effect Effects 0.000 description 12
- 230000000977 initiatory effect Effects 0.000 description 12
- 238000012935 Averaging Methods 0.000 description 11
- 230000004044 response Effects 0.000 description 11
- 238000000605 extraction Methods 0.000 description 10
- 239000000758 substrate Substances 0.000 description 10
- 238000013528 artificial neural network Methods 0.000 description 9
- 238000003860 storage Methods 0.000 description 9
- 238000007667 floating Methods 0.000 description 7
- 238000009826 distribution Methods 0.000 description 6
- 238000004519 manufacturing process Methods 0.000 description 6
- 239000002184 metal Substances 0.000 description 6
- 230000000881 depressing effect Effects 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 5
- 239000013598 vector Substances 0.000 description 5
- 230000001276 controlling effect Effects 0.000 description 4
- 230000009977 dual effect Effects 0.000 description 4
- 239000000975 dye Substances 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 238000010348 incorporation Methods 0.000 description 4
- 239000010410 layer Substances 0.000 description 4
- 239000011241 protective layer Substances 0.000 description 4
- 230000009467 reduction Effects 0.000 description 4
- 230000011218 segmentation Effects 0.000 description 4
- 230000003321 amplification Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 3
- 230000000295 complement effect Effects 0.000 description 3
- 230000002596 correlated effect Effects 0.000 description 3
- 239000011159 matrix material Substances 0.000 description 3
- 238000003199 nucleic acid amplification method Methods 0.000 description 3
- 238000012360 testing method Methods 0.000 description 3
- 230000009466 transformation Effects 0.000 description 3
- 230000004913 activation Effects 0.000 description 2
- 238000000429 assembly Methods 0.000 description 2
- 230000000712 assembly Effects 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 230000015556 catabolic process Effects 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 238000006731 degradation reaction Methods 0.000 description 2
- 230000004069 differentiation Effects 0.000 description 2
- 230000002708 enhancing effect Effects 0.000 description 2
- 239000011521 glass Substances 0.000 description 2
- 238000009434 installation Methods 0.000 description 2
- 230000035515 penetration Effects 0.000 description 2
- 238000002834 transmittance Methods 0.000 description 2
- 238000009827 uniform distribution Methods 0.000 description 2
- 238000011179 visual inspection Methods 0.000 description 2
- 241000533950 Leucojum Species 0.000 description 1
- 238000007792 addition Methods 0.000 description 1
- 230000002238 attenuated effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 229910000149 boron phosphate Inorganic materials 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000010224 classification analysis Methods 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000012790 confirmation Methods 0.000 description 1
- 238000005314 correlation function Methods 0.000 description 1
- 238000013480 data collection Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000000151 deposition Methods 0.000 description 1
- 230000000994 depressogenic effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005530 etching Methods 0.000 description 1
- 230000001815 facial effect Effects 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- -1 laminated Substances 0.000 description 1
- 239000002650 laminated plastic Substances 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000005055 memory storage Effects 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 239000003973 paint Substances 0.000 description 1
- 230000000149 penetrating effect Effects 0.000 description 1
- 229920001296 polysiloxane Polymers 0.000 description 1
- 230000002207 retinal effect Effects 0.000 description 1
- 238000010187 selection method Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
- 229910052710 silicon Inorganic materials 0.000 description 1
- 239000010703 silicon Substances 0.000 description 1
- 238000005549 size reduction Methods 0.000 description 1
- 230000000153 supplemental effect Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/10544—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
- G06K7/10712—Fixed beam scanning
- G06K7/10722—Photodetector array or CCD scanning
- G06K7/10752—Exposure time control
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/10544—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
- G06K7/10712—Fixed beam scanning
- G06K7/10722—Photodetector array or CCD scanning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/10544—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
- G06K7/10712—Fixed beam scanning
- G06K7/10722—Photodetector array or CCD scanning
- G06K7/10732—Light sources
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/10544—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
- G06K7/10792—Special measures in relation to the object to be scanned
- G06K7/10801—Multidistance reading
- G06K7/10811—Focalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/12—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using a selected wavelength, e.g. to sense red marks and ignore blue marks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/14—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/14—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
- G06K7/1404—Methods for optical code recognition
- G06K7/1408—Methods for optical code recognition the method being specifically adapted for the type of code
- G06K7/1417—2D bar codes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/14—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
- G06K7/1404—Methods for optical code recognition
- G06K7/1439—Methods for optical code recognition including a method step for retrieval of the optical code
- G06K7/1443—Methods for optical code recognition including a method step for retrieval of the optical code locating of the code in an image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/10—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
- H04N23/84—Camera processing pipelines; Components thereof for processing colour signals
- H04N23/843—Demosaicing, e.g. interpolating colour pixel values
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/10—Circuitry of solid-state image sensors [SSIS]; Control thereof for transforming different wavelengths into image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/10—Circuitry of solid-state image sensors [SSIS]; Control thereof for transforming different wavelengths into image signals
- H04N25/11—Arrangement of colour filter arrays [CFA]; Filter mosaics
- H04N25/13—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements
- H04N25/133—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements including elements passing panchromatic light, e.g. filters passing white light
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/10—Circuitry of solid-state image sensors [SSIS]; Control thereof for transforming different wavelengths into image signals
- H04N25/11—Arrangement of colour filter arrays [CFA]; Filter mosaics
- H04N25/13—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements
- H04N25/134—Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements based on three different wavelength filter elements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/40—Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled
- H04N25/42—Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled by switching between different modes of operation using different resolutions or aspect ratios, e.g. switching between interlaced and non-interlaced mode
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/50—Control of the SSIS exposure
- H04N25/53—Control of the integration time
- H04N25/532—Control of the integration time by controlling global shutters in CMOS SSIS
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/76—Addressed sensors, e.g. MOS or CMOS sensors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
Definitions
- the present invention relates to optical readers in general and specifically, to an optical reader having a reduced specular reflection read failures.
- Performance of an optical reader may be hindered where an optical reader is operated to read bar code symbols or other indicia of a substrate having a “shiny” surface.
- substrates can include, e.g., metal, glass, and laminated plastic.
- Light rays emanating from a reader that are projected on a highly reflective shiny surface of a substrate may be substantially entirely reflected directly onto a reader image sensor array.
- Artisans skilled in the art of optical readers regard a “specular reflection” read condition to have occurred where a substantial percentage of light rays are reflected from a substrate and directed onto a reader image sensor array.
- Light rays are said to be reflected at a “specular” angle when light rays are reflected from a substrate at about the angle of incidence. Specular reflection tends to saturate a reader image sensor array to cause decoding failures.
- FIG. 1 a is an electrical block diagram of a hand held optical reader of the invention including a hybrid monochrome and a color sensing solid state image sensor array;
- FIG. 1 b is a block diagram of an alternative image sensor array which may be incorporated into an optical reader according to the invention
- FIG. 1 c is a schematic block diagram illustrating an RF communications circuit according to the invention.
- FIG. 1 d is a schematic block diagram illustrating a display according to the invention.
- FIG. 1 e is a schematic view illustrating incorporation of a decode circuit, a signature autodiscrimination circuit, a demosaicing circuit, and a fusion circuit into an optical reader according to the invention
- FIGS. 2 a - 2 d are various partial exploded top views of an embodiment of a solid state image sensor array according to the invention.
- FIG. 3 a is a cutaway exploded side view of a monochrome pixel according to one embodiment of the invention.
- FIG. 3 b is a top view of the pixel shown in FIG. 3 a;
- FIG. 3 c is a cutaway exploded side view of a color sensitive pixel in one embodiment of the invention.
- FIG. 3 d is a top view of the pixel shown in FIG. 3 c;
- FIG. 4 a is an electrical block diagram of an embodiment of an image sensor according to the invention.
- FIG. 4 b is an electrical block diagram of an image sensor array of the invention showing incorporation of reset control lines in the image sensor array;
- FIG. 4 c is a timing diagram illustrating coordinated exposure control timing pulses and reset control timing pulses according to the invention.
- FIGS. 5 a - 5 e are various partial exploded top views of an embodiment of a solid state image sensor array according to the invention.
- FIG. 5 f is a top perspective view of an image sensor integrated circuit chip incorporating an image sensor array according to the invention with an exploded view portion illustrating a pixel pattern of color sensitive “clusters” of pixels which pattern may be distributed throughout the array;
- FIGS. 5 g - 5 i are top perspective views of image sensor integrated circuit chips incorporating a linear bar code symbol optimized image sensor array according to the invention with respective exploded view portions illustrating pixel patterns including “zones” of monochrome pixels and “zones” of color sensitive pixels;
- FIG. 5 j is a top perspective view of an image sensor integrated circuit chip incorporating a linear symbol optimized image sensor array according to the invention.
- FIG. 6 a is a cutaway exploded side view of a monochrome pixel according to one embodiment of the invention.
- FIG. 6 b is a top view of the pixel shown in FIG. 6 a;
- FIG. 6 c is a cutaway exploded side view of a color sensitive pixel in one embodiment of the invention.
- FIG. 6 d is a top view of the pixel shown in FIG. 6 c;
- FIG. 7 a is an electrical block diagram of an embodiment of an image sensor according to the invention.
- FIG. 7 b is an electrical block diagram of an image sensor array of the invention showing incorporation of reset control lines in the image sensor array;
- FIGS. 7 c and 7 d are schematic top views illustrating alternative configurations for a reset control system including separate sets of reset control lines for resetting a first subset of rows of pixels independent of resetting second subset of rows of pixels of an image sensor array according to the invention;
- FIG. 8 a is an exploded perspective view of an imaging module according to the invention.
- FIGS. 8 b and 8 c are front and side views, respectively, of the imaging module shown in FIG. 8 a;
- FIG. 8 d shows an illumination and aiming pattern which may be projected by an optical reader according to the invention
- FIG. 8 e is a top view of an alternative imaging module incorporating a laser based aiming pattern generating system
- FIG. 8 f is a front view of a polarizer plate which may be included as part of an imaging module herein, e.g., the imaging middle shown in FIG. 8 a;
- FIGS. 9 a and 9 b are physical form views of various hand held optical readers according to the invention.
- FIG. 9 c is a perspective view of a hand held mobile telephone (a “cell phone”) which may incorporate a hybrid monochrome and color image sensor array according to the invention and which may be configured according to the invention;
- FIG. 10 is a schematic view of a system incorporating a plurality of optical readers according to the invention.
- FIG. 11 is an application schematic view illustrating an optical reader according to the invention being operated to capture image data representing a parcel that carries a plurality of bar code symbols;
- FIG. 12 a is an application schematic view illustrating a first optical reader according to the invention and a second remotely located optical reader according to the invention being operated to take first and second digital pictures of a parcel at first and second locations that are a distance apart for purposes of determining whether the parcel was damaged during delivery from the first location to the second location;
- FIG. 12 b is another application schematic view illustrating an optical reader being used to take a color picture of a delivery vehicle
- FIG. 13 a is an application schematic diagram according to the invention illustrating an optical reader according to the invention being used to read bar codes of a vehicle and to take color pictures of a vehicle;
- FIG. 13 b is a view of a VIN sticker which may be disposed on the vehicle of FIG. 13 a;
- FIG. 13 c is a view of a VIN plate which may be disposed on the vehicle of FIG. 13 a ;
- FIG. 13 d is a view of a vehicle registration sticker which may be disposed on the vehicle of FIG. 13 a;
- FIG. 13 e is a view of an optical reader programmed to display a GUI form assisting an application wherein an optical reader, according to the invention, is utilized to decode bar code symbols and to take color pictures of a vehicle;
- FIGS. 14 a - 14 c are various flow diagrams illustrating the invention.
- FIGS. 14 d - 14 f are additional flow diagrams illustrating examples of operation of an optical reader according to the invention in an indicia decode mode of operation;
- FIGS. 14 g and 14 h are additional flow diagrams illustrating examples of operation of an optical reader according to the invention in a picture taking mode of operation;
- FIG. 14 i is a flow diagram illustrating operation of a fusion circuit of an optical reader according to the invention which processes monochrome and color image data to produce a high resolution visual display color frame of image data;
- FIGS. 15 a - 15 e are various image capture initiation control signal timing diagrams illustrating the invention.
- FIGS. 16 a - 16 c illustrate various pixelized frames of image data which may be captured by an optical reader according to the invention
- FIG. 17 a is an electrical block diagram of an optical reader according to the invention having a plurality of imaging modules
- FIGS. 17 b and 17 c illustrate alternative hardware blocks that can be utilized with the electrical circuit of FIG. 17 a;
- FIGS. 17 d and 17 e illustrate imaging modules which may be utilized with the reader of FIG. 17 a;
- FIGS. 17 f and 17 g illustrate exemplary optical readers incorporating a pair of imaging modules
- FIG. 18 a is a schematic view of a cyan-magenta-yellow (CMY) image sensor array in accordance with the invention which may be incorporated into an optical reader according to the invention and which may be controlled to generate both a decode frame of image data and a visual display color frame of image data;
- CY cyan-magenta-yellow
- FIG. 19 a is a schematic view of a hybrid monochrome and polarizer image sensor array in accordance with the invention which may be incorporated in an optical reader according to the invention;
- FIG. 19 b is a top perspective view of a hybrid monochrome and polarizer image sensor array according to the invention with an exploded view section illustrating a pattern of light polarizing pixels that may be distributed throughout the image sensor array;
- FIG. 19 c is a flow diagram illustrating an exemplary operational mode of an optical reader according to the invention which incorporates a hybrid monochrome and polarizer image sensor array according to the invention;
- FIGS. 20 a and 20 b are top perspective views of a monochrome polarizer and color sensitive image sensor array according to the invention with an exploded view section illustrating a pattern of light polarizing pixels and color sensitive pixels that may be distributed throughout the array;
- FIG. 21 is a schematic view of an image sensor integrated circuit chip incorporating an image sensor array having color sensitive pixels disposed therein with two different periods of distribution;
- FIG. 22 a is a schematic block diagram of an autodiscrimination circuit which may be utilized with the invention.
- FIG. 22 b is a process for practicing principles of the invention including automatically discriminating between different dataform types
- FIG. 22 c shows one embodiment of a plurality of curvelent detector maps which may be utilized with the invention.
- FIG. 22 d shows another embodiment of a plurality of curvelent detector maps which may be utilized with the invention.
- FIG. 22 e is a diagrammatic representation of a histogram analysis which may be performed in one embodiment of the invention.
- FIGS. 22 f - 22 i are diagrammatic representations of an image data segmentation process according to embodiments of the invention.
- an optical reader having an image sensor array.
- the image sensor array can include a first subset of pixels and a second subset of pixels.
- the first subset of pixels can be devoid of light polarizing filter elements, and the second subset of pixels can be light polarizing pixels including light polarizing filter elements.
- An optical reader can be configured to selectively read out image data from an image sensor array's light polarizing pixels.
- An optical reader image sensor array of the invention can include light polarizing pixels, each light polarizing pixel having a light polarizing filter element (light polarizing filter) that significantly attenuates polarized light rays generated from an appropriately polarized light source and reflected at a specular angle; thus, reducing the contribution of specularly reflected light rays to generated image signals from the polarizing pixels.
- a first subset of pixels of an optical reader image sensor array are monochrome pixels and a second subset of pixels are light polarizing pixels.
- image data corresponding to the light polarizing pixels can be selectively transferred to a decode circuit, either by way of selecting reading out image data from the light polarizing pixels, or by selectively extracting image data corresponding to light polarizing pixels from a frame of image data including image data in addition to image data corresponding to light polarizing pixels.
- a picture taking optical reader having a hybrid monochrome and color (monocolor) solid state image sensor array.
- the hybrid image sensor array comprises a plurality of pixels including a first subset of pixels and a second subset of pixels, wherein the first subset of pixels are monochrome pixels and the second subset of pixels are color sensitive pixels having wavelength selective color filter elements.
- the monochrome first subset of pixels is formed in a checkerboard pattern, and voids are formed at the corners of pixels of the first subset, such that combinations of voids of adjacent pixels define open areas. Pixels of the color sensitive second subset of pixels are formed at the open areas, and wavelength selective filter elements are formed on pixels of the second subset but not on pixels of the first subset.
- an optical reader solid state image sensor array includes a plurality of pixels formed in a plurality of rows on an IC chip in a checkerboard pattern wherein each pixel has approximately the same dimension.
- the majority of pixels of the image sensor array are monochrome pixels of the first subset.
- Color sensitive pixels of the second subset are at spaced apart positions and are uniformly or substantially uniformly distributed throughout the image sensor array.
- a hybrid monochrome and color sensing solid state image sensor array of the invention may be incorporated in an imaging module which, in addition to having an image sensor array constructed in accordance with the invention includes such elements as an imaging lens, an illumination assembly including a field illumination assembly, an aiming illumination assembly and a support member for supporting the above elements.
- An imaging module in turn, may be incorporated into a hand held housing which encapsulates and supports the imaging assembly.
- the image sensor array in one embodiment can be made to have selectively addressable pixels.
- the image sensor array is constructed to have selectively addressable pixels, pixels of the first subset of pixels can be selectively addressed independent of the second subset of pixels so that image data corresponding to the first subset of pixels is selectively read out independent of the second subset of pixels.
- Image sensor arrays having selective read out capability can be provided utilizing alternative fabrication technologies.
- an optical reader includes separate and independently controllable reset control lines for resetting monochrome pixels and color sensitive pixels of the image sensor array.
- monochrome pixels may be driven into reset.
- color sensitive pixels may be driven into reset.
- Driving pixels not being selectively addressed for image data read out into a reset state reduces cross-talk between pixels of the image sensor array.
- an optical reader By incorporating within a single low cost image sensor array a combination of monochrome pixels and color sensitive pixels, an optical reader according to the invention provides indicia decoding performance approximately equal to the performance of an optical reader having an all monochrome image sensor array, and picture taking performance (i.e., the ability to obtain visual display quality color frames of image data) approximately equal to or superior to that of a digital camera incorporating an all color pixel image sensor array, wherein each pixel of the array includes a wavelength selective filter element.
- Reader 100 includes a solid state image sensor array 182 A, incorporated on an image sensor integrated circuit chip 1082 A shown in FIG. 1 a as a CMOS image sensor integrated circuit (IC) chip.
- image sensor array 182 A includes a plurality of pixels and wavelength sensitive color filter elements associated with a color sensitive subset of the pixels, wherein the remaining pixels external to the color sensitive subset of pixels are devoid of associated wavelength selective filter elements. Because image sensor array 182 A includes both monochrome pixels and color sensitive pixels, image sensor array 182 A may be termed a hybrid monochrome and color image sensor array.
- Reader 100 further includes a processor IC chip 548 and a control circuit 552 .
- Control circuit 552 in the embodiment of FIG. 1 a is shown as being provided by a central processing unit (CPU) of processor IC chip 548 .
- control circuit 552 may be provided by e.g., a programmable logic function execution device such as a field programmable gate array (FPGA) or an application specific integrated circuit (ASIC).
- Imaging lens 212 focuses images onto an active surface of image sensor array 182 A and together with image sensor array 182 A forms an imaging assembly 200 .
- Control circuit 552 executes picture taking and indicia decoding algorithms in accordance with instructions stored in program memory EPROM 562 which together with RAM 560 and Flash memory 564 forms a reader memory 566 .
- Reader memory 566 is in communication with processor IC chip 548 via system bus 570 .
- Main processor IC chip 548 may be a multifunctional IC chip such as an XSCALE PXA25x processor IC chip including central processing unit (CPU) 552 .
- Reader 100 further includes a field programmable gate array (FPGA) 580 .
- FPGA field programmable gate array
- FPGA 580 receives digital image data from image sensor IC chip 1082 A and transfers that image data into RAM 560 so that the image data can be further processed (e.g., by the decoding of a bar code symbol).
- Processor IC chip 548 can include an integrated frame grabber.
- processor IC chip 548 can be an XSCALE PXA27X processor IC chip with “Quick Capture Camera Interface” available from INTEL. Where processor IC chip 548 includes an integrated frame grabber, the integrated frame grabber may provide the frame acquisition functionality of FPGA 580 .
- Reader 100 further includes an illumination assembly 104 and a manual trigger 216 .
- FIG. 1 a includes an on-chip control/timing circuit 1092 , an on-chip gain circuit 1084 , an on-chip analog-to-digital converter 1086 and an on-chip line driver 1090 .
- An image sensor array which is incorporated into optical reader 100 may take on a variety of forms.
- reader 100 includes first image sensor array 182 A. However, as indicated by hardware block 208 , the image sensor array 182 A may be replaced.
- reader 100 incorporates image sensor array 182 B.
- optical reader 100 incorporates more than one image sensor array.
- image sensor arrays which may be incorporated into reader 100 are described herein.
- Radio frequency communication interface 571 may include one or more radio transceivers. Referring to the schematic diagram of FIG. 1 c , radio frequency communication interface 571 may include one or more of an 802.11 radio transceiver 5712 , a Bluetooth radio transceiver 5714 , a GSM/GPS radio transceiver 5716 or a WIMAX (802.16) radio transceiver 5718 . Radio frequency communication interface 571 facilitates wireless communication of data between device 100 and a spaced apart device 150 .
- RF radio frequency
- I/O communication interface 572 includes one or more serial or parallel hard-wired communication interfaces facilitating communication with a spaced apart device 150 as will be described further in connection with FIG. 10 .
- I/O communication interface 572 may include one or more of an Ethernet communication interface, a universal serial bus (USB) interface, or an RS-232 communication interface.
- Optical reader 100 may further include a keyboard 508 for entering data, a pointer mover 512 for moving a pointer of a graphical user interface (GUI) and a trigger 216 for initiating bar code reading and/or picture taking.
- Optical reader 100 may also include a display 504 , such as a monochrome or color LED display and a touch screen 504 T overlaid over display 504 .
- display 504 may include a display screen 5042 coupled to display controller 5044 for displaying color image data.
- Display controller 5044 receives a visual display color frame of image data from control circuit 552 , and reformats that data for display depending on the particular requirements of display screen 5042 , including the pixel resolution of display screen 5042 .
- All of the components of FIG. 1 a can be encapsulated and supported by a hand held housing 101 , e.g., as shown in FIGS. 9 a - 9 c . Additional features and functions of the components of reader 100 shown in FIG. 1 a are described herein.
- optical reader 100 may be regarded as having various processing circuits (modules).
- Indicia decode circuit 1702 receives image data and decodes decodable indicia therein such as bar code indicia and OCR character data.
- Optical reader 100 can be configured so that indicia decode module 1702 decodes such bar code symbols UPC/EAN, Code 11, Code 39, Code 128, Codabar, Interleaved 2 of 5, MSI, PDF417, MicroPDF417, Code 16K, Code 49, MaxiCode, Aztec, Aztec Mesa, Data Matrix, Qcode, QR Code, UCC Composite, Snowflake, Vericode, Dataglyphs, RSS, BC 412, Code 93, Codablock, Postnet (US), BPO4 State, Canadian 4 State, Japanese Post, KIX (Dutch Post), Planet Code and the like, and such OCR character forms as OCR A, OCR B, and the like.
- indicia decode module 1702 decodes such bar code symbols UPC/EAN, Code 11, Code 39, Code 128, Codabar, Interleaved 2 of 5, MSI, PDF417, MicroPDF417, Code 16K, Code 49, MaxiCode, Aztec, Aztec Mesa, Data Matrix, Qcode,
- Autodiscrimination circuit 1704 processes received image data and distinguishes between handwritten character data and decodable indicia.
- Autodiscrimination circuit 1704 may include indicia decode circuit 1702 .
- Autodiscrimination circuit 1704 and indicia decode circuit 1702 may be physically embodied by a combination of control circuit 552 and memory 566 .
- control circuit 552 operating under the control of a program stored in memory 562 may process image data stored in memory 560 to decode decodable indicia therein or to discriminate between handwritten character data and decodable indicia. Further aspects of indicia decode circuit 1702 and autodiscrimination circuit 1704 are described in copending U.S. patent application Ser. No.
- optical reader 100 may further include a demosaicing circuit 1706 , and a fusion circuit 1708 .
- Demosaicing circuit 1706 receives as an input a color filter array image data frame (e.g., a Bayer pattern image) and produces as an output a visual display color frame of image data.
- Fusion circuit 1708 receives as inputs both monochrome and color image data and produces as an output a visual display color frame of image data having a spatial resolution at or on the order of the pixel resolution of the optical reader's hybrid monochrome and color image sensor array.
- circuits 1706 and 1708 may be physically embodied by the combination of control circuit 552 and memory 566 .
- Control circuit 552 as well as circuits 1702 , 1704 , 1706 , and 1708 may be incorporated within hand held housing 101 (e.g., as shown in FIGS. 9 a - 9 c ) or else one or more of circuits 552 , 1702 , 1704 , 1706 , and 1708 can be incorporated in a housing of a spaced apart device 150 as described in connection with FIG. 10 .
- a visual display color frame of image data as referred to herein is an image frame including a set of color indicating data at each of a plurality of pixel positions, wherein each set of color indicating data represents a color at a discrete position of a target 1850 (shown in FIG. 8 d ).
- Each set of color indicating data includes three color values, e.g., a color scale value representing red, a color scale value representing blue, and a color scale value representing green.
- the set of color indicating data for each pixel position may include a cyan value, a magenta value and a value representing yellow.
- the set of color indicating data for each pixel position of a visual display color frame of image data output by demosaicing circuit 1706 or fusion circuit 1708 are RGB data sets including 24 bits of information, wherein the first 8 bits represent a red color scale value (red value) for the pixel position, the second 8 bits represent a green color scale value (green value) for the pixel position and the third 8 bits represent a blue color scale value (blue value) for the pixel position.
- a major feature of the invention is the construction of the optical reader's image sensor array various embodiments of which are shown and described throughout several views including the views of FIGS. 2 a - 7 d.
- FIG. 1 a A first embodiment of a hybrid monochrome and color sensitive (monocolor) solid state image sensor array is shown and described in FIG. 1 a and FIGS. 2 a - 4 b.
- solid state image sensor array 182 A includes a monochrome first subset of pixels 250 M and a color sensitive second subset of pixels 250 C.
- the first subset of monochrome pixels 250 M is formed in a checkerboard pattern and voids 253 as shown in FIG. 2 a are formed at the corners of pixels of the first subset, such that combinations of voids, e.g., voids 253 - 1 , 253 - 2 , 253 - 3 , 253 - 4 of adjacent pixels define open areas, e.g., open area 255 , each open area bounded by four pixels of the first subset.
- pixels 250 C forming a second subset of pixels 250 C are disposed in the open areas 255 , and wavelength selective filter elements, e.g., filter element 260 C, 260 M, as shown in FIG. 2 b are formed on pixels of the second subset but not on pixels of the first subset.
- Monochrome pixels 250 M as described herein are devoid of color filter elements (color filters).
- Pixels of the first monochrome pixel subset are in the shape of twelve sided polygons. The pixels are cross-shaped as seen from the top view that is indicated by FIGS. 2 a - 2 d (the monochrome pixels are square shaped as modified by the presence of voids 253 ). Pixels of the color sensitive second subset are square as seen from a top view.
- color sensitive pixels 250 C of image sensor array 182 A include either a cyan (Cy) filter element 260 C or magenta (Mg) filter element 260 M.
- color sensitive pixels 250 C of image sensor array 182 A include either a red filter element 260 R, a green filter element 260 G or a blue color filter element 260 B (RGB filters).
- the color sensitive pixels 250 C can be distributed throughout image sensor array 182 according to a Bayer pattern wherein there are N blue pixels, N red pixels and 2N green pixels.
- Color filter elements of any image sensor array pixel as described herein can be deposited on the major body of color sensitive pixels 250 C by way of a depository process.
- visual display color image data can be obtained utilizing either the version of image sensor array 182 A shown in FIG. 2 b or the version of image sensor array 182 A shown in FIG. 2 c , or another version of image sensor array 182 A such as a version including cyan, magenta and yellow (CMY) color sensitive pixels.
- CMY magenta and yellow
- image sensor array 182 A may include microlenses 320 for directing of light rays incident on image sensor array 182 A. Further aspects of microlenses 320 , including monochrome pixels, microlenses 320 M, and color sensitive pixel microlenses 320 C are described herein.
- FIGS. 3 a - 3 d Exploded physical form views of an image sensor pixel array 182 A, where array 182 A is configured to operate in a global electronic shutter operating mode are shown and described in FIGS. 3 a - 3 d .
- a monochrome pixel 250 M of image sensor array 182 A is shown in FIGS. 3 a and 3 b .
- Monochrome pixel 250 M includes a photodetector 302 which may be of photodiode or photogate construction, a transfer gate 304 , a floating diffusion 306 , a reset transistor 307 including reset gate 308 , a row select transistor 309 including row select gate 310 and a source follower amplifier transistor 311 including amplifier gate 312 .
- An important feature of pixel 250 M is opaque optical shield 316 .
- Opaque optical shield 316 typically comprising metal, shields light rays from components of pixel 250 M other than photodetector 302 . Accordingly, pixels from each of several rows of image sensor array 182 A can be simultaneously exposed to light in a global electronic shutter operating mode without the light rays modulating charges stored in floating diffusion 306 or another storage region. Further aspects of image sensor arrays capable of operating in a global electronic shutter operating mode are described in U.S. patent application Ser. No. 11/077,975 incorporated herein by reference. Referring to additional aspects of pixel 250 M, pixel 250 M includes microlens 320 which may be disposed on light transmissive protective layer 322 . Microlens 320 collects light from a larger surface area than photodetector 302 and directs light toward photodetector 302 .
- Color sensitive pixel 250 C of image sensor array 182 A is described with reference to FIGS. 3 c and 3 d .
- Color sensitive pixel 250 C is similar in construction to monochrome pixel 250 M.
- Color sensitive pixel 250 C includes a photodetector 302 which may be of photodiode or photogate construction, a transfer gate 304 for transferring charge from photodetector 250 C, a floating diffusion 306 , a reset transistor 307 including reset gate 308 , a row select transistor 309 including row select gate 310 and a source follower transistor amplifier 311 including amplifier gate 312 .
- Color sensitive pixel 250 C also includes opaque shield 320 which shields light from light sensitive components of pixel 250 C other than photodetector 302 .
- Pixel 250 C may also include microlens 320 for increasing the amount of light incident on photodetector 302 .
- color sensitive pixel 250 C includes a wavelength selective color filter element 260 formed thereon.
- Wavelength selective color filter element 260 may be disposed intermediate microlens 320 and protective layer 322 .
- each color sensitive pixel 250 C has four adjacent monochrome pixels 250 M.
- Microlenses 320 as shown in FIGS. 3 a and 3 c are also shown in the view of FIG. 2 d .
- Monochrome pixel microlens 320 , 320 M and color sensitive microlens 320 , 320 C may be formed on a microlens array included a plurality of microlenses.
- microlenses 320 C of color sensitive pixels 250 , 250 C have very little (e.g., less than 3.4%) of overlap relative to microlenses 320 M.
- Color sensitive pixel 250 C of image sensor array 182 A as best seen by a comparison between FIGS. 3 b and 3 d and consumes a smaller surface area than pixel 250 M.
- pixel 250 M includes an area, as seen from a top view, of about 12 nm by 12 nm while pixel 250 C includes an area, as seen from a top view, of about 6 ⁇ m by 6 ⁇ m.
- pixel 250 M includes a top surface area of about 6 ⁇ m by 6 ⁇ m, while pixel 250 C includes a top surface area of about 3 ⁇ m or 3 ⁇ m. Size reductions of pixel 250 M or pixel 250 , 250 C may be made at low cost by reducing the number of transistors of pixel 250 M and/or pixel 250 C.
- a transistor count of a pixel 250 C of image sensor array 182 A may readily be reduced by eliminating optically shielded floating diffusion 306 in which charges are stored on a temporary basis to facilitate global electronic shutter operation. Accordingly, in one embodiment, monochrome pixels 250 M of image sensor array 182 A have more transistors than color sensitive pixels 250 C but are capable of being exposed on a global electronic shutter basis, whereas color sensitive pixels 250 C have fewer transistors than monochrome pixels 250 M but are not capable of being exposed on a global electronic shutter basis.
- the relatively larger monochrome pixels 250 M have a transistor count sufficient to facilitate global shutter operation, but the relatively smaller color sensitive pixels 250 C are passive pixels requiring off-pixel amplification, and comprise a single transistor each. Further aspects of global electronic shutter and rolling shutter operations relative to image sensor arrays which may be incorporated into reader 100 are described herein.
- image sensor array 182 A is an active pixel image sensor array of complementary metal oxide semiconductor (CMOS) construction such that each pixel 250 M, 250 C, whether from the monochrome first subset of pixels or the color sensitive second subset of pixels is an active pixel including a pixel amplifier 311 for amplifying signals corresponding to light incident on photosensitive region 252 .
- Each pixel 250 M, 250 C may also include an optically shielded storage element 306 .
- Image sensor array 182 A further includes two-dimensional grid of interconnects 262 which are in electrical communication with respective column circuitry 270 and row circuitry 296 . Row circuitry 296 and column circuitry 270 enable such processing and operational tasks as selectively addressing pixels, decoding pixels, amplification of signals, analog-to-digital conversion, applying timing, read out and reset signals and the like.
- image sensor array 182 A includes separate reset control lines for monochrome pixels 250 M and color pixels 250 C.
- image sensor array 182 A may be constructed so that image sensor array 182 A has a first set of reset control lines 262 R-M for resetting monochrome pixels 250 M and a second set of reset control lines 262 R-C for resetting color pixels 250 C.
- optical reader 100 selectively reads out a windowed frame of image data comprising image data from monochrome pixels 250 M. In other operating modes, optical reader 100 selectively reads out a windowed frame of image data comprising image data from color pixels 250 C.
- a reset control timing pulse can be applied to image sensor array 182 A during the time that a windowed frame of image data is being captured to reset pixels of image sensor array 182 A that are not being selectively addressed for image data read out.
- an exposure control timing pulse 354 can be coordinated with a reset control timing pulse 370 .
- exposure control timing pulse 354 may control exposure of monochrome pixels 250 M of image sensor array 182 A (or alternatively, color pixels 250 C) of image sensor array 182 A, while reset control timing pulse 370 drives pixels not being selectively addressed into a reset state.
- reset control timing pulse 370 drives pixels not being selectively addressed into a reset state.
- pixels are reset, charges built up on pixels tend to be drained out of the pixels. Further, it is believed that photons entering pixels driven into reset may be refracted so that fewer photons become incident on neighboring pixels being exposed for image data read out. Accordingly, coordinating the timing of an exposure control pulse 354 for exposing selectively addressed pixels and a reset control timing pulse 370 for resetting pixels not being selectively addressed reduces cross talk between pixels.
- image sensor array 182 A may be constructed so that the presence of multiple reset control lines 162 R-C, 162 R-M do not substantially decrease the fill factor of pixels of image sensor array 182 A.
- FIG. 4 b shows a schematic top view of multiple reset control lines 162 R-M, 162 R-C incorporated in image sensor array 182 , 182 A.
- control lines 162 R-M, 162 R-C can be incorporated in image sensor array 182 A in a layered manner so that for a substantial portion of image sensor array 182 A, control lines 164 R-M have x, y positions that coincide with x, y positions of control line 164 R-C (axes are defined in FIG.
- Control lines 164 R-C in the embodiments of FIG. 4 b are installed at a different height (a different Z axis position) within image sensor array 182 A such that control lines 162 R-M and 162 R-C, for substantial length of the control lines, have common x, y positions. Installing the multiple control lines to be on top of one another so that the control lines have a common x, y axis position within image sensor array 182 A reduces the amount of fill factor degradation which would otherwise result from installation of an additional set of reset control lines within image sensor array 182 A.
- image sensor array 182 B includes a plurality of square shaped pixels (as seen from a top view) in a checkerboard pattern, each of the pixels having substantially the same dimensions.
- Each pixel 250 M, 250 C of image sensor array 182 B may be constructed to have approximately the same top surface dimensions as seen from the top views of FIGS. 5 a - 5 i and approximately the same side view cross-sectional dimensions as seen from the cross-sectional views of FIGS. 6 a - 6 d .
- Image sensor array 182 B is similar to the construction of a standard off-the-shelf monochrome image sensor array except that select ones of the pixels of the image sensor array have an associated wavelength selective color filter element.
- Solid state image sensor array 182 B includes a plurality of pixels formed in a plurality of rows.
- a monochrome first subset of pixels 250 M comprise the majority of pixels of the array.
- Wavelength selective color filter elements 260 are included in the second subset of color sensitive pixels 250 C.
- the color sensitive second subset of pixels 250 C comprises pixels at spaced apart pixel positions uniformly distributed or substantially uniformly distributed throughout the plurality of pixels forming the image sensor array 182 B.
- image sensor array 182 B can be provided by including an appropriately designed color filter array on an image sensor array of an MT9M111 Digital Clarity SOC 1.3 megapixel CMOS image sensor IC chip of the type available from Micron, Inc., an MT9V022 image sensor IC chip also available from Micron, Inc. or a VV6600 1.3 megapixel CMOS image sensor IC chip of the type available from STMicroelectronics.
- image sensor IC chips which can be utilized to provide image sensor array 182 B include MT9M413 image sensor IC chip available from Micron, Inc., a KAC-0311 image sensor IC chip manufactured by Kodak, Inc. and a KAI-0340 image sensor IC chip also manufactured by Kodak, Inc. Operational aspects of the referenced KAI-0340 image sensor IC chip are described further herein.
- MT9M413 image sensor IC chip available from Micron, Inc.
- KAC-0311 image sensor IC chip manufactured by Kodak, Inc. a KAI-0340 image sensor IC chip also manufactured by Kodak, Inc. Operational aspects of the referenced KAI-0340 image sensor IC chip are described further herein.
- Various manufacturer product description materials respecting certain of the above image sensor IC chips are appended to Provisional Patent Application No. [not yet assigned] filed Jun. 22, 2005 (Express Mail Label No. EV554216715US) and Provisional Patent Application No. [not yet assigned]
- CMOS image sensor IC chips manufactured by Micron, Inc., and KAC-0311 image sensor IC chip by Kodak, Inc. are CMOS image sensor IC chips that may be operated in a global electronic shutter mode such that all rows of pixels subject to image data read out have common exposure periods; that is, all rows of pixels subject to image data read out for reading out a frame of image data (i.e., full frame or “windowed frame”) have a common exposure start time and a common exposure stop time.
- an exposure control timing pulse as will be described herein is applied to the image sensor array.
- Exposure of each row of pixels subject to image data read out begins at the leading edge of the exposure control timing pulse and ends at the falling edge of the exposure control timing pulse.
- Micron, Inc. uses the trademark TRUESNAP with reference to a global electronic shutter operating mode.
- wavelength selective color filter elements (filters) formed on color sensitive pixels 250 , 250 C may be a combination of cyan filter elements 260 C and magenta color filter elements 260 M.
- wavelength sensitive filters of color sensitive pixels 250 C may also be a combination of red filter elements 260 R, green filter elements 260 G and blue filter elements 260 B. Because cyan and magenta filters require only one dye and not two dyes (as in red green and blue filters), the version of FIG. 5 b allows more light to pass through to a photodetector (e.g., photodetector 302 as shown in FIG. 6 c ) and exhibits a higher signal to noise ratio than the embodiment of FIG. 5 b . Nevertheless, the version of FIG. 5 a may be preferred for certain applications.
- hybrid monochrome and color image sensor 182 B can be made by including an appropriately designed color filter array on a commonly available, off-the-shelf image sensor array in a standardly known checkerboard pattern, each pixel of the array having substantially the same dimensions.
- a larger portion of image sensor array 182 B is shown in FIG. 5 c , where pixels designated by the letter “c” are color sensitive pixels 250 C and pixels not designated by the letter “c” are monochrome pixels 250 M.
- each color sensitive pixel 250 C has eight adjacent monochrome pixels 250 M (two side adjacent, one top adjacent, one bottom adjacent and four corner adjacent).
- image sensor array 182 B including a subset of monochrome pixels 250 M and a subset of color sensitive pixels 250 C, wherein each pixel of the image sensor array has substantially equal dimensions are shown and described in connection with FIGS. 5 f - 5 j.
- image sensor array 182 B includes the first subset of monochrome pixels 250 M and a second subset of color sensitive pixels 250 C.
- the color sensitive pixels 250 C of image sensor array 182 B in the version of FIG. 5 f are formed in clusters such as cluster 257 R, cluster 257 G and cluster 257 B.
- Each cluster 257 in the version of FIG. 5 f includes a plurality of pixels in successive horizontally adjacent pixel positions, such that each pixel of the cluster is horizontally adjacent to at least one other color sensitive pixel.
- Color sensitive clusters of pixels are distributed uniformly or substantially uniformly throughout image sensor array 182 B.
- Clusters may be formed in accordance with the standardized color filter pattern such as an RGB Bayer pattern or a cyan-magenta-yellow (CMY) pattern.
- Each cluster may have a plurality of pixels with each pixel of every individual cluster having a filter element of the same wavelength rating.
- clusters are distributed throughout image sensor array 182 B in a pattern that is accordance with the pattern of Bayer color filter array.
- Cluster 257 G includes three horizontally adjacent green pixels.
- Cluster 257 R includes three horizontally adjacent red pixels.
- Cluster 257 B includes three horizontally adjacent blue pixels.
- the version of image sensor array 182 B including a distribution of color sensitive pixels in horizontally arranged clusters as shown in FIG. 5 f is particularly useful where it is desired to include in image sensor array 182 B separate and independently controllable reset control lines 262 R-M and 262 R-C for separately and independently resetting monochrome pixels of image sensor array 182 B and color sensitive pixels of image sensory array 182 B without increasing the thickness of image sensor array 182 B.
- image sensor array 182 B having a subset of monochrome pixels in a subset of color sensitive pixels may be configured to include “zones” of monochrome pixels and “zones” of color sensitive pixels.
- a “zone” of pixels herein is a collection of positionally related pixels at a specified area of an image sensor array each having a color filter element or alternatively, each being without a color element. Examples of zones described herein comprise all pixels of one row of pixels or all pixels of each of several consecutive rows of pixels.
- image sensor array 182 B includes two color sensitive zones of pixels 2500 C and a single monochrome zone of pixels 2500 M.
- Each zone of pixels comprises a plurality of horizontally, vertically or diagonally adjacent pixels.
- the plurality of pixels of a monochrome zone of pixels, e.g., zone 2500 M are all devoid of a color sensitive filter element.
- the plurality of adjacent pixels in a color sensitive zone of pixels, e.g., zone 2500 C, all include a color sensitive filter element.
- monochrome zone of pixels 2500 M is interposed between a pair of color sensitive zones of pixels 2500 C.
- Monochrome zone of pixels 2500 M in the version of FIG. 5 g comprises a single row of pixels of image sensor array 182 B at or approximately the center of image sensor array 182 B.
- the first color sensitive zone of pixels of an image sensor array 182 B includes all pixels from the row of pixels of zone 2500 M up to the top row of image sensor array 182 B.
- the second color sensitive zone of pixels 2500 C in the version of FIG. 5 g includes all pixels from all rows from the center row monochrome zone of pixels 2500 M down to the bottom row of pixels of image sensor array 182 B.
- the color filter elements of color sensitive pixels 250 C of image sensor array 182 B may be formed in a standard color filter pattern, e.g., an RGM Bayer color filter pattern or a CMY pattern.
- FIG. 5 h another version of image sensor array 182 B is shown and described.
- the version of FIG. 5 h is similar to the version of FIG. 5 g except that the monochrome zone of pixels 2500 M is expanded to include ten consecutive rows of pixels at the center or approximately the center of image sensor array 182 B.
- a single color sensitive zone of pixels 2500 C is interposed between two relatively small width monochrome zones of pixels 2500 M formed at the top and bottom of image sensor array 182 B respectively.
- the first monochrome zone of pixels 2500 M comprises the first ten rows of pixels of image sensor array 182 B and a second monochrome zone of pixels 2500 M includes pixels of the bottom ten rows of image sensor array 182 B.
- Color sensitive zone of pixels 2500 C in the version of FIG. 5 i includes all pixels of the array excluding the first ten and the last ten rows of pixels of image sensor array 182 B.
- the pixels of the color sensitive zones 2500 C shown may include color filter elements in accordance with the pattern of a standardized color filter array, e.g., an RGB Bayer pattern or a CMY pattern.
- image sensor array 182 B shown in FIG. 5 j is similar in construction to the version of FIG. 5 g except that the version of FIG. 5 j includes additional monochrome zones of pixels 2500 M.
- image sensor array 182 B includes a pair of diagonal zones of monochrome pixels 2500 M-D extending through a center (actual or approximate) of image sensor array 182 B and a vertically extending zone of monochrome pixels 2500 M-V extending through a center of image sensor array 182 B.
- the linear zones of monochrome pixels 2500 M shown in the version of FIG. 5 j may include a minor dimension equal to one pixel width or more than one pixel width.
- the vertically extending monochrome zone of pixels 2500 M of FIG. 5 j may include pixel positions of one column of pixels or of a plurality of columns of pixels.
- the diagonally extending linear monochrome zones of pixels 2500 M of FIG. 5 g may include pixel positions of a single diagonal row of pixels or alternatively, of a plurality of diagonal rows of pixels.
- image sensor array 182 B shown in FIGS. 5 g - 5 j are particularly well suited for use in picture taking optical readers which in bar code decoding applications are expected to decode linear bar code symbols.
- the image sensor arrays of FIGS. 5 g - 5 j may be referred to as linear symbol optimized image sensor arrays.
- image data corresponding to monochrome zones of pixels 2500 M in the versions of FIGS. 5 g - 5 j can be selectively addressed and read out independently of image data from rows from color sensitive zones 2500 C of pixels.
- control circuit 552 may selectively address pixels of monochrome zones 2500 M and read out image data from monochrome zones of pixels 2500 M as shown in FIGS. 5 g - 5 i and transfer such image data to indicia decode circuit 1702 for decoding of a linear bar code symbol.
- control circuit 552 may selectively address pixels of a color sensitive zone or zones of pixels 2500 C and selectively read out image data from color sensitive zone or zones 2500 C and process such color image data into a visual display color frame of image data.
- the processing as will be explained further herein may include such steps as executing a demosaicing routine to convert color filter pattern image data into a visual display format and interpolation of color pixel values corresponding to the missing pixel positions at the pixel positions occupied by a monochrome zone or zones 2500 M of pixels.
- FIGS. 6 a - 6 d exploded physical form view of pixels of image sensor array 182 , 182 B are shown.
- a monochrome pixel 250 M of image sensor array 182 B is shown in FIGS. 6 a and 6 b .
- Pixel 250 M includes a photodetector 302 which may be of photodiode or photogate construction, a transfer gate 304 , a floating diffusion 306 , a reset transistor 307 including reset gate 308 , a row select transistor 309 including row select gate 310 and a source follower amplifier transistor 311 including amplifier gate 312 .
- An important feature of pixel 250 M is opaque optical shield 316 .
- Opaque optical shield 316 typically comprising metal, shields light rays from components of pixel 250 M other than photodetector 302 . Accordingly, pixels from each of several rows of image sensor array 182 A can be simultaneously exposed to light in a global electronic shutter operating mode without the light rays modulating charges stored in floating diffusion 306 or another storage region. Further aspects of image sensor arrays capable of operating in a global electronic shutter operating mode are described in U.S. patent application Ser. No. 11/077,975 incorporated herein by reference. Referring to additional aspects of pixel 250 M, pixel 250 M includes microlens 320 which may be disposed on light transmissive protective layer 322 . Microlens 320 collects light from a larger surface area than photodetector 302 and directs light toward photodetector 302 .
- Color sensitive pixel 250 C of image sensor array 182 B is described with reference to FIGS. 6 c and 6 d .
- Color sensitive pixel 250 C is similar in construction to monochrome pixel 250 M.
- Color sensitive pixel 250 C includes a photodetector 302 which may be of photodiode or photogate construction, a transfer gate 304 for transferring charge from photodetector 250 C, a floating diffusion 306 , a reset transistor 307 including reset gate 308 , a row select transistor 309 including row select gate 310 and a source follower transistor amplifier 311 including amplifier gate 312 .
- Color sensitive pixel 250 C also includes opaque shield 320 which shields light from light sensitive components of pixel 250 C other than photodetector 302 .
- Pixel 250 C may also include microlens 320 for increasing the amount of light incident on photodetector 302 .
- color sensitive pixel 250 C includes a wavelength selective color filter element 260 formed thereon.
- Wavelength selective color filter element 260 may be disposed intermediate microlens 320 and protective layer 322 .
- Image sensor array 182 B may be of CMOS construction and may be an active pixel image sensor array such that each pixel 250 of image sensor array 182 B includes a pixel amplifier 311 . Each pixel 250 of image sensor array may further have a photosensitive region 252 and an optically shielded storage element 306 . Image sensor array 182 B further includes a two-dimensional grid of interconnects 262 which are in electrical communication with respective column circuitry 270 and row circuitry 296 . Row circuitry 296 and column circuitry 270 enable such processing and operational tasks as selectively addressing pixels, decoding pixels, amplification of signals, analog-to-digital conversion, and applying timing, read out and reset signals.
- Image sensor array 182 B may have multiple sets of reset control lines so that monochrome pixels 250 M of image sensor array 182 B can be reset independently of color sensitive pixels 250 C of image sensor array 182 B as described previously in connection with the description of image sensor array 182 B.
- control lines 262 R-M, 262 R-C can be incorporated in image sensor array 182 B in a layered manner so that for a substantial portion of image sensor array 182 B, control lines 262 R-M have x, y positions that coincide with x, y positions of control line 262 R-C (axes are defined in FIG. 8 a ). Control lines 262 R-C in the embodiment of FIG.
- image sensor array 182 B may be configured to include separate and independent reset control lines for separately and independently resetting monochrome pixels 250 M and color sensitive pixels 250 C without increasing the overall thickness of image sensor array 182 B. While disposing reset control lines on top of one another as described in connection with FIGS. 4 b and 7 b provides significant advantages; the image sensor array is made thicker with such arrangement which adds to manufacturing costs. Referring to FIG.
- FIG. 7 c a version of image sensor array 182 B is illustrated having a first set of reset control lines 262 , 262 R-M for resetting monochrome pixels 250 M and a second set of reset control lines 262 , 262 R-C for resetting color sensitive pixels 250 C of image sensor array 182 B.
- the reset control line configuration of FIG. 7 c may be utilized with the color sensitive pixel distribution shown in FIG. 5 f to provide an image sensor array 182 B having separate and independently controllable reset control lines for separately resetting monochrome pixels 250 M and color sensitive pixels 250 C and which exhibits a thickness equal to a thickness of a commonly available off-the-shelf image sensor array.
- FIG. 7 c a version of image sensor array 182 B is illustrated having a first set of reset control lines 262 , 262 R-M for resetting monochrome pixels 250 M and a second set of reset control lines 262 , 262 R-C for resetting color sensitive pixels 250 C of image sensor array 182
- the reset control lines of monochrome pixel rows are electrically connected together and the reset control lines of rows of pixels including color sensitive pixels are electrically connected together.
- the commonly connected reset control lines of the monochrome pixel rows are designated with the reference numeral 262 , 262 R-M, while the commonly reset control lines of the rows including color sensitive pixels are designated with the reference numeral 262 , 262 R-C.
- every fourth row of pixels of image sensor array 182 B includes clusters of color sensitive pixels 257 R, 257 G, 257 B. As shown in FIG.
- adjacent monochrome pixels 250 M-A adjacent to an end pixel, e.g., pixel 250 C-E of a color sensitive pixel cluster, e.g., cluster 257 R are not driven into reset during exposure periods of color sensitive pixels 250 C.
- image data corresponding only to a center pixel 250 C-I of each color sensitive horizontally arranged cluster (and not the end pixels 250 C-E) may be selectively addressed during read out of color image data.
- the presence of each lateral color filter element at the end pixels 250 C-E, which are not addressed for image data read out reduces the effect of cross talk attributable to photons entering image sensor array 182 B at an angle through end pixels 250 C, 250 C-E.
- image sensor array 182 B includes a plurality of rows of pixels including all monochrome pixels 250 M followed by a plurality of rows of pixels that include color sensitive pixels 250 C only.
- the monochrome rows of pixels 250 M form a first subset of pixels and the color sensitive pixels 250 C form a second subset of pixels.
- the reset control lines for resetting the first subset of pixels can be made separate and independent of the reset control lines for controlling the second subset of pixels by electrically connecting the reset control lines of the first subset of pixels together and then separately electrically connecting together the reset control lines of the second subset of pixels.
- the common control lines of the first subset of monochrome pixels 250 M in the version of FIG. 7 d are designated with reference numeral 262 , 262 R-M while the common control lines of the second subset of color sensitive pixels 250 C in the version of FIG. 7 d are designated with the reference numeral 262 , 262 R-C. It will be seen that the configuration of FIG.
- color sensitive pixels 250 C may be driven to reset during exposure periods for monochrome pixels 250 M by application of a common reset control signal on reset control line 262 , 262 R-M during exposure of color sensitive pixels 250 C for read out of color image data.
- color sensitive pixels 250 C may be driven into reset by application of a reset control signal on common reset control line 262 , 262 R-C during exposure periods of monochrome pixels 250 M for read out of image data from monochrome pixels 250 M.
- FIGS. 2 a - 4 c image sensor array 182 A
- FIGS. 5 a - 7 d image sensor array 182 B
- an image sensor array which may be incorporated into optical reader 100 (that is, features which can be incorporated in the image sensor array, whether of the embodiment labeled 182 A, the embodiment labeled 182 B, or another embodiment such as CMY image sensor array 182 C, RGB image sensor array 182 D, monochrome linear image sensor array 182 E, monochrome area image sensor array 182 F, monochrome and polarizer image sensor array 182 G, or monochrome color and polarizer image sensor array 182 H) are now described.
- Optical reader 100 can be programmed or otherwise be configured to selectively address a first plurality of pixels in an image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H independently of selectively addressing a second plurality of pixels of the image sensor array so that image data can be read out of the first plurality of pixels independently of the second plurality of pixels.
- optical reader 100 selectively addresses the first subset of pixels and reads out image data from first subset of pixels independently of the second color sensitive subset of pixels.
- optical reader 100 selectively addresses the second subset of pixels and reads out image data from the second subset of pixels independently of the first subset of pixels 250 M.
- optical reader 100 selectively addresses and selectively reads out only a subset of pixels of an image sensor array
- the resulting frame of image data read out of the image sensor array may be referred to as a “windowed frame” of image data.
- the frame rate of the image sensor array is normally increased relative to a normal frame rate of the image sensor array.
- Image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H can be configured to have a rolling shutter operating mode and a global shutter operating mode.
- a rolling shutter operating mode When a rolling shutter operating mode is entered, rows of pixels of image sensor array are exposed sequentially.
- the term “rolling” shutter is used because when in a rolling shutter operating mode an exposure period for a row of pixels generally begins prior to a time an exposure period for a previous row has ended.
- transistor components for example, transfer gates 304 and reset gates 308 of the array as shown in the embodiments of FIGS. 3 a and 6 a
- transistor components for example, transfer gates 304 and reset gates 308 of the array as shown in the embodiments of FIGS. 3 a and 6 a
- forming an electronic shutter an image sensor array are controlled in a coordinated manner so that a plurality of rows of pixels are exposed simultaneously and have common exposure periods.
- electronic shutter components of the array are controlled so that the common exposure period for each of the plurality of rows of pixels begins at a common exposure start time (via control of reset gates 308 ) and ends at a common exposure stop time (via control of transfer gates 304 ).
- each pixel of the array may store a charge in an optically shielded storage region during the common exposure period.
- an exposure control timing pulse 354 , 354 ′, 354 ′′, 354 ′′′ can be applied to an image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H, as is described in further detail in connection with the timing diagrams of FIGS. 15 a - 15 e .
- An exposure control timing pulse 354 , 354 ′, 354 ′′, 354 ′′′ controls the timing for exposure of each row of pixels of image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H being exposed.
- the exposure period for each row of pixels the image sensor array being subject to image data read out begins at the leading edge of exposure control timing pulse 354 , 354 ′, 354 ′′, 354 ′′′ and ends at the falling edge of exposure control timing pulse 354 , 354 ′, 354 ′′, 354 ′′′.
- each pixel of the array is equipped with additional circuit elements as is described herein.
- Image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H of optical reader 100 may be constructed to be operated in a rolling shutter mode of operation only; that is, in one specific embodiment an image sensor array of optical reader 100 can only be controlled to expose pixels of the image sensor array on a rolling shutter basis and cannot be controlled so that pixels of image sensor array are exposed on a global electronic shutter basis.
- an image sensor array incorporated in optical reader 100 is constructed to be operated in a global electronic shutter operational mode only and is incapable of operation in a rolling shutter mode.
- Image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H can be constructed to be operated in either of a global electronic shutter operation mode or a rolling shutter operational mode.
- an image sensor array incorporated in optical reader 100 is constructed to be operational in either of a rolling shutter operational mode or a global shutter operational mode
- the switching between rolling shutter and global shutter operational modes may be made in response to a receipt of operator instructions to change the shutter mode.
- the switching between rolling shutter and global shutter operational modes may also be automatic and dynamic in response to the sensing of a predetermined criteria being satisfied.
- An image sensor array equipped optical reader 100 having both rolling shutter and global shutter operational modes is described in U.S.
- Image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H can be constructed so that certain pixels of the image sensor array are capable of being exposed on either a rolling shutter basis or a global shutter basis and certain other pixels of the image sensor array are capable of being exposed only on a rolling shutter basis and are not capable of being exposed on a global electronic shutter basis.
- image sensor array 182 A, and image sensor array 182 B it may be advantageous to incorporate into an image sensor array of optical reader 100 separately controllable reset control lines 262 , 262 R-M and 262 , 262 R-C for resetting monochrome pixels separately and independently of color sensitive pixels to thereby reduce pixel cross talk. It will be understood that it may be advantageous to incorporate separately and independently controllable reset control lines into an image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H according to the invention whenever image data is selectively read out of a first subset of image data and it is desired to reduce cross talk from pixels of the image sensor array external to the first subset of pixels.
- optical reader 100 incorporating a cyan-magenta-yellow (CMY) image sensor array 182 C as shown in FIG. 18 c
- CCMY cyan-magenta-yellow
- the cyan and magenta pixels can be set to reset to eliminate electron diffusion cross talk and to reduce photon penetration cross talk.
- hand held optical reader 100 incorporates a hybrid monochrome and polarizer image sensor array 182 G as shown in FIG. 19 b or the monochrome color and polarizer image sensor array 182 H as shown in FIGS.
- image sensor array 182 it may be beneficial to incorporate into image sensor array 182 separately controllable reset control lines for controlling the reset of pixels external to the polarizing pixels so that when the polarizing pixels are exposed for read out of image data from the polarizing pixels, the remaining pixels of the image sensor array are set to reset to reduce cross talk from the pixels external to the polarizing pixels.
- an image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H is conveniently provided by a CMOS image sensor array fabricated utilizing complementary metal-oxide-silicone integrated circuit fabrication technologies
- an image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H may also be a charge coupled device (CCD) image sensor array, or a CID image sensor array or an image sensor array of another fabrication technology.
- CCD charge coupled device
- a read out of a “windowed frame” of image data which may also be referred to as an image region of interest (ROI).
- ROI image region of interest
- An example of a CCD image sensor array integrated circuit chip having windowing capability is the KODAK KAI-0340 image sensor array IC chip available from Eastman Kodak Corporation of Rochester, N.Y.
- the KAI-0340 image sensor array IC chip has various operational modes that are selectable utilizing various input switch setups.
- setting a SW1 switch to the HIGH position causes charge in outer vertical resisters of the image sensor array to be dumped before it reaches the horizontal register, facilitating the selective read out of image data from center columns of the array only.
- Setting the SW2 switch of the KAI-0340 image sensor array chip changes diode transfer clock timing such that only charge from the center rows is transferred to vertical registers, facilitating the selective read out of image data from center rows of the image sensor array only. Accordingly, where image sensor array 182 B is configured according to the version shown in FIG.
- image data from the center rows can be read out by selecting a preconfigured operating mode of the image sensor array chip.
- Additional “windowed frame” patterns can be selectively read out of a CCD image sensor array by varying the speed of a pixel clock timing control timing pulse that controls the speed with which a pixel is clocked.
- Invalid or null data can be clocked out of a CCD pixel by speeding up a pixel clock signal.
- Varying a pixel clock control signal between valid data yielding rates and invalid data yielding rates during the reading out of image data from a CCD image sensor array yields a windowed frame of image data comprising valid image data clocked out at normal speed and invalid image data clocked out at high speed.
- Image data can also be selectively read out of a CCD image sensor array by selectively gating to output circuitry of the CCD image sensor array image data corresponding to select pixels of the image sensor array. It will be seen that for any application described herein wherein a windowed frame of image data is read by selective addressing of pixels from a CMOS image array, a CCD image sensor array supporting windowing capability may be substituted therefore to provide selective read out functionality.
- FIGS. 8 a - 8 c Additional aspects of the invention are described with reference to the physical form views of FIGS. 8 a - 8 c and the physical form views 9 a , 9 b and 9 c .
- an imaging module onto which an image sensor chip may be incorporated is described.
- hand held housings for supporting and encapsulating an imaging module including an image sensor chip are described.
- An optical reader 100 of the invention may include an imaging module such as imaging module 1802 A.
- Imaging module 1802 A as shown in FIGS. 8 a - 8 c incorporates certain features of an IT4000 imaging module herein and additional features.
- IT4000 imaging modules are available from Hand Held Products, Inc. of Skaneateles Falls, N.Y.
- Imaging module 1802 A includes first circuit board 1804 carrying light sources 160 a , 160 b , while second circuit board 1806 carries light sources 160 c , 160 d , 160 e , 160 f , 160 g , 160 h , 160 i , 160 j , 160 k , 160 l , 160 m , 160 n , 160 o , 160 p , 160 q , 160 r , 160 s , and 160 t (hereinafter 160 c through 160 t ).
- First circuit board 1804 also carries image sensor array 182 , which is integrated onto image sensor IC chip 1082 . Image sensor IC chip 1082 and image sensor array 182 in FIG.
- FIGS. 8 a - 8 d are generically labeled with the reference numerals “ 1082 ” and “ 182 ” respectively in FIGS. 8 a - 8 d to indicate that any one of the specifically described image sensor IC chips 1082 A, 1082 B, 1082 C, 1082 D, 1082 E, 1082 F, 1082 G, 1082 H described herein or any one of the specifically described image sensor arrays 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H described herein may be incorporated into imaging module 1802 A.
- the various image sensor IC chips and image sensor arrays can also be incorporated in another imaging module described herein such as imaging module 1802 B, 1802 C, 1802 D, and 1802 E.
- Imaging module 1802 C shown in FIG. 8 e is a laser aiming IT4300 imaging module of the type available from Hand Held Products, Inc.
- the laser aiming IT4300 imaging module includes a plurality of illumination LEDs e.g., LED 160 , and an aiming pattern generator comprising a laser diode assembly 1872 in combination with a diffractive element 1873 , wherein the diffractive element of the imaging module diffracts laser light from the laser diode assembly to project a two-dimensional aiming pattern onto a substrate, s.
- Imaging module 1802 A also includes support assembly 1810 including lens holder 1812 , which holds lens barrel 1814 that carries imaging lens 212 that focuses images onto an active surface of image sensor array 182 .
- Lens 212 may be e.g., a single lens (a lens singlet), a lens doublet or a lens triplet.
- Light sources 160 a , 160 b are aiming illumination light sources whereas light sources 160 c through 160 t are illumination light sources.
- illumination light sources 160 c through 160 t project a two-dimensional illumination pattern 1830 over a substrate, s, that carries a decodable indicia such as a bar code symbol 1835 whereas aiming illumination light sources 160 a , 160 b project an aiming pattern 1838 .
- a decodable indicia such as a bar code symbol 1835
- aiming illumination light sources 160 a , 160 b project an aiming pattern 1838 .
- aiming illumination light sources 160 a , 160 b is shaped by slit apertures 1840 in combination with lenses 1842 which image slits 1840 onto substrate, s, to form aiming pattern 1838 which in the embodiment of FIGS. 8 a - 8 c is a line pattern 1838 .
- Illumination pattern 1830 substantially corresponds to a full frame field of view of image reader 100 designated by box 1850 .
- the present field of view of optical reader 100 is herein referred to as the “target” of optical reader 100 .
- Aiming pattern 1838 is in the form of a line that extends horizontally across a center of field of view of image reader 100 .
- Illumination pattern 1830 may be projected when all of illumination light sources 160 c through 160 t are operated simultaneously. Illumination pattern 1830 may also be projected when a subset of light sources 160 c through 160 t are simultaneously energized. Illumination pattern 1830 may also be projected when only one of light sources 160 c through 160 t is energized such as LED 160 s or LED 160 t . LEDs 160 s and 160 t of imaging module 1802 have a wider projection angle than LEDs 160 c through 160 t .
- a illumination assembly 104 includes LEDs 160 a , 160 b , LEDs 160 c through 160 t and slit apertures 1840 in combination with lenses 1842 .
- a reader imaging module may be incorporated into one of a hand held housing as shown in FIGS. 9 a , 9 b and 9 c .
- hand held housing 101 is a gun style housing.
- hand held housing 101 supporting imaging module 1802 is in the form factor of a portable data terminal (PDT).
- PDT portable data terminal
- hand held housing 101 supporting imaging module is in the form factor of a mobile telephone, often referred to as a “cell phone.”
- optical reader 100 is a cell phone
- optical reader 100 is configured to send voice data over GSM/GPRS transceiver 571 to GSM/GPRS network 198 ( FIG.
- optical reader 100 may be configured so that an operator inputs telephone numbers via keyboard 508 .
- the specific imaging module 1802 A described in connection with FIGS. 8 a - 8 c may be incorporated in the optical reader shown in FIG. 9 a or the optical reader 100 shown in FIG. 9 b or the optical reader 100 shown in FIG. 9 c .
- housing 101 supports and encapsulates imaging module 1802 B an imaging module of construction similar to imaging module 1802 A, except that only two light sources 160 are incorporated into the imaging module. Housing 101 of the reader of FIG.
- imaging module 1802 which is generically labeled element 1802 to indicate that any one of the specific imager modules described herein, e.g., 1802 , 1802 A, 1802 B, 1802 D, 1802 E may be incorporated into an optical reader according to the invention.
- optical reader 100 may incorporate a graphical user interface (GUI) 3170 enabling selection between various operating modes.
- GUI 3170 an operator moves pointer 3172 to a selected icon and clicks on the icon to configure optical reader 100 in accordance with an operating mode associated with the selected icon.
- Reader 100 may include pointer mover 512 (otherwise termed a navigation matrix) to facilitate movement of the pointer 3172 .
- Buttons 512 B of pointer mover 512 facilitate selection of an icon of a GUI interface that is supported by incorporating a multitasking operating system (OS) into reader 100 such as WINDOWS CE.
- GUI 3172 may be developed using various open standard languages as HTML/Java or XML/Java.
- GUI 3170 includes a plurality of virtual selection buttons 3152 , 3154 , 3156 , 3158 , 3162 , 3164 .
- Selection of rolling shutter icon 3152 configures reader 100 so that during a next exposure period image sensor array 182 is operated in a rolling shutter mode.
- Selection of global shutter icon 3154 configures optical reader 100 so that during a next exposure period image sensor array 182 is operated in a global electronic shutter mode.
- Selection of decode icon 3162 drives optical reader 100 into an indicia decode mode so that a next time a trigger signal is received, optical reader 100 captures a frame of image data and attempts to decode a bar code symbol or other decodable indicia (e.g., an OCR character) represented therein and outputs a decoded out message to display 504 , or a spaced apart device 150 , as is described with reference to FIG. 10 .
- a bar code symbol or other decodable indicia e.g., an OCR character
- Selection of image capture (which may otherwise be referred to a picture taking) icon 3164 configures optical reader 100 so that next time a trigger signal is received, optical reader 100 captures image data and outputs the image data to one or more of a display 504 , a specified memory address, or to a spaced apart device 150 without attempting to decode decodable indicia therein.
- Optical reader 100 may also be constructed so that optical reader 100 can be configured in accordance with a selected operating mode by sending to reader 100 a serial command from a spaced apart device, or by the reading of a specially configured programming bar code symbol.
- Optical reader 100 is configured so that optical reader 100 receives a trigger signal when manual trigger 216 is manually depressed by an operator. Optical reader 100 may also be configured so that a trigger signal is received by the sensing of an object in the proximity of reader 100 or by the sending of a serial trigger command to reader 100 from a spaced apart device, 150 , as shown in FIG. 10 .
- FIGS. 14 a , 14 b , and 14 c A flow diagram illustrating operation of optical reader 100 in one embodiment is described with reference to FIGS. 14 a , 14 b , and 14 c .
- an operator selects between an indicia decode mode and a picture taking mode.
- an operator may select icon 3162 ( FIG. 9 b ) to drive optical reader 100 into an indicia decode mode, or alternatively icon 3164 to drive optical reader 100 into a digital picture taking mode of operation.
- These modes may also be selected by sending to reader 100 a serial command from a spaced apart device 150 or by the reading of a programming bar code symbol.
- optical reader 100 executes an indicia decode process 1102 .
- a picture taking mode optical reader 100 executes picture taking process 1400 .
- a trigger signal is received by one of the methods described (depressing trigger 216 , object sensing, serial trigger command) to commence a decode process.
- control circuit 552 of optical reader 100 captures a plurality of “parameter determination” or test frames of image data.
- the frames of image data captured at step 1203 are not subject to indicia decode processing, but rather, are processed for parameter determination (e.g., exposure, gain, illumination).
- parameter determining step 1203 may be avoided.
- control circuit 552 may apply parameters determined from a previous image capture operation rather than determining parameters at step 1203 .
- control circuit 552 obtains a decode frame of image data details of which are explained herein.
- control circuit 552 may send an illumination control signal to illumination assembly 104 and various image capture initiation control signals to control/timing circuit 1092 of image sensor chip 1082 (labeled generically to refer to any of the image sensor chips described herein).
- control circuit 552 may send to illumination assembly 104 an illumination control timing pulse 350 to energize at least one light source 160 such that illumination pattern 1830 is projected (as shown in FIG. 8 d ).
- Control circuit 552 may also send to image sensor IC chip 1082 an exposure control timing pulse 354 and a read out control timing pulse 368 , and a reset control timing pulse 370 (that is, control circuit 552 sends appropriate signals to image sensor IC chip 1082 to initiate exposure control timing pulse 354 , read out control timing pulse 368 and reset control timing pulse 370 ).
- the exposure control timing pulse 354 begins after and finishes before the illumination control timing pulse 350 .
- the read out control timing pulse 368 begins at the conclusion of the illumination control timing pulse 350 .
- the illumination control timing pulse 350 ′ begins after and finishes before the exposure control timing pulse 354 ′.
- the read out control timing pulse 368 ′ begins at the conclusion of the exposure control timing pulse 354 ′.
- the exposure control timing pulse and the illumination control timing pulse overlap each other while occurring sequentially. In one such embodiment as shown in FIG.
- this sequential operation can include the illumination control timing pulse 350 ′′ starting, the exposure control timing pulse 354 ′′ starting, the illumination control timing signal pulse 350 ′′ ending, and then the exposure control timing pulse 354 ′′ ending.
- the read out control timing pulse 368 ′′ begins at the conclusion of the exposure control timing pulse 354 ′′.
- the sequential operation can include the exposure control timing pulse 354 ′′′ starting, the illumination control timing pulse 350 ′′′ starting, the exposure control timing pulse 354 ′′′ ending, and then the illumination control timing signal pulse 350 ′′′ ending.
- the read out control timing pulse 368 ′′′ begins at the conclusion of the illumination control timing signal pulse 350 ′′′.
- Each illumination control timing pulse 350 , 350 ′, 350 ′′, 350 ′′′ described herein may comprise a plurality of short duration individual pulses, sometimes referred to as a “strobed” pulse, as is indicated by FIG. 15 e.
- exposure control timing pulse 354 is received by an image sensor IC chip and optical reader 100 is configured in a global electronic shutter operating mode
- pixels from several rows of image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H are simultaneously exposed to light for the duration of the pulse. That is, when optical reader 100 is configured in an global electronic shutter operating mode, each of several rows of image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H that are subject to image data read out have common exposure periods such that the exposure period for each row of pixels the image sensor array subject to image data read out begins at a common exposure start time and ends at a common exposure stop time.
- the exposure period for each row of pixels image sensor array 182 A, 182 B, 182 C, 182 D, 182 E, 182 F, 182 G, 182 H subject to image data read out begins at the leading edge of exposure control timing pulse 354 , 354 ′, 354 ′′, 354 ′′′ and ends at the falling edge of exposure control timing pulse 354 , 354 ′, 354 ′′, 354 ′′′.
- image sensor IC chip 1082 B When read out control timing pulse 368 is received by image sensor IC chip 1082 B, image data is read out from the image sensor array. Image signals corresponding to pixels of the image sensor array are converted into digital form by analog-to-digital converter 1086 and transferred into memory 560 by FPGA 580 .
- Optical reader 100 may be configured so that at step 1204 when reading out image data, optical reader 100 reads a “windowed frame” of image data.
- a windowed frame of image data may be read out by selectively addressing pixels of a desired region of interest or window.
- a windowed frame of image data read out during frame capture at step 1204 may comprise pixel values corresponding to all or substantially all monochrome pixels 250 M of an image sensor array.
- a reset control timing pulse 370 for resetting pixels that are not selectively addressed may be coordinated with the exposure control timing pulse 354 for controlling exposure of pixels that are selectively addressed for reading out a windowed frame of image data.
- reset control timing pulse 3709 is applied to reset color pixels of the image sensor array 182 while exposure control timing pulse 354 is applied to enable exposure of monochrome pixels of the image sensor array.
- an image sensor array may be configured to include a reset control line grid specifically adapted to enable resetting of color pixels. Applying reset control pulse 370 to drive color pixels into reset while monochrome pixels are being exposed to light can be expected to eliminate electron diffusion cross talk and to reduce cross talk resulting from light rays angularly entering the color pixels during exposure.
- frames are obtained at step 1204 , they are obtained in a form suitable to facilitate indicia decoding such as bar code symbol decoding or OCR decoding.
- control circuit 552 at step 1204 may store gray scale values into RAM 560 , each pixel value representing an intensity of light at a particular monochrome pixel of image sensor array 182 A, 182 B.
- the frame of image data obtained at step 1204 may include e.g., 8 bit gray scale pixel values, 10 bit gray scale pixel values or 12 bit gray scale pixel values.
- control circuit 552 at step 1204 may obtain a decode frame of image data including color image data.
- decode circuit 1702 is configured to decode color encoded bar code symbols, it is advantageous for control circuit 552 to obtain a decode frame of image data including color image data at step 1204 .
- control circuit 552 may carry out a plurality of alternative processes in obtaining a decode frame of image data.
- optical reader 100 at step 1204 may simply capture a single windowed frame of image data which has been described herein above.
- control circuit 552 may carry out process step 1204 by selectively addressing monochrome pixels 250 M of a hybrid monochrome and color image sensor array such as image sensor array 182 A or image sensor array 182 B and reading out image data from monochrome pixels 250 M only; i.e., a windowed frame of image data comprising image data from monochrome pixels 250 M only.
- the obtaining a decode frame step 1204 may be carried out in the alternative by the execution of steps 1206 and 1207 .
- optical reader 100 may generate a frame of image data that includes image data corresponding to monochrome pixels 250 M and color sensitive pixels 250 C and at step 1207 image sensor array 182 A, 182 B may convert pixel values of the frame generated at step 1206 into gray scale values.
- the frame generated at step 1206 may be generated by exposing color and monochrome pixels of image sensor array 182 A, 182 B during a single exposure period, and reading out image data from both color and monochrome pixels 250 M, 250 C of image sensor array 182 A, 182 B during a single pixel read out period.
- control circuit 552 of optical reader 100 may combine image data from two different frames such as two successive frames, wherein a first of the captured frame is a windowed frame of image data including image data from color sensitive pixels 250 C only and a second of the frames is a frame of image data including image data read out from monochrome pixels 250 M only.
- optical reader 100 may also obtain a decode frame at step 1204 by carrying out steps 1208 and step 1209 .
- optical reader 100 may capture a windowed frame of image data including image data corresponding to monochrome pixels 250 M only and at step 1209 control circuit 552 may interpolate pixel values corresponding to color pixel positions of image sensor array 182 A, 182 B utilizing the monochrome pixel values from the windowed monochrome frame captured at step 1208 .
- control circuit 552 may capture a gray scale pixel value frame 5202 as illustrated in FIG. 16 a that includes a gray scale pixel value for each monochrome pixel position of the image sensor array 182 A, 182 B.
- Optical reader 100 may interpolate a monochrome pixel value for any “missing pixel” color pixel position of the frame 5202 .
- Pixel positions P 11 , P 31 , P 51 , P 12 , P 22 , P 32 , P 42 , P 52 , P 13 , P 33 , P 63 . . . are pixel positions corresponding to monochrome pixels 250 M of image sensor array 182 for which individual frame image data has been read out.
- Pixel positions P 21 , P 41 , P 23 , P 43 , . . . are missing pixel positions corresponding to color sensitive pixels, 250 C of image sensor array 182 B. Referring to the frame of image data represented at FIG.
- an optical reader 100 may calculate a gray scale pixel value for color pixel positions, e.g., position P 23 , by averaging the gray scale values for each pixel position that is adjacent to pixel position P 23 and each pixel position that is corner adjacent to color pixel position P 23 .
- a gray scale value for color pixel position P 23 can be interpolated by averaging pixel values of pixel positions P 12 , P 22 , P 32 , P 13 , P 33 , P 14 , P 24 , P 34 .
- a pixel value for “missing pixel” position P 23 can also be interpolated utilizing more than 8 neighboring pixel positions.
- corner adjacent pixels may be weighted less than side, top or bottom adjacent pixels during averaging.
- only four surrounding pixels are averaged; namely, the top and bottom adjacent pixels and the two side adjacent pixels adjacent to the pixel position for which a gray scale value is being interpolated.
- only two pixels are used for averaging; namely either the two side adjacent pixels adjacent to the pixel position being interpolated or the top and bottom adjacent pixels.
- a two-dimensional image representation of a linear bar code symbol can be expected to have several consecutive pixel positions along a column with similar gray scale values, if the representation of the symbol is oriented with 0° or 180° angle of rotation (i.e., the symbol is right side up or upside down).
- the symbol representation has a 90° or 280° angle of rotation, several consecutive pixel positions along rows of pixel positions can be expected to have similar valued gray scale values. Accordingly, it can be seen that interpolating pixel values of adjacent pixel position values running in the direction of bars in a symbol representation yields truer edge information than utilizing all surrounding pixel positions for interpolation.
- the correlation between a pair of horizontally oriented scan lines is calculated along with a correlation between a pair of vertically oriented scan lines.
- the two correlation measurements are then compared. If row scan lines are more closely correlated, column adjacent pixels are selected for interpolation. If column scan lines are more closely correlated, row adjacent pixels are selected for interpolation.
- An exemplary set of code for calculating a first derivative correlation for a pair of scan lines (horizontal or vertical) is presented by Table 1.
- optical reader 100 transfers the frame of image data obtained at step 1204 , to an indicia decode circuit 1702 which may be a bar code symbol decoding circuit or autodiscrimination circuit 1704 including an indicia decode circuit 1702 .
- decode circuit 1702 decodes 1D and 2D bar code symbols and OCR characters.
- Autodiscrimination circuit 1704 may decode 1D and 2D bar code symbols and OCR characters (decodable indicia) and automatically discriminate between decodable indicia and handwritten characters.
- autodiscrimination circuit 1704 In the event that autodiscrimination circuit 1704 recognizes the presence of handwritten character information, autodiscrimination circuit 1704 automatically outputs to display 504 and/or a spaced apart device 150 image data representing the handwritten character image data. Further details of indicia decode circuit 1702 and autodiscrimination circuit 1704 are described in copending U.S. patent application Ser. No. 11/077,975, filed Mar. 11, 2005, incorporated by reference and U.S. application Ser. No. 10/958,779, filed Oct. 5, 2004, also incorporated herein by reference.
- indicia decoding accuracy is expected to increase with an increase in the percentage of monochrome pixels of image sensor array 182 A, 182 B.
- control circuit 552 may read image data along scan lines, such as scan lines defined by a horizontal row of pixel positions to determine the relative widths of bars and spaces of a symbol and then decode the symbol through table lookup to determine a set of decoded character data corresponding to the bar space information.
- control circuit 552 receives a decoded output message from decode circuit 1702 or autodiscrimination circuit 1704 .
- the message received by control circuit 552 at step 1212 may be e.g., a decoded bar code message or a set of decoded OCR characters.
- optical reader 100 outputs a decoded out message.
- control circuit 552 may send decoded out bar code data and/or decoded OCR data to display 504 or to a spaced apart device 150 or to a data storage memory location of reader 100 , or system 145 as described in FIG. 10 .
- Optical reader 100 may be part of a system 145 and may be included in a local area network (LAN) 170 which comprises, in addition to reader 100 , such spaced apart devices as other portable readers 100 ′, 100 ′′, network access point 174 , personal computers 172 and central server 176 that are spaced apart from hand held housing 101 of reader 100 , all of which are connected together via backbone 177 .
- Server 176 in turn is in communication with a variety of additional spaced apart devices 150 that are spaced apart from hand held housing 101 of reader 100 and which through server 176 are in communication with optical reader 100 .
- Server 176 may be connected via gateways 179 , 180 and network 181 to a first distant remote local area network 185 located miles to thousands of miles away from local area network 170 and a second distant local area network 2170 also located miles to thousands of miles away from local area network.
- Network 170 may be located at a supplier's warehouse.
- Network 2170 may be located at a delivery destination; and network 185 may be located at a data processing/data archiving facility.
- Network 185 can be configured to assemble, store and maintain in server 184 various web pages accessible with use of optical reader 100 , that summarize data that has been collected by various optical readers 100 , 100 ′, 100 ′′, 100 R.
- Server 176 may alternatively or redundantly be connected to remote network 185 via private communication line 190 .
- IP network 181 may be the Internet or a virtual private network (VPN).
- Remote LAN 185 may include a personal computer 186 and a remote server 184 connected via backbone 191 .
- Remote LAN 185 may also include a wireless communication access point 193 .
- Remote LAN 185 may also include a personal data assistant (PDA) 189 .
- Remote LAN 2170 may include a server 2176 , connected to IP network 181 via gateway 2179 , backbone 2177 , access point 2174 , PC 2172 , and optical reader 100 , 100 R.
- System 145 may be configured so that a display equipped device, e.g., device 100 ′, 172 , 186 , 189 automatically displays data, such as decoded out bar code message data of a visual display color image frame of image data, received from optical reader 100 on its associated display 1504 when receiving that data.
- a display equipped device e.g., device 100 ′, 172 , 186 , 189 automatically displays data, such as decoded out bar code message data of a visual display color image frame of image data, received from optical reader 100 on its associated display 1504 when receiving that data.
- All of the steps of process 1200 are carried out automatically in response to the receipt of a trigger signal.
- the steps of process 1200 continue automatically until a stop condition is satisfied.
- a stop condition may be e.g., the receipt of a trigger stop signal as may be generated by release of trigger 216 or the successful decoding of a predetermined number of bar code symbols.
- control circuit 552 may repeatedly attempt to obtain image data and attempt to decode decodable indicia therein until a stop condition is satisfied.
- Interpolating monochrome pixel values for “missing pixels” pixel positions is particularly advantageous where a hybrid monochrome and color image sensor array according to the invention includes a high number of color sensitive pixels distributed at spaced apart pixel positions throughout image sensor array.
- control circuit 552 may obtain a decode frame of image data at step 1204 that is suitable for transferring to decode circuit 1702 by simply reading from an image sensor array image data from monochrome pixels 250 M without interpolation of any pixel values at “missing pixel” pixel positions where an image sensor array of reader 100 is constructed in accordance with image sensor array 182 A, and monochrome pixels 250 M form a complete checkerboard pattern, (there are no “missing” monochrome pixels positions in the M ⁇ N matrix of monochrome pixels within image sensor array 182 A). Accordingly, where optical reader 100 includes image sensor 182 A, the decode frame of image data at step 1204 is advantageously obtained by reading out from image sensor array 182 A image data from monochrome pixels 250 M without interpolation of any further monochrome pixel values.
- optical reader 100 includes a linear bar code symbol optimized image sensor array 182 B of one of the versions described in connection with FIGS. 5 g - 5 j .
- image sensor array 182 B includes at least one “zone” of monochrome pixels 2500 M and at least one “zone” of color sensitive pixels 2500 C.
- Monochrome zone of pixels 2500 M of a linear symbol decoding optimized version of image sensor array 182 B generally comprises an elongated line of monochrome of pixels 250 M having a minor dimension of one to a limited number of rows of pixels (which may be diagonal rows).
- optical reader 100 includes a linear symbol optimized version of image sensor array 182 B a reduced area decode frame of image data at step 1204 without any “missing pixel” pixel positions can be obtained by selectively addressing pixels of monochrome zone 2500 M and selectively reading out of image data from the pixels of monochrome zone 2500 M without reading any image data from the pixels of color sensitive zone of pixels 2500 C.
- control circuit 552 at step 1204 may obtain a reduced area monochrome frame of image data at step 1204 ( FIG. 14 b ) and then transfer the reduced area monochrome frame of image data to decode circuit 1702 at step 1210 .
- a reduced area frame of image data is further explained with reference to FIG. 11 illustrating an application where an optical reader 100 is utilized to collect decoded bar code data and image data from a parcel 1260 carrying various bar code symbols, e.g., linear bar code symbol 1266 and two-dimensional bar code symbol 1270 .
- a full area frame of image data that may be obtained by optical reader 100 represents the scene area indicated by rectangular border 1276 of FIG. 11 .
- a full area frame of image data is obtained.
- optical reader 100 obtains a reduced area frame of image data
- a reduced area frame of image data representing the reduced scene area as indicated by dashed-in border 1278 may be obtained.
- optical reader 100 may incorporate one of the linear symbol optimized image sensor arrays 182 B as shown in FIGS. 5 g and 5 h .
- a reduced area frame of image data representing the reduced scene area 1278 may be obtained by reading out image data from monochrome zone of thin center line monochrome zone 2500 M of image sensor array 182 B according to one of the versions of FIGS. 5 g and 5 h . It is seen with reference to FIG.
- Imaging module 1802 such as module 1802 A ( FIG. 8 a ) of reader 100 can be constructed so that aiming pattern 1838 ( FIG. 8 d ) is projected onto scene area 1278 at expected reading angles, while aiming light sources 160 a , 160 b , and remaining light sources 160 c - 160 t are energized simultaneously during the time that pixels of zone 2500 M are exposed for read out of image data from zone 2500 M.
- decode circuit 1702 may process the thin line reduced area decode frame of image data to decode linear bar code symbol 1266 by calculating the bar space widths of the bars and spaces of linear bar code symbol 1266 and then determining the characters of the symbol through table lookup.
- optical reader 100 may be configured so that aiming pattern 1838 ( FIG.
- optical reader 100 when obtaining the reduced area decode frame of image data at step 1204 may be significantly reduced relative to the frame rate of optical reader 100 when obtaining a full frame of image data. Accordingly, a method of the invention where optical reader 100 at step 1204 obtains a reduced area frame of image data which is transferred to decode circuit 1702 is optimized for fast (“snappy”) decoding. As has been described herein, color sensitive pixels 250 C may be set to reset while monochrome pixels 250 M are exposed for selective read out of image data from monochrome pixels 250 M.
- optical reader 100 can be configured so that where indicia decode circuit 1702 cannot successfully decode a bar code symbol via processing of a reduced area frame of image data or where control circuit 552 is programmed to search and decode multiple bar code symbols, control circuit 552 executes return line ( FIG. 14 b ) to re-execute the obtaining of a decode frame of image data at step 1204 .
- control circuit 552 executes step 1204 the second time, control circuit 552 captures a frame of image data that represents a scene area that is larger than the scene area represented by the frame obtained during the first execution of step 1204 .
- the decode frame of image data obtained by a second execution of step 1204 may be a full area image data frame representing the full field of view of optical reader 100 indicated by dashed-in border 1276 of FIG. 11 .
- control circuit 552 during the second execution of obtain step 1204 may selectively read out image data from the green pixels of color sensitive zones of image sensor array of 2500 C and interpolate green pixels values at non-green pixel positions utilizing the green pixel values so that the decode frame of image data obtained at step 1204 includes all green pixel values. Further, the missing pixel positions corresponding to monochrome zone 2500 M can be filled in utilizing the image data obtained during the previous execution of step 1204 as scaled based on a relationship between the color scale values of pixels corresponding to zone 2500 M and the color scale values of pixels surrounding zone 2500 M.
- larger area green image data is transferred to indicia decode circuit 1702 .
- Indicia decode circuit 1702 may attempt to decode linear bar code symbol 1266 and all other bar code symbols such as two-dimensional bar code symbol 1270 that may be represented in the image obtained during the second execution of step 1204 .
- optical reader 100 incorporating a linear symbol decode optimized to image sensor array 182 B may attempt to decode linear symbol 1266 utilizing small area image representing area 1278 and then subsequently attempt to decode a two-dimensional bar code symbol, e.g., symbol 1270 , utilizing a larger area frame of image data representing scene area 1276 .
- control circuit 552 obtains a reduced area frame of image data, attempts to decode, then subsequently obtains a larger frame of image data and attempts to decode utilizing the larger image may be practiced utilizing an “all monochrome” image sensor array 182 F as shown and described in connection with FIG. 17 b .
- optical reader 100 incorporates an all monochrome image sensor array 182 F as shown in FIG. 17 b , it is particularly useful to set monochrome pixels 250 M to reset that are not being selected for read out of a reduced area decode frame of image data at step 1204 during exposure periods for selected monochrome pixels 250 M that are being selectively addressed for image data read out.
- Monochrome pixels 250 M transmit more light than color sensitive pixels 250 C. Therefore, resetting monochrome pixels 250 M that are not selectively addressed and which are adjacent to a region of interest during an exposure period can be expected to have a greater benefit in terms of improving the overall signal to noise ratio of reader 100 than resetting color sensitive pixels 250 C that are adjacent to a region of interest during an exposure period.
- optical reader 100 it may be advantageous to utilize optical reader 100 to obtain a visual display color frame of image data representing parcel 1260 .
- parcel 1260 may include a damaged area 1272 .
- Obtaining a visual display color frame of image data corresponding to parcel 1260 creates a record documenting parcel damage.
- different optical readers 100 and 100 R at different locations A and B located miles apart along a delivery route may be utilized to document physical transformations of parcel 1260 when parcel 1260 is carried along a delivery route.
- Optical reader 100 at location A including LAN 170 ( FIG. 10 ) may be operated to obtain a visual display color frame of image data of parcel 1260 when parcel 1260 is located at location A.
- the color frame may automatically be transferred to remote server 184 ( FIG. 10 ) having a database 187 of color frames of image data that are indexed by a parcel identifier decoded in a parcel bar code symbol 1266 which identifier is also transmitted to remote server 184 automatically when optical reader 100 reads bar code symbol 1266 .
- remote optical reader 100 , 100 R ( FIG. 10 ) may be utilized to again decode bar code symbol 1266 and capture visual display color frame of image data representing parcel 1266 and automatically transfer the parcel identifier corresponding to bar code 1266 and the color frame of image data to remote server 184 .
- the color frame of image data transmitted to remote server 184 from location B will include a representation of damaged area 1272 that is not included in the color frame of image data transmitted to remote server 184 from location A. Accordingly, a person (for example, at PC 172 viewing web pages of server 184 ) reviewing the parcel identifier indexed color frame data of database 187 can determine that the damage to parcel 1260 occurred during the time that the parcel was delivered from location A to location B.
- optical reader 100 can also be utilized to take color pictures of a delivery vehicle 1282 that carried parcel 1260 from location A to location B. In the example of FIG. 12 b , a picture being taken by optical reader 100 has the field of view indicated by rectangle 1286 .
- the field of view encompasses parcel 1260 , and delivery vehicle 1282 , including a license plate 1284 .
- Trigger 216 can be actuated a first time to decode bar code symbols 1266 , 1267 and then an additional time or times to have a picture of parcel 1260 and/or vehicle 1272 including a picture of license plate 1284 .
- the decoded bar code data and multiple color frames of image data may be associated with one another into a single transaction data set, and then via a packet based transmission scheme, the transaction data set may be sent to remote server 184 , which may organize the data into viewable web pages viewable at PC 172 .
- Optical reader 100 which may be incorporated in hand held housing 101 , can be configured so that all of the data of the transaction data set is sent to remote server 184 in response to a single command input to optical reader 100 via a user interface of reader 100 (e.g., 3170 ). Further aspects of optical reader 100 operating in a picture taking mode of operation are described with reference to the flow diagrams of FIGS. 14 c , 14 g and 14 h.
- a digital picture taking process 1400 is executed when optical reader 100 is configured to operate in a picture taking mode of operation.
- a picture taking mode of operation may be selected, e.g., by clicking on “image capture” icon 3164 ( FIG. 9 b ) and at step 1104 picture taking process 1400 is executed.
- optical reader 100 at step 1402 receives a trigger signal as may be generated e.g., by depression of a manual trigger an object in the proximity of reader 100 being sensed or receipt of a serial command.
- control circuit 552 captures a plurality of “test” or parameter determination frames of image data.
- the frames of image data captured at step 1403 are not output for visual display; but rather are processed in order to determine operational parameters (exposure setting, gain illumination).
- step 1404 can be avoided and control circuit 552 can instead load operational parameters that were derived during a past image capture operation.
- control circuit 552 obtains a “visual display” image frame of image data.
- a visual display color frame of image data is one that is generated for visual display on a display and may include three color scale values for each pixel position of the plurality of pixel positions of the frame.
- a visual display frame after being obtained is sent to a display for visual display of an image or to a memory location for future display.
- the image data obtained at step 1404 is not transferred to decode circuit 1702 .
- An image captured as part of obtaining at step 1404 may be one that is captured in accordance with the timing diagrams of FIGS. 15 a - 15 e .
- the control signals input into image sensor IC chip 1082 for the capture of a frame of image data may not include illumination control timing pulse e.g., pulse 350 .
- illumination control timing pulse e.g., pulse 350
- an object subject to image capture by optical reader 100 during a picture taking mode will be a long range image (an object will be subject to image capture is one that is one to several feet from image reader).
- Light from light sources 160 may have little affect on an image captured that corresponds to a long range object; thus, optical reader 100 , in one embodiment may not send an illumination control timing pulse at step 1404 .
- optical reader 100 can have a plurality of operator selectable configuration settings. Optical reader 100 can be configured so that activation of button 3150 toggles through a sequence of options one of which may be selected by actuation of a key of keyboard 508 .
- Configuration setting 1 is a base line setting wherein there is no differentiation between monochrome read out and color image data read out imaging parameters.
- Configuration setting 2 has been described above. With configuration setting 2, there is no illumination during read out of color image data at step 1404 .
- Configuration setting 3 has also been described above. With configuration setting 3, illumination intensity is increased for read out of color image data. With configuration setting 4, illumination intensity for read out of monochrome image data can be increased.
- illumination pattern 1830 and aiming pattern 1838 can be projected simultaneously during read out of monochrome image data corresponding to a monochrome zone 2500 M of pixels.
- configuration setting 5 exposure time is boosted for read out of color image data and with configuration setting 6, gain is boosted for read out of color image data.
- Configuration setting 3 is highly useful where optical reader 100 includes a long distance flash illumination light source 160 , 160 X or where optical reader 100 is used for picture taking at close range.
- optical reader 100 may carry out a variety of alternative processes.
- optical reader 100 may obtain a visual display color frame of image data utilizing image data read out from color sensitive pixels 250 C only.
- control circuit 552 obtains a visual display color frame of image data utilizing image data derived by reading out of image data from both monochrome pixels and color sensitive pixels of image sensor array 182 .
- control circuit 552 at step 1405 captures a windowed frame of image data by selectively addressing color pixels 250 C and by selectively reading out image data from color pixels 250 C of image sensor array 182 A, 182 B.
- image sensor array 182 A, 182 B may include a separate reset control grid for resetting monochrome pixels 250 M independent of color sensitive pixels 250 C.
- monochrome pixels 250 M may be reset with use of a reset control timing pulse 370 , 370 ′, 370 ′′, 370 ′′′ ( FIGS. 15 a - 15 d ).
- Coordinating a reset control timing pulse 370 , 370 ′, 370 ′′, 370 ′′′ for resetting monochrome pixels with an exposure control timing pulse 354 , 354 ′, 354 ′′, 354 ′′′ for controlling exposure of color sensitive pixels 250 , 250 C reduces cross talk resulting from light rays entering monochrome pixels 250 M, i.e., by eliminating electron diffusion cross talk and by reducing cross talk attributable to light rays angularly penetrating through monochrome 250 M.
- optical reader 100 automatically transfers the color filter array image data frame captured at step 1405 to a demosaicing circuit 1706 ( FIG. 1 e ).
- demosaicing circuit 1706 outputs a visual display color frame of image data.
- display 504 is a color display configured to receive red, green and blue (RGB) signals for each pixel of display 504
- demosaicing circuit 1706 at step 1406 may generate RGB color scale values for each pixel of display 504 so that a frame output by demosaicing circuit 1706 is compatible with display 504 .
- the color scale values may comprise e.g., 8 bits, 10 bits, or 12 bits of data.
- optical reader 100 receives a visual display color frame of image data from demosaicing circuit 1706 .
- optical reader 100 may read out an RGB Bayer pattern frame of image data as shown in FIG. 16 b .
- optical reader 100 at step 1405 captures a 320 ⁇ 256 Bayer pattern of pixels.
- Demosaicing circuit 170 processes the Bayer pattern frame 1502 as shown in FIG. 16 b to output a visual display color frame of image data including a 320 ⁇ 256 color image, wherein each pixel of the frame includes a red color scale value, a green color scale value, and a blue color scale value.
- demosaicing circuit 1706 for each pixel of the Bayer pattern color filter array image data frame 5204 , interpolates red, green, and blue values. Referring to frame 5204 shown in FIG. 16 b , optical reader 100 determines a red value for red pixel position P 32 simply by reading the color scale value of pixel position P 32 .
- Optical reader 100 determines a green value for red pixel P 32 by averaging the values of green pixel positions P 31 , P 22 , P 42 and P 33 .
- Optical reader 100 may interpolate a blue value for red pixel position P 32 by averaging the values of blue pixel positions P 14 P 41 , P 23 , P 43 . It will be seen that red, green, and blue values can be determined for each pixel position interpolating the pixel values as necessary. With increased processing speeds, it is possible to utilize dozens or more surrounding pixel values for interpolation of a red, green, or blue pixel for each pixel position.
- the accuracy with which color scale values for each pixel position may be interpolated can be enhanced by utilizing monochrome pixel values in the color scale value interpolation process.
- color scale values at green pixel positions P 31 , P 22 , P 42 , P 33 may be averaged for interpolating a green color scale value at pixel position P 32 .
- monochrome pixel values at positions P 33 , P 22 , P 42 , P 33 may be utilized for enhancing the interpolation of a green pixel value at position P 32 .
- the monochrome pixel values at positions P 33 , P 22 , P 42 , P 33 may be interpolated from monochrome pixel values by one of the monochrome pixel interpolation methods described herein. Then, the color scale value at each pixel position, P 32 , P 22 , P 42 , P 33 may be offset by a value, Delta, equal to the difference between the interpolated monochrome pixel values at the position being interpolated and the monochrome pixel value at the position contributing to the interpolation calculation. Thus, a green color scale value at position P 32 may be calculated according to Eq. A.
- G ⁇ ( P 32 ) [ G ⁇ ( P 31 ) + Delta 31 ] + [ G ⁇ ( P 22 ) + Delta 22 ] + ⁇ [ G ⁇ ( P 42 ) + Delta 42 ] + [ G ⁇ ( P 33 ) + Delta 33 ] 4 ( Eq . ⁇ A )
- Delta 31 M(P 32 ) ⁇ M(P 31 )
- Delta 22 M(P 32 ) ⁇ M(P 22 )
- Delta 42 M(P 32 ) ⁇ M(P 42 )
- Delta 33 M(P 32 ⁇ M(P 33 ).
- a blue color scale value at position P 42 may be interpolated according to the formula of Equation B.
- step 1 there will normally be two neighborhood color or “C-set” pixels where blue or red values are interpolated at a green pixel position, and in other cases four neighborhood color pixels.
- optical reader 100 executing steps 1405 and 1406 is explained with reference to FIG. 16 c .
- optical reader 100 at step 1405 reads out a color filter array frame 5206 as shown in FIG. 16 c .
- Color filter array image data frame 5206 includes a 320 ⁇ 256 pattern of Cy-Mg pixel values.
- Demosaicing circuit 1706 may process image data frame 5206 into a visual display frame such as a visual display color frame of image data where each pixel position of frame 5206 is represented by a combination of red, green and blue values.
- optical reader 100 may first calculate white, cyan and magenta values for each pixel position of frame 5206 . Where an original pixel position such as pixel position P 53 ( FIG. 16 c ) is a cyan pixel, the cyan value is determined by directly reading the pixel value of the cyan pixel.
- a magenta value for cyan pixel at position P 53 is calculated by interpolation utilizing the magenta values of surrounding pixel positions of magenta pixels such as positions P 52 , P 43 , P 63 , P 54 ( FIG. 16 c ).
- a white value for cyan pixel at position P 35 is calculated by interpolation using pixel values from monochrome pixel positions that surround cyan pixel P 53 .
- a supplemental frame including monochrome pixel values may be captured, e.g., successively before or after frame 5206 is captured for purposes of interpolating a white value for each color pixel of the color filter array windowed frame 5206 .
- the color filter array frame 5206 captured at step 1405 may include monochrome pixel image data for purposes of interpolating a white value for each color pixel value.
- white, cyan and magenta values are calculated for each pixel of frame 5206 , the white, cyan, and magenta values are readily converted into red, green, and blue values.
- display 504 can be configured to be responsive to white, cyan and magenta signals for each pixel of display 504 .
- a set of transform equations for transforming a set of white, cyan and magenta values for a given pixel of a frame into a set of red, green and blue values for that pixel is given as follows.
- an original color filter array frame is processed into a visual display color frame of image data of reduced spatial resolution (a reduced spatial resolution 320 ⁇ 256 visual display color frame of image data may be produced using a hybrid monochrome and color image sensor array having a 1280 ⁇ 1024 pixel resolution).
- a process for producing a high spatial resolution visual display color image is described. In the process described relative to the flow diagram of FIG.
- optical reader 100 utilizes image data from both monochrome pixels 250 M and color pixels 250 C from a hybrid monochrome and color image sensor array such as image sensor array 182 A or image sensor array 182 B in the generation of a visual display color image having spatial resolution equal to or on the order of the overall pixel resolution of the image sensor array.
- control circuit 552 captures a color filter array image data frame by selectively addressing color pixels 250 C of an image sensor array and selectively reading out image data from color sensitive pixels 250 M.
- the frame of image data captured at step 1408 is a windowed frame of image data.
- the monochrome pixels of image sensor array 182 A, 182 B may be reset using reset control timing pulse 370 , 370 ′, 370 ′′, 370 ′′′ at the time that exposure control timing pulse 354 , 354 ′, 354 ′′, 354 ′′′ is applied to expose color pixels for capture of a color filter pattern image frame at step 1408 .
- control circuit 552 captures a monochrome frame of image data by selectively addressing monochrome pixels 250 M of array 182 A, 182 B and selectively reading out image data from monochrome 280 M pixels.
- the frame of image data captured at step 1409 is a windowed frame of image data.
- the color pixels of image sensor array 182 may be reset using reset control timing pulse 370 , 370 ′, 370 ′′, 370 ′′′ at the time that exposure control timing pulse 354 , 354 ′, 354 ′′, 354 ′′′ is applied to expose monochrome pixels for capture of a monochrome, typically gray scale or binarized image frame at step 1409 .
- control circuit 552 transfers both the color filter array frame captured at step 1408 and the monochrome image frame captured at step 1409 to fusion circuit 1708 .
- Fusion circuit 1708 takes as inputs the color filter array image data frame and the monochrome image data frame and processes them into a high resolution visual display color frame of image data.
- fusion circuit 1708 may be physically embodied by the combination of a control circuit provided by a CPU 552 operating in combination with memory 566 that stores an executable program.
- the specific process described with reference to FIG. 14 i is executed utilizing an optical reader 100 including substantially uniform dimensional pixel image sensor array 182 B.
- control circuit 552 generates color filter array image data and monochrome gray scale image data.
- control circuit 552 may execute step 1442 by reading out from image sensor array 182 B a single frame of image data comprising both monochrome image data and color image data. Control circuit 552 may also execute step 1442 by successively capturing a first monochrome frame comprising monochrome image data and then a second color frame comprising color image data. Control circuit 552 at step 1442 may drive monochrome pixels 250 M into reset during an exposure period for read out of color image data from color sensitive pixels 250 C. When generating a frame of monochrome image data at step 1442 , control circuit may interpolate monochrome pixel values for “missing pixel” positions occupied by color sensitive pixels 250 C.
- control circuit 552 generates an RGB image having resolution equal to the color sensitive subset of pixels of image sensor array 182 B.
- each pixel of the image is represented by a red color value, a green color value and a blue color value.
- the RGB image generated at step 1446 may have the same characteristics as the visual display image received by optical reader 100 at step 1407 of the alternative process described in connection with FIG. 14 g .
- a color filter array image captured at step 1442 is a Bayer pattern image
- the RGB image generated at step 1446 is derived by executing a demosaicing routine as described herein.
- the RGB image generated at step 1446 is derived by way of a transformation process as described herein in connection with equations 1, 2 and 3.
- process 1440 which may be executed by fusion circuit 1708 control circuit 552 at step 1450 expands the pixel count of the RGB image generated at step 1446 so that the pixel count of the color image is equal to the pixel count of the monochrome image captured at step 1442 (at step 1442 monochrome pixels from the captured monochrome image may be interpolated as described with FIG. 16 a ).
- control circuit 552 executes step 1450 , the monochrome gray scale image generated at step 1442 and the color image at that stage of the processing have equal numbers of pixels such that each pixel position e.g., pixel of the monochrome image has a corresponding pixel position in the color image.
- control circuit 552 at step 1450 expands each pixel into a 2 ⁇ 2 pixel block.
- control circuit 552 at step 1450 expands each pixel into a 3 ⁇ 3 pixel block.
- control circuit 552 at step 1450 expands each pixel into a 4 ⁇ 4 pixel pixel block.
- control circuit 552 calculates an intensity value I c for each pixel position of the expanded color image.
- Control circuit 552 at step 1454 calculates an intensity value for each pixel position of the expanded color image according to the formula.
- Control circuit 552 at step 1460 then calculates an intensity value delta, D, for each pixel position, (Px, Py) utilizing a monochrome image intensity value I m and an expanded image color intensity value, I c , at each pixel position.
- Control circuit 552 at step 1460 may calculate an intensity value delta for each pixel position of the monochrome and expanded color image according to the formula
- control circuit 552 updates the RGB data set color scale values of the expanded RGB color image using the set of formulas
- control circuit 552 truncates RGB data set color scale values that are greater than 255 (where an 8 bit gray scale is used). After control circuit 552 truncates RGB values greater than 255, control circuit 552 at step 1770 outputs a visual display color frame of image data having a spatial resolution equal to or approximately equal to the overall pixel resolution of image sensor array 182 B.
- the visual display color frame of image data output at step 1770 may have a number of RGB data sets equal to the overall pixel count (e.g., monochrome pixels plus color sensitive pixels) of image sensor array 182 B.
- optical reader 100 receives from fusion circuit 1708 a high resolution visual display color frame of image data.
- the visual display color frame of image data received at step 1411 may include a pixel resolution equal to or on the order of the pixel resolution of image sensor array 182 B.
- Optical reader 100 may be regarded as having received a visual display color frame of image data when fusion circuit 1708 outputs a visual display color frame of image data at step 1470 .
- control circuit 552 When executing process 1440 , control circuit 552 fuses monochrome and color image data to produce a high resolution visual display color frame of image data. When executing the alternative process described with reference to the flow diagram of FIG. 14 j , control circuit 552 fuses monochrome and color image data in such a manner that color reproduction is optimized.
- Optical reader 100 in another aspect may incorporate the structure shown in FIG. 21 .
- the image sensor array 182 B is constructed such that center pixels 2072 are optimized for providing image data yielding increased decoding accuracy while outer pixels 2074 are optimized for providing image data yielding increased color reproduction accuracy.
- control circuit 552 at step 1412 outputs the visual display color frame of image data obtained at step 1404 .
- control circuit 552 may output a visual display color frame of image data to display 504 for visual observation by an operator or to a designated color frame storage memory location of reader 100 such as a designated frame memory storage location of Flash memory 564 or to another frame memory location of system 145 .
- control circuit 552 at step 1410 may also send a visual display color frame of image data to spaced apart device 150 , as shown in FIG. 10 .
- optical reader 100 For sending a frame of image data to a spaced apart device 150 , optical reader 100 , the spaced apart device 150 , and a communication link there between may be configured to transmit data packets in accordance with a protocol of the TCP/IP suite of protocols. Further, optical reader 100 may format the visual display color frame of image data obtained at step 1412 in a suitable image file format (e.g., .BMP, .TIFF, .PDF, .JPG, .GIF) and optical reader 100 may automatically send the visual display color frame of image data at step 1412 utilizing the file transfer protocol (FTP).
- a suitable image file format e.g., .BMP, .TIFF, .PDF, .JPG, .GIF
- FTP file transfer protocol
- Optical reader 100 at output step 1212 may format the visual display color frame of image data in a suitable image file format (e.g., .BMP, .TIFF, .PDF, .JPG, .GIF) when storing the visual display color frame of image data in memory 566 (which can be incorporated in hand held housing 101 ) or when sending the visual display color frame of image data to a spaced apart device 150 for storage.
- Optical reader 100 may also transmit a visual display color frame of image data utilizing a suitable markup language such as .XML. Referring to FIG.
- system 145 may be configured so that when a display equipped spaced apart device 150 receives a visual display color frame of image data from optical reader 100 , the spaced apart device 150 automatically displays that received visual display color frame of image data on a display 1504 associated with that device.
- Optical reader 100 can be configured so that all the steps of process 1400 are carried out automatically in response to receipt of a trigger signal until a stop condition is satisfied.
- a stop condition may be the receipt of a trigger stop signal such as may be generated by the release of trigger 216 .
- optical reader 100 can be configured to carry out indicia decoding or picture taking with a single actuation of a reader control button.
- optical reader 100 can be configured so that actuation of virtual button 3162 both configures reader 100 to decode and simultaneously generates a trigger signal to immediately commence image capture and decoding.
- Optical reader 100 can also be configured so that actuation of virtual icon button 3164 both configures a reader 100 for picture taking and simultaneously generates a trigger signal to immediately commence image capture.
- process 1200 and process 1400 may be carried out in the alternative, process 1200 and process 1400 may also be executed contemporaneously.
- control circuit 552 may obtain a visual display color frame of image data at step 1404 .
- Control circuit 552 may obtain a color frame of image data as a decode frame at step 1204 and then outputs that frame at step 1212 as visual display color frame of image data.
- Control circuit 552 at step 1412 may output a visual display color frame of image data and contemporaneously transfer that frame of image data to decode circuit 1702 .
- reader 100 may be configured so that whenever control circuit 552 obtains a decode frame at step 1204 , control circuit 552 may store that frame for later processing, which processing may include processing for generating a visual display color frame of image data and which processing may be responsive to an operator input command to perform such processing.
- Optical reader 100 may also be configured so that when control circuit 552 obtains a visual display color frame of image data at step 1404 , control circuit may store that frame for further processing, which processing may include transferring that frame to decode circuit 1702 or autodiscrimination circuit 1704 , and which processing may be responsive to an operator input command to perform such processing.
- optical reader 100 includes a pair of imaging modules 1802 D and 1802 E.
- Imaging module 1802 D is a color imaging module having color image sensor array 182 D.
- Color image sensor array 182 D includes a Bayer pattern color filter with one of red, green or blue wavelength selective filter disposed on each pixel.
- Imaging module 1802 E as shown in FIG. 17 e is a monochrome imaging module having a one-dimensional solid state image sensor array 182 E.
- 17 a , 17 e , 17 f , and 17 g includes an M ⁇ 1 (one row) array of monochrome (without color filter) pixels.
- One-dimensional image sensor array 182 E may also include and M ⁇ N array of pixels, where M>>N, e.g., an M ⁇ 2 (2 rows) of pixels.
- optical reader 100 of FIG. 17 a has many of the same components as shown in optical reader 100 of FIG. 1 a .
- optical reader 100 of FIG. 17 a includes a control circuit 552 provided in the example by a CPU, which operates under the control of program data stored in EPROM 562 .
- Control circuit 552 is in communication with a memory unit 566 that includes in addition to EPROM 562 , RAM 560 , and Flash memory 564 .
- Control circuit 552 further receives input control data from various user input devices such as manual trigger 216 , pointer controller 512 , keyboard 508 and touch screen 504 T.
- Control circuit 552 may also output data such as decoded output data and visual display image data to color display 504 .
- control circuit 552 may control either image sensor array 182 E or image sensor array 182 D.
- control circuit 552 sends various image capture initiating control signals to one-dimensional image sensor array 182 E.
- image sensor array 182 E sends analog image signals to signal processing circuit 591 which among various processing functions amplifies the signals and feeds the signals to analog-to-digital converter 592 .
- Analog-to-digital converter 592 converts the signals into digital form and routes the digital image data to FPGA 593 which under the control of control circuit 552 , manages the transfer of the digital information into RAM 560 , where the monochrome image data can be accessed for decoding processing by control circuit 552 .
- control circuit 552 sends appropriate image capture initiation control signals (e.g., exposure, read out) to image sensor chip 1082 .
- FPGA 580 receives digital image data from image sensor IC chip 1082 , 1082 D and under the control of control circuit 552 manages the transfer of color image data into RAM 560 .
- Illumination assembly 104 for each module 1802 D, 1802 E may be controlled during image acquisition as explained with reference to the timing diagrams of FIGS. 15 a - 15 e.
- Optical reader 100 as shown in FIGS. 17 a - 17 g may be operated in accordance with the flow diagram of FIGS. 17 a - 17 g . Namely, by a suitable selection method such as by depressing icon 3162 or icon 3164 ( FIG. 9 b ) one of a decode mode of operation and a color image capture mode of operation can be selected. However, in the dual imaging module embodiment of FIGS. 17 a - 17 g , the imaging module which is utilized for capturing image data depends on which mode (indicia decoding, or picture taking) is selected. If the indicia decode mode is selected at step 1100 ( FIG.
- optical reader 100 proceeds to step 1102 to execute indicia decode process 1200 ( FIG. 14 a ).
- control circuit 552 obtains a decode frame of image data. If the picture taking mode of operation is selected at step 1100 ( FIG. 14 a ) and a trigger signal is received, control circuit 552 proceeds to step 1404 ( FIG. 14 c ) to obtain a visual display color frame of image data.
- reader 100 includes two imaging modules, one color such as module 1802 , 1802 D having color image sensor array 182 , 182 D and one monochrome such as module 1802 , 1802 E having monochrome image sensor 182 , 182 E
- the particular image sensor array 182 to which control circuit 552 sends control signals for initiating image capture depends on whether optical reader 100 is operating in a decode mode of operation or a picture taking mode of operation.
- reader 100 at step 1204 sends image capture initiation control signals to monochrome one-dimensional image sensor array 182 , 182 E to initiate image capture without sending any image capture initiation control signals to color image sensor array 182 , 182 D if the reader 100 is operating in a decode mode operation.
- Reader 100 at step 1404 sends image capture initiation control signals to color image sensor array 182 , 182 D without sending any image capture initiation control signals to monochrome image sensor array 182 , 182 E if reader 100 is operating in a picture taking mode operation.
- optical reader 100 is in an indicia decode mode and receives a trigger signal
- a monochrome frame of image data is sent to RAM 560 for further processing by decode circuit 1702 ( FIG. 10 ).
- optical reader 100 is in a picture taking mode and receives a control signal
- a color image is sent to RAM 560 .
- the color image if a Bayer pattern image is subject to a demosaicing process as described herein for generating a visual display color frame of image data which visual display color frame of image data may be output by control circuit 552 e.g., to display 504 and/or a designated memory address of system 145 (e.g., memory 566 or another memory such as a memory of a spaced apart device 150 ), and/or to a display 1504 of a spaced apart device 150 of system 145 ( FIG. 5 ).
- the type of image capture depends on a selected operating mode.
- an indicia decode mode is selected, a monochrome gray scale image well suited for decode processing is captured.
- a picture taking mode is selected, a color image is captured which is well suited for visual display.
- FIGS. 17 b - 17 g illustrate that the hardware block 598 of reader 100 shown in FIGS. 17 a , 17 f , and 17 g may be replaced with alternative hardware blocks.
- hardware block 398 which in FIG. 17 a includes a CCD one-dimensional solid state image sensor array 182 E and off-board signal processing circuit 591 , analog-to-digital converter 592 and FPGA 593 may be replaced by a hardware block including an CMOS image sensor IC chip 1082 F including a monochrome image sensor array 182 F.
- Image sensor IC chip 1082 , 1082 F is of construction similar to image sensor IC chip 1082 , 1082 A and IC chip 1082 , 1082 D except that image sensor array 182 F of chip 1082 F includes monochrome pixels 250 , 250 M only and is devoid of color sensitive pixels 250 , 250 C.
- FIG. 17 c illustrates that imaging assembly hardware block 598 can be replaced with a laser scanning bar code engine 594 and an associated decode circuit 595 .
- Laser scanning bar code engine 594 and associated decode circuit 595 may be available in a package known as an SE 923 decoded out scan engine available from Symbol Technologies.
- steps 1210 , 1212 , 1214 of decode process 1200 are carried out by decode circuit 595 .
- FIGS. 17 d and 17 e Exemplary imaging modules supporting various types of image sensor IC chips are shown in FIGS. 17 d and 17 e .
- FIG. 17 d shows an exemplary imaging module for supporting image sensor IC chip 182 D.
- Imaging module 1082 D includes the elements shown and described with reference to FIGS. 8 a - 8 d except that imaging module 1082 D includes image sensor IC chip 182 D and further except that certain light sources are optionally deleted.
- Imaging module 1082 E includes the elements shown and described with reference to FIGS. 8 a - 8 e except that imaging module 1082 E includes one-dimensional monochrome image sensor chip 182 E and further except that certain light sources of illumination block 104 are optionally deleted.
- aiming pattern 1838 FIG.
- illumination assembly 104 of an imaging module may include a flash illumination light source, 160 , 160 X ( FIG. 9 a ). It may be particularly useful to incorporate a flash illumination into illumination assembly 104 , where an imaging module 1082 is used primarily for capture of visual display color image.
- FIGS. 17 f and 17 g construction views of dual imaging module readers incorporated in various optical reader housings are shown and described.
- a gun style optical reader 100 is shown having color two-dimensional imaging module 1802 D and one-dimensional monochrome imaging module 1082 E supported therein.
- a portable data terminal (PDT) optical reader 100 is shown having color two-dimensional imaging module 1802 D and one-dimensional monochrome imaging module 1802 E supported therein.
- the dual modules can also be installed in other types of housings such as cell phone housings ( FIG. 9 c ) and personal data assistant housings (PDAs).
- FIGS. 9 c cell phone housings
- PDAs personal data assistant housings
- imaging modules 1802 are supported by struts 597 formed on an interior wall 1802 of housing 101 .
- the modules 1802 in each example are in communication with a main printed circuit board 599 which includes various electrical components including processor IC chip 548 .
- the optical reader 100 of FIGS. 17 a - 17 g is operated in the following manner.
- An operator actuates color image sensor array 182 D to take a color picture of parcel 1260 ( FIGS. 11 and 12 ) carrying a bar code symbol 1266 , 1270 .
- Such actuation may be carried out, e.g., by depressing decode button 3164 and then trigger 216 or button 3164 only.
- An operator then actuates monochrome image sensor array 182 E (or alternatively image sensor array 182 F, or laser scan engine 594 ) to decode bar code symbol 1266 , 1270 .
- Such actuation may be carried out e.g., by depressing button 3162 and then trigger 216 or by depressing button 3162 only.
- control circuit 552 which may be incorporated in hand held housing 101 , may transmit a visual display color frame of image data representing parcel 1260 and decoded out messages corresponding to one or more of symbols 1266 , 1270 to remote server 184 ( FIG. 10 ).
- System 145 can be configured so that such transmission is automatic in response to trigger signals being received, or optical reader 100 can be configured so that associated color picture data and decoded out bar code message data are transmitted in response to receipt of a user-initiated command input into a user-interface of optical reader 100 to transmit associated picture and decoded bar code message data.
- optical reader 100 having the hardware components shown in FIG. 1 a may be modified to include an image sensor array 182 C as shown and described in connection with FIG. 18 a .
- optical reader 100 includes cyan-magenta-yellow (CMY) color filter array 182 C.
- Each pixel 250 of image sensor array 182 C includes a color filter element; namely one of a cyan color filter element, a magenta color filter element or a yellow color filter element.
- Yellow color filter elements have excellent light transmittance (approaching the transmittance of a monochrome pixel). Further, it is seen that in accordance with the CMY color filter pattern shown in FIG. 18 a that approximately 50% of all pixels of image sensor array 182 C are yellow pixels (pixels having a yellow light wavelength sensitive filter element). In the specific example of FIG. 18 a , image sensor array 182 C having cyan, magenta and yellow pixels is devoid of green pixels. However, image sensor arrays are available which have green pixels in addition to cyan, magenta and yellow pixels. Image sensor array 182 C may be incorporated into an optical reader 100 that operates in accordance with the picture taking mode/indicia decode mode flow diagram described in connection with FIG. 14 a .
- optical reader 100 including CMY color image sensor array 182 C obtains a decode frame of image data whereas when optical reader 100 including image sensor array 182 C is driven into a picture taking mode of operation, optical reader 100 obtains a visual display color image frame of image data as described in connection with FIG. 14 c herein.
- an optical reader including a CMY image sensor array 182 C as shown in FIG. 18 a may obtain image data in a manner that depends on which operational mode (indicia code or picture taking) is selected.
- control circuit 552 of optical reader 100 can selectively address yellow color pixels of CMY image sensor array 182 C and selectively read out image data only from yellow colored pixels of image sensor array 182 C.
- control circuit 552 at step 1204 may interpolate missing pixel values corresponding to the pixel positions of magenta and cyan pixels of image sensor array 182 C. After interpolating the missing pixel positions, control circuit 552 at step 1210 may transfer the interpolated decode frame to one of indicia decode circuit 1702 or autodiscrimination circuit 1704 .
- image sensor array 182 C may include separate and independent reset control lines for facilitating the reset of magenta (labeled “Mg”) and cyan (labeled “Cy”) pixels independent from the resetting of yellow pixels (labeled “Y”). Accordingly, when image data at step 1204 is read out selectively from yellow pixels, the magenta and cyan pixels of image sensor array 182 C may be driven into reset to eliminate electron diffusion cross talk and to reduce cross talk attributable to photons entering image sensor array 182 C through magenta and cyan color pixels 250 C.
- an optical reader including image sensor array 182 C may simply read out image data from all of the pixels of the array 182 C and execute a simple demosaicing algorithm to convert a single color value for each pixel of image sensor array 182 C into a visual display color image wherein each pixel of image sensor array 182 C is represented by a data set including three color scale values, e.g., a cyan color scale value, a magenta color scale value and a yellow color scale value.
- Control circuit 552 at step 1404 where the reader includes a CMY image sensor array 182 C may transform the CMY visual display image into an RGB visual display image utilizing a CMY to RGB transformation process as described herein.
- optical reader 100 may be hindered where optical reader 100 is operated to read bar code symbols or other indicia disposed on a substrate having a shiny surface (e.g., metal, glass, laminated, plastic, etc.).
- a substrate having a shiny surface e.g., metal, glass, laminated, plastic, etc.
- Light rays emanating from light sources 160 of reader 100 that are projected on a highly reflective shiny surface of a substrate, s may be substantially entirely reflected directly on to image sensor array 182 .
- “Specular” reflection is said to occur where a substantial percentage of light rays are reflected and directed onto image sensor array 182 .
- Light rays are said to be reflected at a “specular angle” when light rays are reflected from a surface at about the angle of incidence.
- Specular reflection tends to saturate image sensor array 182 to cause decoding failures.
- the optical reader 100 described in connection with FIGS. 19 a - c is configured so that read errors resulting from specular reflection are reduced.
- hardware block 208 shown in FIG. 1 a as including a hybrid monochrome in color image sensor array 182 A can be replaced with hardware block 208 as shown in FIG. 19 a including a hybrid monochrome and polarizer filter image sensor array 182 G.
- Image sensor array 182 G includes a first subset of monochrome pixels 250 M and a second subset of light polarizing pixels 250 P.
- Light polarizing pixels 250 P of image sensor array 182 G include light polarizing filter elements 261 (alternatively termed “light polarizing filters,” or simply “light polarizers”) typically formed at each polarizing pixel 250 P in the position of filter 260 as shown in the color pixel views of FIGS. 3 c and 6 c .
- Light polarizing filter elements 261 of image sensor array 182 G, 182 H can be deposited onto the major body of light polarizing pixels 250 P by way of a depositing process.
- Light polarizing filter elements 261 of image sensor array 182 G can be constructed to attenuate polarized light rays generated from an appropriately polarized light source and reflected at a specular angle. Accordingly, polarized light rays incident on the image sensor array on the polarizing pixels 250 P are attenuated significantly; thus, reducing the contribution of specularly reflected light rays to generated image signals from the polarizing pixels 250 P.
- optical reader 100 including image sensor array 182 G may be configured to selectively address light polarizing pixels 250 P and selectively read out image data from light polarizing pixels 250 P to generate image data for subjecting to decoding which is likely to result in successful reading of bar codes or other indicia notwithstanding the image data being obtained during specular reflections read conditions.
- FIG. 19 b a perspective view of light polarizing image sensor array 182 G is shown with an exploded view showing a pattern which may be repeated throughout the array.
- light polarizing pixels 250 P may also be distributed throughout image sensor array 182 G in a uniform or substantially uniform distribution pattern other than the pattern shown in FIG. 19 b .
- optical reader 100 may be operated in a mode in which optical reader 100 captures image data by selectively addressing polarizing pixels 250 P and selectively reading out image data from light polarizing pixels 250 P only.
- Optical reader 100 may be configured to have a reduced specular reflection read error decode mode.
- Optical reader 100 can be configured so that when button 3156 is actuated, optical reader 100 receives a trigger signal to obtain image data that is likely to result in successful reading notwithstanding specular reflection reading conditions.
- optical reader 100 at step 1902 may receive a trigger signal to commence operation in a reduced specular reflection read error decode mode.
- the trigger signal may be received pursuant to a manual control by an operator such as an actuation of control button 3156 .
- Control circuit 552 may also be configured to receive the trigger signal at step 1902 when control circuit 552 automatically senses a predetermined condition such as a saturation condition.
- Control circuit 552 at step 1902 may determine that a saturation condition is present by analysis of image data at step 1204 ( FIG. 14 b ) during normal decoding operations so that when a saturation condition is detected, optical reader 100 automatically commences operation in a reduced specular reflection read error decode mode.
- control circuit 552 may determine that a saturation condition is present when an average white value of monochrome image data is below a predetermined level.
- optical reader 100 obtains a specular reflection read condition decode frame of image data.
- Control circuit 552 obtains a specular reflection condition decode frame of image data at step 1902 by selectively addressing light polarizing pixels 250 P of image sensor array 182 G and selectively reading out image data from light polarizing pixels 250 P only.
- image sensor array 182 G may include separate reset control lines for resetting monochrome pixels 250 M separately and independently of light polarizing pixels 250 P.
- Image sensor array 182 G may have separate sets of reset control lines as described in connection with image sensor array 182 G, particularly in connection with FIG. 7 a.
- control circuit 552 when control circuit 552 selectively addresses light polarizing pixels 250 P for read out of image data from light polarizing pixels 250 P, control circuit 552 drives monochrome pixels 250 M into reset. Resetting of monochrome pixels 250 M is synchronized with the exposure period for exposing light polarizing pixels 250 P as described herein. Driving monochrome pixels 250 M into reset while light polarizing pixels 250 P are exposed eliminates electron diffusion cross talk and reduces cross talk resulting from photon penetration to image sensor array 182 G.
- control circuit 552 may interpolate pixel values at pixel positions corresponding to missing pixel positions.
- control circuit 552 transfers the specular reflection condition decode frame of image data obtained at step 1904 to indicia decode circuits 1702 or autodiscrimination circuit 1704 as are described in connection with FIG. 1 e.
- control circuit 552 receives decoded output data output by decode circuit 1702 or signature autodiscrimination circuit 1704 .
- control circuit 552 outputs decoded out data, e.g., by transferring decoded out data to an on reader display 504 or to a spaced apart display 1504 or else stores decoded data in appropriate memory address location of system 145 ( FIG. 10 ).
- control circuit 552 selectively reads out monochrome pixel image data from monochrome pixels 250 M and selectively reads out image data from light polarizing pixels 250 P.
- An optical reader including hybrid monochrome and light polarizing image sensor array 182 G may also be operated without selectively reading out image data from image sensor array 182 G.
- An optical reader incorporating hybrid monochrome and light polarizing image sensor array 182 G can be operated to decode decodable indicia and to take pictures in accordance with the process described with reference to the flow diagrams of FIGS. 14 a , 14 b , and 14 c .
- obtaining a decode frame of image data step 1204 , FIG.
- control circuit 552 may read out image data from all pixels of hybrid monochrome and light polarizing image sensor array 182 G including image data from all monochrome pixels 250 M and all light polarizing pixels 250 P in a single frame capture step.
- the full frame monochrome and light polarizer pixel image data can also be captured with two frame capture steps.
- control circuit 552 may transfer to decode circuit 1702 or autodiscrimination circuit 1704 the full frame of monochrome and polarized pixel image data obtained at step 1204 .
- control circuit 552 may, after step 1210 , transfer a subset of full frame of image data originally transferred at step 1210 . Namely, after step 1210 , if decoding or autodiscrimination fails, control circuit 552 may transfer to decode circuit 1702 , or autodiscrimination circuit 1704 a reduced resolution image extracted from a full frame image by selectively extracting monochrome image data from the full frame of image data.
- the reduced resolution frame of image data includes only image data corresponding to light polarizing pixels 250 P of image sensor array 182 G.
- the failure of decode circuit 1702 to decode or autodiscrimination circuit to recognize may be regarded as a determination by control circuit 552 that a saturation condition is present.
- FIGS. 20 a and 20 b show an image sensor array 182 H including a first subset of monochrome pixels 250 M, a second subset of color sensitive pixels 250 C and a third subset of light polarizing pixels 250 P.
- Image sensor array 182 H may include three separate sets of reset control lines to enable separate and independent of resetting of monochrome pixels 250 M, of color sensitive pixels 250 C and of light polarizing pixels 250 P.
- Image sensor array 182 H may be incorporated in hand held optical reader 100 and may be substituted for hardware block 208 as shown in FIG. 1 a .
- Optical reader 100 incorporating image sensor array 182 H may have operating modes in which optical reader separately addresses monochrome pixels 250 M for read out of image data from monochrome pixels 250 M only.
- Optical reader 100 including image sensor array 182 H may also have an operating mode in which optical reader 100 selectively addresses color sensitive pixels 250 C and selectively reads out image data from color sensitive 250 C.
- Optical reader 100 may also have an operating mode in which optical reader 100 selectively addresses light polarizing pixels 250 P and selectively reads out image data from light polarizing pixels 250 P.
- Optical reader 100 may obtain a full frame of image data including monochrome, color and light polarizing pixels image data (obtained with one, two, or three frame capture steps) and then utilize the image data on an as needed basis. For example, if a decode attempt utilizing the full frame image data fails, optical reader 100 may selectively extract light polarizing pixel image data from the full frame image data and transfer the extracted image data to decode circuit 1702 .
- optical reader 100 including image sensor array 182 H selectively reads out image data from monochrome pixels 250 M in obtaining a decode frame of image data for transferring to a decode circuit 1702 under normal read conditions.
- Optical reader 100 selectively reads out image data from color sensitive pixels 250 C when obtaining image data for use when obtaining a visual display color frame of image data.
- Optical reader 100 selectively reads out image data from light polarizing pixels 250 P, or selectively extracts image data corresponding to pixels 250 P from a frame of image data when optical reader 100 senses that a specular reflection is present or when an operator pursuant to operator control drives optical reader 100 into a reduced specular reflection read error decode mode of operation.
- An optical reader 100 including image sensor array 182 H may operate in accordance with the picture taking and decode mode flow diagram as described in connection with FIG. 14 a and may execute the reduced specular reflection read error decode mode decoding process described in connection with FIG. 19 c.
- optical reader 100 may incorporate emit optics light polarizers (which may alternatively be termed “light polarizing filter elements” or “light polarizing filters”).
- a reader imaging module e.g., module 1802 A can include an optical plate 1962 as shown in FIG. 8 f which may be disposed forwardly of circuit board 1806 as shown in FIG. 8 a .
- Optical plate 1962 can incorporate light polarizers 1963 which polarize light from light sources 160 S, 160 T, that can be selectively energized when capturing images utilizing polarizing image sensor array 182 G, 182 H.
- Light polarizers 1963 can be cross-polarized relative to the polarizing filter elements 261 of image sensor array 182 G, 182 H.
- Optical plate 1962 can include other such elements as optical diffusers (not shown) for diffusing light rays emitted by light sources 160 C- 160 T.
- Indicia decode circuit 1702 when receiving image data transferred by control circuit 552 may search the image data for markers, such as a quiet zone, indicative of the presence of a dataform, such as a one or two-dimensional bar code. If a potential decodable indicia (dataform) is located, the decode circuit 1702 applies one or more indicia decoding algorithms to the image data. If the decode attempt is successful, the optical reader outputs decoded dataform data. All of the circuits (modules) described with reference to FIG. 22 a can be incorporated in housing 101 . Further, all of the circuits of FIG. 22 a can be embodied by the combination of control circuit 552 and memory 566 .
- Optical reader 100 may also include an autodiscriminating circuit 1704 .
- autodiscriminating circuit 1704 may incorporate a decode circuit 1702 and an image processing and analysis circuit 21208 , that are in communication with one another.
- the image processing and analysis circuit 21208 comprises a feature extraction circuit 21212 , a generalized classifier circuit 21216 , a signature data processing circuit 21218 , an OCR decode circuit 21222 , and a graphics analysis circuit 21224 that are in communication with each other.
- the feature extraction circuit 21212 comprises a binarizer circuit 21226 , a line thinning circuit 21228 , and a convolution circuit 21230 that are in communication with each other.
- FIG. 22 b shows a process 21300 for employing one embodiment of the invention utilizing the autodiscrimination circuit shown in FIG. 22 a .
- the process 21300 comprises an image reader recording an actuation event (step 21302 ), such as a receipt of a trigger signal, and in response at step 21304 , collecting (obtaining) image data from a target with the optical reader 100 .
- the collecting of image data step may be in accordance with step 1204 ( FIG. 14 b ).
- the image data is transferred (step 21308 ) to the decode circuit 1702 .
- the dataform decode circuit searches (step 21310 ) the image data for markers, such as a quiet zone, indicative of the presence of a dataform, such as a one or two-dimensional bar code.
- the decode circuit 1702 applies (step 21314 ) one or more dataform decoding algorithms to the ensuing image data. If the decode attempt is successful, the optical reader 100 outputs (step 21318 ) decoded dataform data and signals (step 21322 ) a successful read with an alert, such as a beep tone.
- the image data is transferred (step 21326 ) to the image processing and analysis circuit 21208 .
- the image data is processed in parallel with the attempt to decode the dataform data.
- the process that completes first i.e., dataform decode attempt or the image processing
- outputs its data e.g., a decoded bar code or a captured signature
- the image data is processed in response to the decoding of the dataform.
- a bar code encodes item information such as shipping label number and information indicating that a signature should be captured.
- the image data is processed by the feature extraction circuit 21212 .
- the feature extraction circuit generates numeric outputs that are indicative of the texture of the image data.
- the texture of the image data refers to the characteristics of the type of data contained in the image data.
- Common types of texture include one or two-dimensional bar code texture, signature texture, graphics texture, typed text texture, handwritten text texture, drawing or image texture, photograph texture, and the like.
- sub-categories of texture are sometimes capable of being identified.
- the image data is processed (step 21328 ) by the binarizer circuit 21226 .
- the binarizer circuit 21226 binarizes the grey level image into a binary image according to the local thresholding and target image size normalization.
- the image data is processed (step 21332 ) by the line thinning circuit 21228 to reduce multi-pixel thick line segments into single pixel thick lines.
- the image data is processed (step 21336 ) by the convolution circuit 21230 .
- the convolution circuit 21230 convolves the processed image data with one or more detector maps designed according to the invention to identify various textural features in the image data.
- the convolution circuit 21230 generates a pair of numbers, the mean and variance (or standard deviation), for each convolved detector map.
- FIG. 22 c shows a set of 12 2 ⁇ 3 binary curvelet detector maps 21250 used to detect curved elements present in image data.
- the mean value and the variance generated provide an indication of the presence or density of elements in the binarized line thinned image data having similar shapes to the curvelet detector maps 21250 .
- the 12 curvelet detector maps 21250 generate a total of 24 numbers. According to one embodiment, these 24 numbers are representative of the curved or signature texture of the processed image data.
- Further processing of the image data includes the outputs from the feature extraction circuit 21212 being fed (step 21340 ) into the generalized classified circuit 21216 .
- the generalized classifier circuit 21216 uses the numbers generated by the feature extraction circuit as inputs to a neural network, a mean square error classifier or the like. These tools are used to classify the image data into general categories.
- different neural network configurations are contemplated in accordance with the invention to achieve different operational optimizations and characteristics.
- the input layer has 24 nodes for the 12 pairs of mean and variance outputs generated by a convolution circuit 21230 employing the 12 curvelet detector maps 21250 .
- a convolution circuit 21230 employing the 12 curvelet detector maps 21250 .
- the 20 curvelet detector maps 21260 shown in FIG. 22 d are used by the convolution circuit 21230 .
- the 20 curvelet detector maps 21260 include the original 12 curvelet detector maps 21250 of FIG. 22 c .
- the additional 8 pixel maps 21260 are used to provide orientation information regarding the signature.
- the input layer has 40 nodes for the 20 pairs of mean and variance outputs generated by a convolution circuit 21230 employing the 20 curvelet detector maps 21260 .
- the neural network of this embodiment there are two hidden layers of 40 nodes and 20 nodes respectively, one output node to report the positive or negative existence of a signature, and 8 output nodes to report the degree of orientation of the signature.
- the generalized classifier circuit 21216 is capable of classifying data into an expanded collection of categories. For example, in some embodiments the generalized classifier circuit 21216 specifies whether the image data contains various data types such as a signature; a dataform; handwritten text; typed text; machine readable text; OCR data; graphics; pictures; images; forms such as shipping manifest, bill of lading, ID cards, and the like; fingerprints, biometrics such as fingerprints, facial images, retinal scans and the like, and/or other types of identifiers. In further additional embodiments, the generalized classifier circuit 21216 specifies whether the image data includes various combinations of these data types.
- the general classifier circuit 21216 specifies whether the image data contains a specified type of data or not.
- the image processing and analysis circuit 21208 is contained within an identification circuit that outputs an affirmative or negative response depending on the presence or absence of the specified data type, such as a signature or a biometric in the image data.
- image data is transferred (step 21344 ) to the signature data processing circuit 21218 .
- the signature data processing circuit 21218 is used to detect the boundaries of the signature in the image data.
- the signature boundary is detected using a histogram analysis. As shown in FIG. 22 e , a histogram analysis consists of a series of one-dimensional slices along horizontal and vertical directions defined relative to the orientation of the signature. In one embodiment, the value for each one-dimensional slice corresponds to the number of black (i.e., zero valued) pixels along that pixel slice.
- some specified region of the full frame of image data such as a central region is captured for signature analysis.
- the histogram analysis provides a two-dimensional plot of the density of data element pixels in the image data. The boundary of the signature is determined with respect to a minimum density that must be achieved for a certain number of sequential slices.
- the histogram analysis searches inwardly along both horizontal and vertical directions until the pixel density rises above a predefined cutoff threshold. So that the signature data is not inadvertently cropped, it is common to use low cutoff threshold values.
- the signature data processing circuit 21218 crops the image data and extracts the signature image data.
- the cropping is performed by an image modification circuit that generates modified image data in which a portion of the image data not including the signature has been deleted.
- various compression techniques are employed to reduce the memory requirements for the signature image data.
- One such technique includes the encoding of the signature image data by run length encoding. According to this technique, the length of each run of similar binarized values (i.e., the length of each run of 1 or 0) for each scan line is recorded as a means of reconstructing a bit map.
- Another encoding technique treats the signature image data as a data structure where the elements of the data structure consist of vectors.
- the signature is broken down into a collection of vectors.
- the position of each vector in combination with the length and orientation of each vector is used to reconstruct the original signature.
- the encoding process generates a new vector whenever the curvature for a continuous pixel run exceeds a specified value.
- a further compression technique employs B-Spline curve fitting. This technique has the capacity to robustly accommodate curvature and scaling issues.
- the signature image data or a compressed or encoded version of the signature image data is stored locally on a dedicated memory device.
- the local memory device can be a detachable memory device such as a CompactFlash memory card or the like described in more detail below.
- the signature image data is stored in a volatile or non-volatile portion of general purpose memory and downloaded at a future time.
- the signature image data can be transmitted via wired or wireless means either at the time of capture or at a later point, such as when a data collection session has been completed.
- the signature data processing circuit 21218 does not perform a histogram analysis but simply stores in memory the entire image or a compressed version once the presence of a signature has been determined.
- the initial image analysis is performed on a lower resolution image. Once the presence of a signature is determined in this embodiment, a higher resolution image is taken. In one embodiment, a signature extraction histogram analysis is performed on this image.
- the image is stored in memory in either compressed or original format. In some embodiments, the image data is combined with other data to form a record for a particular item such as a package or shipping envelope.
- some of the additional data that can be collected by the optical reader 100 and stored with or separate from the signature data includes but is not limited to dataform data, handwritten text data, typed text data, graphics data, image or picture data, and the like.
- the image processing and analysis circuit 21208 can be designed to perform specialized tasks for different data types. For example, if the generalized classifier circuit 21216 determines that the image data contains typed or machine readable text, the image data can be collected, possibly histogram analyzed, and stored or alternatively, the image data can be transferred to the OCR decoding circuit 21222 . Similarly, if the generalized classifier circuit 21216 determines that the image data includes a graphic element, the image data can be transferred to the graphics analysis circuit 21224 for processing.
- the graphics analysis circuit 21224 is configured to recognize and decode predefined graphics. In one such embodiment, the graphics analysis can include determining which, if any, boxes have been selected in the billing and shipping instructions on a shipping label.
- the graphics analysis can include locating and decoding the typed or handwritten text contained in the zip code box on a shipping label.
- the optical reader 100 can be configured to automatically attempt decode operations in addition to the dataform decode, such as OCR decoding or graphics decoding, prior to the activation of the feature extraction circuit 21212 .
- the image processing and analysis circuit 21208 segments the image data into regions and performs a feature extraction and general classification analysis on each region.
- the standard rectangular image data window is divided into four equal sized sub-rectangles.
- the segmentation consists of overlapping regions so that the total area of the segmented regions is larger than that of the complete field of the image data.
- FIG. 22 g there are seven shown overlapping regions where each identifying numeral is shown in the center of its region.
- the segmentation consists of sample regions (shown as cross-hatched) within the complete field of the image data.
- the sampled regions can be based on a preloaded user template that, for example, identifies regions of interest such as a signature region and/or a bar code region, in for example, a shipping label.
- the segmentation process is used to identify the location of a signature in image data the might include additional elements such as dataforms including bar code dataforms, text, graphics, images and the like.
- the generalized classifier circuit 21216 classifies the contents of each region of the segmented image data. The region containing the signature is then extracted by the signature data processing circuit 21218 . In one embodiment if multiple regions are indicated as containing signature data, the signature data processing circuit 21218 analyzes the arrangement of these regions to identify the region most likely to contain the image data. In a further embodiment when multiple regions are indicated as containing signature data, the image processing and analysis circuit 21208 establishes a feedback loop where additional segmented regions are generated and analyzed until a single segmented region containing signature data is located.
- FIGS. 13 a - 13 e Various applications which may be carried out by any of the optical readers 100 that have been described herein have been described with reference to FIGS. 10 , 11 , 12 a and 12 b . Another application which can be carried out with any optical reader 100 described herein is described with reference to FIGS. 13 a - 13 e .
- a motor vehicle 1282 is shown which may be a delivery vehicle or a passenger vehicle.
- Vehicle 1282 has a license plate 1314 , a vehicle identification number (VIN) sticker 1306 , typically located on the driver's side door jam.
- the VIN sticker 1306 carries a printed VIN number 1308 and a bar code symbol 1310 .
- a VIN number is an alphanumeric unique vehicle identification number assigned at the time of manufacture of the vehicle.
- Vehicle 1282 may further include a VIN plate 1314 ( FIG. 13 c ) carrying the characters of the VIN number etched on a metal plate and located under the vehicle windshield 1351 , and a vehicle registration sticker 1320 .
- Vehicle 1282 has a plurality of machine readable vehicle identifiers.
- the characters of license plate 1284 can be OCR decoded by optical reader.
- VIN sticker 1308 has a VIN bar code 1310 and registration sticker 1320 which may include a plurality of bar code symbols 1322 , 1324 encoding the vehicle registration number and possibly redundantly encoding the VIN number of vehicle 1282 .
- the characters etched on VIN plate 1314 can also be subject to OCR decoding by optical reader 100 .
- VIN characters of VIN sticker 1306 can be subject to OCR decoding by optical reader 100 . It may be advantageous to utilize an optical reader including light polarizing pixels 250 P having light polarizing filter elements 261 when reading VIN plate 1314 given that specular reflection read conditions are more prevalent when decoding indicia encoded by etching on metal surface.
- LAN 170 is a LAN at an automobile insurance claim center
- LAN 185 is a distant data archiving center operated by the automobile insurance provider
- LAN 2170 is a LAN apart from LAN 170 and LAN 185 and may be located, e.g., at a claim center of the insurance provider other than the claim center in which LAN 170 is located.
- Optical reader 100 may be configured so that when an operator actuates a designated user interface control button such as button 3158 ( FIG. 9 b ) an auto insurance application form 1362 is displayed on display 504 which aids an operator of optical reader 100 in entering data into reader 100 .
- Form 1362 first prompts an operator to read several machine readable identifiers of vehicle 1282 .
- Form 1362 prompts an operator to read VIN bar code symbol 1310 , then characters of VIN plate 1314 , then the first registration sticker bar code symbol 1310 , then the second registration sticker bar code symbol 1324 , then the character of the license plate 1284 .
- the text corresponding to each identifier may be highlighted when data corresponding to the identifier is read.
- optical reader 100 When data corresponding to identifier decode section 1363 of form 1362 is being entered, optical reader 100 is in a decode mode of operation such that actuation of trigger 216 causes optical reader 100 to obtain a decode frame at step 1204 and transfer the decode frame to decode circuit 1702 .
- the decode frame may contain monochrome image data read from a hybrid monochrome image sensor array 182 , 182 A.
- the decode frame at step 1204 is obtained by actuation of the imaging assembly within block 598 ( FIG. 17 a ).
- an operator toggles to line 1365 and clicks an appropriate key of keyboard 508 to indicate that identifier decoding is complete.
- Form 1362 then prompts an operator to take pictures of vehicle 1282 for purposes of making a record of the damage to vehicle 1282 .
- the inventor discovered that the incorporation of color filter elements into an image sensor array 182 of optical reader 100 facilitates the obtaining of visual display frames of image data that accurately record damage to a vehicle. With visual display color frames of image data corresponding to vehicle 1282 being stored and/or displayed for visual display, damage to vehicle 1282 can readily be assessed by visual inspection of the visual display frames when displayed on a display 504 , 1504 .
- Section 1364 of display form 1362 prompts an operator to take several color pictures of vehicle.
- optical reader 100 is in a picture taking mode such that actuation of trigger 216 causes a visual display frame of image data to be obtained at step 1404 ( FIG. 14 c ).
- the visual display frame of image data may be output to e.g., a storage device and/or a display device.
- an operator may use optical reader 100 to take several color pictures of damaged area 1370 of vehicle 1282 .
- control circuit 552 may selectively read out color image data from color sensitive pixels 250 C as described herein and possibly utilize monochrome image data for enhancement of the information content of the color image data.
- control circuit 552 at step 1404 may actuate color image sensor array 182 D for execution of obtain step 1404 .
- control circuit 552 When an operator inputs a confirmation that all necessary pictures of vehicle 1282 have been taken by toggling to line 1367 and clicking an appropriate key of keyboard 508 , control circuit 552 , which may be incorporated in hand held housing 101 , may format obtained visual display color frames of image data in one or more suitably image file formats, (e.g., .BMP, .TIFF, .PDF, .JPG, .GIF) assemble all the collected decoded vehicle identifier data and all of the visual display color frames of image data corresponding to vehicle 1282 into a transaction data set, and send the transaction data set to distant remote server 184 . Control circuit 552 may date/time stamp the transaction data set on sending.
- suitably image file formats e.g., .BMP, .TIFF, .PDF, .JPG, .GIF
- the File Transfer Protocol may be utilized to send the transaction data set or another suitable file transferring protocol configured to carry associated decoded vehicle identifier data (such as decoded VIN bar code data and decode vehicle registration bar code data) and color image data.
- Server 184 may store the received transaction data set into a database as indicated by database 187 including similar information from other vehicles at other claim centers.
- Server 184 may be configured to create viewable web pages summarizing the transaction set data (e.g., the date/time stamped combined VIN, registration number, license plate number and record-of damage visual display color frames of image data). These web pages may be viewed using any PC in communication with IP network, e.g., PC 172 and PC 2172 .
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Electromagnetism (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Toxicology (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Color Television Image Signal Generators (AREA)
- Solid State Image Pick-Up Elements (AREA)
- Image Input (AREA)
- Transforming Light Signals Into Electric Signals (AREA)
Abstract
There is provided in one embodiment an optical reader having an image sensor array. In one embodiment, the image sensor array can include a first subset of pixels and a second subset of pixels. The first subset of pixels can be devoid of light polarizing filter elements, and the second subset of pixels can be light polarizing pixels including light polarizing filter elements. An optical reader can be configured to selectively read out image data from an image sensor array's light polarizing pixels.
Description
- This application is a divisional of U.S. patent application Ser. No. 11/445,930 filed Jun. 2, 2006 entitled, “Optical Reader Having Reduced Specular Reflection Read Failures” which claims priority under 35 U.S.C. §119(e) to U.S. Provisional Patent Application No. 60,687,606, filed Jun. 3, 2005 and titled “Digital Picture Taking Optical Reader Having Hybrid Monochrome And Color Image Sensor Array,” (now expired) and to U.S. Provisional Patent Application No. 60/690,268, filed Jun. 14, 2005 and titled “Digital Picture Taking Optical Reader Having Hybrid Monochrome And Color Image Sensor Array,” (now expired) and to U.S. Provisional Patent Application No. 60/692,890 filed Jun. 22, 2005, entitled “Digital Picture Taking Optical Reader Having Hybrid Monochrome And Color Image Sensor Array” (now expired) and to U.S. Provisional Patent Application No. 60/694,371 filed Jun. 27, 2005 entitled “Digital Picture Taking Optical Reader Having Hybrid Monochrome And Color Image Sensor Array” (now expired). The referenced U.S. patent application Ser. No. 11/445,930 and all of the aforementioned provisional patent applications are herein incorporated by reference in their entirety. The aforementioned U.S. patent application Ser. No. 11/445,930 is also related to U.S. patent application Ser. No. 11/174,447 filed Jun. 30, 2005 entitled, “Digital Picture Taking Optical Reader Having Hybrid Monochrome And Color Image Sensor” (now U.S. Patent Publication No. 2006/0274171) which is also incorporated herein by reference in its entirety.
- The present invention relates to optical readers in general and specifically, to an optical reader having a reduced specular reflection read failures.
- Performance of an optical reader may be hindered where an optical reader is operated to read bar code symbols or other indicia of a substrate having a “shiny” surface. Such substrates can include, e.g., metal, glass, and laminated plastic. Light rays emanating from a reader that are projected on a highly reflective shiny surface of a substrate may be substantially entirely reflected directly onto a reader image sensor array. Artisans skilled in the art of optical readers regard a “specular reflection” read condition to have occurred where a substantial percentage of light rays are reflected from a substrate and directed onto a reader image sensor array. Light rays are said to be reflected at a “specular” angle when light rays are reflected from a substrate at about the angle of incidence. Specular reflection tends to saturate a reader image sensor array to cause decoding failures.
- There is a need for an optical reader configured so that specular reflection read errors are reduced.
- The objects and features of the invention can be better understood with reference to the drawings described below, and the claims.
-
FIG. 1 a is an electrical block diagram of a hand held optical reader of the invention including a hybrid monochrome and a color sensing solid state image sensor array; -
FIG. 1 b is a block diagram of an alternative image sensor array which may be incorporated into an optical reader according to the invention; -
FIG. 1 c is a schematic block diagram illustrating an RF communications circuit according to the invention; -
FIG. 1 d is a schematic block diagram illustrating a display according to the invention; -
FIG. 1 e is a schematic view illustrating incorporation of a decode circuit, a signature autodiscrimination circuit, a demosaicing circuit, and a fusion circuit into an optical reader according to the invention; -
FIGS. 2 a-2 d are various partial exploded top views of an embodiment of a solid state image sensor array according to the invention; -
FIG. 3 a is a cutaway exploded side view of a monochrome pixel according to one embodiment of the invention; -
FIG. 3 b is a top view of the pixel shown inFIG. 3 a; -
FIG. 3 c is a cutaway exploded side view of a color sensitive pixel in one embodiment of the invention; -
FIG. 3 d is a top view of the pixel shown inFIG. 3 c; -
FIG. 4 a is an electrical block diagram of an embodiment of an image sensor according to the invention; -
FIG. 4 b is an electrical block diagram of an image sensor array of the invention showing incorporation of reset control lines in the image sensor array; -
FIG. 4 c is a timing diagram illustrating coordinated exposure control timing pulses and reset control timing pulses according to the invention; -
FIGS. 5 a-5 e are various partial exploded top views of an embodiment of a solid state image sensor array according to the invention; -
FIG. 5 f is a top perspective view of an image sensor integrated circuit chip incorporating an image sensor array according to the invention with an exploded view portion illustrating a pixel pattern of color sensitive “clusters” of pixels which pattern may be distributed throughout the array; -
FIGS. 5 g-5 i are top perspective views of image sensor integrated circuit chips incorporating a linear bar code symbol optimized image sensor array according to the invention with respective exploded view portions illustrating pixel patterns including “zones” of monochrome pixels and “zones” of color sensitive pixels; -
FIG. 5 j is a top perspective view of an image sensor integrated circuit chip incorporating a linear symbol optimized image sensor array according to the invention; -
FIG. 6 a is a cutaway exploded side view of a monochrome pixel according to one embodiment of the invention; -
FIG. 6 b is a top view of the pixel shown inFIG. 6 a; -
FIG. 6 c is a cutaway exploded side view of a color sensitive pixel in one embodiment of the invention; -
FIG. 6 d is a top view of the pixel shown inFIG. 6 c; -
FIG. 7 a is an electrical block diagram of an embodiment of an image sensor according to the invention; -
FIG. 7 b is an electrical block diagram of an image sensor array of the invention showing incorporation of reset control lines in the image sensor array; -
FIGS. 7 c and 7 d are schematic top views illustrating alternative configurations for a reset control system including separate sets of reset control lines for resetting a first subset of rows of pixels independent of resetting second subset of rows of pixels of an image sensor array according to the invention; -
FIG. 8 a is an exploded perspective view of an imaging module according to the invention; -
FIGS. 8 b and 8 c are front and side views, respectively, of the imaging module shown inFIG. 8 a; -
FIG. 8 d shows an illumination and aiming pattern which may be projected by an optical reader according to the invention; -
FIG. 8 e is a top view of an alternative imaging module incorporating a laser based aiming pattern generating system; -
FIG. 8 f is a front view of a polarizer plate which may be included as part of an imaging module herein, e.g., the imaging middle shown inFIG. 8 a; -
FIGS. 9 a and 9 b are physical form views of various hand held optical readers according to the invention; -
FIG. 9 c is a perspective view of a hand held mobile telephone (a “cell phone”) which may incorporate a hybrid monochrome and color image sensor array according to the invention and which may be configured according to the invention; -
FIG. 10 is a schematic view of a system incorporating a plurality of optical readers according to the invention; -
FIG. 11 is an application schematic view illustrating an optical reader according to the invention being operated to capture image data representing a parcel that carries a plurality of bar code symbols; -
FIG. 12 a is an application schematic view illustrating a first optical reader according to the invention and a second remotely located optical reader according to the invention being operated to take first and second digital pictures of a parcel at first and second locations that are a distance apart for purposes of determining whether the parcel was damaged during delivery from the first location to the second location; -
FIG. 12 b is another application schematic view illustrating an optical reader being used to take a color picture of a delivery vehicle; -
FIG. 13 a is an application schematic diagram according to the invention illustrating an optical reader according to the invention being used to read bar codes of a vehicle and to take color pictures of a vehicle; -
FIG. 13 b is a view of a VIN sticker which may be disposed on the vehicle ofFIG. 13 a; -
FIG. 13 c is a view of a VIN plate which may be disposed on the vehicle ofFIG. 13 a; -
FIG. 13 d is a view of a vehicle registration sticker which may be disposed on the vehicle ofFIG. 13 a; -
FIG. 13 e is a view of an optical reader programmed to display a GUI form assisting an application wherein an optical reader, according to the invention, is utilized to decode bar code symbols and to take color pictures of a vehicle; -
FIGS. 14 a-14 c are various flow diagrams illustrating the invention; -
FIGS. 14 d-14 f are additional flow diagrams illustrating examples of operation of an optical reader according to the invention in an indicia decode mode of operation; -
FIGS. 14 g and 14 h are additional flow diagrams illustrating examples of operation of an optical reader according to the invention in a picture taking mode of operation; -
FIG. 14 i is a flow diagram illustrating operation of a fusion circuit of an optical reader according to the invention which processes monochrome and color image data to produce a high resolution visual display color frame of image data; -
FIGS. 15 a-15 e are various image capture initiation control signal timing diagrams illustrating the invention; -
FIGS. 16 a-16 c illustrate various pixelized frames of image data which may be captured by an optical reader according to the invention; -
FIG. 17 a is an electrical block diagram of an optical reader according to the invention having a plurality of imaging modules; -
FIGS. 17 b and 17 c illustrate alternative hardware blocks that can be utilized with the electrical circuit ofFIG. 17 a; -
FIGS. 17 d and 17 e illustrate imaging modules which may be utilized with the reader ofFIG. 17 a; -
FIGS. 17 f and 17 g illustrate exemplary optical readers incorporating a pair of imaging modules; -
FIG. 18 a is a schematic view of a cyan-magenta-yellow (CMY) image sensor array in accordance with the invention which may be incorporated into an optical reader according to the invention and which may be controlled to generate both a decode frame of image data and a visual display color frame of image data; -
FIG. 19 a is a schematic view of a hybrid monochrome and polarizer image sensor array in accordance with the invention which may be incorporated in an optical reader according to the invention; -
FIG. 19 b is a top perspective view of a hybrid monochrome and polarizer image sensor array according to the invention with an exploded view section illustrating a pattern of light polarizing pixels that may be distributed throughout the image sensor array; -
FIG. 19 c is a flow diagram illustrating an exemplary operational mode of an optical reader according to the invention which incorporates a hybrid monochrome and polarizer image sensor array according to the invention; -
FIGS. 20 a and 20 b are top perspective views of a monochrome polarizer and color sensitive image sensor array according to the invention with an exploded view section illustrating a pattern of light polarizing pixels and color sensitive pixels that may be distributed throughout the array; -
FIG. 21 is a schematic view of an image sensor integrated circuit chip incorporating an image sensor array having color sensitive pixels disposed therein with two different periods of distribution; -
FIG. 22 a is a schematic block diagram of an autodiscrimination circuit which may be utilized with the invention; -
FIG. 22 b is a process for practicing principles of the invention including automatically discriminating between different dataform types; -
FIG. 22 c shows one embodiment of a plurality of curvelent detector maps which may be utilized with the invention; -
FIG. 22 d shows another embodiment of a plurality of curvelent detector maps which may be utilized with the invention; -
FIG. 22 e is a diagrammatic representation of a histogram analysis which may be performed in one embodiment of the invention; -
FIGS. 22 f-22 i are diagrammatic representations of an image data segmentation process according to embodiments of the invention. - There is provided in one embodiment an optical reader having an image sensor array. In one embodiment, the image sensor array can include a first subset of pixels and a second subset of pixels. The first subset of pixels can be devoid of light polarizing filter elements, and the second subset of pixels can be light polarizing pixels including light polarizing filter elements. An optical reader can be configured to selectively read out image data from an image sensor array's light polarizing pixels.
- An optical reader image sensor array of the invention can include light polarizing pixels, each light polarizing pixel having a light polarizing filter element (light polarizing filter) that significantly attenuates polarized light rays generated from an appropriately polarized light source and reflected at a specular angle; thus, reducing the contribution of specularly reflected light rays to generated image signals from the polarizing pixels. In one embodiment, a first subset of pixels of an optical reader image sensor array are monochrome pixels and a second subset of pixels are light polarizing pixels. For decoding decodable indicia in specular reflection read conditions, image data corresponding to the light polarizing pixels can be selectively transferred to a decode circuit, either by way of selecting reading out image data from the light polarizing pixels, or by selectively extracting image data corresponding to light polarizing pixels from a frame of image data including image data in addition to image data corresponding to light polarizing pixels.
- In another embodiment, there is provided a picture taking optical reader having a hybrid monochrome and color (monocolor) solid state image sensor array. The hybrid image sensor array comprises a plurality of pixels including a first subset of pixels and a second subset of pixels, wherein the first subset of pixels are monochrome pixels and the second subset of pixels are color sensitive pixels having wavelength selective color filter elements.
- In one embodiment, the monochrome first subset of pixels is formed in a checkerboard pattern, and voids are formed at the corners of pixels of the first subset, such that combinations of voids of adjacent pixels define open areas. Pixels of the color sensitive second subset of pixels are formed at the open areas, and wavelength selective filter elements are formed on pixels of the second subset but not on pixels of the first subset.
- In another embodiment, an optical reader solid state image sensor array includes a plurality of pixels formed in a plurality of rows on an IC chip in a checkerboard pattern wherein each pixel has approximately the same dimension. The majority of pixels of the image sensor array are monochrome pixels of the first subset. Color sensitive pixels of the second subset are at spaced apart positions and are uniformly or substantially uniformly distributed throughout the image sensor array. Color sensitive pixels may be distributed in the array in a specific pattern of uniform distribution such as a period of P=2, where every other pixel of every other row of the image sensor array is a color sensitive pixel, or a period of P=4 where, for every fourth row of pixels of the array, every fourth pixel is a color sensitive pixel.
- A hybrid monochrome and color sensing solid state image sensor array of the invention may be incorporated in an imaging module which, in addition to having an image sensor array constructed in accordance with the invention includes such elements as an imaging lens, an illumination assembly including a field illumination assembly, an aiming illumination assembly and a support member for supporting the above elements. An imaging module, in turn, may be incorporated into a hand held housing which encapsulates and supports the imaging assembly.
- Utilizing complementary metal-oxide-silicon (CMOS) integrated circuit fabrication technologies the image sensor array in one embodiment can be made to have selectively addressable pixels. Where the image sensor array is constructed to have selectively addressable pixels, pixels of the first subset of pixels can be selectively addressed independent of the second subset of pixels so that image data corresponding to the first subset of pixels is selectively read out independent of the second subset of pixels. Image sensor arrays having selective read out capability can be provided utilizing alternative fabrication technologies.
- In a further aspect, an optical reader according to the invention includes separate and independently controllable reset control lines for resetting monochrome pixels and color sensitive pixels of the image sensor array. During exposure periods for exposing color sensitive pixels, monochrome pixels may be driven into reset. During exposure periods for exposing monochrome pixels, color sensitive pixels may be driven into reset. Driving pixels not being selectively addressed for image data read out into a reset state reduces cross-talk between pixels of the image sensor array.
- By incorporating within a single low cost image sensor array a combination of monochrome pixels and color sensitive pixels, an optical reader according to the invention provides indicia decoding performance approximately equal to the performance of an optical reader having an all monochrome image sensor array, and picture taking performance (i.e., the ability to obtain visual display quality color frames of image data) approximately equal to or superior to that of a digital camera incorporating an all color pixel image sensor array, wherein each pixel of the array includes a wavelength selective filter element.
- An electrical block diagram of an
optical reader 100 according to the invention is shown inFIG. 1 a.Reader 100 includes a solid stateimage sensor array 182A, incorporated on an image sensor integratedcircuit chip 1082A shown inFIG. 1 a as a CMOS image sensor integrated circuit (IC) chip. In an important aspect, as will be described herein,image sensor array 182A includes a plurality of pixels and wavelength sensitive color filter elements associated with a color sensitive subset of the pixels, wherein the remaining pixels external to the color sensitive subset of pixels are devoid of associated wavelength selective filter elements. Becauseimage sensor array 182A includes both monochrome pixels and color sensitive pixels,image sensor array 182A may be termed a hybrid monochrome and color image sensor array.Reader 100 further includes aprocessor IC chip 548 and acontrol circuit 552.Control circuit 552 in the embodiment ofFIG. 1 a is shown as being provided by a central processing unit (CPU) ofprocessor IC chip 548. In other embodiments,control circuit 552 may be provided by e.g., a programmable logic function execution device such as a field programmable gate array (FPGA) or an application specific integrated circuit (ASIC).Imaging lens 212 focuses images onto an active surface ofimage sensor array 182A and together withimage sensor array 182A forms animaging assembly 200.Control circuit 552 executes picture taking and indicia decoding algorithms in accordance with instructions stored inprogram memory EPROM 562 which together withRAM 560 andFlash memory 564 forms areader memory 566.Reader memory 566 is in communication withprocessor IC chip 548 viasystem bus 570. Mainprocessor IC chip 548 may be a multifunctional IC chip such as an XSCALE PXA25x processor IC chip including central processing unit (CPU) 552.Reader 100 further includes a field programmable gate array (FPGA) 580. Operating under the control ofcontrol circuit 552,FPGA 580 receives digital image data from imagesensor IC chip 1082A and transfers that image data intoRAM 560 so that the image data can be further processed (e.g., by the decoding of a bar code symbol).Processor IC chip 548 can include an integrated frame grabber. For example,processor IC chip 548 can be an XSCALE PXA27X processor IC chip with “Quick Capture Camera Interface” available from INTEL. Whereprocessor IC chip 548 includes an integrated frame grabber, the integrated frame grabber may provide the frame acquisition functionality ofFPGA 580.Reader 100 further includes anillumination assembly 104 and amanual trigger 216. Imagesensor IC chip 1082A in the embodiment ofFIG. 1 a includes an on-chip control/timing circuit 1092, an on-chip gain circuit 1084, an on-chip analog-to-digital converter 1086 and an on-chip line driver 1090. An image sensor array which is incorporated intooptical reader 100 may take on a variety of forms. InFIG. 1 areader 100 includes firstimage sensor array 182A. However, as indicated byhardware block 208, theimage sensor array 182A may be replaced. For example, in the embodiment ofFIG. 1 b,reader 100 incorporatesimage sensor array 182B. In other embodiments,optical reader 100 incorporates more than one image sensor array. Various embodiments of image sensor arrays which may be incorporated intoreader 100 are described herein. - In a further aspect,
reader 100 includes a radio frequency (RF)communication interface 571. Radiofrequency communication interface 571 may include one or more radio transceivers. Referring to the schematic diagram ofFIG. 1 c, radiofrequency communication interface 571 may include one or more of an 802.11radio transceiver 5712, aBluetooth radio transceiver 5714, a GSM/GPS radio transceiver 5716 or a WIMAX (802.16)radio transceiver 5718. Radiofrequency communication interface 571 facilitates wireless communication of data betweendevice 100 and a spaced apartdevice 150. I/O communication interface 572 includes one or more serial or parallel hard-wired communication interfaces facilitating communication with a spaced apartdevice 150 as will be described further in connection withFIG. 10 . I/O communication interface 572 may include one or more of an Ethernet communication interface, a universal serial bus (USB) interface, or an RS-232 communication interface.Optical reader 100 may further include akeyboard 508 for entering data, apointer mover 512 for moving a pointer of a graphical user interface (GUI) and atrigger 216 for initiating bar code reading and/or picture taking.Optical reader 100 may also include adisplay 504, such as a monochrome or color LED display and atouch screen 504T overlaid overdisplay 504. As shown in the schematic block diagram ofFIG. 1 d,display 504 may include adisplay screen 5042 coupled todisplay controller 5044 for displaying color image data.Display controller 5044 receives a visual display color frame of image data fromcontrol circuit 552, and reformats that data for display depending on the particular requirements ofdisplay screen 5042, including the pixel resolution ofdisplay screen 5042. All of the components ofFIG. 1 a can be encapsulated and supported by a hand heldhousing 101, e.g., as shown inFIGS. 9 a-9 c. Additional features and functions of the components ofreader 100 shown inFIG. 1 a are described herein. - Referring to
FIG. 1 e,optical reader 100 may be regarded as having various processing circuits (modules). Indicia decodecircuit 1702 receives image data and decodes decodable indicia therein such as bar code indicia and OCR character data.Optical reader 100 can be configured so that indicia decodemodule 1702 decodes such bar code symbols UPC/EAN, Code 11, Code 39, Code 128, Codabar,Interleaved 2 of 5, MSI, PDF417, MicroPDF417, Code 16K, Code 49, MaxiCode, Aztec, Aztec Mesa, Data Matrix, Qcode, QR Code, UCC Composite, Snowflake, Vericode, Dataglyphs, RSS, BC 412, Code 93, Codablock, Postnet (US), BPO4 State, Canadian 4 State, Japanese Post, KIX (Dutch Post), Planet Code and the like, and such OCR character forms as OCR A, OCR B, and the like.Autodiscrimination circuit 1704 processes received image data and distinguishes between handwritten character data and decodable indicia.Autodiscrimination circuit 1704 may include indicia decodecircuit 1702.Autodiscrimination circuit 1704 and indicia decodecircuit 1702 may be physically embodied by a combination ofcontrol circuit 552 andmemory 566. Specifically,control circuit 552 operating under the control of a program stored inmemory 562 may process image data stored inmemory 560 to decode decodable indicia therein or to discriminate between handwritten character data and decodable indicia. Further aspects of indicia decodecircuit 1702 andautodiscrimination circuit 1704 are described in copending U.S. patent application Ser. No. 10/958,779 entitled, System And Method To Automatically Discriminate Between A Signature And A Barcode, filed Oct. 5, 2004 and U.S. patent application Ser. No. 11/077,975, filed Mar. 11, 2005 entitled, Bar Code Reading Device With Global Electronic Shutter Control, both of which are incorporated herein by reference. As will be described further herein,optical reader 100 may further include ademosaicing circuit 1706, and afusion circuit 1708.Demosaicing circuit 1706 receives as an input a color filter array image data frame (e.g., a Bayer pattern image) and produces as an output a visual display color frame of image data.Fusion circuit 1708 receives as inputs both monochrome and color image data and produces as an output a visual display color frame of image data having a spatial resolution at or on the order of the pixel resolution of the optical reader's hybrid monochrome and color image sensor array. Likecircuit circuits control circuit 552 andmemory 566.Control circuit 552 as well ascircuits FIGS. 9 a-9 c) or else one or more ofcircuits device 150 as described in connection withFIG. 10 . - A visual display color frame of image data as referred to herein, in one embodiment is an image frame including a set of color indicating data at each of a plurality of pixel positions, wherein each set of color indicating data represents a color at a discrete position of a target 1850 (shown in
FIG. 8 d). Each set of color indicating data includes three color values, e.g., a color scale value representing red, a color scale value representing blue, and a color scale value representing green. Alternatively, the set of color indicating data for each pixel position may include a cyan value, a magenta value and a value representing yellow. - In one specific example, the set of color indicating data for each pixel position of a visual display color frame of image data output by
demosaicing circuit 1706 orfusion circuit 1708 are RGB data sets including 24 bits of information, wherein the first 8 bits represent a red color scale value (red value) for the pixel position, the second 8 bits represent a green color scale value (green value) for the pixel position and the third 8 bits represent a blue color scale value (blue value) for the pixel position. - A major feature of the invention is the construction of the optical reader's image sensor array various embodiments of which are shown and described throughout several views including the views of
FIGS. 2 a-7 d. - A first embodiment of a hybrid monochrome and color sensitive (monocolor) solid state image sensor array is shown and described in
FIG. 1 a andFIGS. 2 a-4 b. - Referring to
FIG. 1 a andFIGS. 2 a-4 b, solid stateimage sensor array 182A includes a monochrome first subset ofpixels 250M and a color sensitive second subset ofpixels 250C. The first subset ofmonochrome pixels 250M is formed in a checkerboard pattern and voids 253 as shown inFIG. 2 a are formed at the corners of pixels of the first subset, such that combinations of voids, e.g., voids 253-1, 253-2, 253-3, 253-4 of adjacent pixels define open areas, e.g.,open area 255, each open area bounded by four pixels of the first subset. With further reference toimage sensor array 182A,pixels 250C forming a second subset ofpixels 250C are disposed in theopen areas 255, and wavelength selective filter elements, e.g.,filter element FIG. 2 b are formed on pixels of the second subset but not on pixels of the first subset.Monochrome pixels 250M as described herein are devoid of color filter elements (color filters). Pixels of the first monochrome pixel subset are in the shape of twelve sided polygons. The pixels are cross-shaped as seen from the top view that is indicated byFIGS. 2 a-2 d (the monochrome pixels are square shaped as modified by the presence of voids 253). Pixels of the color sensitive second subset are square as seen from a top view. - In the version of
image sensor array 182A shown inFIG. 2 b, colorsensitive pixels 250C ofimage sensor array 182A include either a cyan (Cy)filter element 260C or magenta (Mg)filter element 260M. In the version ofFIG. 2 c, colorsensitive pixels 250C ofimage sensor array 182A include either ared filter element 260R, agreen filter element 260G or a bluecolor filter element 260B (RGB filters). The colorsensitive pixels 250C can be distributed throughoutimage sensor array 182 according to a Bayer pattern wherein there are N blue pixels, N red pixels and 2N green pixels. Color filter elements of any image sensor array pixel as described herein can be deposited on the major body of colorsensitive pixels 250C by way of a depository process. As will be explained herein, visual display color image data can be obtained utilizing either the version ofimage sensor array 182A shown inFIG. 2 b or the version ofimage sensor array 182A shown inFIG. 2 c, or another version ofimage sensor array 182A such as a version including cyan, magenta and yellow (CMY) color sensitive pixels. Because cyan and magenta filters require only one dye and not two dyes (as in red, green, and blue filters) a version ofimage sensor array 182A including cyan and magenta filter elements in place of red, green and blue filter elements allows more light to pass through to a photodetector of the pixels and exhibits a higher signal to noise ratio than a version including red, green and blue filters. Nevertheless, an image sensor array having a combination of red, green and blue (RGB) filter elements may be preferred for certain applications. Referring toFIG. 2 d,image sensor array 182A may includemicrolenses 320 for directing of light rays incident onimage sensor array 182A. Further aspects ofmicrolenses 320, including monochrome pixels,microlenses 320M, and colorsensitive pixel microlenses 320C are described herein. - Exploded physical form views of an image
sensor pixel array 182A, wherearray 182A is configured to operate in a global electronic shutter operating mode are shown and described inFIGS. 3 a-3 d. Amonochrome pixel 250M ofimage sensor array 182A is shown inFIGS. 3 a and 3 b.Monochrome pixel 250M includes aphotodetector 302 which may be of photodiode or photogate construction, atransfer gate 304, a floatingdiffusion 306, areset transistor 307 includingreset gate 308, a rowselect transistor 309 including rowselect gate 310 and a sourcefollower amplifier transistor 311 includingamplifier gate 312. An important feature ofpixel 250M is opaqueoptical shield 316. Opaqueoptical shield 316, typically comprising metal, shields light rays from components ofpixel 250M other thanphotodetector 302. Accordingly, pixels from each of several rows ofimage sensor array 182A can be simultaneously exposed to light in a global electronic shutter operating mode without the light rays modulating charges stored in floatingdiffusion 306 or another storage region. Further aspects of image sensor arrays capable of operating in a global electronic shutter operating mode are described in U.S. patent application Ser. No. 11/077,975 incorporated herein by reference. Referring to additional aspects ofpixel 250M,pixel 250M includesmicrolens 320 which may be disposed on light transmissiveprotective layer 322.Microlens 320 collects light from a larger surface area thanphotodetector 302 and directs light towardphotodetector 302. - A color
sensitive pixel 250C ofimage sensor array 182A is described with reference toFIGS. 3 c and 3 d. Colorsensitive pixel 250C is similar in construction tomonochrome pixel 250M. Colorsensitive pixel 250C includes aphotodetector 302 which may be of photodiode or photogate construction, atransfer gate 304 for transferring charge fromphotodetector 250C, a floatingdiffusion 306, areset transistor 307 includingreset gate 308, a rowselect transistor 309 including rowselect gate 310 and a sourcefollower transistor amplifier 311 includingamplifier gate 312. Colorsensitive pixel 250C also includesopaque shield 320 which shields light from light sensitive components ofpixel 250C other thanphotodetector 302.Pixel 250C may also includemicrolens 320 for increasing the amount of light incident onphotodetector 302. In addition to the above elements colorsensitive pixel 250C includes a wavelength selectivecolor filter element 260 formed thereon. Wavelength selectivecolor filter element 260 may be disposedintermediate microlens 320 andprotective layer 322. In the versions ofFIGS. 2 a-2 d, it is seen that each colorsensitive pixel 250C has four adjacentmonochrome pixels 250M. -
Microlenses 320 as shown inFIGS. 3 a and 3 c are also shown in the view ofFIG. 2 d.Monochrome pixel microlens sensitive microlens sensitive pixels 250C are disposed in open areas defined by voids of checkerboard pattern of a first monochrome subset ofpixels 250M,microlenses 320C of colorsensitive pixels - Color
sensitive pixel 250C ofimage sensor array 182A as best seen by a comparison betweenFIGS. 3 b and 3 d and consumes a smaller surface area thanpixel 250M. In one version,pixel 250M includes an area, as seen from a top view, of about 12 nm by 12 nm whilepixel 250C includes an area, as seen from a top view, of about 6 μm by 6 μm. In another version,pixel 250M includes a top surface area of about 6 μm by 6 μm, whilepixel 250C includes a top surface area of about 3 μm or 3 μm. Size reductions ofpixel 250M orpixel pixel 250M and/orpixel 250C. - A transistor count of a
pixel 250C ofimage sensor array 182A may readily be reduced by eliminating optically shielded floatingdiffusion 306 in which charges are stored on a temporary basis to facilitate global electronic shutter operation. Accordingly, in one embodiment,monochrome pixels 250M ofimage sensor array 182A have more transistors than colorsensitive pixels 250C but are capable of being exposed on a global electronic shutter basis, whereas colorsensitive pixels 250C have fewer transistors thanmonochrome pixels 250M but are not capable of being exposed on a global electronic shutter basis. In yet another embodiment with reference toimage sensor array 182A having smaller dimensioned color sensitive pixels than monochrome pixels, the relatively largermonochrome pixels 250M have a transistor count sufficient to facilitate global shutter operation, but the relatively smaller colorsensitive pixels 250C are passive pixels requiring off-pixel amplification, and comprise a single transistor each. Further aspects of global electronic shutter and rolling shutter operations relative to image sensor arrays which may be incorporated intoreader 100 are described herein. - Referring to
FIG. 4 a, a high level electrical block diagram ofimage sensor array 182A is shown. According to one version,image sensor array 182A is an active pixel image sensor array of complementary metal oxide semiconductor (CMOS) construction such that eachpixel pixel amplifier 311 for amplifying signals corresponding to light incident onphotosensitive region 252. Eachpixel storage element 306.Image sensor array 182A further includes two-dimensional grid ofinterconnects 262 which are in electrical communication withrespective column circuitry 270 androw circuitry 296.Row circuitry 296 andcolumn circuitry 270 enable such processing and operational tasks as selectively addressing pixels, decoding pixels, amplification of signals, analog-to-digital conversion, applying timing, read out and reset signals and the like. - Among the control lines forming
interconnect grid 262 ofimage sensor array 182A are pixel reset control lines. When pixels are reset by application of an appropriate control signal on a reset control line, residual charges which have accumulated on the pixels are connected temporarily to VDD so that built up charges on pixels of the image sensor array drain out of the pixels. In accordance with the invention,image sensor array 182A includes separate reset control lines formonochrome pixels 250M andcolor pixels 250C. Referring toFIG. 4 b,image sensor array 182A may be constructed so thatimage sensor array 182A has a first set ofreset control lines 262R-M for resettingmonochrome pixels 250M and a second set ofreset control lines 262R-C for resettingcolor pixels 250C. - In certain operating modes
optical reader 100 selectively reads out a windowed frame of image data comprising image data frommonochrome pixels 250M. In other operating modes,optical reader 100 selectively reads out a windowed frame of image data comprising image data fromcolor pixels 250C. In accordance with the invention, a reset control timing pulse can be applied toimage sensor array 182A during the time that a windowed frame of image data is being captured to reset pixels ofimage sensor array 182A that are not being selectively addressed for image data read out. As shown by the timing diagram ofFIG. 4 c, an exposurecontrol timing pulse 354 can be coordinated with a resetcontrol timing pulse 370. - With further reference to
FIG. 4 c, exposurecontrol timing pulse 354 may control exposure ofmonochrome pixels 250M ofimage sensor array 182A (or alternatively,color pixels 250C) ofimage sensor array 182A, while resetcontrol timing pulse 370 drives pixels not being selectively addressed into a reset state. When pixels are reset, charges built up on pixels tend to be drained out of the pixels. Further, it is believed that photons entering pixels driven into reset may be refracted so that fewer photons become incident on neighboring pixels being exposed for image data read out. Accordingly, coordinating the timing of anexposure control pulse 354 for exposing selectively addressed pixels and a resetcontrol timing pulse 370 for resetting pixels not being selectively addressed reduces cross talk between pixels. - Referring again to the view of
FIG. 4 b,image sensor array 182A may be constructed so that the presence of multiple reset control lines 162R-C, 162R-M do not substantially decrease the fill factor of pixels ofimage sensor array 182A.FIG. 4 b shows a schematic top view of multiple reset control lines 162R-M, 162R-C incorporated inimage sensor array image sensor array 182A in a layered manner so that for a substantial portion ofimage sensor array 182A, control lines 164R-M have x, y positions that coincide with x, y positions of control line 164R-C (axes are defined inFIG. 8 a). Control lines 164R-C in the embodiments ofFIG. 4 b are installed at a different height (a different Z axis position) withinimage sensor array 182A such that control lines 162R-M and 162R-C, for substantial length of the control lines, have common x, y positions. Installing the multiple control lines to be on top of one another so that the control lines have a common x, y axis position withinimage sensor array 182A reduces the amount of fill factor degradation which would otherwise result from installation of an additional set of reset control lines withinimage sensor array 182A. - An alternative construction for an image sensor array according to the invention is described with reference to
FIGS. 5 a-7 b. In the embodiment ofFIGS. 5 a-7 bimage sensor array 182B includes a plurality of square shaped pixels (as seen from a top view) in a checkerboard pattern, each of the pixels having substantially the same dimensions. Eachpixel image sensor array 182B may be constructed to have approximately the same top surface dimensions as seen from the top views ofFIGS. 5 a-5 i and approximately the same side view cross-sectional dimensions as seen from the cross-sectional views ofFIGS. 6 a-6 d.Image sensor array 182B is similar to the construction of a standard off-the-shelf monochrome image sensor array except that select ones of the pixels of the image sensor array have an associated wavelength selective color filter element. Solid stateimage sensor array 182B includes a plurality of pixels formed in a plurality of rows. In the version ofFIGS. 5 a-5 e, a monochrome first subset ofpixels 250M comprise the majority of pixels of the array. Wavelength selectivecolor filter elements 260 are included in the second subset of colorsensitive pixels 250C. The color sensitive second subset ofpixels 250C comprises pixels at spaced apart pixel positions uniformly distributed or substantially uniformly distributed throughout the plurality of pixels forming theimage sensor array 182B. In the embodiment ofFIGS. 5 a and 5 b, every other pixel in every other row of pixels (e.g.,pixel row image sensor array 182B can be provided by including an appropriately designed color filter array on an image sensor array of an MT9M111 Digital Clarity SOC 1.3 megapixel CMOS image sensor IC chip of the type available from Micron, Inc., an MT9V022 image sensor IC chip also available from Micron, Inc. or a VV6600 1.3 megapixel CMOS image sensor IC chip of the type available from STMicroelectronics. Other image sensor IC chips which can be utilized to provideimage sensor array 182B include MT9M413 image sensor IC chip available from Micron, Inc., a KAC-0311 image sensor IC chip manufactured by Kodak, Inc. and a KAI-0340 image sensor IC chip also manufactured by Kodak, Inc. Operational aspects of the referenced KAI-0340 image sensor IC chip are described further herein. Various manufacturer product description materials respecting certain of the above image sensor IC chips are appended to Provisional Patent Application No. [not yet assigned] filed Jun. 22, 2005 (Express Mail Label No. EV554216715US) and Provisional Patent Application No. [not yet assigned] filed Jun. 27, 2005 (Express Mail Label No. EV554216661US) which are incorporated herein by reference. The above commercially sold image sensor IC chips can be utilized (with additions or replacements of filter elements as are necessary) to provide any one ofimage sensor arrays - The above referenced MT9V022 and MT9M413 image sensor IC chips manufactured by Micron, Inc., and KAC-0311 image sensor IC chip by Kodak, Inc. are CMOS image sensor IC chips that may be operated in a global electronic shutter mode such that all rows of pixels subject to image data read out have common exposure periods; that is, all rows of pixels subject to image data read out for reading out a frame of image data (i.e., full frame or “windowed frame”) have a common exposure start time and a common exposure stop time. For global electronic shutter operation, an exposure control timing pulse, as will be described herein is applied to the image sensor array. Exposure of each row of pixels subject to image data read out begins at the leading edge of the exposure control timing pulse and ends at the falling edge of the exposure control timing pulse. In its technical literature, Micron, Inc. uses the trademark TRUESNAP with reference to a global electronic shutter operating mode.
- Referring to
FIG. 5 b, wavelength selective color filter elements (filters) formed on colorsensitive pixels cyan filter elements 260C and magentacolor filter elements 260M. As shown inFIG. 5 a, wavelength sensitive filters of colorsensitive pixels 250C may also be a combination ofred filter elements 260R,green filter elements 260G andblue filter elements 260B. Because cyan and magenta filters require only one dye and not two dyes (as in red green and blue filters), the version ofFIG. 5 b allows more light to pass through to a photodetector (e.g.,photodetector 302 as shown inFIG. 6 c) and exhibits a higher signal to noise ratio than the embodiment ofFIG. 5 b. Nevertheless, the version ofFIG. 5 a may be preferred for certain applications. - In the embodiment of
FIGS. 5 a-7 d, hybrid monochrome andcolor image sensor 182B can be made by including an appropriately designed color filter array on a commonly available, off-the-shelf image sensor array in a standardly known checkerboard pattern, each pixel of the array having substantially the same dimensions. A larger portion ofimage sensor array 182B is shown inFIG. 5 c, where pixels designated by the letter “c” are colorsensitive pixels 250C and pixels not designated by the letter “c” aremonochrome pixels 250M. In the example ofFIG. 5 c, color sensitive pixels are formed onarray 182B with a period of P=2, meaning the every other pixel of every other row of pixels is a colorsensitive pixel 250C. In the version ofFIG. 5 d, color sensitive pixels are formed onarray 182B with a period of P=3, meaning that every third pixel of every third row is a colorsensitive pixel 250C. In the version ofFIG. 5 e, color sensitive pixels, c, are formed with a period of P=4, meaning that every fourth pixel from every fourth row of pixels is a colorsensitive pixel 250C. In the versions ofFIGS. 5 a-5 e, each colorsensitive pixel 250C has eight adjacentmonochrome pixels 250M (two side adjacent, one top adjacent, one bottom adjacent and four corner adjacent). - Additional views of
image sensor array 182B including a subset ofmonochrome pixels 250M and a subset of colorsensitive pixels 250C, wherein each pixel of the image sensor array has substantially equal dimensions are shown and described in connection withFIGS. 5 f-5 j. - Referring to the version of
FIG. 5 f,image sensor array 182B includes the first subset ofmonochrome pixels 250M and a second subset of colorsensitive pixels 250C. The colorsensitive pixels 250C ofimage sensor array 182B in the version ofFIG. 5 f are formed in clusters such ascluster 257R,cluster 257G andcluster 257B. - Each cluster 257 in the version of
FIG. 5 f includes a plurality of pixels in successive horizontally adjacent pixel positions, such that each pixel of the cluster is horizontally adjacent to at least one other color sensitive pixel. Color sensitive clusters of pixels are distributed uniformly or substantially uniformly throughoutimage sensor array 182B. Clusters may be formed in accordance with the standardized color filter pattern such as an RGB Bayer pattern or a cyan-magenta-yellow (CMY) pattern. Each cluster may have a plurality of pixels with each pixel of every individual cluster having a filter element of the same wavelength rating. In the specific version shown inFIG. 5 f, clusters are distributed throughoutimage sensor array 182B in a pattern that is accordance with the pattern of Bayer color filter array. -
Cluster 257G includes three horizontally adjacent green pixels.Cluster 257R includes three horizontally adjacent red pixels.Cluster 257B includes three horizontally adjacent blue pixels. As will be described further in connection withFIG. 7 c, the version ofimage sensor array 182B including a distribution of color sensitive pixels in horizontally arranged clusters as shown inFIG. 5 f is particularly useful where it is desired to include inimage sensor array 182B separate and independently controllablereset control lines 262R-M and 262R-C for separately and independently resetting monochrome pixels ofimage sensor array 182B and color sensitive pixels of imagesensory array 182B without increasing the thickness ofimage sensor array 182B. - Referring now to the versions of
image sensor array 182B shown inFIG. 5 g-5 j,image sensor array 182B having a subset of monochrome pixels in a subset of color sensitive pixels may be configured to include “zones” of monochrome pixels and “zones” of color sensitive pixels. A “zone” of pixels herein is a collection of positionally related pixels at a specified area of an image sensor array each having a color filter element or alternatively, each being without a color element. Examples of zones described herein comprise all pixels of one row of pixels or all pixels of each of several consecutive rows of pixels. In the version ofFIG. 5 g,image sensor array 182B includes two color sensitive zones ofpixels 2500C and a single monochrome zone ofpixels 2500M. Each zone of pixels comprises a plurality of horizontally, vertically or diagonally adjacent pixels. The plurality of pixels of a monochrome zone of pixels, e.g.,zone 2500M are all devoid of a color sensitive filter element. The plurality of adjacent pixels in a color sensitive zone of pixels, e.g.,zone 2500C, all include a color sensitive filter element. - Referring to the version of
FIG. 5 g, monochrome zone ofpixels 2500M is interposed between a pair of color sensitive zones ofpixels 2500C. Monochrome zone ofpixels 2500M in the version ofFIG. 5 g comprises a single row of pixels ofimage sensor array 182B at or approximately the center ofimage sensor array 182B. The first color sensitive zone of pixels of animage sensor array 182B includes all pixels from the row of pixels ofzone 2500M up to the top row ofimage sensor array 182B. The second color sensitive zone ofpixels 2500C in the version ofFIG. 5 g includes all pixels from all rows from the center row monochrome zone ofpixels 2500M down to the bottom row of pixels ofimage sensor array 182B. The color filter elements of colorsensitive pixels 250C ofimage sensor array 182B may be formed in a standard color filter pattern, e.g., an RGM Bayer color filter pattern or a CMY pattern. - Referring to
FIG. 5 h, another version ofimage sensor array 182B is shown and described. The version ofFIG. 5 h is similar to the version ofFIG. 5 g except that the monochrome zone ofpixels 2500M is expanded to include ten consecutive rows of pixels at the center or approximately the center ofimage sensor array 182B. - In the version of
image sensor array 182B as shown inFIG. 5 i, a single color sensitive zone ofpixels 2500C is interposed between two relatively small width monochrome zones ofpixels 2500M formed at the top and bottom ofimage sensor array 182B respectively. In the version ofimage sensor array 182B shown inFIG. 5 a, the first monochrome zone ofpixels 2500M comprises the first ten rows of pixels ofimage sensor array 182B and a second monochrome zone ofpixels 2500M includes pixels of the bottom ten rows ofimage sensor array 182B. Color sensitive zone ofpixels 2500C in the version ofFIG. 5 i includes all pixels of the array excluding the first ten and the last ten rows of pixels ofimage sensor array 182B. In the versions ofFIG. 5 h andFIG. 5 i, the pixels of the colorsensitive zones 2500C shown may include color filter elements in accordance with the pattern of a standardized color filter array, e.g., an RGB Bayer pattern or a CMY pattern. - The version of
image sensor array 182B shown inFIG. 5 j is similar in construction to the version ofFIG. 5 g except that the version ofFIG. 5 j includes additional monochrome zones ofpixels 2500M. In the version ofFIG. 5 jimage sensor array 182B includes a pair of diagonal zones ofmonochrome pixels 2500M-D extending through a center (actual or approximate) ofimage sensor array 182B and a vertically extending zone ofmonochrome pixels 2500M-V extending through a center ofimage sensor array 182B. The linear zones ofmonochrome pixels 2500M shown in the version ofFIG. 5 j may include a minor dimension equal to one pixel width or more than one pixel width. For example, the vertically extending monochrome zone ofpixels 2500M ofFIG. 5 j may include pixel positions of one column of pixels or of a plurality of columns of pixels. Likewise, the diagonally extending linear monochrome zones ofpixels 2500M ofFIG. 5 g may include pixel positions of a single diagonal row of pixels or alternatively, of a plurality of diagonal rows of pixels. - It will be seen that the versions of
image sensor array 182B shown inFIGS. 5 g-5 j are particularly well suited for use in picture taking optical readers which in bar code decoding applications are expected to decode linear bar code symbols. The image sensor arrays ofFIGS. 5 g-5 j may be referred to as linear symbol optimized image sensor arrays. As will be described in further detail herein, image data corresponding to monochrome zones ofpixels 2500M in the versions ofFIGS. 5 g-5 j can be selectively addressed and read out independently of image data from rows from colorsensitive zones 2500C of pixels. In bar code decoding applications,control circuit 552 may selectively address pixels ofmonochrome zones 2500M and read out image data from monochrome zones ofpixels 2500M as shown inFIGS. 5 g-5 i and transfer such image data to indicia decodecircuit 1702 for decoding of a linear bar code symbol. For picture taking applications,control circuit 552 may selectively address pixels of a color sensitive zone or zones ofpixels 2500C and selectively read out image data from color sensitive zone orzones 2500C and process such color image data into a visual display color frame of image data. The processing as will be explained further herein may include such steps as executing a demosaicing routine to convert color filter pattern image data into a visual display format and interpolation of color pixel values corresponding to the missing pixel positions at the pixel positions occupied by a monochrome zone orzones 2500M of pixels. - In
FIGS. 6 a-6 d, exploded physical form view of pixels ofimage sensor array monochrome pixel 250M ofimage sensor array 182B is shown inFIGS. 6 a and 6 b.Pixel 250M includes aphotodetector 302 which may be of photodiode or photogate construction, atransfer gate 304, a floatingdiffusion 306, areset transistor 307 includingreset gate 308, a rowselect transistor 309 including rowselect gate 310 and a sourcefollower amplifier transistor 311 includingamplifier gate 312. An important feature ofpixel 250M is opaqueoptical shield 316. Opaqueoptical shield 316, typically comprising metal, shields light rays from components ofpixel 250M other thanphotodetector 302. Accordingly, pixels from each of several rows ofimage sensor array 182A can be simultaneously exposed to light in a global electronic shutter operating mode without the light rays modulating charges stored in floatingdiffusion 306 or another storage region. Further aspects of image sensor arrays capable of operating in a global electronic shutter operating mode are described in U.S. patent application Ser. No. 11/077,975 incorporated herein by reference. Referring to additional aspects ofpixel 250M,pixel 250M includesmicrolens 320 which may be disposed on light transmissiveprotective layer 322.Microlens 320 collects light from a larger surface area thanphotodetector 302 and directs light towardphotodetector 302. - A color
sensitive pixel 250C ofimage sensor array 182B is described with reference toFIGS. 6 c and 6 d. Colorsensitive pixel 250C is similar in construction tomonochrome pixel 250M. Colorsensitive pixel 250C includes aphotodetector 302 which may be of photodiode or photogate construction, atransfer gate 304 for transferring charge fromphotodetector 250C, a floatingdiffusion 306, areset transistor 307 includingreset gate 308, a rowselect transistor 309 including rowselect gate 310 and a sourcefollower transistor amplifier 311 includingamplifier gate 312. Colorsensitive pixel 250C also includesopaque shield 320 which shields light from light sensitive components ofpixel 250C other thanphotodetector 302.Pixel 250C may also includemicrolens 320 for increasing the amount of light incident onphotodetector 302. In addition to the above elements, colorsensitive pixel 250C includes a wavelength selectivecolor filter element 260 formed thereon. Wavelength selectivecolor filter element 260 may be disposedintermediate microlens 320 andprotective layer 322. - A high level electrical block diagram of
image sensor array 182B is shown inFIG. 7 a.Image sensor array 182B may be of CMOS construction and may be an active pixel image sensor array such that eachpixel 250 ofimage sensor array 182B includes apixel amplifier 311. Eachpixel 250 of image sensor array may further have aphotosensitive region 252 and an optically shieldedstorage element 306.Image sensor array 182B further includes a two-dimensional grid ofinterconnects 262 which are in electrical communication withrespective column circuitry 270 androw circuitry 296.Row circuitry 296 andcolumn circuitry 270 enable such processing and operational tasks as selectively addressing pixels, decoding pixels, amplification of signals, analog-to-digital conversion, and applying timing, read out and reset signals. - Reset control lines of
interconnect grid 262 are shown inFIG. 7 b.Image sensor array 182B may have multiple sets of reset control lines so thatmonochrome pixels 250M ofimage sensor array 182B can be reset independently of colorsensitive pixels 250C ofimage sensor array 182B as described previously in connection with the description ofimage sensor array 182B. According to the invention,control lines 262R-M, 262R-C can be incorporated inimage sensor array 182B in a layered manner so that for a substantial portion ofimage sensor array 182B,control lines 262R-M have x, y positions that coincide with x, y positions ofcontrol line 262R-C (axes are defined inFIG. 8 a). Control lines 262R-C in the embodiment ofFIG. 7 b are installed at a different height (a different Z axis position) withinimage sensor array 182B relative to controllines 262R-C such that control lines 262R-M and 262R-C, for substantial length of the control lines, have common x, y positions. Installing the multiple control lines to be on top of one another so that the control lines have a common x, y axis position withinimage sensor array 182B reduces the amount of fill factor degradation which would otherwise result from installation of an additional set of reset control lines withinimage sensor array 182B. - Referring to
FIGS. 7 c and 7 d,image sensor array 182B may be configured to include separate and independent reset control lines for separately and independently resettingmonochrome pixels 250M and colorsensitive pixels 250C without increasing the overall thickness ofimage sensor array 182B. While disposing reset control lines on top of one another as described in connection withFIGS. 4 b and 7 b provides significant advantages; the image sensor array is made thicker with such arrangement which adds to manufacturing costs. Referring toFIG. 7 c, a version ofimage sensor array 182B is illustrated having a first set ofreset control lines monochrome pixels 250M and a second set ofreset control lines sensitive pixels 250C ofimage sensor array 182B. The reset control line configuration ofFIG. 7 c may be utilized with the color sensitive pixel distribution shown inFIG. 5 f to provide animage sensor array 182B having separate and independently controllable reset control lines for separately resettingmonochrome pixels 250M and colorsensitive pixels 250C and which exhibits a thickness equal to a thickness of a commonly available off-the-shelf image sensor array. In the version ofimage sensor array 182B shown inFIG. 7 c, the reset control lines of monochrome pixel rows are electrically connected together and the reset control lines of rows of pixels including color sensitive pixels are electrically connected together. The commonly connected reset control lines of the monochrome pixel rows are designated with thereference numeral reference numeral FIG. 5 f andFIG. 7 c, every fourth row of pixels ofimage sensor array 182B includes clusters of colorsensitive pixels FIG. 7 c, withreset control lines sensitive pixels 250C electrically connected together, all rows ofimage sensor array 182B including colorsensitive pixels 250C may be driven into reset by application of a reset control signal on commonreset control line monochrome pixels 250M (the monochrome row of pixels) can be driven into reset by applying a reset control signal on common monochrome pixel resetcontrol line image sensor array 182B shown inFIG. 7 c,monochrome pixels 250M ofimage sensor array 182B may be driven into reset whenpixels 250C are exposed for image data read out of color image data. - It is noted that with the configuration of
FIG. 7 c, adjacentmonochrome pixels 250M-A adjacent to an end pixel, e.g.,pixel 250C-E of a color sensitive pixel cluster, e.g.,cluster 257R are not driven into reset during exposure periods of colorsensitive pixels 250C. However, according to the invention in one example, image data corresponding only to acenter pixel 250C-I of each color sensitive horizontally arranged cluster (and not theend pixels 250C-E) may be selectively addressed during read out of color image data. The presence of each lateral color filter element at theend pixels 250C-E, which are not addressed for image data read out, reduces the effect of cross talk attributable to photons enteringimage sensor array 182B at an angle throughend pixels - Another configuration for providing separately and independently resetting
monochrome pixels 250M and colorsensitive pixels 250C ofimage sensor array 182B is shown and described with reference toFIG. 7 d. In the version ofFIG. 7 d,image sensor array 182B includes a plurality of rows of pixels including allmonochrome pixels 250M followed by a plurality of rows of pixels that include colorsensitive pixels 250C only. The monochrome rows ofpixels 250M form a first subset of pixels and the colorsensitive pixels 250C form a second subset of pixels. The reset control lines for resetting the first subset of pixels can be made separate and independent of the reset control lines for controlling the second subset of pixels by electrically connecting the reset control lines of the first subset of pixels together and then separately electrically connecting together the reset control lines of the second subset of pixels. The common control lines of the first subset ofmonochrome pixels 250M in the version ofFIG. 7 d are designated withreference numeral sensitive pixels 250C in the version ofFIG. 7 d are designated with thereference numeral FIG. 7 d facilitating separate and independent control ofmonochrome pixels 250M and colorsensitive pixels 250C can be utilized with the line art symbol optimized versions ofimage sensor array 182B shown and described inFIGS. 5 g-5 i having “zones” of monochrome or alternatively colorsensitive pixels 250C that extend entire rows ofimage sensor array 182B. - Referring to
FIG. 7 d, colorsensitive pixels 250C may be driven to reset during exposure periods formonochrome pixels 250M by application of a common reset control signal onreset control line sensitive pixels 250C for read out of color image data. Similarly colorsensitive pixels 250C may be driven into reset by application of a reset control signal on commonreset control line monochrome pixels 250M for read out of image data frommonochrome pixels 250M. - Features respecting specific embodiments of an image sensor array according to the invention have been described in connection with the views of
FIGS. 2 a-4 c (image sensor array 182A), and the views ofFIGS. 5 a-7 d (image sensor array 182B). General features of an image sensor array which may be incorporated into optical reader 100 (that is, features which can be incorporated in the image sensor array, whether of the embodiment labeled 182A, the embodiment labeled 182B, or another embodiment such as CMYimage sensor array 182C, RGBimage sensor array 182D, monochrome linearimage sensor array 182E, monochrome areaimage sensor array 182F, monochrome and polarizerimage sensor array 182G, or monochrome color and polarizerimage sensor array 182H) are now described. -
Optical reader 100 can be programmed or otherwise be configured to selectively address a first plurality of pixels in animage sensor array optical reader 100 selectively addresses the first subset of pixels and reads out image data from first subset of pixels independently of the second color sensitive subset of pixels. In another operating modeoptical reader 100 selectively addresses the second subset of pixels and reads out image data from the second subset of pixels independently of the first subset ofpixels 250M. Whereoptical reader 100 selectively addresses and selectively reads out only a subset of pixels of an image sensor array, the resulting frame of image data read out of the image sensor array may be referred to as a “windowed frame” of image data. When a windowed frame of image data is read out, the frame rate of the image sensor array is normally increased relative to a normal frame rate of the image sensor array. -
Image sensor array - When operated in a global electronic shutter operating mode, pixels from several rows of an image sensor array are exposed simultaneously. That is, when operated in a global electronic shutter operating mode, transistor components (for example, transfer
gates 304 and resetgates 308 of the array as shown in the embodiments ofFIGS. 3 a and 6 a) forming an electronic shutter an image sensor array are controlled in a coordinated manner so that a plurality of rows of pixels are exposed simultaneously and have common exposure periods. In a global electronic shutter operating mode, electronic shutter components of the array are controlled so that the common exposure period for each of the plurality of rows of pixels begins at a common exposure start time (via control of reset gates 308) and ends at a common exposure stop time (via control of transfer gates 304). As explained herein, each pixel of the array may store a charge in an optically shielded storage region during the common exposure period. For facilitating a global electronic shutter operating mode, an exposurecontrol timing pulse image sensor array FIGS. 15 a-15 e. An exposurecontrol timing pulse image sensor array control timing pulse control timing pulse -
Image sensor array optical reader 100 may be constructed to be operated in a rolling shutter mode of operation only; that is, in one specific embodiment an image sensor array ofoptical reader 100 can only be controlled to expose pixels of the image sensor array on a rolling shutter basis and cannot be controlled so that pixels of image sensor array are exposed on a global electronic shutter basis. In another embodiment, an image sensor array incorporated inoptical reader 100 is constructed to be operated in a global electronic shutter operational mode only and is incapable of operation in a rolling shutter mode. -
Image sensor array optical reader 100 is constructed to be operational in either of a rolling shutter operational mode or a global shutter operational mode, the switching between rolling shutter and global shutter operational modes may be made in response to a receipt of operator instructions to change the shutter mode. The switching between rolling shutter and global shutter operational modes may also be automatic and dynamic in response to the sensing of a predetermined criteria being satisfied. An image sensor array equippedoptical reader 100 having both rolling shutter and global shutter operational modes is described in U.S. patent application Ser. No. 11/077,975, filed Mar. 11, 2005 entitled, Bar Code Reading Device With Global Electronic Shutter Control, which is incorporated herein by reference. An image sensor array constructed to be operated in either of a rolling shutter or global shutter operational mode is described in U.S. Pat. No. 6,552,323 entitled, “Image Sensor With A Shared Output Signal Line” which is incorporated by reference. -
Image sensor array - It has been described with reference specifically to
image sensor array 182A, andimage sensor array 182B that it may be advantageous to incorporate into an image sensor array ofoptical reader 100 separately controllablereset control lines image sensor array optical reader 100 incorporating a cyan-magenta-yellow (CMY)image sensor array 182C as shown inFIG. 18 c, it may be advantageous to incorporate separate reset control lines for resetting magenta and cyan pixels separately from yellow pixels so that when yellow pixels are exposed for read out of decode frame yellow pixel image data for transmission to decode circuit 1792, the remainder of the pixels of the array, i.e., the cyan and magenta pixels can be set to reset to eliminate electron diffusion cross talk and to reduce photon penetration cross talk. When hand heldoptical reader 100 incorporates a hybrid monochrome and polarizerimage sensor array 182G as shown inFIG. 19 b or the monochrome color and polarizerimage sensor array 182H as shown inFIGS. 20 a and 20 b, it may be beneficial to incorporate intoimage sensor array 182 separately controllable reset control lines for controlling the reset of pixels external to the polarizing pixels so that when the polarizing pixels are exposed for read out of image data from the polarizing pixels, the remaining pixels of the image sensor array are set to reset to reduce cross talk from the pixels external to the polarizing pixels. - While an
image sensor array image sensor array image sensor array 182B is configured according to the version shown inFIG. 5 h having center rows of monochrome pixels defining amonochrome pixel zone 2500M and where the image sensor array is a CCD KAI-0340 image sensor array, image data from the center rows can be read out by selecting a preconfigured operating mode of the image sensor array chip. Additional “windowed frame” patterns can be selectively read out of a CCD image sensor array by varying the speed of a pixel clock timing control timing pulse that controls the speed with which a pixel is clocked. Invalid or null data can be clocked out of a CCD pixel by speeding up a pixel clock signal. Varying a pixel clock control signal between valid data yielding rates and invalid data yielding rates during the reading out of image data from a CCD image sensor array yields a windowed frame of image data comprising valid image data clocked out at normal speed and invalid image data clocked out at high speed. Image data can also be selectively read out of a CCD image sensor array by selectively gating to output circuitry of the CCD image sensor array image data corresponding to select pixels of the image sensor array. It will be seen that for any application described herein wherein a windowed frame of image data is read by selective addressing of pixels from a CMOS image array, a CCD image sensor array supporting windowing capability may be substituted therefore to provide selective read out functionality. - Additional aspects of the invention are described with reference to the physical form views of
FIGS. 8 a-8 c and the physical form views 9 a, 9 b and 9 c. In the physical views ofFIGS. 8 a-8 c, an imaging module onto which an image sensor chip may be incorporated is described. With reference toFIGS. 9 a, 9 b and 9 c, hand held housings for supporting and encapsulating an imaging module including an image sensor chip are described. - An
optical reader 100 of the invention, as shown in the embodiment ofFIGS. 8 a-8 c, may include an imaging module such asimaging module 1802A.Imaging module 1802A as shown inFIGS. 8 a-8 c incorporates certain features of an IT4000 imaging module herein and additional features. IT4000 imaging modules are available from Hand Held Products, Inc. of Skaneateles Falls, N.Y.Imaging module 1802A includesfirst circuit board 1804 carrying light sources 160 a, 160 b, whilesecond circuit board 1806 carries light sources 160 c, 160 d, 160 e, 160 f, 160 g, 160 h, 160 i, 160 j, 160 k, 160 l, 160 m, 160 n, 160 o, 160 p, 160 q, 160 r, 160 s, and 160 t (hereinafter 160 c through 160 t).First circuit board 1804 also carriesimage sensor array 182, which is integrated onto imagesensor IC chip 1082. Imagesensor IC chip 1082 andimage sensor array 182 inFIG. 8 a are generically labeled with the reference numerals “1082” and “182” respectively inFIGS. 8 a-8 d to indicate that any one of the specifically described imagesensor IC chips image sensor arrays imaging module 1802A. The various image sensor IC chips and image sensor arrays can also be incorporated in another imaging module described herein such asimaging module Imaging module 1802C shown inFIG. 8 e is a laser aiming IT4300 imaging module of the type available from Hand Held Products, Inc. The laser aiming IT4300 imaging module includes a plurality of illumination LEDs e.g.,LED 160, and an aiming pattern generator comprising alaser diode assembly 1872 in combination with adiffractive element 1873, wherein the diffractive element of the imaging module diffracts laser light from the laser diode assembly to project a two-dimensional aiming pattern onto a substrate, s.Imaging module 1802A also includessupport assembly 1810 includinglens holder 1812, which holdslens barrel 1814 that carriesimaging lens 212 that focuses images onto an active surface ofimage sensor array 182.Lens 212 may be e.g., a single lens (a lens singlet), a lens doublet or a lens triplet. Light sources 160 a, 160 b are aiming illumination light sources whereas light sources 160 c through 160 t are illumination light sources. Referring toFIG. 8 d, illumination light sources 160 c through 160 t project a two-dimensional illumination pattern 1830 over a substrate, s, that carries a decodable indicia such as abar code symbol 1835 whereas aiming illumination light sources 160 a, 160 b project an aimingpattern 1838. In the embodiments shown and described in connection withFIGS. 8 a-8 c, light from aiming illumination light sources 160 a, 160 b is shaped byslit apertures 1840 in combination withlenses 1842 which image slits 1840 onto substrate, s, to form aimingpattern 1838 which in the embodiment ofFIGS. 8 a-8 c is aline pattern 1838.Illumination pattern 1830 substantially corresponds to a full frame field of view ofimage reader 100 designated bybox 1850. The present field of view ofoptical reader 100 is herein referred to as the “target” ofoptical reader 100. Aimingpattern 1838 is in the form of a line that extends horizontally across a center of field of view ofimage reader 100.Illumination pattern 1830 may be projected when all of illumination light sources 160 c through 160 t are operated simultaneously.Illumination pattern 1830 may also be projected when a subset of light sources 160 c through 160 t are simultaneously energized.Illumination pattern 1830 may also be projected when only one of light sources 160 c through 160 t is energized such as LED 160 s or LED 160 t. LEDs 160 s and 160 t ofimaging module 1802 have a wider projection angle than LEDs 160 c through 160 t. In anoptical reader 100 incorporatingimaging module 1802 A illumination assembly 104 includes LEDs 160 a, 160 b, LEDs 160 c through 160 t and slitapertures 1840 in combination withlenses 1842. - A reader imaging module may be incorporated into one of a hand held housing as shown in
FIGS. 9 a, 9 b and 9 c. In the embodiment ofFIG. 9 a, hand heldhousing 101 is a gun style housing. In the embodiment ofFIG. 9 b, hand heldhousing 101 supportingimaging module 1802 is in the form factor of a portable data terminal (PDT). In the embodiment ofFIG. 9 c, hand heldhousing 101 supporting imaging module is in the form factor of a mobile telephone, often referred to as a “cell phone.” Whenoptical reader 100 is a cell phone,optical reader 100 is configured to send voice data over GSM/GPRS transceiver 571 to GSM/GPRS network 198 (FIG. 10 ) and to receive over GSM/GPRS transceiver 571 voice data from GSM/GPRS network 198. Further, whereoptical reader 100 is a cell phone,optical reader 100 may be configured so that an operator inputs telephone numbers viakeyboard 508. Thespecific imaging module 1802A described in connection withFIGS. 8 a-8 c may be incorporated in the optical reader shown inFIG. 9 a or theoptical reader 100 shown inFIG. 9 b or theoptical reader 100 shown inFIG. 9 c. However, in the embodiment shown inFIG. 9 a,housing 101 supports and encapsulatesimaging module 1802B an imaging module of construction similar toimaging module 1802A, except that only twolight sources 160 are incorporated into the imaging module.Housing 101 of the reader ofFIG. 9 b supportsimaging module 1802 which is generically labeledelement 1802 to indicate that any one of the specific imager modules described herein, e.g., 1802, 1802A, 1802B, 1802D, 1802E may be incorporated into an optical reader according to the invention. - Referring to further aspects of
optical reader 100,optical reader 100 may incorporate a graphical user interface (GUI) 3170 enabling selection between various operating modes. WithGUI 3170 an operator movespointer 3172 to a selected icon and clicks on the icon to configureoptical reader 100 in accordance with an operating mode associated with the selected icon.Reader 100 may include pointer mover 512 (otherwise termed a navigation matrix) to facilitate movement of thepointer 3172.Buttons 512B ofpointer mover 512 facilitate selection of an icon of a GUI interface that is supported by incorporating a multitasking operating system (OS) intoreader 100 such as WINDOWS CE.GUI 3172 may be developed using various open standard languages as HTML/Java or XML/Java. - In the embodiment of
FIG. 9 b,GUI 3170 includes a plurality ofvirtual selection buttons shutter icon 3152 configuresreader 100 so that during a next exposure periodimage sensor array 182 is operated in a rolling shutter mode. Selection ofglobal shutter icon 3154 configuresoptical reader 100 so that during a next exposure periodimage sensor array 182 is operated in a global electronic shutter mode. - Selection of
decode icon 3162 drivesoptical reader 100 into an indicia decode mode so that a next time a trigger signal is received,optical reader 100 captures a frame of image data and attempts to decode a bar code symbol or other decodable indicia (e.g., an OCR character) represented therein and outputs a decoded out message to display 504, or a spaced apartdevice 150, as is described with reference toFIG. 10 . Selection of image capture (which may otherwise be referred to a picture taking)icon 3164 configuresoptical reader 100 so that next time a trigger signal is received,optical reader 100 captures image data and outputs the image data to one or more of adisplay 504, a specified memory address, or to a spaced apartdevice 150 without attempting to decode decodable indicia therein.Optical reader 100 may also be constructed so thatoptical reader 100 can be configured in accordance with a selected operating mode by sending to reader 100 a serial command from a spaced apart device, or by the reading of a specially configured programming bar code symbol. -
Optical reader 100 is configured so thatoptical reader 100 receives a trigger signal whenmanual trigger 216 is manually depressed by an operator.Optical reader 100 may also be configured so that a trigger signal is received by the sensing of an object in the proximity ofreader 100 or by the sending of a serial trigger command toreader 100 from a spaced apart device, 150, as shown inFIG. 10 . - A flow diagram illustrating operation of
optical reader 100 in one embodiment is described with reference toFIGS. 14 a, 14 b, and 14 c. Atstep 1100 an operator selects between an indicia decode mode and a picture taking mode. Atstep 1100 an operator may select icon 3162 (FIG. 9 b) to driveoptical reader 100 into an indicia decode mode, or alternativelyicon 3164 to driveoptical reader 100 into a digital picture taking mode of operation. These modes may also be selected by sending to reader 100 a serial command from a spaced apartdevice 150 or by the reading of a programming bar code symbol. If an indicia decode mode of operation is selected,optical reader 100 executes anindicia decode process 1102. If a picture taking mode is selected,optical reader 100 executespicture taking process 1400. - An example of an
indicia decode process 1200 is described with reference toFIG. 14 b. At step 1202 a trigger signal is received by one of the methods described (depressing trigger 216, object sensing, serial trigger command) to commence a decode process. Atstep 1203,control circuit 552 ofoptical reader 100 captures a plurality of “parameter determination” or test frames of image data. The frames of image data captured atstep 1203 are not subject to indicia decode processing, but rather, are processed for parameter determination (e.g., exposure, gain, illumination). Alternatively,parameter determining step 1203 may be avoided. For example,control circuit 552 may apply parameters determined from a previous image capture operation rather than determining parameters atstep 1203. Atstep 1204control circuit 552 obtains a decode frame of image data details of which are explained herein. - For the capturing of frames of image data (i.e., “test” frames and/or frames for use in decoding, picture taking or other processing or storage) control circuit 552 (
FIG. 1 a) may send an illumination control signal toillumination assembly 104 and various image capture initiation control signals to control/timing circuit 1092 of image sensor chip 1082 (labeled generically to refer to any of the image sensor chips described herein). - Image capture initiation control signals are described in greater detail with reference to
FIGS. 15 a-15 e. For the capturing of image data,control circuit 552 may send toillumination assembly 104 an illuminationcontrol timing pulse 350 to energize at least onelight source 160 such thatillumination pattern 1830 is projected (as shown inFIG. 8 d).Control circuit 552 may also send to imagesensor IC chip 1082 an exposurecontrol timing pulse 354 and a read outcontrol timing pulse 368, and a reset control timing pulse 370 (that is,control circuit 552 sends appropriate signals to imagesensor IC chip 1082 to initiate exposurecontrol timing pulse 354, read outcontrol timing pulse 368 and reset control timing pulse 370). - In one embodiment as shown in
FIG. 15 a, the exposurecontrol timing pulse 354 begins after and finishes before the illuminationcontrol timing pulse 350. The read outcontrol timing pulse 368 begins at the conclusion of the illuminationcontrol timing pulse 350. In another embodiment as shown inFIG. 15 b, the illuminationcontrol timing pulse 350′ begins after and finishes before the exposurecontrol timing pulse 354′. In this embodiment, the read outcontrol timing pulse 368′ begins at the conclusion of the exposurecontrol timing pulse 354′. In further embodiments the exposure control timing pulse and the illumination control timing pulse overlap each other while occurring sequentially. In one such embodiment as shown inFIG. 15 c, this sequential operation can include the illuminationcontrol timing pulse 350″ starting, the exposurecontrol timing pulse 354″ starting, the illumination control timingsignal pulse 350″ ending, and then the exposurecontrol timing pulse 354″ ending. In this embodiment, the read outcontrol timing pulse 368″ begins at the conclusion of the exposurecontrol timing pulse 354″. In a further such embodiment as shown inFIG. 15 d, the sequential operation can include the exposurecontrol timing pulse 354″′ starting, the illuminationcontrol timing pulse 350″′ starting, the exposurecontrol timing pulse 354″′ ending, and then the illumination control timingsignal pulse 350″′ ending. In this embodiment, the read outcontrol timing pulse 368″′ begins at the conclusion of the illumination control timingsignal pulse 350″′. Each illuminationcontrol timing pulse FIG. 15 e. - When exposure
control timing pulse 354 is received by an image sensor IC chip andoptical reader 100 is configured in a global electronic shutter operating mode, pixels from several rows ofimage sensor array optical reader 100 is configured in an global electronic shutter operating mode, each of several rows ofimage sensor array image sensor array control timing pulse control timing pulse - When read out
control timing pulse 368 is received by imagesensor IC chip 1082B, image data is read out from the image sensor array. Image signals corresponding to pixels of the image sensor array are converted into digital form by analog-to-digital converter 1086 and transferred intomemory 560 byFPGA 580. -
Optical reader 100 may be configured so that atstep 1204 when reading out image data,optical reader 100 reads a “windowed frame” of image data. As indicated herein, a windowed frame of image data may be read out by selectively addressing pixels of a desired region of interest or window. A windowed frame of image data read out during frame capture atstep 1204 may comprise pixel values corresponding to all or substantially allmonochrome pixels 250M of an image sensor array. With further reference to the timing diagrams ofFIGS. 15 a, 15 b, 15 c and 15 d, a resetcontrol timing pulse 370 for resetting pixels that are not selectively addressed may be coordinated with the exposurecontrol timing pulse 354 for controlling exposure of pixels that are selectively addressed for reading out a windowed frame of image data. Thus, for reading out a monochrome frame of image data from a hybrid monochrome and color image sensor array, e.g.,image sensor array 182A orimage sensor array 182B, reset control timing pulse 3709 is applied to reset color pixels of theimage sensor array 182 while exposurecontrol timing pulse 354 is applied to enable exposure of monochrome pixels of the image sensor array. To facilitate resetting of color pixels of an image sensor array independent of resetting of monochrome pixels, an image sensor array may be configured to include a reset control line grid specifically adapted to enable resetting of color pixels. Applyingreset control pulse 370 to drive color pixels into reset while monochrome pixels are being exposed to light can be expected to eliminate electron diffusion cross talk and to reduce cross talk resulting from light rays angularly entering the color pixels during exposure. - When frames are obtained at
step 1204, they are obtained in a form suitable to facilitate indicia decoding such as bar code symbol decoding or OCR decoding. With the windowed frame of image data read out atstep 1204 from a hybrid monochrome and colorimage sensor array sensitive pixels 250C,control circuit 552 atstep 1204 may store gray scale values intoRAM 560, each pixel value representing an intensity of light at a particular monochrome pixel ofimage sensor array step 1204 may include e.g., 8 bit gray scale pixel values, 10 bit gray scale pixel values or 12 bit gray scale pixel values. Since numerous legacy bar code decoding and OCR decoding schemes are designed to operate on monochrome gray scale image data or binarized image data derived from gray scale image data, the selective addressing ofmonochrome pixels 250M in the capturing of a monochrome image frame yields a frame that is well suited for being subjected to indicia decoding processing. Of course, in certain applications,control circuit 552 atstep 1204 may obtain a decode frame of image data including color image data. For example, wheredecode circuit 1702 is configured to decode color encoded bar code symbols, it is advantageous forcontrol circuit 552 to obtain a decode frame of image data including color image data atstep 1204. - In the execution of
step 1204,control circuit 552 may carry out a plurality of alternative processes in obtaining a decode frame of image data. Referring to the flow diagram ofFIG. 14 d,optical reader 100 atstep 1204 may simply capture a single windowed frame of image data which has been described herein above. As indicated byprocess step 1205 ofFIG. 14 d,control circuit 552 may carry outprocess step 1204 by selectively addressingmonochrome pixels 250M of a hybrid monochrome and color image sensor array such asimage sensor array 182A orimage sensor array 182B and reading out image data frommonochrome pixels 250M only; i.e., a windowed frame of image data comprising image data frommonochrome pixels 250M only. - Referring to the flow diagram of
FIG. 14 e, the obtaining adecode frame step 1204 may be carried out in the alternative by the execution ofsteps step 1206optical reader 100 may generate a frame of image data that includes image data corresponding tomonochrome pixels 250M and colorsensitive pixels 250C and atstep 1207image sensor array step 1206 into gray scale values. The frame generated atstep 1206 may be generated by exposing color and monochrome pixels ofimage sensor array monochrome pixels image sensor array step 1206control circuit 552 ofoptical reader 100 may combine image data from two different frames such as two successive frames, wherein a first of the captured frame is a windowed frame of image data including image data from colorsensitive pixels 250C only and a second of the frames is a frame of image data including image data read out frommonochrome pixels 250M only. - Referring to the flow diagram of
FIG. 14 f,optical reader 100 may also obtain a decode frame atstep 1204 by carrying outsteps 1208 andstep 1209. Atstep 1208optical reader 100 may capture a windowed frame of image data including image data corresponding tomonochrome pixels 250M only and atstep 1209control circuit 552 may interpolate pixel values corresponding to color pixel positions ofimage sensor array step 1208. For example,control circuit 552 may capture a gray scalepixel value frame 5202 as illustrated inFIG. 16 a that includes a gray scale pixel value for each monochrome pixel position of theimage sensor array Optical reader 100 may interpolate a monochrome pixel value for any “missing pixel” color pixel position of theframe 5202. Referring toframe 5202,frame 5202 is a gray scale frame of image data captured by selecting reading out image data from animage sensor array 182B constructed in accordance withFIGS. 4 a-7 b (Period=2). - Pixel positions P11, P31, P51, P12, P22, P32, P42, P52, P13, P33, P63 . . . are pixel positions corresponding to
monochrome pixels 250M ofimage sensor array 182 for which individual frame image data has been read out. Pixel positions P21, P41, P23, P43, . . . are missing pixel positions corresponding to color sensitive pixels, 250C ofimage sensor array 182B. Referring to the frame of image data represented atFIG. 16 a anoptical reader 100 may calculate a gray scale pixel value for color pixel positions, e.g., position P23, by averaging the gray scale values for each pixel position that is adjacent to pixel position P23 and each pixel position that is corner adjacent to color pixel position P23. For example, referring to the frame represented inFIG. 16 a, a gray scale value for color pixel position P23 can be interpolated by averaging pixel values of pixel positions P12, P22, P32, P13, P33, P14, P24, P34. A pixel value for “missing pixel” position P23 can also be interpolated utilizing more than 8 neighboring pixel positions. Also, corner adjacent pixels may be weighted less than side, top or bottom adjacent pixels during averaging. In one simple averaging method, only four surrounding pixels are averaged; namely, the top and bottom adjacent pixels and the two side adjacent pixels adjacent to the pixel position for which a gray scale value is being interpolated. In a still further interpolation method, only two pixels are used for averaging; namely either the two side adjacent pixels adjacent to the pixel position being interpolated or the top and bottom adjacent pixels. A two-dimensional image representation of a linear bar code symbol can be expected to have several consecutive pixel positions along a column with similar gray scale values, if the representation of the symbol is oriented with 0° or 180° angle of rotation (i.e., the symbol is right side up or upside down). If the symbol representation has a 90° or 280° angle of rotation, several consecutive pixel positions along rows of pixel positions can be expected to have similar valued gray scale values. Accordingly, it can be seen that interpolating pixel values of adjacent pixel position values running in the direction of bars in a symbol representation yields truer edge information than utilizing all surrounding pixel positions for interpolation. - In one method of the invention, the correlation between a pair of horizontally oriented scan lines is calculated along with a correlation between a pair of vertically oriented scan lines. The two correlation measurements are then compared. If row scan lines are more closely correlated, column adjacent pixels are selected for interpolation. If column scan lines are more closely correlated, row adjacent pixels are selected for interpolation. An exemplary set of code for calculating a first derivative correlation for a pair of scan lines (horizontal or vertical) is presented by Table 1.
-
TABLE 1 Exemplary Code For Performing First Derivative Correlation Calculation %OneDcorelate: correlates two 1D 1st derivative signals to report the %correlation %input a,b: 1D array %output c: 1st derivative correlation function c=OneDcorrelate(a,b) % “diff” is the first derivative calculation. %for an input array a=[ai]n i=1 then diff(a) =[ai − ai+1]n−1 i=1 da=diff(double(a)); db=diff(double(b)); n=length(da); c=0; for i=1:n c=c+da(i)*db(i); end [End Table 1] - A set of code for interpolating missing color pixel position values by one of three methods (simple averaging, first derivative correlation, and simple correlations) wherein “M-set” refers to the monochrome set of pixels is presented in Table 2.
-
TABLE 2 Exemplary Code For Interpolating Missing Pixels Corresponding To Color Pixel Positions % MsetInterpolation: interpolates missing M-set pixels %input I_Mset: M-set image %input method: 1:first derivative correlation; 2: simple correlation; 3: %simple averaging %input p: sample period %output Im: interpolated monochrome image function Im=MsetInterpolation(I_Mset,method,p) Isz=size(I_Mset); %M-set topology % {circumflex over ( )} % MMMMMMMMM % MxMxMxMxM % MMMMMMMMM % MxMxMxMxM %(MMMMMMMMM) % v Im=double(I_Mset); m=Isz(1); n=Isz(2); %correlated averaging for i=p:p:m for j=p:p:n if i+1 <=m & j+1 <=n if method == 2 %simple correlation if abs(Im(i−1,j)−Im(i+1,j)) < abs(Im(i,j−1)−Im(i,j+1)) Im(i,j)=(Im(i−1,j)+Im(i+1,j))/2; else Im(i,j)=(Im(i,j−1)+Im(i,j+1))/2; end else if method==1 %first derivative correlation if OneDcorrelate(Im(i−1,j−1:j+1),Im(i+1,j−1 :j+1)) > OneDcorrelate(Im(i−1:i+1,j− 1),Im(i−1:i+1,j+1)) Im(i,j)=(Im(i−1,j)+Im(i+1,j))/2; else Im(i,j)=(Im(i,j−1)+Im(i,j+1))/2; end else %method==3 %simple averaging Im(i,j)=(Im(i−1,j)+Im(i+1,j)+Im(i,j−1)+Im(i,j+1))/4; end else if i+1 <=m & j+1 > n Im(i,j)=(Im(i−1,j)+Im(i+1,j))/2; else if i+1 > m & j+1 <=n Im(i,j)=(Im(i,j−1)+Im(i,j+1))/2; else if i+1 > m & j+1 > n Im(i,j)=(Im(i−1,j)+Im(i,j−1))/2; end end end Im=uint8(Im); [End Table 2] - At
step 1210,optical reader 100 transfers the frame of image data obtained atstep 1204, to anindicia decode circuit 1702 which may be a bar code symbol decoding circuit orautodiscrimination circuit 1704 including anindicia decode circuit 1702. In one embodiment,decode circuit 1702 decodes 1D and 2D bar code symbols and OCR characters.Autodiscrimination circuit 1704 may decode 1D and 2D bar code symbols and OCR characters (decodable indicia) and automatically discriminate between decodable indicia and handwritten characters. In the event thatautodiscrimination circuit 1704 recognizes the presence of handwritten character information,autodiscrimination circuit 1704 automatically outputs to display 504 and/or a spaced apartdevice 150 image data representing the handwritten character image data. Further details of indicia decodecircuit 1702 andautodiscrimination circuit 1704 are described in copending U.S. patent application Ser. No. 11/077,975, filed Mar. 11, 2005, incorporated by reference and U.S. application Ser. No. 10/958,779, filed Oct. 5, 2004, also incorporated herein by reference. - In general, indicia decoding accuracy is expected to increase with an increase in the percentage of monochrome pixels of
image sensor array image sensor array 182B having a period of P=2, every other row of pixels ofimage sensor array 182B are all monochrome pixels. Thus, horizontal scan lines can be launched through horizontal rows of pixel values of a frame of image data obtained utilizing a P=2image sensor array 182B during attempts to decode a linear bar code symbol without substantial reduction in performance relative to a frame obtained utilizing an all monochrome pixel image sensor array. For decoding linear bar code symbols,control circuit 552 may read image data along scan lines, such as scan lines defined by a horizontal row of pixel positions to determine the relative widths of bars and spaces of a symbol and then decode the symbol through table lookup to determine a set of decoded character data corresponding to the bar space information. - At
step 1212control circuit 552 receives a decoded output message fromdecode circuit 1702 orautodiscrimination circuit 1704. The message received bycontrol circuit 552 atstep 1212 may be e.g., a decoded bar code message or a set of decoded OCR characters. Atstep 1214optical reader 100 outputs a decoded out message. Atstep 1214control circuit 552 may send decoded out bar code data and/or decoded OCR data to display 504 or to a spaced apartdevice 150 or to a data storage memory location ofreader 100, orsystem 145 as described inFIG. 10 . - Examples of spaced apart
devices 150 which may be in communication withoptical reader 100 are shown and described in connection withFIG. 10 .Optical reader 100 may be part of asystem 145 and may be included in a local area network (LAN) 170 which comprises, in addition toreader 100, such spaced apart devices as otherportable readers 100′, 100″,network access point 174, personal computers 172 and central server 176 that are spaced apart from hand heldhousing 101 ofreader 100, all of which are connected together viabackbone 177. Server 176 in turn is in communication with a variety of additional spaced apartdevices 150 that are spaced apart from hand heldhousing 101 ofreader 100 and which through server 176 are in communication withoptical reader 100. Server 176 may be connected viagateways network 181 to a first distant remotelocal area network 185 located miles to thousands of miles away fromlocal area network 170 and a second distantlocal area network 2170 also located miles to thousands of miles away from local area network.Network 170 may be located at a supplier's warehouse.Network 2170 may be located at a delivery destination; andnetwork 185 may be located at a data processing/data archiving facility.Network 185 can be configured to assemble, store and maintain in server 184 various web pages accessible with use ofoptical reader 100, that summarize data that has been collected by variousoptical readers remote network 185 viaprivate communication line 190.IP network 181 may be the Internet or a virtual private network (VPN).Remote LAN 185 may include a personal computer 186 and a remote server 184 connected viabackbone 191.Remote LAN 185 may also include a wirelesscommunication access point 193.Remote LAN 185 may also include a personal data assistant (PDA) 189.Remote LAN 2170 may include aserver 2176, connected toIP network 181 via gateway 2179, backbone 2177,access point 2174, PC 2172, andoptical reader 100, 100R.System 145 may be configured so that a display equipped device, e.g.,device 100′, 172, 186, 189 automatically displays data, such as decoded out bar code message data of a visual display color image frame of image data, received fromoptical reader 100 on its associateddisplay 1504 when receiving that data. - All of the steps of
process 1200 are carried out automatically in response to the receipt of a trigger signal. The steps ofprocess 1200 continue automatically until a stop condition is satisfied. A stop condition may be e.g., the receipt of a trigger stop signal as may be generated by release oftrigger 216 or the successful decoding of a predetermined number of bar code symbols. As indicated byreturn line 1211 ofFIG. 14 b,control circuit 552 may repeatedly attempt to obtain image data and attempt to decode decodable indicia therein until a stop condition is satisfied. - Interpolating monochrome pixel values for “missing pixels” pixel positions is particularly advantageous where a hybrid monochrome and color image sensor array according to the invention includes a high number of color sensitive pixels distributed at spaced apart pixel positions throughout image sensor array. In other instances as alluded to previously in connection with the flow diagram of
FIG. 14 b,control circuit 552 may obtain a decode frame of image data atstep 1204 that is suitable for transferring to decodecircuit 1702 by simply reading from an image sensor array image data frommonochrome pixels 250M without interpolation of any pixel values at “missing pixel” pixel positions where an image sensor array ofreader 100 is constructed in accordance withimage sensor array 182A, andmonochrome pixels 250M form a complete checkerboard pattern, (there are no “missing” monochrome pixels positions in the M×N matrix of monochrome pixels withinimage sensor array 182A). Accordingly, whereoptical reader 100 includesimage sensor 182A, the decode frame of image data atstep 1204 is advantageously obtained by reading out fromimage sensor array 182A image data frommonochrome pixels 250M without interpolation of any further monochrome pixel values. - It may also be useful to obtain a decode frame of image data at
step 1204 without interpolation of monochrome pixel values whereoptical reader 100 includes a linear bar code symbol optimizedimage sensor array 182B of one of the versions described in connection withFIGS. 5 g-5 j. In the versions ofimage sensor array 182B shown and described in connection withFIGS. 5 g-5 j,image sensor array 182B includes at least one “zone” ofmonochrome pixels 2500M and at least one “zone” of colorsensitive pixels 2500C. Monochrome zone ofpixels 2500M of a linear symbol decoding optimized version ofimage sensor array 182B generally comprises an elongated line of monochrome ofpixels 250M having a minor dimension of one to a limited number of rows of pixels (which may be diagonal rows). Whereoptical reader 100 includes a linear symbol optimized version ofimage sensor array 182B a reduced area decode frame of image data atstep 1204 without any “missing pixel” pixel positions can be obtained by selectively addressing pixels ofmonochrome zone 2500M and selectively reading out of image data from the pixels ofmonochrome zone 2500M without reading any image data from the pixels of color sensitive zone ofpixels 2500C. More specifically, whereoptical reader 100 comprises a linear symbol optimized version ofimage sensor array 182B,control circuit 552 atstep 1204 may obtain a reduced area monochrome frame of image data at step 1204 (FIG. 14 b) and then transfer the reduced area monochrome frame of image data to decodecircuit 1702 atstep 1210. A reduced area frame of image data is further explained with reference toFIG. 11 illustrating an application where anoptical reader 100 is utilized to collect decoded bar code data and image data from aparcel 1260 carrying various bar code symbols, e.g., linearbar code symbol 1266 and two-dimensionalbar code symbol 1270. A full area frame of image data that may be obtained byoptical reader 100 represents the scene area indicated byrectangular border 1276 ofFIG. 11 . Where image data from all pixels ofimage sensor array 182B are read out, a full area frame of image data is obtained. Whenoptical reader 100 obtains a reduced area frame of image data, a reduced area frame of image data representing the reduced scene area as indicated by dashed-inborder 1278 may be obtained. In the example ofFIG. 11 ,optical reader 100 may incorporate one of the linear symbol optimizedimage sensor arrays 182B as shown inFIGS. 5 g and 5 h. A reduced area frame of image data representing the reducedscene area 1278 may be obtained by reading out image data from monochrome zone of thin centerline monochrome zone 2500M ofimage sensor array 182B according to one of the versions ofFIGS. 5 g and 5 h. It is seen with reference toFIG. 11 that whenoptical reader 100 obtains a reduced area frame of image data atstep 1204 representing the reducedscene area 1278, the reduced area frame of image data, though reduced, can be of sufficient size to include a representation of linearbar code symbol 1266 carried byparcel 1260.Imaging module 1802 such asmodule 1802A (FIG. 8 a) ofreader 100 can be constructed so that aiming pattern 1838 (FIG. 8 d) is projected ontoscene area 1278 at expected reading angles, while aiming light sources 160 a, 160 b, and remaining light sources 160 c-160 t are energized simultaneously during the time that pixels ofzone 2500M are exposed for read out of image data fromzone 2500M. Simultaneously projecting aimingpattern 1838 andillumination pattern 1830 ontoscene area 1278 improves the signal strength of image data corresponding to pixels ofzone 2500M. After receiving the thin line reduced area frame of image data atstep 1210,decode circuit 1702 may process the thin line reduced area decode frame of image data to decode linearbar code symbol 1266 by calculating the bar space widths of the bars and spaces of linearbar code symbol 1266 and then determining the characters of the symbol through table lookup. In a further aspect,optical reader 100 may be configured so that aiming pattern 1838 (FIG. 8 d) is projected horizontally at a center of a field ofview 1276 ofoptical reader 100 to coincide witharea 1278 represented by the reduced area image obtained atstep 1204 to aid an operator in obtaining an image that includes a representation of linearbar code symbol 1266. The frame rate ofoptical reader 100 when obtaining the reduced area decode frame of image data atstep 1204 may be significantly reduced relative to the frame rate ofoptical reader 100 when obtaining a full frame of image data. Accordingly, a method of the invention whereoptical reader 100 atstep 1204 obtains a reduced area frame of image data which is transferred to decodecircuit 1702 is optimized for fast (“snappy”) decoding. As has been described herein, colorsensitive pixels 250C may be set to reset whilemonochrome pixels 250M are exposed for selective read out of image data frommonochrome pixels 250M. - With further reference to the application view of
FIG. 11 , it is seen that the reduced area frame of image data representingscene area 1278 may not include a complete representation of linearbar code symbol 1266 and it is further seen thatparcel 1260 may include or carry additional bar code symbols such as two-dimensionalbar code symbol 1270 that is part ofpostal area 1268 ofparcel 1260. According to the invention in another aspect,optical reader 100 can be configured so that where indicia decodecircuit 1702 cannot successfully decode a bar code symbol via processing of a reduced area frame of image data or wherecontrol circuit 552 is programmed to search and decode multiple bar code symbols,control circuit 552 executes return line (FIG. 14 b) to re-execute the obtaining of a decode frame of image data atstep 1204. However, whencontrol circuit 552 executesstep 1204 the second time,control circuit 552 captures a frame of image data that represents a scene area that is larger than the scene area represented by the frame obtained during the first execution ofstep 1204. The decode frame of image data obtained by a second execution ofstep 1204 may be a full area image data frame representing the full field of view ofoptical reader 100 indicated by dashed-inborder 1276 ofFIG. 11 . Wherecolor zones 2500C ofimage sensor array 182B are distributed in a Bayer pattern,control circuit 552 during the second execution of obtainstep 1204 may selectively read out image data from the green pixels of color sensitive zones of image sensor array of 2500C and interpolate green pixels values at non-green pixel positions utilizing the green pixel values so that the decode frame of image data obtained atstep 1204 includes all green pixel values. Further, the missing pixel positions corresponding tomonochrome zone 2500M can be filled in utilizing the image data obtained during the previous execution ofstep 1204 as scaled based on a relationship between the color scale values of pixels corresponding to zone 2500M and the color scale values ofpixels surrounding zone 2500M. Atstep 1210, larger area green image data is transferred to indicia decodecircuit 1702. Indicia decodecircuit 1702 may attempt to decode linearbar code symbol 1266 and all other bar code symbols such as two-dimensionalbar code symbol 1270 that may be represented in the image obtained during the second execution ofstep 1204. With reference to the application view ofFIG. 11 ,optical reader 100 incorporating a linear symbol decode optimized to imagesensor array 182B may attempt to decodelinear symbol 1266 utilizing small areaimage representing area 1278 and then subsequently attempt to decode a two-dimensional bar code symbol, e.g.,symbol 1270, utilizing a larger area frame of image data representingscene area 1276. It will be seen that the method described wherecontrol circuit 552 obtains a reduced area frame of image data, attempts to decode, then subsequently obtains a larger frame of image data and attempts to decode utilizing the larger image may be practiced utilizing an “all monochrome”image sensor array 182F as shown and described in connection withFIG. 17 b. Whereoptical reader 100 incorporates an all monochromeimage sensor array 182F as shown inFIG. 17 b, it is particularly useful to setmonochrome pixels 250M to reset that are not being selected for read out of a reduced area decode frame of image data atstep 1204 during exposure periods for selectedmonochrome pixels 250M that are being selectively addressed for image data read out. -
Monochrome pixels 250M transmit more light than colorsensitive pixels 250C. Therefore, resettingmonochrome pixels 250M that are not selectively addressed and which are adjacent to a region of interest during an exposure period can be expected to have a greater benefit in terms of improving the overall signal to noise ratio ofreader 100 than resetting colorsensitive pixels 250C that are adjacent to a region of interest during an exposure period. - With still further reference to the application view of
FIG. 11 , it may be advantageous to utilizeoptical reader 100 to obtain a visual display color frame of imagedata representing parcel 1260. For example,parcel 1260 may include a damagedarea 1272. Obtaining a visual display color frame of image data corresponding to parcel 1260 creates a record documenting parcel damage. Referring to the application view ofFIG. 12 a, differentoptical readers 100 and 100R at different locations A and B located miles apart along a delivery route may be utilized to document physical transformations ofparcel 1260 whenparcel 1260 is carried along a delivery route.Optical reader 100 at location A including LAN 170 (FIG. 10 ) may be operated to obtain a visual display color frame of image data ofparcel 1260 whenparcel 1260 is located at location A. Further, the color frame may automatically be transferred to remote server 184 (FIG. 10 ) having adatabase 187 of color frames of image data that are indexed by a parcel identifier decoded in a parcelbar code symbol 1266 which identifier is also transmitted to remote server 184 automatically whenoptical reader 100 readsbar code symbol 1266. At location B remoteoptical reader 100, 100R (FIG. 10 ) may be utilized to again decodebar code symbol 1266 and capture visual display color frame of imagedata representing parcel 1266 and automatically transfer the parcel identifier corresponding tobar code 1266 and the color frame of image data to remote server 184. With reference to the application view ofFIG. 12 a the color frame of image data transmitted to remote server 184 from location B will include a representation of damagedarea 1272 that is not included in the color frame of image data transmitted to remote server 184 from location A. Accordingly, a person (for example, at PC 172 viewing web pages of server 184) reviewing the parcel identifier indexed color frame data ofdatabase 187 can determine that the damage to parcel 1260 occurred during the time that the parcel was delivered from location A to location B. Referring toFIG. 12 b,optical reader 100 can also be utilized to take color pictures of adelivery vehicle 1282 that carriedparcel 1260 from location A to location B. In the example ofFIG. 12 b, a picture being taken byoptical reader 100 has the field of view indicated byrectangle 1286. The field of view encompassesparcel 1260, anddelivery vehicle 1282, including alicense plate 1284.Trigger 216 can be actuated a first time to decodebar code symbols 1266, 1267 and then an additional time or times to have a picture ofparcel 1260 and/orvehicle 1272 including a picture oflicense plate 1284. The decoded bar code data and multiple color frames of image data may be associated with one another into a single transaction data set, and then via a packet based transmission scheme, the transaction data set may be sent to remote server 184, which may organize the data into viewable web pages viewable at PC 172.Optical reader 100, which may be incorporated in hand heldhousing 101, can be configured so that all of the data of the transaction data set is sent to remote server 184 in response to a single command input tooptical reader 100 via a user interface of reader 100 (e.g., 3170). Further aspects ofoptical reader 100 operating in a picture taking mode of operation are described with reference to the flow diagrams ofFIGS. 14 c, 14 g and 14 h. - Referring again to the flow diagram of
FIG. 14 a, a digitalpicture taking process 1400 is executed whenoptical reader 100 is configured to operate in a picture taking mode of operation. Atstep 1100, a picture taking mode of operation may be selected, e.g., by clicking on “image capture” icon 3164 (FIG. 9 b) and atstep 1104picture taking process 1400 is executed. - Referring to the steps of
picture taking process 1400,optical reader 100 atstep 1402 receives a trigger signal as may be generated e.g., by depression of a manual trigger an object in the proximity ofreader 100 being sensed or receipt of a serial command. Atstep 1403control circuit 552 captures a plurality of “test” or parameter determination frames of image data. The frames of image data captured atstep 1403 are not output for visual display; but rather are processed in order to determine operational parameters (exposure setting, gain illumination). Alternatively,step 1404 can be avoided andcontrol circuit 552 can instead load operational parameters that were derived during a past image capture operation. Atstep 1404control circuit 552 obtains a “visual display” image frame of image data. A visual display color frame of image data is one that is generated for visual display on a display and may include three color scale values for each pixel position of the plurality of pixel positions of the frame. A visual display frame after being obtained is sent to a display for visual display of an image or to a memory location for future display. In the embodiment ofFIG. 14 c, the image data obtained atstep 1404 is not transferred to decodecircuit 1702. - An image captured as part of obtaining at
step 1404 may be one that is captured in accordance with the timing diagrams ofFIGS. 15 a-15 e. In an alternative embodiment, the control signals input into imagesensor IC chip 1082 for the capture of a frame of image data may not include illumination control timing pulse e.g.,pulse 350. In many applications an object subject to image capture byoptical reader 100 during a picture taking mode will be a long range image (an object will be subject to image capture is one that is one to several feet from image reader). Light fromlight sources 160 may have little affect on an image captured that corresponds to a long range object; thus,optical reader 100, in one embodiment may not send an illumination control timing pulse atstep 1404. - However, depending on the application, it may be desirable to increase the illumination intensity of
optical reader 100 during capture of color image data relative to the intensity during capture of monochrome image data to compensate for the signal reduction affect ofcolor filter elements optical reader 100 can have a plurality of operator selectable configuration settings.Optical reader 100 can be configured so that activation of button 3150 toggles through a sequence of options one of which may be selected by actuation of a key ofkeyboard 508. As shown by Table 3, where e=exposure, g=gain, and i=illumination intensity, a selection of a configuration setting can result in a differentiation between the imaging parameters ofreader 100 during read out of monochrome image data atstep 1204 and the imaging parameters ofreader 100 during read out of color image data atstep 1404. Configuration setting 1 is a base line setting wherein there is no differentiation between monochrome read out and color image data read out imaging parameters. Configuration setting 2 has been described above. With configuration setting 2, there is no illumination during read out of color image data atstep 1404. Configuration setting 3 has also been described above. With configuration setting 3, illumination intensity is increased for read out of color image data. With configuration setting 4, illumination intensity for read out of monochrome image data can be increased. For example, as described herein,illumination pattern 1830 and aimingpattern 1838 can be projected simultaneously during read out of monochrome image data corresponding to amonochrome zone 2500M of pixels. With configuration setting 5, exposure time is boosted for read out of color image data and with configuration setting 6, gain is boosted for read out of color image data. Configuration setting 3 is highly useful whereoptical reader 100 includes a long distance flashillumination light source 160, 160X or whereoptical reader 100 is used for picture taking at close range. -
TABLE 3 Imaging Parameters When Imaging Parameters When Reading Out Monochrome Reading Out Color Image Image Data At Decode Frame Data At Visual Display Configuration Obtain Step 1204Obtain Step 14041 e = e0 e = e0 g = g0 g = g0 i = i0 i = i0 2 e = e0 e = e0 g = g0 g = g0 i = i0 i = 0 (Illumination Off) 3 e = e0 e = e0 g = g0 g = g0 i = i0 i = i1, i1 > i0 4 e = e0 e = e0 g = g0 g = g0 i = i2, i2 > i0 i = i0 5 e = e0 e = e1, e1 > e0 g = g0 g = g0 i = i0 i = i0 6 e = e0 e = e0 g = g0 g = g1, g1 > g0 i = i0 i = i0 - In executing the obtaining visual display color frame of
image data step 1404,optical reader 100 may carry out a variety of alternative processes. With reference to the flow diagram ofFIG. 14 g, a process is described whereinoptical reader 100 may obtain a visual display color frame of image data utilizing image data read out from colorsensitive pixels 250C only. With reference to the flow diagram ofFIG. 14 h, a process is described whereincontrol circuit 552 obtains a visual display color frame of image data utilizing image data derived by reading out of image data from both monochrome pixels and color sensitive pixels ofimage sensor array 182. - Referring to the flow diagram of
FIG. 14 g,control circuit 552 atstep 1405 captures a windowed frame of image data by selectively addressingcolor pixels 250C and by selectively reading out image data fromcolor pixels 250C ofimage sensor array image sensor array monochrome pixels 250M independent of colorsensitive pixels 250C. Atstep 1405 while color sensitive pixels are exposed for read out of image data,monochrome pixels 250M may be reset with use of a resetcontrol timing pulse FIGS. 15 a-15 d). Coordinating a resetcontrol timing pulse control timing pulse sensitive pixels monochrome pixels 250M, i.e., by eliminating electron diffusion cross talk and by reducing cross talk attributable to light rays angularly penetrating throughmonochrome 250M. - At
step 1406,optical reader 100 automatically transfers the color filter array image data frame captured atstep 1405 to a demosaicing circuit 1706 (FIG. 1 e). Taking as an input a color filter array image data frame,demosaicing circuit 1706 outputs a visual display color frame of image data. Wheredisplay 504 is a color display configured to receive red, green and blue (RGB) signals for each pixel ofdisplay 504,demosaicing circuit 1706 atstep 1406 may generate RGB color scale values for each pixel ofdisplay 504 so that a frame output bydemosaicing circuit 1706 is compatible withdisplay 504. The color scale values may comprise e.g., 8 bits, 10 bits, or 12 bits of data. Atstep 1407,optical reader 100 receives a visual display color frame of image data fromdemosaicing circuit 1706. - A particular example of
optical reader 100 executingstep 1404 is described with reference toFIG. 16 b. Atstep 1406 whereoptical reader 100 includes a hybrid monochrome colorimage sensor array FIG. 2 c andFIG. 5 a,optical reader 100 may read out an RGB Bayer pattern frame of image data as shown inFIG. 16 b. Where a reader image sensor array is provided byimage sensor array 182B including a 1280×1024, array of pixels, with a 320×256 subset array (P=4) of colorsensitive pixels array 182Boptical reader 100 atstep 1405 captures a 320×256 Bayer pattern of pixels.Demosaicing circuit 170 processes the Bayer pattern frame 1502 as shown inFIG. 16 b to output a visual display color frame of image data including a 320×256 color image, wherein each pixel of the frame includes a red color scale value, a green color scale value, and a blue color scale value. In such an embodiment,demosaicing circuit 1706, for each pixel of the Bayer pattern color filter arrayimage data frame 5204, interpolates red, green, and blue values. Referring to frame 5204 shown inFIG. 16 b,optical reader 100 determines a red value for red pixel position P32 simply by reading the color scale value of pixel position P32. Optical reader 100 determines a green value for red pixel P32 by averaging the values of green pixel positions P31, P22, P42 and P33. Optical reader 100 may interpolate a blue value for red pixel position P32 by averaging the values of blue pixel positions P14 P41, P23, P43. It will be seen that red, green, and blue values can be determined for each pixel position interpolating the pixel values as necessary. With increased processing speeds, it is possible to utilize dozens or more surrounding pixel values for interpolation of a red, green, or blue pixel for each pixel position. - In another aspect of the invention, the accuracy with which color scale values for each pixel position may be interpolated can be enhanced by utilizing monochrome pixel values in the color scale value interpolation process. Referring to red pixel position P32 of
frame 5204, it has been described that color scale values at green pixel positions P31, P22, P42, P33 may be averaged for interpolating a green color scale value at pixel position P32. In another method, monochrome pixel values at positions P33, P22, P42, P33 may be utilized for enhancing the interpolation of a green pixel value at position P32. The monochrome pixel values at positions P33, P22, P42, P33 may be interpolated from monochrome pixel values by one of the monochrome pixel interpolation methods described herein. Then, the color scale value at each pixel position, P32, P22, P42, P33 may be offset by a value, Delta, equal to the difference between the interpolated monochrome pixel values at the position being interpolated and the monochrome pixel value at the position contributing to the interpolation calculation. Thus, a green color scale value at position P32 may be calculated according to Eq. A. -
- Where Delta31=M(P32)−M(P31), Delta22=M(P32)−M(P22), Delta42=M(P32)−M(P42), Delta33=M(P32−M(P33).
- Similarly, a blue color scale value at position P42 may be interpolated according to the formula of Equation B.
-
- An exemplary algorithm for interpolating a color scale value at a color pixel position utilizing monochrome pixel values is presented in Table 4 where “C-set” refers to color pixel values and “M-set” refers to monochrome pixel values.
-
TABLE 4 Algorithm For Interpolating Color Scale Values Utilizing Monochrome Image Data 1) For each color pixel C for interpolation, select the missing color neighborhood C-set pixel values Ci and select the corresponding neighborhood M-set pixel values Mi. Select corresponding M-set pixel value M to color pixel C. 2) let C=0 3) for i=1 to n where n is the number of neighborhood pixel Ci 4) C=C+Ci+M−Mi 5) end 6) C=C/n - Regarding
step 1, it is noted that there will normally be two neighborhood color or “C-set” pixels where blue or red values are interpolated at a green pixel position, and in other cases four neighborhood color pixels. - Another particular example of
optical reader 100 executingsteps FIG. 16 c. Where a reader image sensor array is provided byimage sensor array 182B including 1280×1024 array of pixels, and a period P=4 of color sensitive pixels formed with a Cyan-Magenta (Cy-Mg, or “CM”) color filter array as shown inFIG. 5 b,optical reader 100 atstep 1405 reads out a colorfilter array frame 5206 as shown inFIG. 16 c. Color filter arrayimage data frame 5206 includes a 320×256 pattern of Cy-Mg pixel values.Demosaicing circuit 1706 may processimage data frame 5206 into a visual display frame such as a visual display color frame of image data where each pixel position offrame 5206 is represented by a combination of red, green and blue values. In processing the Cy-Mg visualdisplay color frame 5206 into a frame of image data including red, green, and blue values for each pixel position,optical reader 100 may first calculate white, cyan and magenta values for each pixel position offrame 5206. Where an original pixel position such as pixel position P53 (FIG. 16 c) is a cyan pixel, the cyan value is determined by directly reading the pixel value of the cyan pixel. A magenta value for cyan pixel at position P53 is calculated by interpolation utilizing the magenta values of surrounding pixel positions of magenta pixels such as positions P52, P43, P63, P54 (FIG. 16 c). A white value for cyan pixel at position P35 is calculated by interpolation using pixel values from monochrome pixel positions that surround cyan pixel P53. Referring toFIG. 5 b, a supplemental frame including monochrome pixel values may be captured, e.g., successively before or afterframe 5206 is captured for purposes of interpolating a white value for each color pixel of the color filter arraywindowed frame 5206. Alternatively, the colorfilter array frame 5206 captured atstep 1405 may include monochrome pixel image data for purposes of interpolating a white value for each color pixel value. When white, cyan and magenta values are calculated for each pixel offrame 5206, the white, cyan, and magenta values are readily converted into red, green, and blue values. Alternatively, display 504 can be configured to be responsive to white, cyan and magenta signals for each pixel ofdisplay 504. A set of transform equations for transforming a set of white, cyan and magenta values for a given pixel of a frame into a set of red, green and blue values for that pixel is given as follows. -
R=W−Cy (Eq. 1) -
G=Mg+Cy−W (Eq. 2) -
B=W−Mg (Eq. 3) - In the process described relative to the flow diagram of
FIG. 14 g, an original color filter array frame is processed into a visual display color frame of image data of reduced spatial resolution (a reducedspatial resolution 320×256 visual display color frame of image data may be produced using a hybrid monochrome and color image sensor array having a 1280×1024 pixel resolution). With reference toFIG. 14 h, a process for producing a high spatial resolution visual display color image is described. In the process described relative to the flow diagram ofFIG. 14 h,optical reader 100 utilizes image data from bothmonochrome pixels 250M andcolor pixels 250C from a hybrid monochrome and color image sensor array such asimage sensor array 182A orimage sensor array 182B in the generation of a visual display color image having spatial resolution equal to or on the order of the overall pixel resolution of the image sensor array. - At
step 1408control circuit 552 captures a color filter array image data frame by selectively addressingcolor pixels 250C of an image sensor array and selectively reading out image data from colorsensitive pixels 250M. The frame of image data captured atstep 1408 is a windowed frame of image data. For reduction of cross talk resulting from light enteringmonochrome pixels 250M, the monochrome pixels ofimage sensor array control timing pulse control timing pulse step 1408. - At
step 1409control circuit 552 captures a monochrome frame of image data by selectively addressingmonochrome pixels 250M ofarray step 1409 is a windowed frame of image data. For reduction of cross-talk resulting from light enteringcolor pixels 250C the color pixels ofimage sensor array 182 may be reset using resetcontrol timing pulse control timing pulse step 1409. - At
step 1410control circuit 552 transfers both the color filter array frame captured atstep 1408 and the monochrome image frame captured atstep 1409 tofusion circuit 1708.Fusion circuit 1708 takes as inputs the color filter array image data frame and the monochrome image data frame and processes them into a high resolution visual display color frame of image data. - Referring to
FIG. 14 i, an example of theprocess 1440 which may be executed by fusion circuit 1708 (FIG. 1 e) to process a combination of monochrome image data and color image data into a visual display color frame of image data is described. As explained with reference toFIG. 1 e,fusion circuit 1708 may be physically embodied by the combination of a control circuit provided by aCPU 552 operating in combination withmemory 566 that stores an executable program. The specific process described with reference toFIG. 14 i is executed utilizing anoptical reader 100 including substantially uniform dimensional pixelimage sensor array 182B. Atstep 1442 ofprocess 1440control circuit 552 generates color filter array image data and monochrome gray scale image data. Whereoptical reader 100 includesimage sensor array 182B,control circuit 552 may executestep 1442 by reading out fromimage sensor array 182B a single frame of image data comprising both monochrome image data and color image data.Control circuit 552 may also executestep 1442 by successively capturing a first monochrome frame comprising monochrome image data and then a second color frame comprising color image data.Control circuit 552 atstep 1442 may drivemonochrome pixels 250M into reset during an exposure period for read out of color image data from colorsensitive pixels 250C. When generating a frame of monochrome image data atstep 1442, control circuit may interpolate monochrome pixel values for “missing pixel” positions occupied by colorsensitive pixels 250C. - At
step 1446control circuit 552 generates an RGB image having resolution equal to the color sensitive subset of pixels ofimage sensor array 182B. In an RGB image, each pixel of the image is represented by a red color value, a green color value and a blue color value. The RGB image generated atstep 1446 may have the same characteristics as the visual display image received byoptical reader 100 atstep 1407 of the alternative process described in connection withFIG. 14 g. Where a color filter array image captured atstep 1442 is a Bayer pattern image, the RGB image generated atstep 1446 is derived by executing a demosaicing routine as described herein. Where the color filter array image captured atstep 1442 is a CMY image or a CM image (cyan and magenta only) image as described in connection withFIGS. 2 b and 2 d, the RGB image generated atstep 1446 is derived by way of a transformation process as described herein in connection withequations fusion circuit 1708control circuit 552 atstep 1450 expands the pixel count of the RGB image generated atstep 1446 so that the pixel count of the color image is equal to the pixel count of the monochrome image captured at step 1442 (atstep 1442 monochrome pixels from the captured monochrome image may be interpolated as described withFIG. 16 a). Whencontrol circuit 552 executesstep 1450, the monochrome gray scale image generated atstep 1442 and the color image at that stage of the processing have equal numbers of pixels such that each pixel position e.g., pixel of the monochrome image has a corresponding pixel position in the color image. With reference toimage sensor array 182B having a period of P=2, there are four times as many monochrome pixels as there are color sensitive pixels. Accordingly, withimage sensor array 182B having a period of P=2,control circuit 552 atstep 1450 expands each pixel into a 2×2 pixel block. Whereimage sensor array 182B has a period P=3,control circuit 552 atstep 1450 expands each pixel into a 3×3 pixel block. Whereimage sensor array 182B includes the period of P=4,control circuit 552 atstep 1450 expands each pixel into a 4×4 pixel pixel block. At step 1454control circuit 552 calculates an intensity value Ic for each pixel position of the expanded color image.Control circuit 552 at step 1454 calculates an intensity value for each pixel position of the expanded color image according to the formula. -
I c=0.299R+0.587G+0.144B (Eq. 4) -
Control circuit 552 atstep 1460 then calculates an intensity value delta, D, for each pixel position, (Px, Py) utilizing a monochrome image intensity value Im and an expanded image color intensity value, Ic, at each pixel position.Control circuit 552 atstep 1460 may calculate an intensity value delta for each pixel position of the monochrome and expanded color image according to the formula -
D(P x ,P y)=I m(P x ,P y)−I c(P x ,P y) (Eq. 5) - At
step 1464,control circuit 552 updates the RGB data set color scale values of the expanded RGB color image using the set of formulas -
R′(P x ,P y)=R(P x ,P y)+D(P x ,P y) (Eq. 6) -
G′(P x ,P y)=G(P x ,P y)+D(P x ,P y) (Eq. 7) -
B′(P x ,P y)=B(P x ,P y)+D(P x ,P y) (Eq. 8) - At
step 1468control circuit 552 truncates RGB data set color scale values that are greater than 255 (where an 8 bit gray scale is used). Aftercontrol circuit 552 truncates RGB values greater than 255,control circuit 552 at step 1770 outputs a visual display color frame of image data having a spatial resolution equal to or approximately equal to the overall pixel resolution ofimage sensor array 182B. The visual display color frame of image data output at step 1770 may have a number of RGB data sets equal to the overall pixel count (e.g., monochrome pixels plus color sensitive pixels) ofimage sensor array 182B. - At
step 1411optical reader 100 receives from fusion circuit 1708 a high resolution visual display color frame of image data. The visual display color frame of image data received atstep 1411 may include a pixel resolution equal to or on the order of the pixel resolution ofimage sensor array 182B.Optical reader 100 may be regarded as having received a visual display color frame of image data whenfusion circuit 1708 outputs a visual display color frame of image data atstep 1470. - When executing
process 1440,control circuit 552 fuses monochrome and color image data to produce a high resolution visual display color frame of image data. When executing the alternative process described with reference to the flow diagram ofFIG. 14 j,control circuit 552 fuses monochrome and color image data in such a manner that color reproduction is optimized. - In general, increasing the percentage of
monochrome pixels 250M inimage sensor array sensitive pixels 250C in image sensor array increases color reproduction accuracy. Because of the light transmissivity of monochrome pixels, an image obtained utilizing an image sensor array having a higher percentage ofmonochrome pixels 250M has a higher signal to noise ratio than an image obtained utilizing an image sensor array having a smaller percentage ofmonochrome pixels 250M. Accordingly, an image obtained utilizing an image sensor array having a higher percentage of monochrome pixels often produces an image with greater detail and improved overall visual quality. -
Optical reader 100 in another aspect may incorporate the structure shown inFIG. 21 . InFIG. 21 ,center pixels 2072 of readerimage sensor array 182B have a higher percentage ofmonochrome pixels 250M, i.e., a period of P=4, as shown inFIG. 5 e, whileouter pixels 2074 have a lower percentage ofmonochrome pixels 250M, i.e., a period of P=2, as shown and described inFIG. 5 c. Theimage sensor array 182B is constructed such thatcenter pixels 2072 are optimized for providing image data yielding increased decoding accuracy whileouter pixels 2074 are optimized for providing image data yielding increased color reproduction accuracy. - Referring to further steps of
process 1400,control circuit 552 atstep 1412 outputs the visual display color frame of image data obtained atstep 1404. Atstep 1412control circuit 552 may output a visual display color frame of image data to display 504 for visual observation by an operator or to a designated color frame storage memory location ofreader 100 such as a designated frame memory storage location ofFlash memory 564 or to another frame memory location ofsystem 145. Wherecontrol circuit 552 is incorporated in hand heldhousing 101,control circuit 552 atstep 1410 may also send a visual display color frame of image data to spaced apartdevice 150, as shown inFIG. 10 . For sending a frame of image data to a spaced apartdevice 150,optical reader 100, the spaced apartdevice 150, and a communication link there between may be configured to transmit data packets in accordance with a protocol of the TCP/IP suite of protocols. Further,optical reader 100 may format the visual display color frame of image data obtained atstep 1412 in a suitable image file format (e.g., .BMP, .TIFF, .PDF, .JPG, .GIF) andoptical reader 100 may automatically send the visual display color frame of image data atstep 1412 utilizing the file transfer protocol (FTP).Optical reader 100 atoutput step 1212 may format the visual display color frame of image data in a suitable image file format (e.g., .BMP, .TIFF, .PDF, .JPG, .GIF) when storing the visual display color frame of image data in memory 566 (which can be incorporated in hand held housing 101) or when sending the visual display color frame of image data to a spaced apartdevice 150 for storage.Optical reader 100 may also transmit a visual display color frame of image data utilizing a suitable markup language such as .XML. Referring toFIG. 10 ,system 145 may be configured so that when a display equipped spaced apartdevice 150 receives a visual display color frame of image data fromoptical reader 100, the spaced apartdevice 150 automatically displays that received visual display color frame of image data on adisplay 1504 associated with that device. -
Optical reader 100 can be configured so that all the steps ofprocess 1400 are carried out automatically in response to receipt of a trigger signal until a stop condition is satisfied. A stop condition may be the receipt of a trigger stop signal such as may be generated by the release oftrigger 216. - In the embodiments of
FIGS. 14 a-14 c, two actuations of a reader control button are made to carry out an indicia decode process and two actuations of a reader control button are made to carry out a picture taking process (one actuation ofbutton 3162 orbutton 3164 to configure thereader 100 and then another actuation oftrigger 216 to capture an image). It will be understood thatoptical reader 100 can be configured to carry out indicia decoding or picture taking with a single actuation of a reader control button. For example,optical reader 100 can be configured so that actuation ofvirtual button 3162 both configuresreader 100 to decode and simultaneously generates a trigger signal to immediately commence image capture and decoding.Optical reader 100 can also be configured so that actuation ofvirtual icon button 3164 both configures areader 100 for picture taking and simultaneously generates a trigger signal to immediately commence image capture. - While
process 1200 andprocess 1400 may be carried out in the alternative,process 1200 andprocess 1400 may also be executed contemporaneously. For example, whilecontrol circuit 552 obtains a decode frame atstep 1204,control circuit 552 may obtain a visual display color frame of image data atstep 1404.Control circuit 552 may obtain a color frame of image data as a decode frame atstep 1204 and then outputs that frame atstep 1212 as visual display color frame of image data.Control circuit 552 atstep 1412 may output a visual display color frame of image data and contemporaneously transfer that frame of image data to decodecircuit 1702. In general,reader 100 may be configured so that whenevercontrol circuit 552 obtains a decode frame atstep 1204,control circuit 552 may store that frame for later processing, which processing may include processing for generating a visual display color frame of image data and which processing may be responsive to an operator input command to perform such processing.Optical reader 100 may also be configured so that whencontrol circuit 552 obtains a visual display color frame of image data atstep 1404, control circuit may store that frame for further processing, which processing may include transferring that frame to decodecircuit 1702 orautodiscrimination circuit 1704, and which processing may be responsive to an operator input command to perform such processing. - Another embodiment of the invention is described with reference to
FIGS. 17 a-17 g. In the embodiment ofFIGS. 17 a-17 goptical reader 100 includes a pair ofimaging modules Imaging module 1802D is a color imaging module having colorimage sensor array 182D. Colorimage sensor array 182D includes a Bayer pattern color filter with one of red, green or blue wavelength selective filter disposed on each pixel.Imaging module 1802E as shown inFIG. 17 e is a monochrome imaging module having a one-dimensional solid stateimage sensor array 182E. One-dimensional monochromeimage sensor array 182E in the embodiment ofFIGS. 17 a, 17 e, 17 f, and 17 g includes an M×1 (one row) array of monochrome (without color filter) pixels. One-dimensionalimage sensor array 182E may also include and M×N array of pixels, where M>>N, e.g., an M×2 (2 rows) of pixels. - The
reader 100 shown in the electrical block diagram ofFIG. 17 a has many of the same components as shown inoptical reader 100 ofFIG. 1 a. Namely,optical reader 100 ofFIG. 17 a includes acontrol circuit 552 provided in the example by a CPU, which operates under the control of program data stored inEPROM 562.Control circuit 552 is in communication with amemory unit 566 that includes in addition toEPROM 562,RAM 560, andFlash memory 564.Control circuit 552 further receives input control data from various user input devices such asmanual trigger 216,pointer controller 512,keyboard 508 andtouch screen 504T.Control circuit 552 may also output data such as decoded output data and visual display image data tocolor display 504. For capturing image data,control circuit 552 may control eitherimage sensor array 182E orimage sensor array 182D. For capturing one-dimensional image data corresponding to a one-dimensional slice image of a target,control circuit 552 sends various image capture initiating control signals to one-dimensionalimage sensor array 182E. In response to the image capture initiation control signals,image sensor array 182E sends analog image signals to signalprocessing circuit 591 which among various processing functions amplifies the signals and feeds the signals to analog-to-digital converter 592. Analog-to-digital converter 592 converts the signals into digital form and routes the digital image data to FPGA 593 which under the control ofcontrol circuit 552, manages the transfer of the digital information intoRAM 560, where the monochrome image data can be accessed for decoding processing bycontrol circuit 552. For capturing two-dimensional frames of color image data,control circuit 552 sends appropriate image capture initiation control signals (e.g., exposure, read out) toimage sensor chip 1082.FPGA 580 receives digital image data from imagesensor IC chip control circuit 552 manages the transfer of color image data intoRAM 560.Illumination assembly 104 for eachmodule FIGS. 15 a-15 e. -
Optical reader 100 as shown inFIGS. 17 a-17 g may be operated in accordance with the flow diagram ofFIGS. 17 a-17 g. Namely, by a suitable selection method such as by depressingicon 3162 or icon 3164 (FIG. 9 b) one of a decode mode of operation and a color image capture mode of operation can be selected. However, in the dual imaging module embodiment ofFIGS. 17 a-17 g, the imaging module which is utilized for capturing image data depends on which mode (indicia decoding, or picture taking) is selected. If the indicia decode mode is selected at step 1100 (FIG. 14 a) and a trigger signal is received,optical reader 100 proceeds to step 1102 to execute indicia decode process 1200 (FIG. 14 a). Atstep 1204 of indicia decodeprocess 1200,control circuit 552 obtains a decode frame of image data. If the picture taking mode of operation is selected at step 1100 (FIG. 14 a) and a trigger signal is received,control circuit 552 proceeds to step 1404 (FIG. 14 c) to obtain a visual display color frame of image data. Wherereader 100 includes two imaging modules, one color such asmodule image sensor array module monochrome image sensor image sensor array 182 to whichcontrol circuit 552 sends control signals for initiating image capture depends on whetheroptical reader 100 is operating in a decode mode of operation or a picture taking mode of operation. With reference to thereader 100 ofFIGS. 17 a-17 g, and the flow diagrams ofFIGS. 14 a, 14 b and 14 c,reader 100 atstep 1204 sends image capture initiation control signals to monochrome one-dimensionalimage sensor array image sensor array reader 100 is operating in a decode mode operation.Reader 100 atstep 1404 sends image capture initiation control signals to colorimage sensor array image sensor array reader 100 is operating in a picture taking mode operation. Accordingly, whereoptical reader 100 is in an indicia decode mode and receives a trigger signal, a monochrome frame of image data is sent to RAM 560 for further processing by decode circuit 1702 (FIG. 10 ). Whereoptical reader 100 is in a picture taking mode and receives a control signal, a color image is sent to RAM 560. The color image if a Bayer pattern image is subject to a demosaicing process as described herein for generating a visual display color frame of image data, which visual display color frame of image data may be output bycontrol circuit 552 e.g., to display 504 and/or a designated memory address of system 145 (e.g.,memory 566 or another memory such as a memory of a spaced apart device 150), and/or to adisplay 1504 of a spaced apartdevice 150 of system 145 (FIG. 5 ). - When an image is captured by the two imaging module readers of
FIGS. 17 a-17 g, the type of image capture (monochrome or color) depends on a selected operating mode. When an indicia decode mode is selected, a monochrome gray scale image well suited for decode processing is captured. When a picture taking mode is selected, a color image is captured which is well suited for visual display. - Further aspects of a dual imaging module reader are described with reference to
FIGS. 17 b-17 g.FIGS. 17 b and 17 c illustrate that thehardware block 598 ofreader 100 shown inFIGS. 17 a, 17 f, and 17 g may be replaced with alternative hardware blocks. As indicated byFIG. 17 b,hardware block 398 which inFIG. 17 a includes a CCD one-dimensional solid stateimage sensor array 182E and off-boardsignal processing circuit 591, analog-to-digital converter 592 andFPGA 593 may be replaced by a hardware block including an CMOS imagesensor IC chip 1082F including a monochromeimage sensor array 182F. Imagesensor IC chip sensor IC chip IC chip image sensor array 182F ofchip 1082F includesmonochrome pixels sensitive pixels FIG. 17 c illustrates that imagingassembly hardware block 598 can be replaced with a laser scanningbar code engine 594 and an associateddecode circuit 595. Laser scanningbar code engine 594 and associateddecode circuit 595 may be available in a package known as an SE 923 decoded out scan engine available from Symbol Technologies. In the embodiment ofFIG. 17 c, steps 1210, 1212, 1214 ofdecode process 1200 are carried out bydecode circuit 595. - Exemplary imaging modules supporting various types of image sensor IC chips are shown in
FIGS. 17 d and 17 e.FIG. 17 d shows an exemplary imaging module for supporting imagesensor IC chip 182D.Imaging module 1082D includes the elements shown and described with reference toFIGS. 8 a-8 d except thatimaging module 1082D includes imagesensor IC chip 182D and further except that certain light sources are optionally deleted. Imaging module 1082E includes the elements shown and described with reference toFIGS. 8 a-8 e except that imaging module 1082E includes one-dimensional monochromeimage sensor chip 182E and further except that certain light sources ofillumination block 104 are optionally deleted. Withmodule 1802E, aiming pattern 1838 (FIG. 8 d) may serve as an aiming and illumination pattern. Further, it will be noted thatillumination assembly 104 of an imaging module herein may include a flash illumination light source, 160, 160X (FIG. 9 a). It may be particularly useful to incorporate a flash illumination intoillumination assembly 104, where animaging module 1082 is used primarily for capture of visual display color image. - Referring to
FIGS. 17 f and 17 g construction views of dual imaging module readers incorporated in various optical reader housings are shown and described. InFIG. 17 f a gun styleoptical reader 100 is shown having color two-dimensional imaging module 1802D and one-dimensional monochrome imaging module 1082E supported therein. InFIG. 17 g a portable data terminal (PDT)optical reader 100 is shown having color two-dimensional imaging module 1802D and one-dimensionalmonochrome imaging module 1802E supported therein. The dual modules can also be installed in other types of housings such as cell phone housings (FIG. 9 c) and personal data assistant housings (PDAs). In the examples ofFIGS. 17 f and 17 g,imaging modules 1802 are supported bystruts 597 formed on aninterior wall 1802 ofhousing 101. Viaribbon connectors 598, themodules 1802 in each example are in communication with a main printedcircuit board 599 which includes various electrical components includingprocessor IC chip 548. - In one application, the
optical reader 100 ofFIGS. 17 a-17 g is operated in the following manner. An operator actuates colorimage sensor array 182D to take a color picture of parcel 1260 (FIGS. 11 and 12 ) carrying abar code symbol decode button 3164 and then trigger 216 orbutton 3164 only. An operator then actuates monochromeimage sensor array 182E (or alternativelyimage sensor array 182F, or laser scan engine 594) to decodebar code symbol button 3162 and then trigger 216 or by depressingbutton 3162 only. Further,control circuit 552, which may be incorporated in hand heldhousing 101, may transmit a visual display color frame of imagedata representing parcel 1260 and decoded out messages corresponding to one or more ofsymbols FIG. 10 ).System 145 can be configured so that such transmission is automatic in response to trigger signals being received, oroptical reader 100 can be configured so that associated color picture data and decoded out bar code message data are transmitted in response to receipt of a user-initiated command input into a user-interface ofoptical reader 100 to transmit associated picture and decoded bar code message data. - With further reference to the reader electrical block diagram shown in
FIG. 1 a, various useful optical reader embodiments may be yielded by reconfiguration ofhardware block 208 including an image sensor array. With reference toFIG. 18 a, anoptical reader 100 having the hardware components shown inFIG. 1 a may be modified to include animage sensor array 182C as shown and described in connection withFIG. 18 a. In the embodiment ofFIG. 18 a,optical reader 100 includes cyan-magenta-yellow (CMY)color filter array 182C. Eachpixel 250 ofimage sensor array 182C includes a color filter element; namely one of a cyan color filter element, a magenta color filter element or a yellow color filter element. Yellow color filter elements have excellent light transmittance (approaching the transmittance of a monochrome pixel). Further, it is seen that in accordance with the CMY color filter pattern shown inFIG. 18 a that approximately 50% of all pixels ofimage sensor array 182C are yellow pixels (pixels having a yellow light wavelength sensitive filter element). In the specific example ofFIG. 18 a,image sensor array 182C having cyan, magenta and yellow pixels is devoid of green pixels. However, image sensor arrays are available which have green pixels in addition to cyan, magenta and yellow pixels.Image sensor array 182C may be incorporated into anoptical reader 100 that operates in accordance with the picture taking mode/indicia decode mode flow diagram described in connection withFIG. 14 a. That is, when driven into a indicia decode mode of operation as described in connection withFIG. 14 b,optical reader 100 including CMY colorimage sensor array 182C obtains a decode frame of image data whereas whenoptical reader 100 includingimage sensor array 182C is driven into a picture taking mode of operation,optical reader 100 obtains a visual display color image frame of image data as described in connection withFIG. 14 c herein. - According to the invention, an optical reader including a CMY
image sensor array 182C as shown inFIG. 18 a may obtain image data in a manner that depends on which operational mode (indicia code or picture taking) is selected. Whereoptical reader 100 including CMYimage sensor array 182C obtains a decode frame of image data atstep 1204,control circuit 552 ofoptical reader 100 can selectively address yellow color pixels of CMYimage sensor array 182C and selectively read out image data only from yellow colored pixels ofimage sensor array 182C. With further reference to a reader includingimage sensor array 182C,control circuit 552 atstep 1204 may interpolate missing pixel values corresponding to the pixel positions of magenta and cyan pixels ofimage sensor array 182C. After interpolating the missing pixel positions,control circuit 552 atstep 1210 may transfer the interpolated decode frame to one of indicia decodecircuit 1702 orautodiscrimination circuit 1704. - In a further aspect of the optical reader described in connection with
FIG. 18 a including a CMY colorimage sensor array 182C,image sensor array 182C may include separate and independent reset control lines for facilitating the reset of magenta (labeled “Mg”) and cyan (labeled “Cy”) pixels independent from the resetting of yellow pixels (labeled “Y”). Accordingly, when image data atstep 1204 is read out selectively from yellow pixels, the magenta and cyan pixels ofimage sensor array 182C may be driven into reset to eliminate electron diffusion cross talk and to reduce cross talk attributable to photons enteringimage sensor array 182C through magenta andcyan color pixels 250C. - When obtaining a visual display color frame of image data as described at
step 1404 of the flow diagramFIG. 14 c an optical reader includingimage sensor array 182C may simply read out image data from all of the pixels of thearray 182C and execute a simple demosaicing algorithm to convert a single color value for each pixel ofimage sensor array 182C into a visual display color image wherein each pixel ofimage sensor array 182C is represented by a data set including three color scale values, e.g., a cyan color scale value, a magenta color scale value and a yellow color scale value. -
Control circuit 552 atstep 1404 where the reader includes a CMYimage sensor array 182C may transform the CMY visual display image into an RGB visual display image utilizing a CMY to RGB transformation process as described herein. - The performance of
optical reader 100 may be hindered whereoptical reader 100 is operated to read bar code symbols or other indicia disposed on a substrate having a shiny surface (e.g., metal, glass, laminated, plastic, etc.). Light rays emanating fromlight sources 160 ofreader 100 that are projected on a highly reflective shiny surface of a substrate, s, may be substantially entirely reflected directly on toimage sensor array 182. “Specular” reflection is said to occur where a substantial percentage of light rays are reflected and directed ontoimage sensor array 182. Light rays are said to be reflected at a “specular angle” when light rays are reflected from a surface at about the angle of incidence. Specular reflection tends to saturateimage sensor array 182 to cause decoding failures. Theoptical reader 100 described in connection withFIGS. 19 a-c is configured so that read errors resulting from specular reflection are reduced. As shown and described in connection withFIG. 19 a,hardware block 208 shown inFIG. 1 a as including a hybrid monochrome in colorimage sensor array 182A can be replaced withhardware block 208 as shown inFIG. 19 a including a hybrid monochrome and polarizer filterimage sensor array 182G. -
Image sensor array 182G includes a first subset ofmonochrome pixels 250M and a second subset of lightpolarizing pixels 250P. Lightpolarizing pixels 250P ofimage sensor array 182G include light polarizing filter elements 261 (alternatively termed “light polarizing filters,” or simply “light polarizers”) typically formed at eachpolarizing pixel 250P in the position offilter 260 as shown in the color pixel views ofFIGS. 3 c and 6 c. Lightpolarizing filter elements 261 ofimage sensor array polarizing pixels 250P by way of a depositing process. Lightpolarizing filter elements 261 ofimage sensor array 182G can be constructed to attenuate polarized light rays generated from an appropriately polarized light source and reflected at a specular angle. Accordingly, polarized light rays incident on the image sensor array on thepolarizing pixels 250P are attenuated significantly; thus, reducing the contribution of specularly reflected light rays to generated image signals from thepolarizing pixels 250P. - According to the invention,
optical reader 100 includingimage sensor array 182G may be configured to selectively address lightpolarizing pixels 250P and selectively read out image data from lightpolarizing pixels 250P to generate image data for subjecting to decoding which is likely to result in successful reading of bar codes or other indicia notwithstanding the image data being obtained during specular reflections read conditions. - Referring to
FIG. 19 b, a perspective view of light polarizingimage sensor array 182G is shown with an exploded view showing a pattern which may be repeated throughout the array. In the version ofFIG. 19 b, lightpolarizing pixels 250P having light polarizinglight filter elements 261 are uniformly distributed throughoutimage sensor array 182G with a period of P=2. It will be understood that lightpolarizing pixels 250P may also be distributed throughoutimage sensor array 182G in a uniform or substantially uniform distribution pattern other than the pattern shown inFIG. 19 b. For example, lightpolarizing pixels 250P may be distributed throughoutimage sensor array 182G with a distribution pattern of P=3, (as described in connection withFIG. 5 d showing a hybrid monochrome and color image sensor array) or a distribution pattern having the period of P=4, as illustrated with reference to hybrid monochrome and color image sensor array as shown inFIG. 5 e. - Referring to the view of
FIG. 9 b,optical reader 100 may be operated in a mode in whichoptical reader 100 captures image data by selectively addressingpolarizing pixels 250P and selectively reading out image data from lightpolarizing pixels 250P only.Optical reader 100 may be configured to have a reduced specular reflection read error decode mode.Optical reader 100 can be configured so that whenbutton 3156 is actuated,optical reader 100 receives a trigger signal to obtain image data that is likely to result in successful reading notwithstanding specular reflection reading conditions. - Referring to the flow diagram of
FIG. 19 c,optical reader 100 atstep 1902 may receive a trigger signal to commence operation in a reduced specular reflection read error decode mode. The trigger signal may be received pursuant to a manual control by an operator such as an actuation ofcontrol button 3156.Control circuit 552 may also be configured to receive the trigger signal atstep 1902 whencontrol circuit 552 automatically senses a predetermined condition such as a saturation condition.Control circuit 552 atstep 1902 may determine that a saturation condition is present by analysis of image data at step 1204 (FIG. 14 b) during normal decoding operations so that when a saturation condition is detected,optical reader 100 automatically commences operation in a reduced specular reflection read error decode mode. In a specific embodiment of the invention,control circuit 552 may determine that a saturation condition is present when an average white value of monochrome image data is below a predetermined level. - At
step 1904optical reader 100 obtains a specular reflection read condition decode frame of image data.Control circuit 552 obtains a specular reflection condition decode frame of image data atstep 1902 by selectively addressing lightpolarizing pixels 250P ofimage sensor array 182G and selectively reading out image data from lightpolarizing pixels 250P only. In another aspect ofimage sensor array 182G that may be incorporated inoptical reader 100,image sensor array 182G may include separate reset control lines for resettingmonochrome pixels 250M separately and independently of lightpolarizing pixels 250P.Image sensor array 182G may have separate sets of reset control lines as described in connection withimage sensor array 182G, particularly in connection withFIG. 7 a. - Accordingly, when
control circuit 552 selectively addresses lightpolarizing pixels 250P for read out of image data from lightpolarizing pixels 250P,control circuit 552 drivesmonochrome pixels 250M into reset. Resetting ofmonochrome pixels 250M is synchronized with the exposure period for exposing lightpolarizing pixels 250P as described herein. Drivingmonochrome pixels 250M into reset while lightpolarizing pixels 250P are exposed eliminates electron diffusion cross talk and reduces cross talk resulting from photon penetration to imagesensor array 182G. - At
step 1904,control circuit 552 may interpolate pixel values at pixel positions corresponding to missing pixel positions. At step 1906control circuit 552 transfers the specular reflection condition decode frame of image data obtained atstep 1904 to indicia decodecircuits 1702 orautodiscrimination circuit 1704 as are described in connection withFIG. 1 e. - At
step 1908control circuit 552 receives decoded output data output bydecode circuit 1702 orsignature autodiscrimination circuit 1704. Atstep 1910control circuit 552 outputs decoded out data, e.g., by transferring decoded out data to an onreader display 504 or to a spaced apart display 1504 or else stores decoded data in appropriate memory address location of system 145 (FIG. 10 ). - A process has been described with reference to the flow diagram of
FIG. 19 c whereincontrol circuit 552 selectively reads out monochrome pixel image data frommonochrome pixels 250M and selectively reads out image data from lightpolarizing pixels 250P. An optical reader including hybrid monochrome and light polarizingimage sensor array 182G may also be operated without selectively reading out image data fromimage sensor array 182G. An optical reader incorporating hybrid monochrome and light polarizingimage sensor array 182G can be operated to decode decodable indicia and to take pictures in accordance with the process described with reference to the flow diagrams ofFIGS. 14 a, 14 b, and 14 c. In obtaining a decode frame of image data (step 1204,FIG. 14 b),control circuit 552 may read out image data from all pixels of hybrid monochrome and light polarizingimage sensor array 182G including image data from allmonochrome pixels 250M and all lightpolarizing pixels 250P in a single frame capture step. The full frame monochrome and light polarizer pixel image data can also be captured with two frame capture steps. Atstep 1210,control circuit 552 may transfer to decodecircuit 1702 orautodiscrimination circuit 1704 the full frame of monochrome and polarized pixel image data obtained atstep 1204. Ifdecode circuit 1702 orautodiscrimination circuit 1704 fails to decode or fails to detect the presence of handwritten characters,control circuit 552 may, afterstep 1210, transfer a subset of full frame of image data originally transferred atstep 1210. Namely, afterstep 1210, if decoding or autodiscrimination fails,control circuit 552 may transfer to decodecircuit 1702, or autodiscrimination circuit 1704 a reduced resolution image extracted from a full frame image by selectively extracting monochrome image data from the full frame of image data. The reduced resolution frame of image data includes only image data corresponding to lightpolarizing pixels 250P ofimage sensor array 182G. The failure ofdecode circuit 1702 to decode or autodiscrimination circuit to recognize may be regarded as a determination bycontrol circuit 552 that a saturation condition is present. - The elements of a hybrid monochrome and color image sensor array (such as
image sensor array image sensor array 182G into a single image sensor array.FIGS. 20 a and 20 b show animage sensor array 182H including a first subset ofmonochrome pixels 250M, a second subset of colorsensitive pixels 250C and a third subset of lightpolarizing pixels 250P.Image sensor array 182H may include three separate sets of reset control lines to enable separate and independent of resetting ofmonochrome pixels 250M, of colorsensitive pixels 250C and of lightpolarizing pixels 250P.Image sensor array 182H may be incorporated in hand heldoptical reader 100 and may be substituted forhardware block 208 as shown inFIG. 1 a.Optical reader 100 incorporatingimage sensor array 182H may have operating modes in which optical reader separately addressesmonochrome pixels 250M for read out of image data frommonochrome pixels 250M only.Optical reader 100 includingimage sensor array 182H may also have an operating mode in whichoptical reader 100 selectively addresses colorsensitive pixels 250C and selectively reads out image data from color sensitive 250C.Optical reader 100 may also have an operating mode in whichoptical reader 100 selectively addresses lightpolarizing pixels 250P and selectively reads out image data from lightpolarizing pixels 250P.Optical reader 100 may obtain a full frame of image data including monochrome, color and light polarizing pixels image data (obtained with one, two, or three frame capture steps) and then utilize the image data on an as needed basis. For example, if a decode attempt utilizing the full frame image data fails,optical reader 100 may selectively extract light polarizing pixel image data from the full frame image data and transfer the extracted image data to decodecircuit 1702. - In general,
optical reader 100 includingimage sensor array 182H selectively reads out image data frommonochrome pixels 250M in obtaining a decode frame of image data for transferring to adecode circuit 1702 under normal read conditions.Optical reader 100 selectively reads out image data from colorsensitive pixels 250C when obtaining image data for use when obtaining a visual display color frame of image data.Optical reader 100 selectively reads out image data from lightpolarizing pixels 250P, or selectively extracts image data corresponding topixels 250P from a frame of image data whenoptical reader 100 senses that a specular reflection is present or when an operator pursuant to operator control drivesoptical reader 100 into a reduced specular reflection read error decode mode of operation. Anoptical reader 100 includingimage sensor array 182H may operate in accordance with the picture taking and decode mode flow diagram as described in connection withFIG. 14 a and may execute the reduced specular reflection read error decode mode decoding process described in connection withFIG. 19 c. - For enhancing the performance of an optical reader according to the invention, having an image sensor array such as
image sensor array optical reader 100 may incorporate emit optics light polarizers (which may alternatively be termed “light polarizing filter elements” or “light polarizing filters”). For example, a reader imaging module, e.g.,module 1802A can include anoptical plate 1962 as shown inFIG. 8 f which may be disposed forwardly ofcircuit board 1806 as shown inFIG. 8 a.Optical plate 1962 can incorporatelight polarizers 1963 which polarize light fromlight sources 160S, 160T, that can be selectively energized when capturing images utilizing polarizingimage sensor array Light polarizers 1963 can be cross-polarized relative to thepolarizing filter elements 261 ofimage sensor array Optical plate 1962 can include other such elements as optical diffusers (not shown) for diffusing light rays emitted by light sources 160C-160T. - Further aspects of indicia decode
circuit module 1702 andautodiscrimination circuit module 1704 are described with reference toFIGS. 22 a-22 i. Indicia decode circuit 1702 (which may be a bar code symbol dataform decode circuit) when receiving image data transferred bycontrol circuit 552 may search the image data for markers, such as a quiet zone, indicative of the presence of a dataform, such as a one or two-dimensional bar code. If a potential decodable indicia (dataform) is located, thedecode circuit 1702 applies one or more indicia decoding algorithms to the image data. If the decode attempt is successful, the optical reader outputs decoded dataform data. All of the circuits (modules) described with reference toFIG. 22 a can be incorporated inhousing 101. Further, all of the circuits ofFIG. 22 a can be embodied by the combination ofcontrol circuit 552 andmemory 566. -
Optical reader 100 may also include anautodiscriminating circuit 1704. Referring toFIG. 22 a,autodiscriminating circuit 1704 may incorporate adecode circuit 1702 and an image processing andanalysis circuit 21208, that are in communication with one another. - As shown in this embodiment, the image processing and
analysis circuit 21208 comprises afeature extraction circuit 21212, ageneralized classifier circuit 21216, a signaturedata processing circuit 21218, anOCR decode circuit 21222, and agraphics analysis circuit 21224 that are in communication with each other. In addition, as shown inFIG. 22 a, thefeature extraction circuit 21212 comprises abinarizer circuit 21226, aline thinning circuit 21228, and aconvolution circuit 21230 that are in communication with each other. -
FIG. 22 b shows aprocess 21300 for employing one embodiment of the invention utilizing the autodiscrimination circuit shown inFIG. 22 a. Theprocess 21300 comprises an image reader recording an actuation event (step 21302), such as a receipt of a trigger signal, and in response atstep 21304, collecting (obtaining) image data from a target with theoptical reader 100. The collecting of image data step may be in accordance with step 1204 (FIG. 14 b). After collection, the image data is transferred (step 21308) to thedecode circuit 1702. The dataform decode circuit searches (step 21310) the image data for markers, such as a quiet zone, indicative of the presence of a dataform, such as a one or two-dimensional bar code. If a potential dataform is located, thedecode circuit 1702 applies (step 21314) one or more dataform decoding algorithms to the ensuing image data. If the decode attempt is successful, theoptical reader 100 outputs (step 21318) decoded dataform data and signals (step 21322) a successful read with an alert, such as a beep tone. - In one embodiment if the decode attempt is not successful, the image data is transferred (step 21326) to the image processing and
analysis circuit 21208. In another embodiment, the image data is processed in parallel with the attempt to decode the dataform data. In one such embodiment, the process that completes first (i.e., dataform decode attempt or the image processing) outputs its data (e.g., a decoded bar code or a captured signature) and the other parallel process is terminated. In a further embodiment, the image data is processed in response to the decoding of the dataform. In one such embodiment, a bar code encodes item information such as shipping label number and information indicating that a signature should be captured. - Within the image processing and
analysis circuit 21208, the image data is processed by thefeature extraction circuit 21212. In general, the feature extraction circuit generates numeric outputs that are indicative of the texture of the image data. As indicated above, the texture of the image data refers to the characteristics of the type of data contained in the image data. Common types of texture include one or two-dimensional bar code texture, signature texture, graphics texture, typed text texture, handwritten text texture, drawing or image texture, photograph texture, and the like. Within any category of texture, sub-categories of texture are sometimes capable of being identified. - As part of the processing of the image data by the
feature extraction circuit 21212, the image data is processed (step 21328) by thebinarizer circuit 21226. Thebinarizer circuit 21226 binarizes the grey level image into a binary image according to the local thresholding and target image size normalization. With the image data binarized, the image data is processed (step 21332) by theline thinning circuit 21228 to reduce multi-pixel thick line segments into single pixel thick lines. With binarized line thinned image data, the image data is processed (step 21336) by theconvolution circuit 21230. - In general, the
convolution circuit 21230 convolves the processed image data with one or more detector maps designed according to the invention to identify various textural features in the image data. In one embodiment, theconvolution circuit 21230 generates a pair of numbers, the mean and variance (or standard deviation), for each convolved detector map.FIG. 22 c shows a set of 12 2×3 binary curvelet detector maps 21250 used to detect curved elements present in image data. As each of the curvelet detector maps 21250 is convolved with the image data, the mean value and the variance generated provide an indication of the presence or density of elements in the binarized line thinned image data having similar shapes to the curvelet detector maps 21250. As each pixel map generates a pair of numbers, the 12 curvelet detector maps 21250 generate a total of 24 numbers. According to one embodiment, these 24 numbers are representative of the curved or signature texture of the processed image data. - Further processing of the image data includes the outputs from the
feature extraction circuit 21212 being fed (step 21340) into the generalizedclassified circuit 21216. Thegeneralized classifier circuit 21216 uses the numbers generated by the feature extraction circuit as inputs to a neural network, a mean square error classifier or the like. These tools are used to classify the image data into general categories. In embodiments employing neural networks, different neural network configurations are contemplated in accordance with the invention to achieve different operational optimizations and characteristics. In one embodiment employing a neural network, thegeneralized classifier circuit 21212 includes a 24+12+6+1=43 nodes Feedforward, Back Propagation Multilayer neural network. The input layer has 24 nodes for the 12 pairs of mean and variance outputs generated by aconvolution circuit 21230 employing the 12 curvelet detector maps 21250. In the neural network of this embodiment, there are two hidden layers of 12 nodes and 6 nodes respectively. There is also one output node to report the positive or negative existence of a signature. - In another embodiment employing a neural network, the 20 curvelet detector maps 21260 shown in
FIG. 22 d are used by theconvolution circuit 21230. As shown, the 20 curvelet detector maps 21260 include the original 12 curvelet detector maps 21250 ofFIG. 22 c. The additional 8 pixel maps 21260 are used to provide orientation information regarding the signature. In one embodiment employing the 20 curvelet detector maps 21260, thegeneralized classifier circuit 21212 is a 40+40+20+9=109 nodes Feedforward, Back Propagation Multiplayer neural network. The input layer has 40 nodes for the 20 pairs of mean and variance outputs generated by aconvolution circuit 21230 employing the 20 curvelet detector maps 21260. In the neural network of this embodiment, there are two hidden layers of 40 nodes and 20 nodes respectively, one output node to report the positive or negative existence of a signature, and 8 output nodes to report the degree of orientation of the signature. The eight output nodes provide 28=256 possible orientation states. Therefore, the orientation angle is given in degrees between 0 and 360 in increments of 1.4 degrees. - In some embodiments, the
generalized classifier circuit 21216 is capable of classifying data into an expanded collection of categories. For example, in some embodiments thegeneralized classifier circuit 21216 specifies whether the image data contains various data types such as a signature; a dataform; handwritten text; typed text; machine readable text; OCR data; graphics; pictures; images; forms such as shipping manifest, bill of lading, ID cards, and the like; fingerprints, biometrics such as fingerprints, facial images, retinal scans and the like, and/or other types of identifiers. In further additional embodiments, thegeneralized classifier circuit 21216 specifies whether the image data includes various combinations of these data types. In some embodiments, thegeneral classifier circuit 21216 specifies whether the image data contains a specified type of data or not. In one such embodiment the image processing andanalysis circuit 21208 is contained within an identification circuit that outputs an affirmative or negative response depending on the presence or absence of the specified data type, such as a signature or a biometric in the image data. - In one embodiment once the presence of a signature has been confirmed and its general orientation determined, image data is transferred (step 21344) to the signature
data processing circuit 21218. In one embodiment, the signaturedata processing circuit 21218 is used to detect the boundaries of the signature in the image data. In one embodiment, the signature boundary is detected using a histogram analysis. As shown inFIG. 22 e, a histogram analysis consists of a series of one-dimensional slices along horizontal and vertical directions defined relative to the orientation of the signature. In one embodiment, the value for each one-dimensional slice corresponds to the number of black (i.e., zero valued) pixels along that pixel slice. In some embodiments if no bar codes have been decoded, then some specified region of the full frame of image data, such as a central region is captured for signature analysis. Once completed, the histogram analysis provides a two-dimensional plot of the density of data element pixels in the image data. The boundary of the signature is determined with respect to a minimum density that must be achieved for a certain number of sequential slices. In one embodiment, the histogram analysis searches inwardly along both horizontal and vertical directions until the pixel density rises above a predefined cutoff threshold. So that the signature data is not inadvertently cropped, it is common to use low cutoff threshold values. - In one embodiment, once the boundaries of the signature have been determined, the signature
data processing circuit 21218 crops the image data and extracts the signature image data. In one such embodiment, the cropping is performed by an image modification circuit that generates modified image data in which a portion of the image data not including the signature has been deleted. In other embodiments, various compression techniques are employed to reduce the memory requirements for the signature image data. One such technique includes the encoding of the signature image data by run length encoding. According to this technique, the length of each run of similar binarized values (i.e., the length of each run of 1 or 0) for each scan line is recorded as a means of reconstructing a bit map. Another encoding technique treats the signature image data as a data structure where the elements of the data structure consist of vectors. According this encoding technique, the signature is broken down into a collection of vectors. The position of each vector in combination with the length and orientation of each vector is used to reconstruct the original signature. In one such embodiment, the encoding process generates a new vector whenever the curvature for a continuous pixel run exceeds a specified value. A further compression technique employs B-Spline curve fitting. This technique has the capacity to robustly accommodate curvature and scaling issues. - In various embodiments, the signature image data or a compressed or encoded version of the signature image data is stored locally on a dedicated memory device. In one such embodiment, the local memory device can be a detachable memory device such as a CompactFlash memory card or the like described in more detail below. In another embodiment, the signature image data is stored in a volatile or non-volatile portion of general purpose memory and downloaded at a future time. In a further embodiment, the signature image data can be transmitted via wired or wireless means either at the time of capture or at a later point, such as when a data collection session has been completed.
- In another embodiment, the signature
data processing circuit 21218 does not perform a histogram analysis but simply stores in memory the entire image or a compressed version once the presence of a signature has been determined. In a further embodiment to save processing time, the initial image analysis is performed on a lower resolution image. Once the presence of a signature is determined in this embodiment, a higher resolution image is taken. In one embodiment, a signature extraction histogram analysis is performed on this image. Next, the image is stored in memory in either compressed or original format. In some embodiments, the image data is combined with other data to form a record for a particular item such as a package or shipping envelope. As mentioned above, some of the additional data that can be collected by theoptical reader 100 and stored with or separate from the signature data includes but is not limited to dataform data, handwritten text data, typed text data, graphics data, image or picture data, and the like. - As part of its operations, the image processing and
analysis circuit 21208 can be designed to perform specialized tasks for different data types. For example, if thegeneralized classifier circuit 21216 determines that the image data contains typed or machine readable text, the image data can be collected, possibly histogram analyzed, and stored or alternatively, the image data can be transferred to theOCR decoding circuit 21222. Similarly, if thegeneralized classifier circuit 21216 determines that the image data includes a graphic element, the image data can be transferred to thegraphics analysis circuit 21224 for processing. In one embodiment, thegraphics analysis circuit 21224 is configured to recognize and decode predefined graphics. In one such embodiment, the graphics analysis can include determining which, if any, boxes have been selected in the billing and shipping instructions on a shipping label. In a further embodiment, the graphics analysis can include locating and decoding the typed or handwritten text contained in the zip code box on a shipping label. In an alternative embodiment, theoptical reader 100 can be configured to automatically attempt decode operations in addition to the dataform decode, such as OCR decoding or graphics decoding, prior to the activation of thefeature extraction circuit 21212. - In another embodiment, the image processing and
analysis circuit 21208 segments the image data into regions and performs a feature extraction and general classification analysis on each region. In one embodiment as shown inFIG. 22 f, the standard rectangular image data window is divided into four equal sized sub-rectangles. In another embodiment shown inFIG. 22 g, the segmentation consists of overlapping regions so that the total area of the segmented regions is larger than that of the complete field of the image data. InFIG. 22 g there are seven shown overlapping regions where each identifying numeral is shown in the center of its region. In a further embodiment shown inFIGS. 22 h and 22 i, the segmentation consists of sample regions (shown as cross-hatched) within the complete field of the image data. In another embodiment, the sampled regions can be based on a preloaded user template that, for example, identifies regions of interest such as a signature region and/or a bar code region, in for example, a shipping label. - In one embodiment, the segmentation process is used to identify the location of a signature in image data the might include additional elements such as dataforms including bar code dataforms, text, graphics, images and the like. In one such embodiment the
generalized classifier circuit 21216 classifies the contents of each region of the segmented image data. The region containing the signature is then extracted by the signaturedata processing circuit 21218. In one embodiment if multiple regions are indicated as containing signature data, the signaturedata processing circuit 21218 analyzes the arrangement of these regions to identify the region most likely to contain the image data. In a further embodiment when multiple regions are indicated as containing signature data, the image processing andanalysis circuit 21208 establishes a feedback loop where additional segmented regions are generated and analyzed until a single segmented region containing signature data is located. - Additional image processing operations which may be carried out by
optical reader 100 are described in U.S. patent application Ser. No. 10/958,779, filed Oct. 5, 2004 entitled, “System And Method To Automatically Discriminate Between A Signature And A Barcode” and incorporated herein by reference in its entirety. - Various applications which may be carried out by any of the
optical readers 100 that have been described herein have been described with reference toFIGS. 10 , 11, 12 a and 12 b. Another application which can be carried out with anyoptical reader 100 described herein is described with reference toFIGS. 13 a-13 e. InFIG. 13 a amotor vehicle 1282 is shown which may be a delivery vehicle or a passenger vehicle.Vehicle 1282 has alicense plate 1314, a vehicle identification number (VIN)sticker 1306, typically located on the driver's side door jam. TheVIN sticker 1306 carries a printedVIN number 1308 and abar code symbol 1310. A VIN number is an alphanumeric unique vehicle identification number assigned at the time of manufacture of the vehicle.Vehicle 1282 may further include a VIN plate 1314 (FIG. 13 c) carrying the characters of the VIN number etched on a metal plate and located under thevehicle windshield 1351, and avehicle registration sticker 1320.Vehicle 1282 has a plurality of machine readable vehicle identifiers. Specifically, the characters oflicense plate 1284 can be OCR decoded by optical reader. Further,VIN sticker 1308 has aVIN bar code 1310 andregistration sticker 1320 which may include a plurality ofbar code symbols vehicle 1282. The characters etched onVIN plate 1314 can also be subject to OCR decoding byoptical reader 100. Further, the VIN characters ofVIN sticker 1306 can be subject to OCR decoding byoptical reader 100. It may be advantageous to utilize an optical reader including lightpolarizing pixels 250P having lightpolarizing filter elements 261 when readingVIN plate 1314 given that specular reflection read conditions are more prevalent when decoding indicia encoded by etching on metal surface. - In an application for utilizing
optical reader 100 relative tovehicle 1282, several identifiers ofvehicle 1282 may be decoded and several color pictures ofvehicle 1282 may be taken. The decoded message data together with the color picture data may then be uploaded to a remote server 184 (FIG. 10 ) which archives and creates accessible web pages containing reports summarizing the identifier and picture information. In one application LAN 170 (FIG. 10 ) is a LAN at an automobile insurance claim center,LAN 185 is a distant data archiving center operated by the automobile insurance provider andLAN 2170 is a LAN apart fromLAN 170 andLAN 185 and may be located, e.g., at a claim center of the insurance provider other than the claim center in whichLAN 170 is located. -
Optical reader 100 may be configured so that when an operator actuates a designated user interface control button such as button 3158 (FIG. 9 b) an auto insurance application form 1362 is displayed ondisplay 504 which aids an operator ofoptical reader 100 in entering data intoreader 100. Form 1362 first prompts an operator to read several machine readable identifiers ofvehicle 1282. Form 1362 prompts an operator to read VINbar code symbol 1310, then characters ofVIN plate 1314, then the first registration stickerbar code symbol 1310, then the second registration stickerbar code symbol 1324, then the character of thelicense plate 1284. The text corresponding to each identifier may be highlighted when data corresponding to the identifier is read. When data corresponding to identifier decode section 1363 of form 1362 is being entered,optical reader 100 is in a decode mode of operation such that actuation oftrigger 216 causesoptical reader 100 to obtain a decode frame atstep 1204 and transfer the decode frame to decodecircuit 1702. The decode frame may contain monochrome image data read from a hybrid monochromeimage sensor array optical reader 100 has separate picture taking and decoding imaging assemblies as described in connection withFIGS. 17 a-17 g, the decode frame atstep 1204 is obtained by actuation of the imaging assembly within block 598 (FIG. 17 a). When entry of decoded vehicle identifier information is complete, an operator toggles to line 1365 and clicks an appropriate key ofkeyboard 508 to indicate that identifier decoding is complete. Form 1362 then prompts an operator to take pictures ofvehicle 1282 for purposes of making a record of the damage tovehicle 1282. The inventor discovered that the incorporation of color filter elements into animage sensor array 182 ofoptical reader 100 facilitates the obtaining of visual display frames of image data that accurately record damage to a vehicle. With visual display color frames of image data corresponding tovehicle 1282 being stored and/or displayed for visual display, damage tovehicle 1282 can readily be assessed by visual inspection of the visual display frames when displayed on adisplay optical reader 100 is in a picture taking mode such that actuation oftrigger 216 causes a visual display frame of image data to be obtained at step 1404 (FIG. 14 c). The visual display frame of image data may be output to e.g., a storage device and/or a display device. When data corresponding to form section 1364 is being entered, an operator may useoptical reader 100 to take several color pictures of damagedarea 1370 ofvehicle 1282. While executing obtainstep 1404,control circuit 552 may selectively read out color image data from colorsensitive pixels 250C as described herein and possibly utilize monochrome image data for enhancement of the information content of the color image data. Whereoptical reader 100 includes a pair of imaging assemblies as described in connection withFIGS. 17 a-17 g,control circuit 552 atstep 1404 may actuate colorimage sensor array 182D for execution of obtainstep 1404. When an operator inputs a confirmation that all necessary pictures ofvehicle 1282 have been taken by toggling to line 1367 and clicking an appropriate key ofkeyboard 508,control circuit 552, which may be incorporated in hand heldhousing 101, may format obtained visual display color frames of image data in one or more suitably image file formats, (e.g., .BMP, .TIFF, .PDF, .JPG, .GIF) assemble all the collected decoded vehicle identifier data and all of the visual display color frames of image data corresponding tovehicle 1282 into a transaction data set, and send the transaction data set to distant remote server 184.Control circuit 552 may date/time stamp the transaction data set on sending. The File Transfer Protocol (FTP) may be utilized to send the transaction data set or another suitable file transferring protocol configured to carry associated decoded vehicle identifier data (such as decoded VIN bar code data and decode vehicle registration bar code data) and color image data. Server 184 may store the received transaction data set into a database as indicated bydatabase 187 including similar information from other vehicles at other claim centers. Server 184 may be configured to create viewable web pages summarizing the transaction set data (e.g., the date/time stamped combined VIN, registration number, license plate number and record-of damage visual display color frames of image data). These web pages may be viewed using any PC in communication with IP network, e.g., PC 172 and PC 2172. - While the present invention has necessarily been described with reference to a number of specific embodiments, it will be understood that the time, spirit, and scope of the present invention should be determined only with reference to the following claims:
Claims (10)
1. An optical reading device, said optical reading device comprising:
a solid state image sensor array having a plurality of pixels, the plurality of pixels being formed in a plurality of rows of pixels, wherein the solid state image sensor array includes a first subset of pixels, a second subset of pixels, and a third subset of pixels, the first subset of pixels being monochrome pixels, the second subset of pixels being color sensitive pixels, and a third subset of pixels being light polarizing pixels;
an imaging lens focusing an image onto said solid state image sensor array;
a hand held housing encapsulating said solid state image sensor array;
wherein said optical reading device is operative to read out a frame of image data from the solid state image sensor array, the frame of image data having image data corresponding to at least one of the first subset of pixels, the second subset of pixels, and the third subset of pixels.
2. The optical reading device of claim 1 , wherein the optical reading device is operative so that for read out of the frame of image data, the optical reading device reads out image data from pixels of more than one of the first, second, and third subset of pixels.
3. The optical reading device of claim 1 , wherein the optical reading device is operative so that for read out of the frame of image data, the optical reading device selectively addresses pixels of the first subset of pixels and selectively reads out image data from pixels of the first subset of pixels.
4. The optical reading device of claim 1 , wherein the optical reading device is operative so that for read out of the frame of image data, the optical reading device selectively addresses pixels of the second subset of pixels and selectively reads out image data from pixels of the second subset of pixels.
5. The optical reading device of claim 1 , wherein the optical reading device is operative so that for read out of the frame of image data, the optical reading device selectively addresses pixels of the third subset of pixels and selectively reads out image data from pixels of the third subset of pixels.
6. The optical reading device of claim 1 , wherein the optical reading device is operative so that for read out of the frame of image data, the optical reading device reads out image data from pixels of more than one of the first, second, and third subset of pixels, and wherein the optical reading device is operative to selectively extract from the frame of image data image data corresponding to the third subset of pixels.
7. The optical reading device of claim 1 , wherein said optical reading device includes separate reset control lines for facilitating independent reset of pixels of said first subset of pixels, said second subset of pixels, and said third subset of pixels.
8. The optical reading device of claim 1 , wherein the optical reading device is operative to sense a saturation condition.
9. The optical reading device of claim 1 , wherein the optical reading device is operative to sense a saturation condition by analysis of image data.
10. The optical reading device of claim 1 , wherein the optical reading device includes an illumination assembly.
Priority Applications (22)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/853,090 US8196839B2 (en) | 2005-06-03 | 2010-08-09 | Optical reader having reduced specular reflection read failures |
US13/493,348 US8720785B2 (en) | 2005-06-03 | 2012-06-11 | Apparatus having hybrid monochrome and color image sensor array |
US14/221,874 US9058527B2 (en) | 2005-06-03 | 2014-03-21 | Apparatus having hybrid monochrome and color image sensor array |
US14/684,609 US9454686B2 (en) | 2005-06-03 | 2015-04-13 | Apparatus having hybrid monochrome and color image sensor array |
US15/244,683 US10002272B2 (en) | 2005-06-03 | 2016-08-23 | Apparatus having hybrid monochrome and color image sensor array |
US15/980,213 US10691907B2 (en) | 2005-06-03 | 2018-05-15 | Apparatus having hybrid monochrome and color image sensor array |
US16/874,217 US10949634B2 (en) | 2005-06-03 | 2020-05-14 | Apparatus having hybrid monochrome and color image sensor array |
US17/167,464 US11238251B2 (en) | 2005-06-03 | 2021-02-04 | Apparatus having hybrid monochrome and color image sensor array |
US17/302,324 US11238252B2 (en) | 2005-06-03 | 2021-04-30 | Apparatus having hybrid monochrome and color image sensor array |
US17/523,500 US20220067317A1 (en) | 2005-06-03 | 2021-11-10 | Apparatus having hybrid monochrome and color image sensor array |
US17/523,231 US11625550B2 (en) | 2005-06-03 | 2021-11-10 | Apparatus having hybrid monochrome and color image sensor array |
US17/825,742 US20220284204A1 (en) | 2005-06-03 | 2022-05-26 | Apparatus having hybrid monochrome and color image sensor array |
US17/825,726 US11604933B2 (en) | 2005-06-03 | 2022-05-26 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,708 US12026580B2 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,682 US12073283B2 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,696 US12001913B2 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,712 US20230147821A1 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,689 US12020111B2 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/168,191 US20230196041A1 (en) | 2005-06-03 | 2023-02-13 | Apparatus having hybrid monochrome and color image sensor array |
US18/181,946 US12001914B2 (en) | 2005-06-03 | 2023-03-10 | Apparatus having hybrid monochrome and color image sensor array |
US18/304,065 US20230252252A1 (en) | 2005-06-03 | 2023-04-20 | Apparatus having hybrid monochrome and color image sensor array |
US18/304,088 US12236312B2 (en) | 2005-06-03 | 2023-04-20 | Apparatus having hybrid monochrome and color image sensor array |
Applications Claiming Priority (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US68760605P | 2005-06-03 | 2005-06-03 | |
US69026805P | 2005-06-14 | 2005-06-14 | |
US69289005P | 2005-06-22 | 2005-06-22 | |
US69437105P | 2005-06-27 | 2005-06-27 | |
US11/445,930 US7770799B2 (en) | 2005-06-03 | 2006-06-02 | Optical reader having reduced specular reflection read failures |
US12/853,090 US8196839B2 (en) | 2005-06-03 | 2010-08-09 | Optical reader having reduced specular reflection read failures |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/445,930 Division US7770799B2 (en) | 2005-06-03 | 2006-06-02 | Optical reader having reduced specular reflection read failures |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/493,348 Division US8720785B2 (en) | 2005-06-03 | 2012-06-11 | Apparatus having hybrid monochrome and color image sensor array |
Publications (2)
Publication Number | Publication Date |
---|---|
US20110049245A1 true US20110049245A1 (en) | 2011-03-03 |
US8196839B2 US8196839B2 (en) | 2012-06-12 |
Family
ID=37572408
Family Applications (23)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/445,930 Active 2028-08-18 US7770799B2 (en) | 2005-06-03 | 2006-06-02 | Optical reader having reduced specular reflection read failures |
US12/853,090 Active US8196839B2 (en) | 2005-06-03 | 2010-08-09 | Optical reader having reduced specular reflection read failures |
US13/493,348 Active US8720785B2 (en) | 2005-06-03 | 2012-06-11 | Apparatus having hybrid monochrome and color image sensor array |
US14/221,874 Active US9058527B2 (en) | 2005-06-03 | 2014-03-21 | Apparatus having hybrid monochrome and color image sensor array |
US14/684,609 Active US9454686B2 (en) | 2005-06-03 | 2015-04-13 | Apparatus having hybrid monochrome and color image sensor array |
US15/244,683 Active US10002272B2 (en) | 2005-06-03 | 2016-08-23 | Apparatus having hybrid monochrome and color image sensor array |
US15/980,213 Active US10691907B2 (en) | 2005-06-03 | 2018-05-15 | Apparatus having hybrid monochrome and color image sensor array |
US16/874,217 Active US10949634B2 (en) | 2005-06-03 | 2020-05-14 | Apparatus having hybrid monochrome and color image sensor array |
US17/167,464 Active US11238251B2 (en) | 2005-06-03 | 2021-02-04 | Apparatus having hybrid monochrome and color image sensor array |
US17/302,324 Active US11238252B2 (en) | 2005-06-03 | 2021-04-30 | Apparatus having hybrid monochrome and color image sensor array |
US17/523,231 Active US11625550B2 (en) | 2005-06-03 | 2021-11-10 | Apparatus having hybrid monochrome and color image sensor array |
US17/523,500 Pending US20220067317A1 (en) | 2005-06-03 | 2021-11-10 | Apparatus having hybrid monochrome and color image sensor array |
US17/825,726 Active US11604933B2 (en) | 2005-06-03 | 2022-05-26 | Apparatus having hybrid monochrome and color image sensor array |
US17/825,742 Pending US20220284204A1 (en) | 2005-06-03 | 2022-05-26 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,696 Active US12001913B2 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,682 Active 2026-10-15 US12073283B2 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,708 Active US12026580B2 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,689 Active US12020111B2 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,712 Pending US20230147821A1 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/168,191 Pending US20230196041A1 (en) | 2005-06-03 | 2023-02-13 | Apparatus having hybrid monochrome and color image sensor array |
US18/181,946 Active US12001914B2 (en) | 2005-06-03 | 2023-03-10 | Apparatus having hybrid monochrome and color image sensor array |
US18/304,088 Active US12236312B2 (en) | 2005-06-03 | 2023-04-20 | Apparatus having hybrid monochrome and color image sensor array |
US18/304,065 Pending US20230252252A1 (en) | 2005-06-03 | 2023-04-20 | Apparatus having hybrid monochrome and color image sensor array |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/445,930 Active 2028-08-18 US7770799B2 (en) | 2005-06-03 | 2006-06-02 | Optical reader having reduced specular reflection read failures |
Family Applications After (21)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/493,348 Active US8720785B2 (en) | 2005-06-03 | 2012-06-11 | Apparatus having hybrid monochrome and color image sensor array |
US14/221,874 Active US9058527B2 (en) | 2005-06-03 | 2014-03-21 | Apparatus having hybrid monochrome and color image sensor array |
US14/684,609 Active US9454686B2 (en) | 2005-06-03 | 2015-04-13 | Apparatus having hybrid monochrome and color image sensor array |
US15/244,683 Active US10002272B2 (en) | 2005-06-03 | 2016-08-23 | Apparatus having hybrid monochrome and color image sensor array |
US15/980,213 Active US10691907B2 (en) | 2005-06-03 | 2018-05-15 | Apparatus having hybrid monochrome and color image sensor array |
US16/874,217 Active US10949634B2 (en) | 2005-06-03 | 2020-05-14 | Apparatus having hybrid monochrome and color image sensor array |
US17/167,464 Active US11238251B2 (en) | 2005-06-03 | 2021-02-04 | Apparatus having hybrid monochrome and color image sensor array |
US17/302,324 Active US11238252B2 (en) | 2005-06-03 | 2021-04-30 | Apparatus having hybrid monochrome and color image sensor array |
US17/523,231 Active US11625550B2 (en) | 2005-06-03 | 2021-11-10 | Apparatus having hybrid monochrome and color image sensor array |
US17/523,500 Pending US20220067317A1 (en) | 2005-06-03 | 2021-11-10 | Apparatus having hybrid monochrome and color image sensor array |
US17/825,726 Active US11604933B2 (en) | 2005-06-03 | 2022-05-26 | Apparatus having hybrid monochrome and color image sensor array |
US17/825,742 Pending US20220284204A1 (en) | 2005-06-03 | 2022-05-26 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,696 Active US12001913B2 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,682 Active 2026-10-15 US12073283B2 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,708 Active US12026580B2 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,689 Active US12020111B2 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/050,712 Pending US20230147821A1 (en) | 2005-06-03 | 2022-10-28 | Apparatus having hybrid monochrome and color image sensor array |
US18/168,191 Pending US20230196041A1 (en) | 2005-06-03 | 2023-02-13 | Apparatus having hybrid monochrome and color image sensor array |
US18/181,946 Active US12001914B2 (en) | 2005-06-03 | 2023-03-10 | Apparatus having hybrid monochrome and color image sensor array |
US18/304,088 Active US12236312B2 (en) | 2005-06-03 | 2023-04-20 | Apparatus having hybrid monochrome and color image sensor array |
US18/304,065 Pending US20230252252A1 (en) | 2005-06-03 | 2023-04-20 | Apparatus having hybrid monochrome and color image sensor array |
Country Status (1)
Country | Link |
---|---|
US (23) | US7770799B2 (en) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100259638A1 (en) * | 2009-04-09 | 2010-10-14 | Hand Held Products, Inc. | Imaging terminal having color correction |
US20100316291A1 (en) * | 2009-06-11 | 2010-12-16 | Shulan Deng | Imaging terminal having data compression |
US20110135144A1 (en) * | 2009-07-01 | 2011-06-09 | Hand Held Products, Inc. | Method and system for collecting voice and image data on a remote device and coverting the combined data |
US20110163166A1 (en) * | 2005-03-11 | 2011-07-07 | Hand Held Products, Inc. | Image reader comprising cmos based image sensor array |
US8196839B2 (en) | 2005-06-03 | 2012-06-12 | Hand Held Products, Inc. | Optical reader having reduced specular reflection read failures |
US8600167B2 (en) | 2010-05-21 | 2013-12-03 | Hand Held Products, Inc. | System for capturing a document in an image signal |
US8636215B2 (en) | 2011-06-27 | 2014-01-28 | Hand Held Products, Inc. | Decodable indicia reading terminal with optical filter |
US8640960B2 (en) | 2011-06-27 | 2014-02-04 | Honeywell International Inc. | Optical filter for image and barcode scanning |
US8646694B2 (en) | 2008-12-16 | 2014-02-11 | Hand Held Products, Inc. | Indicia reading terminal including frame processing |
US8720781B2 (en) | 2005-03-11 | 2014-05-13 | Hand Held Products, Inc. | Image reader having image sensor array |
US8720784B2 (en) | 2005-06-03 | 2014-05-13 | Hand Held Products, Inc. | Digital picture taking optical reader having hybrid monochrome and color image sensor array |
US8727223B2 (en) | 2006-06-09 | 2014-05-20 | Hand Held Products, Inc. | Indicia reading apparatus having image sensor array |
US8881983B2 (en) | 2011-12-13 | 2014-11-11 | Honeywell International Inc. | Optical readers and methods employing polarization sensing of light from decodable indicia |
US8978983B2 (en) | 2012-06-01 | 2015-03-17 | Honeywell International, Inc. | Indicia reading apparatus having sequential row exposure termination times |
US8978981B2 (en) | 2012-06-27 | 2015-03-17 | Honeywell International Inc. | Imaging apparatus having imaging lens |
US8985459B2 (en) | 2011-06-30 | 2015-03-24 | Metrologic Instruments, Inc. | Decodable indicia reading terminal with combined illumination |
US9047531B2 (en) | 2010-05-21 | 2015-06-02 | Hand Held Products, Inc. | Interactive user interface for capturing a document in an image signal |
US9251392B2 (en) | 2012-06-01 | 2016-02-02 | Honeywell International, Inc. | Indicia reading apparatus |
Families Citing this family (153)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8352400B2 (en) | 1991-12-23 | 2013-01-08 | Hoffberg Steven M | Adaptive pattern recognition based controller apparatus and method and human-factored interface therefore |
US7966078B2 (en) | 1999-02-01 | 2011-06-21 | Steven Hoffberg | Network media appliance system and method |
US6666377B1 (en) | 2000-07-18 | 2003-12-23 | Scott C. Harris | Bar code data entry device |
US7748620B2 (en) | 2002-01-11 | 2010-07-06 | Hand Held Products, Inc. | Transaction terminal including imaging module |
US9092841B2 (en) * | 2004-06-09 | 2015-07-28 | Cognex Technology And Investment Llc | Method and apparatus for visual detection and inspection of objects |
US20030222147A1 (en) | 2002-06-04 | 2003-12-04 | Hand Held Products, Inc. | Optical reader having a plurality of imaging modules |
US8596542B2 (en) | 2002-06-04 | 2013-12-03 | Hand Held Products, Inc. | Apparatus operative for capture of image data |
US7707039B2 (en) | 2004-02-15 | 2010-04-27 | Exbiblio B.V. | Automatic modification of web pages |
US8442331B2 (en) | 2004-02-15 | 2013-05-14 | Google Inc. | Capturing text from rendered documents using supplemental information |
US7812860B2 (en) | 2004-04-01 | 2010-10-12 | Exbiblio B.V. | Handheld device for capturing text from both a document printed on paper and a document displayed on a dynamic display device |
US8799303B2 (en) | 2004-02-15 | 2014-08-05 | Google Inc. | Establishing an interactive environment for rendered documents |
US10635723B2 (en) | 2004-02-15 | 2020-04-28 | Google Llc | Search engines and systems with handheld document data capture devices |
US20060041484A1 (en) | 2004-04-01 | 2006-02-23 | King Martin T | Methods and systems for initiating application processes by data capture from rendered documents |
US8521772B2 (en) | 2004-02-15 | 2013-08-27 | Google Inc. | Document enhancement system and method |
US20080313172A1 (en) | 2004-12-03 | 2008-12-18 | King Martin T | Determining actions involving captured information and electronic content associated with rendered documents |
WO2008028674A2 (en) * | 2006-09-08 | 2008-03-13 | Exbiblio B.V. | Optical scanners, such as hand-held optical scanners |
US9116890B2 (en) | 2004-04-01 | 2015-08-25 | Google Inc. | Triggering actions in response to optically or acoustically capturing keywords from a rendered document |
US7894670B2 (en) | 2004-04-01 | 2011-02-22 | Exbiblio B.V. | Triggering actions in response to optically or acoustically capturing keywords from a rendered document |
US9143638B2 (en) | 2004-04-01 | 2015-09-22 | Google Inc. | Data capture from rendered documents using handheld device |
US9008447B2 (en) | 2004-04-01 | 2015-04-14 | Google Inc. | Method and system for character recognition |
US8621349B2 (en) | 2004-04-01 | 2013-12-31 | Google Inc. | Publishing techniques for adding value to a rendered document |
US8793162B2 (en) | 2004-04-01 | 2014-07-29 | Google Inc. | Adding information or functionality to a rendered document via association with an electronic counterpart |
US8146156B2 (en) | 2004-04-01 | 2012-03-27 | Google Inc. | Archive of text captures from rendered documents |
US20070300142A1 (en) | 2005-04-01 | 2007-12-27 | King Martin T | Contextual dynamic advertising based upon captured rendered text |
US7990556B2 (en) | 2004-12-03 | 2011-08-02 | Google Inc. | Association of a portable scanner with input/output and storage devices |
US8713418B2 (en) | 2004-04-12 | 2014-04-29 | Google Inc. | Adding value to a rendered document |
US9460346B2 (en) | 2004-04-19 | 2016-10-04 | Google Inc. | Handheld device for capturing text from both a document printed on paper and a document displayed on a dynamic display device |
US8620083B2 (en) | 2004-12-03 | 2013-12-31 | Google Inc. | Method and system for character recognition |
US8489624B2 (en) | 2004-05-17 | 2013-07-16 | Google, Inc. | Processing techniques for text capture from a rendered document |
US8874504B2 (en) | 2004-12-03 | 2014-10-28 | Google Inc. | Processing techniques for visual capture data from a rendered document |
US8127247B2 (en) | 2004-06-09 | 2012-02-28 | Cognex Corporation | Human-machine-interface and method for manipulating data in a machine vision system |
US8243986B2 (en) | 2004-06-09 | 2012-08-14 | Cognex Technology And Investment Corporation | Method and apparatus for automatic visual event detection |
US20050276445A1 (en) | 2004-06-09 | 2005-12-15 | Silver William M | Method and apparatus for automatic visual detection, recording, and retrieval of events |
US8891852B2 (en) | 2004-06-09 | 2014-11-18 | Cognex Technology And Investment Corporation | Method and apparatus for configuring and testing a machine vision detector |
US8346620B2 (en) | 2004-07-19 | 2013-01-01 | Google Inc. | Automatic modification of web pages |
US7204420B2 (en) * | 2004-08-31 | 2007-04-17 | Symbol Technologies, Inc. | Scanner and method for eliminating specular reflection |
US7720315B2 (en) | 2004-11-12 | 2010-05-18 | Cognex Technology And Investment Corporation | System and method for displaying and using non-numeric graphic elements to control and monitor a vision system |
US9292187B2 (en) | 2004-11-12 | 2016-03-22 | Cognex Corporation | System, method and graphical user interface for displaying and controlling vision system operating parameters |
US7636449B2 (en) | 2004-11-12 | 2009-12-22 | Cognex Technology And Investment Corporation | System and method for assigning analysis parameters to vision detector using a graphical interface |
US7963448B2 (en) | 2004-12-22 | 2011-06-21 | Cognex Technology And Investment Corporation | Hand held machine vision method and apparatus |
US9552506B1 (en) | 2004-12-23 | 2017-01-24 | Cognex Technology And Investment Llc | Method and apparatus for industrial identification mark verification |
US20070063048A1 (en) * | 2005-09-14 | 2007-03-22 | Havens William H | Data reader apparatus having an adaptive lens |
US8781532B2 (en) * | 2005-09-19 | 2014-07-15 | Google Inc. | Customized data retrieval applications for mobile devices providing interpretation of markup language data |
US20070153121A1 (en) * | 2005-11-18 | 2007-07-05 | Juan Pertierra | Video data acquisition system |
JP4442624B2 (en) * | 2006-06-22 | 2010-03-31 | 株式会社デンソーウェーブ | Optical information reader |
US8108176B2 (en) | 2006-06-29 | 2012-01-31 | Cognex Corporation | Method and apparatus for verifying two dimensional mark quality |
US7984854B2 (en) * | 2006-07-17 | 2011-07-26 | Cognex Corporation | Method and apparatus for multiplexed symbol decoding |
JP2008118434A (en) * | 2006-11-06 | 2008-05-22 | Fujifilm Corp | Solid-state imaging device and imaging apparatus |
US8403225B2 (en) * | 2006-11-17 | 2013-03-26 | Hand Held Products, Inc. | Vehicle license plate indicia scanning |
US8583100B2 (en) | 2007-01-25 | 2013-11-12 | Adc Telecommunications, Inc. | Distributed remote base station system |
US8737454B2 (en) | 2007-01-25 | 2014-05-27 | Adc Telecommunications, Inc. | Modular wireless communications platform |
US8794526B2 (en) * | 2007-06-04 | 2014-08-05 | Hand Held Products, Inc. | Indicia reading terminal processing plurality of frames of image data responsively to trigger signal activation |
US7762464B2 (en) * | 2007-06-28 | 2010-07-27 | Symbol Technologies, Inc. | Control of specular reflection in imaging reader |
US8950673B2 (en) * | 2007-08-30 | 2015-02-10 | Symbol Technologies, Inc. | Imaging system for reading target with multiple symbols |
US8638363B2 (en) | 2009-02-18 | 2014-01-28 | Google Inc. | Automatically capturing information, such as capturing information using a document-aware device |
US9734376B2 (en) * | 2007-11-13 | 2017-08-15 | Cognex Corporation | System and method for reading patterns using multiple image frames |
US7926721B2 (en) * | 2008-03-26 | 2011-04-19 | Hand Held Products, Inc. | Processing scheme for an indicia reader |
CN102187665A (en) * | 2008-09-04 | 2011-09-14 | 独立行政法人科学技术振兴机构 | Frame rate converting apparatus and corresponding point estimating apparatus, corresponding point estimating method and corresponding point estimating program |
US8601421B2 (en) * | 2008-10-16 | 2013-12-03 | Lockheed Martin Corporation | Small, adaptable, real-time, scalable image processing chip |
US20100123009A1 (en) * | 2008-11-20 | 2010-05-20 | Datalogic Scanning Inc. | High-resolution interpolation for color-imager-based optical code readers |
US8783573B2 (en) | 2008-12-02 | 2014-07-22 | Hand Held Products, Inc. | Indicia reading terminal having plurality of optical assemblies |
WO2010088465A1 (en) * | 2009-02-02 | 2010-08-05 | Gentex Corporation | Improved digital image processing and systems incorporating the same |
US8118226B2 (en) * | 2009-02-11 | 2012-02-21 | Datalogic Scanning, Inc. | High-resolution optical code imaging using a color imager |
US8800874B2 (en) | 2009-02-20 | 2014-08-12 | Datalogic ADC, Inc. | Systems and methods of optical code reading using a color imager |
US8447066B2 (en) | 2009-03-12 | 2013-05-21 | Google Inc. | Performing actions based on capturing information from rendered documents, such as documents under copyright |
WO2010105246A2 (en) | 2009-03-12 | 2010-09-16 | Exbiblio B.V. | Accessing resources based on capturing information from a rendered document |
US9183425B2 (en) * | 2009-04-09 | 2015-11-10 | Hand Held Products, Inc. | Image sensor pixel array having output response curve including logarithmic pattern for image sensor based terminal |
US8295601B2 (en) * | 2009-08-12 | 2012-10-23 | Hand Held Products, Inc. | Indicia reading terminal having multiple exposure periods and methods for same |
US20110080500A1 (en) * | 2009-10-05 | 2011-04-07 | Hand Held Products, Inc. | Imaging terminal, imaging sensor having multiple reset and/or multiple read mode and methods for operating the same |
JP5310483B2 (en) * | 2009-10-28 | 2013-10-09 | 株式会社リコー | Imaging device |
US9081799B2 (en) | 2009-12-04 | 2015-07-14 | Google Inc. | Using gestalt information to identify locations in printed information |
US9323784B2 (en) | 2009-12-09 | 2016-04-26 | Google Inc. | Image search using text-based elements within the contents of images |
US20110163165A1 (en) * | 2010-01-07 | 2011-07-07 | Metrologic Instruments, Inc. | Terminal having illumination and focus control |
US8640958B2 (en) | 2010-01-21 | 2014-02-04 | Honeywell International, Inc. | Indicia reading terminal including optical filter |
US8346478B2 (en) * | 2010-03-25 | 2013-01-01 | GM Global Technology Operations LLC | Location and mobile-adaptation of wireless access-points using map-based navigation |
US8345117B2 (en) * | 2010-06-30 | 2013-01-01 | Hand Held Products, Inc. | Terminal outputting monochrome image data and color image data |
JP5275312B2 (en) * | 2010-09-22 | 2013-08-28 | 株式会社東芝 | Image sensor module and image sensor |
US8387881B2 (en) | 2010-12-01 | 2013-03-05 | Hand Held Products, Inc. | Terminal with screen reading mode |
WO2012074526A1 (en) * | 2010-12-02 | 2012-06-07 | 3M Innovative Properties Company | Methods and systems for enhancing read accuracy in automated license plate reader systems |
WO2012075608A1 (en) | 2010-12-09 | 2012-06-14 | Metrologic Instruments, Inc. | Indicia encoding system with integrated purchase and payment information |
US9418270B2 (en) | 2011-01-31 | 2016-08-16 | Hand Held Products, Inc. | Terminal with flicker-corrected aimer and alternating illumination |
US8678286B2 (en) | 2011-01-31 | 2014-03-25 | Honeywell Scanning & Mobility | Method and apparatus for reading optical indicia using a plurality of data sources |
US8520080B2 (en) | 2011-01-31 | 2013-08-27 | Hand Held Products, Inc. | Apparatus, system, and method of use of imaging assembly on mobile terminal |
US8561903B2 (en) | 2011-01-31 | 2013-10-22 | Hand Held Products, Inc. | System operative to adaptively select an image sensor for decodable indicia reading |
US8537245B2 (en) | 2011-03-04 | 2013-09-17 | Hand Held Products, Inc. | Imaging and decoding device with quantum dot imager |
BR112013028972A2 (en) | 2011-05-12 | 2017-02-07 | Olive Medical Corp | pixel frame area optimization using a stacking scheme for a hybrid image sensor with minimal vertical interconnect elements |
US8626600B2 (en) | 2011-09-15 | 2014-01-07 | Ebay Inc. | Personalized item trading card generation and management |
US8646692B2 (en) | 2011-09-30 | 2014-02-11 | Hand Held Products, Inc. | Devices and methods employing dual target auto exposure |
US8608071B2 (en) | 2011-10-17 | 2013-12-17 | Honeywell Scanning And Mobility | Optical indicia reading terminal with two image sensors |
US8629926B2 (en) | 2011-11-04 | 2014-01-14 | Honeywell International, Inc. | Imaging apparatus comprising image sensor array having shared global shutter circuitry |
US8526720B2 (en) * | 2011-11-17 | 2013-09-03 | Honeywell International, Inc. | Imaging terminal operative for decoding |
US8628013B2 (en) | 2011-12-13 | 2014-01-14 | Honeywell International Inc. | Apparatus comprising image sensor array and illumination control |
US9651499B2 (en) | 2011-12-20 | 2017-05-16 | Cognex Corporation | Configurable image trigger for a vision system and method for using the same |
US9177983B2 (en) * | 2012-01-23 | 2015-11-03 | Omnivision Technologies, Inc. | Image sensor with optical filters having alternating polarization for 3D imaging |
CN104115211B (en) | 2012-02-14 | 2017-09-22 | 金泰克斯公司 | High dynamic range imaging system |
US8777108B2 (en) | 2012-03-23 | 2014-07-15 | Honeywell International, Inc. | Cell phone reading mode using image timer |
US20130249811A1 (en) * | 2012-03-23 | 2013-09-26 | Microsoft Corporation | Controlling a device with visible light |
MX2014011647A (en) | 2012-03-30 | 2014-10-24 | 3M Innovative Properties Co | Retroreflective articles having a machine-readable code. |
US9158954B2 (en) * | 2012-05-15 | 2015-10-13 | Intermec Ip, Corp. | Systems and methods to read machine-readable symbols |
US9437046B2 (en) * | 2012-06-06 | 2016-09-06 | Sodyo Ltd. | Anchors for location-based navigation and augmented reality applications |
US8944313B2 (en) | 2012-06-29 | 2015-02-03 | Honeywell International Inc. | Computer configured to display multimedia content |
JP6461790B2 (en) | 2012-07-26 | 2019-01-30 | デピュー シンセス プロダクツ, インコーポレーテッドDePuy Synthes Products, Inc. | Camera system using minimum area monolithic CMOS image sensor |
CN104619237B (en) | 2012-07-26 | 2018-03-30 | 德普伊辛迪斯制品公司 | The pulse modulated illumination schemes of YCBCR in light deficiency environment |
EP2992810B1 (en) | 2012-07-26 | 2021-11-10 | DePuy Synthes Products, Inc. | Continuous video in a light deficient environment |
CN104010183B (en) * | 2012-11-21 | 2017-03-01 | 豪威科技股份有限公司 | Array system including at least one bayer-like video camera and the method for association |
US9038903B2 (en) | 2012-11-30 | 2015-05-26 | Symbol Technologies, Inc. | Method and apparatus for controlling illumination |
CA2906975A1 (en) | 2013-03-15 | 2014-09-18 | Olive Medical Corporation | Minimize image sensor i/o and conductor counts in endoscope applications |
CN105246394B (en) | 2013-03-15 | 2018-01-12 | 德普伊新特斯产品公司 | It is synchronous without the imaging sensor of input clock and data transfer clock |
JP6422937B2 (en) | 2013-03-15 | 2018-11-14 | デピュイ・シンセス・プロダクツ・インコーポレイテッド | Endoscope sensing in a light controlled environment |
AU2014233193B2 (en) | 2013-03-15 | 2018-11-01 | DePuy Synthes Products, Inc. | Controlling the integral light energy of a laser pulse |
WO2014144947A1 (en) | 2013-03-15 | 2014-09-18 | Olive Medical Corporation | Super resolution and color motion artifact correction in a pulsed color imaging system |
WO2014145008A2 (en) | 2013-03-15 | 2014-09-18 | Olive Medical Corporation | Viewing trocar with intergrated prism for use with angled endoscope |
US9600703B2 (en) | 2013-03-15 | 2017-03-21 | Cognex Corporation | Systems and methods for sorting image acquisition settings for pattern stitching and decoding using multiple captured images |
US9104932B2 (en) | 2013-03-15 | 2015-08-11 | Cognex Corporation | Systems and methods for pattern stitching and decoding using multiple captured images |
US9033237B1 (en) | 2013-10-26 | 2015-05-19 | Symbol Technologies, Inc. | Decoding DPM indicia with polarized illumination |
US9237847B2 (en) | 2014-02-11 | 2016-01-19 | Welch Allyn, Inc. | Ophthalmoscope device |
US9211064B2 (en) | 2014-02-11 | 2015-12-15 | Welch Allyn, Inc. | Fundus imaging system |
WO2015143453A1 (en) | 2014-03-21 | 2015-09-24 | Olive Medical Corporation | Card edge connector for an imaging sensor |
US9367782B2 (en) | 2014-05-29 | 2016-06-14 | Konica Minolta Laboratory U.S.A., Inc. | High capacity 2D color barcode and method for decoding the same |
JP6363903B2 (en) * | 2014-07-31 | 2018-07-25 | 株式会社キーエンス | Optical information reader |
JP6519142B2 (en) * | 2014-10-28 | 2019-05-29 | 株式会社リコー | Processing apparatus, image reading apparatus, and image forming apparatus |
EP3241041B1 (en) * | 2014-12-30 | 2020-06-17 | General Electric Company | X-ray detector assembly |
US11045088B2 (en) | 2015-02-27 | 2021-06-29 | Welch Allyn, Inc. | Through focus retinal image capturing |
US10799115B2 (en) | 2015-02-27 | 2020-10-13 | Welch Allyn, Inc. | Through focus retinal image capturing |
USD771631S1 (en) * | 2015-06-02 | 2016-11-15 | Hand Held Products, Inc. | Mobile computer housing |
US9454688B1 (en) * | 2015-07-20 | 2016-09-27 | Konica Minolta Laboratory U.S.A., Inc. | High capacity 2D color barcode design and decoding method |
US10136804B2 (en) | 2015-07-24 | 2018-11-27 | Welch Allyn, Inc. | Automatic fundus image capture system |
ES2937272T3 (en) * | 2015-09-15 | 2023-03-27 | Pepperl & Fuchs Se | Apparatus and method for providing a plot or plot sequence for detection by a detector |
US10772495B2 (en) | 2015-11-02 | 2020-09-15 | Welch Allyn, Inc. | Retinal image capturing |
US10499269B2 (en) | 2015-11-12 | 2019-12-03 | Commscope Technologies Llc | Systems and methods for assigning controlled nodes to channel interfaces of a controller |
JP2017099616A (en) * | 2015-12-01 | 2017-06-08 | ソニー株式会社 | Surgical control device, surgical control method and program, and surgical system |
US9946943B2 (en) * | 2015-12-07 | 2018-04-17 | Delta Id, Inc. | Methods and apparatuses for birefringence based biometric authentication |
US10413179B2 (en) | 2016-01-07 | 2019-09-17 | Welch Allyn, Inc. | Infrared fundus imaging system |
JP6908357B2 (en) * | 2016-06-21 | 2021-07-28 | 株式会社ミツトヨ | Position identification device and position identification method |
US10573234B2 (en) | 2016-09-14 | 2020-02-25 | Apple Inc. | Systems and methods for in-frame sensing and adaptive sensing control |
US10545242B2 (en) | 2016-09-14 | 2020-01-28 | Apple Inc. | Systems and methods for in-frame sensing and adaptive sensing control |
US10602926B2 (en) | 2016-09-29 | 2020-03-31 | Welch Allyn, Inc. | Through focus retinal image capturing |
CN107886026B (en) * | 2016-09-30 | 2019-12-13 | 腾讯科技(深圳)有限公司 | graphic code processing method and device |
US10285589B2 (en) | 2016-09-30 | 2019-05-14 | Welch Allyn, Inc. | Fundus image capture system |
WO2018150336A1 (en) * | 2017-02-14 | 2018-08-23 | Atracsys Sàrl | High-speed optical tracking with compression and/or cmos windowing |
US20190020411A1 (en) * | 2017-07-13 | 2019-01-17 | Qualcomm Incorporated | Methods and apparatus for efficient visible light communication (vlc) with reduced data rate |
US10079255B1 (en) * | 2017-08-04 | 2018-09-18 | GM Global Technology Operations LLC | Color filter array apparatus |
US10242240B1 (en) * | 2017-12-28 | 2019-03-26 | Symbol Technologies, Llc | Decoded imaging engine using standard un-decoded engine interface |
US11373088B2 (en) * | 2017-12-30 | 2022-06-28 | Intel Corporation | Machine learning accelerator mechanism |
US10567724B2 (en) | 2018-04-10 | 2020-02-18 | GM Global Technology Operations LLC | Dynamic demosaicing of camera pixels |
US11096574B2 (en) | 2018-05-24 | 2021-08-24 | Welch Allyn, Inc. | Retinal image capturing |
US10841488B2 (en) | 2018-08-03 | 2020-11-17 | Qualcomm Incorporated | Combined monochrome and chromatic camera sensor |
JP2020167495A (en) * | 2019-03-29 | 2020-10-08 | ソニーセミコンダクタソリューションズ株式会社 | Solid-state image sensor and image sensor |
KR102625261B1 (en) * | 2019-10-21 | 2024-01-12 | 삼성전자주식회사 | Image device |
US11176342B2 (en) | 2020-01-15 | 2021-11-16 | Datalogic Usa Inc. | Produce or item recognition by hybrid 3D camera and/or multi-spectrum illumination |
US12108172B2 (en) * | 2020-02-14 | 2024-10-01 | Sony Group Corporation | Vehicle control system using imaging device capable of object detection |
US11295104B2 (en) * | 2020-06-10 | 2022-04-05 | Zebra Technologies Corporation | Methods and apparatus to read barcodes on reflective surfaces |
US11995900B2 (en) * | 2021-11-12 | 2024-05-28 | Zebra Technologies Corporation | Method on identifying indicia orientation and decoding indicia for machine vision systems |
Citations (94)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3716699A (en) * | 1971-04-02 | 1973-02-13 | A Eckert | Method and apparatus for optical code reading |
US4253447A (en) * | 1978-10-16 | 1981-03-03 | Welch Allyn, Inc. | Color endoscope with charge coupled device and television viewing |
US4261344A (en) * | 1979-09-24 | 1981-04-14 | Welch Allyn, Inc. | Color endoscope |
USRE31289E (en) * | 1978-10-16 | 1983-06-28 | Welch Allyn, Inc. | Color endoscope with charge coupled device and television viewing |
US4390895A (en) * | 1980-10-07 | 1983-06-28 | Tokyo Shibaura Denki Kabushiki Kaisha | Color image pick-up apparatus |
USRE31290E (en) * | 1978-10-16 | 1983-06-28 | Welch Allyn, Inc. | Color endoscope |
US4437112A (en) * | 1980-02-15 | 1984-03-13 | Canon Kabushiki Kaisha | Solid-state color imaging apparatus |
US4491865A (en) * | 1982-09-29 | 1985-01-01 | Welch Allyn, Inc. | Image sensor assembly |
US4516017A (en) * | 1982-01-20 | 1985-05-07 | Nippondenso Co., Ltd. | High-sensitive optical reading apparatus and method of reading optical information |
US4523224A (en) * | 1982-12-27 | 1985-06-11 | Welch Allyn Inc. | Color filter wheel synchronizer |
US4806776A (en) * | 1980-03-10 | 1989-02-21 | Kley Victor B | Electrical illumination and detecting apparatus |
US4807981A (en) * | 1984-04-28 | 1989-02-28 | Kabushiki Kaisha Toshiba | Spatial frequency filter |
US4823186A (en) * | 1986-12-19 | 1989-04-18 | Fuji Photo Film Co., Ltd. | Color video signal generating device using monochrome and color image sensors having different resolutions to form a luminance signal |
US5018006A (en) * | 1985-10-31 | 1991-05-21 | Canon Kabushiki Kaisha | Multi-plate type image pickup apparatus having picture elements for producing color and luminance signals |
US5019699A (en) * | 1988-08-31 | 1991-05-28 | Norand Corporation | Hand-held optical character reader with means for instantaneously reading information from a predetermined area at an optical sensing area |
US5222477A (en) * | 1991-09-30 | 1993-06-29 | Welch Allyn, Inc. | Endoscope or borescope stereo viewing system |
US5278642A (en) * | 1992-02-26 | 1994-01-11 | Welch Allyn, Inc. | Color imaging system |
US5291008A (en) * | 1992-01-10 | 1994-03-01 | Welch Allyn, Inc. | Optical assembly and apparatus employing same using an aspherical lens and an aperture stop |
US5305122A (en) * | 1988-08-31 | 1994-04-19 | Canon Kabushiki Kaisha | Image reading and processing apparatus suitable for use as a color hand-held scanner |
US5308962A (en) * | 1991-11-01 | 1994-05-03 | Welch Allyn, Inc. | Reduced power scanner for reading indicia |
US5393965A (en) * | 1990-11-13 | 1995-02-28 | Symbol Technologies, Inc. | Flexible merchandise checkout and inventory management system |
US5399846A (en) * | 1990-01-05 | 1995-03-21 | Symbol Technologies, Inc. | Systems utilizing a high density two dimensional bar code symbology |
US5401944A (en) * | 1990-11-20 | 1995-03-28 | Symbol Technologies, Inc. | Traveler security and luggage control system |
US5406032A (en) * | 1992-02-13 | 1995-04-11 | The Boeing Company | Apparatus and method for supporting wire bundles within a structure |
US5410141A (en) * | 1989-06-07 | 1995-04-25 | Norand | Hand-held data capture system with interchangable modules |
US5506619A (en) * | 1995-03-17 | 1996-04-09 | Eastman Kodak Company | Adaptive color plan interpolation in single sensor color electronic camera |
US5513264A (en) * | 1994-04-05 | 1996-04-30 | Metanetics Corporation | Visually interactive encoding and decoding of dataforms |
US5521366A (en) * | 1994-07-26 | 1996-05-28 | Metanetics Corporation | Dataform readers having controlled and overlapped exposure integration periods |
US5591955A (en) * | 1993-05-11 | 1997-01-07 | Laser; Vadim | Portable data file readers |
US5602377A (en) * | 1995-03-01 | 1997-02-11 | Metanetics Corporation | Bar code dataform scanning and labeling apparatus and method |
US5602379A (en) * | 1993-05-31 | 1997-02-11 | Kabushiki Kaisha Tec | Optical code reader |
US5629734A (en) * | 1995-03-17 | 1997-05-13 | Eastman Kodak Company | Adaptive color plan interpolation in single sensor color electronic camera |
US5714745A (en) * | 1995-12-20 | 1998-02-03 | Metanetics Corporation | Portable data collection device with color imaging assembly |
US5717195A (en) * | 1996-03-05 | 1998-02-10 | Metanetics Corporation | Imaging based slot dataform reader |
US5739518A (en) * | 1995-05-17 | 1998-04-14 | Metanetics Corporation | Autodiscrimination for dataform decoding and standardized recording |
US5756981A (en) * | 1992-02-27 | 1998-05-26 | Symbol Technologies, Inc. | Optical scanner for reading and decoding one- and-two-dimensional symbologies at variable depths of field including memory efficient high speed image processing means and high accuracy image analysis means |
US5877487A (en) * | 1995-06-21 | 1999-03-02 | Asahi Kogaku Kogyo Kabushiki Kaisha | Data symbol reading device |
US6010073A (en) * | 1996-08-01 | 2000-01-04 | Datalogic S.P.A. | Portable apparatus for reading an optical code |
US6010070A (en) * | 1997-06-16 | 2000-01-04 | Nippon Electric Industry Co., Ltd. | Code reading device and method with variable light signal storage time |
US6019286A (en) * | 1995-06-26 | 2000-02-01 | Metanetics Corporation | Portable data collection device with dataform decoding and image capture capability |
US6045047A (en) * | 1995-01-17 | 2000-04-04 | Welch Allyn Data Collection, Inc. | Two-dimensional part reader having a focussing guide |
US6045238A (en) * | 1998-10-09 | 2000-04-04 | Welch Allyn Inc. | Illumination assembly for an optical viewing device |
US6049813A (en) * | 1993-02-26 | 2000-04-11 | Intermec Ip Corp. | Portable work station-type data collection system |
US6053408A (en) * | 1997-12-02 | 2000-04-25 | Telxon Corporation | Multi-focal length imaging based portable dataform reader |
US6053407A (en) * | 1995-05-31 | 2000-04-25 | Metanetics Corporation | Maxicode data extraction using spatial domain features |
US6060722A (en) * | 1995-05-15 | 2000-05-09 | Havens; William H. | Optical reader having illumination assembly including improved aiming pattern generator |
US6062475A (en) * | 1997-06-25 | 2000-05-16 | Metanetics Corporation | Portable data collection device including color imaging dataform reader assembly |
US6062477A (en) * | 1998-04-30 | 2000-05-16 | Ncr Corporation | Apparatus and method of operating a retail terminal having a single-orientation base assembly and a multiple-orientation base assembly |
US6179208B1 (en) * | 1997-01-31 | 2001-01-30 | Metanetics Corporation | Portable data collection device with variable focusing module for optic assembly |
US6186404B1 (en) * | 1998-05-29 | 2001-02-13 | Welch Allyn Data Collection, Inc. | Security document voiding system |
US6223988B1 (en) * | 1996-10-16 | 2001-05-01 | Omniplanar, Inc | Hand-held bar code reader with laser scanning and 2D image capture |
US6234394B1 (en) * | 1995-03-20 | 2001-05-22 | Symbol Technologies, Inc. | Triggered optical reader |
US6340114B1 (en) * | 1998-06-12 | 2002-01-22 | Symbol Technologies, Inc. | Imaging engine and method for code readers |
US6375075B1 (en) * | 1999-10-18 | 2002-04-23 | Intermec Ip Corp. | Method and apparatus for reading machine-readable symbols including color symbol elements |
US6386452B1 (en) * | 1998-11-26 | 2002-05-14 | Fuji Photo Film Co., Ltd. | Image reading device with improved controller |
US20030004827A1 (en) * | 1998-04-27 | 2003-01-02 | Wang Ynjiun P. | Payment system |
US20030018897A1 (en) * | 2001-07-20 | 2003-01-23 | Psc Scanning, Inc. | Video identification verification system and method for a self-checkout system |
US20030022147A1 (en) * | 1993-06-04 | 2003-01-30 | Paul E. Segall | Solutions for use as plasma expanders and substitutes |
US20030019934A1 (en) * | 1998-07-08 | 2003-01-30 | Hand Held Products, Inc. | Optical reader aiming assembly comprising aperture |
US6522441B1 (en) * | 2000-11-28 | 2003-02-18 | Psc Scanning, Inc. | Micro-optical system for an auto-focus scanner having an improved depth of field |
US20030034394A1 (en) * | 1999-10-04 | 2003-02-20 | Hand Held Products, Inc. | Optical reader comprising finely adjustable lens assembly |
US6533168B1 (en) * | 1999-05-27 | 2003-03-18 | Peter N. Ching | Method and apparatus for computer-readable purchase receipts using multi-dimensional bar codes |
US20030062413A1 (en) * | 1999-10-04 | 2003-04-03 | Hand Held Products, Inc. | Optical reader comprising multiple color illumination |
US6552323B2 (en) * | 2000-12-06 | 2003-04-22 | Eastman Kodak Company | Image sensor with a shared output signal line |
US6561428B2 (en) * | 1997-10-17 | 2003-05-13 | Hand Held Products, Inc. | Imaging device having indicia-controlled image parsing mode |
US20030089775A1 (en) * | 2001-05-21 | 2003-05-15 | Welch Allyn Data Collection, Inc. | Display-equipped optical reader having decode failure image display feedback mode |
US20040004128A1 (en) * | 1996-09-03 | 2004-01-08 | Hand Held Products, Inc. | Optical reader system comprising digital conversion circuit |
US6676016B1 (en) * | 2000-05-04 | 2004-01-13 | Ncr Corporation | Retail terminal configured as consumer gateway to electronic billing application |
US6681994B1 (en) * | 1988-08-31 | 2004-01-27 | Intermec Ip Corp. | Method and apparatus for optically reading information |
US20040020990A1 (en) * | 2002-06-04 | 2004-02-05 | Havens William H. | Optical reader having a plurality of imaging modules |
US6694064B1 (en) * | 1999-11-19 | 2004-02-17 | Positive Systems, Inc. | Digital aerial image mosaic method and apparatus |
US20040031851A1 (en) * | 2002-03-20 | 2004-02-19 | Bianculli Thomas D. | Image capture system and method using a common imaging array |
US6695209B1 (en) * | 1999-10-04 | 2004-02-24 | Psc Scanning, Inc. | Triggerless optical reader with signal enhancement features |
US20040035933A1 (en) * | 2002-06-11 | 2004-02-26 | Havens William H. | Long range optical reader |
US20040046881A1 (en) * | 2001-04-12 | 2004-03-11 | Nikon Corporation | Imaging device |
US6714239B2 (en) * | 1997-10-29 | 2004-03-30 | Eastman Kodak Company | Active pixel sensor with programmable color balance |
US6722569B2 (en) * | 2001-07-13 | 2004-04-20 | Welch Allyn Data Collection, Inc. | Optical reader having a color imager |
US20050001035A1 (en) * | 2003-05-12 | 2005-01-06 | Thomas Hawley | Picture taking optical reader |
US20050041296A1 (en) * | 2000-08-07 | 2005-02-24 | Taiwan Semiconductor Manufacturing Co., Ltd. | High transmittance overcoat for microlens arrays in semiconductor color imagers |
US6877664B1 (en) * | 2000-07-11 | 2005-04-12 | Datalogic S.P.A. | Device and optical element for the aiming and the visual indication of reading area of a coded information reader |
US6880759B2 (en) * | 2003-05-23 | 2005-04-19 | Symagery Microsystems Inc. | Optical reader station |
US20050103854A1 (en) * | 2003-11-13 | 2005-05-19 | Metrologic Instruments, Inc. | Hand-supportable digital imaging-based bar code symbol reader supporting narrow-area and wide-area modes of illumination and image capture |
US20060011724A1 (en) * | 2004-07-15 | 2006-01-19 | Eugene Joseph | Optical code reading system and method using a variable resolution imaging sensor |
US20060016335A1 (en) * | 2004-07-22 | 2006-01-26 | Kaz, Incorporated | Air cleaner |
US7009638B2 (en) * | 2001-05-04 | 2006-03-07 | Vexcel Imaging Gmbh | Self-calibrating, digital, large format camera with single or multiple detector arrays and single or multiple optical systems |
US20060071081A1 (en) * | 2004-10-05 | 2006-04-06 | Ynjiun Wang | System and method to automatically discriminate between a signature and a barcode |
US7044378B2 (en) * | 2002-12-18 | 2006-05-16 | Symbol Technologies, Inc. | System and method for imaging and decoding optical codes using at least two different imaging settings |
US20070002153A1 (en) * | 2005-06-29 | 2007-01-04 | Bart Dierickx | Hue preservation |
US7159783B2 (en) * | 2002-03-28 | 2007-01-09 | Hand Held Products, Inc. | Customizable optical reader |
US7187442B2 (en) * | 2004-04-30 | 2007-03-06 | Optimum Technologies, Inc. | Polarized optical probes |
US20070108284A1 (en) * | 2005-11-17 | 2007-05-17 | Hand Held Products, Inc. | Optical reading device with programmable parameter control |
US7219841B2 (en) * | 2004-11-05 | 2007-05-22 | Hand Held Products, Inc. | Device and system for verifying quality of bar codes |
US7327504B2 (en) * | 2002-12-06 | 2008-02-05 | Eastman Kodak Company | Method of detecting clipped image pixels |
US7343865B2 (en) * | 2006-01-17 | 2008-03-18 | Schuert Technologies Llc | Rackable twin sheet pallet |
Family Cites Families (630)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
SE103286C1 (en) | ||||
US3378633A (en) | 1965-06-24 | 1968-04-16 | Stanford Research Inst | Monochrome photography system for color television |
US6860428B1 (en) | 1998-09-11 | 2005-03-01 | Robotic Vision Systems Inc. | Optical symbologies imager |
US3684868A (en) * | 1970-10-29 | 1972-08-15 | Ncr Co | Color bar code tag reader with light-emitting diodes |
US3716669A (en) | 1971-05-14 | 1973-02-13 | Japan Eng Dev Co | Mapping rectifier for generating polarstereographic maps from satellite scan signals |
US3949363A (en) | 1974-06-28 | 1976-04-06 | Recognition Equipment, Incorporated | Bar-Code/MICR/OCR merge |
US3971065A (en) | 1975-03-05 | 1976-07-20 | Eastman Kodak Company | Color imaging array |
JPS6056030B2 (en) | 1975-10-24 | 1985-12-07 | 松下電器産業株式会社 | Color solid-state imaging device |
US4047203A (en) | 1976-05-12 | 1977-09-06 | Eastman Kodak Company | Color imaging array |
US4613895A (en) | 1977-03-24 | 1986-09-23 | Eastman Kodak Company | Color responsive imaging device employing wavelength dependent semiconductor optical absorption |
US4350418A (en) | 1978-09-14 | 1982-09-21 | Canon Kabushiki Kaisha | Camera provided with automatic focus adjusting device |
US4353641A (en) | 1979-01-23 | 1982-10-12 | Armando M. Merlo | Color photography print with exposure data and means for making |
US4402088A (en) | 1981-04-09 | 1983-08-30 | Recognition Equipment Incorporated | OCR And bar code reading using area array |
US4500914A (en) | 1981-08-01 | 1985-02-19 | Sharp Kabushiki Kaisha | Color imaging array and color imaging device |
US4570057A (en) | 1981-12-28 | 1986-02-11 | Norand Corporation | Instant portable bar code reader |
US5288985A (en) | 1981-12-28 | 1994-02-22 | Norand Corporation | Instant portable bar code reader |
JPS5940630A (en) | 1982-08-30 | 1984-03-06 | Minolta Camera Co Ltd | 35mm lens shutter camera capable of automatic focusing |
EP0119862B1 (en) | 1983-03-17 | 1990-12-27 | Nec Corporation | Solid-state color imaging apparatus |
US4546379A (en) | 1983-04-21 | 1985-10-08 | Welch Allyn, Inc. | Independent color adjustment for a video system |
US4630307A (en) | 1984-09-10 | 1986-12-16 | Eastman Kodak Company | Signal processing method and apparatus for sampled image signals |
WO1986001965A2 (en) | 1984-09-10 | 1986-03-27 | Eastman Kodak Company | Signal rpocessing method and apparatus for sampled color image signal |
WO1986001678A2 (en) | 1984-09-10 | 1986-03-27 | Eastman Kodak Company | Single-chip solid-state color image sensor and camera incorporating such a sensor |
US4605956A (en) | 1984-09-10 | 1986-08-12 | Eastman Kodak Company | Single-chip electronic color camera with color-dependent birefringent optical spatial frequency filter and red and blue signal interpolating circuit |
US4760411A (en) | 1985-08-06 | 1988-07-26 | Konishiroku Photo Industry Co., Ltd. | Electrostatic recording apparatus |
US4724394A (en) | 1985-10-11 | 1988-02-09 | Brunswick Corporation | Gas detection by ion mobility segregation |
US4793689A (en) | 1985-12-26 | 1988-12-27 | Canon Kabushiki Kaisha | Lens barrel with vibration wave motor |
US4724521A (en) | 1986-01-14 | 1988-02-09 | Veri-Fone, Inc. | Method for operating a local terminal to execute a downloaded application program |
IL78675A (en) | 1986-05-02 | 1993-02-21 | Scitex Corp Ltd | Color separation scanner |
US5892971A (en) | 1986-08-08 | 1999-04-06 | Norand Corporation | Portable data processing device having an indicia reader and a multi-tasking operating system capable of executing battery monitoring instructions while concurrently executing application programs |
US4877949A (en) | 1986-08-08 | 1989-10-31 | Norand Corporation | Hand-held instant bar code reader system with automated focus based on distance measurements |
US5576529A (en) | 1986-08-08 | 1996-11-19 | Norand Technology Corporation | Hand-held optically readable information set reader focus with operation over a range of distances |
US5227614A (en) | 1986-08-15 | 1993-07-13 | Norand Corporation | Core computer processor module, and peripheral shell module assembled to form a pocket size data capture unit |
US5949056A (en) | 1986-09-10 | 1999-09-07 | Norand Corporation | Method and apparatus for optically reading an information pattern |
US5089885A (en) | 1986-11-14 | 1992-02-18 | Video Jukebox Network, Inc. | Telephone access display system with remote monitoring |
JPS63185285A (en) | 1987-01-28 | 1988-07-30 | Canon Inc | Solid-state image pickup device |
US4858020A (en) | 1987-01-28 | 1989-08-15 | Canon Kabushiki Kaisha | Image sensing device |
US4877948A (en) | 1987-06-25 | 1989-10-31 | Krueger Loren L | Combination bar code and mark-sense reader |
US5369778A (en) | 1987-08-21 | 1994-11-29 | Wang Laboratories, Inc. | Data processor that customizes program behavior by using a resource retrieval capability |
US4774565A (en) | 1987-08-24 | 1988-09-27 | Polaroid Corporation | Method and apparatus for reconstructing missing color samples |
US4794239A (en) | 1987-10-13 | 1988-12-27 | Intermec Corporation | Multitrack bar code and associated decoding method |
US4854302A (en) * | 1987-11-12 | 1989-08-08 | Welch Allyn, Inc. | Video equipped endoscope with needle probe |
US5340786A (en) | 1988-02-12 | 1994-08-23 | Mitsui Petrochemical Industries, Ltd. | Olefin polymerization catalyst and process for the polymerization of olefins |
US4874936A (en) | 1988-04-08 | 1989-10-17 | United Parcel Service Of America, Inc. | Hexagonal, information encoding article, process and system |
US4862253A (en) * | 1988-07-20 | 1989-08-29 | Welch Allyn, Inc. | Apparatus for converting a video processor |
US5841121A (en) | 1988-08-31 | 1998-11-24 | Norand Technology Corporation | Hand-held optically readable character set reader having automatic focus control for operation over a range of distances |
US6102293A (en) | 1991-11-04 | 2000-08-15 | Symbol Technologies, Inc. | Optical scanning system minimizing average laser power and improving aiming |
US5710417A (en) | 1988-10-21 | 1998-01-20 | Symbol Technologies, Inc. | Bar code reader for reading both one dimensional and two dimensional symbologies with programmable resolution |
US4853774A (en) * | 1988-10-28 | 1989-08-01 | Welch Allyn, Inc. | Auxiliary light apparatus for borescope |
US4915483A (en) | 1989-01-04 | 1990-04-10 | Lockheed Missiles & Space Company, Inc. | Apochromatic lens triplets with aspherical lens elements |
US4958064A (en) | 1989-01-30 | 1990-09-18 | Image Recognition Equipment Corporation | Bar code locator for video scanner/reader system |
US4962419A (en) | 1989-02-13 | 1990-10-09 | Eastman Kodak Company | Detail processing method and apparatus providing uniform processing of horizontal and vertical detail components |
US5113251A (en) | 1989-02-23 | 1992-05-12 | Fuji Xerox Co. | Editing control system and area editing system for image processing equipment |
US5073954A (en) | 1989-02-28 | 1991-12-17 | Electrocom Automation, Inc. | Bar code location and recognition processing system |
CA1334218C (en) | 1989-03-01 | 1995-01-31 | Jerome Swartz | Hand-held laser scanning for reading two dimensional bar codes |
CA1329263C (en) | 1989-03-01 | 1994-05-03 | Mark Krichever | Bar code scanner |
JPH07105857B2 (en) | 1989-03-16 | 1995-11-13 | 富士ゼロックス株式会社 | Image reader self-diagnosis system |
US5124539A (en) | 1989-06-16 | 1992-06-23 | Symbol Technologies, Inc. | Scan pattern generators for bar code symbol readers |
US5200599A (en) | 1989-06-16 | 1993-04-06 | Symbol Technologies, Inc | Symbol readers with changeable scan direction |
US5059779A (en) | 1989-06-16 | 1991-10-22 | Symbol Technologies, Inc. | Scan pattern generators for bar code symbol readers |
US5040064A (en) | 1989-10-02 | 1991-08-13 | Eastman Kodak Company | Method of processing sampled signal valves produced by a color imaging device |
US4941456A (en) * | 1989-10-05 | 1990-07-17 | Welch Allyn, Inc. | Portable color imager borescope |
US4957346A (en) | 1989-10-06 | 1990-09-18 | Welch Allyn, Inc. | Illumination system for portable color imager borescope |
US4963756A (en) | 1989-10-13 | 1990-10-16 | Hewlett-Packard Company | Focused line identifier for a bar code reader |
US5410649A (en) | 1989-11-17 | 1995-04-25 | Texas Instruments Incorporated | Imaging computer system and network |
US5337361C1 (en) * | 1990-01-05 | 2001-05-15 | Symbol Technologies Inc | Record with encoded data |
US5059146A (en) | 1990-02-22 | 1991-10-22 | Welch Allyn, Inc. | Method of adjusting a light source for color temperature and chromaticity |
US5144190A (en) | 1990-02-22 | 1992-09-01 | Welch Allyn, Inc. | Light source having desired color temperature and chromaticity |
JP2878376B2 (en) | 1990-02-28 | 1999-04-05 | キヤノン株式会社 | Photoelectric conversion device |
JP2839107B2 (en) | 1990-02-28 | 1998-12-16 | キヤノン株式会社 | Photoelectric conversion device |
US5155343A (en) | 1990-03-28 | 1992-10-13 | Chandler Donald G | Omnidirectional bar code reader with method and apparatus for detecting and scanning a bar code symbol |
US5323233A (en) | 1990-07-31 | 1994-06-21 | Canon Kabushiki Kaisha | Image signal processing apparatus having a color filter with offset luminance filter elements |
US5135160A (en) | 1990-08-17 | 1992-08-04 | Opticon, Inc. | Portable bar code reader utilizing pulsed LED array |
JPH04111132A (en) | 1990-08-31 | 1992-04-13 | Fujitsu Ltd | Multi-pointer control method and multi-pointer control device |
US5627359A (en) | 1991-09-17 | 1997-05-06 | Metrologic Instruments, Inc. | Laser code symbol scanner employing optical filtering system having narrow band-pass characteristics and spatially separated optical filter elements with laser light collection optics arranged along laser light return path disposed therebetween |
US5340973A (en) | 1990-09-17 | 1994-08-23 | Metrologic Instruments, Inc. | Automatic laser scanning system and method of reading bar code symbols using same |
US6732929B2 (en) | 1990-09-10 | 2004-05-11 | Metrologic Instruments, Inc. | Led-based planar light illumination beam generation module employing a focal lens for reducing the image size of the light emmiting surface of the led prior to beam collimation and planarization |
US6076733A (en) | 1993-11-24 | 2000-06-20 | Metrologic Instruments, Inc. | Web-based system and method for enabling a viewer to access and display HTML-encoded documents located on the world wide web (WWW) by reading URL-encoded bar code symbols printed on a web-based information resource guide |
US5192856A (en) | 1990-11-19 | 1993-03-09 | An Con Genetics, Inc. | Auto focusing bar code reader |
CA2056842A1 (en) | 1991-02-07 | 1992-08-08 | Jay M. Eastman | System for scanning and reading symbols, such as bar codes, which is adapted to be configured in a hand held unit and can automatically read symbols on objects which are placed inthe detection zone of the unit |
JP3327400B2 (en) | 1991-02-27 | 2002-09-24 | ソニー株式会社 | Exposure control circuit |
US5196684A (en) | 1991-03-25 | 1993-03-23 | Opticon, Inc. | Method and apparatus for improving the throughput of a CCD bar code scanner/decoder |
US5296690A (en) | 1991-03-28 | 1994-03-22 | Omniplanar, Inc. | System for locating and determining the orientation of bar codes in a two-dimensional image |
US5378883A (en) | 1991-07-19 | 1995-01-03 | Omniplanar Inc. | Omnidirectional wide range hand held bar code reader |
US5262623A (en) | 1991-09-04 | 1993-11-16 | Omniplanar, Inc. | Method and apparatus for distinguishing a preferred bar code or the like |
CA2053460A1 (en) | 1991-09-13 | 1993-03-14 | Eugene Bernard Joseph | Analog waveform decoder |
US5452379A (en) | 1991-09-13 | 1995-09-19 | Meadowbrook Industries, Ltd. | Image capture and storage techniques in association with optical mark reading |
US5883375A (en) | 1991-09-17 | 1999-03-16 | Metrologic Instruments, Inc. | Bar code symbol scanner having fixed and hand-held modes |
US5811828A (en) | 1991-09-17 | 1998-09-22 | Norand Corporation | Portable reader system having an adjustable optical focusing means for reading optical information over a substantial range of distances |
US5832289A (en) | 1991-09-20 | 1998-11-03 | Shaw; Venson M. | System for estimating worst time duration required to execute procedure calls and looking ahead/preparing for the next stack operation of the forthcoming procedure calls |
EP0536481A2 (en) | 1991-10-09 | 1993-04-14 | Photographic Sciences Corporation | Bar code reading instrument and selctively orientable graphics display which facilitates the operation of the instrument |
JPH05120173A (en) | 1991-10-24 | 1993-05-18 | Hokkaido Nippon Denki Software Kk | Starting management system for terminal equipment |
EP0621970A1 (en) | 1992-01-17 | 1994-11-02 | Norand Corporation | Portable optical reader system |
US6385352B1 (en) | 1994-10-26 | 2002-05-07 | Symbol Technologies, Inc. | System and method for reading and comparing two-dimensional images |
US6347163B2 (en) | 1994-10-26 | 2002-02-12 | Symbol Technologies, Inc. | System for reading two-dimensional images using ambient and/or projected light |
US5354977A (en) | 1992-02-27 | 1994-10-11 | Alex Roustaei | Optical scanning head |
US5777314A (en) | 1992-02-27 | 1998-07-07 | Symbol | Optical scanner with fixed focus optics |
US5786582A (en) | 1992-02-27 | 1998-07-28 | Symbol Technologies, Inc. | Optical scanner for reading and decoding one- and two-dimensional symbologies at variable depths of field |
US5349172A (en) | 1992-02-27 | 1994-09-20 | Alex Roustaei | Optical scanning head |
JP2932815B2 (en) | 1992-02-28 | 1999-08-09 | 株式会社デンソー | Optical information reader |
US5821523A (en) | 1992-03-12 | 1998-10-13 | Bunte; Alan G. | Combined code reader and digital camera using a common photodetector |
WO1993019441A1 (en) | 1992-03-20 | 1993-09-30 | Commonwealth Scientific And Industrial Research Organisation | An object monitoring system |
DE4209536C3 (en) | 1992-03-24 | 2000-10-05 | Stuttgart Mikroelektronik | Image cell for an image sensor chip |
US5483624A (en) | 1992-03-27 | 1996-01-09 | Monarch Marking Systems, Inc. | Programmable hand held labeler |
US5751844A (en) | 1992-04-20 | 1998-05-12 | International Business Machines Corporation | Method and apparatus for image acquisition with adaptive compensation for image exposure variation |
US5323327A (en) | 1992-05-01 | 1994-06-21 | Storage Technology Corporation | On-the-fly cataloging of library cell contents in an automated robotic tape library |
JPH05316410A (en) | 1992-05-14 | 1993-11-26 | Sanyo Electric Co Ltd | Automatic exposure adjustment device |
US5374956A (en) | 1992-05-29 | 1994-12-20 | Eastman Kodak Company | Electronic imaging apparatus with dithered color filter array |
US6189793B1 (en) | 1992-06-12 | 2001-02-20 | Metrologic Instruments, Inc. | Automatic laser projection scanner with improved activation controlling mechanism |
JPH064229A (en) | 1992-06-18 | 1994-01-14 | Tokyo Electric Co Ltd | Information transmitting method by parallel interface |
JPH064191A (en) | 1992-06-23 | 1994-01-14 | Ricoh Co Ltd | Method and device for controlling external interface |
US5475207A (en) | 1992-07-14 | 1995-12-12 | Spectra-Physics Scanning Systems, Inc. | Multiple plane scanning system for data reading applications |
US5406062A (en) | 1992-07-31 | 1995-04-11 | Alps Electric Co., Ltd. | Sensitivity adjustment circuit for bar code scanner and method therefor |
US5343028A (en) | 1992-08-10 | 1994-08-30 | United Parcel Service Of America, Inc. | Method and apparatus for detecting and decoding bar code symbols using two-dimensional digital pixel images |
JP2833975B2 (en) | 1992-09-28 | 1998-12-09 | オリンパス光学工業株式会社 | Dot code |
JPH11328303A (en) | 1992-09-28 | 1999-11-30 | Olympus Optical Co Ltd | Information reproducing device |
KR100225112B1 (en) | 1992-09-28 | 1999-10-15 | 기시모토 마사도시 | DOT CODE AND INFORMATION RECORDING / REPRODUCING SYSTEM FOR RECORDING / REPRODUCING THE SAME |
US20020044689A1 (en) | 1992-10-02 | 2002-04-18 | Alex Roustaei | Apparatus and method for global and local feature extraction from digital images |
US5654533A (en) * | 1992-10-26 | 1997-08-05 | Kabushiki Kaisha Tec | Apparatus and method for reading two-dimensional symbols |
JP2790401B2 (en) | 1992-12-03 | 1998-08-27 | アルプス電気株式会社 | Barcode scanner sensitivity adjustment method |
US5298643A (en) * | 1992-12-22 | 1994-03-29 | Enzon, Inc. | Aryl imidate activated polyalkylene oxides |
JP4035173B2 (en) | 1993-01-18 | 2008-01-16 | キヤノン株式会社 | Control apparatus and control method |
JP3290738B2 (en) | 1993-03-15 | 2002-06-10 | オリンパス光学工業株式会社 | Barcode detection device |
US6655597B1 (en) | 2000-06-27 | 2003-12-02 | Symbol Technologies, Inc. | Portable instrument for electro-optically reading indicia and for projecting a bit-mapped color image |
JPH06301523A (en) | 1993-04-15 | 1994-10-28 | Fuji Facom Corp | Execution program setting device |
US6811088B2 (en) | 1993-05-28 | 2004-11-02 | Symbol Technologies, Inc. | Portable data collection system |
US6853293B2 (en) | 1993-05-28 | 2005-02-08 | Symbol Technologies, Inc. | Wearable communication system |
JP3316704B2 (en) | 1993-06-10 | 2002-08-19 | 株式会社ニコン | Projection exposure apparatus, scanning exposure method, and element manufacturing method |
JPH0721300A (en) | 1993-06-30 | 1995-01-24 | At & T Global Inf Solutions Internatl Inc | Bar-code reader |
JP2720924B2 (en) | 1993-09-21 | 1998-03-04 | 富士ゼロックス株式会社 | Image signal encoding device |
US5414538A (en) | 1993-10-07 | 1995-05-09 | Xerox Corporation | Image-dependent exposure enhancement |
JP3387173B2 (en) * | 1993-10-27 | 2003-03-17 | ソニー株式会社 | Electromagnetic drive |
JPH07121376A (en) | 1993-10-28 | 1995-05-12 | Fujitsu Ltd | Down load information preparing device |
US5347374A (en) | 1993-11-05 | 1994-09-13 | Xerox Corporation | Cascaded image processing using histogram prediction |
JP3644042B2 (en) | 1993-11-15 | 2005-04-27 | ソニー株式会社 | Multitask processing device |
JPH07182261A (en) | 1993-12-21 | 1995-07-21 | Oki Electric Ind Co Ltd | Information processor |
US6166768A (en) * | 1994-01-28 | 2000-12-26 | California Institute Of Technology | Active pixel sensor array with simple floating gate pixels |
US6486503B1 (en) * | 1994-01-28 | 2002-11-26 | California Institute Of Technology | Active pixel sensor array with electronic shuttering |
JPH07219863A (en) | 1994-02-03 | 1995-08-18 | Hitachi Ltd | Business distribution system |
US5932862A (en) | 1994-03-04 | 1999-08-03 | Welch Allyn, Inc. | Optical reader having improved scanning-decoding features |
US5463214A (en) | 1994-03-04 | 1995-10-31 | Welch Allyn, Inc. | Apparatus for optimizing throughput in decoded-output scanners and method of using same |
US5825006A (en) | 1994-03-04 | 1998-10-20 | Welch Allyn, Inc. | Optical reader having improved autodiscrimination features |
US5929418A (en) | 1994-03-04 | 1999-07-27 | Welch Allyn, Inc. | Optical reader having improved menuing features |
US20040206821A1 (en) | 1994-03-04 | 2004-10-21 | Andrew Longacre | Autodiscriminating bar code reading apparatus having solid state image sensor |
US5942741A (en) | 1994-03-04 | 1999-08-24 | Welch Allyn, Inc. | Apparatus for optimizing throughput in decoded-output scanners and method of using same |
US7387253B1 (en) | 1996-09-03 | 2008-06-17 | Hand Held Products, Inc. | Optical reader system comprising local host processor and optical reader |
US20040262394A1 (en) | 1994-03-04 | 2004-12-30 | Andrew Longacre | Optical reader having solid-state image sensor |
US5900613A (en) | 1994-03-04 | 1999-05-04 | Welch Allyn, Inc. | Optical reader having improved reprogramming features |
JP3507549B2 (en) | 1994-05-02 | 2004-03-15 | 富士写真フイルム株式会社 | Exposure determination method |
US6141496A (en) | 1994-06-06 | 2000-10-31 | The Erie Ceramic Arts Company | Electrically heated air fresheners |
US5917945A (en) | 1994-06-15 | 1999-06-29 | Metanetics Corporation | Recognizing dataforms in image areas |
CA2149476A1 (en) | 1994-06-21 | 1995-12-22 | James Michael Magee | Capability engine method and apparatus for a microkernel data processing system |
US5672858A (en) | 1994-06-30 | 1997-09-30 | Symbol Technologies Inc. | Apparatus and method for reading indicia using charge coupled device and scanning laser beam technology |
US5932872A (en) | 1994-07-01 | 1999-08-03 | Jeffrey H. Price | Autofocus system for scanning microscopy having a volume image formation |
US5940163A (en) * | 1994-07-19 | 1999-08-17 | Electro Plasma Inc. | Photon coupled color flat panel display and method of manufacture |
US5811784A (en) | 1995-06-26 | 1998-09-22 | Telxon Corporation | Extended working range dataform reader |
US5572006A (en) | 1994-07-26 | 1996-11-05 | Metanetics Corporation | Automatic exposure single frame imaging systems |
US5763864A (en) * | 1994-07-26 | 1998-06-09 | Meta Holding Corporation | Dataform reader including dual laser and imaging reading assemblies |
US5702059A (en) | 1994-07-26 | 1997-12-30 | Meta Holding Corp. | Extended working range dataform reader including fuzzy logic image control circuitry |
US5815200A (en) | 1994-07-26 | 1998-09-29 | Metanetics Corporation | Extended working range dataform reader with reduced power consumption |
US5662586A (en) | 1994-08-18 | 1997-09-02 | Welch Allyn, Inc. | Hand held diagnostic instrument with video imaging |
US5527262A (en) * | 1994-08-18 | 1996-06-18 | Welch Allyn, Inc. | Hand-held diagnostic dental probe with video imaging |
DE69514874T2 (en) | 1994-08-23 | 2000-07-20 | Hewlett-Packard Co., Palo Alto | Sensor arrangement for an optical image scanner which supplies gray value and color signals |
US5781708A (en) | 1994-09-13 | 1998-07-14 | Intermec Technology, Inc. | Integral bar code printer and reader system and method of operation |
JPH08106393A (en) | 1994-10-06 | 1996-04-23 | Matsushita Electric Ind Co Ltd | Portable terminal system |
EP0788635B1 (en) | 1994-10-25 | 1998-04-29 | United Parcel Service Of America, Inc. | Method and apparatus for a portable non-contact label imager |
PT788634E (en) | 1994-10-25 | 2000-08-31 | United Parcel Service Inc | AUTOMATIC ELECTRONIC CAMERA FOR LABEL IMAGE COLLECTION |
JPH08147398A (en) | 1994-11-24 | 1996-06-07 | Nippondenso Co Ltd | Portable optical information reader |
US5702058A (en) | 1994-12-01 | 1997-12-30 | Calmar Inc. | Dual foamer nozzle assembly for trigger sprayer |
US5581071A (en) | 1994-12-06 | 1996-12-03 | International Business Machines Corporation | Barcode scanner with adjustable light source intensity |
US5814803A (en) * | 1994-12-23 | 1998-09-29 | Spectra-Physics Scanning Systems, Inc. | Image reader with multi-focus lens |
US5770847A (en) | 1994-12-23 | 1998-06-23 | Spectra-Physics Scanning Systems, Inc. | Bar code reader with multi-focus lens |
US5703348A (en) | 1994-12-26 | 1997-12-30 | Kabushiki Kaisha Tec | Hand-held optical code reader |
EP0722148A2 (en) | 1995-01-10 | 1996-07-17 | Welch Allyn, Inc. | Bar code reader |
US5786586A (en) * | 1995-01-17 | 1998-07-28 | Welch Allyn, Inc. | Hand-held optical reader having a detachable lens-guide assembly |
US5978610A (en) | 1995-03-14 | 1999-11-02 | Asahi Kogaku Kogyo Kabushiki Kaisha | Exposure control apparatus for electronic development type camera |
US6311896B1 (en) | 1995-03-20 | 2001-11-06 | Symbol Technologies, Inc. | Compact bar code scanner |
JPH0991368A (en) | 1995-07-20 | 1997-04-04 | Fujitsu Ltd | Optical reader |
US5780834A (en) | 1995-05-15 | 1998-07-14 | Welch Allyn, Inc. | Imaging and illumination optics assembly |
US5784102A (en) * | 1995-05-15 | 1998-07-21 | Welch Allyn, Inc. | Optical reader having improved interactive image sensing and control circuitry |
JPH08315059A (en) * | 1995-05-16 | 1996-11-29 | Asahi Optical Co Ltd | Data symbol reader |
JPH08329180A (en) | 1995-06-05 | 1996-12-13 | Asahi Optical Co Ltd | Data symbol reader |
US5680542A (en) | 1995-06-07 | 1997-10-21 | Motorola, Inc. | Method and apparatus for synchronizing data in a host memory with data in target MCU memory |
US6069696A (en) | 1995-06-08 | 2000-05-30 | Psc Scanning, Inc. | Object recognition system and method |
JPH096891A (en) | 1995-06-21 | 1997-01-10 | Asahi Optical Co Ltd | Data symbol reader and data symbol read system |
JP3391940B2 (en) | 1995-06-26 | 2003-03-31 | キヤノン株式会社 | Illumination device and exposure device |
US5783811A (en) * | 1995-06-26 | 1998-07-21 | Metanetics Corporation | Portable data collection device with LED targeting and illumination assembly |
US5703349A (en) | 1995-06-26 | 1997-12-30 | Metanetics Corporation | Portable data collection device with two dimensional imaging assembly |
US5818028A (en) | 1995-06-26 | 1998-10-06 | Telxon Corporation | Portable data collection device with two dimensional imaging assembly |
JP3803240B2 (en) | 1995-07-20 | 2006-08-02 | 富士通株式会社 | Optical reader |
DE69624714T2 (en) | 1995-08-11 | 2003-08-07 | Toshiba Kawasaki Kk | IMAGE RECORDING SYSTEM, INTEGRATED SOLID IMAGE RECORDING SEMICONDUCTOR CIRCUIT |
US5640002A (en) | 1995-08-15 | 1997-06-17 | Ruppert; Jonathan Paul | Portable RF ID tag and barcode reader |
GB2308267B (en) | 1995-08-25 | 2000-06-28 | Psc Inc | Optical reader with imaging array having reduced pattern density |
JPH0983841A (en) | 1995-09-06 | 1997-03-28 | Nisca Corp | Structure for universal head for television camera and panning tilt camera |
US5691773A (en) | 1995-09-12 | 1997-11-25 | Metanetics Corporation | Anti-hand-jittering dataform readers and methods |
US5796089A (en) | 1995-09-21 | 1998-08-18 | Symbol Technologies, Inc. | Bar code scanner with simplified auto-focus capability |
US5979763A (en) | 1995-10-13 | 1999-11-09 | Metanetics Corporation | Sub-pixel dataform reader with dynamic noise margins |
US6463173B1 (en) | 1995-10-30 | 2002-10-08 | Hewlett-Packard Company | System and method for histogram-based image contrast enhancement |
US6636635B2 (en) | 1995-11-01 | 2003-10-21 | Canon Kabushiki Kaisha | Object extraction method, and image sensing apparatus using the method |
JPH09134403A (en) | 1995-11-07 | 1997-05-20 | Sumitomo Electric Ind Ltd | Optical information reader |
US6714969B1 (en) | 1995-11-17 | 2004-03-30 | Symbol Technologies, Inc. | Mobile terminal with integrated host application software |
JP3441580B2 (en) | 1995-12-14 | 2003-09-02 | 富士通株式会社 | Reader |
US6254003B1 (en) * | 1995-12-18 | 2001-07-03 | Welch Allyn Data Collection, Inc. | Optical reader exposure control apparatus comprising illumination level detection circuitry |
US5831254A (en) | 1995-12-18 | 1998-11-03 | Welch Allyn, Inc. | Exposure control apparatus for use with optical readers |
DE69733248T2 (en) | 1996-01-22 | 2006-01-19 | California Institute Of Technology, Pasadena | ACTIVE PICTURE ELEMENT SENSOR MATRIX WITH ELECTRONIC TRANSMISSION |
US6575368B1 (en) | 1996-01-31 | 2003-06-10 | Psc Scanning, Inc. | Multiple aperture data reader for multi-mode operation |
US6473519B1 (en) | 1996-02-21 | 2002-10-29 | Hand Held Products, Inc. | Check reader |
US5652621A (en) | 1996-02-23 | 1997-07-29 | Eastman Kodak Company | Adaptive color plane interpolation in single sensor color electronic camera |
US5818023A (en) | 1996-03-05 | 1998-10-06 | Metanetics Corporation | Portable ID card verification apparatus |
US5773810A (en) * | 1996-03-29 | 1998-06-30 | Welch Allyn, Inc. | Method for generating real time degree of focus signal for handheld imaging device |
US5793033A (en) * | 1996-03-29 | 1998-08-11 | Metanetics Corporation | Portable data collection device with viewing assembly |
US5896297A (en) | 1996-04-15 | 1999-04-20 | Valerino, Sr.; Fred M. | Robotube delivery system |
US5986297A (en) | 1996-05-22 | 1999-11-16 | Eastman Kodak Company | Color active pixel sensor with electronic shuttering, anti-blooming and low cross-talk |
US5988506A (en) | 1996-07-16 | 1999-11-23 | Galore Scantec Ltd. | System and method for reading and decoding two dimensional codes of high density |
JPH1063515A (en) | 1996-08-26 | 1998-03-06 | Nippon Telegr & Teleph Corp <Ntt> | Scheduling method for multitask control |
US5825002A (en) | 1996-09-05 | 1998-10-20 | Symbol Technologies, Inc. | Device and method for secure data updates in a self-checkout system |
US6084528A (en) | 1996-09-05 | 2000-07-04 | Symbol Technologies, Inc. | Intranet scanning terminal system |
US5979757A (en) | 1996-09-05 | 1999-11-09 | Symbol Technologies, Inc. | Method and system for presenting item information using a portable data terminal |
JPH10106919A (en) | 1996-09-26 | 1998-04-24 | Canon Inc | Semiconductor aligner and its operation control method |
US5901243A (en) | 1996-09-30 | 1999-05-04 | Hewlett-Packard Company | Dynamic exposure control in single-scan digital input devices |
US6460766B1 (en) | 1996-10-28 | 2002-10-08 | Francis Olschafskie | Graphic symbols and method and system for identification of same |
US5818975A (en) | 1996-10-28 | 1998-10-06 | Eastman Kodak Company | Method and apparatus for area selective exposure adjustment |
US6015088A (en) | 1996-11-05 | 2000-01-18 | Welch Allyn, Inc. | Decoding of real time video imaging |
CN1287188C (en) | 1996-11-08 | 2006-11-29 | Ncs皮尔逊股份有限公司 | Optical scanning with calibrated pixel output |
US5917913A (en) * | 1996-12-04 | 1999-06-29 | Wang; Ynjiun Paul | Portable electronic authorization devices and methods therefor |
US6473126B1 (en) | 1996-12-09 | 2002-10-29 | Canon Kabushiki Kaisha | Focusing information detecting device, focus detecting device and camera utilizing the same |
US5936224A (en) | 1996-12-11 | 1999-08-10 | Intermec Ip Corporation | Method and apparatus for reading machine-readable symbols by employing a combination of multiple operators and/or processors |
JP3129215B2 (en) | 1996-12-20 | 2001-01-29 | 株式会社デンソー | 2D code reader |
US6094221A (en) | 1997-01-02 | 2000-07-25 | Andersion; Eric C. | System and method for using a scripting language to set digital camera device features |
JP3758266B2 (en) | 1997-01-08 | 2006-03-22 | 株式会社デンソー | Code reader |
US5932741A (en) | 1997-01-31 | 1999-08-03 | Hoechst Marion Roussel | Imidazole-5-carboxylic acid compounds |
JPH10223875A (en) | 1997-02-06 | 1998-08-21 | Toshiba Corp | Solid-state image sensing device, driving method thereof and electronic still camera |
EP0858212B1 (en) | 1997-02-10 | 2002-05-29 | Fill Factory | Method for obtaining a high dynamic range read-out signal of a CMOS-based pixel structure and such CMOS-based pixel structure |
EP0883187A1 (en) | 1997-06-04 | 1998-12-09 | Interuniversitair Micro-Elektronica Centrum Vzw | A detector for electromagnetic radiation, pixel structure with high sensitivity using such detector and method of manufacturing such detector |
US5992744A (en) | 1997-02-18 | 1999-11-30 | Welch Allyn, Inc. | Optical reader having multiple scanning assemblies with simultaneously decoded outputs |
US6097839A (en) | 1997-03-10 | 2000-08-01 | Intermec Ip Corporation | Method and apparatus for automatic discriminating and locating patterns such as finder patterns, or portions thereof, in machine-readable symbols |
US7131591B1 (en) * | 1997-03-28 | 2006-11-07 | Symbol Technologies, Inc. | Link cost minimization for portable wireless terminal connectable to the internet |
US6142934A (en) | 1997-04-04 | 2000-11-07 | Welch Allyn, Inc. | Objective lens system for imaging instrument |
JPH10283204A (en) | 1997-04-04 | 1998-10-23 | Konica Corp | Multi-task processing method, multi-task processor and recording medium recording task |
US5909559A (en) | 1997-04-04 | 1999-06-01 | Texas Instruments Incorporated | Bus bridge device including data bus of first width for a first processor, memory controller, arbiter circuit and second processor having a different second data width |
JPH10283207A (en) | 1997-04-10 | 1998-10-23 | Toshiba Corp | Computer system, execution control method for multi-program and storage medium |
US6223986B1 (en) | 1997-04-17 | 2001-05-01 | Psc Scanning, Inc. | Aiming aid for optical data reading |
US6435411B1 (en) | 1997-04-21 | 2002-08-20 | Intermec Ip Corp. | Optoelectronic device for acquisition of images, in particular of bar codes |
US5880451A (en) | 1997-04-24 | 1999-03-09 | United Parcel Service Of America, Inc. | System and method for OCR assisted bar code decoding |
WO1998050814A1 (en) | 1997-05-05 | 1998-11-12 | Roustaei Alexander R | Optical scanner and image reader for reading images and decoding optical information including one and two dimensional symbologies at variable depth of field |
US6057554A (en) | 1997-05-12 | 2000-05-02 | Plesko; George A. | Reflective switch |
US5920061A (en) | 1997-05-29 | 1999-07-06 | Metanetics Corporation | Portable data collection device including imaging assembly with modular high density dataform reader assembly |
US6088058A (en) | 1997-05-29 | 2000-07-11 | Foveon, Inc. | Capacitive coupled bipolar active pixel imager having overflow protection and electronic shutter |
JPH1127485A (en) | 1997-07-03 | 1999-01-29 | Ricoh Co Ltd | Picture reader |
US6459495B1 (en) | 1997-07-15 | 2002-10-01 | Silverbrook Research Pty Ltd | Dot center tracking in optical storage systems using ink dots |
US6195150B1 (en) | 1997-07-15 | 2001-02-27 | Silverbrook Research Pty Ltd | Pseudo-3D stereoscopic images and output device |
JPH1141494A (en) | 1997-07-18 | 1999-02-12 | Fuji Electric Co Ltd | Camera identification device and image monitoring system using the device |
US6097835A (en) | 1997-07-23 | 2000-08-01 | Lockheed Martin Corporation | Projective pan sharpening methods and apparatus |
JPH1165859A (en) | 1997-08-11 | 1999-03-09 | Tec Corp | Method and device for managing operation program and storage medium where program for operation program management is recorded |
JP3036480B2 (en) | 1997-08-25 | 2000-04-24 | 日本電気株式会社 | Bar code reader and focus control method thereof |
US6147767A (en) | 1997-09-05 | 2000-11-14 | Comtec Informations Systems, Inc. | Computer system for a printer which stores and operates multiple application programs |
US6976626B2 (en) | 1997-09-16 | 2005-12-20 | Metrologic Instruments, Inc. | Wireless bar code symbol driven portable data terminal (PDT) system adapted for single handed operation |
US7028899B2 (en) | 1999-06-07 | 2006-04-18 | Metrologic Instruments, Inc. | Method of speckle-noise pattern reduction and apparatus therefore based on reducing the temporal-coherence of the planar laser illumination beam before it illuminates the target object by applying temporal phase modulation techniques during the transmission of the plib towards the target |
US6128414A (en) | 1997-09-29 | 2000-10-03 | Intermec Ip Corporation | Non-linear image processing and automatic discriminating method and apparatus for images such as images of machine-readable symbols |
DE19744230B4 (en) | 1997-10-07 | 2007-10-25 | Robert Bosch Gmbh | Control units for a system and method for operating a control unit |
US6298175B1 (en) | 1997-10-17 | 2001-10-02 | Welch Allyn Data Collection, Inc. | Object sensor system comprising controlled light source |
US6298176B2 (en) | 1997-10-17 | 2001-10-02 | Welch Allyn Data Collection, Inc. | Symbol-controlled image data reading system |
US5949052A (en) | 1997-10-17 | 1999-09-07 | Welch Allyn, Inc. | Object sensor system for stationary position optical reader |
US20070267501A1 (en) | 2006-05-18 | 2007-11-22 | Hand Held Products, Inc. | Multipurpose optical reader |
US5992753A (en) | 1997-10-30 | 1999-11-30 | Metanetics Corporation | Hand held dataform reader utilizing binarization process for dataform and signature area capture |
US6105869A (en) * | 1997-10-31 | 2000-08-22 | Microscan Systems, Incorporated | Symbol reading device including optics for uniformly illuminating symbology |
US5914476A (en) * | 1997-11-04 | 1999-06-22 | Welch Allyn, Inc. | Optical reader configured to accurately and rapidly read multiple symbols |
JP3873157B2 (en) | 1997-11-13 | 2007-01-24 | カシオ計算機株式会社 | Electronic camera device and imaging method |
ATE225968T1 (en) | 1997-11-17 | 2002-10-15 | Datalogic Spa | PORTABLE OPTICAL CODE READER HAVING A DEVICE FOR MONITORING THE CHARGE STATUS OF THE READER BATTERY |
US6603508B1 (en) | 1997-11-19 | 2003-08-05 | Ricoh Company, Ltd. | Method, system, and computer readable medium for controlling a digital camera to avoid blurring of a photograph |
US6318635B1 (en) | 1997-12-02 | 2001-11-20 | Telxon Corporation | Multi-focal length imaging based portable dataform reader |
JP2001526430A (en) | 1997-12-08 | 2001-12-18 | エール.ルースタエ アレキサンダー | Single chip symbol reader with smart sensor |
US20020050518A1 (en) | 1997-12-08 | 2002-05-02 | Roustaei Alexander R. | Sensor array |
US6669093B1 (en) | 1997-12-19 | 2003-12-30 | Telxon Corporation | Hand-held dataform reader having multiple target area illumination sources for independent reading of superimposed dataforms |
JPH11194929A (en) | 1997-12-26 | 1999-07-21 | Nec Corp | Client-side program management method in client-server system |
US5969326A (en) | 1998-01-14 | 1999-10-19 | Intermec Ip Corp. | Method and apparatus of autodiscriminating in symbol reader employing prioritized and updated table of symbologies |
US6102295A (en) * | 1998-01-14 | 2000-08-15 | Intermec Ip Corp. | Method and apparatus for decoding symbols by declaring erasures of element characteristics |
US6721008B2 (en) | 1998-01-22 | 2004-04-13 | Eastman Kodak Company | Integrated CMOS active pixel digital camera |
US6123263A (en) | 1998-01-29 | 2000-09-26 | Meta Holdings Corporation | Hand held dataform reader having strobing ultraviolet light illumination assembly for reading fluorescent dataforms |
JPH11230777A (en) | 1998-02-18 | 1999-08-27 | Tdk Corp | Portable gps navigation system |
US20020079370A1 (en) | 1998-03-11 | 2002-06-27 | Wood Robert J. | Illumination source for optical reading apparatus |
US6809766B1 (en) | 1998-03-11 | 2004-10-26 | Micro Technology, Inc. | Look ahead rolling shutter system in CMOS sensors |
US6003008A (en) | 1998-03-20 | 1999-12-14 | Skyteller L.L.C. | Point of sale device |
US6223190B1 (en) | 1998-04-13 | 2001-04-24 | Flashpoint Technology, Inc. | Method and system for producing an internet page description file on a digital imaging device |
US5965875A (en) | 1998-04-24 | 1999-10-12 | Foveon, Inc. | Color separation in an active pixel cell imaging array using a triple-well structure |
JPH11312211A (en) | 1998-04-30 | 1999-11-09 | Toshiba Tec Corp | Symbol reader |
JPH11312212A (en) | 1998-04-30 | 1999-11-09 | Toshiba Tec Corp | Symbol reader |
US6119941A (en) | 1998-05-04 | 2000-09-19 | Intermec Ip Corp. | Automated help instructions for automatically or adaptively configuring a hand-held device, such as a bar code reader or hand-held personal computer |
US6750906B1 (en) | 1998-05-08 | 2004-06-15 | Cirrus Logic, Inc. | Histogram-based automatic gain control method and system for video applications |
JP3109592B2 (en) | 1998-06-03 | 2000-11-20 | 日本電気株式会社 | Barcode reader and barcode reading method |
US6250551B1 (en) | 1998-06-12 | 2001-06-26 | Symbol Technologies, Inc. | Autodiscrimination and line drawing techniques for code readers |
US6315204B1 (en) | 1998-06-17 | 2001-11-13 | Motorola, Inc. | Optical code reader and methods and articles therefor |
IT245067Y1 (en) | 1998-06-19 | 2002-03-19 | 4P Srl | MULTI-FUNCTION HANDHELD ELECTRONIC PROCESSOR STRUCTURE. |
JP2000006475A (en) | 1998-06-25 | 2000-01-11 | Canon Inc | Image processor, control method therefor and memory medium |
US6275388B1 (en) | 1998-07-08 | 2001-08-14 | Welch Allyn Data Collection, Inc. | Image sensor mounting system |
US6659350B2 (en) | 2000-11-01 | 2003-12-09 | Hand Held Products | Adjustable illumination system for a barcode scanner |
US20040000592A1 (en) | 2002-02-20 | 2004-01-01 | Welch Allyn, Inc. | Adjustable illumination system for a barcode scanner |
US6164544A (en) | 1998-07-08 | 2000-12-26 | Welch Allyn Data Collection, Inc. | Adjustable illumination system for a barcode scanner |
US6601768B2 (en) | 2001-03-08 | 2003-08-05 | Welch Allyn Data Collection, Inc. | Imaging module for optical reader comprising refractive diffuser |
US20030209603A1 (en) | 1998-07-08 | 2003-11-13 | Welch Allyn Data Collection, Inc. | Optical assembly for barcode scanner |
US6547139B1 (en) | 1998-07-10 | 2003-04-15 | Welch Allyn Data Collection, Inc. | Method and apparatus for extending operating range of bar code scanner |
US6097856A (en) | 1998-07-10 | 2000-08-01 | Welch Allyn, Inc. | Apparatus and method for reducing imaging errors in imaging systems having an extended depth of field |
US6176429B1 (en) | 1998-07-17 | 2001-01-23 | Psc Scanning, Inc. | Optical reader with selectable processing characteristics for reading data in multiple formats |
US6106869A (en) | 1998-07-23 | 2000-08-22 | T-Rex Products, Inc. | Method of using a terrarium substrate |
US6191406B1 (en) | 1998-07-24 | 2001-02-20 | Eastman Kodak Company | Data reader and reader system having visible centerless targeting |
US6075240A (en) * | 1998-07-30 | 2000-06-13 | Nec Usa, Inc. | Hand-held plastic optical fiber linear scanner for reading color images formed on a surface |
US6184534B1 (en) | 1998-08-04 | 2001-02-06 | Eastman Kodak Company | Method of pulsing light emitting diodes for reading fluorescent indicia, data reader, and system |
US7098871B1 (en) | 1998-08-05 | 2006-08-29 | Microvision, Inc. | Optical scanning system with correction |
US6062455A (en) | 1998-08-06 | 2000-05-16 | Anthony C. Giannuzzi | Cartridge strip magazine for powder-actuated fastener setting tool |
JP2000056625A (en) | 1998-08-07 | 2000-02-25 | Ricoh Co Ltd | Image forming system |
US6879340B1 (en) | 1998-08-19 | 2005-04-12 | Micron Technology Inc. | CMOS imager with integrated non-volatile memory |
JP4328399B2 (en) | 1998-08-31 | 2009-09-09 | キヤノン株式会社 | Image processing apparatus, image processing method, and storage medium |
US6661521B1 (en) | 1998-09-11 | 2003-12-09 | Robotic Vision Systems, Inc. | Diffuse surface illumination apparatus and methods |
US6598797B2 (en) | 1998-09-11 | 2003-07-29 | Jason J. Lee | Focus and illumination analysis algorithm for imaging device |
US6283374B1 (en) | 1998-09-11 | 2001-09-04 | Robotic Vision Systems, Inc. | Symbology imaging and reading apparatus and method |
US6098887A (en) | 1998-09-11 | 2000-08-08 | Robotic Vision Systems, Inc. | Optical focusing device and method |
US6665012B1 (en) | 1998-09-22 | 2003-12-16 | Pixim, Inc. | Process-scalable high spatial resolution and low bit resolution CMOS area image sensor |
US6336587B1 (en) | 1998-10-19 | 2002-01-08 | Symbol Technologies, Inc. | Optical code reader for producing video displays and measuring physical parameters of objects |
JP3080932B2 (en) | 1998-10-22 | 2000-08-28 | 静岡日本電気株式会社 | Color image reader |
US6493114B1 (en) | 1998-10-22 | 2002-12-10 | Syscan Technology (Shenzhen) Co., Ltd. | Adaptive timing control of light integration process in one-dimensional CMOS image sensors |
US6575367B1 (en) | 1998-11-05 | 2003-06-10 | Welch Allyn Data Collection, Inc. | Image data binarization methods enabling optical reader to read fine print indicia |
US6264105B1 (en) | 1998-11-05 | 2001-07-24 | Welch Allyn Data Collection, Inc. | Bar code reader configured to read fine print barcode symbols |
JP3994595B2 (en) | 1998-11-16 | 2007-10-24 | 株式会社デンソー | Bar code reading method and recording medium |
US6109526A (en) | 1998-11-17 | 2000-08-29 | Intermec Ip Corp. | Optical and passive electromagnetic reader for reading machine-readable symbols, such as bar codes, and reading wireless tags, such as radio frequency tags, and corresponding method |
JP3592107B2 (en) | 1998-11-27 | 2004-11-24 | キヤノン株式会社 | Solid-state imaging device and camera |
US6850278B1 (en) | 1998-11-27 | 2005-02-01 | Canon Kabushiki Kaisha | Solid-state image pickup apparatus |
JP3592106B2 (en) | 1998-11-27 | 2004-11-24 | キヤノン株式会社 | Solid-state imaging device and camera |
US6157027A (en) | 1998-12-01 | 2000-12-05 | Nec Usa, Inc. | Modular optical fiber color image scanner with all-optical scanner head having side-coupled light guide for providing illumination light to the scanner head |
US6857572B2 (en) | 1998-12-03 | 2005-02-22 | Metrologic Instruments, Inc. | Automatically-activated hand-supportable laser scanning bar code symbol reading system with omnidirectional and unidirectional scanning modes in addition to a data transmission activation switch |
US6082619A (en) | 1998-12-16 | 2000-07-04 | Matsushita Electric Industrial Co., Ltd. | Method for locating and reading a two-dimensional barcode |
US6565003B1 (en) | 1998-12-16 | 2003-05-20 | Matsushita Electric Industrial Co., Ltd. | Method for locating and reading a two-dimensional barcode |
JP4096434B2 (en) | 1998-12-22 | 2008-06-04 | ぺんてる株式会社 | Barcode image detector |
US6611289B1 (en) | 1999-01-15 | 2003-08-26 | Yanbin Yu | Digital cameras using multiple sensors with multiple lenses |
JP2000207228A (en) | 1999-01-18 | 2000-07-28 | Star Micronics Co Ltd | Motor controller |
US6950139B2 (en) | 1999-01-22 | 2005-09-27 | Nikon Corporation | Image reading device and storage medium storing control procedure for image reading device |
US6247645B1 (en) | 1999-01-25 | 2001-06-19 | International Business Machines Corporation | Optical reader with combined housing and light pipe |
US6836288B1 (en) | 1999-02-09 | 2004-12-28 | Linvatec Corporation | Automatic exposure control system and method |
US6806847B2 (en) | 1999-02-12 | 2004-10-19 | Fisher-Rosemount Systems Inc. | Portable computer in a process control environment |
JP2000236326A (en) | 1999-02-15 | 2000-08-29 | Nippon Telegr & Teleph Corp <Ntt> | Light terminal control system and method therefor |
JP2002537932A (en) | 1999-03-10 | 2002-11-12 | デベックス (プロプライエタリイ) リミテッド | Imaging equipment |
JP3609977B2 (en) | 1999-07-15 | 2005-01-12 | シャープ株式会社 | Level shift circuit and image display device |
JP2000293622A (en) | 1999-04-09 | 2000-10-20 | Canon Inc | Method and device for image processing and storage medium |
JP3662769B2 (en) | 1999-05-12 | 2005-06-22 | 株式会社東研 | Code reading apparatus and method for color image |
AUPQ439299A0 (en) | 1999-12-01 | 1999-12-23 | Silverbrook Research Pty Ltd | Interface system |
AUPQ056099A0 (en) | 1999-05-25 | 1999-06-17 | Silverbrook Research Pty Ltd | A method and apparatus (pprint01) |
US6357659B1 (en) | 1999-06-03 | 2002-03-19 | Psc Scanning, Inc. | Hands free optical scanner trigger |
JP2000349984A (en) | 1999-06-04 | 2000-12-15 | Fujitsu Ltd | Image reading device and image processing device |
US7369161B2 (en) | 1999-06-08 | 2008-05-06 | Lightsurf Technologies, Inc. | Digital camera device providing improved methodology for rapidly taking successive pictures |
US6969352B2 (en) | 1999-06-22 | 2005-11-29 | Teratech Corporation | Ultrasound probe with integrated electronics |
US6313917B1 (en) | 1999-07-02 | 2001-11-06 | Ncr Corporation | Produce data collector and produce recognition system |
US6788411B1 (en) | 1999-07-08 | 2004-09-07 | Ppt Vision, Inc. | Method and apparatus for adjusting illumination angle |
US6778210B1 (en) | 1999-07-15 | 2004-08-17 | Olympus Optical Co., Ltd. | Image pickup apparatus with blur compensation |
US7079230B1 (en) | 1999-07-16 | 2006-07-18 | Sun Chemical B.V. | Portable authentication device and method of authenticating products or product packaging |
US7768552B1 (en) | 1999-07-23 | 2010-08-03 | Hewlett-Packard Development Company, L.P. | Digital still camera with still and motion image capabilities |
US6754037B1 (en) | 1999-07-28 | 2004-06-22 | Storage Technology Corporation | Small library horseshoe architecture |
JP2001067230A (en) | 1999-08-30 | 2001-03-16 | Toshiba Tec Corp | Store system |
US20010047394A1 (en) | 1999-09-10 | 2001-11-29 | Kloba David D. | System, method, and computer program product for executing scripts on mobile devices |
US7128270B2 (en) | 1999-09-17 | 2006-10-31 | Silverbrook Research Pty Ltd | Scanning device for coded data |
US7605940B2 (en) | 1999-09-17 | 2009-10-20 | Silverbrook Research Pty Ltd | Sensing device for coded data |
US6688525B1 (en) | 1999-09-22 | 2004-02-10 | Eastman Kodak Company | Apparatus and method for reading a coded pattern |
US7270274B2 (en) | 1999-10-04 | 2007-09-18 | Hand Held Products, Inc. | Imaging module comprising support post for optical reader |
EP1226541A2 (en) | 1999-10-04 | 2002-07-31 | Welch Allyn Data Collection, Inc. | Imaging module for optical reader |
US20030029917A1 (en) | 1999-10-04 | 2003-02-13 | Hand Held Products, Inc. | Optical reader for imaging module |
US6585159B1 (en) | 1999-11-02 | 2003-07-01 | Welch Allyn Data Collection, Inc. | Indicia sensor system for optical reader |
EP1102323B1 (en) | 1999-11-19 | 2012-08-15 | CSEM Centre Suisse d'Electronique et de Microtechnique SA - Recherche et Développement | Method for detecting electromagnetic radiation using an optoelectronic sensor |
US6728391B1 (en) | 1999-12-03 | 2004-04-27 | United Parcel Service Of America, Inc. | Multi-resolution label locator |
JP4419234B2 (en) | 1999-12-15 | 2010-02-24 | 株式会社デンソー | Optical information reader |
JP2001177742A (en) | 1999-12-20 | 2001-06-29 | Minolta Co Ltd | Electronic camera |
US20020135683A1 (en) | 1999-12-20 | 2002-09-26 | Hideo Tamama | Digital still camera system and method |
US6836289B2 (en) | 1999-12-20 | 2004-12-28 | Texas Instruments Incorporated | Digital still camera architecture with red and blue interpolation using green as weighting factors |
US6892067B1 (en) | 1999-12-30 | 2005-05-10 | Nokia Corporation | Script based interfaces for mobile phones |
GB2358261B (en) | 2000-01-17 | 2004-06-09 | Advanced Risc Mach Ltd | Data processing with native and interpreted program instruction words |
US6469289B1 (en) | 2000-01-21 | 2002-10-22 | Symagery Microsystems Inc. | Ambient light detection technique for an imaging array |
US6262804B1 (en) * | 2000-02-04 | 2001-07-17 | X-Rite, Incorporated | Handheld color measurement instrument |
EP1128661A1 (en) | 2000-02-22 | 2001-08-29 | Asulab S.A. | Method for driving a CMOS sensor |
US7137555B2 (en) | 2000-02-28 | 2006-11-21 | Psc Scanning, Inc. | Multi-format bar code reader |
JP4272789B2 (en) | 2000-03-01 | 2009-06-03 | 富士フイルム株式会社 | Radiation imaging equipment |
CA2301345A1 (en) | 2000-03-17 | 2001-09-17 | Semiconductor Insights Inc. | Frame capture |
US6489798B1 (en) | 2000-03-30 | 2002-12-03 | Symagery Microsystems Inc. | Method and apparatus for testing image sensing circuit arrays |
JP4419269B2 (en) | 2000-04-18 | 2010-02-24 | 富士通株式会社 | Two-dimensional code extraction method |
EP1152471A3 (en) | 2000-04-28 | 2004-04-07 | Eastman Kodak Company | Image sensor pixel for configurable output |
US7129979B1 (en) | 2000-04-28 | 2006-10-31 | Eastman Kodak Company | Image sensor pixel for global electronic shuttering |
JP2001326847A (en) | 2000-05-16 | 2001-11-22 | Fuji Photo Film Co Ltd | Image pickup device |
US6933488B2 (en) | 2000-06-08 | 2005-08-23 | California Institute Of Technology | Variable electronic shutter in CMOS imager with improved anti smearing techniques |
US6637655B1 (en) | 2000-06-08 | 2003-10-28 | Metrologic Instruments, Inc. | Automatic range adjustment techniques for stand-mountable bar code scanners |
JP2001350685A (en) | 2000-06-08 | 2001-12-21 | Nec Corp | Service switching system and communication system including it |
US6665384B2 (en) | 2000-06-19 | 2003-12-16 | General Electric Company | Methods and apparatus for appliance communication interface |
JP2002042052A (en) | 2000-07-28 | 2002-02-08 | Olympus Optical Co Ltd | Symbol information reading device |
US6456798B1 (en) | 2000-08-09 | 2002-09-24 | Eastman Kodak Company | Barcode and data storage arrangement on a photographic element |
US6576883B1 (en) | 2000-09-25 | 2003-06-10 | Hewlett-Packard Development Company, L.P. | Non-linear gain compensation for optical scanners |
AU2001292946A1 (en) | 2000-09-26 | 2002-04-08 | Advantage 3D Llc | Method and system for generation, storage and distribution of omni-directional object views |
JP3395770B2 (en) | 2000-09-29 | 2003-04-14 | ミノルタ株式会社 | Digital still camera |
US7148923B2 (en) | 2000-09-30 | 2006-12-12 | Hand Held Products, Inc. | Methods and apparatus for automatic exposure control |
JP2002108618A (en) | 2000-10-02 | 2002-04-12 | Techno Core:Kk | Module selection activation system |
US6937774B1 (en) | 2000-10-24 | 2005-08-30 | Lockheed Martin Corporation | Apparatus and method for efficiently increasing the spatial resolution of images |
US7262799B2 (en) | 2000-10-25 | 2007-08-28 | Canon Kabushiki Kaisha | Image sensing apparatus and its control method, control program, and storage medium |
US6813046B1 (en) | 2000-11-07 | 2004-11-02 | Eastman Kodak Company | Method and apparatus for exposure control for a sparsely sampled extended dynamic range image sensing device |
JP2004514225A (en) | 2000-11-17 | 2004-05-13 | ウエイトマン、ヤコブ | Application for mobile digital cameras to distinguish text and image information in images |
US6646246B1 (en) | 2000-11-21 | 2003-11-11 | Eastman Kodak Company | Method and system of noise removal for a sparsely sampled extended dynamic range image sensing device |
US7464877B2 (en) | 2003-11-13 | 2008-12-16 | Metrologic Instruments, Inc. | Digital imaging-based bar code symbol reading system employing image cropping pattern generator and automatic cropped image processor |
US7540424B2 (en) | 2000-11-24 | 2009-06-02 | Metrologic Instruments, Inc. | Compact bar code symbol reading system employing a complex of coplanar illumination and imaging stations for omni-directional imaging of objects within a 3D imaging volume |
US7490774B2 (en) | 2003-11-13 | 2009-02-17 | Metrologic Instruments, Inc. | Hand-supportable imaging based bar code symbol reader employing automatic light exposure measurement and illumination control subsystem integrated therein |
US7077319B2 (en) | 2000-11-24 | 2006-07-18 | Metrologic Instruments, Inc. | Imaging engine employing planar light illumination and linear imaging |
US8682077B1 (en) | 2000-11-28 | 2014-03-25 | Hand Held Products, Inc. | Method for omnidirectional processing of 2D images including recognizable characters |
US6917381B2 (en) | 2000-11-30 | 2005-07-12 | Intel Corporation | Color filter array and color interpolation algorithm |
US6749120B2 (en) | 2000-12-11 | 2004-06-15 | Cpo Technologies Corp. | Method and apparatus for scanning electronic barcodes |
US7268924B2 (en) | 2001-01-22 | 2007-09-11 | Hand Held Products, Inc. | Optical reader having reduced parameter determination delay |
US7270273B2 (en) | 2001-01-22 | 2007-09-18 | Hand Held Products, Inc. | Optical reader having partial frame operating mode |
EP1717728B1 (en) | 2001-01-22 | 2010-09-01 | Hand Held Products, Inc. | Optical reader having partial frame operating mode |
US6637658B2 (en) | 2001-01-22 | 2003-10-28 | Welch Allyn, Inc. | Optical reader having partial frame operating mode |
US6820813B2 (en) | 2001-01-23 | 2004-11-23 | Symbol Technologies, Inc. | Adapter unit having an ergonomic grip for a personal digital assistant |
US6943831B2 (en) | 2001-01-24 | 2005-09-13 | Eastman Kodak Company | Method and apparatus to extend the effective dynamic range of an image sensing device and use residual images |
JP2002240913A (en) | 2001-02-13 | 2002-08-28 | Matsushita Electric Ind Co Ltd | Article control system using mark and tag |
US7382911B1 (en) | 2001-02-16 | 2008-06-03 | Hand Held Products, Inc. | Identification card reader |
JP2002359718A (en) | 2001-03-02 | 2002-12-13 | Canon Inc | Image processing device, information processing method, and control program |
US6476865B1 (en) | 2001-03-07 | 2002-11-05 | Eastman Kodak Company | Sparsely sampled image sensing device with color and luminance photosites |
US6832729B1 (en) | 2001-03-23 | 2004-12-21 | Zih Corp. | Portable data collection device for reading fluorescent indicia |
US6978038B2 (en) | 2001-04-13 | 2005-12-20 | The Code Corporation | Systems and methods for pixel gain compensation in machine-readable graphical codes |
US6619547B2 (en) | 2001-04-30 | 2003-09-16 | The Code Corporation | Image-based graphical code reader device with multi-functional optical element and converging laser targeting |
US20020158127A1 (en) | 2001-04-30 | 2002-10-31 | Pulnix America, Inc. | Matrix code reader |
US20030206150A1 (en) | 2001-05-02 | 2003-11-06 | Hand Held Products, Inc. | Optical reader comprising keyboard |
US7554067B2 (en) | 2001-05-07 | 2009-06-30 | Panavision Imaging Llc | Scanning imager employing multiple chips with staggered pixels |
US7111787B2 (en) | 2001-05-15 | 2006-09-26 | Hand Held Products, Inc. | Multimode image capturing and decoding optical reader |
US6942151B2 (en) | 2001-05-15 | 2005-09-13 | Welch Allyn Data Collection, Inc. | Optical reader having decoding and image capturing functionality |
JP4531294B2 (en) | 2001-06-05 | 2010-08-25 | オリンパス株式会社 | Symbol information reader |
JP4761491B2 (en) | 2001-06-06 | 2011-08-31 | キヤノン株式会社 | Solid-state imaging device and imaging system using the same |
WO2003001435A1 (en) | 2001-06-22 | 2003-01-03 | Emblaze Systems, Ltd | Image based object identification |
JP2003017677A (en) | 2001-06-28 | 2003-01-17 | Canon Inc | Image pickup device |
JP2003015628A (en) | 2001-06-28 | 2003-01-17 | Riso Kagaku Corp | Handwritten character font creation device, handwritten character font creation program, and character entry form for handwritten character font creation |
US7331523B2 (en) | 2001-07-13 | 2008-02-19 | Hand Held Products, Inc. | Adaptive optical image reader |
JP2003032434A (en) | 2001-07-13 | 2003-01-31 | Canon Inc | Image reader, control method for the same, control program and computer readable storage medium |
US6834807B2 (en) | 2001-07-13 | 2004-12-28 | Hand Held Products, Inc. | Optical reader having a color imager |
US7224389B2 (en) | 2001-07-16 | 2007-05-29 | Cypress Semiconductor Corporation (Belgium) Bvba | Method to adjust the signal level of an active pixel and corresponding active pixel |
US20030117491A1 (en) | 2001-07-26 | 2003-06-26 | Dov Avni | Apparatus and method for controlling illumination in an in-vivo imaging device |
GB0118448D0 (en) | 2001-07-27 | 2001-09-19 | Hewlett Packard Co | Apparatus for composing a song and method |
US20030114206A1 (en) | 2001-08-24 | 2003-06-19 | United Parcel Service Of America, Inc. | Portable data acquisition and management system and associated device and method |
US6895219B2 (en) | 2001-08-27 | 2005-05-17 | Symbol Technologies Inc. | Dual use of FFT circuity in imagers and transceivers |
US6400123B1 (en) | 2001-08-29 | 2002-06-04 | Hewlett-Packard Company | Battery fuel gauging using battery chemistry identification |
JP2003116059A (en) | 2001-10-02 | 2003-04-18 | Canon Inc | Imaging device and image processing apparatus |
SE0103286L (en) | 2001-10-03 | 2003-04-04 | Anoto Ab | Optical sensor device and method for controlling its exposure time |
JP4046971B2 (en) | 2001-10-26 | 2008-02-13 | キヤノン株式会社 | Bar code reading apparatus and method |
JP4532800B2 (en) | 2001-11-08 | 2010-08-25 | キヤノン株式会社 | Imaging apparatus and system |
JP2003158680A (en) | 2001-11-20 | 2003-05-30 | Fuji Photo Film Co Ltd | Image pickup device |
JP4086523B2 (en) | 2001-12-04 | 2008-05-14 | キヤノン株式会社 | Image reading apparatus, subject collation system, subject recognition system, and image reading method |
US6789737B2 (en) | 2001-12-19 | 2004-09-14 | Riotec Corp. | Light source mechanism of barcode scanner |
US7748620B2 (en) | 2002-01-11 | 2010-07-06 | Hand Held Products, Inc. | Transaction terminal including imaging module |
US7121470B2 (en) | 2002-01-11 | 2006-10-17 | Hand Held Products, Inc. | Transaction terminal having elongated finger recess |
JP2003215440A (en) | 2002-01-24 | 2003-07-30 | Casio Comput Co Ltd | Autofocus method and autofocus device |
US7057654B2 (en) | 2002-02-26 | 2006-06-06 | Eastman Kodak Company | Four color image sensing apparatus |
JP3561709B2 (en) * | 2002-03-01 | 2004-09-02 | 三洋電機株式会社 | Mobile phone |
JP4374822B2 (en) | 2002-03-07 | 2009-12-02 | 富士ゼロックス株式会社 | PRINT CONTROL DEVICE, PRINT CONTROL METHOD, PRINT DEVICE, AND PRINT CONTROL PROGRAM |
US20030168512A1 (en) | 2002-03-07 | 2003-09-11 | Hand Held Products, Inc. | Optical reader having position responsive decode launch circuit |
US7055747B2 (en) | 2002-06-11 | 2006-06-06 | Hand Held Products, Inc. | Long range optical reader |
JP2002268201A (en) | 2002-03-11 | 2002-09-18 | Dainippon Printing Co Ltd | Method for producing phase shift photomask and blank for the same |
US6976629B2 (en) | 2002-03-20 | 2005-12-20 | Symbol Technologies, Inc. | Image capture system and method |
SE523681C2 (en) | 2002-04-05 | 2004-05-11 | Integrated Vision Prod | System and sensor for mapping properties of an object |
CN1647120A (en) | 2002-04-11 | 2005-07-27 | 传感电子公司 | Portable handheld electronic article surveillance and scanner device |
US6824059B2 (en) | 2002-04-30 | 2004-11-30 | Hewlett-Packard Development Company, L.P. | Apparatus for capturing images and barcodes |
US7012643B2 (en) | 2002-05-08 | 2006-03-14 | Ball Aerospace & Technologies Corp. | One chip, low light level color camera |
US8596542B2 (en) | 2002-06-04 | 2013-12-03 | Hand Held Products, Inc. | Apparatus operative for capture of image data |
US7086596B2 (en) | 2003-01-09 | 2006-08-08 | Hand Held Products, Inc. | Decoder board for an optical reader utilizing a plurality of imaging formats |
US20030222147A1 (en) | 2002-06-04 | 2003-12-04 | Hand Held Products, Inc. | Optical reader having a plurality of imaging modules |
US20030222144A1 (en) | 2002-06-04 | 2003-12-04 | Hand Held Products, Inc. | Manufacturing methods for a decoder board for an optical reader utilizing a plurality of imaging formats |
US7394484B2 (en) | 2002-06-10 | 2008-07-01 | Shimadzu Corporation | Photographic apparatus and photographic method using same |
JP4290935B2 (en) | 2002-07-18 | 2009-07-08 | オリンパス株式会社 | Electronic imaging device |
WO2004036162A1 (en) | 2002-07-26 | 2004-04-29 | Olympus Corporation | Image processing system |
US7421145B2 (en) | 2002-07-31 | 2008-09-02 | Transpacific Ip Ltd. | Image correction method, apparatus, article and image |
US20040023397A1 (en) | 2002-08-05 | 2004-02-05 | Rakesh Vig | Tamper-resistant authentication mark for use in product or product packaging authentication |
US7028911B2 (en) | 2002-08-07 | 2006-04-18 | Shenzhen Syscan Technology Co. Limited | Methods and systems for encoding and decoding data in 2D symbology |
US7110028B1 (en) | 2002-08-13 | 2006-09-19 | Foveon, Inc. | Electronic shutter using buried layers and active pixel sensor and array employing same |
TWI268397B (en) | 2002-08-23 | 2006-12-11 | Pentax Corp | Digital camera |
US20040188644A1 (en) | 2002-09-17 | 2004-09-30 | Quad/Tech, Inc. | Method and apparatus for visually inspecting a substrate on a printing press |
JP4403687B2 (en) | 2002-09-18 | 2010-01-27 | ソニー株式会社 | Solid-state imaging device and drive control method thereof |
US7173663B2 (en) | 2002-10-31 | 2007-02-06 | Freescale Semiconductor, Inc. | Automatic exposure control system for a digital camera |
JP2004159155A (en) | 2002-11-07 | 2004-06-03 | Rohm Co Ltd | Area image sensor |
CN100388764C (en) | 2002-11-07 | 2008-05-14 | 罗姆股份有限公司 | planar image sensor |
JP2004173172A (en) | 2002-11-22 | 2004-06-17 | Fuji Photo Film Co Ltd | Voice reproduction system and electronic camera |
CN100334499C (en) | 2002-12-03 | 2007-08-29 | 矽峰光电科技股份有限公司 | Image pickup apparatus and exposure control method thereof |
US7083096B2 (en) | 2002-12-18 | 2006-08-01 | Symbol Technologies, Inc. | Miniature auto focus piezo actuator system |
US7025272B2 (en) | 2002-12-18 | 2006-04-11 | Symbol Technologies, Inc. | System and method for auto focusing an optical code reader |
JP4226895B2 (en) | 2002-12-27 | 2009-02-18 | 株式会社オプトエレクトロニクス | Optical information reader |
US7195164B2 (en) | 2003-01-03 | 2007-03-27 | Symbol Technologies, Inc. | Optical code reading device having more than one imaging engine |
US7073715B2 (en) | 2003-02-13 | 2006-07-11 | Symbol Technologies, Inc. | Interface for interfacing an imaging engine to an optical code reader |
JP2004274698A (en) | 2003-02-19 | 2004-09-30 | Seiko Epson Corp | Manufacturing method of receiving apparatus and receiving apparatus manufactured using the method |
JP2004264457A (en) | 2003-02-28 | 2004-09-24 | Fuji Photo Optical Co Ltd | Super-high magnifying zoom lens |
US20040196514A1 (en) | 2003-03-05 | 2004-10-07 | Koji Tanimoto | Image sensor unit |
US7471315B2 (en) | 2003-03-14 | 2008-12-30 | Aptina Imaging Corporation | Apparatus and method for detecting and compensating for illuminant intensity changes within an image |
DE20304412U1 (en) | 2003-03-19 | 2003-06-12 | Schott Glas, 55122 Mainz | Control unit for number of lighting units used in a mixed mode for microscopy lighting |
US7276025B2 (en) | 2003-03-20 | 2007-10-02 | Welch Allyn, Inc. | Electrical adapter for medical diagnostic instruments using LEDs as illumination sources |
WO2004090761A1 (en) | 2003-04-07 | 2004-10-21 | Silverbrook Research Pty Ltd | Locations based promotions |
US7574447B2 (en) | 2003-04-08 | 2009-08-11 | United Parcel Service Of America, Inc. | Inbound package tracking systems and methods |
US20040212723A1 (en) | 2003-04-22 | 2004-10-28 | Malcolm Lin | Image pickup apparatus and operating method |
JP4321103B2 (en) | 2003-04-25 | 2009-08-26 | 株式会社デンソーウェーブ | Information code reading apparatus and information code reading method |
JP2004328657A (en) | 2003-04-28 | 2004-11-18 | Toshiba Corp | Image input device, image input method and program |
US7237772B2 (en) | 2003-05-01 | 2007-07-03 | Canon Kabushiki Kaisha | Image forming apparatus with control of sheet carrier to compensate for sheet conveying distance |
US20070241195A1 (en) | 2006-04-18 | 2007-10-18 | Hand Held Products, Inc. | Optical reading device with programmable LED control |
JP2005004718A (en) | 2003-05-16 | 2005-01-06 | Canon Inc | Signal processor and controlling method |
JP2004347163A (en) | 2003-05-20 | 2004-12-09 | Yamaha Corp | Cooking system |
JP4426273B2 (en) | 2003-05-22 | 2010-03-03 | イノテック株式会社 | MOS type solid-state imaging device and manufacturing method thereof |
US7075049B2 (en) | 2003-06-11 | 2006-07-11 | Micron Technology, Inc. | Dual conversion gain imagers |
US8989516B2 (en) | 2007-09-18 | 2015-03-24 | Fotonation Limited | Image processing method and apparatus |
US7636486B2 (en) | 2004-11-10 | 2009-12-22 | Fotonation Ireland Ltd. | Method of determining PSF using multiple instances of a nominally similar scene |
US8264576B2 (en) | 2007-03-05 | 2012-09-11 | DigitalOptics Corporation Europe Limited | RGBW sensor array |
US20130010138A1 (en) | 2003-06-26 | 2013-01-10 | Petronel Bigioi | Digital Camera with an Image Processor |
US9160897B2 (en) | 2007-06-14 | 2015-10-13 | Fotonation Limited | Fast motion estimation method |
US8417055B2 (en) | 2007-03-05 | 2013-04-09 | DigitalOptics Corporation Europe Limited | Image processing method and apparatus |
US7639889B2 (en) | 2004-11-10 | 2009-12-29 | Fotonation Ireland Ltd. | Method of notifying users regarding motion artifacts based on image analysis |
JP2005022802A (en) | 2003-07-01 | 2005-01-27 | Canon Electronics Inc | Cargo delivery system |
JP2005050506A (en) | 2003-07-15 | 2005-02-24 | Shin Etsu Chem Co Ltd | Substrate for magnetic recording medium, method for manufacturing the same and magnetic recording medium |
EP1499111B1 (en) | 2003-07-15 | 2015-01-07 | Canon Kabushiki Kaisha | Image sensiting apparatus, image processing apparatus, and control method thereof |
US7156311B2 (en) | 2003-07-16 | 2007-01-02 | Scanbuy, Inc. | System and method for decoding and analyzing barcodes using a mobile device |
WO2005012997A2 (en) | 2003-07-25 | 2005-02-10 | Color Kinetics, Inc. | Photography methods and systems |
US7428997B2 (en) | 2003-07-29 | 2008-09-30 | Microvision, Inc. | Method and apparatus for illuminating a field-of-view and capturing an image |
US7044377B2 (en) | 2003-08-01 | 2006-05-16 | Symbol Technologies Inc. | Plug-and-play imaging and illumination engine for an optical code reader |
US7277562B2 (en) | 2003-08-01 | 2007-10-02 | Cross Match Technologies, Inc. | Biometric imaging capture system and method |
US7021542B2 (en) | 2003-08-01 | 2006-04-04 | Symbol Technologies, Inc. | Imaging and illumination engine for an optical code reader |
JP4578792B2 (en) * | 2003-09-26 | 2010-11-10 | 富士通セミコンダクター株式会社 | Solid-state imaging device |
US7014114B2 (en) | 2003-10-02 | 2006-03-21 | Symbol Technologies, Inc. | Image capture device for and method of electro-optically reading indicia at low ambient light levels |
US6971579B2 (en) | 2003-10-02 | 2005-12-06 | Symbol Technologies, Inc. | Dynamic omni-directional scan pattern generating arrangement and method in electro-optical readers |
JP2005128413A (en) | 2003-10-27 | 2005-05-19 | Pentax Corp | Lighting control device |
US7472831B2 (en) | 2003-11-13 | 2009-01-06 | Metrologic Instruments, Inc. | System for detecting image light intensity reflected off an object in a digital imaging-based bar code symbol reading device |
US7841533B2 (en) | 2003-11-13 | 2010-11-30 | Metrologic Instruments, Inc. | Method of capturing and processing digital images of an object within the field of view (FOV) of a hand-supportable digitial image capture and processing system |
JP4586026B2 (en) | 2003-11-13 | 2010-11-24 | メトロロジック インスツルメンツ インコーポレイテッド | Hand-supportable imaging-based barcode symbol reader that supports narrow and wide modes of illumination and image capture |
JP4439888B2 (en) | 2003-11-27 | 2010-03-24 | イノテック株式会社 | MOS type solid-state imaging device and driving method thereof |
US20050117017A1 (en) | 2003-12-01 | 2005-06-02 | Baer Richard L. | System and method for imaging regions of interest |
US7364081B2 (en) * | 2003-12-02 | 2008-04-29 | Hand Held Products, Inc. | Method and apparatus for reading under sampled bar code symbols |
JP4194544B2 (en) | 2003-12-05 | 2008-12-10 | キヤノン株式会社 | Solid-state imaging device and driving method of solid-state imaging device |
EP1544673B1 (en) | 2003-12-17 | 2008-11-05 | Agfa-Gevaert HealthCare GmbH | Radiation image storage medium and system for processing such medium |
US7501634B1 (en) | 2003-12-19 | 2009-03-10 | Massachusetts Institute Of Technology | Method and system for distribution of an exposure control signal for focal plane arrays |
US6956544B2 (en) | 2003-12-22 | 2005-10-18 | Motorola, Inc. | Dual mode display |
US20060001761A1 (en) | 2003-12-23 | 2006-01-05 | Tessera, Inc. | Hermetically sealed image sensor module and method of fabricating same |
US7036735B2 (en) | 2004-01-21 | 2006-05-02 | The Code Corporation | Graphical code reader having illumination LEDs of different wavelengths |
ATE523854T1 (en) | 2004-01-23 | 2011-09-15 | Intermec Ip Corp | AUTOFOCUS BARCODE SCANNERS AND THE LIKE WITH MICROFLUIDIC LENSES |
US7087883B2 (en) | 2004-02-04 | 2006-08-08 | Omnivision Technologies, Inc. | CMOS image sensor using shared transistors between pixels with dual pinned photodiode |
KR101110009B1 (en) | 2004-02-27 | 2012-02-06 | 교세라 가부시키가이샤 | Imaging Device and Image Generation Method |
US7333145B2 (en) | 2004-03-05 | 2008-02-19 | Micron Technology, Inc. | Camera module |
US7502505B2 (en) | 2004-03-15 | 2009-03-10 | Microsoft Corporation | High-quality gradient-corrected linear interpolation for demosaicing of color images |
JP2004213689A (en) | 2004-03-15 | 2004-07-29 | Canon Inc | Image input device and fingerprint recognition device |
US7303126B2 (en) | 2004-03-18 | 2007-12-04 | Symbol Technologies, Inc. | System and method for sensing ambient light in an optical code reader |
US7130047B2 (en) * | 2004-04-30 | 2006-10-31 | Optimum Technologies, Inc. | Method of producing polarizers for polarized optical probes |
JP2005354610A (en) | 2004-06-14 | 2005-12-22 | Canon Inc | Image processing apparatus, image processing method and image processing program |
US7376265B2 (en) | 2004-06-17 | 2008-05-20 | Seiko Epson Corporation | Segmentation-based hybrid compression scheme for scanned documents |
US7303131B2 (en) | 2004-07-30 | 2007-12-04 | Symbol Technologies, Inc. | Automatic focusing system for imaging-based bar code reader |
US7347371B2 (en) | 2004-07-30 | 2008-03-25 | Symbol Technologies, Inc. | Aiming pattern for imaging-based bar code readers |
JP4169001B2 (en) | 2004-08-19 | 2008-10-22 | 株式会社デンソーウェーブ | Optical information reader |
US7083098B2 (en) | 2004-08-24 | 2006-08-01 | Symbol Technologies, Inc. | Motion detection in imaging reader |
US7204420B2 (en) | 2004-08-31 | 2007-04-17 | Symbol Technologies, Inc. | Scanner and method for eliminating specular reflection |
US7446753B2 (en) | 2004-09-10 | 2008-11-04 | Hand Held Products, Inc. | Hand held computer device |
JP2006128880A (en) | 2004-10-27 | 2006-05-18 | Funai Electric Co Ltd | Dvd recorder connected to ieee 1394 serial bus, and digital video recording apparatus connected to ieee 1394 serial bus |
US7639888B2 (en) | 2004-11-10 | 2009-12-29 | Fotonation Ireland Ltd. | Method and apparatus for initiating subsequent exposures based on determination of motion blurring artifacts |
KR100614653B1 (en) | 2004-11-18 | 2006-08-22 | 삼성전자주식회사 | CMOS image sensor with global exposure without problems of white point and overflow and its manufacturing method |
US20060109084A1 (en) | 2004-11-24 | 2006-05-25 | Mark Yarvis | Mesh networking with RFID communications |
EP1828958B1 (en) | 2004-12-01 | 2012-05-16 | Datalogic ADC, Inc. | Triggering illumination for a data reader |
US7234641B2 (en) | 2004-12-01 | 2007-06-26 | Datalogic Scanning, Inc. | Illumination pulsing method for a data reader |
JP2006165975A (en) * | 2004-12-07 | 2006-06-22 | Konica Minolta Photo Imaging Inc | Image pickup element, image pickup device and image processing method |
US7204418B2 (en) | 2004-12-08 | 2007-04-17 | Symbol Technologies, Inc. | Pulsed illumination in imaging reader |
WO2006062818A2 (en) | 2004-12-08 | 2006-06-15 | Symbol Technologies, Inc. | Swipe imager scan engine |
US20060138223A1 (en) | 2004-12-23 | 2006-06-29 | Schar Brian A | Shipping information acquisition device and usage |
US7518645B2 (en) | 2005-01-06 | 2009-04-14 | Goodrich Corp. | CMOS active pixel sensor with improved dynamic range and method of operation |
US7398927B2 (en) | 2005-01-26 | 2008-07-15 | Datalogic Scanning, Inc. | Data reader and methods for imaging targets subject to specular reflection |
US7256944B2 (en) * | 2005-02-18 | 2007-08-14 | Eastman Kodak Company | Compact image capture assembly using multiple lenses and image sensors to provide an extended zoom range |
US8654201B2 (en) | 2005-02-23 | 2014-02-18 | Hewlett-Packard Development Company, L.P. | Method for deblurring an image |
US7689465B1 (en) | 2005-03-10 | 2010-03-30 | Amazon Technologies, Inc. | System and method for visual verification of order processing |
US7568628B2 (en) | 2005-03-11 | 2009-08-04 | Hand Held Products, Inc. | Bar code reading device with global electronic shutter control |
US7611060B2 (en) | 2005-03-11 | 2009-11-03 | Hand Held Products, Inc. | System and method to automatically focus an image reader |
EP4236341B1 (en) | 2005-03-11 | 2025-04-30 | Hand Held Products, Inc. | Digital picture taking optical reader having hybrid monochrome and color image sensor array |
JP2008533591A (en) | 2005-03-11 | 2008-08-21 | ハンド ヘルド プロダクツ インコーポレーティッド | Bar code reader with global electronic shutter control |
US7780089B2 (en) | 2005-06-03 | 2010-08-24 | Hand Held Products, Inc. | Digital picture taking optical reader having hybrid monochrome and color image sensor array |
US7416125B2 (en) | 2005-03-24 | 2008-08-26 | Hand Held Products, Inc. | Synthesis decoding and methods of use thereof |
US20090212911A1 (en) | 2005-04-29 | 2009-08-27 | Sang Cheol Min | Radio frequency identification data processing system |
US7383994B2 (en) | 2005-05-03 | 2008-06-10 | Datalogic Scanning, Inc. | Methods and systems for forming images of moving optical codes |
US7770799B2 (en) | 2005-06-03 | 2010-08-10 | Hand Held Products, Inc. | Optical reader having reduced specular reflection read failures |
CN201117008Y (en) | 2005-06-27 | 2008-09-17 | 手持产品公司 | Bar code reading device, marking reading device and data acquisition device |
US7717342B2 (en) | 2005-08-26 | 2010-05-18 | Hand Held Products, Inc. | Data collection device having dynamic access to multiple wireless networks |
US20070063048A1 (en) | 2005-09-14 | 2007-03-22 | Havens William H | Data reader apparatus having an adaptive lens |
US7559473B2 (en) | 2005-09-15 | 2009-07-14 | Symbol Technologies Inc. | Sensor for activating an aiming pattern of a handheld scanner |
WO2007070855A2 (en) | 2005-12-14 | 2007-06-21 | Welch Allyn, Inc. | Medical device wireless adapter |
US7717343B2 (en) * | 2006-01-12 | 2010-05-18 | Hand Held Products, Inc. | High-efficiency illumination in data collection devices |
US20070158535A1 (en) | 2006-01-12 | 2007-07-12 | Cory Watkins | Color imaging using monochrome imagers |
US7699227B2 (en) | 2006-01-13 | 2010-04-20 | Hand Held Products, Inc. | Optical reader |
US9159059B2 (en) | 2006-03-03 | 2015-10-13 | Hand Held Products, Inc. | Method of operating a terminal |
JP4141477B2 (en) | 2006-04-10 | 2008-08-27 | シャープ株式会社 | Image processing device |
US8150163B2 (en) | 2006-04-12 | 2012-04-03 | Scanbuy, Inc. | System and method for recovering image detail from multiple image frames in real-time |
IES20070229A2 (en) | 2006-06-05 | 2007-10-03 | Fotonation Vision Ltd | Image acquisition method and apparatus |
US7784696B2 (en) | 2006-06-09 | 2010-08-31 | Hand Held Products, Inc. | Indicia reading apparatus having image sensing and processing circuit |
US7740176B2 (en) | 2006-06-09 | 2010-06-22 | Hand Held Products, Inc. | Indicia reading apparatus having reduced trigger-to-read time |
US7983373B2 (en) | 2007-02-07 | 2011-07-19 | Vintomie Networks B.V., Llc | Clock distribution for 10GBase-T analog front end |
US7773118B2 (en) | 2007-03-25 | 2010-08-10 | Fotonation Vision Limited | Handheld article with movement discrimination |
KR101116682B1 (en) | 2007-03-29 | 2012-03-07 | 삼성전자주식회사 | Image forming apparatus and control method thereof |
JP4360416B2 (en) | 2007-04-06 | 2009-11-11 | セイコーエプソン株式会社 | Image compression method, image compression apparatus, and program |
EP2149108B1 (en) | 2007-04-23 | 2015-02-25 | FotoNation Limited | Detection and estimation of movement in a camera |
US8134749B2 (en) | 2007-05-25 | 2012-03-13 | Sharp Laboratories Of America, Inc. | Multi-bit-depth error diffusion for the reproduction of color or monochrome images |
US8794526B2 (en) | 2007-06-04 | 2014-08-05 | Hand Held Products, Inc. | Indicia reading terminal processing plurality of frames of image data responsively to trigger signal activation |
US20080309770A1 (en) | 2007-06-18 | 2008-12-18 | Fotonation Vision Limited | Method and apparatus for simulating a camera panning effect |
US8169518B2 (en) | 2007-08-14 | 2012-05-01 | Fujifilm Corporation | Image pickup apparatus and signal processing method |
US20090073516A1 (en) | 2007-09-14 | 2009-03-19 | Kabushiki Kaisha Toshiba | Image forming apparatus |
US7903302B2 (en) | 2007-10-01 | 2011-03-08 | Kabushiki Kaisha Toshiba | Image reading apparatus and image reading method |
US8028919B2 (en) | 2007-10-12 | 2011-10-04 | Symbol Technologies, Inc. | Imaging bar code reader with single prism focus adjustment |
US7874483B2 (en) | 2007-11-14 | 2011-01-25 | Hand Held Products, Inc. | Encoded information reading terminal with wireless path selection capability |
JP4971956B2 (en) | 2007-11-27 | 2012-07-11 | キヤノン株式会社 | Flicker correction apparatus, flicker correction method, and imaging apparatus |
US20090201400A1 (en) | 2008-02-08 | 2009-08-13 | Omnivision Technologies, Inc. | Backside illuminated image sensor with global shutter and storage capacitor |
US8179859B2 (en) | 2008-02-21 | 2012-05-15 | Wang Ynjiun P | Roaming encoded information reading terminal |
JP5035025B2 (en) | 2008-03-03 | 2012-09-26 | ソニー株式会社 | Image processing apparatus, flicker reduction method, imaging apparatus, and flicker reduction program |
EP2109306B1 (en) | 2008-04-07 | 2015-03-11 | CMOSIS nv | Pixel array with global shutter |
US8783573B2 (en) | 2008-12-02 | 2014-07-22 | Hand Held Products, Inc. | Indicia reading terminal having plurality of optical assemblies |
TWI422020B (en) | 2008-12-08 | 2014-01-01 | Sony Corp | Solid-state imaging device |
US8191785B2 (en) | 2009-03-05 | 2012-06-05 | Hand Held Products, Inc. | Encoded information reading terminal operating in infrastructure mode and ad-hoc mode |
US9183425B2 (en) | 2009-04-09 | 2015-11-10 | Hand Held Products, Inc. | Image sensor pixel array having output response curve including logarithmic pattern for image sensor based terminal |
US8218027B2 (en) | 2009-04-09 | 2012-07-10 | Hand Held Products, Inc. | Imaging terminal having color correction |
US20100302420A1 (en) | 2009-05-28 | 2010-12-02 | Symbol Technologies, Inc. | Method and apparatus for reducing the visual discomfort of the illumination generated by imaging scanners |
US20100309340A1 (en) | 2009-06-03 | 2010-12-09 | Border John N | Image sensor having global and rolling shutter processes for respective sets of pixels of a pixel array |
US20100316291A1 (en) | 2009-06-11 | 2010-12-16 | Shulan Deng | Imaging terminal having data compression |
CN102404510B (en) | 2009-06-16 | 2015-07-01 | 英特尔公司 | Camera applications in handheld device |
WO2011014516A1 (en) | 2009-07-28 | 2011-02-03 | University Of Pittsburgh - Of The Commonwealth System Of Higher Education | Estrogen antagonists as treatments for sclerosing disorders |
US8141784B2 (en) | 2009-09-25 | 2012-03-27 | Hand Held Products, Inc. | Encoded information reading terminal with user-configurable multi-protocol wireless communication interface |
US20110080500A1 (en) | 2009-10-05 | 2011-04-07 | Hand Held Products, Inc. | Imaging terminal, imaging sensor having multiple reset and/or multiple read mode and methods for operating the same |
US8302868B2 (en) | 2010-01-15 | 2012-11-06 | Metrologic Instruments, Inc. | Parallel decoding scheme for an indicia reader |
CN102884539B (en) | 2010-03-12 | 2016-08-03 | 日升研发控股有限责任公司 | System and method for product identification |
US8345117B2 (en) | 2010-06-30 | 2013-01-01 | Hand Held Products, Inc. | Terminal outputting monochrome image data and color image data |
US8196834B2 (en) | 2010-07-12 | 2012-06-12 | Symbol Technologies, Inc. | Scan engine with integrated object sensor in electro-optical readers |
US9122939B2 (en) | 2010-09-16 | 2015-09-01 | Datalogic ADC, Inc. | System and method for reading optical codes on reflective surfaces while minimizing flicker perception of pulsed illumination |
US8534556B2 (en) | 2010-10-26 | 2013-09-17 | Symbol Technologies, Inc. | Arrangement for and method of reducing vertical parallax between an aiming pattern and an imaging field of view in a linear imaging reader |
US8387881B2 (en) | 2010-12-01 | 2013-03-05 | Hand Held Products, Inc. | Terminal with screen reading mode |
US9418270B2 (en) | 2011-01-31 | 2016-08-16 | Hand Held Products, Inc. | Terminal with flicker-corrected aimer and alternating illumination |
US8823846B2 (en) | 2011-05-17 | 2014-09-02 | Altasens, Inc. | Pausing digital readout of an optical sensor array |
US20130021507A1 (en) | 2011-07-22 | 2013-01-24 | Hand Held Products, Inc. | Image processors and methods for processing image data |
US20130113967A1 (en) | 2011-11-04 | 2013-05-09 | Honeywell International Inc. Doing Business As (D.B.A.) Honeywell Scanning & Mobility | Apparatus comprising image sensor array having global shutter shared by a plurality of pixels |
US8629926B2 (en) | 2011-11-04 | 2014-01-14 | Honeywell International, Inc. | Imaging apparatus comprising image sensor array having shared global shutter circuitry |
CN104395910B (en) | 2012-03-23 | 2017-05-03 | Opto电子有限公司 | Image reading device capable of producing illumination including a continuous, low-intensity level illumination component and one or more pulsed, high-intensity level illumination components |
US9843757B2 (en) | 2015-06-24 | 2017-12-12 | Stmicroelectronics (Research & Development) Limited | Optical electronic device including enhanced global shutter pixel array and related methods |
US9654712B2 (en) | 2015-10-07 | 2017-05-16 | Semiconductor Components Industries, Llc | Pixels with a global shutter and high dynamic range |
-
2006
- 2006-06-02 US US11/445,930 patent/US7770799B2/en active Active
-
2010
- 2010-08-09 US US12/853,090 patent/US8196839B2/en active Active
-
2012
- 2012-06-11 US US13/493,348 patent/US8720785B2/en active Active
-
2014
- 2014-03-21 US US14/221,874 patent/US9058527B2/en active Active
-
2015
- 2015-04-13 US US14/684,609 patent/US9454686B2/en active Active
-
2016
- 2016-08-23 US US15/244,683 patent/US10002272B2/en active Active
-
2018
- 2018-05-15 US US15/980,213 patent/US10691907B2/en active Active
-
2020
- 2020-05-14 US US16/874,217 patent/US10949634B2/en active Active
-
2021
- 2021-02-04 US US17/167,464 patent/US11238251B2/en active Active
- 2021-04-30 US US17/302,324 patent/US11238252B2/en active Active
- 2021-11-10 US US17/523,231 patent/US11625550B2/en active Active
- 2021-11-10 US US17/523,500 patent/US20220067317A1/en active Pending
-
2022
- 2022-05-26 US US17/825,726 patent/US11604933B2/en active Active
- 2022-05-26 US US17/825,742 patent/US20220284204A1/en active Pending
- 2022-10-28 US US18/050,696 patent/US12001913B2/en active Active
- 2022-10-28 US US18/050,682 patent/US12073283B2/en active Active
- 2022-10-28 US US18/050,708 patent/US12026580B2/en active Active
- 2022-10-28 US US18/050,689 patent/US12020111B2/en active Active
- 2022-10-28 US US18/050,712 patent/US20230147821A1/en active Pending
-
2023
- 2023-02-13 US US18/168,191 patent/US20230196041A1/en active Pending
- 2023-03-10 US US18/181,946 patent/US12001914B2/en active Active
- 2023-04-20 US US18/304,088 patent/US12236312B2/en active Active
- 2023-04-20 US US18/304,065 patent/US20230252252A1/en active Pending
Patent Citations (99)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3716699A (en) * | 1971-04-02 | 1973-02-13 | A Eckert | Method and apparatus for optical code reading |
US4253447A (en) * | 1978-10-16 | 1981-03-03 | Welch Allyn, Inc. | Color endoscope with charge coupled device and television viewing |
USRE31289E (en) * | 1978-10-16 | 1983-06-28 | Welch Allyn, Inc. | Color endoscope with charge coupled device and television viewing |
USRE31290E (en) * | 1978-10-16 | 1983-06-28 | Welch Allyn, Inc. | Color endoscope |
US4261344A (en) * | 1979-09-24 | 1981-04-14 | Welch Allyn, Inc. | Color endoscope |
US4437112A (en) * | 1980-02-15 | 1984-03-13 | Canon Kabushiki Kaisha | Solid-state color imaging apparatus |
US4806776A (en) * | 1980-03-10 | 1989-02-21 | Kley Victor B | Electrical illumination and detecting apparatus |
US4390895A (en) * | 1980-10-07 | 1983-06-28 | Tokyo Shibaura Denki Kabushiki Kaisha | Color image pick-up apparatus |
US4516017A (en) * | 1982-01-20 | 1985-05-07 | Nippondenso Co., Ltd. | High-sensitive optical reading apparatus and method of reading optical information |
US4491865A (en) * | 1982-09-29 | 1985-01-01 | Welch Allyn, Inc. | Image sensor assembly |
US4523224A (en) * | 1982-12-27 | 1985-06-11 | Welch Allyn Inc. | Color filter wheel synchronizer |
US4807981A (en) * | 1984-04-28 | 1989-02-28 | Kabushiki Kaisha Toshiba | Spatial frequency filter |
US5018006A (en) * | 1985-10-31 | 1991-05-21 | Canon Kabushiki Kaisha | Multi-plate type image pickup apparatus having picture elements for producing color and luminance signals |
US4823186A (en) * | 1986-12-19 | 1989-04-18 | Fuji Photo Film Co., Ltd. | Color video signal generating device using monochrome and color image sensors having different resolutions to form a luminance signal |
US5305122A (en) * | 1988-08-31 | 1994-04-19 | Canon Kabushiki Kaisha | Image reading and processing apparatus suitable for use as a color hand-held scanner |
US5019699A (en) * | 1988-08-31 | 1991-05-28 | Norand Corporation | Hand-held optical character reader with means for instantaneously reading information from a predetermined area at an optical sensing area |
US6681994B1 (en) * | 1988-08-31 | 2004-01-27 | Intermec Ip Corp. | Method and apparatus for optically reading information |
US5410141A (en) * | 1989-06-07 | 1995-04-25 | Norand | Hand-held data capture system with interchangable modules |
US5504322A (en) * | 1990-01-05 | 1996-04-02 | Symbol Technologies, Inc. | High density two dimensional bar code symbology |
US5399846A (en) * | 1990-01-05 | 1995-03-21 | Symbol Technologies, Inc. | Systems utilizing a high density two dimensional bar code symbology |
US5393965A (en) * | 1990-11-13 | 1995-02-28 | Symbol Technologies, Inc. | Flexible merchandise checkout and inventory management system |
US5401944A (en) * | 1990-11-20 | 1995-03-28 | Symbol Technologies, Inc. | Traveler security and luggage control system |
US5222477A (en) * | 1991-09-30 | 1993-06-29 | Welch Allyn, Inc. | Endoscope or borescope stereo viewing system |
US5308962A (en) * | 1991-11-01 | 1994-05-03 | Welch Allyn, Inc. | Reduced power scanner for reading indicia |
US5291008A (en) * | 1992-01-10 | 1994-03-01 | Welch Allyn, Inc. | Optical assembly and apparatus employing same using an aspherical lens and an aperture stop |
US5406032A (en) * | 1992-02-13 | 1995-04-11 | The Boeing Company | Apparatus and method for supporting wire bundles within a structure |
US5278642A (en) * | 1992-02-26 | 1994-01-11 | Welch Allyn, Inc. | Color imaging system |
US5756981A (en) * | 1992-02-27 | 1998-05-26 | Symbol Technologies, Inc. | Optical scanner for reading and decoding one- and-two-dimensional symbologies at variable depths of field including memory efficient high speed image processing means and high accuracy image analysis means |
US6049813A (en) * | 1993-02-26 | 2000-04-11 | Intermec Ip Corp. | Portable work station-type data collection system |
US5591955A (en) * | 1993-05-11 | 1997-01-07 | Laser; Vadim | Portable data file readers |
US5602379A (en) * | 1993-05-31 | 1997-02-11 | Kabushiki Kaisha Tec | Optical code reader |
US20030022147A1 (en) * | 1993-06-04 | 2003-01-30 | Paul E. Segall | Solutions for use as plasma expanders and substitutes |
US5513264A (en) * | 1994-04-05 | 1996-04-30 | Metanetics Corporation | Visually interactive encoding and decoding of dataforms |
US5521366A (en) * | 1994-07-26 | 1996-05-28 | Metanetics Corporation | Dataform readers having controlled and overlapped exposure integration periods |
US6045047A (en) * | 1995-01-17 | 2000-04-04 | Welch Allyn Data Collection, Inc. | Two-dimensional part reader having a focussing guide |
US5602377A (en) * | 1995-03-01 | 1997-02-11 | Metanetics Corporation | Bar code dataform scanning and labeling apparatus and method |
US5629734A (en) * | 1995-03-17 | 1997-05-13 | Eastman Kodak Company | Adaptive color plan interpolation in single sensor color electronic camera |
US5506619A (en) * | 1995-03-17 | 1996-04-09 | Eastman Kodak Company | Adaptive color plan interpolation in single sensor color electronic camera |
US6234394B1 (en) * | 1995-03-20 | 2001-05-22 | Symbol Technologies, Inc. | Triggered optical reader |
US6060722A (en) * | 1995-05-15 | 2000-05-09 | Havens; William H. | Optical reader having illumination assembly including improved aiming pattern generator |
US5739518A (en) * | 1995-05-17 | 1998-04-14 | Metanetics Corporation | Autodiscrimination for dataform decoding and standardized recording |
US6053407A (en) * | 1995-05-31 | 2000-04-25 | Metanetics Corporation | Maxicode data extraction using spatial domain features |
US5877487A (en) * | 1995-06-21 | 1999-03-02 | Asahi Kogaku Kogyo Kabushiki Kaisha | Data symbol reading device |
US6019286A (en) * | 1995-06-26 | 2000-02-01 | Metanetics Corporation | Portable data collection device with dataform decoding and image capture capability |
US5714745A (en) * | 1995-12-20 | 1998-02-03 | Metanetics Corporation | Portable data collection device with color imaging assembly |
US5717195A (en) * | 1996-03-05 | 1998-02-10 | Metanetics Corporation | Imaging based slot dataform reader |
US6010073A (en) * | 1996-08-01 | 2000-01-04 | Datalogic S.P.A. | Portable apparatus for reading an optical code |
US20040004128A1 (en) * | 1996-09-03 | 2004-01-08 | Hand Held Products, Inc. | Optical reader system comprising digital conversion circuit |
US6223988B1 (en) * | 1996-10-16 | 2001-05-01 | Omniplanar, Inc | Hand-held bar code reader with laser scanning and 2D image capture |
US6179208B1 (en) * | 1997-01-31 | 2001-01-30 | Metanetics Corporation | Portable data collection device with variable focusing module for optic assembly |
US6010070A (en) * | 1997-06-16 | 2000-01-04 | Nippon Electric Industry Co., Ltd. | Code reading device and method with variable light signal storage time |
US6062475A (en) * | 1997-06-25 | 2000-05-16 | Metanetics Corporation | Portable data collection device including color imaging dataform reader assembly |
US7222789B2 (en) * | 1997-10-17 | 2007-05-29 | Hand Held Products, Inc. | Bar code reading device having image processing mode |
US6561428B2 (en) * | 1997-10-17 | 2003-05-13 | Hand Held Products, Inc. | Imaging device having indicia-controlled image parsing mode |
US6714239B2 (en) * | 1997-10-29 | 2004-03-30 | Eastman Kodak Company | Active pixel sensor with programmable color balance |
US6053408A (en) * | 1997-12-02 | 2000-04-25 | Telxon Corporation | Multi-focal length imaging based portable dataform reader |
US20030004827A1 (en) * | 1998-04-27 | 2003-01-02 | Wang Ynjiun P. | Payment system |
US6062477A (en) * | 1998-04-30 | 2000-05-16 | Ncr Corporation | Apparatus and method of operating a retail terminal having a single-orientation base assembly and a multiple-orientation base assembly |
US6186404B1 (en) * | 1998-05-29 | 2001-02-13 | Welch Allyn Data Collection, Inc. | Security document voiding system |
US6340114B1 (en) * | 1998-06-12 | 2002-01-22 | Symbol Technologies, Inc. | Imaging engine and method for code readers |
US20030019934A1 (en) * | 1998-07-08 | 2003-01-30 | Hand Held Products, Inc. | Optical reader aiming assembly comprising aperture |
US6045238A (en) * | 1998-10-09 | 2000-04-04 | Welch Allyn Inc. | Illumination assembly for an optical viewing device |
US6386452B1 (en) * | 1998-11-26 | 2002-05-14 | Fuji Photo Film Co., Ltd. | Image reading device with improved controller |
US6533168B1 (en) * | 1999-05-27 | 2003-03-18 | Peter N. Ching | Method and apparatus for computer-readable purchase receipts using multi-dimensional bar codes |
US20030034394A1 (en) * | 1999-10-04 | 2003-02-20 | Hand Held Products, Inc. | Optical reader comprising finely adjustable lens assembly |
US20030062413A1 (en) * | 1999-10-04 | 2003-04-03 | Hand Held Products, Inc. | Optical reader comprising multiple color illumination |
US6695209B1 (en) * | 1999-10-04 | 2004-02-24 | Psc Scanning, Inc. | Triggerless optical reader with signal enhancement features |
US6375075B1 (en) * | 1999-10-18 | 2002-04-23 | Intermec Ip Corp. | Method and apparatus for reading machine-readable symbols including color symbol elements |
US6694064B1 (en) * | 1999-11-19 | 2004-02-17 | Positive Systems, Inc. | Digital aerial image mosaic method and apparatus |
US6676016B1 (en) * | 2000-05-04 | 2004-01-13 | Ncr Corporation | Retail terminal configured as consumer gateway to electronic billing application |
US6877664B1 (en) * | 2000-07-11 | 2005-04-12 | Datalogic S.P.A. | Device and optical element for the aiming and the visual indication of reading area of a coded information reader |
US20050041296A1 (en) * | 2000-08-07 | 2005-02-24 | Taiwan Semiconductor Manufacturing Co., Ltd. | High transmittance overcoat for microlens arrays in semiconductor color imagers |
US6522441B1 (en) * | 2000-11-28 | 2003-02-18 | Psc Scanning, Inc. | Micro-optical system for an auto-focus scanner having an improved depth of field |
US6552323B2 (en) * | 2000-12-06 | 2003-04-22 | Eastman Kodak Company | Image sensor with a shared output signal line |
US20040046881A1 (en) * | 2001-04-12 | 2004-03-11 | Nikon Corporation | Imaging device |
US7009638B2 (en) * | 2001-05-04 | 2006-03-07 | Vexcel Imaging Gmbh | Self-calibrating, digital, large format camera with single or multiple detector arrays and single or multiple optical systems |
US20030089775A1 (en) * | 2001-05-21 | 2003-05-15 | Welch Allyn Data Collection, Inc. | Display-equipped optical reader having decode failure image display feedback mode |
US6722569B2 (en) * | 2001-07-13 | 2004-04-20 | Welch Allyn Data Collection, Inc. | Optical reader having a color imager |
US20030018897A1 (en) * | 2001-07-20 | 2003-01-23 | Psc Scanning, Inc. | Video identification verification system and method for a self-checkout system |
US20040031851A1 (en) * | 2002-03-20 | 2004-02-19 | Bianculli Thomas D. | Image capture system and method using a common imaging array |
US6889904B2 (en) * | 2002-03-20 | 2005-05-10 | Symbol Technologies, Inc. | Image capture system and method using a common imaging array |
US7159783B2 (en) * | 2002-03-28 | 2007-01-09 | Hand Held Products, Inc. | Customizable optical reader |
US20040020990A1 (en) * | 2002-06-04 | 2004-02-05 | Havens William H. | Optical reader having a plurality of imaging modules |
US7219843B2 (en) * | 2002-06-04 | 2007-05-22 | Hand Held Products, Inc. | Optical reader having a plurality of imaging modules |
US20040035933A1 (en) * | 2002-06-11 | 2004-02-26 | Havens William H. | Long range optical reader |
US7327504B2 (en) * | 2002-12-06 | 2008-02-05 | Eastman Kodak Company | Method of detecting clipped image pixels |
US7044378B2 (en) * | 2002-12-18 | 2006-05-16 | Symbol Technologies, Inc. | System and method for imaging and decoding optical codes using at least two different imaging settings |
US20050001035A1 (en) * | 2003-05-12 | 2005-01-06 | Thomas Hawley | Picture taking optical reader |
US6880759B2 (en) * | 2003-05-23 | 2005-04-19 | Symagery Microsystems Inc. | Optical reader station |
US20050103854A1 (en) * | 2003-11-13 | 2005-05-19 | Metrologic Instruments, Inc. | Hand-supportable digital imaging-based bar code symbol reader supporting narrow-area and wide-area modes of illumination and image capture |
US20050103864A1 (en) * | 2003-11-13 | 2005-05-19 | Metrologic Instruments, Inc. | Hand-supportable digital imaging-based bar code symbol reading system employing a method of intelligently illuminating an object so as to generate a digital image thereof which is substantially free of noise caused by specular-type reflection of illumination off said object during illumination and imaging operations |
US7187442B2 (en) * | 2004-04-30 | 2007-03-06 | Optimum Technologies, Inc. | Polarized optical probes |
US20060011724A1 (en) * | 2004-07-15 | 2006-01-19 | Eugene Joseph | Optical code reading system and method using a variable resolution imaging sensor |
US20060016335A1 (en) * | 2004-07-22 | 2006-01-26 | Kaz, Incorporated | Air cleaner |
US20060071081A1 (en) * | 2004-10-05 | 2006-04-06 | Ynjiun Wang | System and method to automatically discriminate between a signature and a barcode |
US7219841B2 (en) * | 2004-11-05 | 2007-05-22 | Hand Held Products, Inc. | Device and system for verifying quality of bar codes |
US20070002153A1 (en) * | 2005-06-29 | 2007-01-04 | Bart Dierickx | Hue preservation |
US20070108284A1 (en) * | 2005-11-17 | 2007-05-17 | Hand Held Products, Inc. | Optical reading device with programmable parameter control |
US7343865B2 (en) * | 2006-01-17 | 2008-03-18 | Schuert Technologies Llc | Rackable twin sheet pallet |
Cited By (50)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8978985B2 (en) | 2005-03-11 | 2015-03-17 | Hand Held Products, Inc. | Image reader comprising CMOS based image sensor array |
US12185006B2 (en) | 2005-03-11 | 2024-12-31 | Hand Held Products, Inc. | Image reader comprising CMOS based image sensor array |
US11323649B2 (en) | 2005-03-11 | 2022-05-03 | Hand Held Products, Inc. | Image reader comprising CMOS based image sensor array |
US20110163166A1 (en) * | 2005-03-11 | 2011-07-07 | Hand Held Products, Inc. | Image reader comprising cmos based image sensor array |
US11323650B2 (en) | 2005-03-11 | 2022-05-03 | Hand Held Products, Inc. | Image reader comprising CMOS based image sensor array |
US11317050B2 (en) | 2005-03-11 | 2022-04-26 | Hand Held Products, Inc. | Image reader comprising CMOS based image sensor array |
US10958863B2 (en) | 2005-03-11 | 2021-03-23 | Hand Held Products, Inc. | Image reader comprising CMOS based image sensor array |
US10735684B2 (en) | 2005-03-11 | 2020-08-04 | Hand Held Products, Inc. | Image reader comprising CMOS based image sensor array |
US10721429B2 (en) | 2005-03-11 | 2020-07-21 | Hand Held Products, Inc. | Image reader comprising CMOS based image sensor array |
US10171767B2 (en) | 2005-03-11 | 2019-01-01 | Hand Held Products, Inc. | Image reader comprising CMOS based image sensor array |
US8720781B2 (en) | 2005-03-11 | 2014-05-13 | Hand Held Products, Inc. | Image reader having image sensor array |
US9576169B2 (en) | 2005-03-11 | 2017-02-21 | Hand Held Products, Inc. | Image reader having image sensor array |
US9578269B2 (en) | 2005-03-11 | 2017-02-21 | Hand Held Products, Inc. | Image reader comprising CMOS based image sensor array |
US9465970B2 (en) | 2005-03-11 | 2016-10-11 | Hand Held Products, Inc. | Image reader comprising CMOS based image sensor array |
US8733660B2 (en) | 2005-03-11 | 2014-05-27 | Hand Held Products, Inc. | Image reader comprising CMOS based image sensor array |
US9305199B2 (en) | 2005-03-11 | 2016-04-05 | Hand Held Products, Inc. | Image reader having image sensor array |
US8720784B2 (en) | 2005-06-03 | 2014-05-13 | Hand Held Products, Inc. | Digital picture taking optical reader having hybrid monochrome and color image sensor array |
US10691907B2 (en) | 2005-06-03 | 2020-06-23 | Hand Held Products, Inc. | Apparatus having hybrid monochrome and color image sensor array |
US12236312B2 (en) | 2005-06-03 | 2025-02-25 | Hand Held Products, Inc. | Apparatus having hybrid monochrome and color image sensor array |
US10002272B2 (en) | 2005-06-03 | 2018-06-19 | Hand Held Products, Inc. | Apparatus having hybrid monochrome and color image sensor array |
US12001914B2 (en) | 2005-06-03 | 2024-06-04 | Hand Held Products, Inc. | Apparatus having hybrid monochrome and color image sensor array |
US9058527B2 (en) | 2005-06-03 | 2015-06-16 | Hand Held Products, Inc. | Apparatus having hybrid monochrome and color image sensor array |
US9092654B2 (en) | 2005-06-03 | 2015-07-28 | Hand Held Products, Inc. | Digital picture taking optical reader having hybrid monochrome and color image sensor array |
US9454686B2 (en) | 2005-06-03 | 2016-09-27 | Hand Held Products, Inc. | Apparatus having hybrid monochrome and color image sensor array |
US8196839B2 (en) | 2005-06-03 | 2012-06-12 | Hand Held Products, Inc. | Optical reader having reduced specular reflection read failures |
US8720785B2 (en) | 2005-06-03 | 2014-05-13 | Hand Held Products, Inc. | Apparatus having hybrid monochrome and color image sensor array |
US11238251B2 (en) | 2005-06-03 | 2022-02-01 | Hand Held Products, Inc. | Apparatus having hybrid monochrome and color image sensor array |
US10949634B2 (en) | 2005-06-03 | 2021-03-16 | Hand Held Products, Inc. | Apparatus having hybrid monochrome and color image sensor array |
US9438867B2 (en) | 2005-06-03 | 2016-09-06 | Hand Held Products, Inc. | Digital picture taking optical reader having hybrid monochrome and color image sensor array |
US8727223B2 (en) | 2006-06-09 | 2014-05-20 | Hand Held Products, Inc. | Indicia reading apparatus having image sensor array |
US8646694B2 (en) | 2008-12-16 | 2014-02-11 | Hand Held Products, Inc. | Indicia reading terminal including frame processing |
US8218027B2 (en) | 2009-04-09 | 2012-07-10 | Hand Held Products, Inc. | Imaging terminal having color correction |
US20100259638A1 (en) * | 2009-04-09 | 2010-10-14 | Hand Held Products, Inc. | Imaging terminal having color correction |
US20100316291A1 (en) * | 2009-06-11 | 2010-12-16 | Shulan Deng | Imaging terminal having data compression |
US20110135144A1 (en) * | 2009-07-01 | 2011-06-09 | Hand Held Products, Inc. | Method and system for collecting voice and image data on a remote device and coverting the combined data |
US9521284B2 (en) | 2010-05-21 | 2016-12-13 | Hand Held Products, Inc. | Interactive user interface for capturing a document in an image signal |
US9319548B2 (en) | 2010-05-21 | 2016-04-19 | Hand Held Products, Inc. | Interactive user interface for capturing a document in an image signal |
US9047531B2 (en) | 2010-05-21 | 2015-06-02 | Hand Held Products, Inc. | Interactive user interface for capturing a document in an image signal |
US9451132B2 (en) | 2010-05-21 | 2016-09-20 | Hand Held Products, Inc. | System for capturing a document in an image signal |
US8600167B2 (en) | 2010-05-21 | 2013-12-03 | Hand Held Products, Inc. | System for capturing a document in an image signal |
US8640960B2 (en) | 2011-06-27 | 2014-02-04 | Honeywell International Inc. | Optical filter for image and barcode scanning |
US8636215B2 (en) | 2011-06-27 | 2014-01-28 | Hand Held Products, Inc. | Decodable indicia reading terminal with optical filter |
US9224025B2 (en) | 2011-06-27 | 2015-12-29 | Hand Held Products, Inc. | Decodable indicia reading terminal with optical filter |
US9489557B2 (en) | 2011-06-27 | 2016-11-08 | Hand Held Products, Inc. | Decodable indicia reading terminal with optical filter |
US8985459B2 (en) | 2011-06-30 | 2015-03-24 | Metrologic Instruments, Inc. | Decodable indicia reading terminal with combined illumination |
US8881983B2 (en) | 2011-12-13 | 2014-11-11 | Honeywell International Inc. | Optical readers and methods employing polarization sensing of light from decodable indicia |
US9262661B2 (en) | 2011-12-13 | 2016-02-16 | Honeywell International, Inc. | Optical readers and methods employing polarization sensing of light from decodable indicia |
US8978983B2 (en) | 2012-06-01 | 2015-03-17 | Honeywell International, Inc. | Indicia reading apparatus having sequential row exposure termination times |
US9251392B2 (en) | 2012-06-01 | 2016-02-02 | Honeywell International, Inc. | Indicia reading apparatus |
US8978981B2 (en) | 2012-06-27 | 2015-03-17 | Honeywell International Inc. | Imaging apparatus having imaging lens |
Also Published As
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US12001914B2 (en) | Apparatus having hybrid monochrome and color image sensor array | |
US9438867B2 (en) | Digital picture taking optical reader having hybrid monochrome and color image sensor array |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 12 |