[go: up one dir, main page]

US10380936B2 - Display device, display panel driver, image processing apparatus and image processing method - Google Patents

Display device, display panel driver, image processing apparatus and image processing method Download PDF

Info

Publication number
US10380936B2
US10380936B2 US15/886,565 US201815886565A US10380936B2 US 10380936 B2 US10380936 B2 US 10380936B2 US 201815886565 A US201815886565 A US 201815886565A US 10380936 B2 US10380936 B2 US 10380936B2
Authority
US
United States
Prior art keywords
data
correction
image data
input
starting point
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US15/886,565
Other versions
US20180158214A1 (en
Inventor
Takashi Nose
Hirobumi Furihata
Akio Sugiyama
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Synaptics Inc
Original Assignee
Synaptics Japan GK
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Synaptics Japan GK filed Critical Synaptics Japan GK
Priority to US15/886,565 priority Critical patent/US10380936B2/en
Assigned to SYNAPTICS JAPAN GK reassignment SYNAPTICS JAPAN GK CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: SYNAPTICS DISPLAY DEVICES GK
Assigned to SYNAPTICS DISPLAY DEVICES GK reassignment SYNAPTICS DISPLAY DEVICES GK CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: SYNAPTICS DISPLAY DEVICES KK
Assigned to SYNAPTICS DISPLAY DEVICES KK reassignment SYNAPTICS DISPLAY DEVICES KK ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: NOSE, TAKASHI, FURIHATA, HIROBUMI, SUGIYAMA, AKIO
Publication of US20180158214A1 publication Critical patent/US20180158214A1/en
Application granted granted Critical
Publication of US10380936B2 publication Critical patent/US10380936B2/en
Assigned to SYNAPTICS INCORPORATED reassignment SYNAPTICS INCORPORATED ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SYNAPTICS JAPAN GK
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G3/00Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
    • G09G3/20Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters
    • G09G3/2092Details of a display terminals using a flat panel, the details relating to the control arrangement of the display terminal and to the interfaces thereto
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/001Texturing; Colouring; Generation of texture or colour
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/02Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the way in which colour is displayed
    • G09G5/04Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the way in which colour is displayed using circuits for interfacing with colour displays
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/02Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the way in which colour is displayed
    • G09G5/06Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the way in which colour is displayed using colour palettes, e.g. look-up tables
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2320/00Control of display operating conditions
    • G09G2320/06Adjustment of display parameters
    • G09G2320/066Adjustment of display parameters for control of contrast
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2320/00Control of display operating conditions
    • G09G2320/06Adjustment of display parameters
    • G09G2320/0673Adjustment of display parameters for control of gamma adjustment, e.g. selecting another gamma curve
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/06Colour space transformation

Definitions

  • the present invention relates to a display device, a display panel driver, an image processing apparatus and an image processing method, in particular, to process image data for saturation enhancement of display images in a display device, a display panel driver, an image processing apparatus.
  • the saturation enhancement processing is one of the known image processing techniques applied to image data.
  • a display device which originally has a relatively narrow color gamut for example, performing saturation enhancement processing on image data effectively compensates the color gamut in images actually displayed on the display screen.
  • a liquid crystal display device which uses a white backlight undesirably has a narrow color gamut compared to recent OLED (organic light emitting diode) display devices, and therefore an enlargement of the color gamut is often required to achieve beautiful coloring.
  • Saturation enhancement processing to image data allows enlarging the effective color gamut to meet such a requirement.
  • Saturation enhancement processing is also used to recover the saturation, when the saturation of an image is deteriorated by an auto contrast optimization (ACO).
  • ACO auto contrast optimization
  • a contrast enhancement is achieved in response to characterization data obtained by analyzing image data (for example, the luminance histogram or the average picture level (APL) of the image); note that a contrast enhancement is disclosed in Japanese patent No. 4,198,720 B, for example.
  • the saturation that is, the differences among the grayscale levels of the red color (R), the green color (G) and the blue color (B) may be reduced, because a common contrast enhancement is performed for the red, green and blue colors, as is understood from FIG. 1A .
  • saturation enhancement processing is often performed on image data obtained by a contrast enhancement.
  • FIG. 1B illustrates an example of a system in which saturation enhancement processing is performed after contrast enhancement processing.
  • output image data obtained by the contrast enhancement processing needs to have a wider bit width than that of input image data, to avoid gradation collapse in the contrast enhancement processing.
  • input image data to be subjected to contrast enhancement processing represent the grayscale level of each of the red, green and blue colors with eight bits
  • the output image data of the contrast enhancement processing may be generated as image data which represent the grayscale level of each of the red, green and blue colors with 10 bits.
  • the image data obtained by the saturation enhancement processing need to have a further wider bit width.
  • the output image data of the contrast enhancement processing represent the grayscale level of each of the red, green and blue colors with 10 bits
  • the output image data of the saturation enhancement processing may be generated as image data which represent the grayscale level of each of the red, green and blue colors with 12 bits.
  • the increase in the bit widths of the input and output image data of the saturation enhancement processing undesirably increases the size of the circuit used for the saturation enhancement processing.
  • Japanese Patent Application Publication No. 2010-79119 A discloses a technique in which red-green-blue (RGB) data are transformed into hue-saturation-value (HSV) data and the saturation enhancement is achieved in the HSV color space.
  • Japanese Patent Application Publication No. H06-339017 A discloses a saturation enhancement in which the red (R), green (G), blue (B) values of a saturation-enhanced image are respectively calculated by subtracting products of an enhancement coefficient and the differences between I and the original R, G, and B values from I, where I is the maximum value of the R, G, and B values of each pixel.
  • embodiments of the present invention provide an apparatus and method for image processing which achieve saturation enhancement with a reduce circuit size, and a display panel driver and display device using the same.
  • a display device includes: a display panel and a display panel driver driving the display panel.
  • the display panel driver includes: a correction calculation section which performs correction calculations on input image data to generate saturation-enhanced output image data; a drive circuitry driving the display panel in response to the output image data; and a starting point control section.
  • the correction calculation section generates R data of the output image data by performing a first correction calculation on R data of the input image data, generates G data of the output image data by performing a second correction calculation on G data of the input image data, and generates B data of the output image data by performing a third correction calculation on B data of the input image data.
  • the starting point control section controls a position of a starting point of a first input-output curve corresponding to an input-output relation of the first correction calculation; a position of a starting point of a second input-output curve corresponding to an input-output relation of the second correction calculation; and a position of a starting point of a third input-output curve corresponding to an input-output relation of the third correction calculation.
  • a display panel driver for driving a display panel includes: a correction calculation section which performs correction calculations on input image data to generate saturation-enhanced output image data, a drive circuitry driving the display panel in response to the output image data and a starting point control section.
  • the correction calculation section generates R data of the output image data by performing a first correction calculation on R data of the input image data, generates G data of the output image data by performing a second correction calculation on G data of the input image data, and generates B data of the output image data by performing a third correction calculation on B data of the input image data.
  • the starting point control section controls a position of a starting point of a first input-output curve corresponding to an input-output relation of the first correction calculation; a position of a starting point of a second input-output curve corresponding to an input-output relation of the second correction calculation; and a position of a starting point of a third input-output curve corresponding to an input-output relation of the third correction calculation.
  • an image processing apparatus includes: a correction calculation section which performs correction calculations on input image data to generate saturation-enhanced output image data; and a starting point control section.
  • the correction calculation section generates R data of the output image data by performing a first correction calculation on R data of the input image data, generates G data of the output image data by performing a second correction calculation on G data of the input image data, and generates B data of the output image data by performing a third correction calculation on B data of the input image data.
  • the starting point control section controls a position of a starting point of a first input-output curve corresponding to an input-output relation of the first correction calculation; a position of a starting point of a second input-output curve corresponding to an input-output relation of the second correction calculation; and a position of a starting point of a third input-output curve corresponding to an input-output relation of the third correction calculation.
  • an image processing method includes: generating saturation-enhanced output image data by performing correction calculations on input image data.
  • the step of generating the output image data includes: generating R data of the output image data by performing a first correction calculation on R data of the input image data; generating G data of the output image data by performing a second correction calculation on G data of the input image data; generating B data of the output image data by performing a third correction calculation on B data of the input image data; and controlling a position of a starting point of a first input-output curve corresponding to an input-output relation of the first correction calculation; a position of a starting point of a second input-output curve corresponding to an input-output relation of the second correction calculation; and a position of a starting point of a third input-output curve corresponding to an input-output relation of the third correction calculation.
  • the present invention provides an apparatus and method for image processing which achieve saturation enhancement with a reduce circuit size.
  • FIG. 1A illustrates a reduction in saturation caused by a contrast enhancement
  • FIG. 1B illustrates an example of a system which performs a contrast enhancement and saturation enhancement in series
  • FIG. 2 is a conceptual diagram illustrating saturation enhancement processing in one embodiment of the present invention
  • FIG. 3 is a block diagram illustrating an exemplary configuration of a display device in a first embodiment of the present invention
  • FIG. 4 is a circuit diagram schematically illustrating the configuration of each subpixel
  • FIG. 5 is a block diagram illustrating an exemplary configuration of a driver IC in the first embodiment
  • FIG. 6 is a flowchart illustrating an exemplary procedure of the saturation enhancement processing of the first embodiment
  • FIG. 7 is a graph illustrating an example of a function used to determine an enhancement coefficient in the first embodiment
  • FIG. 8 is a graph illustrating the relation between R data, G data and B data of input image data and R data, G data and B data of output image data in the first embodiment
  • FIG. 9 is a block diagram illustrating an exemplary configuration of a driver IC in a second embodiment of the present invention.
  • FIG. 10 illustrates a gamma curve specified by each correction point data set and contents of a correction calculation (or gamma correction) in accordance with the gamma curve
  • FIG. 11 is a block diagram illustrating an example of the configuration of an approximate gamma correction circuit in the second embodiment
  • FIG. 12 is a block diagram illustrating an example of the configuration of a correction point data calculation circuit in the second embodiment
  • FIG. 13 is a flowchart illustrating the procedure of contrast enhancement processing and saturation enhancement processing in the second embodiment
  • FIG. 14 is a graph illustrating the relation among APL, ⁇ _PIXEL k and a correction point data set CP_L k in one embodiment
  • FIG. 15 a graph illustrating the relation among APL, ⁇ _PIXEL k and a correction point data set CP_L k in another embodiment
  • FIG. 16 is a graph schematically illustrating the shapes of gamma curves respectively corresponding to correction point data sets CP#q and CP#(q+1) and the gamma curve corresponding to the correction point data set CP_L k ;
  • FIG. 17 is a conceptual diagram illustrating a technical significance of modification of the correction point data set CP_L k on the basis of variance data D CHR _ ⁇ 2 ;
  • FIG. 18 illustrates correction point control data CP1_cont k to CP4_cont k .
  • FIG. 19 is a graph illustrating the relation between R data D IN R , G data D IN G and B data D IN B of input image data D IN and R data D OUT R , G data D OUT G and B data D OUT B of output image data D OUT in the second embodiment.
  • FIG. 2 is a conceptual diagram illustrating saturation enhancement processing in one embodiment of the present invention.
  • saturation-enhanced output image data D OUT are generated by performing correction calculations on input image data D IN .
  • the input image data D IN include R data D IN R (data indicating a grayscale level of the red color), G data D IN G (data indicating a grayscale level of the green color) and B data D IN B (data indicating a grayscale level of the blue color).
  • the output image data D OUT include R data D OUT R , G data D OUT G and B data D OUT B .
  • a saturation enhancement is achieved by individually adjusting the position of the starting point of a curve indicating the input-output relation of the correction calculation (that is, the relation between the value of input image data and that of output image data) for each color.
  • a curve indicating the input-output relation of a correction calculation may be referred to as “input-output curve”.
  • starting point means the point on an input-output curve for input image data with the allowed minimum value.
  • the term “input-output curve” means to include the case when the input-output relation is linear, that is, the case when the input-output relation is represented with a straight line.
  • R data D OUT R of output image data D OUT are generated by performing a first correction calculation on R data D IN R of input image data D IN .
  • G data D OUT G of the output image data D OUT are generated by performing a second correction calculation on G data D IN G of the input image data D IN
  • B data D OUT B of the output image data D OUT are generated by performing a third correction calculation on B data D IN B of the input image data D IN .
  • the left graph illustrates the input-output curve of the first correction calculation
  • the center graph illustrates the input-output curve of the second correction calculation
  • the right graph illustrates the input-output curve of the third correction calculation.
  • the saturation enhancement is achieved by controlling the position of the starting point CP0 R of the input-output curve corresponding to the input-output relation of the first correction calculation, the position of the starting point CP0 G of the input-output curve corresponding to the input-output relation of the second correction calculation, and the position of the starting point CP0 B of the input-output curve corresponding to the input-output relation of the third correction calculation.
  • saturation enhancement processing allows enhancing the saturation with a simple calculation. This effectively reduces the size of the circuit used to perform the saturation enhancement processing.
  • the saturation enhancement processing of the present embodiment is easy to be combined with other image processing techniques, such as contrast enhancement.
  • a saturation enhancement is achieved by adjusting the position of the starting point of the input-output curve, while a contrast enhancement is achieved by determining the overall shape of the input-output curve. This feature is also advantageous for reducing the size of the circuit used to perform the saturation enhancement processing.
  • the position of the starting point CP0 R of the input-output curve of the correction calculation performed on the R data D IN R of the input image data D IN is determined in response to the difference D IN R ⁇ Y PIXEL , where Y PIXEL is the luminance value of the certain pixel calculated from the input image data D IN .
  • the position of the starting point CP0 G of the input-output curve of the correction calculation performed on the G data D IN G of the input image data D IN is determined in response to the difference D IN G ⁇ Y PIXEL
  • the position of the starting point CP0 B of the input-output curve of the correction calculation performed on the B data D IN B of the input image data D IN is determined in response to the difference D IN B ⁇ Y PIXEL .
  • the luminance value Y PIXEL is a value determined as a weighted average of the R data D IN R , G data D IN G and B data D IN B , at least one of the differences D IN R ⁇ Y PIXEL , D IN G ⁇ Y PIXEL and D IN B ⁇ Y PIXEL is positive and at least another one is negative. Accordingly, the saturation can be enhanced by determining the positions of the starting points CP0 R , CP0 G and CP0 B on the basis of the differences D IN R ⁇ Y PIXEL , D IN G ⁇ Y PIXEL and D IN B ⁇ Y PIXEL .
  • FIG. 2 illustrates saturation enhancement processing in the case when the difference D IN R ⁇ Y PIXEL is positive and the differences D IN G ⁇ Y PIXEL and D IN B ⁇ Y PIXEL are negative.
  • the starting points CP0 R , CP0 G and CP0 B may be determined in response to the average saturation of the frame image displayed in the certain frame period. This effectively allows an improved saturation enhancement suitable for the average saturation of the frame image.
  • the position of the starting point CP0 R is determined on the basis of the product of the difference D IN R ⁇ Y PIXEL and an enhancement coefficient INST determined as described below
  • the position of the starting point CP0 G is determined on the basis of the product of the difference D IN G ⁇ Y PIXEL and the enhancement coefficient INST
  • the position of the starting point CP0 B is determined on the basis of the product of the difference D IN B ⁇ Y PIXEL and the enhancement coefficient INST.
  • the enhancement coefficient INST is determined as the minimum value of enhancement coefficients INST R , INST G and INST B , where the enhancement coefficients INST R , INST G and INST B are obtained from the R data D IN R , G data D IN G , and B data D IN B of the input image data D IN , respectively.
  • is determined as D IN MAX /2 or the integer closest to D IN MAX /2 (if there are two integers closest to D IN MAX /2, one selected from the two closest integers), where D IN MAX is the allowed maximum value of the R data D IN R , G data D IN G , and B data D IN B .
  • D IN MAX is the allowed maximum value of the R data D IN R , G data D IN G , and B data D IN B .
  • the allowed maximum value D IN MAX is “255” and ⁇ is “127” or “128”.
  • the saturation is enhanced more strongly when the values of the R data D IN R , G data D IN G , and B data D IN B (that is, the grayscale values of the red, green and blue colors) are about a half of the allowed maximum value D IN MAX and this effectively achieves improved saturation enhancement.
  • the positions of the starting points CP0 R , CP0 G and CP0 B may be controlled in response to the average saturation S AVE in the frame image. More specifically, the position of the starting point CP0 R is controlled in response to the product of the enhancement coefficient INST, the difference D IN MAX ⁇ S AVE and the difference D IN R ⁇ Y PIXEL .
  • the position of the starting point CP0 G may be controlled in response to the product of the enhancement coefficient INST, the difference D IN MAX ⁇ S AVE and the difference D IN G ⁇ Y PIXEL
  • the position of the starting point CP0 B may be controlled in response to the product of the enhancement coefficient INST, the difference D IN MAX ⁇ S AVE , and the difference D IN B ⁇ Y PIXEL .
  • FIG. 3 is a block diagram illustrating an exemplary configuration of a display device in a first embodiment of the present invention.
  • the display device of the present embodiment which is configured as a liquid crystal display device denoted by numeral 1 , includes an LCD (liquid crystal display) panel 2 and a driver IC (integrated circuit) 3 .
  • LCD liquid crystal display
  • driver IC integrated circuit
  • the LCD panel 2 includes a display region 5 and a gate line drive circuit 6 (also referred to as GIP (gate in panel) circuit). Disposed in the display region 5 are a plurality of gate lines 7 (also referred to as scan lines or address lines), a plurality of data lines 8 (also referred to as signal lines or source lines) and pixels 9 .
  • the number of the gate lines 7 is v and the number of the data lines 8 is 3h; the pixels 9 are arranged in v rows and h columns in the display region 5 , where v and h are integers equal to or more than two.
  • the horizontal direction of the display region 5 (that is, the direction in which the gate lines 7 are extended) may be referred to as X-axis direction and the vertical direction of the display region 5 (that is, the direction in which the data lines 8 are extended) may be referred to as Y-axis direction.
  • each pixel 9 includes three subpixels: an R subpixel 11 R, a G subpixel 11 G and a B subpixel 11 B, where the R subpixel 11 R is a subpixel corresponding to the red color (that is, a subpixel displaying the red color), the G subpixel 11 G is a subpixel corresponding to the green color (that is, a subpixel displaying the green color) and the B subpixel 11 B is a subpixel corresponding to the blue color (that is, a subpixel displaying the blue color).
  • the R subpixel 11 R, G subpixel 11 G and B subpixel 11 B may be collectively referred to as subpixel 11 if not distinguished from each other.
  • subpixels 11 are arrayed in v rows and 3h columns on the LCD panel 2 .
  • Each subpixel 11 is connected with one corresponding gate line 7 and one corresponding data line 8 .
  • gate lines 7 are sequentially selected and desired drive voltages are written into the subpixels 11 connected with a selected gate line 7 via the data lines 8 . This allows setting the respective subpixels 11 to desired grayscale levels to thereby display a desired image in the display region 5 of the LCD panel 2 .
  • FIG. 4 is a circuit diagram schematically illustrating the configuration of each subpixel 11 .
  • Each subpixel 11 includes a TFT (thin film transistor) 12 and a pixel electrode 13 .
  • the TFT 12 has a gate connected with a gate line 7 , a source connected with a data line 8 and a drain connected with the pixel electrode 13 .
  • the pixel electrode 13 is disposed opposed to the opposing electrode (common electrode) 14 of the LCD panel 2 and the space between each pixel electrode 13 and the opposing electrode 14 is filled with liquid crystal.
  • FIG. 4 illustrates the subpixel 11 as if the opposing electrode 14 may be separately disposed for each subpixel 11 , a person skilled in the art would appreciate that the opposing electrode 14 is actually shared by the subpixels 11 of the entire LCD panel 2 .
  • the driver IC 3 drives the data lines 8 and also generates gate line control signals S GIP for controlling the gate line drive circuit 6 .
  • the drive of the data lines 8 is responsive to input image data D IN and synchronization data D SYNC received from a processor 4 (for example, a CPU (central processing unit)).
  • a processor 4 for example, a CPU (central processing unit)
  • the input image data D IN are image data corresponding to images to be displayed in the display region 5 of the LCD panel 2 , more specifically, data indicating the grayscale levels of each subpixel 11 of each pixel 9 .
  • the input image data D IN represent the grayscale level of each subpixel 11 of each pixel 9 with eight bits.
  • the input image data D IN represent the grayscale levels of each pixel 9 of the LCD panel 2 with 24 bits.
  • the synchronization data D SYNC are used to control the operation timing of the driver IC 3 ; the generation timing of various timing control signals in the driver IC 3 , including the vertical synchronization signal V SYNC and the horizontal synchronization signal H SYNC , is controlled in response to the synchronization data D SYNC .
  • the gate line control signals S GIP are generated in response to the synchronization data D SYNC .
  • the driver IC 3 is mounted on the LCD panel 2 with a surface mounting technology such as a COG (chip on glass) technology.
  • FIG. 5 is a block diagram illustrating an example of the configuration of the driver IC 3 .
  • the driver IC 3 includes an interface circuit 21 , a linear function correction circuit 22 , a color reduction circuit 23 , a latch circuit 24 , a grayscale voltage generator circuit 25 , a data line drive circuit 26 , a timing control circuit 27 , and a starting point control circuit 28 .
  • the interface circuit 21 receives the input image data D IN and synchronization data D SYNC from the processor 4 and forwards the input image data D IN to the linear function correction circuit 22 and the synchronization data D SYNC to the timing control circuit 27 .
  • the linear function correction circuit 22 performs saturation enhancement processing as described above; the linear function correction circuit 22 generates output image data D OUT by performing saturation enhancement processing on the input image data D IN .
  • data indicating the grayscale level of an R subpixel 11 R of input image data D IN may be referred to as input image data D IN R .
  • data indicating the grayscale level of a G subpixel 11 G of input image data D IN may be referred to as input image data D IN G
  • data indicating the grayscale level of a B subpixel 11 B of input image data D IN may be referred to as input image data D IN B
  • data indicating the grayscale level of an R subpixel 11 R of the output image data D OUT may be referred to as output image data D OUT R .
  • data indicating the grayscale level of a G subpixel 11 G of the output image data D OUT may be referred to as output image data D OUT G
  • data indicating the grayscale level of a B subpixel 11 B of the output image data D OUT may be referred to as output image data D OUT B .
  • straight lines are used as the input-output curves of the saturation enhancement processing performed in the linear function correction circuit 22 (that is, the curves indicating the input-output relation between the input image data D IN inputted to the linear function correction circuit 22 and the output image data D OUT outputted from the linear function correction circuit 22 ), and the positions of the starting points of the input-output curves are specified by starting point control data CP0_cont R , CP0_cont G and CP0_cont B received from the starting point control circuit 28 .
  • the starting point control data CP0_cont R specifies the position of the starting point of the input-output curve of the processing to be performed on the R data D IN R of the input image data D IN .
  • the starting point control data CP0_cont G specifies the position of the starting point of the input-output curve of the processing to be performed on the G data D IN G of the input image data D IN and the starting point control data CP0_cont B specifies the position of the starting point of the input-output curve of the processing to be performed on the B data D IN B of the input image data D IN .
  • the number of bits of the R data D OUT R , G data D OUT G and B data D OUT B of the output image data D OUT is larger than that of the R data D IN R , G data D IN G and B data D IN B of the input image data D IN . This effectively avoids losing information of the grayscale levels of pixels in the correction calculation.
  • the R data D OUT R , G data D OUT G and B data D OUT B of the output image data D OUT may be, for example, generated as 10-bit data.
  • the color reduction circuit 23 , the latch circuit 24 , the grayscale voltage generator circuit 25 and the data line drive circuit 26 function in total as a drive circuitry which drives the data lines 8 of the display region 5 of the LCD panel 2 in response to the output image data D OUT generated by the linear function correction circuit 22 .
  • the color reduction circuit 23 performs a color reduction on the output image data D OUT generated by the linear function correction circuit 22 to generate color-reduced image data D OUT _ D .
  • the latch circuit 24 latches the color-reduced image data D OUT _ D from the color reduction circuit 23 in response to a latch signal S STB received from the timing control circuit 27 and forwards the color-reduced image data D OUT _ D to the data line drive circuit 26 .
  • the grayscale voltage generator circuit 25 feeds a set of grayscale voltages to the data line drive circuit 26 .
  • the data line drive circuit 26 drives the data lines 8 of the display region 5 of the LCD panel 2 in response to the color-reduced image data D OUT _ D received from the latch circuit 24 .
  • the data line drive circuit 26 selects desired grayscale voltages from the set of the grayscale voltages received from the grayscale voltage generator circuit 25 in response to color-reduced image data D OUT _ D , and drives the corresponding data lines 8 of the LCD panel 2 to the selected grayscale voltages.
  • the timing control circuit 27 performs timing control of the entire drive IC 3 in response to the synchronization data D SYNC .
  • the timing control circuit 27 generates the latch signal S STB in response to the synchronization data D SYNC and feeds the generated latch signal S STB to the latch circuit 24 .
  • the latch signal S STB is a control signal instructing the latch circuit 24 to latch the color-reduced data D OUT _ D .
  • the timing control circuit 27 generates a frame signal S FRM in response to the synchronization data D SYNC and feeds the generated frame signal S FRM to the starting point control circuit 28 .
  • the frame signal S FRM is a control signal which informs the starting point control circuit 28 of the start of each frame period; the frame signal S FRM is asserted at the beginning of each frame period.
  • the vertical synchronization signal V SYNC generated in response to the synchronization data D SYNC may be used as the frame signal S FRM .
  • the timing control circuit 27 also generates coordinate data D (X, Y) indicating the coordinates of the pixel 9 for which the input image data D IN currently indicate the grayscale levels of the respective subpixels 11 thereof and feeds the generated coordinate data D (X, Y) to the starting point control circuit 28 .
  • the timing control circuit 27 feeds the coordinate data D (X, Y) indicating the coordinates of the certain pixel 9 in the display region 5 to the starting point control circuit 28 .
  • the starting point control circuit 28 controls the saturation enhancement processing performed in the linear function correction circuit 22 .
  • the starting point control circuit 28 generates the above-described starting point control data CP0_cont R , CP0_cont G and CP0_cont B in response to the input image data D IN , to thereby control the position of the starting point of the input-output curve of the processing performed on the input image data D IN for each color.
  • the starting point control circuit 28 includes an average saturation calculation section 28 a , a luminance difference calculation section 28 b , an enhancement coefficient calculation section 28 c , and a starting point control data generator section 28 d.
  • the average saturation calculation section 28 a calculates the average saturation S AVE of each frame image (that is, the image displayed in the display region 5 of the LCD panel 2 in each frame period), from the input image data D IN .
  • the luminance difference calculation section 28 b calculates, for each pixel 9 , the difference Rdist between the R data D IN R and the luminance value Y PIXEL , the difference Gdist between the G data D IN G and the luminance value Y PIXEL , and the difference Bdist between the B data D IN B and the luminance value Y PIXEL , from the input image data D IN associated with each pixel 9 .
  • the enhancement coefficient calculation section 28 c calculates an enhancement coefficient INST from the R data D IN R , G data D IN G , and B data D IN B of the input image data D IN for each pixel 9 .
  • the enhancement coefficient INST is a coefficient indicating the degree of the saturation enhancement in the saturation enhancement processing.
  • the starting point control data generator section 28 d generates the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B in response to the average saturation S AVE calculated by the average saturation calculation section 28 a , the differences Rdist, Gdist, and Bdist calculated by the luminance difference calculation section 28 b and the enhancement coefficient INST calculated by the enhancement coefficient calculation section 28 c .
  • the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B used for the saturation enhancement processing for the certain pixel 9 are calculated on the basis of the average saturation S AVE of the frame image displayed in the certain frame period, the differences Rdist, Gdist, and Bdist and the enhancement coefficient INST calculated from the R data D IN R , G data D IN G , and B data D IN B of the input image data D IN associated with the certain pixel 9 .
  • FIG. 6 is a flowchart illustrating the saturation enhancement processing in the present embodiment, more specifically, the processing performed in the starting point control circuit 28 and the linear function correction circuit 22 .
  • the saturation enhancement processing in the present embodiment includes calculation of starting point control data CP0_cont R , CP0_cont G , and CP0_cont B (steps S 11 to S 14 ) and processing on the input image data D IN on the basis of the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B (step S 15 ).
  • the following processing is performed in the calculation of the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B .
  • the average saturation S AVE of the frame image displayed in each frame period is calculated from the input image data D IN .
  • the saturation S of a certain pixel 9 is calculated as the difference between the maximum value and minimum value of the R data D IN R , G data D IN G , and B data D IN B of the input image data D IN associated with the certain pixel 9 .
  • the average saturation S AVE of a certain frame image is calculated as the average of the saturations of all the pixels 9 in the frame image. More specifically, the average saturation S AVE of a certain frame image is calculated in accordance with the following expressions (1a) and (1b):
  • max(R j , G j , B j ) is the maximum value of R j , G j , and B j and min(R j , G j , B j ) is the minimum value of R j , G j , and B j .
  • Data_count is the number of the pixels 9 in the display region 5 (the number of the pixels 9 in the frame image) and ⁇ represents the sum with respect to all the pixels 9 in the display region 5 .
  • the average saturation S AVE is calculated by the average saturation calculation section 28 a . It is only necessary to calculate the average saturation S AVE once for each frame period.
  • the luminance value Y PIXEL of each pixel 9 is further calculated from the input image data D IN and the difference Rdist between the R data D IN R and the luminance value Y PIXEL , the difference Gdist between the G data D IN G and the luminance value Y PIXEL and the difference Bdist between the B data D IN B and the luminance value Y PIXEL are calculated from the input image data D IN associated with each pixel 9 .
  • the luminance value Y PIXEL and the differences Rdist, Gdist, and Bdist are calculated by the luminance difference calculation section 28 b.
  • the weights a, b, and c used for the calculation of the luminance value Y PIXEL depend on the definition of the luminance value Y PIXEL .
  • the difference Rdist between the R data D IN R and the luminance value Y PIXEL , the difference Gdist between the G data D IN G and the luminance value Y PIXEL , and the difference Bdist between the B data D IN B and the luminance value Y PIXEL are calculated in accordance with the following expressions (3a) to (3c):
  • R dist R j ⁇ Y PIXEL (3a)
  • G dist G j ⁇ Y PIXEL
  • B dist B j ⁇ Y PIXEL .
  • an enhancement coefficient INST associated with each pixel 9 is further calculated from the input image data D IN associated with each pixel 9 .
  • the calculation of the enhancement coefficient INST is achieved by the enhancement coefficient calculation section 28 c as described above.
  • the enhancement coefficient INST is calculated for each pixel 9 as described below: First, enhancement coefficients INST R , INST G , and INST B are respectively calculated for R data D IN R , G data D IN G , and B data D IN B of the input image data D IN associated with each pixel 9 .
  • may be determined as D IN MAX /2 or the integer closest to D IN MAX /2 (if there are two integers closest to D IN MAX /2, one selected from the two integers), where D IN MAX is the allowed maximum value of the R data D IN R , G data D IN G , and B data D IN B .
  • D IN MAX is the allowed maximum value of the R data D IN R , G data D IN G , and B data D IN B .
  • the allowed maximum value D IN MAX is “255” and ⁇ is “127” or “128”.
  • FIG. 7 is a graph illustrating a specific example of a function used to determine the enhancement coefficients INST R , INST G , and INST B .
  • the enhancement coefficients INST R , INST G , and INST B associated with pixel j may be calculated from the values R j , G j , and B j of the R data D IN R , G data D IN G , and B data D IN B of the input image data D IN associated with pixel j in accordance with the following expressions (6a) to (6c):
  • FIG. 7 is a graph illustrating the enhancement coefficients INST R , INST G , and INST B calculated in accordance with expressions (7a) to (7c).
  • Starting point control data CP0_cont R , CP0_cont G , and CP0_cont B are calculated on the basis of the average saturation S AVE , differences Rdist, Gdist, and Bdist and enhancement coefficient INST, which are calculated as described above.
  • the calculation of the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B are achieved by the starting point control data generator section 28 d.
  • the starting point control data CP0_cont R is determined so that the starting point control data CP0_cont R increases as the difference D IN MAX ⁇ S AVE increases and is proportional to the difference Rdist, where the difference D IN MAX ⁇ S AVE is obtained by subtracting the average saturation S AVE from the allowed maximum value D IN MAX .
  • the starting point control data CP0_cont G is determined so that the starting point control data CP0_cont G increases as the difference D IN MAX ⁇ S AVE increases and is proportional to the difference Gdist
  • the starting point control data CP0_cont B is determined so that the starting point control data CP0_cont B increases as the difference D IN MAX ⁇ S AVE increases and is proportional to the difference Bdist.
  • the enhancement coefficient INST is used as a parameter indicating the degree of the saturation enhancement, commonly for the calculations of the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B .
  • the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B may be calculated in accordance with the following expressions (9a) to (9c):
  • CP0_cont R INST ⁇ D IN MAX - S AVE D IN MAX ⁇ Rdist ( 9 ⁇ a )
  • CP0_cont G INST ⁇ D IN MAX - S AVE D IN MAX ⁇ Gdist ( 9 ⁇ b )
  • CP0_const B INST ⁇ D IN MAX - S AVE D IN MAX ⁇ Bdist ( 9 ⁇ c )
  • the input image data D IN are processed in response to the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B , to thereby calculate output image data D OUT .
  • the output image data D OUT are obtained as a result of the saturation enhancement processing to the input image data D IN .
  • the input-output curves of the processing applied to the R data D IN R , G data D IN G , and B data D IN B of the input image data D IN are each defined as a straight line.
  • the starting point of the input-output curves of the processing applied to the R data D IN R of the input image data D IN is specified by the starting point control data CP0_cont R .
  • the starting point of the input-output curves of the processing applied to the G data D IN G of the input image data D IN is specified by the starting point control data CP0_cont G and the starting point of the input-output curves of the processing applied to the B data D IN B of the input image data D IN is specified by the starting point control data CP0_cont B .
  • the end point of the input-output curve of the processing to the input image data D IN (the point on the input-output curve corresponding to input image data D IN with the allowed maximum value D IN MAX ) is determined commonly for the R data D IN R , G data D IN G , and B data D IN B , so that the value of the output image data D OUT is determined as the allowed maximum value D OUT MAX when the value of the input image data D IN is the allowed maximum value D IN MAX .
  • allowed maximum value D OUT MAX depends on the bit width of the R data D OUT R , G data D OUT G , and B data D OUT B of the output image data D OUT .
  • D OUT R D OUT MAX - CP0_cont R D IN MAX ⁇ D IN R + CP0_cont R ( 10 ⁇ a )
  • D OUT G D OUT MAX - CP0_cont G D IN MAX ⁇ D IN G + CP0_cont G ( 10 ⁇ b )
  • D OUT B D OUT MAX - CP0_cont B D IN MAX ⁇ D IN B + CP0_cont B ( 10 ⁇ c )
  • FIG. 8 illustrates the respective relations between the R data D IN R , G data D IN G , and B data D IN B of the input image data D IN and the R data D OUT R , G data D OUT G , and B data D OUT B of the output image data D OUT in the case when the R data D OUT R , G data D OUT G , and B data D OUT B of the output image data D OUT are calculated in accordance with the above expressions (10a) to (10c).
  • the symbol “CP5” denotes the end point of each input-output curve.
  • At least one of the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B is positive and at least another one is negative. This is because the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B are calculated in accordance with expressions (9a) to (9c) and at least one of the differences Rdist, Gdist, and Bdist is positive and at least another one is negative, where the luminance value Y PIXEL is a weighted average of the R data D IN R , G data D IN G , and B data D IN B of the input image data D IN . It would be easily understood from FIG.
  • the output image data D OUT calculated by the linear function correction circuit 22 with the above-described series of expressions are forwarded to the color reduction circuit 23 .
  • a color reduction is performed on the output image data D OUT to generate the color-reduced image data D OUT _ D .
  • the color-reduced image data D OUT _ D are forwarded to the data line drive circuit 26 via the latch circuit 24 and the data lines 8 of the LCD panel 2 are driven in response to the color-reduced image data D OUT _ D .
  • the above-described saturation enhancement processing in the present embodiment effectively achieves saturation enhancement only with simple processing. This effectively contributes the reduction of the circuit size of an image processing circuit used for the saturation enhancement (the starting point control circuit 28 and the linear function correction circuit 22 in this embodiment).
  • the saturation enhancement processing in the present embodiment only causes a reduced change in the luminance value Y PIXEL of each pixel.
  • R j , G j , and B j which are the values of the R data D IN R , G data D IN G , and B data D IN B of the input image data D IN , are 50, 100, and 200, respectively.
  • the luminance value Y PIXEL is obtained in accordance with expression (2c) as follows:
  • the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B are obtained by expressions (9a) to (9c). It should be noted that, with respect to (D IN MAX ⁇ S AVE )/D IN MAX recited in expressions (9a) to (9c), it holds: 0 ⁇ ( D IN MAX ⁇ S AVE )/ D IN MAX ⁇ 1. To discuss the change in the luminance value Y PIXEL in the case when the saturation is most strongly enhanced, let us consider the case when (D IN MAX ⁇ S AVE )/D IN MAX is one; in this case, the saturation is most strongly enhanced.
  • the values of the R data D OUT R , G data D OUT G , and B data D OUT B are represented with 10 bits, taking a value from 0.00 to 255.00 in increments of 0.25.
  • the luminance value Y PIXEL ′ after the saturation enhancement processing is calculated from the thus-obtained output image data D OUT as follows:
  • the saturation enhancement processing in the present embodiment causes only a reduced change in the luminance value, since the luminance value Y PIXEL calculated from the original input image data D IN is 96.59.
  • the reduced change in the luminance value caused by the saturation enhancement processing results from that the enhancement coefficient INST is calculated with a function f(x) and ⁇ , where ⁇ is determined as D IN MAX /2 or a value close to D IN MAX /2, and the function f(x) satisfies the following conditions (a) to (c):
  • the calculation of the enhancement coefficient INST in this way effectively reduces changes in the luminance value in performing the saturation enhancement processing.
  • FIG. 9 is a block diagram illustrating an example of the configuration of the driver IC 3 in a second embodiment.
  • saturation enhancement and contrast enhancement are concurrently achieved, not by a series of processes as illustrated in FIG. 1B .
  • Concurrently performing saturation enhancement and contrast enhancement is effective for reducing the circuit size.
  • an approximate gamma correction circuit 31 is used in place of the linear function correction circuit 22 , and a characterization data calculation circuit 32 and a correction point data calculation circuit 33 are additionally integrated in the driver IC 3 .
  • the approximate gamma correction circuit 31 performs correction calculations based on an approximate expression of the gamma correction on the input image data D IN to generate the output image data D OUT .
  • the input-output curves used in the correction calculations in the approximate gamma correction circuit 31 are each determined as a curve obtained by modifying a gamma curve specified by a certain gamma value so as to achieve contrast enhancement. It should be noted that, in the present embodiment, contrast enhancement is achieved to some degree by using a gamma curve as each input-output curve and additional contrast enhancement is achieved by modifying the shape of the gamma curve.
  • the shapes of the input-output curves used in the correction calculation by the approximate gamma correction circuit 31 are specified by correction point data sets CP_sel R , CP_sel G , and CP_sel B received from the correction point data calculation circuit 33 .
  • the shape of the input-output curve used in the correction calculation for the R data D IN R of the input image data D IN is specified by the correction point data set CP_sel R .
  • the shape of the input-output curve used in the correction calculation for the G data D IN G of the input image data D IN is specified by the correction point data set CP_sel G
  • the shape of the input-output curve used in the correction calculation for the B data D IN B of the input image data D IN is specified by the correction point data set CP_sel B .
  • the characterization data calculation circuit 32 generates characterization data indicating one or more feature quantities of each frame image (the image displayed in the display region 5 in each frame period) on the basis of the input image data D IN .
  • the characterization data include APL data D CHR _ APL indicating the average picture level (APL) of each frame image, and variance data D CHR _ ⁇ 2 indicating the variance of the luminance values of the pixels of each frame image. Note that the calculation of the APL data D CHR _ APL and variance data D CHR _ ⁇ 2 may be performed once for each frame period.
  • the correction point data calculation circuit 33 calculates the above-described correction point data sets CP_sel R , CP_sel G , and CP_sel B in response to the characterization data generated by the characterization data calculation circuit 32 (in the present embodiment, the APL data D CHR _ APL and variance data D CHR _ ⁇ 2 ) and the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B generated by the starting point control circuit 28 .
  • the correction point data calculation circuit 33 functions as a correction control circuitry which controls the shapes of the input-output curves in response to the characterization data (the APL data D CHR _ APL and variance data D CHR _ ⁇ 2 ) generated by the characterization data calculation circuit 32 and adjusts the positions of the starting points of the input-output curves in response to the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B .
  • the characterization data the APL data D CHR _ APL and variance data D CHR _ ⁇ 2
  • FIG. 10 illustrates the input-output curve specified by each correction point data set CP_sel k generated by the correction point data calculation circuit 33 and the contents of the correction calculation in accordance with the input-output curve.
  • Each correction point data set CP_sel k includes correction point data CP0 to CP5.
  • the correction point data CP0 to CP5 are each defined as data indicating a point in a coordinate system in which input image data D IN k are associated with the horizontal axis (or a first axis) and output image data D OUT k are associated with the vertical axis (or a second axis).
  • the correction point data CP0 indicate the position of the starting point of the input-output curve, which may be denoted also by symbol “CP0”, and the correction point data CP5 indicate the position of the end point of the input output curve, which may be denoted also by symbol “CP5”.
  • the correction point data CP2 and CP3 respectively indicate the positions of correction points which the input-output curve passes through near the center of the input-output curves, which may be denoted also by symbols “CP2” and “CP3”, respectively.
  • the correction point data CP1 indicate the position of a correction point located between the correction points CP0 and CP2, which is denoted also by symbol “CP1”, and the correction point data CP4 indicate the position of a correction point located between the correction points CP3 and CP5, which is also denoted by symbol “CP4”.
  • the shape of the gamma curve is specified by appropriately determining the positions of the correction points CP0 to CP0 indicated by the correction point data CP0 to CP5.
  • the shape of a gamma curve as being convex downward by determining the positions of the correction points CP1 to CP4 as being lower than the straight line connecting the both ends of the gamma curve.
  • the output image data D OUT k are generated by performing a gamma correction in accordance with the gamma curve with the shape specified by the correction point data CP0 to CP5 included in the correction point data set CP_sel k .
  • the correction point data CP0 to CP5 of the correction point data sets CP_sel R , CP_sel G and CP_sel B are determined in response to the characterization data (in the present embodiment, APL data D CHR _ APL and variance data D CHR _ ⁇ 2 ) generated by the characterization data calculation circuit 32 .
  • correction point data CP0 of the correction point data sets CP_sel R , CP_sel G and CP_sel B are adjusted in response to the starting point control data CP0_cont R , CP0_cont G and CP0_cont B , respectively.
  • FIG. 11 is a block diagram illustrating an example of the configuration of the approximate gamma correction circuit 31 .
  • the approximate gamma correction circuit 31 includes approximate gamma correction units 34 R, 34 G and 34 B, which are prepared for the R data D IN R , G data D IN G and B data D IN B of the input image data D IN , respectively.
  • the approximate gamma correction unit 34 R performs a correction calculation with an arithmetic expression on the R data D IN R of the input image data D IN to generate the R data D OUT R of the output image data D OUT .
  • the approximate gamma correction units 34 G and 34 B perform correction calculations with arithmetic expressions on the G data D IN G and B data D IN B of the input image data D IN to generate the G data D OUT G and B data D OUT B of the output image data D OUT , respectively.
  • the number of bits of the R data D OUT R , G data D OUT G and B data D OUT B of output image data D OUT is ten bits in the present embodiment; this means that the number of bits of the R data D OUT R , G data D OUT G and B data D OUT B of the output image data D OUT is larger than that of the R data D IN R , G data D IN G and B data D IN B of the input image data D IN .
  • Coefficients of the arithmetic expression used for the correction calculation by the approximate gamma correction unit 34 R are determined depending on the correction point data CP0 to CP5 of the correction point data set CP_sel R .
  • coefficients of the arithmetic expressions used for the correction calculation by the approximate gamma correction units 34 G and 34 B are determined depending on the correction point data CP0 to CP5 of the correction point data set CP_sel G and CP_sel B , respectively.
  • FIG. 12 is a block diagram illustrating an example of the configuration of the correction point data calculation circuit 33 .
  • the correction point data calculation circuit 33 includes: a correction point data set storage register 41 , an interpolation/selection circuit 42 and a correction point data adjustment circuit 43 and a CP0 adjustment circuit 44 .
  • the correction point data set storage register 41 stores therein a plurality of correction point data sets CP#1 to CP#m.
  • the correction point data sets CP#1 to CP#m are used as seed data for determining the above-described correction point data sets CP_sel R , CP_sel G and CP_sel B .
  • Each of the correction point data sets CP#1 to CP#m includes correction point data CP0 to CP5 defined as illustrated in FIG. 10 .
  • the interpolation/selection circuit 42 determines gamma values ⁇ _PIXEL R , ⁇ _PIXEL G and ⁇ _PIXEL B on the basis of the APL data D CHR _ APL received from the characterization data calculation circuit 32 and determines correction point data sets CP_L R , CP_L G and CP_L B corresponding to the gamma values ⁇ _PIXEL R , ⁇ _PIXEL G and ⁇ _PIXEL B thus determined.
  • the gamma value ⁇ _PIXEL R is the gamma value of the gamma curve used for the correction calculation performed on the R data D IN R of the input image data D IN .
  • the gamma value ⁇ _PIXEL G is the gamma value of the gamma curve used for the correction calculation performed on the G data D IN G of the input image data D IN
  • the gamma value ⁇ _PIXEL B is the gamma value of the gamma curve used for the correction calculation performed on the B data D IN B of the input image data D IN .
  • the interpolation/selection circuit 42 may select one of the correction point data sets CP#1 to CP#m on the basis of the gamma value ⁇ _PIXEL k and determine the correction point data set CP_L k as the selected one of the correction point data sets CP#1 to CP#m.
  • the interpolation/selection circuit 42 may determine the correction point data set CP_L k by selecting two of correction point data sets CP#1 to CP#m on the basis of the gamma value ⁇ _PIXEL k and applying an interpolation to the selected two correction point data sets. Details of the determination of the correction point data sets CP_L R , CP_L G and CP_L B are described later.
  • the correction point data sets CP_L R , CP_L G and CP_L B determined by the interpolation/selection circuit 42 are forwarded to the correction point data adjustment circuit 43 .
  • the correction point data adjustment circuit 43 modifies the correction point data sets CP_L R , CP_L G and CP_L B in response to the variance data D CHR _ ⁇ 2 received from the characterization data calculation circuit 32 .
  • This operation of the correction point data adjustment circuit 43 is technically equivalent to processing to modify the shape of the gamma curve specified by the gamma value ⁇ _PIXEL k in response to the variance data D CHR _ ⁇ 2 .
  • the correction point data adjustment circuit 43 modifies the correction point data CP1 and CP4 of the correction point data sets CP_L R , CP_L G and CP_L B in response to the variance data D CHR _ ⁇ 2 , to thereby achieve contrast enhancement.
  • the CP0 adjustment circuit 44 modifies the correction point data sets CP_L R , CP_L G and CP_L B in response to the starting point control data CP0_cont R , CP0_cont G and CP0_cont B received from the starting point control circuit 28 .
  • This operation of the CP0 adjustment circuit 4 is technically equivalent to processing to adjust the positions of the starting points of the input-output curves in response to the starting point control data CP0_cont R , CP0_cont G and CP0_cont B .
  • correction point data sets CP_L R , CP_L G and CP_L B thus modified by the correction point data adjustment circuit 43 and the CP0 adjustment circuit 44 are used as the correction point data sets CP_sel R , CP_sel G and CP_sel B , which are finally fed to the approximate gamma correction circuit 31 .
  • the correction point data sets CP_L R , CP_L G and CP_L B are first modified by the correction point data adjustment circuit 43 to generate correction point data sets CP_CR B , CP_CR G and CP_CR B and the correction point data sets CP_CR B , CP_CR G and CP_CR B are further modified by the CP0 adjustment circuit 44 to thereby generate the correction point data sets CP_sel R , CP_sel G and CP_sel B , which are finally fed to the approximate gamma correction circuit 31 .
  • the correction point data calculation circuit 33 may be configured so that the correction point data sets CP_L R , CP_L G and CP_L B are first modified by the CP0 adjustment circuit 44 and the modified correction point data sets CP_L R , CP_L G and CP_L B are further modified by the correction point data adjustment circuit 43 to generate the correction point data sets CP_sel R , CP_sel G and CP_sel B .
  • the correction point data adjustment circuit 43 and the CP0 adjustment circuit 44 may be combined to form a unified circuit.
  • the unified circuit may modify the correction point data sets CP_L R , CP_L G and CP_L B in response to the variance data D CHR _ ⁇ 2 and the starting point control data CP0_cont R , CP0_cont G and CP0_cont B . Details of the operations of the respective circuits in the correction point data calculation circuit 33 are described later.
  • FIG. 13 is a flowchart illustrating the procedure of the correction calculation performed on the input image data D IN in the second embodiment.
  • the input image data D IN are fed to the starting point control circuit 28 and the starting point control data CP0_cont R , CP0_cont G and CP0_cont B are generated by the starting point control circuit 28 on the basis of the input image data D IN at steps 11 to 14 .
  • the starting point control data CP0_cont R , CP0_cont G and CP0_cont B are generated through the same processing as in the first embodiment.
  • the input image data D IN are also fed to the characterization data calculation circuit 32 and the APL data D CHR _ APL and variance data D CHR _ ⁇ 2 of each frame image are generated by the characterization data calculation circuit 32 from the input image data D IN .
  • the APL data D CHR _ APL indicate the APL of each frame image
  • the variance data D CHR _ ⁇ 2 indicate the variance of the luminance values of the pixels in each frame period.
  • the gamma values to be used for the correction calculation of the input image data D IN associated with each pixel 9 in each frame image are calculated from the APL data D CHR _ APL of each frame image.
  • a gamma value is calculated for each of the R data D IN R , G data D IN G and B data D IN B of the input image data D IN associated with each pixel 9 .
  • ⁇ _STD R , ⁇ _STD G , and ⁇ _STD B may be equal to each other, or different, and ⁇ R , ⁇ G and ⁇ B may be equal to each other, or different. It should be noted that the gamma values ⁇ _PIXEL R , ⁇ _PIXEL G , and ⁇ _PIXEL B are calculated for each frame image.
  • the correction point data sets CP_L R , CP_L G , and CP_L B are selected or determined on the basis of the thus-calculated gamma values ⁇ _PIXEL R , ⁇ _PIXEL G and ⁇ _PIXEL B , respectively.
  • the correction point data sets CP_L R , CP_L G and CP_L B are seed data used for calculating the correction point data sets CP_sel R , CP_sel G , and CP_sel B , which are finally fed to the approximate gamma correction circuit 31 .
  • the correction point data sets CP_L R , CP_L G and CP_L B are determined for each frame image.
  • the correction point data sets CP_L R , CP_L G , and CP_L B are each selected from the correction point data sets CP#1 to CP#m stored in the correction point data set storage register 41 of the correction point data calculation circuit 33 .
  • the correction point data sets CP#1 to CP#m correspond to different gamma values ⁇ and each of the correction point data sets CP#1 to CP#m includes correction point data CP0 to CP5.
  • the correction point data CP0 to CP5 of a correction point data set CP#j corresponding to a certain gamma value ⁇ are determined as follows:
  • D OUT MAX is the allowed maximum value of the R data D OUT R , G data D OUT G and B data D OUT B of the output image data D OUT ; D OUT MAX depends on the number of bits of the R data D OUT R , G data D OUT G and B data D OUT B .
  • the function Gamma [x] D OUT MAX ⁇ ( x/D IN MAX ) ⁇ (13b)
  • the correction point data sets CP#1 to CP#m are determined so that the gamma value ⁇ recited in expression (13b) to which a correction point data set CP#j selected from the correction point data sets CP#1 to CP#m corresponds is increased as j is increased.
  • ⁇ 1 ⁇ 2 ⁇ . . . ⁇ m-1 ⁇ m (14) where ⁇ j is the gamma value corresponding to the correction point data set CP#j.
  • the correction point data set CP_L R is selected from the correction point data sets CP#1 to CP#m on the basis of the gamma value ⁇ _PIXEL R .
  • the correction point data set CP_L R is determined as a correction point data set CP#j with a larger value of j as the gamma value ⁇ _PIXEL R increases.
  • the correction point data sets CP_L G and CP_L B are selected from the correction point data sets CP#1 to CP#m on the basis of the gamma values ⁇ _PIXEL G and ⁇ _PIXEL B , respectively.
  • FIG. 14 is a graph illustrating the relation among the average picture level (APL) of each frame image, ⁇ _PIXEL k and the correction point data set CP_L k in the case when the correction point data set CP_L k is determined in this manner.
  • APL average picture level
  • the correction point data sets CP_L R , CP_L G , and CP_L B may be determined as follows:
  • the correction point data sets CP#1 to CP#m are stored in the correction point data set storage register 41 of the correction point data calculation circuit 33 .
  • the correction point data sets CP#1 to CP#m to be stored in the correction point data set storage register 41 may be fed from the processor 4 to the drive IC 3 as initial settings.
  • two correction point data sets CP#q and CP#(q+1) are selected on the basis of the gamma value ⁇ _PIXEL k (k is any one of “R”, “G” and “B”) from among the correction point data sets CP#1 to CP#m stored in the correction point data set storage register 41 for determining the correction point data set CP_L k , where q is an integer from one to m ⁇ 1.
  • the two correction point data sets CP#q and CP#(q+1) are selected to satisfy the following expression: ⁇ q ⁇ _PIXEL k ⁇ q+1 .
  • correction point data CP0 to CP5 of the correction point data set CP_L k are respectively calculated with an interpolation of correction point data CP0 to CP5 of the selected two correction point data sets CP#q and CP#(q+1).
  • FIG. 15 is a graph illustrating the relation among the average picture level (APL), ⁇ _PIXEL k and the correction point data set CP_L k in the case when the correction point data set CP_L k is determined in this manner.
  • APL average picture level
  • ⁇ _PIXEL k As the average picture level increases, the gamma value ⁇ _PIXEL k is increased and correction point data sets CP#q and CP#(q+1) with a larger value of q are selected.
  • the correction point data set CP_L k is determined to correspond to a gamma value in a range from the gamma value ⁇ q to ⁇ q+1 , which the correction point data sets CP#q and CP#(q+1) respectively correspond to.
  • FIG. 16 is a graph schematically illustrating the shapes of the gamma curves corresponding to the correction point data sets CP#q and CP#(q+1) and the correction point data set CP_L k . Since the correction point data CP ⁇ of the correction point data set CP_L k is obtained through the interpolation of the correction point data CP ⁇ (#q) and CP ⁇ (#(q+1)) of the correction point data sets CP#q and CP#(q+1), the shape of the gamma curve corresponding to the correction point data set CP_L k is determined so that the gamma curve corresponding to the correction point data set CP_L k is located between the gamma curves corresponding to the correction point data sets CP#q and CP#(q+1).
  • the calculation of the correction point data CP0 to CP5 of the correction point data set CP_L k through the interpolation of the correction point data CP0 to CP5 of the correction point data sets CP#q and CP#(q+1) is advantageous for allowing finely adjusting the gamma value used for the gamma correction even when only a reduced number of the correction point data sets CP#1 to CP#m are stored in the correction point data set storage register 41 .
  • a common correction point data set CP_L is selected or calculated from the gamma value ⁇ _PIXEL in a similar way.
  • step S 24 the correction point data sets CP_L R , CP_L G , and CP_L B thus selected or calculated are modified in response to the variance data D CHR _ ⁇ 2 to generate correction point data sets CP_CR R , CP_CR G and CP_CR B .
  • correction point data sets CP_L R , CP_L G and CP_L B each represent the shape of a specific gamma curve
  • modifying the correction point data sets CP_L R , CP_L G and CP_L B in response to the variation data D CHR _ ⁇ 2 is technically equivalent to modifying the shapes of the gamma curves used for the correction calculations in response to the variation data D CHR _ ⁇ 2 .
  • FIG. 17 is a conceptual diagram illustrating the technical significance of modification of the correction point data set CP_L k on the basis of variance data D CHR _ ⁇ 2 .
  • a reduced variance ⁇ 2 of the luminance values of a certain frame image means that the frame image includes an increased number of pixels with luminance values close to the APL of the frame image.
  • the contrast of the frame image corresponding to the input image data D IN is small.
  • the contrast of the frame image corresponding to the input image data D IN is small, it is possible to display the frame image with an improved image quality by performing a correction calculation to enhance the contrast by the approximate gamma correction circuit 31 .
  • the correction point data CP1 and CP4 of the correction point data set CP_L k are adjusted on the basis of the variance data D CHR _ ⁇ 2 in the present embodiment.
  • the correction point data CP1 of the correction point data set CP_L k is modified so that the correction point data CP1 of the correction point data sets CP_CR k , which is obtained as the result of the modification, is decreased as the variance ⁇ 2 indicated by the variance data D CHR _ ⁇ 2 decreases.
  • the correction point data CP4 of the correction point data set CP_L k is, on the other hand, modified so that the correction point data CP4 of the correction point data sets CP_CR k is increased as the variance ⁇ 2 indicated by the variance data D CHR _ ⁇ 2 decreases.
  • Such modification results in that the correction calculation in the approximate gamma correction circuit 31 is performed to enhance the contrast, when the contrast of the image corresponding to the input image data D IN is small.
  • the correction point data CP0, CP2, CP3 and CP5 of the correction point data set CP_L k are not modified in the present embodiment.
  • the correction point data set CP_CR R , CP_CR G , and CP_CR B are modified in response to the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B to generate the correction point data sets CP_sel R , CP_sel G and CP_sel B .
  • the starting points of the input-output curves that is, the correction point data CP0 of the correction point data sets CP_CR R , CP_CR G and CP_CR B are modified in response to the starting point control data CP0_cont R , CP0_cont G and CP0_cont B .
  • modifying the correction point data CP0 of the correction point data sets CP_CR R , CP_CR G , and CP_CR B is technically equivalent to adjusting the positions of the starting points of the gamma curves associated with the correction point data sets CP_CR R , CP_CR G and CP_CR B .
  • the correction point data CP1 to CP4 of the correction point data sets CP_CR R , CP_CR G , and CP_CR B are also modified in accordance with the modification of the correction point data CP0, that is, the adjustment of the positions of the input-output curves. Since the correction point data CP1 to CP4 of the correction point data sets CP_CR R , CP_CR G , and CP_CR B indicates the shapes of the intermediate portions of the input-output curves, modifying the correction point data CP1 to CP4 of the correction point data sets CP_CR R , CP_CR G and CP_CR B is technically equivalent to modifying the shapes of the intermediate portions of the input-output curves.
  • correction point data corresponding to correction points close to the starting points of the input-output curves are subjected to a relatively large modification and correction point data corresponding to correction points close to the end points of the input-output curves (for example, correction point data CP4) are subjected to a relatively small modification.
  • correction point control data CP1_cont k to CP4_cont k which indicate the amounts of modifications of the correction point data CP1 to CP4 of the correction point data set CP_CR k , are calculated in accordance with the following expressions (17a) to (17d):
  • FIG. 18 schematically illustrates the correction point control data CP1_cont k to CP4_cont k thus calculated.
  • symbols “CP5 R ”, “CP5 G ” and “CP5 B ” indicate the end points of the input-output curves used for the correction calculations of the R data D IN R , G data D IN G , and B data D IN B , respectively.
  • the correction point data set CP_sel k is calculated by adding the values of the above-described correction point control data CP1_cont k to CP4_cont k to the values of the correction point data CP1 to CP4 of the correction point data sets CP_CR k .
  • ⁇ R , ⁇ G , and ⁇ B are predetermined proportionality constants. ⁇ R , ⁇ G , and ⁇ B may be equal to each other or different.
  • CP ⁇ _sel k is the value of the correction point data CP ⁇ of the correction point data set CP_sel k and CP ⁇ _L k is the value of the correction point data CP ⁇ of the correction point data set CP_L k .
  • the correction point data CP5 of the correction point data set CP_sel k is determined as equal to the correction point data CP5 of the correction point data set CP_L k .
  • the correction point data sets CP_sel R , CP_sel G , and CP_sel B are also calculated for each pixel 9 .
  • correction point data CP_sel k associated with a certain pixel 9 in a certain frame image are calculated from the correction point data set CP_L k calculated for the frame image, the variance ⁇ 2 of the luminance values described in the variance data D CHR _ ⁇ 2 calculated for the frame image, and the starting point control data CP0_cont R , CP0_cont G and CP0_cont B calculated for the pixel 9 .
  • the correction point data CP5 of the correction point data set CP_sel k are equal to the correction point data CP5 of the correction point data set CP_L k , and expression (23) holds.
  • correction calculations are performed on the R data D IN R , G data D IN G and B data D IN B of the input image data D IN associated with each pixel 9 on the basis of the thus-calculated correction point data sets CP_sel R , CP_sel G and CP_sel B associated with each pixel 9 , respectively, to thereby generate the R data D OUT R , G data D OUT G and B data D OUT B of the output image data D OUT associated with each pixel 9 .
  • the correction calculations are achieved by the approximate gamma correction units 34 R, 34 G and 34 B of the approximate gamma correction circuit 31 , respectively.
  • the R data D OUT R , G data D OUT G and B data D OUT B of the output image data D OUT are calculated from the R data D IN R , G data D IN G and B data D IN B of the input image data D IN in accordance with the following expressions:
  • CP0 to CP5 recited in expressions (29a) to (29c) indicate the correction point data CP0 to CP5 of the correction point data set CP_sel k .
  • D IN Center D IN MAX /2
  • K the parameter given by the above-described expression (13a).
  • D INS , PD INS , and ND INS recited in expressions (29a) to (29c) are values defined as follows: (a) D INS
  • D INS is a value which depends on the input image data D IN k ;
  • D INS is given by the following expressions (30a) and (30b):
  • D INS D IN k (for D IN k ⁇ D IN Center )
  • D INS D IN k +1 ⁇ K (for D IN k >D IN Center )
  • PD INS PD INS
  • ND INS ( K ⁇ D INS ) ⁇ D INS . (32)
  • ND INS is a value calculated by an expression including a term proportional to a square of D IN k .
  • FIG. 19 illustrates the relations between the R data D IN R , G data D IN G , and B data D IN B of the input image data D IN and the R data D OUT R , G data D OUT G , and B data D OUT B of the output image data D OUT , when the R data D OUT R , G data D OUT G , and B data D OUT B of the output image data D OUT are calculated as described above.
  • at least one of the starting point control data CP0_cont R , CP0_cont G , and CP0_cont B is positive and at least another one is negative. In view of this fact, it would be easily understood from FIG. 19 that the above-described processing effectively enhances the saturation.
  • the output image data D OUT which are calculated by the approximate gamma correction circuit 31 with the above-described series of expressions, are forwarded to the color reduction circuit 23 .
  • the color reduction circuit 23 performs a color reduction on the output image data D OUT to generate the color-reduced image data D OUT _ D .
  • the color-reduced image data D OUT _ D are forwarded to the data line drive circuit 26 via the latch circuit 24 and the data lines 8 of the LCD panel 2 are driven in response to the color-reduced image data D OUT _ D .
  • saturation enhancement processing of this embodiment effectively achieves a saturation enhancement with simple processing in which the positions of the input-output curves are adjusted.
  • saturation enhancement and contrast enhancement are concurrently achieved in the approximate gamma correction circuit 31 and this effectively reduces the circuit size (for example, compared to the system in which contrast enhancement and saturation enhancement are performed in series as illustrated in FIG. 1B ).
  • the present invention is not limited to the above-described embodiments; a person skilled in the art would appreciate that the present invention may be implemented with various modifications.
  • the above-described embodiments recite the liquid crystal display device 1 including the LCD panel 2
  • the saturation enhancement and contrast enhancement recited in the above-descried embodiments may be implemented in an image processing apparatus in general.
  • the present invention is applicable to various panel display devices including different display panels (for example, a display device including an OLED (organic light emitting diode) display panel).

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Hardware Design (AREA)
  • Control Of Indicators Other Than Cathode Ray Tubes (AREA)
  • Liquid Crystal Display Device Control (AREA)
  • Liquid Crystal (AREA)

Abstract

A display panel driver includes: a correction calculation section which performs correction calculations on input image data to generate saturation-enhanced output image data and a drive circuitry driving the display panel in response to the output image data and a starting point control section. The correction calculation section generates red (R) data, green (G) data and blue (B) data of the output image data by performing the correction calculations on R data, G data and B data of the input image data, respectively. The starting point control section controls the positions of starting points of the input-output curves of the correction calculations.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
This application is a continuation of co-pending U.S. application Ser. No. 14/617,674, filed Feb. 9, 2015, which claims priority to Japanese Patent Application No. 2014-023879, filed on Feb. 10, 2014. Each of the related applications is incorporated herein by reference.
TECHNICAL FIELD
The present invention relates to a display device, a display panel driver, an image processing apparatus and an image processing method, in particular, to process image data for saturation enhancement of display images in a display device, a display panel driver, an image processing apparatus.
BACKGROUND ART
The saturation enhancement processing is one of the known image processing techniques applied to image data. In a display device which originally has a relatively narrow color gamut, for example, performing saturation enhancement processing on image data effectively compensates the color gamut in images actually displayed on the display screen. More specifically, a liquid crystal display device which uses a white backlight undesirably has a narrow color gamut compared to recent OLED (organic light emitting diode) display devices, and therefore an enlargement of the color gamut is often required to achieve beautiful coloring. Saturation enhancement processing to image data allows enlarging the effective color gamut to meet such a requirement.
Saturation enhancement processing is also used to recover the saturation, when the saturation of an image is deteriorated by an auto contrast optimization (ACO). In a typical auto contrast optimization, a contrast enhancement is achieved in response to characterization data obtained by analyzing image data (for example, the luminance histogram or the average picture level (APL) of the image); note that a contrast enhancement is disclosed in Japanese patent No. 4,198,720 B, for example. In a typical auto contrast optimization, however, the saturation, that is, the differences among the grayscale levels of the red color (R), the green color (G) and the blue color (B) may be reduced, because a common contrast enhancement is performed for the red, green and blue colors, as is understood from FIG. 1A. To address this problem, saturation enhancement processing is often performed on image data obtained by a contrast enhancement.
According to a study of the inventors, however, there is a room for improvement in known saturation enhancement techniques to achieve appropriate saturation enhancement with a reduced circuit size.
Such a situation is especially severe when different image processing (such as contrast enhancement) is performed in series with saturation enhancement processing. FIG. 1B illustrates an example of a system in which saturation enhancement processing is performed after contrast enhancement processing. In order to achieve an improved contrast enhancement, output image data obtained by the contrast enhancement processing needs to have a wider bit width than that of input image data, to avoid gradation collapse in the contrast enhancement processing. When input image data to be subjected to contrast enhancement processing represent the grayscale level of each of the red, green and blue colors with eight bits, for example, the output image data of the contrast enhancement processing may be generated as image data which represent the grayscale level of each of the red, green and blue colors with 10 bits. When the output image data of the contrast enhancement processing are further subjected to saturation enhancement processing, the image data obtained by the saturation enhancement processing need to have a further wider bit width. When the output image data of the contrast enhancement processing represent the grayscale level of each of the red, green and blue colors with 10 bits, for example, the output image data of the saturation enhancement processing may be generated as image data which represent the grayscale level of each of the red, green and blue colors with 12 bits. The increase in the bit widths of the input and output image data of the saturation enhancement processing, however, undesirably increases the size of the circuit used for the saturation enhancement processing.
As a technique potentially related to the present invention, Japanese Patent Application Publication No. 2010-79119 A discloses a technique in which red-green-blue (RGB) data are transformed into hue-saturation-value (HSV) data and the saturation enhancement is achieved in the HSV color space. Japanese Patent Application Publication No. H06-339017 A discloses a saturation enhancement in which the red (R), green (G), blue (B) values of a saturation-enhanced image are respectively calculated by subtracting products of an enhancement coefficient and the differences between I and the original R, G, and B values from I, where I is the maximum value of the R, G, and B values of each pixel.
SUMMARY
Therefore, embodiments of the present invention provide an apparatus and method for image processing which achieve saturation enhancement with a reduce circuit size, and a display panel driver and display device using the same.
Other objectives and new features of the present invention would be understood from the disclosure in the Specification and attached drawings.
In an aspect of the present invention, a display device includes: a display panel and a display panel driver driving the display panel. The display panel driver includes: a correction calculation section which performs correction calculations on input image data to generate saturation-enhanced output image data; a drive circuitry driving the display panel in response to the output image data; and a starting point control section. The correction calculation section generates R data of the output image data by performing a first correction calculation on R data of the input image data, generates G data of the output image data by performing a second correction calculation on G data of the input image data, and generates B data of the output image data by performing a third correction calculation on B data of the input image data. The starting point control section controls a position of a starting point of a first input-output curve corresponding to an input-output relation of the first correction calculation; a position of a starting point of a second input-output curve corresponding to an input-output relation of the second correction calculation; and a position of a starting point of a third input-output curve corresponding to an input-output relation of the third correction calculation.
In another aspect of the present invention, a display panel driver for driving a display panel includes: a correction calculation section which performs correction calculations on input image data to generate saturation-enhanced output image data, a drive circuitry driving the display panel in response to the output image data and a starting point control section. The correction calculation section generates R data of the output image data by performing a first correction calculation on R data of the input image data, generates G data of the output image data by performing a second correction calculation on G data of the input image data, and generates B data of the output image data by performing a third correction calculation on B data of the input image data. The starting point control section controls a position of a starting point of a first input-output curve corresponding to an input-output relation of the first correction calculation; a position of a starting point of a second input-output curve corresponding to an input-output relation of the second correction calculation; and a position of a starting point of a third input-output curve corresponding to an input-output relation of the third correction calculation.
In still another aspect of the present invention, an image processing apparatus includes: a correction calculation section which performs correction calculations on input image data to generate saturation-enhanced output image data; and a starting point control section. The correction calculation section generates R data of the output image data by performing a first correction calculation on R data of the input image data, generates G data of the output image data by performing a second correction calculation on G data of the input image data, and generates B data of the output image data by performing a third correction calculation on B data of the input image data. The starting point control section controls a position of a starting point of a first input-output curve corresponding to an input-output relation of the first correction calculation; a position of a starting point of a second input-output curve corresponding to an input-output relation of the second correction calculation; and a position of a starting point of a third input-output curve corresponding to an input-output relation of the third correction calculation.
In still another aspect of the present invention, an image processing method includes: generating saturation-enhanced output image data by performing correction calculations on input image data. The step of generating the output image data includes: generating R data of the output image data by performing a first correction calculation on R data of the input image data; generating G data of the output image data by performing a second correction calculation on G data of the input image data; generating B data of the output image data by performing a third correction calculation on B data of the input image data; and controlling a position of a starting point of a first input-output curve corresponding to an input-output relation of the first correction calculation; a position of a starting point of a second input-output curve corresponding to an input-output relation of the second correction calculation; and a position of a starting point of a third input-output curve corresponding to an input-output relation of the third correction calculation.
The present invention provides an apparatus and method for image processing which achieve saturation enhancement with a reduce circuit size.
BRIEF DESCRIPTION OF THE DRAWINGS
The above and other advantages and features of the present invention will be more apparent from the following description taken in conjunction with the accompanied drawings, in which:
FIG. 1A illustrates a reduction in saturation caused by a contrast enhancement;
FIG. 1B illustrates an example of a system which performs a contrast enhancement and saturation enhancement in series;
FIG. 2 is a conceptual diagram illustrating saturation enhancement processing in one embodiment of the present invention;
FIG. 3 is a block diagram illustrating an exemplary configuration of a display device in a first embodiment of the present invention;
FIG. 4 is a circuit diagram schematically illustrating the configuration of each subpixel;
FIG. 5 is a block diagram illustrating an exemplary configuration of a driver IC in the first embodiment;
FIG. 6 is a flowchart illustrating an exemplary procedure of the saturation enhancement processing of the first embodiment;
FIG. 7 is a graph illustrating an example of a function used to determine an enhancement coefficient in the first embodiment;
FIG. 8 is a graph illustrating the relation between R data, G data and B data of input image data and R data, G data and B data of output image data in the first embodiment;
FIG. 9 is a block diagram illustrating an exemplary configuration of a driver IC in a second embodiment of the present invention;
FIG. 10 illustrates a gamma curve specified by each correction point data set and contents of a correction calculation (or gamma correction) in accordance with the gamma curve;
FIG. 11 is a block diagram illustrating an example of the configuration of an approximate gamma correction circuit in the second embodiment;
FIG. 12 is a block diagram illustrating an example of the configuration of a correction point data calculation circuit in the second embodiment;
FIG. 13 is a flowchart illustrating the procedure of contrast enhancement processing and saturation enhancement processing in the second embodiment;
FIG. 14 is a graph illustrating the relation among APL, γ_PIXELk and a correction point data set CP_Lk in one embodiment;
FIG. 15 a graph illustrating the relation among APL, γ_PIXELk and a correction point data set CP_Lk in another embodiment;
FIG. 16 is a graph schematically illustrating the shapes of gamma curves respectively corresponding to correction point data sets CP#q and CP#(q+1) and the gamma curve corresponding to the correction point data set CP_Lk;
FIG. 17 is a conceptual diagram illustrating a technical significance of modification of the correction point data set CP_Lk on the basis of variance data DCHR _ σ2;
FIG. 18 illustrates correction point control data CP1_contk to CP4_contk; and
FIG. 19 is a graph illustrating the relation between R data DIN R, G data DIN G and B data DIN B of input image data DIN and R data DOUT R, G data DOUT G and B data DOUT B of output image data DOUT in the second embodiment.
DESCRIPTION OF EMBODIMENTS
The invention will be now described herein with reference to illustrative embodiments. Those skilled in the art would recognize that many alternative embodiments can be accomplished using the teachings of the present invention and that the invention is not limited to the embodiments illustrated for explanatory purposed. It should be noted that same, similar or corresponding elements are denoted by the same or similar reference numerals in the following description.
FIG. 2 is a conceptual diagram illustrating saturation enhancement processing in one embodiment of the present invention. In the saturation enhancement processing in the present embodiment, saturation-enhanced output image data DOUT are generated by performing correction calculations on input image data DIN. Here, the input image data DIN include R data DIN R (data indicating a grayscale level of the red color), G data DIN G (data indicating a grayscale level of the green color) and B data DIN B (data indicating a grayscale level of the blue color). Correspondingly, the output image data DOUT include R data DOUT R, G data DOUT G and B data DOUT B.
In the present embodiment, a saturation enhancement is achieved by individually adjusting the position of the starting point of a curve indicating the input-output relation of the correction calculation (that is, the relation between the value of input image data and that of output image data) for each color. In the following, a curve indicating the input-output relation of a correction calculation may be referred to as “input-output curve”. Note that a “starting point” means the point on an input-output curve for input image data with the allowed minimum value. The term “input-output curve” means to include the case when the input-output relation is linear, that is, the case when the input-output relation is represented with a straight line.
In detail, in the saturation enhancement processing in the present embodiment, R data DOUT R of output image data DOUT are generated by performing a first correction calculation on R data DIN R of input image data DIN. Correspondingly, G data DOUT G of the output image data DOUT are generated by performing a second correction calculation on G data DIN G of the input image data DIN, and B data DOUT B of the output image data DOUT are generated by performing a third correction calculation on B data DIN B of the input image data DIN. In FIG. 2, the left graph illustrates the input-output curve of the first correction calculation, the center graph illustrates the input-output curve of the second correction calculation, and the right graph illustrates the input-output curve of the third correction calculation.
In the present embodiment, the saturation enhancement is achieved by controlling the position of the starting point CP0R of the input-output curve corresponding to the input-output relation of the first correction calculation, the position of the starting point CP0G of the input-output curve corresponding to the input-output relation of the second correction calculation, and the position of the starting point CP0B of the input-output curve corresponding to the input-output relation of the third correction calculation.
Such saturation enhancement processing allows enhancing the saturation with a simple calculation. This effectively reduces the size of the circuit used to perform the saturation enhancement processing. In addition, the saturation enhancement processing of the present embodiment is easy to be combined with other image processing techniques, such as contrast enhancement. For example, a saturation enhancement is achieved by adjusting the position of the starting point of the input-output curve, while a contrast enhancement is achieved by determining the overall shape of the input-output curve. This feature is also advantageous for reducing the size of the circuit used to perform the saturation enhancement processing.
In one embodiment, in saturation enhancement processing with respect to input image data DIN associated with a certain pixel, the position of the starting point CP0R of the input-output curve of the correction calculation performed on the R data DIN R of the input image data DIN is determined in response to the difference DIN R−YPIXEL, where YPIXEL is the luminance value of the certain pixel calculated from the input image data DIN. Correspondingly, the position of the starting point CP0G of the input-output curve of the correction calculation performed on the G data DIN G of the input image data DIN is determined in response to the difference DIN G−YPIXEL, and the position of the starting point CP0B of the input-output curve of the correction calculation performed on the B data DIN B of the input image data DIN is determined in response to the difference DIN B−YPIXEL.
Since the luminance value YPIXEL is a value determined as a weighted average of the R data DIN R, G data DIN G and B data DIN B, at least one of the differences DIN R−YPIXEL, DIN G−YPIXEL and DIN B−YPIXEL is positive and at least another one is negative. Accordingly, the saturation can be enhanced by determining the positions of the starting points CP0R, CP0G and CP0B on the basis of the differences DIN R−YPIXEL, DIN G−YPIXEL and DIN B−YPIXEL. FIG. 2 illustrates saturation enhancement processing in the case when the difference DIN R−YPIXEL is positive and the differences DIN G−YPIXEL and DIN B−YPIXEL are negative.
In one embodiment, in saturation enhancement processing for an image displayed in a certain frame period (or vertical synchronization period), the starting points CP0R, CP0G and CP0B may be determined in response to the average saturation of the frame image displayed in the certain frame period. This effectively allows an improved saturation enhancement suitable for the average saturation of the frame image.
In one embodiment, the position of the starting point CP0R is determined on the basis of the product of the difference DIN R−YPIXEL and an enhancement coefficient INST determined as described below, the position of the starting point CP0G is determined on the basis of the product of the difference DIN G−YPIXEL and the enhancement coefficient INST, and the position of the starting point CP0B is determined on the basis of the product of the difference DIN B−YPIXEL and the enhancement coefficient INST. The enhancement coefficient INST is determined as the minimum value of enhancement coefficients INSTR, INSTG and INSTB, where the enhancement coefficients INSTR, INSTG and INSTB are obtained from the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN, respectively. The enhancement coefficients INSTR, INSTG, and INSTB are calculated with a predetermined function f(x) as follows:
INSTR =f(D IN R),
INSTG =f(D IN G), and
INSTB =f(D IN B),
where f(x) is a function satisfying the following conditions:
(a) f(x) takes the maximum value when x is β;
(b) f(x) monotonically increases as x increases when x is less than β; and
(c) f(x) monotonically decreases as x increases when x is more than β.
β is determined as DIN MAX/2 or the integer closest to DIN MAX/2 (if there are two integers closest to DIN MAX/2, one selected from the two closest integers), where DIN MAX is the allowed maximum value of the R data DIN R, G data DIN G, and B data DIN B. When the R data DIN R, G data DIN G, and B data DIN B are eight bit data, for example, the allowed maximum value DIN MAX is “255” and β is “127” or “128”.
According to such saturation enhancement processing, the saturation is enhanced more strongly when the values of the R data DIN R, G data DIN G, and B data DIN B (that is, the grayscale values of the red, green and blue colors) are about a half of the allowed maximum value DIN MAX and this effectively achieves improved saturation enhancement.
Also in this case, the positions of the starting points CP0R, CP0G and CP0B may be controlled in response to the average saturation SAVE in the frame image. More specifically, the position of the starting point CP0R is controlled in response to the product of the enhancement coefficient INST, the difference DIN MAX−SAVE and the difference DIN R−YPIXEL. Correspondingly, the position of the starting point CP0G may be controlled in response to the product of the enhancement coefficient INST, the difference DIN MAX−SAVE and the difference DIN G−YPIXEL, and the position of the starting point CP0B may be controlled in response to the product of the enhancement coefficient INST, the difference DIN MAX−SAVE, and the difference DIN B−YPIXEL.
In the following, a description is given of more specific embodiments of the present invention.
First Embodiment
FIG. 3 is a block diagram illustrating an exemplary configuration of a display device in a first embodiment of the present invention. The display device of the present embodiment, which is configured as a liquid crystal display device denoted by numeral 1, includes an LCD (liquid crystal display) panel 2 and a driver IC (integrated circuit) 3.
The LCD panel 2 includes a display region 5 and a gate line drive circuit 6 (also referred to as GIP (gate in panel) circuit). Disposed in the display region 5 are a plurality of gate lines 7 (also referred to as scan lines or address lines), a plurality of data lines 8 (also referred to as signal lines or source lines) and pixels 9. In the present embodiment, the number of the gate lines 7 is v and the number of the data lines 8 is 3h; the pixels 9 are arranged in v rows and h columns in the display region 5, where v and h are integers equal to or more than two. In the following, the horizontal direction of the display region 5 (that is, the direction in which the gate lines 7 are extended) may be referred to as X-axis direction and the vertical direction of the display region 5 (that is, the direction in which the data lines 8 are extended) may be referred to as Y-axis direction.
In the present embodiment, each pixel 9 includes three subpixels: an R subpixel 11R, a G subpixel 11G and a B subpixel 11B, where the R subpixel 11R is a subpixel corresponding to the red color (that is, a subpixel displaying the red color), the G subpixel 11G is a subpixel corresponding to the green color (that is, a subpixel displaying the green color) and the B subpixel 11B is a subpixel corresponding to the blue color (that is, a subpixel displaying the blue color). Note that the R subpixel 11R, G subpixel 11G and B subpixel 11B may be collectively referred to as subpixel 11 if not distinguished from each other. In the present embodiment, subpixels 11 are arrayed in v rows and 3h columns on the LCD panel 2. Each subpixel 11 is connected with one corresponding gate line 7 and one corresponding data line 8. In driving respective subpixels 11 of the LCD panel 2, gate lines 7 are sequentially selected and desired drive voltages are written into the subpixels 11 connected with a selected gate line 7 via the data lines 8. This allows setting the respective subpixels 11 to desired grayscale levels to thereby display a desired image in the display region 5 of the LCD panel 2.
FIG. 4 is a circuit diagram schematically illustrating the configuration of each subpixel 11. Each subpixel 11 includes a TFT (thin film transistor) 12 and a pixel electrode 13. The TFT 12 has a gate connected with a gate line 7, a source connected with a data line 8 and a drain connected with the pixel electrode 13. The pixel electrode 13 is disposed opposed to the opposing electrode (common electrode) 14 of the LCD panel 2 and the space between each pixel electrode 13 and the opposing electrode 14 is filled with liquid crystal. Although FIG. 4 illustrates the subpixel 11 as if the opposing electrode 14 may be separately disposed for each subpixel 11, a person skilled in the art would appreciate that the opposing electrode 14 is actually shared by the subpixels 11 of the entire LCD panel 2.
Referring back to FIG. 3, the driver IC 3 drives the data lines 8 and also generates gate line control signals SGIP for controlling the gate line drive circuit 6. The drive of the data lines 8 is responsive to input image data DIN and synchronization data DSYNC received from a processor 4 (for example, a CPU (central processing unit)). It should be noted here that the input image data DIN are image data corresponding to images to be displayed in the display region 5 of the LCD panel 2, more specifically, data indicating the grayscale levels of each subpixel 11 of each pixel 9. In the present embodiment, the input image data DIN represent the grayscale level of each subpixel 11 of each pixel 9 with eight bits. In other words, the input image data DIN represent the grayscale levels of each pixel 9 of the LCD panel 2 with 24 bits. The synchronization data DSYNC are used to control the operation timing of the driver IC 3; the generation timing of various timing control signals in the driver IC 3, including the vertical synchronization signal VSYNC and the horizontal synchronization signal HSYNC, is controlled in response to the synchronization data DSYNC. Also, the gate line control signals SGIP are generated in response to the synchronization data DSYNC. The driver IC 3 is mounted on the LCD panel 2 with a surface mounting technology such as a COG (chip on glass) technology.
FIG. 5 is a block diagram illustrating an example of the configuration of the driver IC 3. The driver IC 3 includes an interface circuit 21, a linear function correction circuit 22, a color reduction circuit 23, a latch circuit 24, a grayscale voltage generator circuit 25, a data line drive circuit 26, a timing control circuit 27, and a starting point control circuit 28.
The interface circuit 21 receives the input image data DIN and synchronization data DSYNC from the processor 4 and forwards the input image data DIN to the linear function correction circuit 22 and the synchronization data DSYNC to the timing control circuit 27.
The linear function correction circuit 22 performs saturation enhancement processing as described above; the linear function correction circuit 22 generates output image data DOUT by performing saturation enhancement processing on the input image data DIN.
In the following, data indicating the grayscale level of an R subpixel 11R of input image data DIN may be referred to as input image data DIN R. Correspondingly, data indicating the grayscale level of a G subpixel 11G of input image data DIN may be referred to as input image data DIN G, and data indicating the grayscale level of a B subpixel 11B of input image data DIN may be referred to as input image data DIN B. Similarly, data indicating the grayscale level of an R subpixel 11R of the output image data DOUT may be referred to as output image data DOUT R. Correspondingly, data indicating the grayscale level of a G subpixel 11G of the output image data DOUT may be referred to as output image data DOUT G, and data indicating the grayscale level of a B subpixel 11B of the output image data DOUT may be referred to as output image data DOUT B.
In the present embodiment, straight lines are used as the input-output curves of the saturation enhancement processing performed in the linear function correction circuit 22 (that is, the curves indicating the input-output relation between the input image data DIN inputted to the linear function correction circuit 22 and the output image data DOUT outputted from the linear function correction circuit 22), and the positions of the starting points of the input-output curves are specified by starting point control data CP0_contR, CP0_contG and CP0_contB received from the starting point control circuit 28. Here, the starting point control data CP0_contR specifies the position of the starting point of the input-output curve of the processing to be performed on the R data DIN R of the input image data DIN. Correspondingly, the starting point control data CP0_contG specifies the position of the starting point of the input-output curve of the processing to be performed on the G data DIN G of the input image data DIN and the starting point control data CP0_contB specifies the position of the starting point of the input-output curve of the processing to be performed on the B data DIN B of the input image data DIN.
The number of bits of the R data DOUT R, G data DOUT G and B data DOUT B of the output image data DOUT is larger than that of the R data DIN R, G data DIN G and B data DIN B of the input image data DIN. This effectively avoids losing information of the grayscale levels of pixels in the correction calculation. In the present embodiment, in which the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN are generated as 8-bit data, the R data DOUT R, G data DOUT G and B data DOUT B of the output image data DOUT may be, for example, generated as 10-bit data.
The color reduction circuit 23, the latch circuit 24, the grayscale voltage generator circuit 25 and the data line drive circuit 26 function in total as a drive circuitry which drives the data lines 8 of the display region 5 of the LCD panel 2 in response to the output image data DOUT generated by the linear function correction circuit 22. Specifically, the color reduction circuit 23 performs a color reduction on the output image data DOUT generated by the linear function correction circuit 22 to generate color-reduced image data DOUT _ D. The latch circuit 24 latches the color-reduced image data DOUT _ D from the color reduction circuit 23 in response to a latch signal SSTB received from the timing control circuit 27 and forwards the color-reduced image data DOUT _ D to the data line drive circuit 26. The grayscale voltage generator circuit 25 feeds a set of grayscale voltages to the data line drive circuit 26. In one embodiment, the number of the grayscale voltages fed from the grayscale voltage generator circuit 25 may be 256 (=28) in view of the configuration in which the grayscale level of each subpixel 11 of each pixel 9 is represented with eight bits. The data line drive circuit 26 drives the data lines 8 of the display region 5 of the LCD panel 2 in response to the color-reduced image data DOUT _ D received from the latch circuit 24. In detail, the data line drive circuit 26 selects desired grayscale voltages from the set of the grayscale voltages received from the grayscale voltage generator circuit 25 in response to color-reduced image data DOUT _ D, and drives the corresponding data lines 8 of the LCD panel 2 to the selected grayscale voltages.
The timing control circuit 27 performs timing control of the entire drive IC 3 in response to the synchronization data DSYNC. In detail, the timing control circuit 27 generates the latch signal SSTB in response to the synchronization data DSYNC and feeds the generated latch signal SSTB to the latch circuit 24. The latch signal SSTB is a control signal instructing the latch circuit 24 to latch the color-reduced data DOUT _ D. Furthermore, the timing control circuit 27 generates a frame signal SFRM in response to the synchronization data DSYNC and feeds the generated frame signal SFRM to the starting point control circuit 28. It should be noted here that the frame signal SFRM is a control signal which informs the starting point control circuit 28 of the start of each frame period; the frame signal SFRM is asserted at the beginning of each frame period. The vertical synchronization signal VSYNC generated in response to the synchronization data DSYNC may be used as the frame signal SFRM. The timing control circuit 27 also generates coordinate data D(X, Y) indicating the coordinates of the pixel 9 for which the input image data DIN currently indicate the grayscale levels of the respective subpixels 11 thereof and feeds the generated coordinate data D(X, Y) to the starting point control circuit 28. When input image data DIN which describe the grayscale levels of the respective subpixels 11 of a certain pixel 9 are fed to the starting point control circuit 28, the timing control circuit 27 feeds the coordinate data D(X, Y) indicating the coordinates of the certain pixel 9 in the display region 5 to the starting point control circuit 28.
The starting point control circuit 28 controls the saturation enhancement processing performed in the linear function correction circuit 22. The starting point control circuit 28 generates the above-described starting point control data CP0_contR, CP0_contG and CP0_contB in response to the input image data DIN, to thereby control the position of the starting point of the input-output curve of the processing performed on the input image data DIN for each color. More specifically, the starting point control circuit 28 includes an average saturation calculation section 28 a, a luminance difference calculation section 28 b, an enhancement coefficient calculation section 28 c, and a starting point control data generator section 28 d.
The average saturation calculation section 28 a calculates the average saturation SAVE of each frame image (that is, the image displayed in the display region 5 of the LCD panel 2 in each frame period), from the input image data DIN.
The luminance difference calculation section 28 b calculates, for each pixel 9, the difference Rdist between the R data DIN R and the luminance value YPIXEL, the difference Gdist between the G data DIN G and the luminance value YPIXEL, and the difference Bdist between the B data DIN B and the luminance value YPIXEL, from the input image data DIN associated with each pixel 9.
The enhancement coefficient calculation section 28 c calculates an enhancement coefficient INST from the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN for each pixel 9. As described later, the enhancement coefficient INST is a coefficient indicating the degree of the saturation enhancement in the saturation enhancement processing.
The starting point control data generator section 28 d generates the starting point control data CP0_contR, CP0_contG, and CP0_contB in response to the average saturation SAVE calculated by the average saturation calculation section 28 a, the differences Rdist, Gdist, and Bdist calculated by the luminance difference calculation section 28 b and the enhancement coefficient INST calculated by the enhancement coefficient calculation section 28 c. In the saturation enhancement processing of input image data DIN associated with a certain pixel 9 with respect to a certain frame period, the starting point control data CP0_contR, CP0_contG, and CP0_contB used for the saturation enhancement processing for the certain pixel 9 are calculated on the basis of the average saturation SAVE of the frame image displayed in the certain frame period, the differences Rdist, Gdist, and Bdist and the enhancement coefficient INST calculated from the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN associated with the certain pixel 9.
FIG. 6 is a flowchart illustrating the saturation enhancement processing in the present embodiment, more specifically, the processing performed in the starting point control circuit 28 and the linear function correction circuit 22. Overall, the saturation enhancement processing in the present embodiment includes calculation of starting point control data CP0_contR, CP0_contG, and CP0_contB (steps S11 to S14) and processing on the input image data DIN on the basis of the starting point control data CP0_contR, CP0_contG, and CP0_contB (step S15).
The following processing is performed in the calculation of the starting point control data CP0_contR, CP0_contG, and CP0_contB.
At step S11, the average saturation SAVE of the frame image displayed in each frame period is calculated from the input image data DIN. The saturation S of a certain pixel 9 is calculated as the difference between the maximum value and minimum value of the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN associated with the certain pixel 9. The average saturation SAVE of a certain frame image is calculated as the average of the saturations of all the pixels 9 in the frame image. More specifically, the average saturation SAVE of a certain frame image is calculated in accordance with the following expressions (1a) and (1b):
S j = max ( R j , G j , B j ) - min ( R j , G j , B j ) ( 1 a ) S AVE = j S j Data_count ( 1 b )
where Rj, Gj, and Bj are the values of the R data DIN R, G data DIN G, and B data DIN B associated with pixel j of the frame image and Sj is the saturation of pixel j. In the above, max(Rj, Gj, Bj) is the maximum value of Rj, Gj, and Bj and min(Rj, Gj, Bj) is the minimum value of Rj, Gj, and Bj. Data_count is the number of the pixels 9 in the display region 5 (the number of the pixels 9 in the frame image) and Σ represents the sum with respect to all the pixels 9 in the display region 5. As described above, the average saturation SAVE is calculated by the average saturation calculation section 28 a. It is only necessary to calculate the average saturation SAVE once for each frame period.
The luminance value YPIXEL of each pixel 9 is further calculated from the input image data DIN and the difference Rdist between the R data DIN R and the luminance value YPIXEL, the difference Gdist between the G data DIN G and the luminance value YPIXEL and the difference Bdist between the B data DIN B and the luminance value YPIXEL are calculated from the input image data DIN associated with each pixel 9. The luminance value YPIXEL and the differences Rdist, Gdist, and Bdist are calculated by the luminance difference calculation section 28 b.
In detail, the luminance value YPIXEL of each pixel j is calculated as a weighted average of the values Rj, Gj, and Bj of the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN in accordance with the following expression (2a):
Y PIXEL =aR j +bG j +cB j,  (2a)
where a, b, and c are weights given to the R data DIN R, G data DIN G, and B data DIN B, respectively. The weights a, b, and c are larger than zero and satisfy the following expression (2b):
a+b+c=1  (2b)
The weights a, b, and c used for the calculation of the luminance value YPIXEL depend on the definition of the luminance value YPIXEL.
According to ITU-R BT.709, for example, the luminance value Y is given by the following expression:
Y PIXEL=0.2126·R j+0.7152·G j+0.0722·B j.  (2c)
This expression implies that the weights a, b and c are defined as follows:
a=0.2126,
b=0.7152, and
c=0.0722.
The difference Rdist between the R data DIN R and the luminance value YPIXEL, the difference Gdist between the G data DIN G and the luminance value YPIXEL, and the difference Bdist between the B data DIN B and the luminance value YPIXEL are calculated in accordance with the following expressions (3a) to (3c):
Rdist=R j −Y PIXEL  (3a)
Gdist=G j −Y PIXEL, and  (3b)
Bdist=B j −Y PIXEL.  (3c)
At step S13, an enhancement coefficient INST associated with each pixel 9 is further calculated from the input image data DIN associated with each pixel 9. The calculation of the enhancement coefficient INST is achieved by the enhancement coefficient calculation section 28 c as described above. The enhancement coefficient INST is calculated for each pixel 9 as described below: First, enhancement coefficients INSTR, INSTG, and INSTB are respectively calculated for R data DIN R, G data DIN G, and B data DIN B of the input image data DIN associated with each pixel 9.
The enhancement coefficients INSTR, INSTG, and INSTB are calculated with a function f(x) as follows:
INSTR =f(D IN R),  (4a)
INSTG =f(D IN G), and  (4b)
INSTB =f(D IN B),  (4c)
where f(x) is a function satisfying the following conditions (a) to (c):
(a) f(x) takes the maximum value when x is β;
(b) f(x) monotonically increases as x increases when x is less than β; and
(c) f(x) monotonically decreases as x increases when x is more than β.
β may be determined as DIN MAX/2 or the integer closest to DIN MAX/2 (if there are two integers closest to DIN MAX/2, one selected from the two integers), where DIN MAX is the allowed maximum value of the R data DIN R, G data DIN G, and B data DIN B. In the present embodiment in which the R data DIN R, G data DIN G, and B data DIN B are eight bit data, for example, the allowed maximum value DIN MAX is “255” and β is “127” or “128”.
Note that expressions (4a) to (4b) can be rewritten as the following expressions (5a) to (5c):
INSTR =f(R j),  (5a)
INSTG =f(G j), and  (5b)
INSTB =f(B j),  (5c)
where Rj, Gj, and Bj are the values of the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN associated with pixel j.
FIG. 7 is a graph illustrating a specific example of a function used to determine the enhancement coefficients INSTR, INSTG, and INSTB. In one embodiment, the enhancement coefficients INSTR, INSTG, and INSTB associated with pixel j may be calculated from the values Rj, Gj, and Bj of the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN associated with pixel j in accordance with the following expressions (6a) to (6c):
INST R = { reg β · R j β if R j β reg β · D IN MAX - R j β if R j > β ( 6 a ) INST G = { reg β · G j β if G j β reg β · D IN MAX - G j β if G j > β ( 6 b ) INST B = { reg β · B j β if B j β reg β · D IN MAX - B j β if B j > β ( 6 c )
where reg is a register value set in an enhancement coefficient setting register (not shown) integrated within the driver IC 3; reg takes a value larger than zero and equal to or less than β. The degree of the saturation enhancement can be adjusted by adjusting the register value reg.
In particular, when the register value reg is equal to β, expressions (6a) to (6c) can be rewritten as the following expressions (7a) to (7c):
INST R = { R j β if R j β D IN MAX - R j β if R j > β ( 7 a ) INST G = { G j β if G j β D IN MAX - G j β if G j > β ( 7 b ) INST B = { B j β if B j β D IN MAX - B j β if B j > β ( 7 c )
FIG. 7 is a graph illustrating the enhancement coefficients INSTR, INSTG, and INSTB calculated in accordance with expressions (7a) to (7c).
The enhancement coefficient INST finally calculated for each pixel 9 is determined as the minimum value of enhancement coefficients INSTR, INSTG, and INSTB, which are obtained for the R data DIN R, G data DIN G, and B data DIN B, respectively. In other words, it holds:
INST=min(INSTR,INSTG,INSTB).  (8)
Referring back to FIG. 6, at step S14, Starting point control data CP0_contR, CP0_contG, and CP0_contB are calculated on the basis of the average saturation SAVE, differences Rdist, Gdist, and Bdist and enhancement coefficient INST, which are calculated as described above. The calculation of the starting point control data CP0_contR, CP0_contG, and CP0_contB are achieved by the starting point control data generator section 28 d.
The starting point control data CP0_contR is determined so that the starting point control data CP0_contR increases as the difference DIN MAX−SAVE increases and is proportional to the difference Rdist, where the difference DIN MAX−SAVE is obtained by subtracting the average saturation SAVE from the allowed maximum value DIN MAX. Correspondingly, the starting point control data CP0_contG is determined so that the starting point control data CP0_contG increases as the difference DIN MAX−SAVE increases and is proportional to the difference Gdist, and the starting point control data CP0_contB is determined so that the starting point control data CP0_contB increases as the difference DIN MAX−SAVE increases and is proportional to the difference Bdist. The enhancement coefficient INST is used as a parameter indicating the degree of the saturation enhancement, commonly for the calculations of the starting point control data CP0_contR, CP0_contG, and CP0_contB.
In one embodiment, the starting point control data CP0_contR, CP0_contG, and CP0_contB may be calculated in accordance with the following expressions (9a) to (9c):
CP0_cont R = INST · D IN MAX - S AVE D IN MAX · Rdist ( 9 a ) CP0_cont G = INST · D IN MAX - S AVE D IN MAX · Gdist ( 9 b ) CP0_const B = INST · D IN MAX - S AVE D IN MAX · Bdist ( 9 c )
At step S15, the input image data DIN are processed in response to the starting point control data CP0_contR, CP0_contG, and CP0_contB, to thereby calculate output image data DOUT. The output image data DOUT are obtained as a result of the saturation enhancement processing to the input image data DIN.
In the processing at step S15, the input-output curves of the processing applied to the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN are each defined as a straight line. The starting point of the input-output curves of the processing applied to the R data DIN R of the input image data DIN is specified by the starting point control data CP0_contR. Correspondingly, the starting point of the input-output curves of the processing applied to the G data DIN G of the input image data DIN is specified by the starting point control data CP0_contG and the starting point of the input-output curves of the processing applied to the B data DIN B of the input image data DIN is specified by the starting point control data CP0_contB. Note that the end point of the input-output curve of the processing to the input image data DIN (the point on the input-output curve corresponding to input image data DIN with the allowed maximum value DIN MAX) is determined commonly for the R data DIN R, G data DIN G, and B data DIN B, so that the value of the output image data DOUT is determined as the allowed maximum value DOUT MAX when the value of the input image data DIN is the allowed maximum value DIN MAX. Note that allowed maximum value DOUT MAX depends on the bit width of the R data DOUT R, G data DOUT G, and B data DOUT B of the output image data DOUT.
More specifically, the values of the R data DOUT R, G data DOUT G, and B data DOUT B of the output image data DOUT are calculated in accordance with the following expressions (10a) to (104
D OUT R = D OUT MAX - CP0_cont R D IN MAX · D IN R + CP0_cont R ( 10 a ) D OUT G = D OUT MAX - CP0_cont G D IN MAX · D IN G + CP0_cont G ( 10 b ) D OUT B = D OUT MAX - CP0_cont B D IN MAX · D IN B + CP0_cont B ( 10 c )
FIG. 8 illustrates the respective relations between the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN and the R data DOUT R, G data DOUT G, and B data DOUT B of the output image data DOUT in the case when the R data DOUT R, G data DOUT G, and B data DOUT B of the output image data DOUT are calculated in accordance with the above expressions (10a) to (10c). In FIG. 8, the symbol “CP5” denotes the end point of each input-output curve. It should be noted that at least one of the starting point control data CP0_contR, CP0_contG, and CP0_contB is positive and at least another one is negative. This is because the starting point control data CP0_contR, CP0_contG, and CP0_contB are calculated in accordance with expressions (9a) to (9c) and at least one of the differences Rdist, Gdist, and Bdist is positive and at least another one is negative, where the luminance value YPIXEL is a weighted average of the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN. It would be easily understood from FIG. 8 that the processing in accordance with expressions (10a) to (10c) effectively enhances the saturation, as at least one of the starting point control data CP0_contR, CP0_contG, and CP0_contB is positive and at least another one is negative.
The output image data DOUT calculated by the linear function correction circuit 22 with the above-described series of expressions are forwarded to the color reduction circuit 23. In the color reduction circuit 23, a color reduction is performed on the output image data DOUT to generate the color-reduced image data DOUT _ D. The color-reduced image data DOUT _ D are forwarded to the data line drive circuit 26 via the latch circuit 24 and the data lines 8 of the LCD panel 2 are driven in response to the color-reduced image data DOUT _ D.
The above-described saturation enhancement processing in the present embodiment effectively achieves saturation enhancement only with simple processing. This effectively contributes the reduction of the circuit size of an image processing circuit used for the saturation enhancement (the starting point control circuit 28 and the linear function correction circuit 22 in this embodiment).
It should be also noted that the saturation enhancement processing in the present embodiment only causes a reduced change in the luminance value YPIXEL of each pixel. In the following, a description is given of the fact that the saturation enhancement processing in the present embodiment only causes a reduced change in the luminance value YPIXEL, with a specific numerical example.
In the following, the case is considered when Rj, Gj, and Bj, which are the values of the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN, are 50, 100, and 200, respectively. In this case, the luminance value YPIXEL is obtained in accordance with expression (2c) as follows:
Y PIXEL = 0.2126 × 50 + 0.7152 × 100 + 0.0722 × 200 = 96.59 97.
It should be noted that the value “97” is obtained by rounding to handle the luminance value YPIXEL as an integer representable with eight bits.
The differences Rdist, Gdist, and Bdist are calculated in accordance with expressions (3a) to (3c) as follows:
Rdist=50−97=−47,
Gdist=100−97=3, and
Bdist=200−97=103.
With expressions (7a) to (7c) for β being 127, the enhancement coefficients INSTR, INSTG, and INSTB are obtained as follows:
INSTR=50/127,
INSTG=100/127, and
INSTB=(255−200)/127=55/127.
Since the enhancement coefficient INST is defined as the minimum value of the enhancement coefficients INSTR, INSTG, and INSTB (see expression (8)), the enhancement coefficient INST is obtained as follows:
INST=50/127.
The starting point control data CP0_contR, CP0_contG, and CP0_contB are obtained by expressions (9a) to (9c). It should be noted that, with respect to (DIN MAX−SAVE)/DIN MAX recited in expressions (9a) to (9c), it holds:
0≤(D IN MAX −S AVE)/D IN MAX≤1.
To discuss the change in the luminance value YPIXEL in the case when the saturation is most strongly enhanced, let us consider the case when (DIN MAX−SAVE)/DIN MAX is one; in this case, the saturation is most strongly enhanced. Under this assumption, the starting point control data CP0_contR, CP0_contG, and CP0_contB are obtained by expressions (9a) to (9c) as follows:
CP0_contR=(50/127)×1×(−47)=−18.50≈−19,
CP0_contG=(50/127)×1×3=1.18≈1, and
CP0_contB=(50/127)×1×103=44.55≈45.
The values of the R data DOUT R, G data DOUT G and B data DOUT B are obtained with expressions (10a) to (10c) as follows:
D OUT R = { ( 255.00 + 19 ) / 255 } × 50 - 19 = 34.72 34.75 , D OUT G = { ( 255.00 + 3 ) / 255 } × 100 + 3 = 101.82 101.75 , and D OUT B = { ( 255.00 - 103 ) / 255 } × 200 + 103 = 222.21 222.25 .
Note that the values of the R data DOUT R, G data DOUT G, and B data DOUT B are represented with 10 bits, taking a value from 0.00 to 255.00 in increments of 0.25.
The luminance value YPIXEL′ after the saturation enhancement processing is calculated from the thus-obtained output image data DOUT as follows:
Y PIXEL = 0.2126 × 34.75 + 0.7152 × 101.75 + 0.0722 × 222.25 = 96.2059
This result proves that the saturation enhancement processing in the present embodiment causes only a reduced change in the luminance value, since the luminance value YPIXEL calculated from the original input image data DIN is 96.59. The reduced change in the luminance value caused by the saturation enhancement processing results from that the enhancement coefficient INST is calculated with a function f(x) and β, where β is determined as DIN MAX/2 or a value close to DIN MAX/2, and the function f(x) satisfies the following conditions (a) to (c):
(a) f(x) takes the maximum value when x is β;
(b) f(x) monotonically increases as x increases when x is less than β; and
(c) f(x) monotonically decreases as x increases when x is more than β.
The calculation of the enhancement coefficient INST in this way effectively reduces changes in the luminance value in performing the saturation enhancement processing.
Second Embodiment
FIG. 9 is a block diagram illustrating an example of the configuration of the driver IC 3 in a second embodiment. In the second embodiment, saturation enhancement and contrast enhancement are concurrently achieved, not by a series of processes as illustrated in FIG. 1B. Concurrently performing saturation enhancement and contrast enhancement is effective for reducing the circuit size. In the present embodiment, in order to concurrently achieve saturation enhancement and contrast enhancement, an approximate gamma correction circuit 31 is used in place of the linear function correction circuit 22, and a characterization data calculation circuit 32 and a correction point data calculation circuit 33 are additionally integrated in the driver IC 3.
The approximate gamma correction circuit 31 performs correction calculations based on an approximate expression of the gamma correction on the input image data DIN to generate the output image data DOUT. In the present embodiment, the input-output curves used in the correction calculations in the approximate gamma correction circuit 31 are each determined as a curve obtained by modifying a gamma curve specified by a certain gamma value so as to achieve contrast enhancement. It should be noted that, in the present embodiment, contrast enhancement is achieved to some degree by using a gamma curve as each input-output curve and additional contrast enhancement is achieved by modifying the shape of the gamma curve.
The shapes of the input-output curves used in the correction calculation by the approximate gamma correction circuit 31 are specified by correction point data sets CP_selR, CP_selG, and CP_selB received from the correction point data calculation circuit 33. The shape of the input-output curve used in the correction calculation for the R data DIN R of the input image data DIN is specified by the correction point data set CP_selR. Correspondingly, the shape of the input-output curve used in the correction calculation for the G data DIN G of the input image data DIN is specified by the correction point data set CP_selG, and the shape of the input-output curve used in the correction calculation for the B data DIN B of the input image data DIN is specified by the correction point data set CP_selB.
The characterization data calculation circuit 32 generates characterization data indicating one or more feature quantities of each frame image (the image displayed in the display region 5 in each frame period) on the basis of the input image data DIN. In the present embodiment, the characterization data include APL data DCHR _ APL indicating the average picture level (APL) of each frame image, and variance data DCHR _ σ2 indicating the variance of the luminance values of the pixels of each frame image. Note that the calculation of the APL data DCHR _ APL and variance data DCHR _ σ2 may be performed once for each frame period.
The correction point data calculation circuit 33 calculates the above-described correction point data sets CP_selR, CP_selG, and CP_selB in response to the characterization data generated by the characterization data calculation circuit 32 (in the present embodiment, the APL data DCHR _ APL and variance data DCHR _ σ2) and the starting point control data CP0_contR, CP0_contG, and CP0_contB generated by the starting point control circuit 28. Overall, the correction point data calculation circuit 33 functions as a correction control circuitry which controls the shapes of the input-output curves in response to the characterization data (the APL data DCHR _ APL and variance data DCHR _ σ2) generated by the characterization data calculation circuit 32 and adjusts the positions of the starting points of the input-output curves in response to the starting point control data CP0_contR, CP0_contG, and CP0_contB.
FIG. 10 illustrates the input-output curve specified by each correction point data set CP_selk generated by the correction point data calculation circuit 33 and the contents of the correction calculation in accordance with the input-output curve. Each correction point data set CP_selk includes correction point data CP0 to CP5. The correction point data CP0 to CP5 are each defined as data indicating a point in a coordinate system in which input image data DIN k are associated with the horizontal axis (or a first axis) and output image data DOUT k are associated with the vertical axis (or a second axis). The correction point data CP0 indicate the position of the starting point of the input-output curve, which may be denoted also by symbol “CP0”, and the correction point data CP5 indicate the position of the end point of the input output curve, which may be denoted also by symbol “CP5”. The correction point data CP2 and CP3 respectively indicate the positions of correction points which the input-output curve passes through near the center of the input-output curves, which may be denoted also by symbols “CP2” and “CP3”, respectively. The correction point data CP1 indicate the position of a correction point located between the correction points CP0 and CP2, which is denoted also by symbol “CP1”, and the correction point data CP4 indicate the position of a correction point located between the correction points CP3 and CP5, which is also denoted by symbol “CP4”. The shape of the gamma curve is specified by appropriately determining the positions of the correction points CP0 to CP0 indicated by the correction point data CP0 to CP5.
As illustrated in FIG. 10, for example, it is possible to specify the shape of a gamma curve as being convex downward by determining the positions of the correction points CP1 to CP4 as being lower than the straight line connecting the both ends of the gamma curve. In the approximate gamma correction circuit 31, the output image data DOUT k are generated by performing a gamma correction in accordance with the gamma curve with the shape specified by the correction point data CP0 to CP5 included in the correction point data set CP_selk. As described later in detail, the correction point data CP0 to CP5 of the correction point data sets CP_selR, CP_selG and CP_selB are determined in response to the characterization data (in the present embodiment, APL data DCHR _ APL and variance data DCHR _ σ2) generated by the characterization data calculation circuit 32. This effectively allows controlling the shapes of the input-output curves used in the correction calculations of the R data DIN R, G data DIN G and B data DIN B of the input image data DIN in response to the characterization data. Furthermore, the correction point data CP0 of the correction point data sets CP_selR, CP_selG and CP_selB are adjusted in response to the starting point control data CP0_contR, CP0_contG and CP0_contB, respectively. This effectively allows adjusting the positions of the starting points of the input-output curves used in the correction calculations of the R data DIN R, G data DIN G and B data DIN B of the input image data DIN, to thereby achieve saturation enhancement.
FIG. 11 is a block diagram illustrating an example of the configuration of the approximate gamma correction circuit 31. The approximate gamma correction circuit 31 includes approximate gamma correction units 34R, 34G and 34B, which are prepared for the R data DIN R, G data DIN G and B data DIN B of the input image data DIN, respectively. The approximate gamma correction unit 34R performs a correction calculation with an arithmetic expression on the R data DIN R of the input image data DIN to generate the R data DOUT R of the output image data DOUT. Correspondingly, the approximate gamma correction units 34G and 34B perform correction calculations with arithmetic expressions on the G data DIN G and B data DIN B of the input image data DIN to generate the G data DOUT G and B data DOUT B of the output image data DOUT, respectively. As described above, the number of bits of the R data DOUT R, G data DOUT G and B data DOUT B of output image data DOUT is ten bits in the present embodiment; this means that the number of bits of the R data DOUT R, G data DOUT G and B data DOUT B of the output image data DOUT is larger than that of the R data DIN R, G data DIN G and B data DIN B of the input image data DIN.
Coefficients of the arithmetic expression used for the correction calculation by the approximate gamma correction unit 34R are determined depending on the correction point data CP0 to CP5 of the correction point data set CP_selR. Correspondingly, coefficients of the arithmetic expressions used for the correction calculation by the approximate gamma correction units 34G and 34B are determined depending on the correction point data CP0 to CP5 of the correction point data set CP_selG and CP_selB, respectively.
FIG. 12 is a block diagram illustrating an example of the configuration of the correction point data calculation circuit 33. In the example illustrated in FIG. 12, the correction point data calculation circuit 33 includes: a correction point data set storage register 41, an interpolation/selection circuit 42 and a correction point data adjustment circuit 43 and a CP0 adjustment circuit 44.
The correction point data set storage register 41 stores therein a plurality of correction point data sets CP#1 to CP#m. The correction point data sets CP#1 to CP#m are used as seed data for determining the above-described correction point data sets CP_selR, CP_selG and CP_selB. Each of the correction point data sets CP#1 to CP#m includes correction point data CP0 to CP5 defined as illustrated in FIG. 10.
The interpolation/selection circuit 42 determines gamma values γ_PIXELR, γ_PIXELG and γ_PIXELB on the basis of the APL data DCHR _ APL received from the characterization data calculation circuit 32 and determines correction point data sets CP_LR, CP_LG and CP_LB corresponding to the gamma values γ_PIXELR, γ_PIXELG and γ_PIXELB thus determined. Here, the gamma value γ_PIXELR is the gamma value of the gamma curve used for the correction calculation performed on the R data DIN R of the input image data DIN. Correspondingly, the gamma value γ_PIXELG is the gamma value of the gamma curve used for the correction calculation performed on the G data DIN G of the input image data DIN and the gamma value γ_PIXELB is the gamma value of the gamma curve used for the correction calculation performed on the B data DIN B of the input image data DIN.
In one embodiment, the interpolation/selection circuit 42 may select one of the correction point data sets CP#1 to CP#m on the basis of the gamma value γ_PIXELk and determine the correction point data set CP_Lk as the selected one of the correction point data sets CP#1 to CP#m. Alternatively, the interpolation/selection circuit 42 may determine the correction point data set CP_Lk by selecting two of correction point data sets CP#1 to CP#m on the basis of the gamma value γ_PIXELk and applying an interpolation to the selected two correction point data sets. Details of the determination of the correction point data sets CP_LR, CP_LG and CP_LB are described later. The correction point data sets CP_LR, CP_LG and CP_LB determined by the interpolation/selection circuit 42 are forwarded to the correction point data adjustment circuit 43.
The correction point data adjustment circuit 43 modifies the correction point data sets CP_LR, CP_LG and CP_LB in response to the variance data DCHR _ σ2 received from the characterization data calculation circuit 32. This operation of the correction point data adjustment circuit 43 is technically equivalent to processing to modify the shape of the gamma curve specified by the gamma value γ_PIXELk in response to the variance data DCHR _ σ2. As described later, in the present embodiment, the correction point data adjustment circuit 43 modifies the correction point data CP1 and CP4 of the correction point data sets CP_LR, CP_LG and CP_LB in response to the variance data DCHR _ σ2, to thereby achieve contrast enhancement.
The CP0 adjustment circuit 44 modifies the correction point data sets CP_LR, CP_LG and CP_LB in response to the starting point control data CP0_contR, CP0_contG and CP0_contB received from the starting point control circuit 28. This operation of the CP0 adjustment circuit 4 is technically equivalent to processing to adjust the positions of the starting points of the input-output curves in response to the starting point control data CP0_contR, CP0_contG and CP0_contB. The correction point data sets CP_LR, CP_LG and CP_LB thus modified by the correction point data adjustment circuit 43 and the CP0 adjustment circuit 44 are used as the correction point data sets CP_selR, CP_selG and CP_selB, which are finally fed to the approximate gamma correction circuit 31.
In the present embodiment, the correction point data sets CP_LR, CP_LG and CP_LB are first modified by the correction point data adjustment circuit 43 to generate correction point data sets CP_CRB, CP_CRG and CP_CRB and the correction point data sets CP_CRB, CP_CRG and CP_CRB are further modified by the CP0 adjustment circuit 44 to thereby generate the correction point data sets CP_selR, CP_selG and CP_selB, which are finally fed to the approximate gamma correction circuit 31. In an alternative embodiment, the correction point data calculation circuit 33 may be configured so that the correction point data sets CP_LR, CP_LG and CP_LB are first modified by the CP0 adjustment circuit 44 and the modified correction point data sets CP_LR, CP_LG and CP_LB are further modified by the correction point data adjustment circuit 43 to generate the correction point data sets CP_selR, CP_selG and CP_selB. The correction point data adjustment circuit 43 and the CP0 adjustment circuit 44 may be combined to form a unified circuit. In this case, the unified circuit may modify the correction point data sets CP_LR, CP_LG and CP_LB in response to the variance data DCHR _ σ2 and the starting point control data CP0_contR, CP0_contG and CP0_contB. Details of the operations of the respective circuits in the correction point data calculation circuit 33 are described later.
FIG. 13 is a flowchart illustrating the procedure of the correction calculation performed on the input image data DIN in the second embodiment. Similarly to the first embodiment, the input image data DIN are fed to the starting point control circuit 28 and the starting point control data CP0_contR, CP0_contG and CP0_contB are generated by the starting point control circuit 28 on the basis of the input image data DIN at steps 11 to 14. The starting point control data CP0_contR, CP0_contG and CP0_contB are generated through the same processing as in the first embodiment.
At step S21, the input image data DIN are also fed to the characterization data calculation circuit 32 and the APL data DCHR _ APL and variance data DCHR _ σ2 of each frame image are generated by the characterization data calculation circuit 32 from the input image data DIN. As described above, the APL data DCHR _ APL indicate the APL of each frame image and the variance data DCHR _ σ2 indicate the variance of the luminance values of the pixels in each frame period.
At step S22, the gamma values to be used for the correction calculation of the input image data DIN associated with each pixel 9 in each frame image are calculated from the APL data DCHR _ APL of each frame image. In the present embodiment, a gamma value is calculated for each of the R data DIN R, G data DIN G and B data DIN B of the input image data DIN associated with each pixel 9. In detail, the gamma value γ_PIXEL R to be used for the correction calculation performed on the R data DIN R of the input image data DIN associated with a certain frame image is calculated by the following expression (11a):
γ_PIXELR=γ_STDR+APL·ηR,  (11a)
where γ_STDR is a given reference gamma value, APL is the average picture level of the certain frame image, and ηR is a given positive proportionality constant. It should be noted that, under the definition with expression (11a), the gamma value γ_PIXEL R increases as the average picture level APL increases.
Correspondingly, the gamma values γ_PIXEL G and γ_PIXEL B to be used for the correction calculations performed on the G data DIN G and B data DIN B of the input image data DIN associated with the certain frame image are calculated by the following expressions (11b) and (11c):
γ_PIXELG=γ_STDG+APL·ηG, and  (11b)
γ_PIXELB=γ_STDB+APL·ηB,  (11c)
where γ_STDG and γ_STDB are given reference gamma values, and ηG and ηB are given positive proportionality constants. γ_STDR, γ_STDG, and γ_STDB may be equal to each other, or different, and ηR, ηG and ηB may be equal to each other, or different. It should be noted that the gamma values γ_PIXELR, γ_PIXELG, and γ_PIXELB are calculated for each frame image.
When γ_STDR, γ_STDG, and γ_STDB are equal to each other and ηR, ηG, and ηB are equal to each other, a common gamma value is calculated for the R data DIN R, G data DIN G and B data DIN B of the input image data DIN; in this case, the common gamma value γ_PIXEL is calculated by the following expression (11d):
γ_PIXEL=γ_STD+APL·η.  (11d)
At step S23, the correction point data sets CP_LR, CP_LG, and CP_LB are selected or determined on the basis of the thus-calculated gamma values γ_PIXELR, γ_PIXELG and γ_PIXELB, respectively. It should be noted that the correction point data sets CP_LR, CP_LG and CP_LB are seed data used for calculating the correction point data sets CP_selR, CP_selG, and CP_selB, which are finally fed to the approximate gamma correction circuit 31. The correction point data sets CP_LR, CP_LG and CP_LB are determined for each frame image.
In one embodiment, the correction point data sets CP_LR, CP_LG, and CP_LB are each selected from the correction point data sets CP#1 to CP#m stored in the correction point data set storage register 41 of the correction point data calculation circuit 33. As described above, the correction point data sets CP#1 to CP#m correspond to different gamma values γ and each of the correction point data sets CP#1 to CP#m includes correction point data CP0 to CP5.
The correction point data CP0 to CP5 of a correction point data set CP#j corresponding to a certain gamma value γ are determined as follows:
( 1 ) For y < 1 , CP 0 = 0 CP 1 = 4 · Gamma [ K / 4 ] - Gamma [ K ] 2 CP 2 = Gamma [ K - 1 ] CP 3 = Gamma [ K ] CP 4 = 2 · Gamma [ ( D IN MAX + K - 1 ) / 2 ] - D OUT MAX CP 5 = D OUT MAX and ( 12 a ) ( 2 ) for y 1 , CP 0 = 0 CP 1 = 2 · Gamma [ K / 2 ] - Gamma [ K ] CP 2 = Gamma [ K - 1 ] CP 3 = Gamma [ K ] CP 4 = 2 · Gamma [ ( D IN MAX + K - 1 ) / 2 ] - D OUT MAX CP 5 = D OUT MAX ( 12 b )
where DIN MAX is the allowed maximum value of the R data DIN R, G data DIN G and B data DIN B of the input image data DIN; DIN MAX depends on the number of bits of the R data DIN R, G data DIN G, and B data DIN B. Similarly, DOUT MAX is the allowed maximum value of the R data DOUT R, G data DOUT G and B data DOUT B of the output image data DOUT; DOUT MAX depends on the number of bits of the R data DOUT R, G data DOUT G and B data DOUT B. K is a constant given by the following expression (13a):
K=(D IN MAX+1)/2.  (13a)
In the above, the function Gamma [x], which is a function corresponding to the strict expression of the gamma correction, is defined by the following expression:
Gamma[x]=D OUT MAX·(x/D IN MAX)γ  (13b)
In the present embodiment, the correction point data sets CP#1 to CP#m are determined so that the gamma value γ recited in expression (13b) to which a correction point data set CP#j selected from the correction point data sets CP#1 to CP#m corresponds is increased as j is increased. In other words, it holds:
γ12< . . . <γm-1m,  (14)
where γj is the gamma value corresponding to the correction point data set CP#j.
In one embodiment, the correction point data set CP_LR is selected from the correction point data sets CP#1 to CP#m on the basis of the gamma value γ_PIXELR. The correction point data set CP_LR is determined as a correction point data set CP#j with a larger value of j as the gamma value γ_PIXELR increases. Correspondingly, the correction point data sets CP_LG and CP_LB are selected from the correction point data sets CP#1 to CP#m on the basis of the gamma values γ_PIXELG and γ_PIXELB, respectively. FIG. 14 is a graph illustrating the relation among the average picture level (APL) of each frame image, γ_PIXELk and the correction point data set CP_Lk in the case when the correction point data set CP_Lk is determined in this manner. As the average picture level of each frame image increases, the gamma value γ_PIXELk is increased and a correction point data set CP#j with a larger value of j is selected as the correction point data set CP_Lk.
In an alternative embodiment, the correction point data sets CP_LR, CP_LG, and CP_LB may be determined as follows: The correction point data sets CP#1 to CP#m are stored in the correction point data set storage register 41 of the correction point data calculation circuit 33. The number of the correction point data sets CP#1 to CP#m stored in the correction point data set storage register 41 is 2P−(Q−1), where P is the number of bits used to describe the average picture level (APL) of each frame image and Q is a predetermined integer equal to more than two and less than P. This implies that m=2P−(Q−1). The correction point data sets CP#1 to CP#m to be stored in the correction point data set storage register 41 may be fed from the processor 4 to the drive IC 3 as initial settings.
Furthermore, two correction point data sets CP#q and CP#(q+1) are selected on the basis of the gamma value γ_PIXELk (k is any one of “R”, “G” and “B”) from among the correction point data sets CP#1 to CP#m stored in the correction point data set storage register 41 for determining the correction point data set CP_Lk, where q is an integer from one to m−1. The two correction point data sets CP#q and CP#(q+1) are selected to satisfy the following expression:
γq<γ_PIXELkq+1.  (15)
The correction point data CP0 to CP5 of the correction point data set CP_Lk are respectively calculated with an interpolation of correction point data CP0 to CP5 of the selected two correction point data sets CP#q and CP#(q+1).
More specifically, the correction point data CP0 to CP5 of the correction point data set CP_Lk (where k is any of “R”, “G” and “B”) are calculated from the correction point data CP0 to CP5 of the selected two correction point data sets CP#q and CP#(q+1) in accordance with the following expressions:
CPα_L k=CPα(#q)+{(CPα(#(q+1))−CPα(#q))/2Q}×APL_PIXEL[Q−1:0],  (16)
where α is an integer from zero to five, CPα_Lk is the correction point data CPα of correction point data set CP_Lk, CPα(#q) is the correction point data CPα of the selected correction point data set CP#q, CPα(#(q+1)) is the correction point data CPα of the selected correction point data set CP#(q+1), and APL[Q−1:0] is the lowest Q bits of the average picture level of each frame image. FIG. 15 is a graph illustrating the relation among the average picture level (APL), γ_PIXELk and the correction point data set CP_Lk in the case when the correction point data set CP_Lk is determined in this manner. As the average picture level increases, the gamma value γ_PIXELk is increased and correction point data sets CP#q and CP#(q+1) with a larger value of q are selected. The correction point data set CP_Lk is determined to correspond to a gamma value in a range from the gamma value γq to γq+1, which the correction point data sets CP#q and CP#(q+1) respectively correspond to.
FIG. 16 is a graph schematically illustrating the shapes of the gamma curves corresponding to the correction point data sets CP#q and CP#(q+1) and the correction point data set CP_Lk. Since the correction point data CPα of the correction point data set CP_Lk is obtained through the interpolation of the correction point data CPα(#q) and CPα(#(q+1)) of the correction point data sets CP#q and CP#(q+1), the shape of the gamma curve corresponding to the correction point data set CP_Lk is determined so that the gamma curve corresponding to the correction point data set CP_Lk is located between the gamma curves corresponding to the correction point data sets CP#q and CP#(q+1). The calculation of the correction point data CP0 to CP5 of the correction point data set CP_Lk through the interpolation of the correction point data CP0 to CP5 of the correction point data sets CP#q and CP#(q+1) is advantageous for allowing finely adjusting the gamma value used for the gamma correction even when only a reduced number of the correction point data sets CP#1 to CP#m are stored in the correction point data set storage register 41.
It should be noted that, when a gamma value γ_PIXEL is calculated commonly for the R data DIN R, G data DIN G and B data DIN B, a common correction point data set CP_L is selected or calculated from the gamma value γ_PIXEL in a similar way.
At step S24, the correction point data sets CP_LR, CP_LG, and CP_LB thus selected or calculated are modified in response to the variance data DCHR _ σ2 to generate correction point data sets CP_CRR, CP_CRG and CP_CRB. It should be noted that, since the correction point data sets CP_LR, CP_LG and CP_LB each represent the shape of a specific gamma curve, modifying the correction point data sets CP_LR, CP_LG and CP_LB in response to the variation data DCHR _ σ2 is technically equivalent to modifying the shapes of the gamma curves used for the correction calculations in response to the variation data DCHR _ σ2.
FIG. 17 is a conceptual diagram illustrating the technical significance of modification of the correction point data set CP_Lk on the basis of variance data DCHR _ σ2. A reduced variance σ2 of the luminance values of a certain frame image means that the frame image includes an increased number of pixels with luminance values close to the APL of the frame image. In other words, the contrast of the frame image corresponding to the input image data DIN is small. When the contrast of the frame image corresponding to the input image data DIN is small, it is possible to display the frame image with an improved image quality by performing a correction calculation to enhance the contrast by the approximate gamma correction circuit 31.
Since the correction point data CP1 and CP4 of the correction point data set CP_Lk largely influence the contrast, the correction point data CP1 and CP4 of the correction point data set CP_Lk are adjusted on the basis of the variance data DCHR _ σ2 in the present embodiment. The correction point data CP1 of the correction point data set CP_Lk is modified so that the correction point data CP1 of the correction point data sets CP_CRk, which is obtained as the result of the modification, is decreased as the variance σ2 indicated by the variance data DCHR _ σ2 decreases. The correction point data CP4 of the correction point data set CP_Lk is, on the other hand, modified so that the correction point data CP4 of the correction point data sets CP_CRk is increased as the variance σ2 indicated by the variance data DCHR _ σ2 decreases. Such modification results in that the correction calculation in the approximate gamma correction circuit 31 is performed to enhance the contrast, when the contrast of the image corresponding to the input image data DIN is small. It should be noted that the correction point data CP0, CP2, CP3 and CP5 of the correction point data set CP_Lk are not modified in the present embodiment.
At step S25, the correction point data set CP_CRR, CP_CRG, and CP_CRB are modified in response to the starting point control data CP0_contR, CP0_contG, and CP0_contB to generate the correction point data sets CP_selR, CP_selG and CP_selB. In the present embodiment, the starting points of the input-output curves, that is, the correction point data CP0 of the correction point data sets CP_CRR, CP_CRG and CP_CRB are modified in response to the starting point control data CP0_contR, CP0_contG and CP0_contB. Since a correction point data CP0 indicates the position of the starting point of a gamma curve (input-output curve), modifying the correction point data CP0 of the correction point data sets CP_CRR, CP_CRG, and CP_CRB is technically equivalent to adjusting the positions of the starting points of the gamma curves associated with the correction point data sets CP_CRR, CP_CRG and CP_CRB.
The correction point data CP1 to CP4 of the correction point data sets CP_CRR, CP_CRG, and CP_CRB are also modified in accordance with the modification of the correction point data CP0, that is, the adjustment of the positions of the input-output curves. Since the correction point data CP1 to CP4 of the correction point data sets CP_CRR, CP_CRG, and CP_CRB indicates the shapes of the intermediate portions of the input-output curves, modifying the correction point data CP1 to CP4 of the correction point data sets CP_CRR, CP_CRG and CP_CRB is technically equivalent to modifying the shapes of the intermediate portions of the input-output curves. Correction point data corresponding to correction points close to the starting points of the input-output curves (for example, correction point data CP1) are subjected to a relatively large modification and correction point data corresponding to correction points close to the end points of the input-output curves (for example, correction point data CP4) are subjected to a relatively small modification. More specifically, correction point control data CP1_contk to CP4_contk, which indicate the amounts of modifications of the correction point data CP1 to CP4 of the correction point data set CP_CRk, are calculated in accordance with the following expressions (17a) to (17d):
CP1_cont k = ( D OUT MAX - CP0_cont k ) D IN MAX · K 2 + CP0_cont k ( 17 a ) CP2_cont k = ( D OUT MAX - CP0_cont k ) D IN MAX · ( K · 1 ) + CP0_cont k ( 17 b ) CP3_cont k = ( D OUT MAX - CP0_cont k ) D IN MAX · K + CP0_cont k ( 17 c ) CP4_cont k = ( D OUT MAX - CP0_cont k ) D IN MAX · ( D IN MAX + K - 1 ) 2 + CP0_cont k ( 17 d )
FIG. 18 schematically illustrates the correction point control data CP1_contk to CP4_contk thus calculated. In FIG. 18, symbols “CP5R”, “CP5G” and “CP5B” indicate the end points of the input-output curves used for the correction calculations of the R data DIN R, G data DIN G, and B data DIN B, respectively. The correction point data set CP_selk is calculated by adding the values of the above-described correction point control data CP1_contk to CP4_contk to the values of the correction point data CP1 to CP4 of the correction point data sets CP_CRk.
As a whole, the correction point data CP0 to CP4 of the correction point data set CP_selk may be calculated in one embodiment by the following expressions:
CP0_selR=CP0_L R+CP0_contR,  (18a)
CP0_selG=CP0_L G+CP0_contG,  (18b)
CP0_selB=CP0_L B+CP0_contB,  (18c)
CP1_selR=CP1_L R−(D IN MAX−σ2)·ξR+CP1_contR,  ((19 a
CP1_selG=CP1_L G−(D IN MAX−σ2)·ξG+CP1_contG,  (19b)
CP1_selB=CP1_L B−(D IN MAX−σ2)·ξB+CP1_contB,  (19c)
CP2_selR=CP2_L R+CP2_contR,  (20a)
CP2_selG=CP2_L G+CP2_contG,  (20b)
CP2_selB=CP2_L B+CP2_contB,  (20c)
CP3_selR=CP3_L R+CP3_contR,  (21a)
CP3_selG=CP3_L G+CP3_contG,  (21b)
CP3_selB=CP3_L B+CP3_contB,  (21c)
CP4_selR=CP4_L R+(D IN MAX−σ2)·ξR+CP4_contR,  (22a)
CP4_selG=CP4_L G+(D IN MAX−σ2)·ξG+CP4_contG, and  (22b)
CP4_selB=CP4_L B+(D IN MAX−σ2)·ξB+CP4_contB,  (22c)
where σ2 is the variance of the luminance values indicated by the variance data DCHR _ σ2, and DIN MAX is the allowed maximum value of the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN. ξR, ξG, and ξB are predetermined proportionality constants. ξR, ξG, and ξB may be equal to each other or different. CPα_selk is the value of the correction point data CPα of the correction point data set CP_selk and CPα_Lk is the value of the correction point data CPα of the correction point data set CP_Lk.
The correction point data CP5 of the correction point data set CP_selk is determined as equal to the correction point data CP5 of the correction point data set CP_Lk.
Since the correction point data CP5 are unchanged in the selection of the correction point data set CP_Lk or calculation of the correction point data set CP_Lk through an interpolation at step S23, the following expression (23) holds:
CP5_selR=CP5_selG=CP5_selB=CP5(=D OUT MAX).   (23)
It should be noted that, since the starting point control data sets CP0_contR, CP0_contG, and CP0_contB are calculated for each pixel 9, the correction point data sets CP_selR, CP_selG, and CP_selB are also calculated for each pixel 9. In other words, correction point data CP_selk associated with a certain pixel 9 in a certain frame image are calculated from the correction point data set CP_Lk calculated for the frame image, the variance σ2 of the luminance values described in the variance data DCHR _ σ2 calculated for the frame image, and the starting point control data CP0_contR, CP0_contG and CP0_contB calculated for the pixel 9.
When a common gamma value γ_PIXEL is calculated for the R data DIN R, G data DIN G and B data DIN B of input image data DIN and a common correction point data set CP_L is selected or calculated, the modification of the correction point data set CP_L on the basis of the variance data DCHR _ σ2 is commonly performed for the R data DIN R, G data DIN G and B data DIN B, while the modification of the correction point data set CP_L on the basis of the starting point data CP0_contR, CP0_contG, and CP0_contB are individually performed for the R data DIN R, G data DIN G and B data DIN B. In this case, the correction point data CP0 to CP4 of the correction point data sets CP_selR, CP_selG and CP_selB are resultantly calculated in accordance with the following expressions:
CP0_selR=CP0_L+CP0_contR,  (24a)
CP0_selG=CP0_L+CP0_contG,  (24b)
CP0_selB=CP0_L+CP0_contB,  (24c)
CP1_selR=CP1_L−(D IN MAX−σ2)·ξ+CP1_contR,  (25a)
CP1_selG=CP1_L−(D IN MAX−σ2)·ξ+CP1_contG,  (25b)
CP1_selB=CP1_L−(D IN MAX−σ2)·ξ+CP1_contB,  (25c)
CP2_selR=CP2_L+CP2_contR,  (26a)
CP2_selG=CP2_L+CP2_contG,  (26b)
CP2_selB=CP2_L+CP2_contB,  (26c)
CP3_selR=CP3_L+CP3_contR,  (27a)
CP3_selG=CP3_L+CP3_contG,  (27b)
CP3_selB=CP3_L+CP3_contB,  (27c)
CP4_selR=CP4_L+(D IN MAX−σ2)·ξ+CP4_contR,  (28a)
CP4_selG=CP4_L+(D IN MAX−σ2)·ξ+CP4_contG, and  (28b)
CP4_selB=CP4_L+(D IN MAX−σ2)·ξ+CP4_contB.  (28c)
Also in this case, the correction point data CP5 of the correction point data set CP_selk are equal to the correction point data CP5 of the correction point data set CP_Lk, and expression (23) holds.
At step S26, correction calculations are performed on the R data DIN R, G data DIN G and B data DIN B of the input image data DIN associated with each pixel 9 on the basis of the thus-calculated correction point data sets CP_selR, CP_selG and CP_selB associated with each pixel 9, respectively, to thereby generate the R data DOUT R, G data DOUT G and B data DOUT B of the output image data DOUT associated with each pixel 9. The correction calculations are achieved by the approximate gamma correction units 34R, 34G and 34B of the approximate gamma correction circuit 31, respectively.
More specifically, in the correction calculations in the approximate gamma correction circuit 31, the R data DOUT R, G data DOUT G and B data DOUT B of the output image data DOUT are calculated from the R data DIN R, G data DIN G and B data DIN B of the input image data DIN in accordance with the following expressions:
( 1 ) For the case when D IN k < D IN Center and CP 1 > CP 0 D OUT k = 2 ( CP 1 - CP 0 ) · PD INS K 2 + ( CP 3 - CP 0 ) D INS K + CP 0 ( 29 a ) ( 2 ) For the case when D IN k < D IN Center and CP 1 CP 0 D OUT k = 2 ( CP 1 - CP 0 ) · ND INS K 2 + ( CP 3 - CP 0 ) D INS K + CP 0 ( 29 b ) and ( 3 ) For the case when D IN k > D IN Center D OUT k = 2 ( CP 4 - CP 2 ) · ND INS K 2 + ( CP 5 - CP 2 ) D INS K + CP 2 ( 29 c )
It should be noted that CP0 to CP5 recited in expressions (29a) to (29c) indicate the correction point data CP0 to CP5 of the correction point data set CP_selk.
In the above, the center data value DIN Center is a value defined by the following expression:
D IN Center =D IN MAX/2,  (29d)
where DIN MAX is the allowed maximum value and K is the parameter given by the above-described expression (13a). Furthermore, DINS, PDINS, and NDINS recited in expressions (29a) to (29c) are values defined as follows:
(a) DINS
DINS is a value which depends on the input image data DIN k; DINS is given by the following expressions (30a) and (30b):
D INS =D IN k (for D IN k <D IN Center)  (30a)
D INS =D IN k+1−K (for D IN k >D IN Center)  (30b)
(b) PDINS
PDINS is defined by the following expression (31a) with a parameter R defined by expression (31b):
PD INS=(K−RR  (31a)
R=K 1/2 ·D INS 1/2  (31b)
As understood from expressions (30a), (30b) and (31b), the parameter R is proportional to a square root of input image data DIN k and therefore PDINS is a value calculated by an expression including a term proportional to a square root of DIN k and a term proportional to DIN k (or one power of DIN k).
(c) NDINS
NDINS is given by the following expression (32):
ND INS=(K−D INSD INS.  (32)
As understood from expressions (30a), (30b) and (32), NDINS is a value calculated by an expression including a term proportional to a square of DIN k.
FIG. 19 illustrates the relations between the R data DIN R, G data DIN G, and B data DIN B of the input image data DIN and the R data DOUT R, G data DOUT G, and B data DOUT B of the output image data DOUT, when the R data DOUT R, G data DOUT G, and B data DOUT B of the output image data DOUT are calculated as described above. In FIG. 19, at least one of the starting point control data CP0_contR, CP0_contG, and CP0_contB is positive and at least another one is negative. In view of this fact, it would be easily understood from FIG. 19 that the above-described processing effectively enhances the saturation.
The output image data DOUT, which are calculated by the approximate gamma correction circuit 31 with the above-described series of expressions, are forwarded to the color reduction circuit 23. The color reduction circuit 23 performs a color reduction on the output image data DOUT to generate the color-reduced image data DOUT _ D. The color-reduced image data DOUT _ D are forwarded to the data line drive circuit 26 via the latch circuit 24 and the data lines 8 of the LCD panel 2 are driven in response to the color-reduced image data DOUT _ D.
The above-described saturation enhancement processing of this embodiment effectively achieves a saturation enhancement with simple processing in which the positions of the input-output curves are adjusted. In addition, in the present embodiment, saturation enhancement and contrast enhancement are concurrently achieved in the approximate gamma correction circuit 31 and this effectively reduces the circuit size (for example, compared to the system in which contrast enhancement and saturation enhancement are performed in series as illustrated in FIG. 1B).
Although embodiments of the present invention are specifically described above, the present invention is not limited to the above-described embodiments; a person skilled in the art would appreciate that the present invention may be implemented with various modifications. Although the above-described embodiments recite the liquid crystal display device 1 including the LCD panel 2, the saturation enhancement and contrast enhancement recited in the above-descried embodiments may be implemented in an image processing apparatus in general. It should be also noted that the present invention is applicable to various panel display devices including different display panels (for example, a display device including an OLED (organic light emitting diode) display panel).

Claims (20)

What is claimed is:
1. A display driver, comprising:
processing circuitry configured to:
control a starting point of an input-output curve between input color data and output color data for performing correction calculations, the starting point of the input-output curve corresponding to an allowed minimum value of the input color data; and
enhance a saturation of received image data by performing a correction calculation for at least one color of a plurality of colors of the image data; and
drive circuitry configured to drive a display panel in response to the enhanced-saturation image data.
2. The display driver according to claim 1, wherein the processing circuitry is further configured to, for each color of the plurality of colors of the image data, control a respective starting point of a respective input-output curve for the color.
3. The display driver according to claim 1, wherein the input-output curve is linear.
4. The display driver according to claim 1, wherein the correction calculation includes a linear function correction based on a linear input-output curve.
5. The display driver according to claim 1, wherein the correction calculation includes an approximate gamma correction based on an approximate expression of gamma correction on the image data.
6. The display driver according to claim 5, wherein the input-output curve comprises a gamma curve modified to achieve contrast enhancement.
7. The display driver according to claim 6, wherein the processing circuitry is further configured to control a shape and the starting point of the gamma curve.
8. The display driver according to claim 6, wherein the processing circuitry is further configured to control one or more points of the gamma curve between the starting point and an end point of the gamma curve.
9. The display driver according to claim 1, wherein the processing circuitry is further configured to enhance a contrast of the image data concurrently with the saturation enhancement.
10. The display driver according to claim 1, wherein the processing circuitry is further configured to enhance a contrast of the image data after the saturation enhancement.
11. The display driver according to claim 1, wherein the plurality of colors of the image data include red, green, and blue.
12. An image processing apparatus, comprising:
starting point control circuitry configured to control a starting point of an input-output curve between input color data and output color data for performing correction calculations, the starting point of the input-output curve corresponding to an allowed minimum value of the input color data; and
correction circuitry configured to enhance a saturation of received image data by performing a correction calculation for at least one color of a plurality of colors of the image data.
13. The image processing apparatus of claim 12, wherein the starting point control circuitry is further configured to, for each color of the plurality of colors of the image data, control a respective starting point of a respective input-output curve for the color.
14. The image processing apparatus of claim 12, wherein the correction calculation includes a linear function correction based on a linear input-output curve.
15. The image processing apparatus of claim 12, wherein the correction calculation includes an approximate gamma correction based on an approximate expression of gamma correction on the image data.
16. The image processing apparatus of claim 12, wherein the correction circuitry is further configured to enhance a contrast of the image data concurrently with the saturation enhancement.
17. The image processing apparatus of claim 12, wherein the correction circuitry is further configured to enhance a contrast of the image data after the saturation enhancement.
18. A method, comprising:
controlling a starting point of an input-output curve between input color data and output color data for performing correction calculations, the starting point of the input-output curve corresponding to an allowed minimum value of the input color data; and
generating, using the starting point, saturation-enhanced image data by performing a correction calculation for at least one color of a plurality of colors of received image data.
19. The method of claim 18, wherein performing the correction calculation further comprises:
enhancing a contrast of the image data concurrently with the saturation enhancement.
20. The method of claim 18, wherein performing the correction calculation further comprises:
enhancing a contrast of the image data after the saturation enhancement.
US15/886,565 2014-02-10 2018-02-01 Display device, display panel driver, image processing apparatus and image processing method Active US10380936B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/886,565 US10380936B2 (en) 2014-02-10 2018-02-01 Display device, display panel driver, image processing apparatus and image processing method

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2014023879 2014-02-10
JP2014023879A JP6360321B2 (en) 2014-02-10 2014-02-10 Display device, display panel driver, image processing device, and image processing method
US14/617,764 US9886779B2 (en) 2014-02-10 2015-02-09 Display device, display panel driver, image processing apparatus and image processing method
US15/886,565 US10380936B2 (en) 2014-02-10 2018-02-01 Display device, display panel driver, image processing apparatus and image processing method

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/617,764 Continuation US9886779B2 (en) 2014-02-10 2015-02-09 Display device, display panel driver, image processing apparatus and image processing method

Publications (2)

Publication Number Publication Date
US20180158214A1 US20180158214A1 (en) 2018-06-07
US10380936B2 true US10380936B2 (en) 2019-08-13

Family

ID=53775423

Family Applications (2)

Application Number Title Priority Date Filing Date
US14/617,764 Active 2035-05-23 US9886779B2 (en) 2014-02-10 2015-02-09 Display device, display panel driver, image processing apparatus and image processing method
US15/886,565 Active US10380936B2 (en) 2014-02-10 2018-02-01 Display device, display panel driver, image processing apparatus and image processing method

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US14/617,764 Active 2035-05-23 US9886779B2 (en) 2014-02-10 2015-02-09 Display device, display panel driver, image processing apparatus and image processing method

Country Status (3)

Country Link
US (2) US9886779B2 (en)
JP (1) JP6360321B2 (en)
CN (2) CN110299099B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11145247B2 (en) * 2018-11-20 2021-10-12 Chengdu Boe Optoelectronics Technology Co., Ltd. Device, system and method for display gamma correction
US20240233595A1 (en) * 2023-01-10 2024-07-11 Lg Electronics Inc. Display device and operation method thereof

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6360321B2 (en) * 2014-02-10 2018-07-18 シナプティクス・ジャパン合同会社 Display device, display panel driver, image processing device, and image processing method
JP6611494B2 (en) * 2015-07-08 2019-11-27 キヤノン株式会社 Image display apparatus and control method thereof
JP6817789B2 (en) * 2016-06-10 2021-01-20 ラピスセミコンダクタ株式会社 Display driver and semiconductor device
US10657873B2 (en) * 2017-01-12 2020-05-19 Synaptics Japan Gk System and method for subpixel rendering and display driver
KR20180092330A (en) * 2017-02-08 2018-08-20 삼성디스플레이 주식회사 Display apparatus and method of driving the same
US10176761B2 (en) 2017-02-23 2019-01-08 Synaptics Incorporated Compressed data transmission in panel display system
US10755392B2 (en) * 2017-07-13 2020-08-25 Mediatek Inc. High-dynamic-range video tone mapping
US11270662B2 (en) * 2020-01-21 2022-03-08 Synaptics Incorporated Device and method for brightness control of display device based on display brightness value encoding parameters beyond brightness
KR20230060767A (en) * 2021-10-28 2023-05-08 주식회사 엘엑스세미콘 Display processor ans method for processing image data for driving display panel
CN114155817B (en) * 2021-10-29 2022-11-01 江苏泽景汽车电子股份有限公司 Color correction method of HUD display system
CN114067275A (en) * 2021-11-26 2022-02-18 新疆爱华盈通信息技术有限公司 Target object reminding method and system in monitoring scene and electronic equipment

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH06339017A (en) 1993-05-31 1994-12-06 Nec Corp Saturation emphasizing method and device for color picture
US20020036716A1 (en) 2000-03-14 2002-03-28 Keiichi Ito Dynamic gamma correction apparatus
JP2003052050A (en) 2001-05-30 2003-02-21 Sharp Corp Color display
US20070268524A1 (en) * 2006-05-17 2007-11-22 Nec Electronics Corporation Display device, display panel driver and method of driving display panel
US20070273776A1 (en) 2006-05-29 2007-11-29 Samsung Techwin Co., Ltd. Apparatus and method of gamma correction in digital image processing device
US20090245676A1 (en) * 2003-11-10 2009-10-01 Huaya Microelectronics, Inc. Contour enhancment unit and method to enhance the contours of an image
US20100079479A1 (en) 2008-09-29 2010-04-01 Renesas Technology Corp. Display drive circuit
US9886779B2 (en) * 2014-02-10 2018-02-06 Synaptics Japan Gk Display device, display panel driver, image processing apparatus and image processing method

Family Cites Families (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0281594A (en) * 1988-09-19 1990-03-22 Fuji Photo Film Co Ltd Color adjusting device for digital picture
JPH02271793A (en) * 1989-04-12 1990-11-06 Mitsubishi Electric Corp Full color display device for liquid crystal display device
JP3408316B2 (en) * 1994-04-18 2003-05-19 日本フィリップス株式会社 Knee circuit
JP3592930B2 (en) * 1998-05-01 2004-11-24 アルパイン株式会社 Automatic dimmer for LCD display
JP4244105B2 (en) * 1999-11-15 2009-03-25 富士フイルム株式会社 Image processing method and apparatus, and recording medium
JP4583666B2 (en) * 2000-06-29 2010-11-17 Jx日鉱日石エネルギー株式会社 Fuel for fuel cell system
KR100500688B1 (en) * 2001-12-31 2005-07-12 비오이 하이디스 테크놀로지 주식회사 Method of driving driver ic in twisted nematic thin film transistor lcd
JP2007311889A (en) * 2006-05-16 2007-11-29 Sony Corp Image correction circuit, image correction method, and image display apparatus
JP5012195B2 (en) * 2006-05-17 2012-08-29 ソニー株式会社 Image correction circuit, image correction method, and image display apparatus
US8369646B2 (en) * 2006-05-17 2013-02-05 Sony Corporation Image correction circuit, image correction method and image display
CN1874527A (en) * 2006-06-09 2006-12-06 北京中星微电子有限公司 Gamma correction unit, and method and equipment for implementing gamma correction
JP4996309B2 (en) * 2007-04-03 2012-08-08 キヤノン株式会社 Color image processing system and method
JP2009237373A (en) * 2008-03-27 2009-10-15 Casio Hitachi Mobile Communications Co Ltd Image quality adjusting device, portable terminal device, and program
US8279235B2 (en) * 2008-07-03 2012-10-02 Wintek Corporation Color correction method and color correcting integrated chip
JP5241031B2 (en) * 2009-12-08 2013-07-17 ルネサスエレクトロニクス株式会社 Display device, display panel driver, and image data processing device
JP5247671B2 (en) * 2009-12-22 2013-07-24 ルネサスエレクトロニクス株式会社 Display data correction device, display panel driver using the same, and display device
US9478173B2 (en) * 2010-08-30 2016-10-25 Qualcomm Incorporated Adaptive color correction for display with backlight modulation
JP5728244B2 (en) * 2011-02-16 2015-06-03 キヤノン株式会社 Image processing apparatus and image processing method
KR20120118383A (en) * 2011-04-18 2012-10-26 삼성전자주식회사 Image compensation device, image processing apparatus and methods thereof
KR101904339B1 (en) * 2012-04-17 2018-10-08 삼성디스플레이 주식회사 Organic light emittng device and metheod of configuring gamma set of the same

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH06339017A (en) 1993-05-31 1994-12-06 Nec Corp Saturation emphasizing method and device for color picture
US20020036716A1 (en) 2000-03-14 2002-03-28 Keiichi Ito Dynamic gamma correction apparatus
JP2003052050A (en) 2001-05-30 2003-02-21 Sharp Corp Color display
US20090245676A1 (en) * 2003-11-10 2009-10-01 Huaya Microelectronics, Inc. Contour enhancment unit and method to enhance the contours of an image
US20070268524A1 (en) * 2006-05-17 2007-11-22 Nec Electronics Corporation Display device, display panel driver and method of driving display panel
US7973973B2 (en) 2006-05-17 2011-07-05 Renesas Electronics Corporation Display device, display panel driver and method of driving display panel
US20070273776A1 (en) 2006-05-29 2007-11-29 Samsung Techwin Co., Ltd. Apparatus and method of gamma correction in digital image processing device
US20100079479A1 (en) 2008-09-29 2010-04-01 Renesas Technology Corp. Display drive circuit
US9886779B2 (en) * 2014-02-10 2018-02-06 Synaptics Japan Gk Display device, display panel driver, image processing apparatus and image processing method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
JP Office Action Application No. 2014-023879, dated Dec. 6, 2017 consists of 4 pages, English translation 4 pages.

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11145247B2 (en) * 2018-11-20 2021-10-12 Chengdu Boe Optoelectronics Technology Co., Ltd. Device, system and method for display gamma correction
US20240233595A1 (en) * 2023-01-10 2024-07-11 Lg Electronics Inc. Display device and operation method thereof

Also Published As

Publication number Publication date
US20150228215A1 (en) 2015-08-13
JP6360321B2 (en) 2018-07-18
US9886779B2 (en) 2018-02-06
CN110299099A (en) 2019-10-01
JP2015152646A (en) 2015-08-24
CN104835438B (en) 2019-07-23
US20180158214A1 (en) 2018-06-07
CN110299099B (en) 2023-05-16
CN104835438A (en) 2015-08-12

Similar Documents

Publication Publication Date Title
US10380936B2 (en) Display device, display panel driver, image processing apparatus and image processing method
US9837045B2 (en) Device and method for color adjustment and gamma correction and display panel driver using the same
KR102528560B1 (en) Display driver, display system and control method of display driver
US10657870B2 (en) Method and device for display color adjustment
US9524664B2 (en) Display device, display panel driver and drive method of display panel
KR101492712B1 (en) Organic light emitting diode display device and method for driving the same
CN105280146B (en) Control signal generating circuit, video display devices and control signal creating method
US9779514B2 (en) Display device, display panel driver and driving method of display panel
US9183797B2 (en) Display device and control method for display device
US11948522B2 (en) Display device with light adjustment for divided areas using an adjustment coefficient
KR102510573B1 (en) Transparent display device and method for driving the same
US10373584B2 (en) Device and method for display color adjustment
US9734772B2 (en) Display device
KR101968911B1 (en) organic light-emitting dIODE DISPLAY DEVICE AND DRIVING METHOD THEREOF
KR101120313B1 (en) Display driving device
KR20170080356A (en) Displya device, method for converting datat and controller therefor
KR102379774B1 (en) Image processing method, image processing circuit and display device using the same
KR101947804B1 (en) Image quality processor, display device including the same, and image quality processing method
KR20180071810A (en) Transparent display device and method for driving the same
WO2006085508A1 (en) Display gradation voltage setting method, display driving method, program, and display

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

AS Assignment

Owner name: SYNAPTICS DISPLAY DEVICES KK, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:NOSE, TAKASHI;FURIHATA, HIROBUMI;SUGIYAMA, AKIO;SIGNING DATES FROM 20141215 TO 20141218;REEL/FRAME:044823/0620

Owner name: SYNAPTICS DISPLAY DEVICES GK, JAPAN

Free format text: CHANGE OF NAME;ASSIGNOR:SYNAPTICS DISPLAY DEVICES KK;REEL/FRAME:045239/0919

Effective date: 20150415

Owner name: SYNAPTICS JAPAN GK, JAPAN

Free format text: CHANGE OF NAME;ASSIGNOR:SYNAPTICS DISPLAY DEVICES GK;REEL/FRAME:045240/0001

Effective date: 20160701

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

AS Assignment

Owner name: SYNAPTICS INCORPORATED, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SYNAPTICS JAPAN GK;REEL/FRAME:067793/0211

Effective date: 20240617