/* GStreamer * Copyright (C) <1999> Erik Walthinsen * Copyright (C) <2003> David Schleef * Copyright (C) 2003 Arwed v. Merkatz * Copyright (C) 2006 Mark Nauwelaerts * Copyright (C) 2010 United States Government, Joshua M. Doe * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ /** * SECTION:element-videolevels * * Convert grayscale video from one bpp/depth combination to another. * * * Example launch line * |[ * gst-launch videotestsrc ! videolevels ! ffmpegcolorspace ! autovideosink * ]| * */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include "gstvideolevels.h" #include #include #include /* GstVideoLevels signals and args */ enum { /* FILL ME */ LAST_SIGNAL }; enum { PROP_0, PROP_LOWIN, PROP_HIGHIN, PROP_LOWOUT, PROP_HIGHOUT, PROP_AUTO, PROP_INTERVAL/*, PROP_NPIXSAT_LOW, PROP_NPIXSAT_HIGH, PROP_ROI_X, PROP_ROI_Y, PROP_ROI_WIDTH, PROP_ROI_HEIGHT */ }; #define DEFAULT_PROP_LOWIN 0.0 #define DEFAULT_PROP_HIGHIN 1.0 #define DEFAULT_PROP_LOWOUT 0.0 #define DEFAULT_PROP_HIGHOUT 1.0 #define DEFAULT_PROP_AUTO 0 #define DEFAULT_PROP_INTERVAL (GST_SECOND / 2) #define DEFAULT_PROP_ROW_X 0 #define DEFAULT_PROP_ROW_Y 0 #define DEFAULT_PROP_ROW_WIDTH 0 #define DEFAULT_PROP_ROW_HEIGHT 0 static const GstElementDetails videolevels_details = GST_ELEMENT_DETAILS ("Video videolevels adjustment", "Filter/Effect/Video", "Adjusts videolevels on a video stream", "Joshua Doe finalize (object); } /** * gst_videolevels_class_init: * @object: #GstVideoLevelsClass. * */ static void gst_videolevels_class_init (GstVideoLevelsClass * object) { GObjectClass *obj_class = G_OBJECT_CLASS (object); GstBaseTransformClass *trans_class = GST_BASE_TRANSFORM_CLASS (object); GST_DEBUG ("class init"); /* Register GObject vmethods */ obj_class->finalize = GST_DEBUG_FUNCPTR (gst_videolevels_finalize); obj_class->set_property = GST_DEBUG_FUNCPTR (gst_videolevels_set_property); obj_class->get_property = GST_DEBUG_FUNCPTR (gst_videolevels_get_property); /* Install GObject properties */ g_object_class_install_property (obj_class, PROP_LOWIN, g_param_spec_double ("low_in", "Lower Input Level", "Lower Input Level", -G_MINDOUBLE, G_MAXDOUBLE, DEFAULT_PROP_LOWIN, G_PARAM_READWRITE)); g_object_class_install_property (obj_class, PROP_HIGHIN, g_param_spec_double ("upper_in", "Upper Input Level", "Upper Input Level", -G_MINDOUBLE, G_MAXDOUBLE, DEFAULT_PROP_HIGHIN, G_PARAM_READWRITE)); g_object_class_install_property (obj_class, PROP_LOWOUT, g_param_spec_double ("low_out", "Lower Output Level", "Lower Output Level", -G_MINDOUBLE, G_MAXDOUBLE, DEFAULT_PROP_LOWOUT, G_PARAM_READWRITE)); g_object_class_install_property (obj_class, PROP_HIGHOUT, g_param_spec_double ("upper_out", "Upper Output Level", "Upper Output Level", -G_MINDOUBLE, G_MAXDOUBLE, DEFAULT_PROP_HIGHOUT, G_PARAM_READWRITE)); g_object_class_install_property (obj_class, PROP_AUTO, g_param_spec_int ("auto", "Auto Adjust", "Auto adjust contrast (0): off, (1): single-shot, (2): continuous", 0, 2, DEFAULT_PROP_AUTO, G_PARAM_READWRITE)); g_object_class_install_property (obj_class, PROP_INTERVAL, g_param_spec_uint64 ("interval", "Interval", "Interval of time between adjustments (in nanoseconds)", 1, G_MAXUINT64, DEFAULT_PROP_INTERVAL, G_PARAM_READWRITE)); /* Register GstBaseTransform vmethods */ trans_class->transform_caps = GST_DEBUG_FUNCPTR (gst_videolevels_transform_caps); trans_class->set_caps = GST_DEBUG_FUNCPTR (gst_videolevels_set_caps); trans_class->transform = GST_DEBUG_FUNCPTR (gst_videolevels_transform); trans_class->transform_ip = GST_DEBUG_FUNCPTR (gst_videolevels_transform_ip); trans_class->get_unit_size = GST_DEBUG_FUNCPTR (gst_videolevels_get_unit_size); /* simply pass the data through if in/out caps are the same */ trans_class->passthrough_on_same_caps = TRUE; } /** * gst_videolevels_init: * @videolevels: GstVideoLevels * @g_class: GstVideoLevelsClass * * Initialize the new element */ static void gst_videolevels_init (GstVideoLevels * videolevels, GstVideoLevelsClass * g_class) { GST_DEBUG_OBJECT (videolevels, "init class instance"); gst_videolevels_reset (videolevels); } /** * gst_videolevels_set_property: * @object: #GObject * @prop_id: guint * @value: #GValue * @pspec: #GParamSpec * */ static void gst_videolevels_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { GstVideoLevels *videolevels = GST_VIDEOLEVELS (object); GST_DEBUG ("setting property %s", pspec->name); switch (prop_id) { case PROP_LOWIN: videolevels->lower_input = g_value_get_double (value); //gst_videolevels_calculate_tables (videolevels); break; case PROP_HIGHIN: videolevels->upper_input = g_value_get_double (value); //gst_videolevels_calculate_tables (videolevels); break; case PROP_LOWOUT: videolevels->lower_output = g_value_get_double (value); //gst_videolevels_calculate_tables (videolevels); break; case PROP_HIGHOUT: videolevels->upper_output = g_value_get_double (value); //gst_videolevels_calculate_tables (videolevels); break; case PROP_AUTO: videolevels->auto_adjust = g_value_get_int (value); break; case PROP_INTERVAL: videolevels->interval = g_value_get_uint64 (value); videolevels->last_auto_timestamp = GST_CLOCK_TIME_NONE; break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } /** * gst_videolevels_get_property: * @object: #GObject * @prop_id: guint * @value: #GValue * @pspec: #GParamSpec * */ static void gst_videolevels_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstVideoLevels *videolevels = GST_VIDEOLEVELS (object); GST_DEBUG ("getting property %s", pspec->name); switch (prop_id) { case PROP_LOWIN: g_value_set_double (value, videolevels->lower_input); break; case PROP_HIGHIN: g_value_set_double (value, videolevels->upper_input); break; case PROP_LOWOUT: g_value_set_double (value, videolevels->lower_output); break; case PROP_HIGHOUT: g_value_set_double (value, videolevels->upper_output); break; case PROP_AUTO: g_value_set_int (value, videolevels->auto_adjust); break; case PROP_INTERVAL: g_value_set_uint64 (value, videolevels->interval); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } /************************************************************************/ /* GstBaseTransform vmethod implementations */ /************************************************************************/ /** * gst_videolevels_transform_caps: * @base: #GstBaseTransform * @direction: #GstPadDirection * @caps: #GstCaps * * Given caps on one side, what caps are allowed on the other * * Returns: #GstCaps allowed on other pad */ static GstCaps * gst_videolevels_transform_caps (GstBaseTransform * base, GstPadDirection direction, GstCaps * caps) { GstVideoLevels *videolevels; GstCaps *newcaps; GstStructure *structure; videolevels = GST_VIDEOLEVELS (base); GST_DEBUG_OBJECT (caps, "transforming caps (from)"); newcaps = gst_caps_copy (caps); /* finish settings caps of the opposite pad */ if (direction == GST_PAD_SINK) { GST_DEBUG ("Pad direction is sink"); gst_caps_set_simple (newcaps, "bpp", G_TYPE_INT, 8, "depth", G_TYPE_INT, 8, NULL); structure = gst_caps_get_structure (newcaps, 0); gst_structure_remove_field (structure, "endianness"); gst_structure_remove_field (structure, "signed"); } else { GValue endianness = {0}; GValue signed_list = {0}; GValue ival = {0}; GST_DEBUG ("Pad direction is src"); gst_caps_set_simple (newcaps, "bpp", GST_TYPE_INT_RANGE, 1, 16, "depth", G_TYPE_INT, 16, NULL); structure = gst_caps_get_structure (newcaps, 0); /* add BIG/LITTLE endianness to caps */ g_value_init (&ival, G_TYPE_INT); g_value_init (&endianness, GST_TYPE_LIST); g_value_set_int (&ival, G_LITTLE_ENDIAN); gst_value_list_append_value (&endianness, &ival); g_value_set_int (&ival, G_BIG_ENDIAN); gst_value_list_append_value (&endianness, &ival); gst_structure_set_value (structure, "endianness", &endianness); /* add signed/unsigned to caps */ g_value_init (&signed_list, GST_TYPE_LIST); g_value_set_int (&ival, FALSE); gst_value_list_append_value (&signed_list, &ival); g_value_set_int (&ival, TRUE); gst_value_list_append_value (&signed_list, &ival); gst_structure_set_value (structure, "signed", &signed_list); } GST_DEBUG_OBJECT (newcaps, "allowed caps are"); return newcaps; } /** * gst_videolevels_set_caps: * base: #GstBaseTransform * incaps: #GstCaps * outcaps: #GstCaps * * Notification of the actual caps set. * * Returns: TRUE on success */ static gboolean gst_videolevels_set_caps (GstBaseTransform * base, GstCaps * incaps, GstCaps * outcaps) { GstVideoLevels *levels; GstStructure *structure; gboolean res; levels = GST_VIDEOLEVELS (base); GST_DEBUG_OBJECT (levels, "set_caps: in %" GST_PTR_FORMAT " out %" GST_PTR_FORMAT, incaps, outcaps); GST_DEBUG_OBJECT (incaps, "incaps"); GST_DEBUG_OBJECT (outcaps, "outcaps"); /* retrieve input caps info */ structure = gst_caps_get_structure (incaps, 0); res = gst_structure_get (structure, "width", G_TYPE_INT, &levels->width, "height", G_TYPE_INT, &levels->height, "bpp", G_TYPE_INT, &levels->bpp_in, "depth", G_TYPE_INT, &levels->depth_in, "endianness", G_TYPE_INT, &levels->endianness_in, NULL); if (!res) return FALSE; /* retrieve input caps sign, or set to false if field doesn't exist */ if (!gst_structure_get (structure, "signed", G_TYPE_BOOLEAN, &levels->is_signed_in)) levels->is_signed_in = FALSE; /* retrieve output bpp and depth */ structure = gst_caps_get_structure (outcaps, 0); res = gst_structure_get (structure, "bpp", G_TYPE_INT, &levels->bpp_out, "depth", G_TYPE_INT, &levels->depth_out, NULL); if (!res) return FALSE; levels->stride_in = GST_ROUND_UP_4 (levels->width * levels->depth_in/8); levels->stride_out = GST_ROUND_UP_4 (levels->width * levels->depth_out/8); levels->lower_input = 0.0; levels->upper_input = (1 << levels->bpp_in) - 1; levels->lower_output = 0.0; levels->upper_output = (1 << levels->bpp_out) - 1; //gst_videolevels_calculate_tables (levels); return res; } /** * gst_videolevels_get_unit_size: * @base: #GstBaseTransform * @caps: #GstCaps * @size: guint size of unit (one frame for video) * * Tells GstBaseTransform the size in bytes of an output frame from the given * caps. * * Returns: TRUE on success */ static gboolean gst_videolevels_get_unit_size (GstBaseTransform * base, GstCaps * caps, guint * size) { GstStructure *structure; gint width; gint height; gint depth; structure = gst_caps_get_structure (caps, 0); /* get proposed caps width, height, and depth to determine frame size */ if (gst_structure_get_int (structure, "width", &width) && gst_structure_get_int (structure, "height", &height) && gst_structure_get_int (structure, "depth", &depth)) { guint stride = GST_ROUND_UP_4 (width*depth/8); /* need 4-byte alignment */ *size = stride * height; GST_DEBUG ("Get unit size %dx%d, stride %u, %u bytes", width, height, stride, *size); return TRUE; } GST_ELEMENT_ERROR (base, CORE, NEGOTIATION, (NULL), ("Incomplete caps, some required field missing")); return FALSE; } /** * gst_videolevels_transform: * @base: #GstBaseTransform * @inbuf: #GstBuffer * @outbuf: #GstBuffer * * Transforms input buffer to output buffer. * * Returns: GST_FLOW_OK on success */ static GstFlowReturn gst_videolevels_transform (GstBaseTransform * base, GstBuffer * inbuf, GstBuffer * outbuf) { GstVideoLevels *videolevels = GST_VIDEOLEVELS (base); gpointer input; gpointer output; gboolean ret; GstClockTimeDiff elapsed; /* We need to lock our videolevels params to prevent segfaults */ GST_BASE_TRANSFORM_LOCK (videolevels); input = GST_BUFFER_DATA (inbuf); output = GST_BUFFER_DATA (outbuf); if (videolevels->auto_adjust == 1) { gst_videolevels_auto_adjust (videolevels, input); videolevels->auto_adjust = 0; } else if (videolevels->auto_adjust == 2) { elapsed = GST_CLOCK_DIFF (videolevels->last_auto_timestamp, inbuf->timestamp); if (videolevels->last_auto_timestamp == GST_CLOCK_TIME_NONE || elapsed >= (GstClockTimeDiff)videolevels->interval || elapsed < 0) { gst_videolevels_auto_adjust (videolevels, input); videolevels->last_auto_timestamp = GST_BUFFER_TIMESTAMP (inbuf); } } ret = gst_videolevels_do_levels (videolevels, input, output); GST_BASE_TRANSFORM_UNLOCK (videolevels); if (ret) return GST_FLOW_OK; else return GST_FLOW_ERROR; } GstFlowReturn gst_videolevels_transform_ip( GstBaseTransform * base, GstBuffer * buf ) { return GST_FLOW_OK; } /************************************************************************/ /* GstVideoLevels method implementations */ /************************************************************************/ /** * gst_videolevels_reset: * @videolevels: #GstVideoLevels * * Reset instance variables and free memory */ static void gst_videolevels_reset(GstVideoLevels* videolevels) { videolevels->width = 0; videolevels->height = 0; videolevels->stride_in = 0; videolevels->bpp_in = 0; videolevels->depth_in = 0; videolevels->endianness_in = 0; videolevels->is_signed_in = FALSE; videolevels->stride_out = 0; videolevels->bpp_out = 0; videolevels->depth_out = 0; videolevels->endianness_out = 0; videolevels->lower_input = DEFAULT_PROP_LOWIN; videolevels->upper_input = DEFAULT_PROP_HIGHIN; videolevels->lower_output = DEFAULT_PROP_LOWOUT; videolevels->upper_output = DEFAULT_PROP_HIGHOUT; g_free (videolevels->lookup_table); videolevels->lookup_table = NULL; videolevels->auto_adjust = DEFAULT_PROP_AUTO; videolevels->interval = DEFAULT_PROP_INTERVAL; videolevels->last_auto_timestamp = GST_CLOCK_TIME_NONE; videolevels->lower_pix_sat = 0.01f; videolevels->upper_pix_sat = 0.01f; videolevels->nbins = 4096; g_free (videolevels->histogram); videolevels->histogram = NULL; } /** * gst_videolevels_calculate_tables * @videolevels: #GstVideoLevels * * Calculate lookup tables based on input and output levels */ static void gst_videolevels_calculate_tables (GstVideoLevels * videolevels) { gint i; guint16 loIn, hiIn; guint8 loOut, hiOut; gdouble slope; guint8 * lut; GST_DEBUG ("calculating lookup table"); if (!videolevels->lookup_table) { videolevels->lookup_table = g_new (guint8, 65536); } lut = (guint8*) videolevels->lookup_table; loIn = (guint16) (videolevels->lower_input * G_MAXUINT16); hiIn = (guint16) (videolevels->upper_input * G_MAXUINT16); loOut = (guint8) (videolevels->lower_output * G_MAXUINT8); hiOut = (guint8) (videolevels->upper_output * G_MAXUINT8); GST_DEBUG ("input levels (%.3f, %.3f), output levels (%.3f, %.3f)",videolevels->lower_input,videolevels->upper_input,videolevels->lower_output,videolevels->upper_output); GST_DEBUG ("input levels (%d, %d), output levels (%d, %d)",loIn,hiIn,loOut,hiOut); if (hiIn == loIn) slope = 0; else slope = (videolevels->upper_output - videolevels->lower_output) / (videolevels->upper_input - videolevels->lower_input); GST_DEBUG ("slope = %f", slope); if (videolevels->endianness_in == G_BYTE_ORDER) { for (i = 0; i < loIn; i++) lut[i] = loOut; for (i = loIn; i < hiIn; i++) lut[i] = loOut + (guint8) ( (i - loIn) * slope); for (i = hiIn; i < 65536; i++) lut[i] = hiOut; } else { for (i = 0; i < loIn; i++) lut[GUINT16_SWAP_LE_BE(i)] = loOut; for (i = loIn; i < hiIn; i++) lut[GUINT16_SWAP_LE_BE(i)] = loOut + (guint8) ((GUINT16_SWAP_LE_BE(i) - loIn) * slope); for (i = hiIn; i < 65536; i++) lut[GUINT16_SWAP_LE_BE(i)] = hiOut; } for (i = 0; i < 65536; i += 128) GST_DEBUG ("lut[%d / %d] = %d", i, GUINT16_SWAP_LE_BE(i), lut[i]); } #define GINT_CLAMP(x, low, high) ((gint)(CLAMP((x),(low),(high)))) #define GUINT8_CLAMP(x, low, high) ((guint8)(CLAMP((x),(low),(high)))) void gst_videolevels_convert_uint16le_to_uint8(GstVideoLevels * videolevels, guint16 * in, guint8 * out) { gint i; const gint size = videolevels->stride_in / 2 * videolevels->height; gdouble m; gdouble b; const guint8 low = (guint8) videolevels->lower_output; const guint8 high = (guint8) videolevels->upper_output; GST_DEBUG ("Applying linear mapping (%.3f, %.3f) -> (%.3f, %.3f)", videolevels->lower_input, videolevels->upper_input, videolevels->lower_output, videolevels->upper_output); if (videolevels->lower_input == videolevels->upper_input) m = 0; else m = (videolevels->upper_output - videolevels->lower_output) / (videolevels->upper_input - videolevels->lower_input); b = videolevels->lower_output - m * videolevels->lower_input; for (i = 0; i < size; i++) out[i] = GUINT8_CLAMP (m * GUINT16_FROM_LE (in[i]) + b, low, high); } void gst_videolevels_convert_uint16be_to_uint8(GstVideoLevels * videolevels, guint16 * in, guint8 * out) { gint i; const gint size = videolevels->stride_in / 2 * videolevels->height; gdouble m; gdouble b; const guint8 low = (guint8) videolevels->lower_output; const guint8 high = (guint8) videolevels->upper_output; GST_DEBUG ("Applying linear mapping (%.3f, %.3f) -> (%.3f, %.3f)", videolevels->lower_input, videolevels->upper_input, videolevels->lower_output, videolevels->upper_output); if (videolevels->lower_input == videolevels->upper_input) m = 0; else m = (videolevels->upper_output - videolevels->lower_output) / (videolevels->upper_input - videolevels->lower_input); b = videolevels->lower_output - m * videolevels->lower_input; for (i = 0; i < size; i++) out[i] = GUINT8_CLAMP (m * GUINT16_FROM_BE (in[i]) + b, low, high); } /** * gst_videolevels_do_levels * @videolevels: #GstVideoLevels * @indata: input data * @outdata: output data * @size: size of data * * Convert frame using previously calculated LUT * * Returns: TRUE on success */ gboolean gst_videolevels_do_levels (GstVideoLevels * videolevels, gpointer indata, gpointer outdata) { if (videolevels->endianness_in == G_LITTLE_ENDIAN) gst_videolevels_convert_uint16le_to_uint8 (videolevels, indata, outdata); else gst_videolevels_convert_uint16be_to_uint8 (videolevels, indata, outdata); return TRUE; //guint8 * dst = outdata; //guint16 * src = indata; //gint r = 0, c = 0; //guint8 * lut = (guint8 *) videolevels->lookup_table; //GST_DEBUG ("Converting frame using LUT"); // //for (r = 0; r < videolevels->height; r++) { // for (c = 0; c < videolevels->width; c++) { // dst[c+r*videolevels->stride_out] = // lut[src[c+r*videolevels->stride_in/2]]; // } //} return TRUE; } /** * gst_videolevels_calculate_histogram * @videolevels: #GstVideoLevels * @data: input frame data * * Calculate histogram of input frame * * Returns: TRUE on success */ gboolean gst_videolevels_calculate_histogram (GstVideoLevels * videolevels, guint16 * data) { gint * hist; gint nbins = videolevels->nbins; gint r; gint c; gfloat factor; factor = nbins/(gfloat)(1 << videolevels->bpp_in); if (videolevels->histogram == NULL) { GST_DEBUG ("First call, allocate memory for histogram (%d bins)", nbins); videolevels->histogram = g_new (gint, nbins); } hist = videolevels->histogram; /* reset histogram */ memset (hist, 0, sizeof(gint)*nbins); GST_DEBUG ("Calculating histogram"); if (!videolevels->is_signed_in) { if (videolevels->endianness_in == G_BYTE_ORDER) { for (r = 0; r < videolevels->height; r++) { for (c = 0; c < videolevels->width; c++) { /* GST_DEBUG ("(%d, %d) = %d, hist[%d] = %d", r, c, data [c + r * videolevels->stride_in / 2], GINT_CLAMP (data [c + r * videolevels->stride_in / 2] * factor, 0, nbins - 1), hist [GINT_CLAMP (data [c + r * videolevels->stride_in / 2] * factor, 0, nbins - 1)] + 1);*/ hist [GINT_CLAMP (data [c + r * videolevels->stride_in / 2] * factor, 0, nbins - 1)]++; } } } else { for (r = 0; r < videolevels->height; r++) { for (c = 0; c < videolevels->width; c++) { hist [GINT_CLAMP (GUINT16_FROM_BE (data [c + r * videolevels->stride_in / 2]) * factor, 0, nbins - 1)]++; } } } } else { if (videolevels->endianness_in == G_BYTE_ORDER) { for (r = 0; r < videolevels->height; r++) { for (c = 0; c < videolevels->width; c++) { hist [GINT_CLAMP ((data [c + r * videolevels->stride_in / 2] + 32767) * factor, 0, nbins - 1)]++; } } } else { for (r = 0; r < videolevels->height; r++) { for (c = 0; c < videolevels->width; c++) { hist [GINT_CLAMP ((GUINT16_FROM_BE (data [c + r * videolevels->stride_in / 2]) + 32767) * factor, 0, nbins - 1)]++; } } } } //for (r = 0; r < videolevels->nbins; r++ ) { // GST_DEBUG ("hist[%5d] = %d",r, hist[r]); //} return TRUE; } /** * gst_videolevels_auto_adjust * @videolevels: #GstVideoLevels * @data: input frame data * * Calculate lower and upper levels based on the histogram of the frame * * Returns: TRUE on success */ gboolean gst_videolevels_auto_adjust (GstVideoLevels * videolevels, guint16 * data) { guint npixsat; guint sum; gint i; gint size; gdouble min = 0.0; gdouble max = (1 << videolevels->bpp_in) - 1.0; gst_videolevels_calculate_histogram (videolevels, data); size = videolevels->width * videolevels->height; /* pixels to saturate on low end */ npixsat = (guint)(videolevels->lower_pix_sat * size); sum = 0; for (i = 0; i < videolevels->nbins; i++) { sum += videolevels->histogram[i]; if (sum > npixsat) { videolevels->lower_input = CLAMP (i / (gdouble)videolevels->nbins * max, min, max); break; } } /* pixels to saturate on high end */ npixsat = (guint)(videolevels->upper_pix_sat * size); sum = 0; for (i = videolevels->nbins - 1; i >= 0; i--) { sum += videolevels->histogram[i]; if (sum > npixsat) { videolevels->upper_input = CLAMP ((i + 1) / (gdouble)videolevels->nbins * max, min, max); break; } } GST_DEBUG ("Contrast stretch with npixsat=%d, (%.3f, %.3f)", npixsat, videolevels->lower_input, videolevels->upper_input); //gst_videolevels_calculate_tables (videolevels); return TRUE; }