From 95e061d93b2e237f3bd460f9811b05320eb28daa Mon Sep 17 00:00:00 2001 From: diego Date: Tue, 6 Oct 2009 08:45:20 +0000 Subject: K&R coding style and whitespace cosmetics git-svn-id: svn://svn.mplayerhq.hu/mplayer/trunk@29755 b3059339-0415-0410-9bf9-f77b7e298cf2 --- libvo/vo_vdpau.c | 332 ++++++++++++++++++++++++++++--------------------------- 1 file changed, 170 insertions(+), 162 deletions(-) diff --git a/libvo/vo_vdpau.c b/libvo/vo_vdpau.c index fffb1b010e..3a6df1195a 100644 --- a/libvo/vo_vdpau.c +++ b/libvo/vo_vdpau.c @@ -379,7 +379,8 @@ static int win_x11_init_vdpau_procs(void) for (dsc = vdp_func; dsc->pointer; dsc++) { vdp_st = vdp_get_proc_address(vdp_device, dsc->id, dsc->pointer); if (vdp_st != VDP_STATUS_OK) { - mp_msg(MSGT_VO, MSGL_ERR, "[vdpau] Error when calling vdp_get_proc_address(function id %d): %s\n", dsc->id, vdp_get_error_string ? vdp_get_error_string(vdp_st) : "?"); + mp_msg(MSGT_VO, MSGL_ERR, "[vdpau] Error when calling vdp_get_proc_address(function id %d): %s\n", + dsc->id, vdp_get_error_string ? vdp_get_error_string(vdp_st) : "?"); return -1; } } @@ -402,7 +403,8 @@ static int win_x11_init_vdpau_flip_queue(void) return 0; } -static int create_vdp_mixer(VdpChromaType vdp_chroma_type) { +static int create_vdp_mixer(VdpChromaType vdp_chroma_type) +{ #define VDP_NUM_MIXER_PARAMETER 3 #define MAX_NUM_FEATURES 5 int i; @@ -443,7 +445,8 @@ static int create_vdp_mixer(VdpChromaType vdp_chroma_type) { &video_mixer); CHECK_ST_ERROR("Error when calling vdp_video_mixer_create") - for (i = 0; i < feature_count; i++) feature_enables[i] = VDP_TRUE; + for (i = 0; i < feature_count; i++) + feature_enables[i] = VDP_TRUE; if (deint < 3) feature_enables[0] = VDP_FALSE; if (feature_count) @@ -459,7 +462,8 @@ static int create_vdp_mixer(VdpChromaType vdp_chroma_type) { } // Free everything specific to a certain video file -static void free_video_specific(void) { +static void free_video_specific(void) +{ int i; VdpStatus vdp_st; @@ -499,22 +503,22 @@ static int create_vdp_decoder(int max_refs) if (decoder != VDP_INVALID_HANDLE) vdp_decoder_destroy(decoder); switch (image_format) { - case IMGFMT_VDPAU_MPEG1: - vdp_decoder_profile = VDP_DECODER_PROFILE_MPEG1; - break; - case IMGFMT_VDPAU_MPEG2: - vdp_decoder_profile = VDP_DECODER_PROFILE_MPEG2_MAIN; - break; - case IMGFMT_VDPAU_H264: - vdp_decoder_profile = VDP_DECODER_PROFILE_H264_HIGH; - mp_msg(MSGT_VO, MSGL_V, "[vdpau] Creating H264 hardware decoder for %d reference frames.\n", max_refs); - break; - case IMGFMT_VDPAU_WMV3: - vdp_decoder_profile = VDP_DECODER_PROFILE_VC1_MAIN; - break; - case IMGFMT_VDPAU_VC1: - vdp_decoder_profile = VDP_DECODER_PROFILE_VC1_ADVANCED; - break; + case IMGFMT_VDPAU_MPEG1: + vdp_decoder_profile = VDP_DECODER_PROFILE_MPEG1; + break; + case IMGFMT_VDPAU_MPEG2: + vdp_decoder_profile = VDP_DECODER_PROFILE_MPEG2_MAIN; + break; + case IMGFMT_VDPAU_H264: + vdp_decoder_profile = VDP_DECODER_PROFILE_H264_HIGH; + mp_msg(MSGT_VO, MSGL_V, "[vdpau] Creating H264 hardware decoder for %d reference frames.\n", max_refs); + break; + case IMGFMT_VDPAU_WMV3: + vdp_decoder_profile = VDP_DECODER_PROFILE_VC1_MAIN; + break; + case IMGFMT_VDPAU_VC1: + vdp_decoder_profile = VDP_DECODER_PROFILE_VC1_ADVANCED; + break; } vdp_st = vdp_decoder_create(vdp_device, vdp_decoder_profile, vid_width, vid_height, max_refs, &decoder); @@ -553,7 +557,7 @@ static int config(uint32_t width, uint32_t height, uint32_t d_width, if (IMGFMT_IS_VDPAU(image_format) && !create_vdp_decoder(2)) return -1; - int_pause = 0; + int_pause = 0; visible_buf = 0; #ifdef CONFIG_GUI @@ -604,26 +608,26 @@ static int config(uint32_t width, uint32_t height, uint32_t d_width, vdp_chroma_type = VDP_CHROMA_TYPE_420; switch (image_format) { - case IMGFMT_YV12: - case IMGFMT_I420: - case IMGFMT_IYUV: - vdp_pixel_format = VDP_YCBCR_FORMAT_YV12; - break; - case IMGFMT_NV12: - vdp_pixel_format = VDP_YCBCR_FORMAT_NV12; - break; - case IMGFMT_YUY2: - vdp_pixel_format = VDP_YCBCR_FORMAT_YUYV; - vdp_chroma_type = VDP_CHROMA_TYPE_422; - break; - case IMGFMT_UYVY: - vdp_pixel_format = VDP_YCBCR_FORMAT_UYVY; - vdp_chroma_type = VDP_CHROMA_TYPE_422; + case IMGFMT_YV12: + case IMGFMT_I420: + case IMGFMT_IYUV: + vdp_pixel_format = VDP_YCBCR_FORMAT_YV12; + break; + case IMGFMT_NV12: + vdp_pixel_format = VDP_YCBCR_FORMAT_NV12; + break; + case IMGFMT_YUY2: + vdp_pixel_format = VDP_YCBCR_FORMAT_YUYV; + vdp_chroma_type = VDP_CHROMA_TYPE_422; + break; + case IMGFMT_UYVY: + vdp_pixel_format = VDP_YCBCR_FORMAT_UYVY; + vdp_chroma_type = VDP_CHROMA_TYPE_422; } if (create_vdp_mixer(vdp_chroma_type)) return -1; - surface_num = 0; + surface_num = 0; vid_surface_num = -1; resize(); @@ -674,8 +678,8 @@ static void draw_osd_I8A8(int x0,int y0, int w,int h, unsigned char *src, // index_data creation, component order - I, A, I, A, ..... for (i = 0; i < h; i++) for (j = 0; j < w; j++) { - index_data[i*2*w + j*2] = src [i*stride+j]; - index_data[i*2*w + j*2 + 1] = -srca[i*stride+j]; + index_data[i*2*w + j*2] = src [i*stride + j]; + index_data[i*2*w + j*2 + 1] = -srca[i*stride + j]; } output_indexed_rect_vid.x0 = x0; @@ -713,7 +717,8 @@ static void draw_osd_I8A8(int x0,int y0, int w,int h, unsigned char *src, CHECK_ST_WARNING("Error when calling vdp_output_surface_render_output_surface") } -static void draw_eosd(void) { +static void draw_eosd(void) +{ VdpStatus vdp_st; VdpOutputSurface output_surface = output_surfaces[surface_num]; VdpOutputSurfaceRenderBlendState blend_state; @@ -727,7 +732,7 @@ static void draw_eosd(void) { blend_state.blend_equation_color = VDP_OUTPUT_SURFACE_RENDER_BLEND_EQUATION_ADD; blend_state.blend_equation_alpha = VDP_OUTPUT_SURFACE_RENDER_BLEND_EQUATION_ADD; - for (i=0; ichanged == 1) goto eosd_skip_upload; - for (j=0; jnext) { // Try to reuse a suitable surface found = -1; - for (j=0; j= i->w && eosd_surfaces[j].h >= i->h) { found = j; @@ -770,7 +776,7 @@ static void generate_eosd(mp_eosd_images_t *imgs) { } // None found, allocate a new surface if (found < 0) { - for (j=0; jnext) { // Render dest, color, etc. - eosd_targets[eosd_render_count].color.alpha = 1.0 - ((i->color >> 0) & 0xff) / 255.0; - eosd_targets[eosd_render_count].color.blue = ((i->color >> 8) & 0xff) / 255.0; - eosd_targets[eosd_render_count].color.green = ((i->color >> 16) & 0xff) / 255.0; - eosd_targets[eosd_render_count].color.red = ((i->color >> 24) & 0xff) / 255.0; - eosd_targets[eosd_render_count].dest.x0 = i->dst_x; - eosd_targets[eosd_render_count].dest.y0 = i->dst_y; - eosd_targets[eosd_render_count].dest.x1 = i->w + i->dst_x; - eosd_targets[eosd_render_count].dest.y1 = i->h + i->dst_y; - eosd_targets[eosd_render_count].source.x0 = 0; - eosd_targets[eosd_render_count].source.y0 = 0; - eosd_targets[eosd_render_count].source.x1 = i->w; - eosd_targets[eosd_render_count].source.y1 = i->h; + eosd_targets[eosd_render_count].color.alpha = 1.0 - ((i->color >> 0) & 0xff) / 255.0; + eosd_targets[eosd_render_count].color.blue = ((i->color >> 8) & 0xff) / 255.0; + eosd_targets[eosd_render_count].color.green = ((i->color >> 16) & 0xff) / 255.0; + eosd_targets[eosd_render_count].color.red = ((i->color >> 24) & 0xff) / 255.0; + eosd_targets[eosd_render_count].dest.x0 = i->dst_x; + eosd_targets[eosd_render_count].dest.y0 = i->dst_y; + eosd_targets[eosd_render_count].dest.x1 = i->w + i->dst_x; + eosd_targets[eosd_render_count].dest.y1 = i->h + i->dst_y; + eosd_targets[eosd_render_count].source.x0 = 0; + eosd_targets[eosd_render_count].source.y0 = 0; + eosd_targets[eosd_render_count].source.x1 = i->w; + eosd_targets[eosd_render_count].source.y1 = i->h; eosd_render_count++; } } @@ -953,19 +959,19 @@ static int query_format(uint32_t format) { int default_flags = VFCAP_CSP_SUPPORTED | VFCAP_CSP_SUPPORTED_BY_HW | VFCAP_HWSCALE_UP | VFCAP_HWSCALE_DOWN | VFCAP_OSD | VFCAP_EOSD | VFCAP_EOSD_UNSCALED; switch (format) { - case IMGFMT_YV12: - case IMGFMT_I420: - case IMGFMT_IYUV: - case IMGFMT_NV12: - case IMGFMT_YUY2: - case IMGFMT_UYVY: - return default_flags | VOCAP_NOSLICES; - case IMGFMT_VDPAU_MPEG1: - case IMGFMT_VDPAU_MPEG2: - case IMGFMT_VDPAU_H264: - case IMGFMT_VDPAU_WMV3: - case IMGFMT_VDPAU_VC1: - return default_flags; + case IMGFMT_YV12: + case IMGFMT_I420: + case IMGFMT_IYUV: + case IMGFMT_NV12: + case IMGFMT_YUY2: + case IMGFMT_UYVY: + return default_flags | VOCAP_NOSLICES; + case IMGFMT_VDPAU_MPEG1: + case IMGFMT_VDPAU_MPEG2: + case IMGFMT_VDPAU_H264: + case IMGFMT_VDPAU_WMV3: + case IMGFMT_VDPAU_VC1: + return default_flags; } return 0; } @@ -1068,7 +1074,7 @@ static int preinit(const char *arg) deint_buffer_past_frames = 0; deint_mpi[0] = deint_mpi[1] = NULL; chroma_deint = 1; - pullup = 0; + pullup = 0; denoise = 0; sharpen = 0; if (subopt_parse(arg, subopts) != 0) { @@ -1123,7 +1129,8 @@ static int preinit(const char *arg) return 0; } -static int get_equalizer(char *name, int *value) { +static int get_equalizer(char *name, int *value) +{ if (!strcasecmp(name, "brightness")) *value = procamp.brightness * 100; else if (!strcasecmp(name, "contrast")) @@ -1137,7 +1144,8 @@ static int get_equalizer(char *name, int *value) { return VO_TRUE; } -static int set_equalizer(char *name, int value) { +static int set_equalizer(char *name, int value) +{ VdpStatus vdp_st; VdpCSCMatrix matrix; static const VdpVideoMixerAttribute attributes[] = {VDP_VIDEO_MIXER_ATTRIBUTE_CSC_MATRIX}; @@ -1166,98 +1174,98 @@ static int set_equalizer(char *name, int value) { static int control(uint32_t request, void *data, ...) { switch (request) { - case VOCTRL_GET_DEINTERLACE: - *(int*)data = deint; - return VO_TRUE; - case VOCTRL_SET_DEINTERLACE: - deint = *(int*)data; - if (deint) - deint = deint_type; - if (deint_type > 2) { - VdpStatus vdp_st; - VdpVideoMixerFeature features[1] = - {deint_type == 3 ? - VDP_VIDEO_MIXER_FEATURE_DEINTERLACE_TEMPORAL : - VDP_VIDEO_MIXER_FEATURE_DEINTERLACE_TEMPORAL_SPATIAL}; - VdpBool feature_enables[1] = {deint ? VDP_TRUE : VDP_FALSE}; - vdp_st = vdp_video_mixer_set_feature_enables(video_mixer, 1, - features, - feature_enables); - CHECK_ST_WARNING("Error changing deinterlacing settings") - deint_buffer_past_frames = 1; - } - return VO_TRUE; - case VOCTRL_PAUSE: - return (int_pause = 1); - case VOCTRL_RESUME: - return (int_pause = 0); - case VOCTRL_QUERY_FORMAT: - return query_format(*(uint32_t *)data); - case VOCTRL_GET_IMAGE: - return get_image(data); - case VOCTRL_DRAW_IMAGE: - return draw_image(data); - case VOCTRL_GUISUPPORT: - return VO_TRUE; - case VOCTRL_BORDER: - vo_x11_border(); - resize(); - return VO_TRUE; - case VOCTRL_FULLSCREEN: - vo_x11_fullscreen(); - resize(); - return VO_TRUE; - case VOCTRL_GET_PANSCAN: - return VO_TRUE; - case VOCTRL_SET_PANSCAN: - resize(); - return VO_TRUE; - case VOCTRL_SET_EQUALIZER: { - va_list ap; - int value; - - va_start(ap, data); - value = va_arg(ap, int); - - va_end(ap); - return set_equalizer(data, value); + case VOCTRL_GET_DEINTERLACE: + *(int*)data = deint; + return VO_TRUE; + case VOCTRL_SET_DEINTERLACE: + deint = *(int*)data; + if (deint) + deint = deint_type; + if (deint_type > 2) { + VdpStatus vdp_st; + VdpVideoMixerFeature features[1] = + {deint_type == 3 ? + VDP_VIDEO_MIXER_FEATURE_DEINTERLACE_TEMPORAL : + VDP_VIDEO_MIXER_FEATURE_DEINTERLACE_TEMPORAL_SPATIAL}; + VdpBool feature_enables[1] = {deint ? VDP_TRUE : VDP_FALSE}; + vdp_st = vdp_video_mixer_set_feature_enables(video_mixer, 1, + features, + feature_enables); + CHECK_ST_WARNING("Error changing deinterlacing settings") + deint_buffer_past_frames = 1; } - case VOCTRL_GET_EQUALIZER: { - va_list ap; - int *value; + return VO_TRUE; + case VOCTRL_PAUSE: + return (int_pause = 1); + case VOCTRL_RESUME: + return (int_pause = 0); + case VOCTRL_QUERY_FORMAT: + return query_format(*(uint32_t *)data); + case VOCTRL_GET_IMAGE: + return get_image(data); + case VOCTRL_DRAW_IMAGE: + return draw_image(data); + case VOCTRL_GUISUPPORT: + return VO_TRUE; + case VOCTRL_BORDER: + vo_x11_border(); + resize(); + return VO_TRUE; + case VOCTRL_FULLSCREEN: + vo_x11_fullscreen(); + resize(); + return VO_TRUE; + case VOCTRL_GET_PANSCAN: + return VO_TRUE; + case VOCTRL_SET_PANSCAN: + resize(); + return VO_TRUE; + case VOCTRL_SET_EQUALIZER: { + va_list ap; + int value; - va_start(ap, data); - value = va_arg(ap, int *); + va_start(ap, data); + value = va_arg(ap, int); - va_end(ap); - return get_equalizer(data, value); - } - case VOCTRL_ONTOP: - vo_x11_ontop(); - return VO_TRUE; - case VOCTRL_UPDATE_SCREENINFO: - update_xinerama_info(); - return VO_TRUE; - case VOCTRL_DRAW_EOSD: - if (!data) - return VO_FALSE; - generate_eosd(data); - draw_eosd(); - return VO_TRUE; - case VOCTRL_GET_EOSD_RES: { - mp_eosd_res_t *r = data; - r->mt = r->mb = r->ml = r->mr = 0; - if (vo_fs) { - r->w = vo_screenwidth; - r->h = vo_screenheight; - r->ml = r->mr = border_x; - r->mt = r->mb = border_y; - } else { - r->w = vo_dwidth; - r->h = vo_dheight; - } - return VO_TRUE; + va_end(ap); + return set_equalizer(data, value); + } + case VOCTRL_GET_EQUALIZER: { + va_list ap; + int *value; + + va_start(ap, data); + value = va_arg(ap, int *); + + va_end(ap); + return get_equalizer(data, value); + } + case VOCTRL_ONTOP: + vo_x11_ontop(); + return VO_TRUE; + case VOCTRL_UPDATE_SCREENINFO: + update_xinerama_info(); + return VO_TRUE; + case VOCTRL_DRAW_EOSD: + if (!data) + return VO_FALSE; + generate_eosd(data); + draw_eosd(); + return VO_TRUE; + case VOCTRL_GET_EOSD_RES: { + mp_eosd_res_t *r = data; + r->mt = r->mb = r->ml = r->mr = 0; + if (vo_fs) { + r->w = vo_screenwidth; + r->h = vo_screenheight; + r->ml = r->mr = border_x; + r->mt = r->mb = border_y; + } else { + r->w = vo_dwidth; + r->h = vo_dheight; } + return VO_TRUE; + } } return VO_NOTIMPL; } -- cgit v1.2.3