diff options
author | oshima@chromium.org <oshima@chromium.org@0039d316-1c4b-4281-b951-d872f2087c98> | 2013-01-16 20:15:53 +0000 |
---|---|---|
committer | oshima@chromium.org <oshima@chromium.org@0039d316-1c4b-4281-b951-d872f2087c98> | 2013-01-16 20:15:53 +0000 |
commit | 4ca987f321d0bedfa28d02ad77edc1fb73e85387 (patch) | |
tree | 7cd1550012c4bdcd9ef15513f16e1fe5043a0b27 /build | |
parent | 64462f6f2229d94957e34e40829c4ddb2f544891 (diff) | |
download | chromium_src-4ca987f321d0bedfa28d02ad77edc1fb73e85387.zip chromium_src-4ca987f321d0bedfa28d02ad77edc1fb73e85387.tar.gz chromium_src-4ca987f321d0bedfa28d02ad77edc1fb73e85387.tar.bz2 |
Add -o option to optimize png file size
The optimize code is based on pngslim. A few steps are omitted, but it runs all other processes. I'll check if it
worth fixing them (for example, advdef didn't make much difference)
Using -o option requires new third party program: optpng, pngout and advdef.
The script asks to install if they're not installed.
So for, I've got ~30% reduction in ash/resources (it's probably because pngslim was never run on this directory), I'd expect less
on other places where pngslim was run before.
Print results
BUG=150046
TEST=none
NOTRY=true
Review URL: https://chromiumcodereview.appspot.com/11905002
git-svn-id: svn://svn.chromium.org/chrome/trunk/src@177202 0039d316-1c4b-4281-b951-d872f2087c98
Diffstat (limited to 'build')
-rwxr-xr-x | build/sanitize-png-files.sh | 282 |
1 files changed, 270 insertions, 12 deletions
diff --git a/build/sanitize-png-files.sh b/build/sanitize-png-files.sh index f13a191..88c704d 100755 --- a/build/sanitize-png-files.sh +++ b/build/sanitize-png-files.sh @@ -3,7 +3,20 @@ # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. -ALL_DIRS=" +# The optimization code is based on pngslim (http://goo.gl/a0XHg) +# and executes a similar pipleline to optimize the png file size. +# The steps that require pngoptimizercl/pngrewrite/deflopt are omitted, +# but this runs all other processes, including: +# 1) various color-dependent optimizations using optipng. +# 2) optimize the number of huffman blocks. +# 3) randomize the huffman table. +# 4) Further optimize using optipng and advdef (zlib stream). +# Due to the step 3), each run may produce slightly different results. +# +# Note(oshima): In my experiment, advdef didn't reduce much. I'm keeping it +# for now as it does not take much time to run. + +readonly ALL_DIRS=" ash/resources ui/resources chrome/app/theme @@ -14,15 +27,205 @@ remoting/resources remoting/webapp " -function sanitize_file { +# Constants used for optimization +readonly MIN_BLOCK_SIZE=128 +readonly LIMIT_BLOCKS=256 +readonly RANDOM_TRIALS=100 + +# Global variables for stats +TOTAL_OLD_BYTES=0 +TOTAL_NEW_BYTES=0 +TOTAL_FILE=0 +PROCESSED_FILE=0 + +declare -a THROBBER_STR=('-' '\\' '|' '/') +THROBBER_COUNT=0 + +# Show throbber character at current cursor position. +function throbber { + echo -ne "${THROBBER_STR[$THROBBER_COUNT]}\b" + let THROBBER_COUNT=($THROBBER_COUNT+1)%4 +} + +# Usage: pngout_loop <file> <png_out_options> ... +# Optimize the png file using pngout with the given options +# using various block split thresholds and filter types. +function pngout_loop { + local file=$1 + shift + local opts=$* + for i in 0 128 256 512; do + for j in $(seq 0 5); do + throbber + pngout -q -k1 -s1 -b$i -f$j $opts $file + done + done +} + +# Usage: process_grayscale <file> +# Optimize grayscale images for all color bit depths. +# +# TODO(oshima): Experiment with -d0 w/o -c0. +function process_grayscale { + echo -n "|gray" + for opt in -d1 -d2 -d4 -d8; do + pngout_loop $file -c0 $opt + done +} + +# Usage: process_grayscale_alpha <file> +# Optimize grayscale images with alpha for all color bit depths. +function process_grayscale_alpha { + echo -n "|gray-a" + pngout_loop $file -c4 + for opt in -d1 -d2 -d4 -d8; do + pngout_loop $file -c3 $opt + done +} + +# Usage: process_rgb <file> +# Optimize rgb images with or without alpha for all color bit depths. +function process_rgb { + echo -n "|rgb" + for opt in -d1 -d2 -d4 -d8; do + pngout_loop $file -c3 $opt + done + pngout_loop $file -c2 + pngout_loop $file -c6 +} + +# Usage: huffman_blocks <file> +# Optimize the huffman blocks. +function huffman_blocks { + local file=$1 + echo -n "|huffman" + local size=$(stat -c%s $file) + let MAX_BLOCKS=$size/$MIN_BLOCK_SIZE + if [ $MAX_BLOCKS -gt $LIMIT_BLOCKS ]; then + MAX_BLOCKS=$LIMIT_BLOCKS + fi + for i in $(seq 2 $MAX_BLOCKS); do + throbber + pngout -q -k1 -ks -s1 -n$i $file + done +} + +# Usage: random_huffman_table_trial <file> +# Try compressing by randomizing the initial huffman table. +# +# TODO(oshima): Try adjusting different parameters for large files to +# reduce runtime. +function random_huffman_table_trial { + echo -n "|random" + local file=$1 + local old_size=$(stat -c%s $file) + for i in $(seq 1 $RANDOM_TRIALS); do + throbber + pngout -q -k1 -ks -s0 -r $file + done + local new_size=$(stat -c%s $file) + if [ $new_size -lt $old_size ]; then + random_huffman_table_trial $file + fi +} + +# Usage: final_comprssion <file> +# Further compress using optipng and advdef. +# TODO(oshima): Experiment with 256. +function final_compression { + echo -n "|final" + local file=$1 + for i in 32k 16k 8k 4k 2k 1k 512; do + throbber + optipng -q -nb -nc -zw$i -zc1-9 -zm1-9 -zs0-3 -f0-5 $file + done + for i in $(seq 1 4); do + throbber + advdef -q -z -$i $file + done + echo -ne "\r" +} + +# Usage: optimize_size <file> +# Performs png file optimization. +function optimize_size { tput el - echo -ne "$1\r" + local file=$1 + echo -n "$file " + + advdef -q -z -4 $file + + pngout -q -s4 -c0 -force $file $file.tmp.png + if [ -f $file.tmp.png ]; then + rm $file.tmp.png + process_grayscale $file + process_grayscale_alpha $file + else + pngout -q -s4 -c4 -force $file $file.tmp.png + if [ -f $file.tmp.png ]; then + rm $file.tmp.png + process_grayscale_alpha $file + else + process_rgb $file + fi + fi + + echo -n "|filter" + optipng -q -zc9 -zm8 -zs0-3 -f0-5 $file + pngout -q -k1 -s1 $file + + huffman_blocks $file + + # TODO(oshima): Experiment with strategy 1. + echo -n "|strategy" + for i in 3 2 0; do + pngout -q -k1 -ks -s$i $file + done + + random_huffman_table_trial $file + + final_compression $file +} + +# Usage: process_file <file> +function process_file { local file=$1 local name=$(basename $file) # -rem alla removes all ancillary chunks except for tRNS pngcrush -d $TMP_DIR -brute -reduce -rem alla $file > /dev/null - mv "$TMP_DIR/$name" "$file" + if [ ! -z "$OPTIMIZE" ]; then + optimize_size $TMP_DIR/$name + fi +} + +# Usage: sanitize_file <file> +function sanitize_file { + local file=$1 + local name=$(basename $file) + local old=$(stat -c%s $file) + local tmp_file=$TMP_DIR/$name + + process_file $file + + local new=$(stat -c%s $tmp_file) + let diff=$old-$new + let TOTAL_OLD_BYTES+=$old + let TOTAL_NEW_BYTES+=$new + let percent=($diff*100)/$old + let TOTAL_FILE+=1 + + tput el + if [ $new -lt $old ]; then + echo -ne "$file : $old => $new ($diff bytes : $percent %)\n" + mv "$tmp_file" "$file" + let PROCESSED_FILE+=1 + else + if [ -z "$OPTIMIZE" ]; then + echo -ne "$file : skipped\r" + fi + rm $tmp_file + fi } function sanitize_dir { @@ -32,17 +235,64 @@ function sanitize_dir { done } +function install_if_not_installed { + local program=$1 + dpkg -s $program > /dev/null 2>&1 + if [ "$?" != "0" ]; then + read -p "Couldn't find $program. Do you want to install? (y/n)" + [ "$REPLY" == "y" ] && sudo apt-get install $program + [ "$REPLY" == "y" ] || exit + fi +} + +function fail_if_not_installed { + local program=$1 + local url=$2 + which $program > /dev/null + if [ $? != 0 ]; then + echo "Couldn't find $program. Please download and install it from $url" + exit 1 + fi +} + +function show_help { + local program=$(basename $0) + echo \ +"Usage: $program [options] dir ... + +$program is a utility to reduce the size of png files by removing +unnecessary chunks and compressing the image. + +Options: + -o Aggressively optimize file size. Warning: this is *VERY* slow and + can take hours to process all files. + -h Print this help text." + exit 1 +} + if [ ! -e ../.gclient ]; then echo "$0 must be run in src directory" exit 1 fi -# Make sure we have pngcrush installed. -dpkg -s pngcrush > /dev/null 2>&1 -if [ "$?" != "0" ]; then - read -p "Couldn't fnd pngcrush. Do you want to install? (y/n)" - [ "$REPLY" == "y" ] && sudo apt-get install pngcrush - [ "$REPLY" == "y" ] || exit +# Parse options +while getopts oh opts +do + case $opts in + o) + OPTIMIZE=true; + shift;; + [h?]) + show_help;; + esac +done + +# Make sure we have all necessary commands installed. +install_if_not_installed pngcrush +if [ ! -z "$OPTIMIZE" ]; then + install_if_not_installed optipng + fail_if_not_installed advdef "http://advancemame.sourceforge.net/comp-download.html" + fail_if_not_installed pngout "http://www.jonof.id.au/kenutils" fi # Create tmp directory for crushed png file. @@ -51,8 +301,8 @@ TMP_DIR=$(mktemp -d) # Make sure we cleanup temp dir trap "rm -rf $TMP_DIR" EXIT -# If no arguments passed, sanitize all directories. -DIRS=$* +# If no directories are specified, sanitize all directories. +DIRS=$@ set ${DIRS:=$ALL_DIRS} for d in $DIRS; do @@ -61,3 +311,11 @@ for d in $DIRS; do echo done +# Print the results. +let diff=$TOTAL_OLD_BYTES-$TOTAL_NEW_BYTES +let percent=$diff*100/$TOTAL_OLD_BYTES +echo "Processed $PROCESSED_FILE files (out of $TOTAL_FILE files)" \ + "in $(date -u -d @$SECONDS +%T)s" +echo "Result : $TOTAL_OLD_BYTES => $TOTAL_NEW_BYTES bytes" \ + "($diff bytes : $percent %)" + |