Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 1 | /* |
| 2 | * Small range coder implementation for lzma. |
| 3 | * Copyright (C) 2006 Aurelien Jacobs <aurel@gnuage.org> |
| 4 | * |
| 5 | * Based on LzmaDecode.c from the LZMA SDK 4.22 (http://www.7-zip.org/) |
| 6 | * Copyright (c) 1999-2005 Igor Pavlov |
| 7 | * |
| 8 | * This program is free software; you can redistribute it and/or |
| 9 | * modify it under the terms of the GNU Lesser General Public |
| 10 | * License as published by the Free Software Foundation; either |
| 11 | * version 2.1 of the License, or (at your option) any later version. |
| 12 | * |
| 13 | * This program is distributed in the hope that it will be useful, |
| 14 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| 15 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| 16 | * Lesser General Public License for more details. |
| 17 | * |
| 18 | * You should have received a copy of the GNU Lesser General Public |
| 19 | * License along with this library; if not, write to the Free Software |
| 20 | * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA |
| 21 | */ |
| 22 | |
| 23 | #include <stdint.h> |
| 24 | |
| 25 | #include "libbb.h" |
| 26 | |
| 27 | #ifndef always_inline |
| 28 | # if defined(__GNUC__) && (__GNUC__ > 3 || __GNUC__ == 3 && __GNUC_MINOR__ >0) |
| 29 | # define always_inline __attribute__((always_inline)) inline |
| 30 | # else |
| 31 | # define always_inline inline |
| 32 | # endif |
| 33 | #endif |
| 34 | |
| 35 | #ifdef CONFIG_FEATURE_LZMA_FAST |
| 36 | # define speed_inline always_inline |
| 37 | #else |
| 38 | # define speed_inline |
| 39 | #endif |
| 40 | |
| 41 | |
| 42 | typedef struct { |
| 43 | int fd; |
| 44 | uint8_t *ptr; |
| 45 | uint8_t *buffer; |
| 46 | uint8_t *buffer_end; |
| 47 | int buffer_size; |
| 48 | uint32_t code; |
| 49 | uint32_t range; |
| 50 | uint32_t bound; |
| 51 | } rc_t; |
| 52 | |
| 53 | |
| 54 | #define RC_TOP_BITS 24 |
| 55 | #define RC_MOVE_BITS 5 |
| 56 | #define RC_MODEL_TOTAL_BITS 11 |
| 57 | |
| 58 | |
Rob Landley | 49a5599 | 2006-03-09 17:51:25 +0000 | [diff] [blame^] | 59 | /* Called twice: once at startup and once in rc_normalize() */ |
| 60 | static void rc_read(rc_t * rc) |
Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 61 | { |
| 62 | rc->buffer_size = read(rc->fd, rc->buffer, rc->buffer_size); |
| 63 | if (rc->buffer_size <= 0) |
| 64 | bb_error_msg_and_die("unexpected EOF"); |
| 65 | rc->ptr = rc->buffer; |
| 66 | rc->buffer_end = rc->buffer + rc->buffer_size; |
| 67 | } |
| 68 | |
Rob Landley | 49a5599 | 2006-03-09 17:51:25 +0000 | [diff] [blame^] | 69 | /* Called once */ |
Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 70 | static always_inline void rc_init(rc_t * rc, int fd, int buffer_size) |
| 71 | { |
| 72 | int i; |
| 73 | |
| 74 | rc->fd = fd; |
| 75 | rc->buffer = malloc(buffer_size); |
| 76 | rc->buffer_size = buffer_size; |
| 77 | rc->buffer_end = rc->buffer + rc->buffer_size; |
| 78 | rc->ptr = rc->buffer_end; |
| 79 | |
| 80 | rc->code = 0; |
| 81 | rc->range = 0xFFFFFFFF; |
| 82 | for (i = 0; i < 5; i++) { |
| 83 | if (rc->ptr >= rc->buffer_end) |
| 84 | rc_read(rc); |
| 85 | rc->code = (rc->code << 8) | *rc->ptr++; |
| 86 | } |
| 87 | } |
| 88 | |
Rob Landley | 49a5599 | 2006-03-09 17:51:25 +0000 | [diff] [blame^] | 89 | /* Called once. TODO: bb_maybe_free() */ |
Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 90 | static always_inline void rc_free(rc_t * rc) |
| 91 | { |
| 92 | if (ENABLE_FEATURE_CLEAN_UP) |
| 93 | free(rc->buffer); |
| 94 | } |
| 95 | |
Rob Landley | 49a5599 | 2006-03-09 17:51:25 +0000 | [diff] [blame^] | 96 | /* Called twice, but one callsite is in speed_inline'd rc_is_bit_0_helper() */ |
| 97 | static void rc_do_normalize(rc_t * rc) |
| 98 | { |
| 99 | if (rc->ptr >= rc->buffer_end) |
| 100 | rc_read(rc); |
| 101 | rc->range <<= 8; |
| 102 | rc->code = (rc->code << 8) | *rc->ptr++; |
| 103 | } |
Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 104 | static always_inline void rc_normalize(rc_t * rc) |
| 105 | { |
| 106 | if (rc->range < (1 << RC_TOP_BITS)) { |
Rob Landley | 49a5599 | 2006-03-09 17:51:25 +0000 | [diff] [blame^] | 107 | rc_do_normalize(rc); |
Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 108 | } |
| 109 | } |
| 110 | |
Rob Landley | 49a5599 | 2006-03-09 17:51:25 +0000 | [diff] [blame^] | 111 | /* Called 9 times */ |
| 112 | /* Why rc_is_bit_0_helper exists? |
| 113 | * Because we want to always expose (rc->code < rc->bound) to optimizer |
| 114 | */ |
| 115 | static speed_inline uint32_t rc_is_bit_0_helper(rc_t * rc, uint16_t * p) |
Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 116 | { |
| 117 | rc_normalize(rc); |
| 118 | rc->bound = *p * (rc->range >> RC_MODEL_TOTAL_BITS); |
Rob Landley | 49a5599 | 2006-03-09 17:51:25 +0000 | [diff] [blame^] | 119 | return rc->bound; |
| 120 | } |
| 121 | static always_inline int rc_is_bit_0(rc_t * rc, uint16_t * p) |
| 122 | { |
| 123 | uint32_t t = rc_is_bit_0_helper(rc, p); |
| 124 | return rc->code < t; |
Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 125 | } |
| 126 | |
Rob Landley | 49a5599 | 2006-03-09 17:51:25 +0000 | [diff] [blame^] | 127 | /* Called ~10 times, but very small, thus inlined */ |
Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 128 | static speed_inline void rc_update_bit_0(rc_t * rc, uint16_t * p) |
| 129 | { |
| 130 | rc->range = rc->bound; |
| 131 | *p += ((1 << RC_MODEL_TOTAL_BITS) - *p) >> RC_MOVE_BITS; |
| 132 | } |
Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 133 | static speed_inline void rc_update_bit_1(rc_t * rc, uint16_t * p) |
| 134 | { |
| 135 | rc->range -= rc->bound; |
| 136 | rc->code -= rc->bound; |
| 137 | *p -= *p >> RC_MOVE_BITS; |
| 138 | } |
| 139 | |
Rob Landley | 49a5599 | 2006-03-09 17:51:25 +0000 | [diff] [blame^] | 140 | /* Called 4 times in unlzma loop */ |
| 141 | static int rc_get_bit(rc_t * rc, uint16_t * p, int *symbol) |
Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 142 | { |
| 143 | if (rc_is_bit_0(rc, p)) { |
| 144 | rc_update_bit_0(rc, p); |
| 145 | *symbol *= 2; |
| 146 | return 0; |
| 147 | } else { |
| 148 | rc_update_bit_1(rc, p); |
| 149 | *symbol = *symbol * 2 + 1; |
| 150 | return 1; |
| 151 | } |
| 152 | } |
| 153 | |
Rob Landley | 49a5599 | 2006-03-09 17:51:25 +0000 | [diff] [blame^] | 154 | /* Called once */ |
Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 155 | static always_inline int rc_direct_bit(rc_t * rc) |
| 156 | { |
| 157 | rc_normalize(rc); |
| 158 | rc->range >>= 1; |
| 159 | if (rc->code >= rc->range) { |
| 160 | rc->code -= rc->range; |
| 161 | return 1; |
| 162 | } |
| 163 | return 0; |
| 164 | } |
| 165 | |
Rob Landley | 49a5599 | 2006-03-09 17:51:25 +0000 | [diff] [blame^] | 166 | /* Called twice */ |
Rob Landley | c1d6990 | 2006-01-20 18:28:50 +0000 | [diff] [blame] | 167 | static speed_inline void |
| 168 | rc_bit_tree_decode(rc_t * rc, uint16_t * p, int num_levels, int *symbol) |
| 169 | { |
| 170 | int i = num_levels; |
| 171 | |
| 172 | *symbol = 1; |
| 173 | while (i--) |
| 174 | rc_get_bit(rc, p + *symbol, symbol); |
| 175 | *symbol -= 1 << num_levels; |
| 176 | } |
| 177 | |
| 178 | /* vi:set ts=4: */ |