# MIT License # # Copyright (c) 2018 Jelle Hermsen # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. # # +-------------------------------------+ # | _ _ | # | __ _ __| |_ _ _ __ ___(_)_ __ | # | / _` |/ _` | | | | '_ \|_ / | '_ \ | # || (_| | (_| | |_| | | | |/ /| | |_) || # | \__, |\__,_|\__,_|_| |_/___|_| .__/ | # | |___/ |_| | # | | # +-------------------------------------+ # | A small GDScript that allows you to | # | browse and uncompress zip files. | # +-------------------------------------+ # Example usage: # # - put gdunzip.gd somewhere in your Godot project # - instance this scrip by using: # var gdunzip = load('res://LOCATION_IN_LIBRARY/gdunzip.gd').new() # - load a zip file: # var loaded = gdunzip.load('res://PATH_TO_ZIP/test.zip') # - if loaded is true you can try to uncompress a file: # var uncompressed = gdunzip.uncompress('PATH_TO_FILE_IN_ZIP/test.txt') # - now you have got a PoolByteArray named "uncompressed" with the # uncompressed data for the given file # # You can iterate over the "files" variable from the gdunzip instance, to # see all the available files: # - for f in gdunzip.files: # print(f['file_name']) # The path of the currently loaded zip file var path # A PoolByteArray with the contents of the zip file var buffer # The size of the currently loaded buffer var buffer_size # A dictionary with the files in the currently loaded zip file. # It maps full file paths to dictionaries with meta file information: # # - compression_method: -1 if uncompressed or File.COMPRESSION_DEFLATE # - file_name: the full path of the compressed file inside the zip # - file_header_offset: the exact byte location of this file's compressed data # inside the zip file # - compressed_size # - uncompressed_size var files = {} # The current position we're at inside the buffer var pos = 0 # An instance of the inner Tinf class, which is a GDScript port of Jørgen # Ibsen's excellent "tiny inflate library" var tinf # Initialize the gdunzip class func _init(): self.tinf = Tinf.new() # Load a zip file with the given path. Returns a boolean denoting success. func _load(path): if path == null: return false self.path = path self.pos = 0 var file = File.new() if !file.file_exists(path): return false file.open(path, File.READ) var file_length = file.get_len() if file.get_32() != 0x04034B50: return false file.seek(0) self.buffer = file.get_buffer(file_length) self.buffer_size = self.buffer.size() file.close() if self.buffer_size < 22: # Definitely not a valid zip file return false # Fill in self.files with all the file data return self._get_files() # Uncompress the given file. Returns false if uncompressing fails, or when the # file isn't available in the currently loaded zip file. func uncompress(file_name): if !(file_name in self.files): return false var f = self.files[file_name] self.pos = f['file_header_offset'] self._skip_file_header() var uncompressed = self._read(f['compressed_size']) if f['compression_method'] == -1: return uncompressed return tinf.tinf_uncompress(f['uncompressed_size'], uncompressed) # Returns a PoolByteArray with the compressed data for the given file. # Returns false if it can't be found. func get_compressed(file_name): if !(file_name in self.files): return false var f = self.files[file_name] self.pos = f['file_header_offset'] self._skip_file_header() return self._read(f['compressed_size']) # Parse the zip file's central directory and fill self.files with all the # file info func _get_files(): # Locate starting position of central directory var eocd_offset = buffer.size() - 22 while ( eocd_offset > 0 && not (buffer[eocd_offset+3] == 0x06 && buffer[eocd_offset+2] == 0x05 && buffer[eocd_offset+1] == 0x4B && buffer[eocd_offset] == 0x50 ) ): eocd_offset -= 1 # Set the central directory start offset self.pos = ( buffer[eocd_offset + 19] << 24 | buffer[eocd_offset + 18] << 16 | buffer[eocd_offset + 17] << 8 | buffer[eocd_offset + 16] ) # Get all central directory records, and fill self.files # with all the file information while ( buffer[pos + 3] == 0x02 && buffer[pos + 2] == 0x01 && buffer[pos + 1] == 0x4B && buffer[pos] == 0x50 ): var raw = _read(46) var header = { 'compression_method': '', 'file_name': '', 'compressed_size': 0, 'uncompressed_size': 0, 'file_header_offset': -1, } if raw[10] == 0 && raw[11] == 0: header['compression_method'] = -1 else: header['compression_method'] = File.COMPRESSION_DEFLATE header['compressed_size'] = ( raw[23] << 24 | raw[22] << 16 | raw[21] << 8 | raw[20] ) header['uncompressed_size'] = ( raw[27] << 24 | raw[26] << 16 | raw[25] << 8 | raw[24] ) header['file_header_offset'] = ( raw[45] << 24 | raw[44] << 16 | raw[43] << 8 | raw[42] ) var file_name_length = raw[29] << 8 | raw[28] var extra_field_length = raw[31] << 8 | raw[30] var comment_length = raw[33] << 8 | raw[32] var raw_end = _read(file_name_length + extra_field_length + comment_length) if !raw_end: return false header['file_name'] = ( raw_end.subarray(0, file_name_length - 1).get_string_from_utf8() ) self.files[header['file_name']] = header return true # Read a given number of bytes from the buffer, and return it as a # PoolByteArray func _read(length): var result = buffer.subarray(pos, pos + length - 1) if result.size() != length: return false pos = pos + length return result # Skip the given number of bytes in the buffer. Advancing "pos". func _skip(length): pos += length # This function skips the file header information, when "pos" points at a file # header inside the buffer. func _skip_file_header(): var raw = _read(30) if !raw: return false var file_name_length = raw[27] << 8 | raw[26] var extra_field_length = raw[29] << 8 | raw[28] var raw_end = _skip(file_name_length + extra_field_length) # The inner Tinf class is a pretty straight port from Jørgen Ibsen's excellent # "tiny inflate library". It's written in an imperative style and I have tried # to stay close to the original. I added two helper functions, and had to make # some minor additions to support "faux pointer arithmetic". # # I have created a TINF_TREE and TINF_DATA dictionary to serve as structs. I # use "duplicate" to make an instance from these structs. class Tinf: # ---------------------------------------- # -- GDscript specific helper functions -- # ---------------------------------------- func make_pool_int_array(size): var pool_int_array = PoolIntArray() pool_int_array.resize(size) return pool_int_array func make_pool_byte_array(size): var pool_byte_array = PoolByteArray() pool_byte_array.resize(size) return pool_byte_array # ------------------------------ # -- internal data structures -- # ------------------------------ var TINF_TREE = { 'table': make_pool_int_array(16), 'trans': make_pool_int_array(288), } var TINF_DATA = { 'source': PoolByteArray(), # sourcePtr is an "int" that's used to point at a location in "source". # I added this since we don't have pointer arithmetic in GDScript. 'sourcePtr': 0, 'tag': 0, 'bitcount': 0, 'dest': PoolByteArray(), 'destLen': 0, # "Faux pointer" to dest. 'destPtr': 0, 'ltree': TINF_TREE.duplicate(), 'dtree': TINF_TREE.duplicate() } const TINF_OK = 0 const TINF_DATA_ERROR = -3 # --------------------------------------------------- # -- uninitialized global data (static structures) -- # --------------------------------------------------- var sltree = TINF_TREE.duplicate() # fixed length/symbol tree var sdtree = TINF_TREE.duplicate() # fixed distance tree var base_tables = { # extra bits and base tables for length codes 'length_bits': make_pool_byte_array(30), 'length_base': make_pool_int_array(30), # extra bits and base tables for distance codes 'dist_bits': make_pool_byte_array(30), 'dist_base': make_pool_int_array(30) } var clcidx = PoolByteArray([ 16, 17, 18, 0, 8, 7, 9, 6, 10, 5, 11, 4, 12, 3, 13, 2, 14, 1, 15]) # ----------------------- # -- utility functions -- # ----------------------- # build extra bits and base tables # bits: PoolByteArray # base: PoolIntArray # delta: int # first: int func tinf_build_bits_base(target, delta, first): var i = 0 var sum = first for i in range(0, delta): base_tables[target + '_bits'][i] = 0 for i in range(0, 30 - delta): base_tables[target + '_bits'][i + delta] = i / delta for i in range(0, 30): base_tables[target + '_base'][i] = sum sum += 1 << base_tables[target + '_bits'][i] # build the fixed huffman trees # lt: TINF_TREE # rt: TINF_TREE func tinf_build_fixed_trees(lt, dt): var i = 0 for i in range(0, 7): lt['table'][i] = 0 lt['table'][7] = 24 lt['table'][8] = 152 lt['table'][9] = 112 for i in range(0, 24): lt['trans'][i] = 256 + i for i in range(0, 144): lt['trans'][24 + i] = i for i in range(0, 8): lt['trans'][24 + 144 + i] = 280 + i for i in range(0, 112): lt['trans'][24 + 144 + 8 + i] = 144 + i for i in range(0, 5): dt['table'][i] = 0 dt['table'][5] = 32 for i in range(0, 32): dt['trans'][i] = i # given an array of code lengths, build a tree # t: TINF_TREE # lengths: PoolByteArray # num: int func tinf_build_tree(t, lengths, num): var offs = make_pool_int_array(16) var i = 0 var sum = 0 # clear code length count table for i in range(0,16): t['table'][i] = 0 # scan symbol lengths, and sum code length counts for i in range(0, num): t['table'][lengths[i]] += 1 t['table'][0] = 0 for i in range(0,16): offs[i] = sum sum += t['table'][i] for i in range(0, num): if lengths[i]: t['trans'][offs[lengths[i]]] = i offs[lengths[i]] += 1 # ---------------------- # -- decode functions -- # ---------------------- # get one bit from source stream # d: TINF_DATA # returns: int func tinf_getbit(d): var bit = 0 d['bitcount'] -= 1 if !(d['bitcount'] + 1) : d['tag'] = d['source'][d['sourcePtr']] d['sourcePtr'] += 1 d['bitcount'] = 7 bit = d['tag'] & 0x01 d['tag'] >>= 1 return bit # read a num bit value from a stream and add base # d: TINF_DATA # num: int # base: int func tinf_read_bits(d, num, base): var val = 0 if num: var limit = 1 << num var mask = 1 while mask < limit: if tinf_getbit(d): val += mask mask *= 2 return val + base # given a data stream and a tree, decode a symbol # d: TINF_DATA # t: TINF_TREE func tinf_decode_symbol(d, t): var sum = 0 var cur = 0 var length = 0 while true: cur = 2 * cur + tinf_getbit(d) length += 1 sum += t['table'][length] cur -= t['table'][length] if cur < 0: break return t['trans'][sum + cur] # given a data stream, decode dynamic trees from it # d: TINF_DATA # lt: TINF_TREE # dt: TINF_TREE func tinf_decode_trees(d, lt, dt): var code_tree = TINF_TREE.duplicate() var lengths = make_pool_byte_array(288 + 32) var hlit = 0 var hdist = 0 var hclen = 0 var i = 0 var num = 0 var length = 0 # get 5 bits HLIT (257-286) hlit = tinf_read_bits(d, 5, 257) # get 5 bits HDIST (1-32) hdist = tinf_read_bits(d, 5, 1) # get 4 bits HCLEN (4-19) hclen = tinf_read_bits(d, 4, 4) for i in range(0, 19): lengths[i] = 0 for i in range(0, hclen): var clen = tinf_read_bits(d, 3, 0) lengths[clcidx[i]] = clen # build code length tree tinf_build_tree(code_tree, lengths, 19) while num < hlit + hdist: var sym = tinf_decode_symbol(d, code_tree) match sym: 16: var prev = lengths[num - 1] length = tinf_read_bits(d, 2, 3) while length != 0: lengths[num] = prev num += 1 length -= 1 17: length = tinf_read_bits(d, 3, 3) while length != 0: lengths[num] = 0 num += 1 length -= 1 18: length = tinf_read_bits(d, 7, 11) while length != 0: lengths[num] = 0 num += 1 length -= 1 _: lengths[num] = sym num += 1 # build dynamic trees tinf_build_tree(lt, lengths, hlit) tinf_build_tree(dt, lengths.subarray(hlit, lengths.size() - 1), hdist) # ----------------------------- # -- block inflate functions -- # ----------------------------- # given a stream and two trees, inflate a block of data # d: TINF_DATA # lt: TINF_TREE # dt: TINF_TREE func tinf_inflate_block_data(d, lt, dt): var start = d['destPtr'] while true: var sym = tinf_decode_symbol(d, lt) if sym == 256: d['destLen'] += d['destPtr'] - start return TINF_OK if sym < 256: d['dest'][d['destPtr']] = sym d['destPtr'] += 1 else: var length = 0 var dist = 0 var offs = 0 var i = 0 var ptr = d['destPtr'] sym -= 257 length = tinf_read_bits(d, base_tables['length_bits'][sym], base_tables['length_base'][sym]) dist = tinf_decode_symbol(d, dt) # possibly get more bits from distance code offs = tinf_read_bits(d, base_tables['dist_bits'][dist], base_tables['dist_base'][dist]) for i in range(0, length): d['dest'][ptr + i] = d['dest'][ptr + (i - offs)] d['destPtr'] += length # inflate an uncompressed block of data */ # d: TINF_DATA func tinf_inflate_uncompressed_block(d): var length = 0 var invlength = 0 var i = 0 # get length length = d['source'][d['sourcePtr'] + 1] length = 256 * length + d['source'][d['sourcePtr']] # get one's complement of length invlength = d['source'][d['sourcePtr'] + 3] invlength = 256 * invlength + d['source'][d['sourcePtr'] + 2] if length != ~invlength & 0x0000FFFF: return TINF_DATA_ERROR d['sourcePtr'] += 4 i = length while i: d['dest'][d['destPtr']] = d['source'][d['sourcePtr']] d['destPtr'] += 1 d['sourcePtr'] += 1 i -= 1 d['bitcount'] = 0 d['destLen'] += length return TINF_OK # inflate a block of data compressed with fixed huffman trees # d: TINF_DATA # returns: int func tinf_inflate_fixed_block(d): # decode block using fixed trees return tinf_inflate_block_data(d, sltree, sdtree) # inflate a block of data compressed with dynamic huffman trees # d: TINF_DATA # returns: int func tinf_inflate_dynamic_block(d): # decode trees from stream tinf_decode_trees(d, d['ltree'], d['dtree']) # decode block using decoded trees return tinf_inflate_block_data(d, d['ltree'], d['dtree']) # ---------------------- # -- public functions -- # ---------------------- func _init(): # build fixed huffman trees tinf_build_fixed_trees(sltree, sdtree) # build extra bits and base tables tinf_build_bits_base('length', 4, 3) tinf_build_bits_base('dist', 2, 1) # fix a special case base_tables['length_bits'][28] = 0 base_tables['length_base'][28] = 258 # inflate stream from source to dest func tinf_uncompress(destLen, source): var d = TINF_DATA.duplicate() var dest = make_pool_byte_array(destLen) var sourceSize = source.size() d['source'] = source d['dest'] = dest destLen = 0 while true: var btype = 0 var res = 0 # Skip final block flag tinf_getbit(d) # read block type (2 bits) btype = tinf_read_bits(d, 2, 0) match btype: 0: # decompress uncompressed block res = tinf_inflate_uncompressed_block(d) 1: # decompress block with fixed huffman trees res = tinf_inflate_fixed_block(d) 2: # decompress block with dynamic huffman trees res = tinf_inflate_dynamic_block(d) _: return false if res != TINF_OK: return false # When we have consumed the entire source, we're done if d['sourcePtr'] >= sourceSize: break return d['dest']