/* $NetBSD: libhfs.c,v 1.3 2007/03/09 22:14:09 dillo Exp $ */ /*- * Copyright (c) 2005, 2007 The NetBSD Foundation, Inc. * All rights reserved. * * This code is derived from software contributed to The NetBSD Foundation * by Yevgeny Binder and Dieter Baron. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. */ /* * All functions and variable types have the prefix "hfs_". All constants * have the prefix "HFS_". * * Naming convention for functions which read/write raw, linear data * into/from a structured form: * * hfs_read/write[d][a]_foo_bar * [d] - read/write from/to [d]isk instead of a memory buffer * [a] - [a]llocate output buffer instead of using an existing one * (not applicable for writing functions) * * Most functions do not have either of these options, so they will read from * or write to a memory buffer, which has been previously allocated by the * caller. */ #include "libhfs.h" /* global private file/folder keys */ hfs_catalog_key_t hfs_gMetadataDirectoryKey; /* contains HFS+ inodes */ hfs_catalog_key_t hfs_gJournalInfoBlockFileKey; hfs_catalog_key_t hfs_gJournalBufferFileKey; hfs_catalog_key_t* hfs_gPrivateObjectKeys[4] = { &hfs_gMetadataDirectoryKey, &hfs_gJournalInfoBlockFileKey, &hfs_gJournalBufferFileKey, NULL}; extern uint16_t be16tohp(void** inout_ptr); extern uint32_t be32tohp(void** inout_ptr); extern uint64_t be64tohp(void** inout_ptr); int hfslib_create_casefolding_table(void); #ifdef DLO_DEBUG #include void dlo_print_key(hfs_catalog_key_t *key) { int i; printf("%ld:[", (long)key->parent_cnid); for (i=0; iname.length; i++) { if (key->name.unicode[i] < 256 && isprint(key->name.unicode[i])) putchar(key->name.unicode[i]); else printf("<%04x>", key->name.unicode[i]); } printf("]"); } #endif void hfslib_init(hfs_callbacks* in_callbacks) { unichar_t temp[256]; if(in_callbacks!=NULL) memcpy(&hfs_gcb, in_callbacks, sizeof(hfs_callbacks)); hfs_gcft = NULL; /* * Create keys for the HFS+ "private" files so we can reuse them whenever * we perform a user-visible operation, such as listing directory contents. */ #define ATOU(str, len) /* quick & dirty ascii-to-unicode conversion */ \ do{ int i; for(i=0; ireadonly = in_readonly; if(hfslib_openvoldevice(out_vol, in_device, in_offset, cbargs) != 0) HFS_LIBERR("could not open device"); /* * Read the volume header. */ buffer = hfslib_malloc(sizeof(hfs_volume_header_t), cbargs); if(buffer==NULL) HFS_LIBERR("could not allocate volume header"); if(hfslib_readd(out_vol, buffer, sizeof(hfs_volume_header_t), HFS_VOLUME_HEAD_RESERVE_SIZE, cbargs)!=0) HFS_LIBERR("could not read volume header"); if(hfslib_read_volume_header(buffer, &(out_vol->vh))==0) HFS_LIBERR("could not parse volume header"); /* * Check the volume signature to see if this is a legitimate HFS+ or HFSX * volume. If so, set the key comparison function pointers appropriately. */ switch(out_vol->vh.signature) { case HFS_SIG_HFSP: out_vol->keycmp = hfslib_compare_catalog_keys_cf; break; case HFS_SIG_HFSX: out_vol->keycmp = NULL; /* will be set below */ break; case HFS_SIG_HFS: HFS_LIBERR("HFS volumes and HFS+ volumes with HFS wrappers are" "not currently supported"); break; default: HFS_LIBERR("unrecognized volume format"); } /* * Read the catalog header. */ buffer2 = hfslib_realloc(buffer, 512, cbargs); if(buffer2==NULL) HFS_LIBERR("could not allocate catalog header node"); buffer = buffer2; /* We are only interested in the node header, so read the first 512 bytes and construct the node descriptor by hand. */ if(hfslib_readd(out_vol, buffer, 512, out_vol->vh.catalog_file.extents[0].start_block *(uint64_t)out_vol->vh.block_size, cbargs) != 0) HFS_LIBERR("could not read catalog header node"); node_recs[0] = (char *)buffer+14; node_rec_sizes[0] = 120; if(hfslib_read_header_node(node_recs, node_rec_sizes, 1, &out_vol->chr, NULL, NULL)==0) HFS_LIBERR("could not parse catalog header node"); /* If this is an HFSX volume, the catalog header specifies the type of * key comparison method (case-folding or binary compare) we should use. */ if(out_vol->keycmp == NULL) { if(out_vol->chr.keycomp_type == HFS_KEY_CASEFOLD) out_vol->keycmp = hfslib_compare_catalog_keys_cf; else if(out_vol->chr.keycomp_type == HFS_KEY_BINARY) out_vol->keycmp = hfslib_compare_catalog_keys_bc; else HFS_LIBERR("undefined key compare method"); } out_vol->catkeysizefieldsize = (out_vol->chr.attributes & HFS_BIG_KEYS_MASK) ? sizeof(uint16_t) : sizeof(uint8_t); /* * Read the extent overflow header. */ /* We are only interested in the node header, so read the first 512 bytes and construct the node descriptor by hand. buffer is already 512 bytes long. */ if(hfslib_readd(out_vol, buffer, 512, out_vol->vh.extents_file.extents[0].start_block *(uint64_t)out_vol->vh.block_size, cbargs) != 0) HFS_LIBERR("could not read extent header node"); node_recs[0] = (char *)buffer+14; node_rec_sizes[0] = 120; if(hfslib_read_header_node(node_recs, node_rec_sizes, 1, &out_vol->ehr, NULL, NULL)==0) HFS_LIBERR("could not parse extent header node"); out_vol->extkeysizefieldsize = (out_vol->ehr.attributes & HFS_BIG_KEYS_MASK) ? sizeof(uint16_t):sizeof(uint8_t); /* * Read the journal info block and journal header (if volume journaled). */ if(out_vol->vh.attributes & (1<vh.journal_info_block * out_vol->vh.block_size, cbargs) != 0) HFS_LIBERR("could not read journal info block"); if(hfslib_read_journal_info(buffer, &out_vol->jib)==0) HFS_LIBERR("could not parse journal info block"); /* journal header */ buffer2 = hfslib_realloc(buffer, sizeof(hfs_journal_header_t),cbargs); if(buffer2==NULL) HFS_LIBERR("could not allocate journal header"); buffer = buffer2; if(hfslib_readd(out_vol, buffer, sizeof(hfs_journal_header_t), out_vol->jib.offset, cbargs) != 0) HFS_LIBERR("could not read journal header"); if(hfslib_read_journal_header(buffer, &out_vol->jh)==0) HFS_LIBERR("could not parse journal header"); out_vol->journaled = 1; } else { out_vol->journaled = 0; } /* * If this volume uses case-folding comparison and the folding table hasn't * been created yet, do that here. (We don't do this in hfslib_init() * because the table is large and we might never even need to use it.) */ if(out_vol->keycmp==hfslib_compare_catalog_keys_cf && hfs_gcft==NULL) result = hfslib_create_casefolding_table(); else result = 0; /* * Find and store the volume name. */ if(hfslib_make_catalog_key(HFS_CNID_ROOT_FOLDER, 0, NULL, &rootkey)==0) HFS_LIBERR("could not make root search key"); if(hfslib_find_catalog_record_with_key(out_vol, &rootkey, (hfs_catalog_keyed_record_t*)&rootthread, cbargs)!=0) HFS_LIBERR("could not find root parent"); memcpy(&out_vol->name, &rootthread.name, sizeof(hfs_unistr255_t)); /* FALLTHROUGH */ error: if(buffer!=NULL) hfslib_free(buffer, cbargs); return result; } void hfslib_close_volume(hfs_volume* in_vol, hfs_callback_args* cbargs) { if(in_vol==NULL) return; hfslib_closevoldevice(in_vol, cbargs); } int hfslib_path_to_cnid(hfs_volume* in_vol, hfs_cnid_t in_cnid, char** out_unicode, uint16_t* out_length, hfs_callback_args* cbargs) { hfs_thread_record_t parent_thread; hfs_cnid_t parent_cnid, child_cnid; char* newpath; char* path; int path_offset = 0; int result; uint16_t* ptr; /* dummy var */ uint16_t uchar; /* dummy var */ uint16_t total_path_length; if(in_vol==NULL || in_cnid==0 || out_unicode==NULL || out_length==NULL) return 1; result = 1; *out_unicode = NULL; *out_length = 0; path = NULL; total_path_length = 0; path = hfslib_malloc(514, cbargs); /* 256 unichars plus a forward slash */ if(path==NULL) return 1; child_cnid = in_cnid; parent_cnid = child_cnid; /* skips loop in case in_cnid is root id */ while(parent_cnid != HFS_CNID_ROOT_FOLDER && parent_cnid != HFS_CNID_ROOT_PARENT) { if(child_cnid!=in_cnid) { newpath = hfslib_realloc(path, 514 + total_path_length*2, cbargs); if(newpath==NULL) goto exit; path = newpath; memmove(path + 514, path + path_offset, total_path_length*2); } parent_cnid = hfslib_find_parent_thread(in_vol, child_cnid, &parent_thread, cbargs); if(parent_cnid==0) goto exit; path_offset = 512 - parent_thread.name.length*2; memcpy(path + path_offset, parent_thread.name.unicode, parent_thread.name.length*2); /* Add a forward slash. The unicode string was specified in big endian * format, so convert to core format if necessary. */ path[512]=0x00; path[513]=0x2F; ptr = (uint16_t*)path + 256; uchar = be16tohp((void*)&ptr); *(ptr-1) = uchar; total_path_length += parent_thread.name.length + 1; child_cnid = parent_cnid; } /* * At this point, 'path' holds a sequence of unicode characters which * represent the absolute path to the given cnid. This string is missing * a terminating null char and an initial forward slash that represents * the root of the filesystem. It most likely also has extra space in * the beginning, due to the fact that we reserve 512 bytes for each path * component and won't usually use all that space. So, we allocate the * final string based on the actual length of the absolute path, plus four * additional bytes (two unichars) for the forward slash and the null char. */ *out_unicode = hfslib_malloc((total_path_length+2)*2, cbargs); if(*out_unicode == NULL) goto exit; /* copy only the bytes that are actually used */ memcpy(*out_unicode+2, path + path_offset, total_path_length*2); /* insert forward slash at start */ (*out_unicode)[0] = 0x00; (*out_unicode)[1] = 0x2F; ptr = (uint16_t*)*out_unicode; uchar = be16tohp((void*)&ptr); *(ptr-1) = uchar; /* insert null char at end */ (*out_unicode)[total_path_length*2+2] = 0x00; (*out_unicode)[total_path_length*2+3] = 0x00; *out_length = total_path_length + 1 /* extra for forward slash */ ; result = 0; exit: if(path!=NULL) hfslib_free(path, cbargs); return result; } hfs_cnid_t hfslib_find_parent_thread( hfs_volume* in_vol, hfs_cnid_t in_child, hfs_thread_record_t* out_thread, hfs_callback_args* cbargs) { hfs_catalog_key_t childkey; if(in_vol==NULL || in_child==0 || out_thread==NULL) return 0; if(hfslib_make_catalog_key(in_child, 0, NULL, &childkey)==0) return 0; if(hfslib_find_catalog_record_with_key(in_vol, &childkey, (hfs_catalog_keyed_record_t*)out_thread, cbargs)!=0) return 0; return out_thread->parent_cnid; } /* * hfslib_find_catalog_record_with_cnid() * * Looks up a catalog record by calling hfslib_find_parent_thread() and * hfslib_find_catalog_record_with_key(). out_key may be NULL; if not, the key * corresponding to this cnid is stuffed in it. Returns 0 on success. */ int hfslib_find_catalog_record_with_cnid( hfs_volume* in_vol, hfs_cnid_t in_cnid, hfs_catalog_keyed_record_t* out_rec, hfs_catalog_key_t* out_key, hfs_callback_args* cbargs) { hfs_cnid_t parentcnid; hfs_thread_record_t parentthread; hfs_catalog_key_t key; if(in_vol==NULL || in_cnid==0 || out_rec==NULL) return 0; parentcnid = hfslib_find_parent_thread(in_vol, in_cnid, &parentthread, cbargs); if(parentcnid == 0) HFS_LIBERR("could not find parent thread for cnid %i", in_cnid); if(hfslib_make_catalog_key(parentthread.parent_cnid, parentthread.name.length, parentthread.name.unicode, &key) == 0) HFS_LIBERR("could not make catalog search key"); if(out_key!=NULL) memcpy(out_key, &key, sizeof(key)); return hfslib_find_catalog_record_with_key(in_vol, &key, out_rec, cbargs); error: return 1; } /* Returns 0 on success, 1 on error, and -1 if record was not found. */ int hfslib_find_catalog_record_with_key( hfs_volume* in_vol, hfs_catalog_key_t* in_key, hfs_catalog_keyed_record_t* out_rec, hfs_callback_args* cbargs) { hfs_node_descriptor_t nd; hfs_extent_descriptor_t* extents; hfs_catalog_keyed_record_t lastrec; hfs_catalog_key_t* curkey; void** recs; void* buffer; uint64_t bytesread; uint32_t curnode; uint16_t* recsizes; uint16_t numextents; uint16_t recnum; int16_t leaftype; int keycompare; int result; if(in_key==NULL || out_rec==NULL || in_vol==NULL) return 1; result = 1; buffer = NULL; curkey = NULL; extents = NULL; recs = NULL; recsizes = NULL; /* The key takes up over half a kb of ram, which is a lot for the BSD * kernel stack. So allocate it in the heap instead to play it safe. */ curkey = hfslib_malloc(sizeof(hfs_catalog_key_t), cbargs); if(curkey==NULL) HFS_LIBERR("could not allocate catalog search key"); buffer = hfslib_malloc(in_vol->chr.node_size, cbargs); if(buffer==NULL) HFS_LIBERR("could not allocate node buffer"); numextents = hfslib_get_file_extents(in_vol, HFS_CNID_CATALOG, HFS_DATAFORK, &extents, cbargs); if(numextents==0) HFS_LIBERR("could not locate fork extents"); nd.num_recs = 0; curnode = in_vol->chr.root_node; #ifdef DLO_DEBUG printf("-> key "); dlo_print_key(in_key); printf("\n"); #endif do { #ifdef DLO_DEBUG printf("--> node %d\n", curnode); #endif if(hfslib_readd_with_extents(in_vol, buffer, &bytesread,in_vol->chr.node_size, curnode * in_vol->chr.node_size, extents, numextents, cbargs)!=0) HFS_LIBERR("could not read catalog node #%i", curnode); if(hfslib_reada_node(buffer, &nd, &recs, &recsizes, HFS_CATALOG_FILE, in_vol, cbargs)==0) HFS_LIBERR("could not parse catalog node #%i", curnode); for(recnum=0; recnum record %d: ", recnum); dlo_print_key(curkey); fflush(stdout); #endif keycompare = in_vol->keycmp(in_key, curkey); #ifdef DLO_DEBUG printf(" %c\n", keycompare < 0 ? '<' : keycompare == 0 ? '=' : '>'); #endif if(keycompare < 0) { /* Check if key is less than *every* record, which should never * happen if the volume is consistent and the key legit. */ if(recnum==0) HFS_LIBERR("all records greater than key"); /* Otherwise, we've found the first record that exceeds our key, * so retrieve the previous record, which is still less... */ memcpy(out_rec, &lastrec, sizeof(hfs_catalog_keyed_record_t)); /* ...unless this is a leaf node, which means we've gone from * a key which is smaller than the search key, in the previous * loop, to a key which is larger, in this loop, and that * implies that our search key does not exist on the volume. */ if(nd.kind==HFS_LEAFNODE) result = -1; break; } else if(keycompare == 0) { /* If leaf node, found an exact match. */ result = 0; break; } else if(recnum==nd.num_recs-1 && keycompare > 0) { /* If leaf node, we've reached the last record with no match, * which means this key is not present on the volume. */ result = -1; break; } memcpy(&lastrec, out_rec, sizeof(hfs_catalog_keyed_record_t)); } if(nd.kind==HFS_INDEXNODE) curnode = out_rec->child; else if(nd.kind==HFS_LEAFNODE) break; hfslib_free_recs(&recs, &recsizes, &nd.num_recs, cbargs); } while(nd.kind!=HFS_LEAFNODE); /* FALLTHROUGH */ error: if(extents!=NULL) hfslib_free(extents, cbargs); hfslib_free_recs(&recs, &recsizes, &nd.num_recs, cbargs); if(curkey!=NULL) hfslib_free(curkey, cbargs); if(buffer!=NULL) hfslib_free(buffer, cbargs); return result; } /* returns 0 on success */ /* XXX Need to look this over and make sure it gracefully handles cases where * XXX the key is not found. */ int hfslib_find_extent_record_with_key(hfs_volume* in_vol, hfs_extent_key_t* in_key, hfs_extent_record_t* out_rec, hfs_callback_args* cbargs) { hfs_node_descriptor_t nd; hfs_extent_descriptor_t* extents; hfs_extent_record_t lastrec; hfs_extent_key_t curkey; void** recs; void* buffer; uint64_t bytesread; uint32_t curnode; uint16_t* recsizes; uint16_t numextents; uint16_t recnum; int keycompare; int result; if(in_vol==NULL || in_key==NULL || out_rec==NULL) return 1; result = 1; buffer = NULL; extents = NULL; recs = NULL; recsizes = NULL; buffer = hfslib_malloc(in_vol->ehr.node_size, cbargs); if(buffer==NULL) HFS_LIBERR("could not allocate node buffer"); numextents = hfslib_get_file_extents(in_vol, HFS_CNID_EXTENTS, HFS_DATAFORK, &extents, cbargs); if(numextents==0) HFS_LIBERR("could not locate fork extents"); nd.num_recs = 0; curnode = in_vol->ehr.root_node; do { hfslib_free_recs(&recs, &recsizes, &nd.num_recs, cbargs); recnum = 0; if(hfslib_readd_with_extents(in_vol, buffer, &bytesread, in_vol->ehr.node_size, curnode * in_vol->ehr.node_size, extents, numextents, cbargs)!=0) HFS_LIBERR("could not read extents overflow node #%i", curnode); if(hfslib_reada_node(buffer, &nd, &recs, &recsizes, HFS_EXTENTS_FILE, in_vol, cbargs)==0) HFS_LIBERR("could not parse extents overflow node #%i",curnode); for(recnum=0; recnum 0)) break; } if(nd.kind==HFS_INDEXNODE) curnode = *((uint32_t *)out_rec); /* out_rec is a node ptr in this case */ else if(nd.kind==HFS_LEAFNODE) break; else HFS_LIBERR("unknwon node type for extents overflow node #%i",curnode); } while(nd.kind!=HFS_LEAFNODE); result = 0; /* FALLTHROUGH */ error: if(buffer!=NULL) hfslib_free(buffer, cbargs); if(extents!=NULL) hfslib_free(extents, cbargs); hfslib_free_recs(&recs, &recsizes, &nd.num_recs, cbargs); return result; } /* out_extents may be NULL. */ uint16_t hfslib_get_file_extents(hfs_volume* in_vol, hfs_cnid_t in_cnid, uint8_t in_forktype, hfs_extent_descriptor_t** out_extents, hfs_callback_args* cbargs) { hfs_extent_descriptor_t* dummy; hfs_extent_key_t extentkey; hfs_file_record_t file; hfs_catalog_key_t filekey; hfs_thread_record_t fileparent; hfs_fork_t fork; hfs_extent_record_t nextextentrec; uint32_t numblocks; uint16_t numextents, n; if(in_vol==NULL || in_cnid==0) return 0; if(out_extents!=NULL) { *out_extents = hfslib_malloc(sizeof(hfs_extent_descriptor_t), cbargs); if(*out_extents==NULL) return 0; } switch(in_cnid) { case HFS_CNID_CATALOG: fork = in_vol->vh.catalog_file; break; case HFS_CNID_EXTENTS: fork = in_vol->vh.extents_file; break; case HFS_CNID_ALLOCATION: fork = in_vol->vh.allocation_file; break; case HFS_CNID_ATTRIBUTES: fork = in_vol->vh.attributes_file; break; case HFS_CNID_STARTUP: fork = in_vol->vh.startup_file; break; default: if(hfslib_find_parent_thread(in_vol, in_cnid, &fileparent, cbargs)==0) goto error; if(hfslib_make_catalog_key(fileparent.parent_cnid, fileparent.name.length, fileparent.name.unicode, &filekey)==0) goto error; if(hfslib_find_catalog_record_with_key(in_vol, &filekey, (hfs_catalog_keyed_record_t*)&file, cbargs)!=0) goto error; /* only files have extents, not folders or threads */ if(file.rec_type!=HFS_REC_FILE) goto error; if(in_forktype==HFS_DATAFORK) fork = file.data_fork; else if(in_forktype==HFS_RSRCFORK) fork = file.rsrc_fork; } numextents = 0; numblocks = 0; memcpy(&nextextentrec, &fork.extents, sizeof(hfs_extent_record_t)); while(1) { for(n=0; n<8; n++) { if(nextextentrec[n].block_count==0) break; numblocks += nextextentrec[n].block_count; } if(out_extents!=NULL) { dummy = hfslib_realloc(*out_extents, (numextents+n) * sizeof(hfs_extent_descriptor_t), cbargs); if(dummy==NULL) goto error; *out_extents = dummy; memcpy(*out_extents + numextents, &nextextentrec, n*sizeof(hfs_extent_descriptor_t)); } numextents += n; if(numblocks >= fork.total_blocks) break; if(hfslib_make_extent_key(in_cnid, in_forktype, numblocks, &extentkey)==0) goto error; if(hfslib_find_extent_record_with_key(in_vol, &extentkey, &nextextentrec, cbargs)!=0) goto error; } goto exit; error: if(out_extents!=NULL && *out_extents!=NULL) { hfslib_free(*out_extents, cbargs); *out_extents = NULL; } return 0; exit: return numextents; } /* * hfslib_get_directory_contents() * * Finds the immediate children of a given directory CNID and places their * CNIDs in an array allocated here. The first child is found by doing a * catalog search that only compares parent CNIDs (ignoring file/folder names) * and skips over thread records. Then the remaining children are listed in * ascending order by name, according to the HFS+ spec, so just read off each * successive leaf node until a different parent CNID is found. * * If out_childnames is not NULL, it will be allocated and set to an array of * hfs_unistr255_t's which correspond to the name of the child with that same * index. * * out_children may be NULL. * * Returns 0 on success. */ int hfslib_get_directory_contents( hfs_volume* in_vol, hfs_cnid_t in_dir, hfs_catalog_keyed_record_t** out_children, hfs_unistr255_t** out_childnames, uint32_t* out_numchildren, hfs_callback_args* cbargs) { hfs_node_descriptor_t nd; hfs_extent_descriptor_t* extents; hfs_catalog_keyed_record_t currec; hfs_catalog_key_t curkey; void** recs; void* buffer; void* ptr; /* temporary pointer for realloc() */ uint64_t bytesread; uint32_t curnode; uint32_t lastnode; uint16_t* recsizes; uint16_t numextents; uint16_t recnum; int16_t leaftype; int keycompare; int result; if(in_vol==NULL || in_dir==0 || out_numchildren==NULL) return 1; result = 1; buffer = NULL; extents = NULL; lastnode = 0; recs = NULL; recsizes = NULL; *out_numchildren = 0; if(out_children!=NULL) *out_children = NULL; if(out_childnames!=NULL) *out_childnames = NULL; buffer = hfslib_malloc(in_vol->chr.node_size, cbargs); if(buffer==NULL) HFS_LIBERR("could not allocate node buffer"); numextents = hfslib_get_file_extents(in_vol, HFS_CNID_CATALOG, HFS_DATAFORK, &extents, cbargs); if(numextents==0) HFS_LIBERR("could not locate fork extents"); nd.num_recs = 0; curnode = in_vol->chr.root_node; while(1) { hfslib_free_recs(&recs, &recsizes, &nd.num_recs, cbargs); recnum = 0; if(hfslib_readd_with_extents(in_vol, buffer, &bytesread, in_vol->chr.node_size, curnode * in_vol->chr.node_size, extents, numextents, cbargs)!=0) HFS_LIBERR("could not read catalog node #%i", curnode); if(hfslib_reada_node(buffer, &nd, &recs, &recsizes, HFS_CATALOG_FILE, in_vol, cbargs)==0) HFS_LIBERR("could not parse catalog node #%i", curnode); for(recnum=0; recnum 0) { /* Descend to child node if we found an exact match, or if * this is the last pointer record. */ curnode = currec.child; break; } lastnode = currec.child; } else { /* * We have now descended down the hierarchy of index nodes into * the leaf node that contains the first catalog record with a * matching parent CNID. Since all leaf nodes are chained * through their flink/blink, we can simply walk forward through * this chain, copying every matching non-thread record, until * we hit a record with a different parent CNID. At that point, * we've retrieved all of our directory's items, if any. */ curnode = nd.flink; if(curkey.parent_cnidvh.attributes & (1<jh.start == in_vol->jh.end); } /* * hfslib_is_private_file() * * Given a file/folder's key and parent CNID, determines if it should be hidden * from the user (e.g., the journal header file or the HFS+ Private Data folder) */ int hfslib_is_private_file(hfs_catalog_key_t *filekey) { hfs_catalog_key_t* curkey = NULL; int i = 0; /* * According to the HFS+ spec to date, all special objects are located in * the root directory of the volume, so don't bother going further if the * requested object is not. */ if(filekey->parent_cnid != HFS_CNID_ROOT_FOLDER) return 0; while((curkey = hfs_gPrivateObjectKeys[i]) != NULL) { /* XXX Always use binary compare here, or use volume's specific key * XXX comparison routine? */ if(filekey->name.length == curkey->name.length && memcmp(filekey->name.unicode, curkey->name.unicode, 2 * curkey->name.length)==0) return 1; i++; } return 0; } /* bool hfslib_is_journal_valid(hfs_volume* in_vol) { - check magic numbers - check Other Things }*/ #if 0 #pragma mark - #pragma mark Major Structures #endif /* * hfslib_read_volume_header() * * Reads in_bytes, formats the data appropriately, and places the result * in out_header, which is assumed to be previously allocated. Returns number * of bytes read, 0 if failed. */ size_t hfslib_read_volume_header(void* in_bytes, hfs_volume_header_t* out_header) { void* ptr; size_t last_bytes_read; int i; if(in_bytes==NULL || out_header==NULL) return 0; ptr = in_bytes; out_header->signature = be16tohp(&ptr); out_header->version = be16tohp(&ptr); out_header->attributes = be32tohp(&ptr); out_header->last_mounting_version = be32tohp(&ptr); out_header->journal_info_block = be32tohp(&ptr); out_header->date_created = be32tohp(&ptr); out_header->date_modified = be32tohp(&ptr); out_header->date_backedup = be32tohp(&ptr); out_header->date_checked = be32tohp(&ptr); out_header->file_count = be32tohp(&ptr); out_header->folder_count = be32tohp(&ptr); out_header->block_size = be32tohp(&ptr); out_header->total_blocks = be32tohp(&ptr); out_header->free_blocks = be32tohp(&ptr); out_header->next_alloc_block = be32tohp(&ptr); out_header->rsrc_clump_size = be32tohp(&ptr); out_header->data_clump_size = be32tohp(&ptr); out_header->next_cnid = be32tohp(&ptr); out_header->write_count = be32tohp(&ptr); out_header->encodings = be64tohp(&ptr); for(i=0;i<8;i++) out_header->finder_info[i] = be32tohp(&ptr); if((last_bytes_read = hfslib_read_fork_descriptor(ptr, &out_header->allocation_file))==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; if((last_bytes_read = hfslib_read_fork_descriptor(ptr, &out_header->extents_file))==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; if((last_bytes_read = hfslib_read_fork_descriptor(ptr, &out_header->catalog_file))==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; if((last_bytes_read = hfslib_read_fork_descriptor(ptr, &out_header->attributes_file))==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; if((last_bytes_read = hfslib_read_fork_descriptor(ptr, &out_header->startup_file))==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; return ((uint8_t*)ptr - (uint8_t*)in_bytes); } /* * hfslib_reada_node() * * Given the pointer to and size of a buffer containing the entire, raw * contents of any b-tree node from the disk, this function will: * * 1. determine the type of node and read its contents * 2. allocate memory for each record and fill it appropriately * 3. set out_record_ptrs_array to point to an array (which it allocates) * which has out_node_descriptor->num_recs many pointers to the * records themselves * 4. allocate out_record_ptr_sizes_array and fill it with the sizes of * each record * 5. return the number of bytes read (i.e., the size of the node) * or 0 on failure * * out_node_descriptor must be allocated by the caller and may not be NULL. * * out_record_ptrs_array and out_record_ptr_sizes_array must both be specified, * or both be NULL if the caller is not interested in reading the records. * * out_record_ptr_sizes_array may be NULL if the caller is not interested in * reading the records, but must not be NULL if out_record_ptrs_array is not. * * in_parent_file is HFS_CATALOG_FILE, HFS_EXTENTS_FILE, or * HFS_ATTRIBUTES_FILE, depending on the special file in which this node * resides. * * inout_volume must have its catnodesize or extnodesize field (depending on * the parent file) set to the correct value if this is an index, leaf, or map * node. If this is a header node, the field will be set to its correct value. */ size_t hfslib_reada_node(void* in_bytes, hfs_node_descriptor_t* out_node_descriptor, void** out_record_ptrs_array[], uint16_t* out_record_ptr_sizes_array[], hfs_btree_file_type in_parent_file, hfs_volume* inout_volume, hfs_callback_args* cbargs) { void* ptr; uint16_t* rec_offsets; size_t last_bytes_read; uint16_t nodesize; uint16_t numrecords; uint16_t free_space_offset; /* offset to free space in node */ int keysizefieldsize; int i; numrecords = 0; rec_offsets = NULL; if(out_record_ptrs_array!=NULL) *out_record_ptrs_array = NULL; if(out_record_ptr_sizes_array!=NULL) *out_record_ptr_sizes_array = NULL; if(in_bytes==NULL || inout_volume==NULL || out_node_descriptor==NULL || (out_record_ptrs_array==NULL && out_record_ptr_sizes_array!=NULL) || (out_record_ptrs_array!=NULL && out_record_ptr_sizes_array==NULL) ) goto error; ptr = in_bytes; out_node_descriptor->flink = be32tohp(&ptr); out_node_descriptor->blink = be32tohp(&ptr); out_node_descriptor->kind = *(((int8_t*)ptr)); ptr = (uint8_t*)ptr + 1; out_node_descriptor->height = *(((uint8_t*)ptr)); ptr = (uint8_t*)ptr + 1; out_node_descriptor->num_recs = be16tohp(&ptr); out_node_descriptor->reserved = be16tohp(&ptr); numrecords = out_node_descriptor->num_recs; /* * To go any further, we will need to know the size of this node, as well * as the width of keyed records' key_len parameters for this btree. If * this is an index, leaf, or map node, inout_volume already has the node * size set in its catnodesize or extnodesize field and the key length set * in the catkeysizefieldsize or extkeysizefieldsize for catalog files and * extent files, respectively. However, if this is a header node, this * information has not yet been determined, so this is the place to do it. */ if(out_node_descriptor->kind == HFS_HEADERNODE) { hfs_header_record_t hr; void* header_rec_offset[1]; uint16_t header_rec_size[1]; /* sanity check to ensure this is a good header node */ if(numrecords!=3) HFS_LIBERR("header node does not have exactly 3 records"); header_rec_offset[0] = ptr; header_rec_size[0] = sizeof(hfs_header_record_t); last_bytes_read = hfslib_read_header_node(header_rec_offset, header_rec_size, 1, &hr, NULL, NULL); if(last_bytes_read==0) HFS_LIBERR("could not read header node"); switch(in_parent_file) { case HFS_CATALOG_FILE: inout_volume->chr.node_size = hr.node_size; inout_volume->catkeysizefieldsize = (hr.attributes & HFS_BIG_KEYS_MASK) ? sizeof(uint16_t):sizeof(uint8_t); break; case HFS_EXTENTS_FILE: inout_volume->ehr.node_size = hr.node_size; inout_volume->extkeysizefieldsize = (hr.attributes & HFS_BIG_KEYS_MASK) ? sizeof(uint16_t):sizeof(uint8_t); break; case HFS_ATTRIBUTES_FILE: default: HFS_LIBERR("invalid parent file type specified"); /* NOTREACHED */ } } switch(in_parent_file) { case HFS_CATALOG_FILE: nodesize = inout_volume->chr.node_size; keysizefieldsize = inout_volume->catkeysizefieldsize; break; case HFS_EXTENTS_FILE: nodesize = inout_volume->ehr.node_size; keysizefieldsize = inout_volume->extkeysizefieldsize; break; case HFS_ATTRIBUTES_FILE: default: HFS_LIBERR("invalid parent file type specified"); /* NOTREACHED */ } /* * Don't care about records so just exit after getting the node descriptor. * Note: This happens after the header node code, and not before it, in * case the caller calls this function and ignores the record data just to * get at the node descriptor, but then tries to call it again on a non- * header node without first setting inout_volume->cat/extnodesize. */ if(out_record_ptrs_array==NULL) return ((uint8_t*)ptr - (uint8_t*)in_bytes); rec_offsets = hfslib_malloc(numrecords * sizeof(uint16_t), cbargs); *out_record_ptr_sizes_array = hfslib_malloc(numrecords * sizeof(uint16_t), cbargs); if(rec_offsets==NULL || *out_record_ptr_sizes_array==NULL) HFS_LIBERR("could not allocate node record offsets"); *out_record_ptrs_array = hfslib_malloc(numrecords * sizeof(void*), cbargs); if(*out_record_ptrs_array==NULL) HFS_LIBERR("could not allocate node records"); last_bytes_read = hfslib_reada_node_offsets((uint8_t*)in_bytes + nodesize - numrecords * sizeof(uint16_t), rec_offsets); if(last_bytes_read==0) HFS_LIBERR("could not read node record offsets"); /* The size of the last record (i.e. the first one listed in the offsets) * must be determined using the offset to the node's free space. */ free_space_offset = be16toh(*(uint16_t*)((uint8_t*)in_bytes + nodesize - (numrecords+1) * sizeof(uint16_t))); (*out_record_ptr_sizes_array)[numrecords-1] = free_space_offset - rec_offsets[0]; for(i=1;ikind == HFS_LEAFNODE || out_node_descriptor->kind == HFS_INDEXNODE) { hfs_catalog_key_t reckey; uint16_t rectype; rectype = out_node_descriptor->kind; last_bytes_read = hfslib_read_catalog_keyed_record(ptr, NULL, &rectype, &reckey, inout_volume); if(last_bytes_read==0) HFS_LIBERR("could not read node record"); if((reckey.key_len + keysizefieldsize) % 2 == 1) { ptr = (uint8_t*)ptr + 1; (*out_record_ptr_sizes_array)[i]--; } if((*out_record_ptr_sizes_array)[i] % 2 == 1) (*out_record_ptr_sizes_array)[i]--; } memcpy((*out_record_ptrs_array)[i], ptr, (*out_record_ptr_sizes_array)[i]); ptr = (uint8_t*)ptr + (*out_record_ptr_sizes_array)[i]; } goto exit; error: hfslib_free_recs(out_record_ptrs_array, out_record_ptr_sizes_array, &numrecords, cbargs); ptr = in_bytes; /* warn("error occurred in hfslib_reada_node()"); */ /* FALLTHROUGH */ exit: if(rec_offsets!=NULL) hfslib_free(rec_offsets, cbargs); return ((uint8_t*)ptr - (uint8_t*)in_bytes); } /* * hfslib_reada_node_offsets() * * Sets out_offset_array to contain the offsets to each record in the node, * in reverse order. Does not read the free space offset. */ size_t hfslib_reada_node_offsets(void* in_bytes, uint16_t* out_offset_array) { void* ptr; if(in_bytes==NULL || out_offset_array==NULL) return 0; ptr = in_bytes; /* * The offset for record 0 (which is the very last offset in the node) is * always equal to 14, the size of the node descriptor. So, once we hit * offset=14, we know this is the last offset. In this way, we don't need * to know the number of records beforehand. */ out_offset_array--; do { out_offset_array++; *out_offset_array = be16tohp(&ptr); } while(*out_offset_array != (uint16_t)14); return ((uint8_t*)ptr - (uint8_t*)in_bytes); } /* hfslib_read_header_node() * * out_header_record and/or out_map_record may be NULL if the caller doesn't * care about their contents. */ size_t hfslib_read_header_node(void** in_recs, uint16_t* in_rec_sizes, uint16_t in_num_recs, hfs_header_record_t* out_hr, void* out_userdata, void* out_map) { void* ptr; int i; if(in_recs==NULL || in_rec_sizes==NULL) return 0; if(out_hr!=NULL) { ptr = in_recs[0]; out_hr->tree_depth = be16tohp(&ptr); out_hr->root_node = be32tohp(&ptr); out_hr->leaf_recs = be32tohp(&ptr); out_hr->first_leaf = be32tohp(&ptr); out_hr->last_leaf = be32tohp(&ptr); out_hr->node_size = be16tohp(&ptr); out_hr->max_key_len = be16tohp(&ptr); out_hr->total_nodes = be32tohp(&ptr); out_hr->free_nodes = be32tohp(&ptr); out_hr->reserved = be16tohp(&ptr); out_hr->clump_size = be32tohp(&ptr); out_hr->btree_type = *(((uint8_t*)ptr)); ptr = (uint8_t*)ptr + 1; out_hr->keycomp_type = *(((uint8_t*)ptr)); ptr = (uint8_t*)ptr + 1; out_hr->attributes = be32tohp(&ptr); for(i=0;i<16;i++) out_hr->reserved2[i] = be32tohp(&ptr); } if(out_userdata!=NULL) { memcpy(out_userdata, in_recs[1], in_rec_sizes[1]); } ptr = (uint8_t*)ptr + in_rec_sizes[1]; /* size of user data record */ if(out_map!=NULL) { memcpy(out_map, in_recs[2], in_rec_sizes[2]); } ptr = (uint8_t*)ptr + in_rec_sizes[2]; /* size of map record */ return ((uint8_t*)ptr - (uint8_t*)in_recs[0]); } /* * hfslib_read_catalog_keyed_record() * * out_recdata can be NULL. inout_rectype must be set to either HFS_LEAFNODE * or HFS_INDEXNODE upon calling this function, and will be set by the * function to one of HFS_REC_FLDR, HFS_REC_FILE, HFS_REC_FLDR_THREAD, or * HFS_REC_FLDR_THREAD upon return if the node is a leaf node. If it is an * index node, inout_rectype will not be changed. */ size_t hfslib_read_catalog_keyed_record( void* in_bytes, hfs_catalog_keyed_record_t* out_recdata, int16_t* inout_rectype, hfs_catalog_key_t* out_key, hfs_volume* in_volume) { void* ptr; size_t last_bytes_read; if(in_bytes==NULL || out_key==NULL || inout_rectype==NULL) return 0; ptr = in_bytes; /* For HFS+, the key length is always a 2-byte number. This is indicated * by the HFS_BIG_KEYS_MASK bit in the attributes field of the catalog * header record. However, we just assume this bit is set, since all HFS+ * volumes should have it set anyway. */ if(in_volume->catkeysizefieldsize == sizeof(uint16_t)) out_key->key_len = be16tohp(&ptr); else if (in_volume->catkeysizefieldsize == sizeof(uint8_t)) { out_key->key_len = *(((uint8_t*)ptr)); ptr = (uint8_t*)ptr + 1; } out_key->parent_cnid = be32tohp(&ptr); last_bytes_read = hfslib_read_unistr255(ptr, &out_key->name); if(last_bytes_read==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; /* don't waste time if the user just wanted the key and/or record type */ if(out_recdata==NULL) { if(*inout_rectype == HFS_LEAFNODE) *inout_rectype = be16tohp(&ptr); else if(*inout_rectype != HFS_INDEXNODE) return 0; /* should not happen if we were given valid arguments */ return ((uint8_t*)ptr - (uint8_t*)in_bytes); } if(*inout_rectype == HFS_INDEXNODE) { out_recdata->child = be32tohp(&ptr); } else { /* first need to determine what kind of record this is */ *inout_rectype = be16tohp(&ptr); out_recdata->type = *inout_rectype; switch(out_recdata->type) { case HFS_REC_FLDR: { out_recdata->folder.flags = be16tohp(&ptr); out_recdata->folder.valence = be32tohp(&ptr); out_recdata->folder.cnid = be32tohp(&ptr); out_recdata->folder.date_created = be32tohp(&ptr); out_recdata->folder.date_content_mod = be32tohp(&ptr); out_recdata->folder.date_attrib_mod = be32tohp(&ptr); out_recdata->folder.date_accessed = be32tohp(&ptr); out_recdata->folder.date_backedup = be32tohp(&ptr); last_bytes_read = hfslib_read_bsd_data(ptr, &out_recdata->folder.bsd); if(last_bytes_read==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; last_bytes_read = hfslib_read_folder_userinfo(ptr, &out_recdata->folder.user_info); if(last_bytes_read==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; last_bytes_read = hfslib_read_folder_finderinfo(ptr, &out_recdata->folder.finder_info); if(last_bytes_read==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; out_recdata->folder.text_encoding = be32tohp(&ptr); out_recdata->folder.reserved = be32tohp(&ptr); } break; case HFS_REC_FILE: { out_recdata->file.flags = be16tohp(&ptr); out_recdata->file.reserved = be32tohp(&ptr); out_recdata->file.cnid = be32tohp(&ptr); out_recdata->file.date_created = be32tohp(&ptr); out_recdata->file.date_content_mod = be32tohp(&ptr); out_recdata->file.date_attrib_mod = be32tohp(&ptr); out_recdata->file.date_accessed = be32tohp(&ptr); out_recdata->file.date_backedup = be32tohp(&ptr); last_bytes_read = hfslib_read_bsd_data(ptr, &out_recdata->file.bsd); if(last_bytes_read==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; last_bytes_read = hfslib_read_file_userinfo(ptr, &out_recdata->file.user_info); if(last_bytes_read==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; last_bytes_read = hfslib_read_file_finderinfo(ptr, &out_recdata->file.finder_info); if(last_bytes_read==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; out_recdata->file.text_encoding = be32tohp(&ptr); out_recdata->file.reserved2 = be32tohp(&ptr); last_bytes_read = hfslib_read_fork_descriptor(ptr, &out_recdata->file.data_fork); if(last_bytes_read==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; last_bytes_read = hfslib_read_fork_descriptor(ptr, &out_recdata->file.rsrc_fork); if(last_bytes_read==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; } break; case HFS_REC_FLDR_THREAD: case HFS_REC_FILE_THREAD: { out_recdata->thread.reserved = be16tohp(&ptr); out_recdata->thread.parent_cnid = be32tohp(&ptr); last_bytes_read = hfslib_read_unistr255(ptr, &out_recdata->thread.name); if(last_bytes_read==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; } break; default: return 1; /* NOTREACHED */ } } return ((uint8_t*)ptr - (uint8_t*)in_bytes); } /* out_rec may be NULL */ size_t hfslib_read_extent_record( void* in_bytes, hfs_extent_record_t* out_rec, hfs_node_kind in_nodekind, hfs_extent_key_t* out_key, hfs_volume* in_volume) { void* ptr; size_t last_bytes_read; if(in_bytes==NULL || out_key==NULL || (in_nodekind!=HFS_LEAFNODE && in_nodekind!=HFS_INDEXNODE)) return 0; ptr = in_bytes; /* For HFS+, the key length is always a 2-byte number. This is indicated * by the HFS_BIG_KEYS_MASK bit in the attributes field of the extent * overflow header record. However, we just assume this bit is set, since * all HFS+ volumes should have it set anyway. */ if(in_volume->extkeysizefieldsize == sizeof(uint16_t)) out_key->key_length = be16tohp(&ptr); else if (in_volume->extkeysizefieldsize == sizeof(uint8_t)) { out_key->key_length = *(((uint8_t*)ptr)); ptr = (uint8_t*)ptr + 1; } out_key->fork_type = *(((uint8_t*)ptr)); ptr = (uint8_t*)ptr + 1; out_key->padding = *(((uint8_t*)ptr)); ptr = (uint8_t*)ptr + 1; out_key->file_cnid = be32tohp(&ptr); out_key->start_block = be32tohp(&ptr); /* don't waste time if the user just wanted the key */ if(out_rec==NULL) return ((uint8_t*)ptr - (uint8_t*)in_bytes); if(in_nodekind==HFS_LEAFNODE) { last_bytes_read = hfslib_read_extent_descriptors(ptr, out_rec); if(last_bytes_read==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; } else { /* XXX: this is completely bogus */ /* (uint32_t*)*out_rec = be32tohp(&ptr); */ uint32_t *ptr_32 = (uint32_t *)out_rec; *ptr_32 = be32tohp(&ptr); /* (*out_rec)[0].start_block = be32tohp(&ptr); */ } return ((uint8_t*)ptr - (uint8_t*)in_bytes); } void hfslib_free_recs( void*** inout_node_recs, uint16_t** inout_rec_sizes, uint16_t* inout_num_recs, hfs_callback_args* cbargs) { uint16_t i; if(inout_num_recs==NULL || *inout_num_recs==0) return; if(inout_node_recs!=NULL && *inout_node_recs!=NULL) { for(i=0;i<*inout_num_recs;i++) { if((*inout_node_recs)[i]!=NULL) { hfslib_free((*inout_node_recs)[i], cbargs); (*inout_node_recs)[i] = NULL; } } hfslib_free(*inout_node_recs, cbargs); *inout_node_recs = NULL; } if(inout_rec_sizes!=NULL && *inout_rec_sizes!=NULL) { hfslib_free(*inout_rec_sizes, cbargs); *inout_rec_sizes = NULL; } *inout_num_recs = 0; } #if 0 #pragma mark - #pragma mark Individual Fields #endif size_t hfslib_read_fork_descriptor(void* in_bytes, hfs_fork_t* out_forkdata) { void* ptr; size_t last_bytes_read; if(in_bytes==NULL || out_forkdata==NULL) return 0; ptr = in_bytes; out_forkdata->logical_size = be64tohp(&ptr); out_forkdata->clump_size = be32tohp(&ptr); out_forkdata->total_blocks = be32tohp(&ptr); if((last_bytes_read = hfslib_read_extent_descriptors(ptr, &out_forkdata->extents))==0) return 0; ptr = (uint8_t*)ptr + last_bytes_read; return ((uint8_t*)ptr - (uint8_t*)in_bytes); } size_t hfslib_read_extent_descriptors( void* in_bytes, hfs_extent_record_t* out_extentrecord) { void* ptr; int i; if(in_bytes==NULL || out_extentrecord==NULL) return 0; ptr = in_bytes; for(i=0;i<8;i++) { (((hfs_extent_descriptor_t*)*out_extentrecord)[i]).start_block = be32tohp(&ptr); (((hfs_extent_descriptor_t*)*out_extentrecord)[i]).block_count = be32tohp(&ptr); } return ((uint8_t*)ptr - (uint8_t*)in_bytes); } size_t hfslib_read_unistr255(void* in_bytes, hfs_unistr255_t* out_string) { void* ptr; uint16_t i, length; if(in_bytes==NULL || out_string==NULL) return 0; ptr = in_bytes; length = be16tohp(&ptr); if(length>255) length = 255; /* hfs+ folder/file names have a limit of 255 chars */ out_string->length = length; for(i=0; iunicode[i] = be16tohp(&ptr); } return ((uint8_t*)ptr - (uint8_t*)in_bytes); } size_t hfslib_read_bsd_data(void* in_bytes, hfs_bsd_data_t* out_perms) { void* ptr; if(in_bytes==NULL || out_perms==NULL) return 0; ptr = in_bytes; out_perms->owner_id = be32tohp(&ptr); out_perms->group_id = be32tohp(&ptr); out_perms->admin_flags = *(((uint8_t*)ptr)); ptr = (uint8_t*)ptr + 1; out_perms->owner_flags = *(((uint8_t*)ptr)); ptr = (uint8_t*)ptr + 1; out_perms->file_mode = be16tohp(&ptr); out_perms->special.inode_num = be32tohp(&ptr); /* this field is a union */ return ((uint8_t*)ptr - (uint8_t*)in_bytes); } size_t hfslib_read_file_userinfo(void* in_bytes, hfs_macos_file_info_t* out_info) { void* ptr; if(in_bytes==NULL || out_info==NULL) return 0; ptr = in_bytes; out_info->file_type = be32tohp(&ptr); out_info->file_creator = be32tohp(&ptr); out_info->finder_flags = be16tohp(&ptr); out_info->location.v = be16tohp(&ptr); out_info->location.h = be16tohp(&ptr); out_info->reserved = be16tohp(&ptr); return ((uint8_t*)ptr - (uint8_t*)in_bytes); } size_t hfslib_read_file_finderinfo( void* in_bytes, hfs_macos_extended_file_info_t* out_info) { void* ptr; if(in_bytes==NULL || out_info==NULL) return 0; ptr = in_bytes; #if 0 #pragma warn Fill in with real code! #endif /* FIXME: Fill in with real code! */ memset(out_info, 0, sizeof(*out_info)); ptr = (uint8_t*)ptr + sizeof(hfs_macos_extended_file_info_t); return ((uint8_t*)ptr - (uint8_t*)in_bytes); } size_t hfslib_read_folder_userinfo(void* in_bytes, hfs_macos_folder_info_t* out_info) { void* ptr; if(in_bytes==NULL || out_info==NULL) return 0; ptr = in_bytes; #if 0 #pragma warn Fill in with real code! #endif /* FIXME: Fill in with real code! */ memset(out_info, 0, sizeof(*out_info)); ptr = (uint8_t*)ptr + sizeof(hfs_macos_folder_info_t); return ((uint8_t*)ptr - (uint8_t*)in_bytes); } size_t hfslib_read_folder_finderinfo( void* in_bytes, hfs_macos_extended_folder_info_t* out_info) { void* ptr; if(in_bytes==NULL || out_info==NULL) return 0; ptr = in_bytes; #if 0 #pragma warn Fill in with real code! #endif /* FIXME: Fill in with real code! */ memset(out_info, 0, sizeof(*out_info)); ptr = (uint8_t*)ptr + sizeof(hfs_macos_extended_folder_info_t); return ((uint8_t*)ptr - (uint8_t*)in_bytes); } size_t hfslib_read_journal_info(void* in_bytes, hfs_journal_info_t* out_info) { void* ptr; int i; if(in_bytes==NULL || out_info==NULL) return 0; ptr = in_bytes; out_info->flags = be32tohp(&ptr); for(i=0; i<8; i++) { out_info->device_signature[i] = be32tohp(&ptr); } out_info->offset = be64tohp(&ptr); out_info->size = be64tohp(&ptr); for(i=0; i<32; i++) { out_info->reserved[i] = be64tohp(&ptr); } return ((uint8_t*)ptr - (uint8_t*)in_bytes); } size_t hfslib_read_journal_header(void* in_bytes, hfs_journal_header_t* out_header) { void* ptr; if(in_bytes==NULL || out_header==NULL) return 0; ptr = in_bytes; out_header->magic = be32tohp(&ptr); out_header->endian = be32tohp(&ptr); out_header->start = be64tohp(&ptr); out_header->end = be64tohp(&ptr); out_header->size = be64tohp(&ptr); out_header->blocklist_header_size = be32tohp(&ptr); out_header->checksum = be32tohp(&ptr); out_header->journal_header_size = be32tohp(&ptr); return ((uint8_t*)ptr - (uint8_t*)in_bytes); } #if 0 #pragma mark - #pragma mark Disk Access #endif /* * hfslib_readd_with_extents() * * This function reads the contents of a file from the volume, given an array * of extent descriptors which specify where every extent of the file is * located (in addition to the usual pread() arguments). out_bytes is presumed * to exist and be large enough to hold in_length number of bytes. Returns 0 * on success. */ int hfslib_readd_with_extents( hfs_volume* in_vol, void* out_bytes, uint64_t* out_bytesread, uint64_t in_length, uint64_t in_offset, hfs_extent_descriptor_t in_extents[], uint16_t in_numextents, hfs_callback_args* cbargs) { uint64_t ext_length, last_offset; uint16_t i; int error; if(in_vol==NULL || out_bytes==NULL || in_extents==NULL || in_numextents==0 || out_bytesread==NULL) return -1; *out_bytesread = 0; last_offset = 0; for(i=0; ivh.block_size; if(in_offset < last_offset+ext_length && in_offset+in_length >= last_offset) { uint64_t isect_start, isect_end; isect_start = max(in_offset, last_offset); isect_end = min(in_offset+in_length, last_offset+ext_length); error = hfslib_readd(in_vol, out_bytes, isect_end-isect_start, isect_start - last_offset + (uint64_t)in_extents[i].start_block * in_vol->vh.block_size, cbargs); if(error!=0) return error; *out_bytesread += isect_end-isect_start; out_bytes = (uint8_t*)out_bytes + isect_end-isect_start; } last_offset += ext_length; } return 0; } #if 0 #pragma mark - #pragma mark Callback Wrappers #endif void hfslib_error(const char* in_format, const char* in_file, int in_line, ...) { va_list ap; if(in_format==NULL) return; if(hfs_gcb.error!=NULL) { va_start(ap, in_line); hfs_gcb.error(in_format, in_file, in_line, ap); va_end(ap); } } void* hfslib_malloc(size_t size, hfs_callback_args* cbargs) { if(hfs_gcb.allocmem!=NULL) return hfs_gcb.allocmem(size, cbargs); return NULL; } void* hfslib_realloc(void* ptr, size_t size, hfs_callback_args* cbargs) { if(hfs_gcb.reallocmem!=NULL) return hfs_gcb.reallocmem(ptr, size, cbargs); return NULL; } void hfslib_free(void* ptr, hfs_callback_args* cbargs) { if(hfs_gcb.freemem!=NULL && ptr!=NULL) hfs_gcb.freemem(ptr, cbargs); } int hfslib_openvoldevice( hfs_volume* in_vol, const char* in_device, uint64_t in_offset, hfs_callback_args* cbargs) { if(hfs_gcb.openvol!=NULL && in_device!=NULL) return hfs_gcb.openvol(in_vol, in_device, in_offset, cbargs); return 1; } void hfslib_closevoldevice(hfs_volume* in_vol, hfs_callback_args* cbargs) { if(hfs_gcb.closevol!=NULL) hfs_gcb.closevol(in_vol, cbargs); } int hfslib_readd( hfs_volume* in_vol, void* out_bytes, uint64_t in_length, uint64_t in_offset, hfs_callback_args* cbargs) { if(in_vol==NULL || out_bytes==NULL) return -1; if(hfs_gcb.read!=NULL) return hfs_gcb.read(in_vol, out_bytes, in_length, in_offset, cbargs); return -1; } #if 0 #pragma mark - #pragma mark Other #endif /* returns key length */ uint16_t hfslib_make_catalog_key( hfs_cnid_t in_parent_cnid, uint16_t in_name_len, unichar_t* in_unicode, hfs_catalog_key_t* out_key) { if(in_parent_cnid==0 || (in_name_len>0 && in_unicode==NULL) || out_key==0) return 0; if(in_name_len>255) in_name_len = 255; out_key->key_len = 6 + 2 * in_name_len; out_key->parent_cnid = in_parent_cnid; out_key->name.length = in_name_len; if(in_name_len>0) memcpy(&out_key->name.unicode, in_unicode, in_name_len*2); return out_key->key_len; } /* returns key length */ uint16_t hfslib_make_extent_key( hfs_cnid_t in_cnid, uint8_t in_forktype, uint32_t in_startblock, hfs_extent_key_t* out_key) { if(in_cnid==0 || out_key==0) return 0; out_key->key_length = HFS_MAX_EXT_KEY_LEN; out_key->fork_type = in_forktype; out_key->padding = 0; out_key->file_cnid = in_cnid; out_key->start_block = in_startblock; return out_key->key_length; } /* case-folding */ int hfslib_compare_catalog_keys_cf ( const void *ap, const void *bp) { const hfs_catalog_key_t *a, *b; unichar_t ac, bc; /* current character from a, b */ unichar_t lc; /* lowercase version of current character */ uint8_t apos, bpos; /* current character indices */ a = (const hfs_catalog_key_t*)ap; b = (const hfs_catalog_key_t*)bp; if(a->parent_cnid != b->parent_cnid) { return (a->parent_cnid - b->parent_cnid); } else { /* * The following code implements the pseudocode suggested by * the HFS+ technote. */ /* * XXX These need to be revised to be endian-independent! */ #define hbyte(x) ((x) >> 8) #define lbyte(x) ((x) & 0x00FF) apos = bpos = 0; while(1) { /* get next valid character from a */ for (lc=0; lc == 0 && apos < a->name.length; apos++) { ac = a->name.unicode[apos]; lc = hfs_gcft[hbyte(ac)]; if(lc==0) lc = ac; else lc = hfs_gcft[lc + lbyte(ac)]; }; ac=lc; /* get next valid character from b */ for (lc=0; lc == 0 && bpos < b->name.length; bpos++) { bc = b->name.unicode[bpos]; lc = hfs_gcft[hbyte(bc)]; if(lc==0) lc = bc; else lc = hfs_gcft[lc + lbyte(bc)]; }; bc=lc; /* on end of string ac/bc are 0, otherwise > 0 */ if (ac != bc || (ac == 0 && bc == 0)) return ac - bc; } #undef hbyte #undef lbyte } } /* binary compare (i.e., not case folding) */ int hfslib_compare_catalog_keys_bc ( const void *a, const void *b) { if(((const hfs_catalog_key_t*)a)->parent_cnid == ((const hfs_catalog_key_t*)b)->parent_cnid) { if(((const hfs_catalog_key_t*)a)->name.length == 0 && ((const hfs_catalog_key_t*)b)->name.length == 0) return 0; if(((const hfs_catalog_key_t*)a)->name.length == 0) return -1; if(((const hfs_catalog_key_t*)b)->name.length == 0) return 1; /* FIXME: This does a byte-per-byte comparison, whereas the HFS spec * mandates a uint16_t chunk comparison. */ return memcmp(((const hfs_catalog_key_t*)a)->name.unicode, ((const hfs_catalog_key_t*)b)->name.unicode, min(((const hfs_catalog_key_t*)a)->name.length, ((const hfs_catalog_key_t*)b)->name.length)); } else { return (((const hfs_catalog_key_t*)a)->parent_cnid - ((const hfs_catalog_key_t*)b)->parent_cnid); } } int hfslib_compare_extent_keys ( const void *a, const void *b) { /* * Comparison order, in descending importance: * * CNID -> fork type -> start block */ if(((const hfs_extent_key_t*)a)->file_cnid == ((const hfs_extent_key_t*)b)->file_cnid) { if(((const hfs_extent_key_t*)a)->fork_type == ((const hfs_extent_key_t*)b)->fork_type) { if(((const hfs_extent_key_t*)a)->start_block == ((const hfs_extent_key_t*)b)->start_block) { return 0; } else { return (((const hfs_extent_key_t*)a)->start_block - ((const hfs_extent_key_t*)b)->start_block); } } else { return (((const hfs_extent_key_t*)a)->fork_type - ((const hfs_extent_key_t*)b)->fork_type); } } else { return (((const hfs_extent_key_t*)a)->file_cnid - ((const hfs_extent_key_t*)b)->file_cnid); } } /* 1+10 tables of 16 rows and 16 columns, each 2 bytes wide = 5632 bytes */ int hfslib_create_casefolding_table(void) { hfs_callback_args cbargs; unichar_t* t; /* convenience */ uint16_t s; /* current subtable * 256 */ uint16_t i; /* current subtable index (0 to 255) */ if(hfs_gcft!=NULL) return 0; /* no sweat, table already exists */ hfslib_init_cbargs(&cbargs); hfs_gcft = hfslib_malloc(5632, &cbargs); if(hfs_gcft==NULL) HFS_LIBERR("could not allocate case folding table"); t = hfs_gcft; /* easier to type :) */ /* * high byte indices */ s = 0 * 256; memset(t, 0x00, 512); t[s+ 0] = 0x0100; t[s+ 1] = 0x0200; t[s+ 3] = 0x0300; t[s+ 4] = 0x0400; t[s+ 5] = 0x0500; t[s+ 16] = 0x0600; t[s+ 32] = 0x0700; t[s+ 33] = 0x0800; t[s+254] = 0x0900; t[s+255] = 0x0a00; /* * table 1 (high byte 0x00) */ s = 1 * 256; for(i=0; i<65; i++) t[s+i] = i; t[s+ 0] = 0xffff; for(i=65; i<91; i++) t[s+i] = i + 0x20; for(i=91; i<256; i++) t[s+i] = i; t[s+198] = 0x00e6; t[s+208] = 0x00f0; t[s+216] = 0x00f8; t[s+222] = 0x00fe; /* * table 2 (high byte 0x01) */ s = 2 * 256; for(i=0; i<256; i++) t[s+i] = i + 0x0100; t[s+ 16] = 0x0111; t[s+ 38] = 0x0127; t[s+ 50] = 0x0133; t[s+ 63] = 0x0140; t[s+ 65] = 0x0142; t[s+ 74] = 0x014b; t[s+ 82] = 0x0153; t[s+102] = 0x0167; t[s+129] = 0x0253; t[s+130] = 0x0183; t[s+132] = 0x0185; t[s+134] = 0x0254; t[s+135] = 0x0188; t[s+137] = 0x0256; t[s+138] = 0x0257; t[s+139] = 0x018c; t[s+142] = 0x01dd; t[s+143] = 0x0259; t[s+144] = 0x025b; t[s+145] = 0x0192; t[s+147] = 0x0260; t[s+148] = 0x0263; t[s+150] = 0x0269; t[s+151] = 0x0268; t[s+152] = 0x0199; t[s+156] = 0x026f; t[s+157] = 0x0272; t[s+159] = 0x0275; t[s+162] = 0x01a3; t[s+164] = 0x01a5; t[s+167] = 0x01a8; t[s+169] = 0x0283; t[s+172] = 0x01ad; t[s+174] = 0x0288; t[s+177] = 0x028a; t[s+178] = 0x028b; t[s+179] = 0x01b4; t[s+181] = 0x01b6; t[s+183] = 0x0292; t[s+184] = 0x01b9; t[s+188] = 0x01bd; t[s+196] = 0x01c6; t[s+197] = 0x01c6; t[s+199] = 0x01c9; t[s+200] = 0x01c9; t[s+202] = 0x01cc; t[s+203] = 0x01cc; t[s+228] = 0x01e5; t[s+241] = 0x01f3; t[s+242] = 0x01f3; /* * table 3 (high byte 0x03) */ s = 3 * 256; for(i=0; i<145; i++) t[s+i] = i + 0x0300; for(i=145; i<170; i++) t[s+i] = i + 0x0320; t[s+162] = 0x03a2; for(i=170; i<256; i++) t[s+i] = i + 0x0300; for(i=226; i<239; i+=2) t[s+i] = i + 0x0301; /* * table 4 (high byte 0x04) */ s = 4 * 256; for(i=0; i<16; i++) t[s+i] = i + 0x0400; t[s+ 2] = 0x0452; t[s+ 4] = 0x0454; t[s+ 5] = 0x0455; t[s+ 6] = 0x0456; t[s+ 8] = 0x0458; t[s+ 9] = 0x0459; t[s+ 10] = 0x045a; t[s+ 11] = 0x045b; t[s+ 15] = 0x045f; for(i=16; i<48; i++) t[s+i] = i + 0x0420; t[s+ 25] = 0x0419; for(i=48; i<256; i++) t[s+i] = i + 0x0400; t[s+195] = 0x04c4; t[s+199] = 0x04c8; t[s+203] = 0x04cc; for(i=96; i<129; i+=2) t[s+i] = i + 0x0401; t[s+118] = 0x0476; for(i=144; i<191; i+=2) t[s+i] = i + 0x0401; /* * table 5 (high byte 0x05) */ s = 5 * 256; for(i=0; i<49; i++) t[s+i] = i + 0x0500; for(i=49; i<87; i++) t[s+i] = i + 0x0530; for(i=87; i<256; i++) t[s+i] = i + 0x0500; /* * table 6 (high byte 0x10) */ s = 6 * 256; for(i=0; i<160; i++) t[s+i] = i + 0x1000; for(i=160; i<198; i++) t[s+i] = i + 0x1030; for(i=198; i<256; i++) t[s+i] = i + 0x1000; /* * table 7 (high byte 0x20) */ s = 7 * 256; for(i=0; i<256; i++) t[s+i] = i + 0x2000; { uint8_t zi[15] = { 12, 13, 14, 15, 42, 43, 44, 45, 46, 106, 107, 108, 109, 110, 111}; for(i=0; i<15; i++) t[s+zi[i]] = 0x0000; } /* * table 8 (high byte 0x21) */ s = 8 * 256; for(i=0; i<96; i++) t[s+i] = i + 0x2100; for(i=96; i<112; i++) t[s+i] = i + 0x2110; for(i=112; i<256; i++) t[s+i] = i + 0x2100; /* * table 9 (high byte 0xFE) */ s = 9 * 256; for(i=0; i<256; i++) t[s+i] = i + 0xFE00; t[s+255] = 0x0000; /* * table 10 (high byte 0xFF) */ s = 10 * 256; for(i=0; i<33; i++) t[s+i] = i + 0xFF00; for(i=33; i<59; i++) t[s+i] = i + 0xFF20; for(i=59; i<256; i++) t[s+i] = i + 0xFF00; return 0; error: return 1; } int hfslib_get_hardlink(hfs_volume *vol, uint32_t inode_num, hfs_catalog_keyed_record_t *rec, hfs_callback_args *cbargs) { hfs_catalog_keyed_record_t metadata; hfs_catalog_key_t key; char name[16]; unichar_t name_uni[16]; int i, len; /* XXX: cache this */ if (hfslib_find_catalog_record_with_key(vol, &hfs_gMetadataDirectoryKey, &metadata, cbargs) != 0 || metadata.type != HFS_REC_FLDR) return -1; len = snprintf(name, sizeof(name), "iNode%d", inode_num); for (i=0; i