| #!/usr/bin/env python |
| # |
| # Copyright 2016 The Android Open Source Project |
| # |
| # Licensed under the Apache License, Version 2.0 (the "License"); |
| # you may not use this file except in compliance with the License. |
| # You may obtain a copy of the License at |
| # |
| # http://www.apache.org/licenses/LICENSE-2.0 |
| # |
| # Unless required by applicable law or agreed to in writing, software |
| # distributed under the License is distributed on an "AS IS" BASIS, |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| # See the License for the specific language governing permissions and |
| # limitations under the License. |
| |
| import os |
| import sys |
| import struct |
| |
| FAT_TABLE_START = 0x200 |
| DEL_MARKER = 0xe5 |
| ESCAPE_DEL_MARKER = 0x05 |
| |
| ATTRIBUTE_READ_ONLY = 0x1 |
| ATTRIBUTE_HIDDEN = 0x2 |
| ATTRIBUTE_SYSTEM = 0x4 |
| ATTRIBUTE_VOLUME_LABEL = 0x8 |
| ATTRIBUTE_SUBDIRECTORY = 0x10 |
| ATTRIBUTE_ARCHIVE = 0x20 |
| ATTRIBUTE_DEVICE = 0x40 |
| |
| LFN_ATTRIBUTES = \ |
| ATTRIBUTE_VOLUME_LABEL | \ |
| ATTRIBUTE_SYSTEM | \ |
| ATTRIBUTE_HIDDEN | \ |
| ATTRIBUTE_READ_ONLY |
| LFN_ATTRIBUTES_BYTE = struct.pack("B", LFN_ATTRIBUTES) |
| |
| MAX_CLUSTER_ID = 0x7FFF |
| |
| def read_le_short(f): |
| "Read a little-endian 2-byte integer from the given file-like object" |
| return struct.unpack("<H", f.read(2))[0] |
| |
| def read_le_long(f): |
| "Read a little-endian 4-byte integer from the given file-like object" |
| return struct.unpack("<L", f.read(4))[0] |
| |
| def read_byte(f): |
| "Read a 1-byte integer from the given file-like object" |
| return struct.unpack("B", f.read(1))[0] |
| |
| def skip_bytes(f, n): |
| "Fast-forward the given file-like object by n bytes" |
| f.seek(n, os.SEEK_CUR) |
| |
| def skip_short(f): |
| "Fast-forward the given file-like object 2 bytes" |
| skip_bytes(f, 2) |
| |
| def skip_byte(f): |
| "Fast-forward the given file-like object 1 byte" |
| skip_bytes(f, 1) |
| |
| def rewind_bytes(f, n): |
| "Rewind the given file-like object n bytes" |
| skip_bytes(f, -n) |
| |
| def rewind_short(f): |
| "Rewind the given file-like object 2 bytes" |
| rewind_bytes(f, 2) |
| |
| class fake_file(object): |
| """ |
| Interface for python file-like objects that we use to manipulate the image. |
| Inheritors must have an idx member which indicates the file pointer, and a |
| size member which indicates the total file size. |
| """ |
| |
| def seek(self, amount, direction=0): |
| "Implementation of seek from python's file-like object interface." |
| if direction == os.SEEK_CUR: |
| self.idx += amount |
| elif direction == os.SEEK_END: |
| self.idx = self.size - amount |
| else: |
| self.idx = amount |
| |
| if self.idx < 0: |
| self.idx = 0 |
| if self.idx > self.size: |
| self.idx = self.size |
| |
| class fat_file(fake_file): |
| """ |
| A file inside of our fat image. The file may or may not have a dentry, and |
| if it does this object knows nothing about it. All we see is a valid cluster |
| chain. |
| """ |
| |
| def __init__(self, fs, cluster, size=None): |
| """ |
| fs: The fat() object for the image this file resides in. |
| cluster: The first cluster of data for this file. |
| size: The size of this file. If not given, we use the total length of the |
| cluster chain that starts from the cluster argument. |
| """ |
| self.fs = fs |
| self.start_cluster = cluster |
| self.size = size |
| |
| if self.size is None: |
| self.size = fs.get_chain_size(cluster) |
| |
| self.idx = 0 |
| |
| def read(self, size): |
| "Read method for pythonic file-like interface." |
| if self.idx + size > self.size: |
| size = self.size - self.idx |
| got = self.fs.read_file(self.start_cluster, self.idx, size) |
| self.idx += len(got) |
| return got |
| |
| def write(self, data): |
| "Write method for pythonic file-like interface." |
| self.fs.write_file(self.start_cluster, self.idx, data) |
| self.idx += len(data) |
| |
| if self.idx > self.size: |
| self.size = self.idx |
| |
| def shorten(name, index): |
| """ |
| Create a file short name from the given long name (with the extension already |
| removed). The index argument gives a disambiguating integer to work into the |
| name to avoid collisions. |
| """ |
| name = "".join(name.split('.')).upper() |
| postfix = "~" + str(index) |
| return name[:8 - len(postfix)] + postfix |
| |
| class fat_dir(object): |
| "A directory in our fat filesystem." |
| |
| def __init__(self, backing): |
| """ |
| backing: A file-like object from which we can read dentry info. Should have |
| an fs member allowing us to get to the underlying image. |
| """ |
| self.backing = backing |
| self.dentries = [] |
| to_read = self.backing.size / 32 |
| |
| self.backing.seek(0) |
| |
| while to_read > 0: |
| (dent, consumed) = self.backing.fs.read_dentry(self.backing) |
| to_read -= consumed |
| |
| if dent: |
| self.dentries.append(dent) |
| |
| def __str__(self): |
| return "\n".join([str(x) for x in self.dentries]) + "\n" |
| |
| def add_dentry(self, attributes, shortname, ext, longname, first_cluster, |
| size): |
| """ |
| Add a new dentry to this directory. |
| attributes: Attribute flags for this dentry. See the ATTRIBUTE_ constants |
| above. |
| shortname: Short name of this file. Up to 8 characters, no dots. |
| ext: Extension for this file. Up to 3 characters, no dots. |
| longname: The long name for this file, with extension. Largely unrestricted. |
| first_cluster: The first cluster in the cluster chain holding the contents |
| of this file. |
| size: The size of this file. Set to 0 for subdirectories. |
| """ |
| new_dentry = dentry(self.backing.fs, attributes, shortname, ext, |
| longname, first_cluster, size) |
| new_dentry.commit(self.backing) |
| self.dentries.append(new_dentry) |
| return new_dentry |
| |
| def make_short_name(self, name): |
| """ |
| Given a long file name, return an 8.3 short name as a tuple. Name will be |
| engineered not to collide with other such names in this folder. |
| """ |
| parts = name.rsplit('.', 1) |
| |
| if len(parts) == 1: |
| parts.append('') |
| |
| name = parts[0] |
| ext = parts[1].upper() |
| |
| index = 1 |
| shortened = shorten(name, index) |
| |
| for dent in self.dentries: |
| assert dent.longname != name, "File must not exist" |
| if dent.shortname == shortened: |
| index += 1 |
| shortened = shorten(name, index) |
| |
| if len(name) <= 8 and len(ext) <= 3 and not '.' in name: |
| return (name.upper().ljust(8), ext.ljust(3)) |
| |
| return (shortened.ljust(8), ext[:3].ljust(3)) |
| |
| def new_file(self, name, data=None): |
| """ |
| Add a new regular file to this directory. |
| name: The name of the new file. |
| data: The contents of the new file. Given as a file-like object. |
| """ |
| size = 0 |
| if data: |
| data.seek(0, os.SEEK_END) |
| size = data.tell() |
| |
| chunk = self.backing.fs.allocate(size or 1) |
| (shortname, ext) = self.make_short_name(name) |
| self.add_dentry(0, shortname, ext, name, chunk, size) |
| |
| if data is None: |
| return |
| |
| data_file = fat_file(self.backing.fs, chunk, size) |
| data.seek(0) |
| data_file.write(data.read()) |
| |
| def new_subdirectory(self, name): |
| """ |
| Create a new subdirectory of this directory with the given name. |
| Returns a fat_dir(). |
| """ |
| chunk = self.backing.fs.allocate(1) |
| (shortname, ext) = self.make_short_name(name) |
| new_dentry = dentry(self.backing.fs, ATTRIBUTE_SUBDIRECTORY, |
| shortname, ext, name, chunk, 0) |
| new_dentry.commit(self.backing) |
| return new_dentry.open_directory() |
| |
| def lfn_checksum(name_data): |
| """ |
| Given the characters of an 8.3 file name (concatenated *without* the dot), |
| Compute a one-byte checksum which needs to appear in corresponding long file |
| name entries. |
| """ |
| assert len(name_data) == 11, "Name data should be exactly 11 characters" |
| name_data = struct.unpack("B" * 11, name_data) |
| |
| result = 0 |
| |
| for char in name_data: |
| last_bit = (result & 1) << 7 |
| result = (result >> 1) | last_bit |
| result += char |
| result = result & 0xFF |
| |
| return struct.pack("B", result) |
| |
| class dentry(object): |
| "A directory entry" |
| def __init__(self, fs, attributes, shortname, ext, longname, |
| first_cluster, size): |
| """ |
| fs: The fat() object for the image we're stored in. |
| attributes: The attribute flags for this dentry. See the ATTRIBUTE_ flags |
| above. |
| shortname: The short name stored in this dentry. Up to 8 characters, no |
| dots. |
| ext: The file extension stored in this dentry. Up to 3 characters, no |
| dots. |
| longname: The long file name stored in this dentry. |
| first_cluster: The first cluster in the cluster chain backing the file |
| this dentry points to. |
| size: Size of the file this dentry points to. 0 for subdirectories. |
| """ |
| self.fs = fs |
| self.attributes = attributes |
| self.shortname = shortname |
| self.ext = ext |
| self.longname = longname |
| self.first_cluster = first_cluster |
| self.size = size |
| |
| def name(self): |
| "A friendly text file name for this dentry." |
| if self.longname: |
| return self.longname |
| |
| if not self.ext or len(self.ext) == 0: |
| return self.shortname |
| |
| return self.shortname + "." + self.ext |
| |
| def __str__(self): |
| return self.name() + " (" + str(self.size) + \ |
| " bytes @ " + str(self.first_cluster) + ")" |
| |
| def is_directory(self): |
| "Return whether this dentry points to a directory." |
| return (self.attributes & ATTRIBUTE_SUBDIRECTORY) != 0 |
| |
| def open_file(self): |
| "Open the target of this dentry if it is a regular file." |
| assert not self.is_directory(), "Cannot open directory as file" |
| return fat_file(self.fs, self.first_cluster, self.size) |
| |
| def open_directory(self): |
| "Open the target of this dentry if it is a directory." |
| assert self.is_directory(), "Cannot open file as directory" |
| return fat_dir(fat_file(self.fs, self.first_cluster)) |
| |
| def longname_records(self, checksum): |
| """ |
| Get the longname records necessary to store this dentry's long name, |
| packed as a series of 32-byte strings. |
| """ |
| if self.longname is None: |
| return [] |
| if len(self.longname) == 0: |
| return [] |
| |
| encoded_long_name = self.longname.encode('utf-16-le') |
| long_name_padding = "\0" * (26 - (len(encoded_long_name) % 26)) |
| padded_long_name = encoded_long_name + long_name_padding |
| |
| chunks = [padded_long_name[i:i+26] for i in range(0, |
| len(padded_long_name), 26)] |
| records = [] |
| sequence_number = 1 |
| |
| for c in chunks: |
| sequence_byte = struct.pack("B", sequence_number) |
| sequence_number += 1 |
| record = sequence_byte + c[:10] + LFN_ATTRIBUTES_BYTE + "\0" + \ |
| checksum + c[10:22] + "\0\0" + c[22:] |
| records.append(record) |
| |
| last = records.pop() |
| last_seq = struct.unpack("B", last[0])[0] |
| last_seq = last_seq | 0x40 |
| last = struct.pack("B", last_seq) + last[1:] |
| records.append(last) |
| records.reverse() |
| |
| return records |
| |
| def commit(self, f): |
| """ |
| Write this dentry into the given file-like object, |
| which is assumed to contain a FAT directory. |
| """ |
| f.seek(0) |
| padded_short_name = self.shortname.ljust(8) |
| padded_ext = self.ext.ljust(3) |
| name_data = padded_short_name + padded_ext |
| longname_record_data = self.longname_records(lfn_checksum(name_data)) |
| record = struct.pack("<11sBBBHHHHHHHL", |
| name_data, |
| self.attributes, |
| 0, |
| 0, |
| 0, |
| 0, |
| 0, |
| 0, |
| 0, |
| 0, |
| self.first_cluster, |
| self.size) |
| entry = "".join(longname_record_data + [record]) |
| |
| record_count = len(longname_record_data) + 1 |
| |
| found_count = 0 |
| |
| while True: |
| record = f.read(32) |
| |
| if record is None or len(record) != 32: |
| break |
| |
| marker = struct.unpack("B", record[0])[0] |
| |
| if marker == DEL_MARKER or marker == 0: |
| found_count += 1 |
| |
| if found_count == record_count: |
| break |
| else: |
| found_count = 0 |
| |
| if found_count != record_count: |
| f.write("\0" * self.fs.bytes_per_cluster) |
| f.seek(-self.fs.bytes_per_cluster, os.SEEK_CUR) |
| else: |
| f.seek(-(record_count * 32), os.SEEK_CUR) |
| f.write(entry) |
| |
| class root_dentry_file(fake_file): |
| """ |
| File-like object for the root directory. The root directory isn't stored in a |
| normal file, so we can't use a normal fat_file object to create a view of it. |
| """ |
| def __init__(self, fs): |
| self.fs = fs |
| self.idx = 0 |
| self.size = fs.root_entries * 32 |
| |
| def read(self, count): |
| f = self.fs.f |
| f.seek(self.fs.data_start() + self.idx) |
| |
| if self.idx + count > self.size: |
| count = self.size - self.idx |
| |
| ret = f.read(count) |
| self.idx += len(ret) |
| return ret |
| |
| def write(self, data): |
| f = self.fs.f |
| f.seek(self.fs.data_start() + self.idx) |
| |
| if self.idx + len(data) > self.size: |
| data = data[:self.size - self.idx] |
| |
| f.write(data) |
| self.idx += len(data) |
| if self.idx > self.size: |
| self.size = self.idx |
| |
| class fat(object): |
| "A FAT image" |
| |
| def __init__(self, path): |
| """ |
| path: Path to an image file containing a FAT file system. |
| """ |
| f = open(path, "r+b") |
| |
| self.f = f |
| |
| f.seek(0xb) |
| bytes_per_sector = read_le_short(f) |
| sectors_per_cluster = read_byte(f) |
| |
| self.bytes_per_cluster = bytes_per_sector * sectors_per_cluster |
| |
| reserved_sectors = read_le_short(f) |
| assert reserved_sectors == 1, \ |
| "Can only handle FAT with 1 reserved sector" |
| |
| fat_count = read_byte(f) |
| assert fat_count == 2, "Can only handle FAT with 2 tables" |
| |
| self.root_entries = read_le_short(f) |
| |
| skip_short(f) # Image size. Sort of. Useless field. |
| skip_byte(f) # Media type. We don't care. |
| |
| self.fat_size = read_le_short(f) * bytes_per_sector |
| self.root = fat_dir(root_dentry_file(self)) |
| |
| def data_start(self): |
| """ |
| Index of the first byte after the FAT tables. |
| """ |
| return FAT_TABLE_START + self.fat_size * 2 |
| |
| def get_chain_size(self, head_cluster): |
| """ |
| Return how many total bytes are in the cluster chain rooted at the given |
| cluster. |
| """ |
| if head_cluster == 0: |
| return 0 |
| |
| f = self.f |
| f.seek(FAT_TABLE_START + head_cluster * 2) |
| |
| cluster_count = 0 |
| |
| while head_cluster <= MAX_CLUSTER_ID: |
| cluster_count += 1 |
| head_cluster = read_le_short(f) |
| f.seek(FAT_TABLE_START + head_cluster * 2) |
| |
| return cluster_count * self.bytes_per_cluster |
| |
| def read_dentry(self, f=None): |
| """ |
| Read and decode a dentry from the given file-like object at its current |
| seek position. |
| """ |
| f = f or self.f |
| attributes = None |
| |
| consumed = 1 |
| |
| lfn_entries = {} |
| |
| while True: |
| skip_bytes(f, 11) |
| attributes = read_byte(f) |
| rewind_bytes(f, 12) |
| |
| if attributes & LFN_ATTRIBUTES != LFN_ATTRIBUTES: |
| break |
| |
| consumed += 1 |
| |
| seq = read_byte(f) |
| chars = f.read(10) |
| skip_bytes(f, 3) # Various hackish nonsense |
| chars += f.read(12) |
| skip_short(f) # Lots more nonsense |
| chars += f.read(4) |
| |
| chars = unicode(chars, "utf-16-le").encode("utf-8") |
| |
| lfn_entries[seq] = chars |
| |
| ind = read_byte(f) |
| |
| if ind == 0 or ind == DEL_MARKER: |
| skip_bytes(f, 31) |
| return (None, consumed) |
| |
| if ind == ESCAPE_DEL_MARKER: |
| ind = DEL_MARKER |
| |
| ind = str(unichr(ind)) |
| |
| if ind == '.': |
| skip_bytes(f, 31) |
| return (None, consumed) |
| |
| shortname = ind + f.read(7).rstrip() |
| ext = f.read(3).rstrip() |
| skip_bytes(f, 15) # Assorted flags, ctime/atime/mtime, etc. |
| first_cluster = read_le_short(f) |
| size = read_le_long(f) |
| |
| lfn = lfn_entries.items() |
| lfn.sort(key=lambda x: x[0]) |
| lfn = reduce(lambda x, y: x + y[1], lfn, "") |
| |
| if len(lfn) == 0: |
| lfn = None |
| else: |
| lfn = lfn.split('\0', 1)[0] |
| |
| return (dentry(self, attributes, shortname, ext, lfn, first_cluster, |
| size), consumed) |
| |
| def read_file(self, head_cluster, start_byte, size): |
| """ |
| Read from a given FAT file. |
| head_cluster: The first cluster in the file. |
| start_byte: How many bytes in to the file to begin the read. |
| size: How many bytes to read. |
| """ |
| f = self.f |
| |
| assert size >= 0, "Can't read a negative amount" |
| if size == 0: |
| return "" |
| |
| got_data = "" |
| |
| while True: |
| size_now = size |
| if start_byte + size > self.bytes_per_cluster: |
| size_now = self.bytes_per_cluster - start_byte |
| |
| if start_byte < self.bytes_per_cluster: |
| size -= size_now |
| |
| cluster_bytes_from_root = (head_cluster - 2) * \ |
| self.bytes_per_cluster |
| bytes_from_root = cluster_bytes_from_root + start_byte |
| bytes_from_data_start = bytes_from_root + self.root_entries * 32 |
| |
| f.seek(self.data_start() + bytes_from_data_start) |
| line = f.read(size_now) |
| got_data += line |
| |
| if size == 0: |
| return got_data |
| |
| start_byte -= self.bytes_per_cluster |
| |
| if start_byte < 0: |
| start_byte = 0 |
| |
| f.seek(FAT_TABLE_START + head_cluster * 2) |
| assert head_cluster <= MAX_CLUSTER_ID, "Out-of-bounds read" |
| head_cluster = read_le_short(f) |
| assert head_cluster > 0, "Read free cluster" |
| |
| return got_data |
| |
| def write_cluster_entry(self, entry): |
| """ |
| Write a cluster entry to the FAT table. Assumes our backing file is already |
| seeked to the correct entry in the first FAT table. |
| """ |
| f = self.f |
| f.write(struct.pack("<H", entry)) |
| skip_bytes(f, self.fat_size - 2) |
| f.write(struct.pack("<H", entry)) |
| rewind_bytes(f, self.fat_size) |
| |
| def allocate(self, amount): |
| """ |
| Allocate a new cluster chain big enough to hold at least the given amount |
| of bytes. |
| """ |
| f = self.f |
| f.seek(FAT_TABLE_START + 4) |
| |
| current = None |
| current_size = 0 |
| free_zones = {} |
| |
| pos = 2 |
| while pos < self.fat_size / 2: |
| data = read_le_short(f) |
| |
| if data == 0 and current is not None: |
| current_size += 1 |
| elif data == 0: |
| current = pos |
| current_size = 1 |
| elif current is not None: |
| free_zones[current] = current_size |
| current = None |
| |
| pos += 1 |
| |
| if current is not None: |
| free_zones[current] = current_size |
| |
| free_zones = free_zones.items() |
| free_zones.sort(key=lambda x: x[1]) |
| |
| grabbed_zones = [] |
| grabbed = 0 |
| |
| while grabbed < amount and len(free_zones) > 0: |
| zone = free_zones.pop() |
| grabbed += zone[1] * self.bytes_per_cluster |
| grabbed_zones.append(zone) |
| |
| if grabbed < amount: |
| return None |
| |
| excess = (grabbed - amount) / self.bytes_per_cluster |
| |
| grabbed_zones[-1] = (grabbed_zones[-1][0], |
| grabbed_zones[-1][1] - excess) |
| |
| out = None |
| grabbed_zones.reverse() |
| |
| for cluster, size in grabbed_zones: |
| entries = range(cluster + 1, cluster + size) |
| entries.append(out or 0xFFFF) |
| out = cluster |
| f.seek(FAT_TABLE_START + cluster * 2) |
| for entry in entries: |
| self.write_cluster_entry(entry) |
| |
| return out |
| |
| def extend_cluster(self, cluster, amount): |
| """ |
| Given a cluster which is the *last* cluster in a chain, extend it to hold |
| at least `amount` more bytes. |
| """ |
| return_cluster = None |
| f = self.f |
| |
| position = FAT_TABLE_START + cluster * 2 |
| f.seek(position) |
| |
| assert read_le_short(f) == 0xFFFF, "Extending from middle of chain" |
| rewind_short(f) |
| |
| while position + 2 < FAT_TABLE_START + self.fat_size and amount > 0: |
| skip_short(f) |
| got = read_le_short(f) |
| rewind_short(f) |
| rewind_short(f) |
| |
| if got != 0: |
| break |
| |
| cluster += 1 |
| return_cluster = return_cluster or cluster |
| position += 2 |
| self.write_cluster_entry(cluster) |
| |
| if amount < 0: |
| self.write_cluster_entry(0xFFFF) |
| return return_cluster |
| |
| new_chunk = self.allocate(amount) |
| f.seek(FAT_TABLE_START + cluster * 2) |
| self.write_cluster_entry(new_chunk) |
| |
| return return_cluster or new_chunk |
| |
| def write_file(self, head_cluster, start_byte, data): |
| """ |
| Write to a given FAT file. |
| |
| head_cluster: The first cluster in the file. |
| start_byte: How many bytes in to the file to begin the write. |
| data: The data to write. |
| """ |
| f = self.f |
| |
| while True: |
| if start_byte < self.bytes_per_cluster: |
| to_write = data[:self.bytes_per_cluster - start_byte] |
| data = data[self.bytes_per_cluster - start_byte:] |
| |
| cluster_bytes_from_root = (head_cluster - 2) * \ |
| self.bytes_per_cluster |
| bytes_from_root = cluster_bytes_from_root + start_byte |
| bytes_from_data_start = bytes_from_root + self.root_entries * 32 |
| |
| f.seek(self.data_start() + bytes_from_data_start) |
| f.write(to_write) |
| |
| if len(data) == 0: |
| return |
| |
| start_byte -= self.bytes_per_cluster |
| |
| if start_byte < 0: |
| start_byte = 0 |
| |
| f.seek(FAT_TABLE_START + head_cluster * 2) |
| next_cluster = read_le_short(f) |
| if next_cluster > MAX_CLUSTER_ID: |
| head_cluster = self.extend_cluster(head_cluster, len(data)) |
| else: |
| head_cluster = next_cluster |
| assert head_cluster > 0, "Cannot write free cluster" |
| |
| def add_item(directory, item): |
| """ |
| Copy a file into the given FAT directory. If the path given is a directory, |
| copy recursively. |
| directory: fat_dir to copy the file in to |
| item: Path of local file to copy |
| """ |
| if os.path.isdir(item): |
| base = os.path.basename(item) |
| if len(base) == 0: |
| base = os.path.basename(item[:-1]) |
| sub = directory.new_subdirectory(base) |
| for next_item in os.listdir(item): |
| add_item(sub, os.path.join(item, next_item)) |
| else: |
| with open(item, 'rb') as f: |
| directory.new_file(os.path.basename(item), f) |
| |
| if __name__ == "__main__": |
| if len(sys.argv) < 3: |
| print("Usage: fat16copy.py <image> <file> [<file> ...]") |
| print("Files are copied into the root of the image.") |
| print("Directories are copied recursively") |
| sys.exit(1) |
| |
| root = fat(sys.argv[1]).root |
| |
| for p in sys.argv[2:]: |
| add_item(root, p) |