btrfs-progs/btrfs-debugfs

297 lines
9.3 KiB
Python
Executable File

#!/usr/bin/env python2
#
# Simple python program to print out all the extents of a single file
# LGPLv2 license
# Copyright Facebook 2014
import sys,os,struct,fcntl,ctypes,stat
# helpers for max ints
maxu64 = (1L << 64) - 1
maxu32 = (1L << 32) - 1
# the inode (like form stat)
BTRFS_INODE_ITEM_KEY = 1
# backref to the directory
BTRFS_INODE_REF_KEY = 12
# backref to the directory v2
BTRFS_INODE_EXTREF_KEY = 13
# xattr items
BTRFS_XATTR_ITEM_KEY = 24
# orphans for list files
BTRFS_ORPHAN_ITEM_KEY = 48
# treelog items for dirs
BTRFS_DIR_LOG_ITEM_KEY = 60
BTRFS_DIR_LOG_INDEX_KEY = 72
# dir items and dir indexes both hold filenames
BTRFS_DIR_ITEM_KEY = 84
BTRFS_DIR_INDEX_KEY = 96
# these are the file extent pointers
BTRFS_EXTENT_DATA_KEY = 108
# csums
BTRFS_EXTENT_CSUM_KEY = 128
# root item for subvols and snapshots
BTRFS_ROOT_ITEM_KEY = 132
# root item backrefs
BTRFS_ROOT_BACKREF_KEY = 144
BTRFS_ROOT_REF_KEY = 156
# each allocated extent has an extent item
BTRFS_EXTENT_ITEM_KEY = 168
# optimized extents for metadata only
BTRFS_METADATA_ITEM_KEY = 169
# backrefs for extents
BTRFS_TREE_BLOCK_REF_KEY = 176
BTRFS_EXTENT_DATA_REF_KEY = 178
BTRFS_EXTENT_REF_V0_KEY = 180
BTRFS_SHARED_BLOCK_REF_KEY = 182
BTRFS_SHARED_DATA_REF_KEY = 184
# one of these for each block group
BTRFS_BLOCK_GROUP_ITEM_KEY = 192
# dev extents records which part of each device is allocated
BTRFS_DEV_EXTENT_KEY = 204
# dev items describe devs
BTRFS_DEV_ITEM_KEY = 216
# one for each chunk
BTRFS_CHUNK_ITEM_KEY = 228
# qgroup info
BTRFS_QGROUP_STATUS_KEY = 240
BTRFS_QGROUP_INFO_KEY = 242
BTRFS_QGROUP_LIMIT_KEY = 244
BTRFS_QGROUP_RELATION_KEY = 246
# records balance progress
BTRFS_BALANCE_ITEM_KEY = 248
# stats on device errors
BTRFS_DEV_STATS_KEY = 249
BTRFS_DEV_REPLACE_KEY = 250
BTRFS_STRING_ITEM_KEY = 253
# in the kernel sources, this is flattened
# btrfs_ioctl_search_args_v2. It includes both the btrfs_ioctl_search_key
# and the buffer. We're using a 64K buffer size.
#
args_buffer_size = 65536
class btrfs_ioctl_search_args(ctypes.Structure):
_pack_ = 1
_fields_ = [ ("tree_id", ctypes.c_ulonglong),
("min_objectid", ctypes.c_ulonglong),
("max_objectid", ctypes.c_ulonglong),
("min_offset", ctypes.c_ulonglong),
("max_offset", ctypes.c_ulonglong),
("min_transid", ctypes.c_ulonglong),
("max_transid", ctypes.c_ulonglong),
("min_type", ctypes.c_uint),
("max_type", ctypes.c_uint),
("nr_items", ctypes.c_uint),
("unused", ctypes.c_uint),
("unused1", ctypes.c_ulonglong),
("unused2", ctypes.c_ulonglong),
("unused3", ctypes.c_ulonglong),
("unused4", ctypes.c_ulonglong),
("buf_size", ctypes.c_ulonglong),
("buf", ctypes.c_ubyte * args_buffer_size),
]
# the search ioctl resturns one header for each item
#
class btrfs_ioctl_search_header(ctypes.Structure):
_pack_ = 1
_fields_ = [ ("transid", ctypes.c_ulonglong),
("objectid", ctypes.c_ulonglong),
("offset", ctypes.c_ulonglong),
("type", ctypes.c_uint),
("len", ctypes.c_uint),
]
# the type field in btrfs_file_extent_item
BTRFS_FILE_EXTENT_INLINE = 0
BTRFS_FILE_EXTENT_REG = 1
BTRFS_FILE_EXTENT_PREALLOC = 2
class btrfs_file_extent_item(ctypes.LittleEndianStructure):
_pack_ = 1
_fields_ = [ ("generation", ctypes.c_ulonglong),
("ram_bytes", ctypes.c_ulonglong),
("compression", ctypes.c_ubyte),
("encryption", ctypes.c_ubyte),
("other_encoding", ctypes.c_ubyte * 2),
("type", ctypes.c_ubyte),
("disk_bytenr", ctypes.c_ulonglong),
("disk_num_bytes", ctypes.c_ulonglong),
("offset", ctypes.c_ulonglong),
("num_bytes", ctypes.c_ulonglong),
]
class btrfs_ioctl_search():
def __init__(self):
self.args = btrfs_ioctl_search_args()
self.args.tree_id = 0
self.args.min_objectid = 0
self.args.max_objectid = maxu64
self.args.min_offset = 0
self.args.max_offset = maxu64
self.args.min_transid = 0
self.args.max_transid = maxu64
self.args.min_type = 0
self.args.max_type = maxu32
self.args.nr_items = 0
self.args.buf_size = args_buffer_size
# magic encoded for x86_64 this is the v2 search ioctl
self.ioctl_num = 3228603409L
# the results of the search get stored into args.buf
def search(self, fd, nritems=65536):
self.args.nr_items = nritems
fcntl.ioctl(fd, self.ioctl_num, self.args, 1)
# this moves the search key forward by one. If the end result is
# still a valid search key (all mins less than all maxes), we return
# True. Otherwise False
#
def advance_search(search):
if search.args.min_offset < maxu64:
search.args.min_offset += 1
elif search.args.min_type < 255:
search.args.min_type += 1
elif search.args.min_objectid < maxu64:
search.args.min_objectid += 1
else:
return False
if search.args.min_offset > search.args.max_offset:
return False
if search.args.min_type > search.args.max_type:
return False
if search.args.min_objectid > search.args.max_objectid:
return False
return True
# given one search_header and one file_item, print the details. This
# also tosses the [disk_bytenr,disk_num_bytes] into extent_hash to record
# which extents were used by this file
#
def print_one_extent(header, fi, extent_hash):
# we're ignoring inline items for now
if fi.type == BTRFS_FILE_EXTENT_INLINE:
# header.len is the length of the item returned. We subtract
# the part of the file item header that is actually used (21 bytes)
# and we get the length of the inlined data.
# this may or may not be compressed
inline_len = header.len - 21
if fi.compression:
ram_bytes = fi.ram_bytes
else:
ram_bytes = inline_len
print "(%Lu %Lu): ram %Lu disk 0 disk_size %Lu -- inline" % \
(header.objectid, header.offset, ram_bytes, inline_len)
extent_hash[-1] = inline_len
return
if fi.disk_bytenr == 0:
tag = " -- hole"
else:
tag = ""
print "(%Lu %Lu): ram %Lu disk %Lu disk_size %Lu%s" % (header.objectid,
header.offset, fi.num_bytes, fi.disk_bytenr, fi.disk_num_bytes, tag)
if fi.disk_bytenr:
extent_hash[fi.disk_bytenr] = fi.disk_num_bytes
# open 'filename' and run the search ioctl against it, printing all the extents
# we find
def print_file_extents(filename):
extent_hash = {}
s = btrfs_ioctl_search()
s.args.min_type = BTRFS_EXTENT_DATA_KEY
s.args.max_type = BTRFS_EXTENT_DATA_KEY
try:
fd = os.open(filename, os.O_RDONLY)
st = os.fstat(fd)
except Exception, e:
sys.stderr.write("Failed to open %s (%s)\n" % (filename, e))
return -1
if not stat.S_ISREG(st.st_mode):
sys.stderr.write("%s not a regular file\n" % filename)
return 0
s.args.min_objectid = st.st_ino
s.args.max_objectid = st.st_ino
size = st.st_size
while True:
try:
s.search(fd)
except Exception, e:
sys.stderr.write("Search ioctl failed for %s (%s)\n" % (filename, e))
return -1
if s.args.nr_items == 0:
break
# p is the results buffer from the kernel
p = ctypes.addressof(s.args.buf)
header = btrfs_ioctl_search_header()
header_size = ctypes.sizeof(header)
h = ctypes.addressof(header)
p_left = args_buffer_size
for x in xrange(0, s.args.nr_items):
# for each item, copy the header from the buffer into
# our header struct.
ctypes.memmove(h, p, header_size)
p += header_size
p_left -= header_size
# this would be a kernel bug it shouldn't be sending malformed
# items
if p_left <= 0:
break
if header.type == BTRFS_EXTENT_DATA_KEY:
fi = btrfs_file_extent_item()
# this would also be a kernel bug
if p_left < ctypes.sizeof(fi):
break
# Copy the file item out of the results buffer
ctypes.memmove(ctypes.addressof(fi), p, ctypes.sizeof(fi))
print_one_extent(header, fi, extent_hash)
p += header.len
p_left -= header.len
if p_left <= 0:
break
s.args.min_offset = header.offset
if not advance_search(s):
break
total_on_disk = 0
total_extents = 0
for x in extent_hash.itervalues():
total_on_disk += x
total_extents += 1
# don't divide by zero
if total_on_disk == 0:
total_on_disk = 1
print "file: %s extents %Lu disk size %Lu logical size %Lu ratio %.2f" % \
(filename, total_extents, total_on_disk, st.st_size,
float(st.st_size) / float(total_on_disk))
return 0
if len(sys.argv) == 1:
sys.stderr.write("Usage: btrfs-debug filename ...\n")
sys.exit(1)
for f in sys.argv[1:]:
print_file_extents(f)