blob: 5585b45c67bd717a27919df353eb8ab4649c5844 [file] [log] [blame]
# Copyright (c) 2014 The Chromium OS Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import os, fcntl, logging, struct, random
from autotest_lib.client.bin import site_utils, test, utils
from autotest_lib.client.common_lib import error
class hardware_TrimIntegrity(test.test):
"""
Performs data integrity trim test on an unmounted partition.
This test will write 1 GB of data and verify that trimmed data are gone and
untrimmed data are unaffected. The verification will be run in 5 passes with
0%, 25%, 50%, 75%, and 100% of data trimmed.
Also, perform 4K random read QD32 before and after trim. We should see some
speed / latency difference if the device firmware trim data properly.
"""
version = 1
FILE_SIZE = 1024 * 1024 * 1024
CHUNK_SIZE = 64 * 1024
TRIM_RATIO = [0, 0.25, 0.5, 0.75, 1]
# Use hash value to check integrity of the random data.
HASH_CMD = 'sha256sum | cut -d" " -f 1'
# 0x1277 is ioctl BLKDISCARD command
IOCTL_TRIM_CMD = 0x1277
IOCTL_NOT_SUPPORT_ERRNO = 95
def _get_hash(self, chunk_count, chunk_size):
"""
Get hash for every chunk of data.
"""
cmd = str('for i in $(seq 0 %d); do dd if=%s of=/dev/stdout bs=%d'
' count=1 skip=$i iflag=direct | %s; done' %
(chunk_count - 1, self._filename, chunk_size, self.HASH_CMD))
return utils.run(cmd).stdout.split()
def _do_trim(self, fd, offset, size):
"""
Invoke ioctl to trim command.
"""
fcntl.ioctl(fd, self.IOCTL_TRIM_CMD, struct.pack('QQ', offset, size))
def run_once(self, filename=None, file_size=FILE_SIZE,
chunk_size=CHUNK_SIZE, trim_ratio=TRIM_RATIO):
"""
Executes the test and logs the output.
@param file_name: file/disk name to test
default: spare partition of internal disk
@param file_size: size of data to test. default: 1GB
@param chunk_size: size of chunk to calculate hash/trim. default: 64KB
@param trim_ratio: list of ratio of file size to trim data
default: [0, 0.25, 0.5, 0.75, 1]
"""
if not filename:
self._filename = site_utils.get_fixed_dst_drive()
if self._filename == site_utils.get_root_device():
self._filename = site_utils.get_free_root_partition()
logging.debug("target device: %s", self._filename)
else:
self._filename = filename
if file_size == 0:
fulldisk = True
file_size = site_utils.get_disk_size(self._filename)
if file_size == 0:
cmd = ('%s seem to have 0 storage block. Is the media present?'
% filename)
raise error.TestError(cmd)
else:
fulldisk = False
# Make file size multiple of 4 * chunk size
file_size -= file_size % (4 * chunk_size)
if fulldisk:
fio_file_size = 0
else:
fio_file_size = file_size
logging.info('filename: %s, filesize: %d', self._filename, file_size)
# Check for trim support in ioctl. Raise TestNAError if not support.
try:
fd = os.open(self._filename, os.O_RDWR, 0666)
self._do_trim(fd, 0, chunk_size)
except IOError, err:
if err.errno == self.IOCTL_NOT_SUPPORT_ERRNO:
raise error.TestNAError("IOCTL Does not support trim.")
else:
raise
finally:
os.close(fd)
# Calculate hash value for zero'ed and one'ed data
cmd = str('dd if=/dev/zero bs=%d count=1 | %s' %
(chunk_size, self.HASH_CMD))
zero_hash = utils.run(cmd).stdout.strip()
cmd = str("dd if=/dev/zero bs=%d count=1 | tr '\\0' '\\xff' | %s" %
(chunk_size, self.HASH_CMD))
one_hash = utils.run(cmd).stdout.strip()
trim_hash = ""
# Write random data to disk
chunk_count = file_size / chunk_size
cmd = str('dd if=/dev/urandom of=%s bs=%d count=%d oflag=direct' %
(self._filename, chunk_size, chunk_count))
utils.run(cmd)
ref_hash = self._get_hash(chunk_count, chunk_size)
# Check read speed/latency when reading real data.
self.job.run_test('hardware_StorageFio',
filesize=fio_file_size,
requirements=[('4k_read_qd32', [])],
tag='before_trim')
# Generate random order of chunk to trim
trim_order = list(range(0, chunk_count))
random.shuffle(trim_order)
trim_status = [False] * chunk_count
# Init stat variable
data_verify_count = 0
data_verify_match = 0
trim_verify_count = 0
trim_verify_zero = 0
trim_verify_one = 0
trim_verify_non_delete = 0
trim_deterministic = True
last_ratio = 0
for ratio in trim_ratio:
# Do trim
begin_trim_chunk = int(last_ratio * chunk_count)
end_trim_chunk = int(ratio * chunk_count)
fd = os.open(self._filename, os.O_RDWR, 0666)
for chunk in trim_order[begin_trim_chunk:end_trim_chunk]:
self._do_trim(fd, chunk * chunk_size, chunk_size)
trim_status[chunk] = True
os.close(fd)
last_ratio = ratio
cur_hash = self._get_hash(chunk_count, chunk_size)
trim_verify_count += int(ratio * chunk_count)
data_verify_count += chunk_count - int(ratio * chunk_count)
# Verify hash
for cur, ref, trim in zip(cur_hash, ref_hash, trim_status):
if trim:
if not trim_hash:
trim_hash = cur
elif cur != trim_hash:
trim_deterministic = False
if cur == zero_hash:
trim_verify_zero += 1
elif cur == one_hash:
trim_verify_one += 1
elif cur == ref:
trim_verify_non_delete += 1
else:
if cur == ref:
data_verify_match += 1
keyval = dict()
keyval['data_verify_count'] = data_verify_count
keyval['data_verify_match'] = data_verify_match
keyval['trim_verify_count'] = trim_verify_count
keyval['trim_verify_zero'] = trim_verify_zero
keyval['trim_verify_one'] = trim_verify_one
keyval['trim_verify_non_delete'] = trim_verify_non_delete
keyval['trim_deterministic'] = trim_deterministic
self.write_perf_keyval(keyval)
# Check read speed/latency when reading from trimmed data.
self.job.run_test('hardware_StorageFio',
filesize=fio_file_size,
requirements=[('4k_read_qd32', [])],
tag='after_trim')
# Raise error when untrimmed data changed only.
# Don't care about trimmed data.
if data_verify_match < data_verify_count:
raise error.TestFail("Fail to verify untrimmed data.")
if trim_verify_non_delete > 0 :
raise error.TestFail("Trimmed data are not deleted.")