aboutsummaryrefslogtreecommitdiff
path: root/testing/sgh_dd.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'testing/sgh_dd.cpp')
-rw-r--r--testing/sgh_dd.cpp5090
1 files changed, 5090 insertions, 0 deletions
diff --git a/testing/sgh_dd.cpp b/testing/sgh_dd.cpp
new file mode 100644
index 00000000..b0704ffa
--- /dev/null
+++ b/testing/sgh_dd.cpp
@@ -0,0 +1,5090 @@
+/*
+ * A utility program for copying files. Specialised for "files" that
+ * represent devices that understand the SCSI command set.
+ *
+ * Copyright (C) 2018-2022 D. Gilbert
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License as published by
+ * the Free Software Foundation; either version 2, or (at your option)
+ * any later version.
+ *
+ * SPDX-License-Identifier: GPL-2.0-or-later
+ *
+ * This program is a specialisation of the Unix "dd" command in which
+ * one or both of the given files is a scsi generic device.
+ * A logical block size ('bs') is assumed to be 512 if not given. This
+ * program complains if 'ibs' or 'obs' are given with some other value
+ * than 'bs'. If 'if' is not given or 'if=-' then stdin is assumed. If
+ * 'of' is not given or 'of=-' then stdout assumed.
+ *
+ * A non-standard argument "bpt" (blocks per transfer) is added to control
+ * the maximum number of blocks in each transfer. The default value is 128.
+ * For example if "bs=512" and "bpt=32" then a maximum of 32 blocks (16 KiB
+ * in this case) are transferred to or from the sg device in a single SCSI
+ * command.
+ *
+ * This version is designed for the linux kernel 2.4, 2.6, 3, 4 and 5 series.
+ *
+ * sgp_dd is a Posix threads specialization of the sg_dd utility. Both
+ * sgp_dd and sg_dd only perform special tasks when one or both of the given
+ * devices belong to the Linux sg driver.
+ *
+ * sgh_dd further extends sgp_dd to use the experimental kernel buffer
+ * sharing feature added in 3.9.02 .
+ * N.B. This utility was previously called sgs_dd but there was already an
+ * archived version of a dd variant called sgs_dd so this utility name was
+ * renamed [20181221]
+ */
+
+static const char * version_str = "2.22 20221020";
+
+#define _XOPEN_SOURCE 600
+#ifndef _GNU_SOURCE
+#define _GNU_SOURCE 1
+#endif
+
+#include <unistd.h>
+#include <fcntl.h>
+#include <stdio.h>
+#include <stdlib.h>
+#include <stdarg.h>
+#include <stdbool.h>
+#include <string.h>
+#include <ctype.h>
+#include <errno.h>
+#include <poll.h>
+#include <limits.h>
+#include <pthread.h>
+#include <signal.h>
+#define __STDC_FORMAT_MACROS 1
+#include <inttypes.h>
+#include <sys/ioctl.h>
+#include <sys/stat.h>
+#include <sys/sysmacros.h>
+#ifndef major
+#include <sys/types.h>
+#endif
+#include <sys/time.h>
+#include <linux/major.h> /* for MEM_MAJOR, SCSI_GENERIC_MAJOR, etc */
+#include <linux/fs.h> /* for BLKSSZGET and friends */
+#include <sys/mman.h> /* for mmap() system call */
+
+#include <vector>
+#include <array>
+#include <atomic> // C++ header replacing <stdatomic.h> also link
+ // needed '-l atomic' . Not anymore??
+#include <random>
+#include <thread> // needed for std::this_thread::yield()
+#include <mutex>
+#include <chrono>
+
+#ifdef HAVE_CONFIG_H
+#include "config.h"
+#endif
+
+#ifdef HAVE_GETRANDOM
+#include <sys/random.h> /* for getrandom() system call */
+#endif
+
+#ifndef HAVE_LINUX_SG_V4_HDR
+/* Kernel uapi header contain __user decorations on user space pointers
+ * to indicate they are unsafe in the kernel space. However glibc takes
+ * all those __user decorations out from headers in /usr/include/linux .
+ * So to stop compile errors when directly importing include/uapi/scsi/sg.h
+ * undef __user before doing that include. */
+#define __user
+
+/* Want to block the original sg.h header from also being included. That
+ * causes lots of multiple definition errors. This will only work if this
+ * header is included _before_ the original sg.h header. */
+#define _SCSI_GENERIC_H /* original kernel header guard */
+#define _SCSI_SG_H /* glibc header guard */
+
+#include "uapi_sg.h" /* local copy of include/uapi/scsi/sg.h */
+
+#else
+#define __user
+#endif /* end of: ifndef HAVE_LINUX_SG_V4_HDR */
+
+#include "sg_lib.h"
+#include "sg_cmds_basic.h"
+#include "sg_io_linux.h"
+#include "sg_unaligned.h"
+#include "sg_pr2serr.h"
+
+
+using namespace std;
+
+#ifdef __GNUC__
+#ifndef __clang__
+#pragma GCC diagnostic ignored "-Wclobbered"
+#endif
+#endif
+
+/* comment out following line to stop ioctl(SG_CTL_FLAGM_SNAP_DEV) */
+#define SGH_DD_SNAP_DEV 1
+
+#ifndef SGV4_FLAG_POLLED
+#define SGV4_FLAG_POLLED 0x800
+#endif
+
+#define DEF_BLOCK_SIZE 512
+#define DEF_BLOCKS_PER_TRANSFER 128
+#define DEF_BLOCKS_PER_2048TRANSFER 32
+#define DEF_SDT_ICT_MS 300
+#define DEF_SDT_CRT_SEC 3
+#define DEF_SCSI_CDBSZ 10
+#define MAX_SCSI_CDBSZ 16
+#define MAX_BPT_VALUE (1 << 24) /* used for maximum bs as well */
+#define MAX_COUNT_SKIP_SEEK (1LL << 48) /* coverity wants upper bound */
+
+#define SENSE_BUFF_LEN 64 /* Arbitrary, could be larger */
+#define READ_CAP_REPLY_LEN 8
+#define RCAP16_REPLY_LEN 32
+
+#define DEF_TIMEOUT 60000 /* 60,000 millisecs == 60 seconds */
+
+#define SGP_READ10 0x28
+#define SGP_PRE_FETCH10 0x34
+#define SGP_PRE_FETCH16 0x90
+#define SGP_VERIFY10 0x2f
+#define SGP_WRITE10 0x2a
+#define DEF_NUM_THREADS 4
+#define MAX_NUM_THREADS 1024 /* was SG_MAX_QUEUE with v3 driver */
+#define DEF_NUM_MRQS 0
+
+#define FT_OTHER 1 /* filetype other than one of the following */
+#define FT_SG 2 /* filetype is sg char device */
+#define FT_DEV_NULL 4 /* either /dev/null, /dev/zero or "." */
+#define FT_ST 8 /* filetype is st char device (tape) */
+#define FT_CHAR 16 /* filetype is st char device (tape) */
+#define FT_BLOCK 32 /* filetype is a block device */
+#define FT_FIFO 64 /* fifo (named or unnamed pipe (stdout)) */
+#define FT_RANDOM_0_FF 128 /* iflag=00, iflag=ff and iflag=random
+ override if=IFILE */
+#define FT_ERROR 256 /* couldn't "stat" file */
+
+#define DEV_NULL_MINOR_NUM 3
+#define DEV_ZERO_MINOR_NUM 5
+
+#define EBUFF_SZ 768
+
+#define PROC_SCSI_SG_VERSION "/proc/scsi/sg/version"
+#define SYS_SCSI_SG_VERSION "/sys/module/sg/version"
+
+struct flags_t {
+ bool append;
+ bool coe;
+ bool defres; /* without this res_sz==bs*bpt */
+ bool dio;
+ bool direct;
+ bool dpo;
+ bool dsync;
+ bool excl;
+ bool ff;
+ bool fua;
+ bool polled; /* formerly called 'hipri' */
+ bool masync; /* more async sg v4 driver flag */
+ bool mrq_immed; /* mrq submit non-blocking */
+ bool mrq_svb; /* mrq shared_variable_block, for sg->sg copy */
+ bool no_dur;
+ bool nocreat;
+ bool noshare;
+ bool no_thresh;
+ bool no_unshare; /* leave it for driver close/release */
+ bool no_waitq; /* dummy, no longer supported, just warn */
+ bool noxfer;
+ bool qhead;
+ bool qtail;
+ bool random;
+ bool mout_if; /* META_OUT_IF flag at mrq level */
+ bool same_fds;
+ bool swait; /* now ignore; kept for backward compatibility */
+ bool v3;
+ bool v4;
+ bool v4_given;
+ bool wq_excl;
+ bool zero;
+ int mmap;
+};
+
+struct global_collection
+{ /* one instance visible to all threads */
+ int infd;
+ int64_t skip;
+ int in_type;
+ int cdbsz_in;
+ int help;
+ int elem_sz;
+ struct flags_t in_flags;
+ // int64_t in_blk; /* -\ next block address to read */
+ // int64_t in_count; /* | blocks remaining for next read */
+ atomic<int64_t> in_rem_count; /* | count of remaining in blocks */
+ atomic<int> in_partial; /* | */
+ atomic<bool> in_stop; /* | */
+ off_t in_st_size; /* Only for FT_OTHER (regular) file */
+ pthread_mutex_t in_mutex; /* -/ */
+ int nmrqs; /* Number of multi-reqs for sg v4 */
+ int outfd;
+ int64_t seek;
+ int out_type;
+ int out2fd;
+ int out2_type;
+ int cdbsz_out;
+ int aen; /* abort every nth command */
+ int m_aen; /* abort mrq every nth command */
+ struct flags_t out_flags;
+ atomic<int64_t> out_blk; /* -\ next block address to write */
+ atomic<int64_t> out_count; /* | blocks remaining for next write */
+ atomic<int64_t> out_rem_count; /* | count of remaining out blocks */
+ atomic<int> out_partial; /* | */
+ atomic<bool> out_stop; /* | */
+ off_t out_st_size; /* Only for FT_OTHER (regular) file */
+ pthread_mutex_t out_mutex; /* | */
+ pthread_cond_t out_sync_cv; /* | hold writes until "in order" */
+ pthread_mutex_t out2_mutex;
+ int bs;
+ int bpt;
+ int cmd_timeout; /* in milliseconds */
+ int outregfd;
+ int outreg_type;
+ int ofsplit;
+ atomic<int> dio_incomplete_count;
+ atomic<int> sum_of_resids;
+ uint32_t sdt_ict; /* stall detection; initial check time (milliseconds) */
+ uint32_t sdt_crt; /* check repetition time (seconds), after first stall */
+ int fail_mask;
+ int verbose;
+ int dry_run;
+ int chkaddr;
+ bool aen_given;
+ bool cdbsz_given;
+ bool is_mrq_i;
+ bool is_mrq_o;
+ bool m_aen_given;
+ bool ofile_given;
+ bool ofile2_given;
+ bool unit_nanosec; /* default duration unit is millisecond */
+ bool mrq_cmds; /* mrq=<NRQS>,C given */
+ bool mrq_async; /* mrq_immed flag given */
+ bool noshare; /* don't use request sharing */
+ bool unbalanced_mrq; /* so _not_ sg->sg request sharing sync mrq */
+ bool verify; /* don't copy, verify like Unix: cmp */
+ bool prefetch; /* for verify: do PF(b),RD(a),V(b)_a_data */
+ bool unshare; /* let close() do file unshare operation */
+ const char * infp;
+ const char * outfp;
+ const char * out2fp;
+};
+
+typedef struct mrq_abort_info
+{
+ int from_tid;
+ int fd;
+ int mrq_id;
+ int debug;
+} Mrq_abort_info;
+
+typedef struct request_element
+{ /* one instance per worker thread */
+ struct global_collection *clp;
+ bool wr;
+ bool has_share;
+ bool both_sg;
+ bool same_sg;
+ bool only_in_sg;
+ bool only_out_sg;
+ // bool mrq_abort_thread_active;
+ int id;
+ int bs;
+ int infd;
+ int outfd;
+ int out2fd;
+ int outregfd;
+ int64_t iblk;
+ int64_t oblk;
+ int num_blks;
+ uint8_t * buffp;
+ uint8_t * alloc_bp;
+ struct sg_io_hdr io_hdr;
+ struct sg_io_v4 io_hdr4[2];
+ uint8_t cmd[MAX_SCSI_CDBSZ];
+ uint8_t sb[SENSE_BUFF_LEN];
+ int dio_incomplete_count;
+ int mmap_active;
+ int resid;
+ int rd_p_id;
+ int rep_count;
+ int rq_id;
+ int mmap_len;
+ int mrq_id;
+ int mrq_index;
+ uint32_t in_mrq_q_blks;
+ uint32_t out_mrq_q_blks;
+ long seed;
+#ifdef HAVE_SRAND48_R /* gcc extension. N.B. non-reentrant version slower */
+ struct drand48_data drand;/* opaque, used by srand48_r and mrand48_r */
+#endif
+ pthread_t mrq_abort_thread_id;
+ Mrq_abort_info mai;
+} Rq_elem;
+
+typedef struct thread_info
+{
+ int id;
+ struct global_collection * gcp;
+ pthread_t a_pthr;
+} Thread_info;
+
+/* Additional parameters for sg_start_io() and sg_finish_io() */
+struct sg_io_extra {
+ bool is_wr2;
+ bool prefetch;
+ bool dout_is_split;
+ int hpv4_ind;
+ int blk_offset;
+ int blks;
+};
+
+#define MONO_MRQ_ID_INIT 0x10000
+
+// typedef vector< pair<int, struct sg_io_v4> > mrq_arr_t;
+typedef array<uint8_t, 32> big_cdb; /* allow up to a 32 byte cdb */
+typedef pair< vector<struct sg_io_v4>, vector<big_cdb> > mrq_arr_t;
+
+
+/* Use this class to wrap C++11 <random> features to produce uniform random
+ * unsigned ints in the range [lo, hi] (inclusive) given a_seed */
+class Rand_uint {
+public:
+ Rand_uint(unsigned int lo, unsigned int hi, unsigned int a_seed)
+ : uid(lo, hi), dre(a_seed) { }
+ /* uid ctor takes inclusive range when integral type */
+
+ unsigned int get() { return uid(dre); }
+
+private:
+ uniform_int_distribution<unsigned int> uid;
+ default_random_engine dre;
+};
+
+static atomic<int> mono_pack_id(1);
+static atomic<int> mono_mrq_id(MONO_MRQ_ID_INIT);
+static atomic<long int> pos_index(0);
+
+static atomic<int> num_ebusy(0);
+static atomic<int> num_start_eagain(0);
+static atomic<int> num_fin_eagain(0);
+static atomic<int> num_abort_req(0);
+static atomic<int> num_abort_req_success(0);
+static atomic<int> num_mrq_abort_req(0);
+static atomic<int> num_mrq_abort_req_success(0);
+static atomic<int> num_miscompare(0);
+static atomic<long> num_waiting_calls(0);
+static atomic<bool> vb_first_time(true);
+static atomic<bool> shutting_down(false);
+
+static sigset_t signal_set;
+static sigset_t orig_signal_set;
+static pthread_t sig_listen_thread_id;
+
+static const char * sg_allow_dio = "/sys/module/sg/parameters/allow_dio";
+
+static void sg_in_rd_cmd(struct global_collection * clp, Rq_elem * rep,
+ mrq_arr_t & def_arr);
+static void sg_out_wr_cmd(Rq_elem * rep, mrq_arr_t & def_arr, bool is_wr2,
+ bool prefetch);
+static bool normal_in_rd(Rq_elem * rep, int blocks);
+static void normal_out_wr(Rq_elem * rep, int blocks);
+static int sg_start_io(Rq_elem * rep, mrq_arr_t & def_arr, int & pack_id,
+ struct sg_io_extra *xtrp);
+static int sg_finish_io(bool wr, Rq_elem * rep, int pack_id,
+ struct sg_io_extra *xtrp);
+static int sg_in_open(struct global_collection *clp, const char *inf,
+ uint8_t **mmpp, int *mmap_len);
+static int sg_out_open(struct global_collection *clp, const char *outf,
+ uint8_t **mmpp, int *mmap_len);
+static int sgh_do_deferred_mrq(Rq_elem * rep, mrq_arr_t & def_arr);
+
+#define STRERR_BUFF_LEN 128
+
+static pthread_mutex_t strerr_mut = PTHREAD_MUTEX_INITIALIZER;
+
+static bool have_sg_version = false;
+static int sg_version = 0;
+static bool sg_version_lt_4 = false;
+static bool sg_version_ge_40045 = false;
+static bool do_sync = false;
+static int do_time = 1;
+static struct global_collection gcoll;
+static struct timeval start_tm;
+static int64_t dd_count = -1;
+static int num_threads = DEF_NUM_THREADS;
+static int exit_status = 0;
+static bool after1 = false;
+
+static const char * my_name = "sgh_dd: ";
+
+static const char * mrq_blk_s = "mrq: ordinary blocking";
+static const char * mrq_vb_s = "mrq: variable blocking";
+static const char * mrq_svb_s = "mrq: shared variable blocking (svb)";
+static const char * mrq_s_nb_s = "mrq: submit of full non-blocking";
+
+
+#ifdef __GNUC__
+static int pr2serr_lk(const char * fmt, ...)
+ __attribute__ ((format (printf, 1, 2)));
+#if 0
+static void pr_errno_lk(int e_no, const char * fmt, ...)
+ __attribute__ ((format (printf, 2, 3)));
+#endif
+#else
+static int pr2serr_lk(const char * fmt, ...);
+#if 0
+static void pr_errno_lk(int e_no, const char * fmt, ...);
+#endif
+#endif
+
+
+static int
+pr2serr_lk(const char * fmt, ...)
+{
+ int n;
+ va_list args;
+
+ pthread_mutex_lock(&strerr_mut);
+ va_start(args, fmt);
+ n = vfprintf(stderr, fmt, args);
+ va_end(args);
+ pthread_mutex_unlock(&strerr_mut);
+ return n;
+}
+
+static void
+usage(int pg_num)
+{
+ if (pg_num > 3)
+ goto page4;
+ else if (pg_num > 2)
+ goto page3;
+ else if (pg_num > 1)
+ goto page2;
+
+ pr2serr("Usage: sgh_dd [bs=BS] [conv=CONVS] [count=COUNT] [ibs=BS] "
+ "[if=IFILE]\n"
+ " [iflag=FLAGS] [obs=BS] [of=OFILE] [oflag=FLAGS] "
+ "[seek=SEEK]\n"
+ " [skip=SKIP] [--help] [--version]\n\n");
+ pr2serr(" [ae=AEN[,MAEN]] [bpt=BPT] [cdbsz=6|10|12|16] "
+ "[coe=0|1]\n"
+ " [dio=0|1] [elemsz_kb=EKB] [fail_mask=FM] "
+ "[fua=0|1|2|3]\n"
+ " [mrq=[I|O,]NRQS[,C]] [noshare=0|1] "
+ "[of2=OFILE2]\n"
+ " [ofreg=OFREG] [ofsplit=OSP] [sdt=SDT] "
+ "[sync=0|1]\n"
+ " [thr=THR] [time=0|1|2[,TO]] [unshare=1|0] "
+ "[verbose=VERB]\n"
+ " [--dry-run] [--prefetch] [-v|-vv|-vvv] "
+ "[--verbose]\n"
+ " [--verify] [--version]\n\n"
+ " where the main options (shown in first group above) are:\n"
+ " bs must be device logical block size (default "
+ "512)\n"
+ " conv comma separated list from: [nocreat,noerror,"
+ "notrunc,\n"
+ " null,sync]\n"
+ " count number of blocks to copy (def: device size)\n"
+ " if file or device to read from (def: stdin)\n"
+ " iflag comma separated list from: [00,coe,defres,dio,"
+ "direct,dpo,\n"
+ " dsync,excl,ff,fua,masync,mmap,mout_if,"
+ "mrq_immed,mrq_svb,\n"
+ " nocreat,nodur,noxfer,null,polled,qhead,"
+ "qtail,\n"
+ " random,same_fds,v3,v4,wq_excl]\n"
+ " of file or device to write to (def: /dev/null "
+ "N.B. different\n"
+ " from dd it defaults to stdout). If 'of=.' "
+ "uses /dev/null\n"
+ " of2 second file or device to write to (def: "
+ "/dev/null)\n"
+ " oflag comma separated list from: [append,<<list from "
+ "iflag>>]\n"
+ " seek block position to start writing to OFILE\n"
+ " skip block position to start reading from IFILE\n"
+ " --help|-h output this usage message then exit\n"
+ " --verify|-x do a verify (compare) operation [def: do a "
+ "copy]\n"
+ " --version|-V output version string then exit\n\n"
+ "Copy IFILE to OFILE, similar to dd command. This utility is "
+ "specialized for\nSCSI devices and uses multiple POSIX threads. "
+ "It expects one or both IFILE\nand OFILE to be sg devices. With "
+ "--verify option does a verify/compare\noperation instead of a "
+ "copy. This utility is Linux specific and uses the\nv4 sg "
+ "driver 'share' capability if available. Use '-hh', '-hhh' or "
+ "'-hhhh'\nfor more information.\n"
+ );
+ return;
+page2:
+ pr2serr("Syntax: sgh_dd [operands] [options]\n\n"
+ " where: operands have the form name=value and are pecular to "
+ "'dd'\n"
+ " style commands, and options start with one or "
+ "two hyphens;\n"
+ " the lesser used operands and option are:\n\n"
+ " ae AEN: abort every n commands (def: 0 --> don't "
+ "abort any)\n"
+ " MAEN: abort every n mrq commands (def: 0 --> "
+ "don't)\n"
+ " [requires commands with > 1 ms duration]\n"
+ " bpt is blocks_per_transfer (default is 128)\n"
+ " cdbsz size of SCSI READ, WRITE or VERIFY cdb_s "
+ "(default is 10)\n"
+ " coe continue on error, 0->exit (def), "
+ "1->zero + continue\n"
+ " dio is direct IO, 1->attempt, 0->indirect IO (def)\n"
+ " elemsz_kb scatter gather list element size in kilobytes "
+ "(def: 32[KB])\n"
+ " fail_mask 1: misuse KEEP_SHARE flag; 0: nothing (def)\n"
+ " fua force unit access: 0->don't(def), 1->OFILE, "
+ "2->IFILE,\n"
+ " 3->OFILE+IFILE\n"
+ " mrq number of cmds placed in each sg call "
+ "(def: 0);\n"
+ " may have trailing ',C', to send bulk cdb_s; "
+ "if preceded\n"
+ " by 'I' then mrq only on IFILE, likewise 'O' "
+ "for OFILE\n"
+ " noshare 0->use request sharing(def), 1->don't\n"
+ " ofreg OFREG is regular file or pipe to send what is "
+ "read from\n"
+ " IFILE in the first half of each shared element\n"
+ " ofsplit split ofile write in two at block OSP (def: 0 "
+ "(no split))\n"
+ " sdt stall detection times: CRT[,ICT]. CRT: check "
+ "repetition\n"
+ " time (after first) in seconds; ICT: initial "
+ "check time\n"
+ " in milliseconds. Default: 3,300 . Use CRT=0 "
+ "to disable\n"
+ " sync 0->no sync(def), 1->SYNCHRONIZE CACHE on OFILE "
+ "after copy\n"
+ " thr is number of threads, must be > 0, default 4, "
+ "max 1024\n"
+ " time 0->no timing, 1->calc throughput(def), "
+ "2->nanosec\n"
+ " precision; TO is command timeout in seconds "
+ "(def: 60)\n"
+ " unshare 0->don't explicitly unshare after share; 1->let "
+ "close do\n"
+ " file unshare (default)\n"
+ " verbose increase verbosity\n"
+ " --chkaddr|-c exits if read block does not contain "
+ "32 bit block\n"
+ " address, used once only checks first "
+ "address in block\n"
+ " --dry-run|-d prepare but bypass copy/read\n"
+ " --prefetch|-p with verify: do pre-fetch first\n"
+ " --verbose|-v increase verbosity of utility\n\n"
+ "Use '-hhh' or '-hhhh' for more information about flags.\n"
+ );
+ return;
+page3:
+ pr2serr("Syntax: sgh_dd [operands] [options]\n\n"
+ " where: 'iflag=<arg>' and 'oflag=<arg>' arguments are listed "
+ "below:\n\n"
+ " 00 use all zeros instead of if=IFILE (only in "
+ "iflags)\n"
+ " 00,ff generates blocks that contain own (32 bit be) "
+ "blk address\n"
+ " append append output to OFILE (assumes OFILE is "
+ "regular file)\n"
+ " coe continue of error (reading, fills with zeros)\n"
+ " defres keep default reserve buffer size (else its "
+ "bs*bpt)\n"
+ " dio sets the SG_FLAG_DIRECT_IO in sg requests\n"
+ " direct sets the O_DIRECT flag on open()\n"
+ " dpo sets the DPO (disable page out) in SCSI READs "
+ "and WRITEs\n"
+ " dsync sets the O_SYNC flag on open()\n"
+ " excl sets the O_EXCL flag on open()\n"
+ " ff use all 0xff bytes instead of if=IFILE (only in "
+ "iflags)\n"
+ " fua sets the FUA (force unit access) in SCSI READs "
+ "and WRITEs\n"
+ " hipri same as 'polled'; 'hipri' name is deprecated\n"
+ " masync set 'more async' flag on this sg device\n"
+ " mmap setup mmap IO on IFILE or OFILE; OFILE only "
+ "with noshare\n"
+ " mmap,mmap when used twice, doesn't call munmap()\n"
+ " mout_if set META_OUT_IF flag on each request\n"
+ " mrq_immed if mrq active, do submit non-blocking (def: "
+ "ordered\n"
+ " blocking)\n"
+ " mrq_svb if mrq and sg->sg copy, do shared_variable_"
+ "blocking\n"
+ " nocreat will fail rather than create OFILE\n"
+ " nodur turns off command duration calculations\n"
+ " noxfer no transfer to/from the user space\n"
+ " no_thresh skip checking per fd max data xfer\n"
+ " null does nothing, placeholder\n"
+ " polled set POLLED flag on command, uses blk_poll() to "
+ "complete\n"
+ " qhead queue new request at head of block queue\n"
+ " qtail queue new request at tail of block queue (def: "
+ "q at head)\n"
+ " random use random data instead of if=IFILE (only in "
+ "iflags)\n"
+ " same_fds each thread uses the same IFILE and OFILE(2) "
+ "file\n"
+ " descriptors (def: each threads has own file "
+ "descriptors)\n"
+ " swait this option is now ignored\n"
+ " v3 use v3 sg interface (def: v3 unless sg driver "
+ "is v4)\n"
+ " v4 use v4 sg interface (def: v3 unless sg driver "
+ "is v4)\n"
+ " wq_excl set SG_CTL_FLAGM_EXCL_WAITQ on this sg fd\n"
+ "\n"
+ "Copies IFILE to OFILE (and to OFILE2 if given). If IFILE and "
+ "OFILE are sg\ndevices 'shared' mode is selected unless "
+ "'noshare' is given to 'iflag=' or\n'oflag='. of2=OFILE2 uses "
+ "'oflag=FLAGS'. When sharing, the data stays in a\nsingle "
+ "in-kernel buffer which is copied (or mmap-ed) to the user "
+ "space\nif the 'ofreg=OFREG' is given. Use '-hhhh' for more "
+ "information.\n"
+ );
+ return;
+page4:
+ pr2serr("pack_id:\n"
+ "These are ascending integers, starting at 1, associated with "
+ "each issued\nSCSI command. When both IFILE and OFILE are sg "
+ "devices, then the READ in\neach read-write pair is issued an "
+ "even pack_id and its WRITE pair is\ngiven the pack_id one "
+ "higher (i.e. an odd number). This enables a\n'dmesg -w' "
+ "user to see that progress is being "
+ "made.\n\n");
+ pr2serr("Debugging:\n"
+ "Apart from using one or more '--verbose' options which gets a "
+ "bit noisy\n'dmesg -w' can give a good overview "
+ "of what is happening.\nThat does a sg driver object tree "
+ "traversal that does minimal locking\nto make sure that each "
+ "traversal is 'safe'. So it is important to note\nthe whole "
+ "tree is not locked. This means for fast devices the overall\n"
+ "tree state may change while the traversal is occurring. For "
+ "example,\nit has been observed that both the read- and write- "
+ "sides of a request\nshare show they are in 'active' state "
+ "which should not be possible.\nIt occurs because the read-"
+ "side probably jumped out of active state and\nthe write-side "
+ "request entered it while some other nodes were being "
+ "printed.\n\n");
+ pr2serr("Busy state:\n"
+ "Busy state (abbreviated to 'bsy' in the dmesg "
+ "output)\nis entered during request setup and completion. It "
+ "is intended to be\na temporary state. It should not block "
+ "but does sometimes (e.g. in\nblock_get_request()). Even so "
+ "that blockage should be short and if not\nthere is a "
+ "problem.\n\n");
+ pr2serr("--verify :\n"
+ "For comparing IFILE with OFILE. Does repeated sequences of: "
+ "READ(ifile)\nand uses data returned to send to VERIFY(ofile, "
+ "BYTCHK=1). So the OFILE\ndevice/disk is doing the actual "
+ "comparison. Stops on first miscompare.\n\n");
+ pr2serr("--prefetch :\n"
+ "Used with --verify option. Prepends a PRE-FETCH(ofile, IMMED) "
+ "to verify\nsequence. This should speed the trailing VERIFY by "
+ "making sure that\nthe data it needs for the comparison is "
+ "already in its cache.\n");
+ return;
+}
+
+static void
+lk_print_command_len(const char *prefix, uint8_t * cmdp, int len, bool lock)
+{
+ if (lock)
+ pthread_mutex_lock(&strerr_mut);
+ if (prefix && *prefix)
+ fputs(prefix, stderr);
+ sg_print_command_len(cmdp, len);
+ if (lock)
+ pthread_mutex_unlock(&strerr_mut);
+}
+
+static void
+lk_chk_n_print3(const char * leadin, struct sg_io_hdr * hp, bool raw_sinfo)
+{
+ pthread_mutex_lock(&strerr_mut);
+ sg_chk_n_print3(leadin, hp, raw_sinfo);
+ pthread_mutex_unlock(&strerr_mut);
+}
+
+static void
+lk_chk_n_print4(const char * leadin, const struct sg_io_v4 * h4p,
+ bool raw_sinfo)
+{
+ pthread_mutex_lock(&strerr_mut);
+ sg_linux_sense_print(leadin, h4p->device_status, h4p->transport_status,
+ h4p->driver_status, (const uint8_t *)h4p->response,
+ h4p->response_len, raw_sinfo);
+ pthread_mutex_unlock(&strerr_mut);
+}
+
+static void
+hex2stderr_lk(const uint8_t * b_str, int len, int no_ascii)
+{
+ pthread_mutex_lock(&strerr_mut);
+ hex2stderr(b_str, len, no_ascii);
+ pthread_mutex_unlock(&strerr_mut);
+}
+
+/* Flags decoded into abbreviations for those that are set, separated by
+ * '|' . */
+static char *
+sg_flags_str(int flags, int b_len, char * b)
+{
+ int n = 0;
+
+ if ((b_len < 1) || (! b))
+ return b;
+ b[0] = '\0';
+ if (SG_FLAG_DIRECT_IO & flags) { /* 0x1 */
+ n += sg_scnpr(b + n, b_len - n, "DIO|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SG_FLAG_MMAP_IO & flags) { /* 0x4 */
+ n += sg_scnpr(b + n, b_len - n, "MMAP|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_YIELD_TAG & flags) { /* 0x8 */
+ n += sg_scnpr(b + n, b_len - n, "YTAG|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SG_FLAG_Q_AT_TAIL & flags) { /* 0x10 */
+ n += sg_scnpr(b + n, b_len - n, "QTAI|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SG_FLAG_Q_AT_HEAD & flags) { /* 0x20 */
+ n += sg_scnpr(b + n, b_len - n, "QHEA|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_DOUT_OFFSET & flags) { /* 0x40 */
+ n += sg_scnpr(b + n, b_len - n, "DOFF|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_EVENTFD & flags) { /* 0x80 */
+ n += sg_scnpr(b + n, b_len - n, "EVFD|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_COMPLETE_B4 & flags) { /* 0x100 */
+ n += sg_scnpr(b + n, b_len - n, "CPL_B4|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_SIGNAL & flags) { /* 0x200 */
+ n += sg_scnpr(b + n, b_len - n, "SIGNAL|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_IMMED & flags) { /* 0x400 */
+ n += sg_scnpr(b + n, b_len - n, "IMM|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_POLLED & flags) { /* 0x800 */
+ n += sg_scnpr(b + n, b_len - n, "POLLED|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_STOP_IF & flags) { /* 0x1000 */
+ n += sg_scnpr(b + n, b_len - n, "STOPIF|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_DEV_SCOPE & flags) { /* 0x2000 */
+ n += sg_scnpr(b + n, b_len - n, "DEV_SC|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_SHARE & flags) { /* 0x4000 */
+ n += sg_scnpr(b + n, b_len - n, "SHARE|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_DO_ON_OTHER & flags) { /* 0x8000 */
+ n += sg_scnpr(b + n, b_len - n, "DO_OTH|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_NO_DXFER & flags) { /* 0x10000 */
+ n += sg_scnpr(b + n, b_len - n, "NOXFER|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_KEEP_SHARE & flags) { /* 0x20000 */
+ n += sg_scnpr(b + n, b_len - n, "KEEP_SH|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_MULTIPLE_REQS & flags) { /* 0x40000 */
+ n += sg_scnpr(b + n, b_len - n, "MRQS|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_ORDERED_WR & flags) { /* 0x80000 */
+ n += sg_scnpr(b + n, b_len - n, "OWR|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_REC_ORDER & flags) { /* 0x100000 */
+ n += sg_scnpr(b + n, b_len - n, "REC_O|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SGV4_FLAG_META_OUT_IF & flags) { /* 0x200000 */
+ n += sg_scnpr(b + n, b_len - n, "MOUT_IF|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (0 == n)
+ n += sg_scnpr(b + n, b_len - n, "<none>");
+fini:
+ if (n < b_len) { /* trim trailing '\' */
+ if ('|' == b[n - 1])
+ b[n - 1] = '\0';
+ } else if ('|' == b[b_len - 1])
+ b[b_len - 1] = '\0';
+ return b;
+}
+
+/* Info field decoded into abbreviations for those bits that are set,
+ * separated by '|' . */
+static char *
+sg_info_str(int info, int b_len, char * b)
+{
+ int n = 0;
+
+ if ((b_len < 1) || (! b))
+ return b;
+ b[0] = '\0';
+ if (SG_INFO_CHECK & info) { /* 0x1 */
+ n += sg_scnpr(b + n, b_len - n, "CHK|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SG_INFO_DIRECT_IO & info) { /* 0x2 */
+ n += sg_scnpr(b + n, b_len - n, "DIO|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SG_INFO_MIXED_IO & info) { /* 0x4 */
+ n += sg_scnpr(b + n, b_len - n, "MIO|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SG_INFO_DEVICE_DETACHING & info) { /* 0x8 */
+ n += sg_scnpr(b + n, b_len - n, "DETA|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SG_INFO_ABORTED & info) { /* 0x10 */
+ n += sg_scnpr(b + n, b_len - n, "ABRT|");
+ if (n >= b_len)
+ goto fini;
+ }
+ if (SG_INFO_MRQ_FINI & info) { /* 0x20 */
+ n += sg_scnpr(b + n, b_len - n, "MRQF|");
+ if (n >= b_len)
+ goto fini;
+ }
+fini:
+ if (n < b_len) { /* trim trailing '\' */
+ if ('|' == b[n - 1])
+ b[n - 1] = '\0';
+ } else if ('|' == b[b_len - 1])
+ b[b_len - 1] = '\0';
+ return b;
+}
+
+static void
+v4hdr_out_lk(const char * leadin, const sg_io_v4 * h4p, int id)
+{
+ char b[80];
+
+ pthread_mutex_lock(&strerr_mut);
+ if (leadin)
+ pr2serr("%s [id=%d]:\n", leadin, id);
+ if (('Q' != h4p->guard) || (0 != h4p->protocol) ||
+ (0 != h4p->subprotocol))
+ pr2serr(" <<<sg_io_v4 _NOT_ properly set>>>\n");
+ pr2serr(" pointers: cdb=%s sense=%s din=%p dout=%p\n",
+ (h4p->request ? "y" : "NULL"), (h4p->response ? "y" : "NULL"),
+ (void *)h4p->din_xferp, (void *)h4p->dout_xferp);
+ pr2serr(" lengths: cdb=%u sense=%u din=%u dout=%u\n",
+ h4p->request_len, h4p->max_response_len, h4p->din_xfer_len,
+ h4p->dout_xfer_len);
+ pr2serr(" flags=0x%x request_extra{pack_id}=%d\n",
+ h4p->flags, h4p->request_extra);
+ pr2serr(" flags set: %s\n", sg_flags_str(h4p->flags, sizeof(b), b));
+ pr2serr(" %s OUT fields:\n", leadin);
+ pr2serr(" response_len=%d driver/transport/device_status="
+ "0x%x/0x%x/0x%x\n", h4p->response_len, h4p->driver_status,
+ h4p->transport_status, h4p->device_status);
+ pr2serr(" info=0x%x din_resid=%u dout_resid=%u spare_out=%u "
+ "dur=%u\n",
+ h4p->info, h4p->din_resid, h4p->dout_resid, h4p->spare_out,
+ h4p->duration);
+ pthread_mutex_unlock(&strerr_mut);
+}
+
+static void
+fetch_sg_version(void)
+{
+ FILE * fp;
+ char b[96];
+
+ have_sg_version = false;
+ sg_version = 0;
+ fp = fopen(PROC_SCSI_SG_VERSION, "r");
+ if (fp && fgets(b, sizeof(b) - 1, fp)) {
+ if (1 == sscanf(b, "%d", &sg_version))
+ have_sg_version = !!sg_version;
+ } else {
+ int j, k, l;
+
+ if (fp)
+ fclose(fp);
+ fp = fopen(SYS_SCSI_SG_VERSION, "r");
+ if (fp && fgets(b, sizeof(b) - 1, fp)) {
+ if (3 == sscanf(b, "%d.%d.%d", &j, &k, &l)) {
+ sg_version = (j * 10000) + (k * 100) + l;
+ have_sg_version = !!sg_version;
+ }
+ }
+ }
+ if (fp)
+ fclose(fp);
+}
+
+static void
+calc_duration_throughput(int contin)
+{
+ struct timeval end_tm, res_tm;
+ double a, b;
+
+ gettimeofday(&end_tm, NULL);
+ res_tm.tv_sec = end_tm.tv_sec - start_tm.tv_sec;
+ res_tm.tv_usec = end_tm.tv_usec - start_tm.tv_usec;
+ if (res_tm.tv_usec < 0) {
+ --res_tm.tv_sec;
+ res_tm.tv_usec += 1000000;
+ }
+ a = res_tm.tv_sec;
+ a += (0.000001 * res_tm.tv_usec);
+ b = (double)gcoll.bs * (dd_count - gcoll.out_rem_count.load());
+ pr2serr("time to %s data %s %d.%06d secs",
+ (gcoll.verify ? "verify" : "copy"), (contin ? "so far" : "was"),
+ (int)res_tm.tv_sec, (int)res_tm.tv_usec);
+ if ((a > 0.00001) && (b > 511))
+ pr2serr(", %.2f MB/sec\n", b / (a * 1000000.0));
+ else
+ pr2serr("\n");
+}
+
+static void
+print_stats(const char * str)
+{
+ int64_t infull;
+
+ if (0 != gcoll.out_rem_count.load())
+ pr2serr(" remaining block count=%" PRId64 "\n",
+ gcoll.out_rem_count.load());
+ infull = dd_count - gcoll.in_rem_count.load();
+ pr2serr("%s%" PRId64 "+%d records in\n", str,
+ infull - gcoll.in_partial.load(), gcoll.in_partial.load());
+
+ if (gcoll.out_type == FT_DEV_NULL)
+ pr2serr("%s0+0 records out\n", str);
+ else {
+ int64_t outfull = dd_count - gcoll.out_rem_count.load();
+
+ pr2serr("%s%" PRId64 "+%d records %s\n", str,
+ outfull - gcoll.out_partial.load(), gcoll.out_partial.load(),
+ (gcoll.verify ? "verified" : "out"));
+ }
+}
+
+static void
+interrupt_handler(int sig)
+{
+ struct sigaction sigact;
+
+ sigact.sa_handler = SIG_DFL;
+ sigemptyset(&sigact.sa_mask);
+ sigact.sa_flags = 0;
+ sigaction(sig, &sigact, NULL);
+ pr2serr("Interrupted by signal,");
+ if (do_time > 0)
+ calc_duration_throughput(0);
+ print_stats("");
+ kill(getpid (), sig);
+}
+
+static void
+siginfo_handler(int sig)
+{
+ if (sig) { ; } /* unused, dummy to suppress warning */
+ pr2serr("Progress report, continuing ...\n");
+ if (do_time > 0)
+ calc_duration_throughput(1);
+ print_stats(" ");
+}
+
+static void
+siginfo2_handler(int sig)
+{
+ struct global_collection * clp = &gcoll;
+
+ if (sig) { ; } /* unused, dummy to suppress warning */
+ pr2serr("Progress report, continuing ...\n");
+ if (do_time > 0)
+ calc_duration_throughput(1);
+ print_stats(" ");
+ pr2serr("Send broadcast on out_sync_cv condition variable\n");
+ pthread_cond_broadcast(&clp->out_sync_cv);
+}
+
+static void
+install_handler(int sig_num, void (*sig_handler) (int sig))
+{
+ struct sigaction sigact;
+ sigaction (sig_num, NULL, &sigact);
+ if (sigact.sa_handler != SIG_IGN)
+ {
+ sigact.sa_handler = sig_handler;
+ sigemptyset (&sigact.sa_mask);
+ sigact.sa_flags = 0;
+ sigaction (sig_num, &sigact, NULL);
+ }
+}
+
+#ifdef SG_LIB_ANDROID
+static void
+thread_exit_handler(int sig)
+{
+ pthread_exit(0);
+}
+#endif
+
+/* Make safe_strerror() thread safe */
+static char *
+tsafe_strerror(int code, char * ebp)
+{
+ char * cp;
+
+ pthread_mutex_lock(&strerr_mut);
+ cp = safe_strerror(code);
+ strncpy(ebp, cp, STRERR_BUFF_LEN);
+ pthread_mutex_unlock(&strerr_mut);
+
+ ebp[strlen(ebp)] = '\0';
+ return ebp;
+}
+
+
+/* Following macro from D.R. Butenhof's POSIX threads book:
+ * ISBN 0-201-63392-2 . [Highly recommended book.] Changed __FILE__
+ * to __func__ */
+#define err_exit(code,text) do { \
+ char strerr_buff[STRERR_BUFF_LEN + 1]; \
+ pr2serr("%s at \"%s\":%d: %s\n", \
+ text, __func__, __LINE__, tsafe_strerror(code, strerr_buff)); \
+ exit(1); \
+ } while (0)
+
+
+static int
+dd_filetype(const char * filename, off_t & st_size)
+{
+ struct stat st;
+ size_t len = strlen(filename);
+
+ if ((1 == len) && ('.' == filename[0]))
+ return FT_DEV_NULL;
+ if (stat(filename, &st) < 0)
+ return FT_ERROR;
+ if (S_ISCHR(st.st_mode)) {
+ if ((MEM_MAJOR == major(st.st_rdev)) &&
+ ((DEV_NULL_MINOR_NUM == minor(st.st_rdev)) ||
+ (DEV_ZERO_MINOR_NUM == minor(st.st_rdev))))
+ return FT_DEV_NULL; /* treat /dev/null + /dev/zero the same */
+ if (SCSI_GENERIC_MAJOR == major(st.st_rdev))
+ return FT_SG;
+ if (SCSI_TAPE_MAJOR == major(st.st_rdev))
+ return FT_ST;
+ return FT_CHAR;
+ } else if (S_ISBLK(st.st_mode))
+ return FT_BLOCK;
+ else if (S_ISFIFO(st.st_mode))
+ return FT_FIFO;
+ st_size = st.st_size;
+ return FT_OTHER;
+}
+
+static inline void
+stop_both(struct global_collection * clp)
+{
+ clp->in_stop = true;
+ clp->out_stop = true;
+}
+
+/* Return of 0 -> success, see sg_ll_read_capacity*() otherwise */
+static int
+scsi_read_capacity(int sg_fd, int64_t * num_sect, int * sect_sz)
+{
+ int res;
+ uint8_t rcBuff[RCAP16_REPLY_LEN] = {};
+
+ res = sg_ll_readcap_10(sg_fd, 0, 0, rcBuff, READ_CAP_REPLY_LEN, false, 0);
+ if (0 != res)
+ return res;
+
+ if ((0xff == rcBuff[0]) && (0xff == rcBuff[1]) && (0xff == rcBuff[2]) &&
+ (0xff == rcBuff[3])) {
+
+ res = sg_ll_readcap_16(sg_fd, 0, 0, rcBuff, RCAP16_REPLY_LEN, false,
+ 0);
+ if (0 != res)
+ return res;
+ *num_sect = sg_get_unaligned_be64(rcBuff + 0) + 1;
+ *sect_sz = sg_get_unaligned_be32(rcBuff + 8);
+ } else {
+ /* take care not to sign extend values > 0x7fffffff */
+ *num_sect = (int64_t)sg_get_unaligned_be32(rcBuff + 0) + 1;
+ *sect_sz = sg_get_unaligned_be32(rcBuff + 4);
+ }
+ return 0;
+}
+
+/* Return of 0 -> success, -1 -> failure. BLKGETSIZE64, BLKGETSIZE and */
+/* BLKSSZGET macros problematic (from <linux/fs.h> or <sys/mount.h>). */
+static int
+read_blkdev_capacity(int sg_fd, int64_t * num_sect, int * sect_sz)
+{
+#ifdef BLKSSZGET
+ if ((ioctl(sg_fd, BLKSSZGET, sect_sz) < 0) && (*sect_sz > 0)) {
+ perror("BLKSSZGET ioctl error");
+ return -1;
+ } else {
+ #ifdef BLKGETSIZE64
+ uint64_t ull;
+
+ if (ioctl(sg_fd, BLKGETSIZE64, &ull) < 0) {
+
+ perror("BLKGETSIZE64 ioctl error");
+ return -1;
+ }
+ *num_sect = ((int64_t)ull / (int64_t)*sect_sz);
+ #else
+ unsigned long ul;
+
+ if (ioctl(sg_fd, BLKGETSIZE, &ul) < 0) {
+ perror("BLKGETSIZE ioctl error");
+ return -1;
+ }
+ *num_sect = (int64_t)ul;
+ #endif
+ }
+ return 0;
+#else
+ *num_sect = 0;
+ *sect_sz = 0;
+ return -1;
+#endif
+}
+
+static int
+system_wrapper(const char * cmd)
+{
+ int res;
+
+ res = system(cmd);
+ if (WIFSIGNALED(res) &&
+ (WTERMSIG(res) == SIGINT || WTERMSIG(res) == SIGQUIT))
+ raise(WTERMSIG(res));
+ return WEXITSTATUS(res);
+}
+
+/* Has an infinite loop doing a timed wait for any signals in sig_set. After
+ * each timeout (300 ms) checks if the most_recent_pack_id atomic integer
+ * has changed. If not after another two timeouts announces a stall has
+ * been detected. If shutting down atomic is true breaks out of loop and
+ * shuts down this thread. Other than that, this thread is normally cancelled
+ * by the main thread, after other threads have exited. */
+static void *
+sig_listen_thread(void * v_clp)
+{
+ bool stall_reported = false;
+ int prev_pack_id = 0;
+ struct timespec ts;
+ struct timespec * tsp = &ts;
+ struct global_collection * clp = (struct global_collection *)v_clp;
+ uint32_t ict_ms = (clp->sdt_ict ? clp->sdt_ict : DEF_SDT_ICT_MS);
+
+ tsp->tv_sec = ict_ms / 1000;
+ tsp->tv_nsec = (ict_ms % 1000) * 1000 * 1000; /* DEF_SDT_ICT_MS */
+ while (1) {
+ int sig_number = sigtimedwait(&signal_set, NULL, tsp);
+
+ if (sig_number < 0) {
+ int err = errno;
+
+ /* EAGAIN implies a timeout */
+ if ((EAGAIN == err) && (clp->sdt_crt > 0)) {
+ int pack_id = mono_pack_id.load();
+
+ if ((pack_id > 0) && (pack_id == prev_pack_id)) {
+ if (! stall_reported) {
+ stall_reported = true;
+ tsp->tv_sec = clp->sdt_crt;
+ tsp->tv_nsec = 0;
+ pr2serr_lk("%s: first stall at pack_id=%d detected\n",
+ __func__, pack_id);
+ } else
+ pr2serr_lk("%s: subsequent stall at pack_id=%d\n",
+ __func__, pack_id);
+ // following command assumes linux bash or similar shell
+ system_wrapper("cat /proc/scsi/sg/debug >> /dev/stderr\n");
+ // system_wrapper("/usr/bin/dmesg\n");
+ } else
+ prev_pack_id = pack_id;
+ } else if (EAGAIN != err)
+ pr2serr_lk("%s: sigtimedwait() errno=%d\n", __func__, err);
+ }
+ if (SIGINT == sig_number) {
+ pr2serr_lk("%sinterrupted by SIGINT\n", my_name);
+ stop_both(clp);
+ pthread_cond_broadcast(&clp->out_sync_cv);
+ sigprocmask(SIG_SETMASK, &orig_signal_set, NULL);
+ raise(SIGINT);
+ break;
+ }
+ if (SIGUSR2 == sig_number) {
+ if (clp->verbose > 2)
+ pr2serr_lk("%s: interrupted by SIGUSR2\n", __func__);
+ break;
+ }
+ if (shutting_down)
+ break;
+ } /* end of while loop */
+ if (clp->verbose > 3)
+ pr2serr_lk("%s: exiting\n", __func__);
+
+ return NULL;
+}
+
+static void *
+mrq_abort_thread(void * v_maip)
+{
+ int res, err;
+ int n = 0;
+ int seed;
+ unsigned int rn;
+ Mrq_abort_info l_mai = *(Mrq_abort_info *)v_maip;
+ struct sg_io_v4 ctl_v4 {};
+
+#ifdef HAVE_GETRANDOM
+ {
+ ssize_t ssz = getrandom(&seed, sizeof(seed), GRND_NONBLOCK);
+
+ if (ssz < (ssize_t)sizeof(seed)) {
+ pr2serr("getrandom() failed, ret=%d\n", (int)ssz);
+ seed = (int)time(NULL);
+ }
+ }
+#else
+ seed = (int)time(NULL); /* use seconds since epoch as proxy */
+#endif
+ if (l_mai.debug)
+ pr2serr_lk("%s: from_id=%d: to abort mrq_pack_id=%d\n", __func__,
+ l_mai.from_tid, l_mai.mrq_id);
+ res = ioctl(l_mai.fd, SG_GET_NUM_WAITING, &n);
+ ++num_waiting_calls;
+ if (res < 0) {
+ err = errno;
+ pr2serr_lk("%s: ioctl(SG_GET_NUM_WAITING) failed: %s [%d]\n",
+ __func__, safe_strerror(err), err);
+ } else if (l_mai.debug)
+ pr2serr_lk("%s: num_waiting=%d\n", __func__, n);
+
+ Rand_uint * ruip = new Rand_uint(5, 500, seed);
+ struct timespec tspec = {0, 4000 /* 4 usecs */};
+ rn = ruip->get();
+ tspec.tv_nsec = rn * 1000;
+ if (l_mai.debug > 1)
+ pr2serr_lk("%s: /dev/urandom seed=0x%x delay=%u microsecs\n",
+ __func__, seed, rn);
+ if (rn >= 20)
+ nanosleep(&tspec, NULL);
+ else if (l_mai.debug > 1)
+ pr2serr_lk("%s: skipping nanosleep cause delay < 20 usecs\n",
+ __func__);
+
+ ctl_v4.guard = 'Q';
+ ctl_v4.flags = SGV4_FLAG_MULTIPLE_REQS;
+ ctl_v4.request_extra = l_mai.mrq_id;
+ ++num_mrq_abort_req;
+ res = ioctl(l_mai.fd, SG_IOABORT, &ctl_v4);
+ if (res < 0) {
+ err = errno;
+ if (ENODATA == err)
+ pr2serr_lk("%s: ioctl(SG_IOABORT) no match on "
+ "MRQ pack_id=%d\n", __func__, l_mai.mrq_id);
+ else
+ pr2serr_lk("%s: MRQ ioctl(SG_IOABORT) failed: %s [%d]\n",
+ __func__, safe_strerror(err), err);
+ } else {
+ ++num_mrq_abort_req_success;
+ if (l_mai.debug > 1)
+ pr2serr_lk("%s: from_id=%d sent ioctl(SG_IOABORT) on MRQ rq_id="
+ "%d, success\n", __func__, l_mai.from_tid,
+ l_mai.mrq_id);
+ }
+ delete ruip;
+ return NULL;
+}
+
+static bool
+sg_share_prepare(int write_side_fd, int read_side_fd, int id, bool vb_b)
+{
+ struct sg_extended_info sei {};
+ struct sg_extended_info * seip = &sei;
+
+ seip->sei_wr_mask |= SG_SEIM_SHARE_FD;
+ seip->sei_rd_mask |= SG_SEIM_SHARE_FD;
+ seip->share_fd = read_side_fd;
+ if (ioctl(write_side_fd, SG_SET_GET_EXTENDED, seip) < 0) {
+ pr2serr_lk("tid=%d: ioctl(EXTENDED(shared_fd=%d), failed "
+ "errno=%d %s\n", id, read_side_fd, errno,
+ strerror(errno));
+ return false;
+ }
+ if (vb_b)
+ pr2serr_lk("%s: tid=%d: ioctl(EXTENDED(shared_fd)) ok, "
+ "read_side_fd=%d, write_side_fd=%d\n", __func__,
+ id, read_side_fd, write_side_fd);
+ return true;
+}
+
+static void
+sg_unshare(int sg_fd, int id, bool vb_b)
+{
+ struct sg_extended_info sei {};
+ struct sg_extended_info * seip = &sei;
+
+ seip->sei_wr_mask |= SG_SEIM_CTL_FLAGS;
+ seip->sei_rd_mask |= SG_SEIM_CTL_FLAGS;
+ seip->ctl_flags_wr_mask |= SG_CTL_FLAGM_UNSHARE;
+ seip->ctl_flags |= SG_CTL_FLAGM_UNSHARE; /* needs to be set to unshare */
+ if (ioctl(sg_fd, SG_SET_GET_EXTENDED, seip) < 0) {
+ pr2serr_lk("tid=%d: ioctl(EXTENDED(UNSHARE), failed errno=%d %s\n",
+ id, errno, strerror(errno));
+ return;
+ }
+ if (vb_b)
+ pr2serr_lk("tid=%d: ioctl(UNSHARE) ok\n", id);
+}
+
+static void
+sg_noshare_enlarge(int sg_fd, bool vb_b)
+{
+ if (sg_version_ge_40045) {
+ struct sg_extended_info sei {};
+ struct sg_extended_info * seip = &sei;
+
+ sei.sei_wr_mask |= SG_SEIM_TOT_FD_THRESH;
+ seip->tot_fd_thresh = 96 * 1024 * 1024;
+ if (ioctl(sg_fd, SG_SET_GET_EXTENDED, seip) < 0) {
+ pr2serr_lk("%s: ioctl(EXTENDED(TOT_FD_THRESH), failed errno=%d "
+ "%s\n", __func__, errno, strerror(errno));
+ return;
+ }
+ if (vb_b)
+ pr2serr_lk("ioctl(TOT_FD_THRESH) ok\n");
+ }
+}
+
+static void
+sg_take_snap(int sg_fd, int id, bool vb_b)
+{
+ struct sg_extended_info sei {};
+ struct sg_extended_info * seip = &sei;
+
+ seip->sei_wr_mask |= SG_SEIM_CTL_FLAGS;
+ seip->sei_rd_mask |= SG_SEIM_CTL_FLAGS;
+ seip->ctl_flags_wr_mask |= SG_CTL_FLAGM_SNAP_DEV;
+ seip->ctl_flags &= ~SG_CTL_FLAGM_SNAP_DEV; /* 0 --> append */
+ if (ioctl(sg_fd, SG_SET_GET_EXTENDED, seip) < 0) {
+ pr2serr_lk("tid=%d: ioctl(EXTENDED(SNAP_DEV), failed errno=%d %s\n",
+ id, errno, strerror(errno));
+ return;
+ }
+ if (vb_b)
+ pr2serr_lk("tid=%d: ioctl(SNAP_DEV) ok\n", id);
+}
+
+static void
+cleanup_in(void * v_clp)
+{
+ struct global_collection * clp = (struct global_collection *)v_clp;
+
+ pr2serr("thread cancelled while in mutex held\n");
+ stop_both(clp);
+ pthread_mutex_unlock(&clp->in_mutex);
+ pthread_cond_broadcast(&clp->out_sync_cv);
+}
+
+static void
+cleanup_out(void * v_clp)
+{
+ struct global_collection * clp = (struct global_collection *)v_clp;
+
+ pr2serr("thread cancelled while out_mutex held\n");
+ stop_both(clp);
+ pthread_mutex_unlock(&clp->out_mutex);
+ pthread_cond_broadcast(&clp->out_sync_cv);
+}
+
+static void inline buffp_onto_next(Rq_elem * rep)
+{
+ struct global_collection * clp = rep->clp;
+
+ if ((clp->nmrqs > 0) && clp->unbalanced_mrq) {
+ ++rep->mrq_index;
+ if (rep->mrq_index >= clp->nmrqs)
+ rep->mrq_index = 0; /* wrap */
+ }
+}
+
+static inline uint8_t *
+get_buffp(Rq_elem * rep)
+{
+ struct global_collection * clp = rep->clp;
+
+ if ((clp->nmrqs > 0) && clp->unbalanced_mrq && (rep->mrq_index > 0))
+ return rep->buffp + (rep->mrq_index * clp->bs * clp->bpt);
+ else
+ return rep->buffp;
+}
+
+static void *
+read_write_thread(void * v_tip)
+{
+ Thread_info * tip;
+ struct global_collection * clp;
+ Rq_elem rel {};
+ Rq_elem * rep = &rel;
+ int n, sz, blocks, status, vb, err, res, wr_blks, c_addr;
+ int num_sg = 0;
+ int64_t my_index;
+ volatile bool stop_after_write = false;
+ bool own_infd = false;
+ bool in_is_sg, in_mmap, out_is_sg, out_mmap;
+ bool own_outfd = false;
+ bool own_out2fd = false;
+ bool share_and_ofreg;
+ mrq_arr_t deferred_arr; /* MRQ deferred array (vector) */
+
+ tip = (Thread_info *)v_tip;
+ clp = tip->gcp;
+ vb = clp->verbose;
+ rep->bs = clp->bs;
+ sz = clp->bpt * rep->bs;
+ c_addr = clp->chkaddr;
+ in_is_sg = (FT_SG == clp->in_type);
+ in_mmap = (in_is_sg && (clp->in_flags.mmap > 0));
+ out_is_sg = (FT_SG == clp->out_type);
+ out_mmap = (out_is_sg && (clp->out_flags.mmap > 0));
+ /* Following clp members are constant during lifetime of thread */
+ rep->clp = clp;
+ rep->id = tip->id;
+ if (vb > 2)
+ pr2serr_lk("%d <-- Starting worker thread\n", rep->id);
+ if (! (in_mmap || out_mmap)) {
+ n = sz;
+ if (clp->unbalanced_mrq)
+ n *= clp->nmrqs;
+ rep->buffp = sg_memalign(n, 0 /* page align */, &rep->alloc_bp,
+ false);
+ if (NULL == rep->buffp)
+ err_exit(ENOMEM, "out of memory creating user buffers\n");
+ }
+ rep->infd = clp->infd;
+ rep->outfd = clp->outfd;
+ rep->out2fd = clp->out2fd;
+ rep->outregfd = clp->outregfd;
+ rep->rep_count = 0;
+ if (clp->unbalanced_mrq && (clp->nmrqs > 0))
+ rep->mrq_index = clp->nmrqs - 1;
+
+ if (rep->infd == rep->outfd) {
+ if (in_is_sg)
+ rep->same_sg = true;
+ } else if (in_is_sg && out_is_sg)
+ rep->both_sg = true;
+ else if (in_is_sg)
+ rep->only_in_sg = true;
+ else if (out_is_sg)
+ rep->only_out_sg = true;
+
+ if (clp->in_flags.random) {
+#ifdef HAVE_GETRANDOM
+ ssize_t ssz = getrandom(&rep->seed, sizeof(rep->seed), GRND_NONBLOCK);
+
+ if (ssz < (ssize_t)sizeof(rep->seed)) {
+ pr2serr_lk("thread=%d: getrandom() failed, ret=%d\n",
+ rep->id, (int)ssz);
+ rep->seed = (long)time(NULL);
+ }
+#else
+ rep->seed = (long)time(NULL); /* use seconds since epoch as proxy */
+#endif
+ if (vb > 1)
+ pr2serr_lk("thread=%d: seed=%ld\n", rep->id, rep->seed);
+#ifdef HAVE_SRAND48_R
+ srand48_r(rep->seed, &rep->drand);
+#else
+ srand48(rep->seed);
+#endif
+ }
+ if (clp->in_flags.same_fds || clp->out_flags.same_fds)
+ ;
+ else {
+ int fd;
+
+ if (in_is_sg && clp->infp) {
+ fd = sg_in_open(clp, clp->infp, (in_mmap ? &rep->buffp : NULL),
+ (in_mmap ? &rep->mmap_len : NULL));
+ if (fd < 0)
+ goto fini;
+ rep->infd = fd;
+ rep->mmap_active = in_mmap ? clp->in_flags.mmap : 0;
+ if (in_mmap && (vb > 4))
+ pr2serr_lk("thread=%d: mmap buffp=%p\n", rep->id, rep->buffp);
+ own_infd = true;
+ ++num_sg;
+ if (vb > 2)
+ pr2serr_lk("thread=%d: opened local sg IFILE\n", rep->id);
+ }
+ if (out_is_sg && clp->outfp) {
+ fd = sg_out_open(clp, clp->outfp, (out_mmap ? &rep->buffp : NULL),
+ (out_mmap ? &rep->mmap_len : NULL));
+ if (fd < 0)
+ goto fini;
+ rep->outfd = fd;
+ if (! rep->mmap_active)
+ rep->mmap_active = out_mmap ? clp->out_flags.mmap : 0;
+ if (out_mmap && (vb > 4))
+ pr2serr_lk("thread=%d: mmap buffp=%p\n", rep->id, rep->buffp);
+ own_outfd = true;
+ ++num_sg;
+ if (vb > 2)
+ pr2serr_lk("thread=%d: opened local sg OFILE\n", rep->id);
+ }
+ if ((FT_SG == clp->out2_type) && clp->out2fp) {
+ fd = sg_out_open(clp, clp->out2fp,
+ (out_mmap ? &rep->buffp : NULL),
+ (out_mmap ? &rep->mmap_len : NULL));
+ if (fd < 0)
+ goto fini;
+ rep->out2fd = fd;
+ own_out2fd = true;
+ if (vb > 2)
+ pr2serr_lk("thread=%d: opened local sg OFILE2\n", rep->id);
+ }
+ }
+ if (vb > 2) {
+ if (in_is_sg && (! own_infd))
+ pr2serr_lk("thread=%d: using global sg IFILE, fd=%d\n", rep->id,
+ rep->infd);
+ if (out_is_sg && (! own_outfd))
+ pr2serr_lk("thread=%d: using global sg OFILE, fd=%d\n", rep->id,
+ rep->outfd);
+ if ((FT_SG == clp->out2_type) && (! own_out2fd))
+ pr2serr_lk("thread=%d: using global sg OFILE2, fd=%d\n", rep->id,
+ rep->out2fd);
+ }
+ if (!sg_version_ge_40045) {
+ if (vb > 4)
+ pr2serr_lk("thread=%d: Skipping share because driver too old\n",
+ rep->id);
+ } else if (clp->noshare) {
+ if (vb > 4)
+ pr2serr_lk("thread=%d: Skipping IFILE share with OFILE due to "
+ "noshare=1\n", rep->id);
+ } else if (sg_version_ge_40045 && in_is_sg && out_is_sg)
+ rep->has_share = sg_share_prepare(rep->outfd, rep->infd, rep->id,
+ vb > 9);
+ if (vb > 9)
+ pr2serr_lk("tid=%d, has_share=%s\n", rep->id,
+ (rep->has_share ? "true" : "false"));
+ share_and_ofreg = (rep->has_share && (rep->outregfd >= 0));
+
+ /* vvvvvvvvvvvvvv Main segment copy loop vvvvvvvvvvvvvvvvvvvvvvv */
+ while (1) {
+ rep->wr = false;
+ my_index = atomic_fetch_add(&pos_index, (long int)clp->bpt);
+ /* Start of READ half of a segment */
+ buffp_onto_next(rep);
+ status = pthread_mutex_lock(&clp->in_mutex);
+ if (0 != status) err_exit(status, "lock in_mutex");
+
+ if (dd_count >= 0) {
+ if (my_index >= dd_count) {
+ status = pthread_mutex_unlock(&clp->in_mutex);
+ if (0 != status) err_exit(status, "unlock in_mutex");
+ if ((clp->nmrqs > 0) && (deferred_arr.first.size() > 0)) {
+ if (vb > 2)
+ pr2serr_lk("thread=%d: tail-end my_index>=dd_count, "
+ "to_do=%u\n", rep->id,
+ (uint32_t)deferred_arr.first.size());
+ res = sgh_do_deferred_mrq(rep, deferred_arr);
+ if (res)
+ pr2serr_lk("%s tid=%d: sgh_do_deferred_mrq failed\n",
+ __func__, rep->id);
+ }
+ break; /* at or beyond end, so leave loop >>>>>>>>>> */
+ } else if ((my_index + clp->bpt) > dd_count)
+ blocks = dd_count - my_index;
+ else
+ blocks = clp->bpt;
+ } else
+ blocks = clp->bpt;
+
+ rep->iblk = clp->skip + my_index;
+ rep->oblk = clp->seek + my_index;
+ rep->num_blks = blocks;
+
+ // clp->in_blk += blocks;
+ // clp->in_count -= blocks;
+
+ pthread_cleanup_push(cleanup_in, (void *)clp);
+ if (in_is_sg)
+ sg_in_rd_cmd(clp, rep, deferred_arr);
+ else {
+ stop_after_write = normal_in_rd(rep, blocks);
+ status = pthread_mutex_unlock(&clp->in_mutex);
+ if (0 != status) err_exit(status, "unlock in_mutex");
+ }
+ if (c_addr && (rep->bs > 3)) {
+ int k, j, off, num;
+ uint32_t addr = (uint32_t)rep->iblk;
+
+ num = (1 == c_addr) ? 4 : (rep->bs - 3);
+ for (k = 0, off = 0; k < blocks; ++k, ++addr, off += rep->bs) {
+ for (j = 0; j < num; j += 4) {
+ if (addr != sg_get_unaligned_be32(rep->buffp + off + j))
+ break;
+ }
+ if (j < num)
+ break;
+ }
+ if (k < blocks) {
+ pr2serr("%s: chkaddr failure at addr=0x%x\n", __func__, addr);
+ exit_status = SG_LIB_CAT_MISCOMPARE;
+ ++num_miscompare;
+ stop_both(clp);
+ }
+ }
+ pthread_cleanup_pop(0);
+ ++rep->rep_count;
+
+ /* Start of WRITE part of a segment */
+ rep->wr = true;
+ status = pthread_mutex_lock(&clp->out_mutex);
+ if (0 != status) err_exit(status, "lock out_mutex");
+
+ /* Make sure the OFILE (+ OFREG) are in same sequence as IFILE */
+ if (clp->in_flags.random)
+ goto skip_force_out_sequence;
+ if ((rep->outregfd < 0) && in_is_sg && out_is_sg)
+ goto skip_force_out_sequence;
+ if (share_and_ofreg || (FT_DEV_NULL != clp->out_type)) {
+ while ((! clp->out_stop.load()) &&
+ (rep->oblk != clp->out_blk.load())) {
+ /* if write would be out of sequence then wait */
+ pthread_cleanup_push(cleanup_out, (void *)clp);
+ status = pthread_cond_wait(&clp->out_sync_cv, &clp->out_mutex);
+ if (0 != status) err_exit(status, "cond out_sync_cv");
+ pthread_cleanup_pop(0);
+ }
+ }
+
+skip_force_out_sequence:
+ if (clp->out_stop.load() || (clp->out_count.load() <= 0)) {
+ if (! clp->out_stop.load())
+ clp->out_stop = true;
+ status = pthread_mutex_unlock(&clp->out_mutex);
+ if (0 != status) err_exit(status, "unlock out_mutex");
+ break; /* stop requested so leave loop >>>>>>>>>> */
+ }
+ if (stop_after_write)
+ clp->out_stop = true;
+
+ clp->out_count -= blocks;
+ clp->out_blk += blocks;
+
+ pthread_cleanup_push(cleanup_out, (void *)clp);
+ if (rep->outregfd >= 0) {
+ res = write(rep->outregfd, get_buffp(rep),
+ rep->bs * rep->num_blks);
+ err = errno;
+ if (res < 0)
+ pr2serr_lk("%s: tid=%d: write(outregfd) failed: %s\n",
+ __func__, rep->id, strerror(err));
+ else if (vb > 9)
+ pr2serr_lk("%s: tid=%d: write(outregfd), fd=%d, num_blks=%d"
+ "\n", __func__, rep->id, rep->outregfd,
+ rep->num_blks);
+ }
+ /* Output to OFILE */
+ wr_blks = rep->num_blks;
+ if (out_is_sg) {
+ sg_out_wr_cmd(rep, deferred_arr, false, clp->prefetch);
+ ++rep->rep_count;
+ } else if (FT_DEV_NULL == clp->out_type) {
+ /* skip actual write operation */
+ wr_blks = 0;
+ clp->out_rem_count -= blocks;
+ status = pthread_mutex_unlock(&clp->out_mutex);
+ if (0 != status) err_exit(status, "unlock out_mutex");
+ } else {
+ normal_out_wr(rep, blocks);
+ status = pthread_mutex_unlock(&clp->out_mutex);
+ if (0 != status) err_exit(status, "unlock out_mutex");
+ ++rep->rep_count;
+ }
+ pthread_cleanup_pop(0);
+
+ /* Output to OFILE2 if sg device */
+ if ((clp->out2fd >= 0) && (FT_SG == clp->out2_type)) {
+ pthread_cleanup_push(cleanup_out, (void *)clp);
+ status = pthread_mutex_lock(&clp->out2_mutex);
+ if (0 != status) err_exit(status, "lock out2_mutex");
+ /* releases out2_mutex mid operation */
+ sg_out_wr_cmd(rep, deferred_arr, true, false);
+
+ pthread_cleanup_pop(0);
+ }
+ if (0 == rep->num_blks) {
+ if ((clp->nmrqs > 0) && (deferred_arr.first.size() > 0)) {
+ if (wr_blks > 0)
+ rep->out_mrq_q_blks += wr_blks;
+ if (vb > 2)
+ pr2serr_lk("thread=%d: tail-end, to_do=%u\n", rep->id,
+ (uint32_t)deferred_arr.first.size());
+ res = sgh_do_deferred_mrq(rep, deferred_arr);
+ if (res)
+ pr2serr_lk("%s tid=%d: sgh_do_deferred_mrq failed\n",
+ __func__, rep->id);
+ }
+ clp->out_stop = true;
+ stop_after_write = true;
+ break; /* read nothing so leave loop >>>>>>>>>> */
+ }
+ // if ((! rep->has_share) && (FT_DEV_NULL != clp->out_type))
+ pthread_cond_broadcast(&clp->out_sync_cv);
+ if (stop_after_write)
+ break; /* leaving main loop >>>>>>>>> */
+ } /* ^^^^^^^^^^ end of main while loop which copies segments ^^^^^^ */
+
+ status = pthread_mutex_lock(&clp->in_mutex);
+ if (0 != status) err_exit(status, "lock in_mutex");
+ if (! clp->in_stop.load())
+ clp->in_stop = true; /* flag other workers to stop */
+ status = pthread_mutex_unlock(&clp->in_mutex);
+ if (0 != status) err_exit(status, "unlock in_mutex");
+
+fini:
+ if ((1 == rep->mmap_active) && (rep->mmap_len > 0)) {
+ if (munmap(rep->buffp, rep->mmap_len) < 0) {
+ err = errno;
+ char bb[STRERR_BUFF_LEN + 1];
+
+ pr2serr_lk("thread=%d: munmap() failed: %s\n", rep->id,
+ tsafe_strerror(err, bb));
+ }
+ if (vb > 4)
+ pr2serr_lk("thread=%d: munmap(%p, %d)\n", rep->id, rep->buffp,
+ rep->mmap_len);
+ rep->mmap_active = 0;
+ }
+ if (rep->alloc_bp) {
+ free(rep->alloc_bp);
+ rep->alloc_bp = NULL;
+ rep->buffp = NULL;
+ }
+
+ if (sg_version_ge_40045) {
+ if (clp->noshare) {
+ if ((clp->nmrqs > 0) && clp->unshare)
+ sg_unshare(rep->infd, rep->id, vb > 9);
+ } else if (in_is_sg && out_is_sg)
+ if (clp->unshare)
+ sg_unshare(rep->infd, rep->id, vb > 9);
+ }
+ if (own_infd && (rep->infd >= 0)) {
+ if (vb && in_is_sg) {
+ ++num_waiting_calls;
+ if (ioctl(rep->infd, SG_GET_NUM_WAITING, &n) >= 0) {
+ if (n > 0)
+ pr2serr_lk("%s: tid=%d: num_waiting=%d prior close(in)\n",
+ __func__, rep->id, n);
+ } else {
+ err = errno;
+ pr2serr_lk("%s: [%d] ioctl(SG_GET_NUM_WAITING) errno=%d: "
+ "%s\n", __func__, rep->id, err, strerror(err));
+ }
+ }
+ close(rep->infd);
+ }
+ if (own_outfd && (rep->outfd >= 0)) {
+ if (vb && out_is_sg) {
+ ++num_waiting_calls;
+ if (ioctl(rep->outfd, SG_GET_NUM_WAITING, &n) >= 0) {
+ if (n > 0)
+ pr2serr_lk("%s: tid=%d: num_waiting=%d prior "
+ "close(out)\n", __func__, rep->id, n);
+ } else {
+ err = errno;
+ pr2serr_lk("%s: [%d] ioctl(SG_GET_NUM_WAITING) errno=%d: "
+ "%s\n", __func__, rep->id, err, strerror(err));
+ }
+ }
+ close(rep->outfd);
+ }
+ if (own_out2fd && (rep->out2fd >= 0))
+ close(rep->out2fd);
+ pthread_cond_broadcast(&clp->out_sync_cv);
+ return stop_after_write ? NULL : clp;
+}
+
+/* N.B. A return of true means it wants to stop the copy. So false is the
+ * 'good' reply (i.e. keep going). */
+static bool
+normal_in_rd(Rq_elem * rep, int blocks)
+{
+ struct global_collection * clp = rep->clp;
+ bool stop_after_write = false;
+ bool same_fds = clp->in_flags.same_fds || clp->out_flags.same_fds;
+ int res;
+
+ if (clp->verbose > 4)
+ pr2serr_lk("%s: tid=%d: iblk=%" PRIu64 ", blocks=%d\n", __func__,
+ rep->id, rep->iblk, blocks);
+ if (FT_RANDOM_0_FF == clp->in_type) {
+ int k, j;
+ const int jbump = sizeof(uint32_t);
+ long rn;
+ uint8_t * bp;
+
+ if (clp->in_flags.zero && clp->in_flags.ff && (rep->bs >= 4)) {
+ uint32_t pos = (uint32_t)rep->iblk;
+ uint32_t off;
+
+ for (k = 0, off = 0; k < blocks; ++k, off += rep->bs, ++pos) {
+ for (j = 0; j < (rep->bs - 3); j += 4)
+ sg_put_unaligned_be32(pos, rep->buffp + off + j);
+ }
+ } else if (clp->in_flags.zero)
+ memset(rep->buffp, 0, blocks * rep->bs);
+ else if (clp->in_flags.ff)
+ memset(rep->buffp, 0xff, blocks * rep->bs);
+ else {
+ for (k = 0, bp = rep->buffp; k < blocks; ++k, bp += rep->bs) {
+ for (j = 0; j < rep->bs; j += jbump) {
+ /* mrand48 takes uniformly from [-2^31, 2^31) */
+#ifdef HAVE_SRAND48_R
+ mrand48_r(&rep->drand, &rn);
+#else
+ rn = mrand48();
+#endif
+ *((uint32_t *)(bp + j)) = (uint32_t)rn;
+ }
+ }
+ }
+ clp->in_rem_count -= blocks;
+ return stop_after_write;
+ }
+ if (! same_fds) { /* each has own file pointer, so we need to move it */
+ int64_t pos = rep->iblk * rep->bs;
+
+ if (lseek64(rep->infd, pos, SEEK_SET) < 0) { /* problem if pipe! */
+ pr2serr_lk("%s: tid=%d: >> lseek64(%" PRId64 "): %s\n", __func__,
+ rep->id, pos, safe_strerror(errno));
+ stop_both(clp);
+ return true;
+ }
+ }
+ /* enters holding in_mutex */
+ while (((res = read(clp->infd, rep->buffp, blocks * rep->bs)) < 0) &&
+ ((EINTR == errno) || (EAGAIN == errno)))
+ std::this_thread::yield();/* another thread may be able to progress */
+ if (res < 0) {
+ char strerr_buff[STRERR_BUFF_LEN + 1];
+
+ if (clp->in_flags.coe) {
+ memset(rep->buffp, 0, rep->num_blks * rep->bs);
+ pr2serr_lk("tid=%d: >> substituted zeros for in blk=%" PRId64
+ " for %d bytes, %s\n", rep->id, rep->iblk,
+ rep->num_blks * rep->bs,
+ tsafe_strerror(errno, strerr_buff));
+ res = rep->num_blks * rep->bs;
+ }
+ else {
+ pr2serr_lk("tid=%d: error in normal read, %s\n", rep->id,
+ tsafe_strerror(errno, strerr_buff));
+ stop_both(clp);
+ return true;
+ }
+ }
+ if (res < blocks * rep->bs) {
+ // int o_blocks = blocks;
+
+ stop_after_write = true;
+ blocks = res / rep->bs;
+ if ((res % rep->bs) > 0) {
+ blocks++;
+ clp->in_partial++;
+ }
+ /* Reverse out + re-apply blocks on clp */
+ // clp->in_blk -= o_blocks;
+ // clp->in_count += o_blocks;
+ rep->num_blks = blocks;
+ // clp->in_blk += blocks;
+ // clp->in_count -= blocks;
+ }
+ clp->in_rem_count -= blocks;
+ return stop_after_write;
+}
+
+static void
+normal_out_wr(Rq_elem * rep, int blocks)
+{
+ int res;
+ struct global_collection * clp = rep->clp;
+
+ /* enters holding out_mutex */
+ if (clp->verbose > 4)
+ pr2serr_lk("%s: tid=%d: oblk=%" PRIu64 ", blocks=%d\n", __func__,
+ rep->id, rep->oblk, blocks);
+ while (((res = write(clp->outfd, rep->buffp, blocks * rep->bs))
+ < 0) && ((EINTR == errno) || (EAGAIN == errno)))
+ std::this_thread::yield();/* another thread may be able to progress */
+ if (res < 0) {
+ char strerr_buff[STRERR_BUFF_LEN + 1];
+
+ if (clp->out_flags.coe) {
+ pr2serr_lk("tid=%d: >> ignored error for out blk=%" PRId64
+ " for %d bytes, %s\n", rep->id, rep->oblk,
+ rep->num_blks * rep->bs,
+ tsafe_strerror(errno, strerr_buff));
+ res = rep->num_blks * rep->bs;
+ }
+ else {
+ pr2serr_lk("tid=%d: error normal write, %s\n", rep->id,
+ tsafe_strerror(errno, strerr_buff));
+ stop_both(clp);
+ return;
+ }
+ }
+ if (res < blocks * rep->bs) {
+ blocks = res / rep->bs;
+ if ((res % rep->bs) > 0) {
+ blocks++;
+ clp->out_partial++;
+ }
+ rep->num_blks = blocks;
+ }
+ clp->out_rem_count -= blocks;
+}
+
+static int
+sg_build_scsi_cdb(uint8_t * cdbp, int cdb_sz, unsigned int blocks,
+ int64_t start_block, bool ver_true, bool write_true,
+ bool fua, bool dpo)
+{
+ int rd_opcode[] = {0x8, 0x28, 0xa8, 0x88};
+ int ve_opcode[] = {0xff /* no VER(6) */, 0x2f, 0xaf, 0x8f};
+ int wr_opcode[] = {0xa, 0x2a, 0xaa, 0x8a};
+ int sz_ind;
+
+ memset(cdbp, 0, cdb_sz);
+ if (ver_true) { /* only support VERIFY(10) */
+ if (cdb_sz < 10) {
+ pr2serr_lk("%sonly support VERIFY(10)\n", my_name);
+ return 1;
+ }
+ cdb_sz = 10;
+ fua = false;
+ cdbp[1] |= 0x2; /* BYTCHK=1 --> sending dout for comparison */
+ cdbp[0] = ve_opcode[1];
+ }
+ if (dpo)
+ cdbp[1] |= 0x10;
+ if (fua)
+ cdbp[1] |= 0x8;
+ switch (cdb_sz) {
+ case 6:
+ sz_ind = 0;
+ cdbp[0] = (uint8_t)(write_true ? wr_opcode[sz_ind] :
+ rd_opcode[sz_ind]);
+ sg_put_unaligned_be24(0x1fffff & start_block, cdbp + 1);
+ cdbp[4] = (256 == blocks) ? 0 : (uint8_t)blocks;
+ if (blocks > 256) {
+ pr2serr_lk("%sfor 6 byte commands, maximum number of blocks is "
+ "256\n", my_name);
+ return 1;
+ }
+ if ((start_block + blocks - 1) & (~0x1fffff)) {
+ pr2serr_lk("%sfor 6 byte commands, can't address blocks beyond "
+ "%d\n", my_name, 0x1fffff);
+ return 1;
+ }
+ if (dpo || fua) {
+ pr2serr_lk("%sfor 6 byte commands, neither dpo nor fua bits "
+ "supported\n", my_name);
+ return 1;
+ }
+ break;
+ case 10:
+ if (! ver_true) {
+ sz_ind = 1;
+ cdbp[0] = (uint8_t)(write_true ? wr_opcode[sz_ind] :
+ rd_opcode[sz_ind]);
+ }
+ sg_put_unaligned_be32((uint32_t)start_block, cdbp + 2);
+ sg_put_unaligned_be16((uint16_t)blocks, cdbp + 7);
+ if (blocks & (~0xffff)) {
+ pr2serr_lk("%sfor 10 byte commands, maximum number of blocks is "
+ "%d\n", my_name, 0xffff);
+ return 1;
+ }
+ break;
+ case 12:
+ sz_ind = 2;
+ cdbp[0] = (uint8_t)(write_true ? wr_opcode[sz_ind] :
+ rd_opcode[sz_ind]);
+ sg_put_unaligned_be32((uint32_t)start_block, cdbp + 2);
+ sg_put_unaligned_be32((uint32_t)blocks, cdbp + 6);
+ break;
+ case 16:
+ sz_ind = 3;
+ cdbp[0] = (uint8_t)(write_true ? wr_opcode[sz_ind] :
+ rd_opcode[sz_ind]);
+ sg_put_unaligned_be64((uint64_t)start_block, cdbp + 2);
+ sg_put_unaligned_be32((uint32_t)blocks, cdbp + 10);
+ break;
+ default:
+ pr2serr_lk("%sexpected cdb size of 6, 10, 12, or 16 but got %d\n",
+ my_name, cdb_sz);
+ return 1;
+ }
+ return 0;
+}
+
+/* Enters this function holding in_mutex */
+static void
+sg_in_rd_cmd(struct global_collection * clp, Rq_elem * rep,
+ mrq_arr_t & def_arr)
+{
+ int status, pack_id;
+
+ while (1) {
+ int res = sg_start_io(rep, def_arr, pack_id, NULL);
+
+ if (1 == res)
+ err_exit(ENOMEM, "sg starting in command");
+ else if (res < 0) {
+ pr2serr_lk("tid=%d: inputting to sg failed, blk=%" PRId64 "\n",
+ rep->id, rep->iblk);
+ status = pthread_mutex_unlock(&clp->in_mutex);
+ if (0 != status) err_exit(status, "unlock in_mutex");
+ stop_both(clp);
+ return;
+ }
+ /* Now release in mutex to let other reads run in parallel */
+ status = pthread_mutex_unlock(&clp->in_mutex);
+ if (0 != status) err_exit(status, "unlock in_mutex");
+
+ res = sg_finish_io(rep->wr, rep, pack_id, NULL);
+ switch (res) {
+ case SG_LIB_CAT_ABORTED_COMMAND:
+ case SG_LIB_CAT_UNIT_ATTENTION:
+ /* try again with same addr, count info */
+ /* now re-acquire in mutex for balance */
+ /* N.B. This re-read could now be out of read sequence */
+ status = pthread_mutex_lock(&clp->in_mutex);
+ if (0 != status) err_exit(status, "lock in_mutex");
+ break; /* will loop again */
+ case SG_LIB_CAT_MEDIUM_HARD:
+ if (0 == clp->in_flags.coe) {
+ pr2serr_lk("error finishing sg in command (medium)\n");
+ if (exit_status <= 0)
+ exit_status = res;
+ stop_both(clp);
+ return;
+ } else {
+ memset(get_buffp(rep), 0, rep->num_blks * rep->bs);
+ pr2serr_lk("tid=%d: >> substituted zeros for in blk=%" PRId64
+ " for %d bytes\n", rep->id, rep->iblk,
+ rep->num_blks * rep->bs);
+ }
+#if defined(__GNUC__)
+#if (__GNUC__ >= 7)
+ __attribute__((fallthrough));
+ /* FALL THROUGH */
+#endif
+#endif
+ case 0:
+ status = pthread_mutex_lock(&clp->in_mutex);
+ if (0 != status) err_exit(status, "lock in_mutex");
+ if (rep->dio_incomplete_count || rep->resid) {
+ clp->dio_incomplete_count += rep->dio_incomplete_count;
+ clp->sum_of_resids += rep->resid;
+ }
+ clp->in_rem_count -= rep->num_blks;
+ status = pthread_mutex_unlock(&clp->in_mutex);
+ if (0 != status) err_exit(status, "unlock in_mutex");
+ return;
+ default:
+ pr2serr_lk("tid=%d: error finishing sg in command (%d)\n",
+ rep->id, res);
+ if (exit_status <= 0)
+ exit_status = res;
+ stop_both(clp);
+ return;
+ }
+ } /* end of while (1) loop */
+}
+
+static bool
+sg_wr_swap_share(Rq_elem * rep, int to_fd, bool before)
+{
+ bool not_first = false;
+ int err = 0;
+ int k;
+ int read_side_fd = rep->infd;
+ struct global_collection * clp = rep->clp;
+ struct sg_extended_info sei {};
+ struct sg_extended_info * seip = &sei;
+
+ if (rep->clp->verbose > 2)
+ pr2serr_lk("%s: tid=%d: to_fd=%d, before=%d\n", __func__, rep->id,
+ to_fd, (int)before);
+ seip->sei_wr_mask |= SG_SEIM_CHG_SHARE_FD;
+ seip->sei_rd_mask |= SG_SEIM_CHG_SHARE_FD;
+ seip->share_fd = to_fd;
+ if (before) {
+ /* clear READ_SIDE_FINI bit: puts read-side in SG_RQ_SHR_SWAP state */
+ seip->sei_wr_mask |= SG_SEIM_CTL_FLAGS;
+ seip->sei_rd_mask |= SG_SEIM_CTL_FLAGS;
+ seip->ctl_flags_wr_mask |= SG_CTL_FLAGM_READ_SIDE_FINI;
+ seip->ctl_flags &= ~SG_CTL_FLAGM_READ_SIDE_FINI;/* would be 0 anyway */
+ }
+ for (k = 0; (ioctl(read_side_fd, SG_SET_GET_EXTENDED, seip) < 0) &&
+ (EBUSY == errno); ++k) {
+ err = errno;
+ if (k > 10000)
+ break;
+ if (! not_first) {
+ if (clp->verbose > 3)
+ pr2serr_lk("tid=%d: ioctl(EXTENDED(change_shared_fd=%d), "
+ "failed errno=%d %s\n", rep->id, read_side_fd, err,
+ strerror(err));
+ not_first = true;
+ }
+ err = 0;
+ std::this_thread::yield();/* another thread may be able to progress */
+ }
+ if (err) {
+ pr2serr_lk("tid=%d: ioctl(EXTENDED(change_shared_fd=%d), failed "
+ "errno=%d %s\n", rep->id, read_side_fd, err, strerror(err));
+ return false;
+ }
+ if (clp->verbose > 15)
+ pr2serr_lk("%s: tid=%d: ioctl(EXTENDED(change_shared_fd)) ok, "
+ "read_side_fd=%d, to_write_side_fd=%d\n", __func__, rep->id,
+ read_side_fd, to_fd);
+ return true;
+}
+
+/* Enters this function holding out_mutex */
+static void
+sg_out_wr_cmd(Rq_elem * rep, mrq_arr_t & def_arr, bool is_wr2, bool prefetch)
+{
+ int res, status, pack_id, nblks;
+ struct global_collection * clp = rep->clp;
+ uint32_t ofsplit = clp->ofsplit;
+ pthread_mutex_t * mutexp = is_wr2 ? &clp->out2_mutex : &clp->out_mutex;
+ struct sg_io_extra xtr {};
+ struct sg_io_extra * xtrp = &xtr;
+ const char * wr_or_ver = clp->verify ? "verify" : "out";
+
+ xtrp->is_wr2 = is_wr2;
+ xtrp->prefetch = prefetch;
+ nblks = rep->num_blks;
+ if (rep->has_share && is_wr2)
+ sg_wr_swap_share(rep, rep->out2fd, true);
+
+ if (prefetch) {
+again:
+ res = sg_start_io(rep, def_arr, pack_id, xtrp);
+ if (1 == res)
+ err_exit(ENOMEM, "sg starting out command");
+ else if (res < 0) {
+ pr2serr_lk("%ssg %s failed, blk=%" PRId64 "\n",
+ my_name, wr_or_ver, rep->oblk);
+ status = pthread_mutex_unlock(mutexp);
+ if (0 != status) err_exit(status, "unlock out_mutex");
+ stop_both(clp);
+ goto fini;
+ }
+ /* Now release in mutex to let other reads run in parallel */
+ status = pthread_mutex_unlock(mutexp);
+ if (0 != status) err_exit(status, "unlock out_mutex");
+
+ res = sg_finish_io(rep->wr, rep, pack_id, xtrp);
+ switch (res) {
+ case SG_LIB_CAT_ABORTED_COMMAND:
+ case SG_LIB_CAT_UNIT_ATTENTION:
+ /* try again with same addr, count info */
+ /* now re-acquire out mutex for balance */
+ /* N.B. This re-write could now be out of write sequence */
+ status = pthread_mutex_lock(mutexp);
+ if (0 != status) err_exit(status, "lock out_mutex");
+ goto again;
+ case SG_LIB_CAT_CONDITION_MET:
+ case 0:
+ status = pthread_mutex_lock(mutexp);
+ if (0 != status) err_exit(status, "unlock out_mutex");
+ break;
+ default:
+ pr2serr_lk("error finishing sg prefetch command (%d)\n", res);
+ if (exit_status <= 0)
+ exit_status = res;
+ stop_both(clp);
+ goto fini;
+ }
+ }
+
+ /* start write (or verify) on current segment on sg device */
+ xtrp->prefetch = false;
+ if ((ofsplit > 0) && (rep->num_blks > (int)ofsplit)) {
+ xtrp->dout_is_split = true;
+ xtrp->blk_offset = 0;
+ xtrp->blks = ofsplit;
+ nblks = ofsplit;
+ xtrp->hpv4_ind = 0;
+ }
+split_upper:
+ while (1) {
+ res = sg_start_io(rep, def_arr, pack_id, xtrp);
+ if (1 == res)
+ err_exit(ENOMEM, "sg starting out command");
+ else if (res < 0) {
+ pr2serr_lk("%ssg %s failed, blk=%" PRId64 "\n", my_name,
+ wr_or_ver, rep->oblk);
+ status = pthread_mutex_unlock(mutexp);
+ if (0 != status) err_exit(status, "unlock out_mutex");
+ stop_both(clp);
+ goto fini;
+ }
+ /* Now release in mutex to let other reads run in parallel */
+ status = pthread_mutex_unlock(mutexp);
+ if (0 != status) err_exit(status, "unlock out_mutex");
+
+ res = sg_finish_io(rep->wr, rep, pack_id, xtrp);
+ switch (res) {
+ case SG_LIB_CAT_ABORTED_COMMAND:
+ case SG_LIB_CAT_UNIT_ATTENTION:
+ /* try again with same addr, count info */
+ /* now re-acquire out mutex for balance */
+ /* N.B. This re-write could now be out of write sequence */
+ status = pthread_mutex_lock(mutexp);
+ if (0 != status) err_exit(status, "lock out_mutex");
+ break; /* loops around */
+ case SG_LIB_CAT_MEDIUM_HARD:
+ if (0 == clp->out_flags.coe) {
+ pr2serr_lk("error finishing sg %s command (medium)\n",
+ wr_or_ver);
+ if (exit_status <= 0)
+ exit_status = res;
+ stop_both(clp);
+ goto fini;
+ } else
+ pr2serr_lk(">> ignored error for %s blk=%" PRId64 " for %d "
+ "bytes\n", wr_or_ver, rep->oblk, nblks * rep->bs);
+#if defined(__GNUC__)
+#if (__GNUC__ >= 7)
+ __attribute__((fallthrough));
+ /* FALL THROUGH */
+#endif
+#endif
+ case SG_LIB_CAT_CONDITION_MET:
+ case 0:
+ if (! is_wr2) {
+ status = pthread_mutex_lock(mutexp);
+ if (0 != status) err_exit(status, "lock out_mutex");
+ if (rep->dio_incomplete_count || rep->resid) {
+ clp->dio_incomplete_count += rep->dio_incomplete_count;
+ clp->sum_of_resids += rep->resid;
+ }
+ clp->out_rem_count -= nblks;
+ status = pthread_mutex_unlock(mutexp);
+ if (0 != status) err_exit(status, "unlock out_mutex");
+ }
+ goto fini;
+ case SG_LIB_CAT_MISCOMPARE:
+ ++num_miscompare;
+ // fall through
+ default:
+ pr2serr_lk("error finishing sg %s command (%d)\n", wr_or_ver,
+ res);
+ if (exit_status <= 0)
+ exit_status = res;
+ stop_both(clp);
+ goto fini;
+ }
+ } /* end of while (1) loop */
+fini:
+ if (xtrp->dout_is_split) { /* set up upper half of split */
+ if ((0 == xtrp->hpv4_ind) && (rep->num_blks > (int)ofsplit)) {
+ xtrp->hpv4_ind = 1;
+ xtrp->blk_offset = ofsplit;
+ xtrp->blks = rep->num_blks - ofsplit;
+ nblks = xtrp->blks;
+ status = pthread_mutex_lock(mutexp);
+ if (0 != status) err_exit(status, "lock out_mutex");
+ goto split_upper;
+ }
+ }
+ if (rep->has_share && is_wr2)
+ sg_wr_swap_share(rep, rep->outfd, false);
+}
+
+static int
+process_mrq_response(Rq_elem * rep, const struct sg_io_v4 * ctl_v4p,
+ const struct sg_io_v4 * a_v4p, int num_mrq,
+ uint32_t & good_inblks, uint32_t & good_outblks)
+{
+ struct global_collection * clp = rep->clp;
+ bool ok, all_good;
+ bool sb_in_co = !!(ctl_v4p->response);
+ int id = rep->id;
+ int resid = ctl_v4p->din_resid;
+ int sres = ctl_v4p->spare_out;
+ int n_subm = num_mrq - ctl_v4p->dout_resid;
+ int n_cmpl = ctl_v4p->info;
+ int n_good = 0;
+ int hole_count = 0;
+ int vb = clp->verbose;
+ int k, j, f1, slen, blen;
+ char b[160];
+
+ blen = sizeof(b);
+ good_inblks = 0;
+ good_outblks = 0;
+ if (vb > 2)
+ pr2serr_lk("[thread_id=%d] %s: num_mrq=%d, n_subm=%d, n_cmpl=%d\n",
+ id, __func__, num_mrq, n_subm, n_cmpl);
+ if (n_subm < 0) {
+ pr2serr_lk("[%d] co.dout_resid(%d) > num_mrq(%d)\n", id,
+ ctl_v4p->dout_resid, num_mrq);
+ return -1;
+ }
+ if (n_cmpl != (num_mrq - resid))
+ pr2serr_lk("[%d] co.info(%d) != (num_mrq(%d) - co.din_resid(%d))\n"
+ "will use co.info\n", id, n_cmpl, num_mrq, resid);
+ if (n_cmpl > n_subm) {
+ pr2serr_lk("[%d] n_cmpl(%d) > n_subm(%d), use n_subm for both\n",
+ id, n_cmpl, n_subm);
+ n_cmpl = n_subm;
+ }
+ if (sres) {
+ pr2serr_lk("[%d] secondary error: %s [%d], info=0x%x\n", id,
+ strerror(sres), sres, ctl_v4p->info);
+ if (E2BIG == sres) {
+ sg_take_snap(rep->infd, id, true);
+ sg_take_snap(rep->outfd, id, true);
+ }
+ }
+ /* Check if those submitted have finished or not. N.B. If there has been
+ * an error then there may be "holes" (i.e. info=0x0) in the array due
+ * to completions being out-of-order. */
+ for (k = 0, j = 0; ((k < num_mrq) && (j < n_subm));
+ ++k, j += f1, ++a_v4p) {
+ slen = a_v4p->response_len;
+ if (! (SG_INFO_MRQ_FINI & a_v4p->info))
+ ++hole_count;
+ ok = true;
+ f1 = !!(a_v4p->info); /* want to skip n_subm count if info is 0x0 */
+ if (SG_INFO_CHECK & a_v4p->info) {
+ ok = false;
+ pr2serr_lk("[%d] a_v4[%d]: SG_INFO_CHECK set [%s]\n", id, k,
+ sg_info_str(a_v4p->info, sizeof(b), b));
+ }
+ if (sg_scsi_status_is_bad(a_v4p->device_status) ||
+ a_v4p->transport_status || a_v4p->driver_status) {
+ ok = false;
+ if (SAM_STAT_CHECK_CONDITION != a_v4p->device_status) {
+ pr2serr_lk("[%d] a_v4[%d]:\n", id, k);
+ if (vb)
+ lk_chk_n_print4(" >>", a_v4p, vb > 4);
+ }
+ }
+ if (slen > 0) {
+ struct sg_scsi_sense_hdr ssh;
+ const uint8_t *sbp = (const uint8_t *)
+ (sb_in_co ? ctl_v4p->response : a_v4p->response);
+
+ if (sg_scsi_normalize_sense(sbp, slen, &ssh) &&
+ (ssh.response_code >= 0x70)) {
+ if (ssh.response_code & 0x1)
+ ok = true;
+ if (vb) {
+ sg_get_sense_str(" ", sbp, slen, false, blen, b);
+ pr2serr_lk("[%d] a_v4[%d]:\n%s\n", id, k, b);
+ }
+ }
+ }
+ if (ok && f1) {
+ ++n_good;
+ if (a_v4p->dout_xfer_len >= (uint32_t)rep->bs)
+ good_outblks += (a_v4p->dout_xfer_len - a_v4p->dout_resid) /
+ rep->bs;
+ if (a_v4p->din_xfer_len >= (uint32_t)rep->bs)
+ good_inblks += (a_v4p->din_xfer_len - a_v4p->din_resid) /
+ rep->bs;
+ }
+ } /* end of request array scan loop */
+ if ((n_subm == num_mrq) || (vb < 3))
+ goto fini;
+ pr2serr_lk("[%d] checking response array _beyond_ number of "
+ "submissions [%d] to num_mrq:\n", id, k);
+ for (all_good = true; k < num_mrq; ++k, ++a_v4p) {
+ if (SG_INFO_MRQ_FINI & a_v4p->info) {
+ pr2serr_lk("[%d] a_v4[%d]: unexpected SG_INFO_MRQ_FINI set [%s]\n",
+ id, k, sg_info_str(a_v4p->info, sizeof(b), b));
+ all_good = false;
+ }
+ if (a_v4p->device_status || a_v4p->transport_status ||
+ a_v4p->driver_status) {
+ pr2serr_lk("[%d] a_v4[%d]:\n", id, k);
+ lk_chk_n_print4(" ", a_v4p, vb > 4);
+ all_good = false;
+ }
+ }
+ if (all_good)
+ pr2serr_lk(" ... all good\n");
+fini:
+ return n_good;
+}
+
+#if 0
+static int
+chk_mrq_response(Rq_elem * rep, const struct sg_io_v4 * ctl_v4p,
+ const struct sg_io_v4 * a_v4p, int nrq,
+ uint32_t * good_inblksp, uint32_t * good_outblksp)
+{
+ struct global_collection * clp = rep->clp;
+ bool ok;
+ int id = rep->id;
+ int resid = ctl_v4p->din_resid;
+ int sres = ctl_v4p->spare_out;
+ int n_subm = nrq - ctl_v4p->dout_resid;
+ int n_cmpl = ctl_v4p->info;
+ int n_good = 0;
+ int vb = clp->verbose;
+ int k, slen, blen;
+ uint32_t good_inblks = 0;
+ uint32_t good_outblks = 0;
+ const struct sg_io_v4 * a_np = a_v4p;
+ char b[80];
+
+ blen = sizeof(b);
+ if (n_subm < 0) {
+ pr2serr_lk("[%d] %s: co.dout_resid(%d) > nrq(%d)\n", id, __func__,
+ ctl_v4p->dout_resid, nrq);
+ return -1;
+ }
+ if (n_cmpl != (nrq - resid))
+ pr2serr_lk("[%d] %s: co.info(%d) != (nrq(%d) - co.din_resid(%d))\n"
+ "will use co.info\n", id, __func__, n_cmpl, nrq, resid);
+ if (n_cmpl > n_subm) {
+ pr2serr_lk("[%d] %s: n_cmpl(%d) > n_subm(%d), use n_subm for both\n",
+ id, __func__, n_cmpl, n_subm);
+ n_cmpl = n_subm;
+ }
+ if (sres) {
+ pr2serr_lk("[%d] %s: secondary error: %s [%d], info=0x%x\n", id,
+ __func__, strerror(sres), sres, ctl_v4p->info);
+ if (E2BIG == sres) {
+ sg_take_snap(rep->infd, id, true);
+ sg_take_snap(rep->outfd, id, true);
+ }
+ }
+ /* Check if those submitted have finished or not */
+ for (k = 0; k < n_subm; ++k, ++a_np) {
+ slen = a_np->response_len;
+ if (! (SG_INFO_MRQ_FINI & a_np->info)) {
+ pr2serr_lk("[%d] %s, a_n[%d]: missing SG_INFO_MRQ_FINI [%s]\n",
+ id, __func__, k, sg_info_str(a_np->info, blen, b));
+ v4hdr_out_lk("a_np", a_np, id);
+ v4hdr_out_lk("cop", ctl_v4p, id);
+ }
+ ok = true;
+ if (SG_INFO_CHECK & a_np->info) {
+ ok = false;
+ pr2serr_lk("[%d] a_n[%d]: SG_INFO_CHECK set [%s]\n", id, k,
+ sg_info_str(a_np->info, sizeof(b), b));
+ }
+ if (sg_scsi_status_is_bad(a_np->device_status) ||
+ a_np->transport_status || a_np->driver_status) {
+ ok = false;
+ if (SAM_STAT_CHECK_CONDITION != a_np->device_status) {
+ pr2serr_lk("[%d] %s, a_n[%d]:\n", id, __func__, k);
+ if (vb)
+ lk_chk_n_print4(" >>", a_np, false);
+ }
+ }
+ if (slen > 0) {
+ struct sg_scsi_sense_hdr ssh;
+ const uint8_t *sbp = (const uint8_t *)a_np->response;
+
+ if (sg_scsi_normalize_sense(sbp, slen, &ssh) &&
+ (ssh.response_code >= 0x70)) {
+ char b[256];
+
+ if (ssh.response_code & 0x1)
+ ok = true;
+ if (vb) {
+ sg_get_sense_str(" ", sbp, slen, false, sizeof(b), b);
+ pr2serr_lk("[%d] %s, a_n[%d]:\n%s\n", id, __func__, k, b);
+ }
+ }
+ }
+ if (ok) {
+ ++n_good;
+ if (a_np->dout_xfer_len >= (uint32_t)rep->bs)
+ good_outblks += (a_np->dout_xfer_len - a_np->dout_resid) /
+ rep->bs;
+ if (a_np->din_xfer_len >= (uint32_t)rep->bs)
+ good_inblks += (a_np->din_xfer_len - a_np->din_resid) /
+ rep->bs;
+ }
+ }
+ if ((n_subm == nrq) || (vb < 3))
+ goto fini;
+ pr2serr_lk("[%d] %s: checking response array beyond number of "
+ "submissions:\n", id, __func__);
+ for (k = n_subm; k < nrq; ++k, ++a_np) {
+ if (SG_INFO_MRQ_FINI & a_np->info)
+ pr2serr_lk("[%d] %s, a_n[%d]: unexpected SG_INFO_MRQ_FINI set\n",
+ id, __func__, k);
+ if (a_np->device_status || a_np->transport_status ||
+ a_np->driver_status) {
+ pr2serr_lk("[%d] %s, a_n[%d]:\n", id, __func__, k);
+ lk_chk_n_print4(" ", a_np, vb > 4);
+ }
+ }
+fini:
+ if (good_inblksp)
+ *good_inblksp = good_inblks;
+ if (good_outblksp)
+ *good_outblksp = good_outblks;
+ return n_good;
+}
+#endif
+
+/* do mrq 'full non-blocking' invocation so both submission and completion
+ * is async (i.e. uses SGV4_FLAG_IMMED flag). This type of mrq is
+ * restricted to a single file descriptor (i.e. the 'fd' argument). */
+static int
+sgh_do_async_mrq(Rq_elem * rep, mrq_arr_t & def_arr, int fd,
+ struct sg_io_v4 * ctlop, int nrq)
+{
+ int half = nrq / 2;
+ int k, res, nwait, half_num, rest, err, num_good, b_len;
+ const int64_t wait_us = 10;
+ uint32_t in_fin_blks, out_fin_blks;
+ struct sg_io_v4 * a_v4p;
+ struct sg_io_v4 hold_ctlo;
+ struct global_collection * clp = rep->clp;
+ char b[80];
+
+ hold_ctlo = *ctlop;
+ b_len = sizeof(b);
+ a_v4p = def_arr.first.data();
+ ctlop->flags = SGV4_FLAG_MULTIPLE_REQS;
+ if (clp->in_flags.polled || clp->out_flags.polled) {
+ /* submit of full non-blocking with POLLED */
+ ctlop->flags |= (SGV4_FLAG_IMMED | SGV4_FLAG_POLLED);
+ if (!after1 && (clp->verbose > 1)) {
+ after1 = true;
+ pr2serr_lk("%s: %s\n", __func__, mrq_s_nb_s);
+ }
+ } else {
+ ctlop->flags |= SGV4_FLAG_IMMED; /* submit non-blocking */
+ if (!after1 && (clp->verbose > 1)) {
+ after1 = true;
+ pr2serr_lk("%s: %s\n", __func__, mrq_s_nb_s);
+ }
+ }
+ if (clp->verbose > 4) {
+ pr2serr_lk("%s: Controlling object _before_ ioctl(SG_IOSUBMIT):\n",
+ __func__);
+ if (clp->verbose > 5)
+ hex2stderr_lk((const uint8_t *)ctlop, sizeof(*ctlop), 1);
+ v4hdr_out_lk("Controlling object before", ctlop, rep->id);
+ }
+ res = ioctl(fd, SG_IOSUBMIT, ctlop);
+ if (res < 0) {
+ err = errno;
+ if (E2BIG == err)
+ sg_take_snap(fd, rep->id, true);
+ pr2serr_lk("%s: ioctl(SG_IOSUBMIT, %s)-->%d, errno=%d: %s\n", __func__,
+ sg_flags_str(ctlop->flags, b_len, b), res, err,
+ strerror(err));
+ return -1;
+ }
+ /* fetch first half */
+ for (k = 0; k < 100000; ++k) {
+ ++num_waiting_calls;
+ res = ioctl(fd, SG_GET_NUM_WAITING, &nwait);
+ if (res < 0) {
+ err = errno;
+ pr2serr_lk("%s: ioctl(SG_GET_NUM_WAITING)-->%d, errno=%d: %s\n",
+ __func__, res, err, strerror(err));
+ return -1;
+ }
+ if (nwait >= half)
+ break;
+ this_thread::sleep_for(chrono::microseconds{wait_us});
+ }
+ ctlop->flags = (SGV4_FLAG_MULTIPLE_REQS | SGV4_FLAG_IMMED);
+ res = ioctl(fd, SG_IORECEIVE, ctlop);
+ if (res < 0) {
+ err = errno;
+ if (ENODATA != err) {
+ pr2serr_lk("%s: ioctl(SG_IORECEIVE, %s),1-->%d, errno=%d: %s\n",
+ __func__, sg_flags_str(ctlop->flags, b_len, b), res,
+ err, strerror(err));
+ return -1;
+ }
+ half_num = 0;
+ } else
+ half_num = ctlop->info;
+ if (clp->verbose > 4) {
+ pr2serr_lk("%s: Controlling object output by ioctl(SG_IORECEIVE),1: "
+ "num_received=%d\n", __func__, half_num);
+ if (clp->verbose > 5)
+ hex2stderr_lk((const uint8_t *)ctlop, sizeof(*ctlop), 1);
+ v4hdr_out_lk("Controlling object after", ctlop, rep->id);
+ if (clp->verbose > 5) {
+ for (k = 0; k < half_num; ++k) {
+ pr2serr_lk("AFTER: def_arr[%d]:\n", k);
+ v4hdr_out_lk("normal v4 object", (a_v4p + k), rep->id);
+ // hex2stderr_lk((const uint8_t *)(a_v4p + k), sizeof(*a_v4p),
+ // 1);
+ }
+ }
+ }
+ in_fin_blks = 0;
+ out_fin_blks = 0;
+ num_good = process_mrq_response(rep, ctlop, a_v4p, half_num, in_fin_blks,
+ out_fin_blks);
+ if (clp->verbose > 2)
+ pr2serr_lk("%s: >>>1 num_good=%d, in_q/fin blks=%u/%u; out_q/fin "
+ "blks=%u/%u\n", __func__, num_good, rep->in_mrq_q_blks,
+ in_fin_blks, rep->out_mrq_q_blks, out_fin_blks);
+
+ if (num_good < 0)
+ res = -1;
+ else if (num_good < half_num) {
+ int resid_blks = rep->in_mrq_q_blks - in_fin_blks;
+
+ if (resid_blks > 0)
+ gcoll.in_rem_count += resid_blks;
+ resid_blks = rep->out_mrq_q_blks - out_fin_blks;
+ if (resid_blks > 0)
+ gcoll.out_rem_count += resid_blks;
+
+ return -1;
+ }
+
+ rest = nrq - half_num;
+ if (rest < 1)
+ goto fini;
+ /* fetch remaining */
+ for (k = 0; k < 100000; ++k) {
+ ++num_waiting_calls;
+ res = ioctl(fd, SG_GET_NUM_WAITING, &nwait);
+ if (res < 0) {
+ pr2serr_lk("%s: ioctl(SG_GET_NUM_WAITING)-->%d, errno=%d: %s\n",
+ __func__, res, errno, strerror(errno));
+ return -1;
+ }
+ if (nwait >= rest)
+ break;
+ this_thread::sleep_for(chrono::microseconds{wait_us});
+ }
+ ctlop = &hold_ctlo;
+ ctlop->din_xferp += (half_num * sizeof(struct sg_io_v4));
+ ctlop->din_xfer_len -= (half_num * sizeof(struct sg_io_v4));
+ ctlop->dout_xferp = ctlop->din_xferp;
+ ctlop->dout_xfer_len = ctlop->din_xfer_len;
+ ctlop->flags = (SGV4_FLAG_MULTIPLE_REQS | SGV4_FLAG_IMMED);
+ res = ioctl(fd, SG_IORECEIVE, ctlop);
+ if (res < 0) {
+ err = errno;
+ if (ENODATA != err) {
+ pr2serr_lk("%s: ioctl(SG_IORECEIVE, %s),2-->%d, errno=%d: %s\n",
+ __func__, sg_flags_str(ctlop->flags, b_len, b), res,
+ err, strerror(err));
+ return -1;
+ }
+ half_num = 0;
+ } else
+ half_num = ctlop->info;
+ if (clp->verbose > 4) {
+ pr2serr_lk("%s: Controlling object output by ioctl(SG_IORECEIVE),2: "
+ "num_received=%d\n", __func__, half_num);
+ if (clp->verbose > 5)
+ hex2stderr_lk((const uint8_t *)ctlop, sizeof(*ctlop), 1);
+ v4hdr_out_lk("Controlling object after", ctlop, rep->id);
+ if (clp->verbose > 5) {
+ for (k = 0; k < half_num; ++k) {
+ pr2serr_lk("AFTER: def_arr[%d]:\n", k);
+ v4hdr_out_lk("normal v4 object", (a_v4p + k), rep->id);
+ // hex2stderr_lk((const uint8_t *)(a_v4p + k), sizeof(*a_v4p),
+ // 1);
+ }
+ }
+ }
+ in_fin_blks = 0;
+ out_fin_blks = 0;
+ num_good = process_mrq_response(rep, ctlop, a_v4p, half_num, in_fin_blks,
+ out_fin_blks);
+ if (clp->verbose > 2)
+ pr2serr_lk("%s: >>>2 num_good=%d, in_q/fin blks=%u/%u; out_q/fin "
+ "blks=%u/%u\n", __func__, num_good, rep->in_mrq_q_blks,
+ in_fin_blks, rep->out_mrq_q_blks, out_fin_blks);
+
+ if (num_good < 0)
+ res = -1;
+ else if (num_good < half_num) {
+ int resid_blks = rep->in_mrq_q_blks - in_fin_blks;
+
+ if (resid_blks > 0)
+ gcoll.in_rem_count += resid_blks;
+ resid_blks = rep->out_mrq_q_blks - out_fin_blks;
+ if (resid_blks > 0)
+ gcoll.out_rem_count += resid_blks;
+
+ res = -1;
+ }
+
+fini:
+ return res;
+}
+
+/* Split def_arr into fd_def_arr and o_fd_arr based on whether each element's
+ * flags field has SGV4_FLAG_DO_ON_OTHER set. If it is set place in
+ * o_fd_def_arr and mask out SGV4_DO_ON_OTHER. Returns number of elements
+ * in o_fd_def_arr. */
+static int
+split_def_arr(const mrq_arr_t & def_arr, mrq_arr_t & fd_def_arr,
+ mrq_arr_t & o_fd_def_arr)
+{
+ int nrq, k;
+ int res = 0;
+ const struct sg_io_v4 * a_v4p;
+
+ a_v4p = def_arr.first.data();
+ nrq = def_arr.first.size();
+
+ for (k = 0; k < nrq; ++k) {
+ int flags;
+ const struct sg_io_v4 * h4p = a_v4p + k;
+
+ flags = h4p->flags;
+ if (flags & SGV4_FLAG_DO_ON_OTHER) {
+ o_fd_def_arr.first.push_back(def_arr.first[k]);
+ o_fd_def_arr.second.push_back(def_arr.second[k]);
+ flags &= ~SGV4_FLAG_DO_ON_OTHER; /* mask out DO_ON_OTHER */
+ o_fd_def_arr.first[res].flags = flags;
+ ++res;
+ } else {
+ fd_def_arr.first.push_back(def_arr.first[k]);
+ fd_def_arr.second.push_back(def_arr.second[k]);
+ }
+ }
+ return res;
+}
+
+/* This function sets up a multiple request (mrq) transaction and sends it
+ * to the pass-through. Returns 0 on success, 1 if ENOMEM error else -1 for
+ * other errors. */
+static int
+sgh_do_deferred_mrq(Rq_elem * rep, mrq_arr_t & def_arr)
+{
+ bool launch_mrq_abort = false;
+ int nrq, k, res, fd, mrq_pack_id, status, id, num_good, b_len;
+ uint32_t in_fin_blks, out_fin_blks;
+ const int max_cdb_sz = 16;
+ struct sg_io_v4 * a_v4p;
+ struct sg_io_v4 ctl_v4 {};
+ uint8_t * cmd_ap = NULL;
+ struct global_collection * clp = rep->clp;
+ const char * iosub_str = "iosub_str";
+ char b[80];
+
+ id = rep->id;
+ b_len = sizeof(b);
+ ctl_v4.guard = 'Q';
+ a_v4p = def_arr.first.data();
+ nrq = def_arr.first.size();
+ if (nrq < 1) {
+ pr2serr_lk("[%d] %s: strange nrq=0, nothing to do\n", id, __func__);
+ return 0;
+ }
+ if (clp->mrq_cmds) {
+ cmd_ap = (uint8_t *)calloc(nrq, max_cdb_sz);
+ if (NULL == cmd_ap) {
+ pr2serr_lk("[%d] %s: no memory for calloc(%d * 16)\n", id,
+ __func__, nrq);
+ return 1;
+ }
+ }
+ for (k = 0; k < nrq; ++k) {
+ struct sg_io_v4 * h4p = a_v4p + k;
+ uint8_t *cmdp = &def_arr.second[k].front();
+
+ if (clp->mrq_cmds) {
+ memcpy(cmd_ap + (k * max_cdb_sz), cmdp, h4p->request_len);
+ h4p->request = 0;
+ } else
+ h4p->request = (uint64_t)cmdp;
+ if (clp->verbose > 5) {
+ pr2serr_lk("%s%s[%d] def_arr[%d]", ((0 == k) ? __func__ : ""),
+ ((0 == k) ? ": " : ""), id, k);
+ if (h4p->din_xferp)
+ pr2serr_lk(" [din=0x%p]:\n", (void *)h4p->din_xferp);
+ else if (h4p->dout_xferp)
+ pr2serr_lk(" [dout=0x%p]:\n", (void *)h4p->dout_xferp);
+ else
+ pr2serr_lk(":\n");
+ hex2stderr_lk((const uint8_t *)h4p, sizeof(*h4p), 1);
+ }
+ }
+ if (rep->both_sg || rep->same_sg)
+ fd = rep->infd; /* assume share to rep->outfd */
+ else if (rep->only_in_sg)
+ fd = rep->infd;
+ else if (rep->only_out_sg)
+ fd = rep->outfd;
+ else {
+ pr2serr_lk("[%d] %s: why am I here? No sg devices\n", id, __func__);
+ res = -1;
+ goto fini;
+ }
+ res = 0;
+ if (clp->mrq_cmds) {
+ ctl_v4.request_len = nrq * max_cdb_sz;
+ ctl_v4.request = (uint64_t)cmd_ap;
+ }
+ ctl_v4.flags = SGV4_FLAG_MULTIPLE_REQS;
+ if (! clp->mrq_async) {
+ ctl_v4.flags |= SGV4_FLAG_STOP_IF;
+ if (clp->in_flags.mrq_svb || clp->out_flags.mrq_svb)
+ ctl_v4.flags |= SGV4_FLAG_SHARE;
+ }
+ ctl_v4.dout_xferp = (uint64_t)a_v4p; /* request array */
+ ctl_v4.dout_xfer_len = nrq * sizeof(*a_v4p);
+ ctl_v4.din_xferp = (uint64_t)a_v4p; /* response array */
+ ctl_v4.din_xfer_len = nrq * sizeof(*a_v4p);
+ mrq_pack_id = atomic_fetch_add(&mono_mrq_id, 1);
+ if ((clp->m_aen > 0) && (MONO_MRQ_ID_INIT != mrq_pack_id) &&
+ (0 == ((mrq_pack_id - MONO_MRQ_ID_INIT) % clp->m_aen))) {
+ launch_mrq_abort = true;
+ if (clp->verbose > 2)
+ pr2serr_lk("[%d] %s: Decide to launch MRQ abort thread, "
+ "mrq_id=%d\n", id, __func__, mrq_pack_id);
+ memset(&rep->mai, 0, sizeof(rep->mai));
+ rep->mai.from_tid = id;
+ rep->mai.mrq_id = mrq_pack_id;
+ rep->mai.fd = fd;
+ rep->mai.debug = clp->verbose;
+
+ status = pthread_create(&rep->mrq_abort_thread_id, NULL,
+ mrq_abort_thread, (void *)&rep->mai);
+ if (0 != status) err_exit(status, "pthread_create, sig...");
+ }
+ ctl_v4.request_extra = launch_mrq_abort ? mrq_pack_id : 0;
+ rep->mrq_id = mrq_pack_id;
+ if (clp->verbose && rep->both_sg && clp->mrq_async)
+ iosub_str = "SG_IOSUBMIT(variable)";
+ if (clp->verbose > 4) {
+ pr2serr_lk("%s: Controlling object _before_ ioctl(%s):\n",
+ __func__, iosub_str);
+ if (clp->verbose > 5)
+ hex2stderr_lk((const uint8_t *)&ctl_v4, sizeof(ctl_v4), 1);
+ v4hdr_out_lk("Controlling object before", &ctl_v4, id);
+ }
+ if (clp->mrq_async && (! rep->both_sg)) {
+ /* do 'submit non-blocking' or 'full non-blocking' mrq */
+ mrq_arr_t fd_def_arr;
+ mrq_arr_t o_fd_def_arr;
+
+ /* need to deconstruct def_arr[] into two separate lists, one for
+ * the source, the other for the destination. */
+ int o_num_fd = split_def_arr(def_arr, fd_def_arr, o_fd_def_arr);
+ int num_fd = fd_def_arr.first.size();
+ if (num_fd > 0) {
+ struct sg_io_v4 fd_ctl = ctl_v4;
+ struct sg_io_v4 * aa_v4p = fd_def_arr.first.data();
+
+ for (k = 0; k < num_fd; ++k) {
+ struct sg_io_v4 * h4p = aa_v4p + k;
+ uint8_t *cmdp = &fd_def_arr.second[k].front();
+
+ if (clp->mrq_cmds) {
+ memcpy(cmd_ap + (k * max_cdb_sz), cmdp, h4p->request_len);
+ h4p->request = 0;
+ } else
+ h4p->request = (uint64_t)cmdp;
+ if (clp->verbose > 5) {
+ pr2serr_lk("[%d] df_def_arr[%d]:\n", id, k);
+ hex2stderr_lk((const uint8_t *)(aa_v4p + k),
+ sizeof(*aa_v4p), 1);
+ }
+ }
+ fd_ctl.dout_xferp = (uint64_t)aa_v4p; /* request array */
+ fd_ctl.dout_xfer_len = num_fd * sizeof(*aa_v4p);
+ fd_ctl.din_xferp = (uint64_t)aa_v4p; /* response array */
+ fd_ctl.din_xfer_len = num_fd * sizeof(*aa_v4p);
+ fd_ctl.request_extra = launch_mrq_abort ? mrq_pack_id : 0;
+ /* this is the source side mrq command */
+ res = sgh_do_async_mrq(rep, fd_def_arr, fd, &fd_ctl, num_fd);
+ rep->in_mrq_q_blks = 0;
+ if (res)
+ goto fini;
+ }
+ if (o_num_fd > 0) {
+ struct sg_io_v4 o_fd_ctl = ctl_v4;
+ struct sg_io_v4 * aa_v4p = o_fd_def_arr.first.data();
+
+ for (k = 0; k < o_num_fd; ++k) {
+ struct sg_io_v4 * h4p = aa_v4p + k;
+ uint8_t *cmdp = &o_fd_def_arr.second[k].front();
+
+ if (clp->mrq_cmds) {
+ memcpy(cmd_ap + (k * max_cdb_sz), cmdp, h4p->request_len);
+ h4p->request = 0;
+ } else
+ h4p->request = (uint64_t)cmdp;
+ if (clp->verbose > 5) {
+ pr2serr_lk("[%d] o_fd_def_arr[%d]:\n", id, k);
+ hex2stderr_lk((const uint8_t *)(aa_v4p + k),
+ sizeof(*aa_v4p), 1);
+ }
+ }
+ o_fd_ctl.dout_xferp = (uint64_t)aa_v4p; /* request array */
+ o_fd_ctl.dout_xfer_len = o_num_fd * sizeof(*aa_v4p);
+ o_fd_ctl.din_xferp = (uint64_t)aa_v4p; /* response array */
+ o_fd_ctl.din_xfer_len = o_num_fd * sizeof(*aa_v4p);
+ o_fd_ctl.request_extra = launch_mrq_abort ? mrq_pack_id : 0;
+ /* this is the destination side mrq command */
+ res = sgh_do_async_mrq(rep, o_fd_def_arr, rep->outfd, &o_fd_ctl,
+ o_num_fd);
+ rep->out_mrq_q_blks = 0;
+ }
+ goto fini;
+ }
+
+try_again:
+ if (clp->unbalanced_mrq) {
+ iosub_str = "SG_IOSUBMIT(variable_blocking)";
+ if (!after1 && (clp->verbose > 1)) {
+ after1 = true;
+ pr2serr_lk("%s: unbalanced %s\n", __func__, mrq_vb_s);
+ }
+ res = ioctl(fd, SG_IOSUBMIT, &ctl_v4);
+ } else {
+ if (clp->mrq_async) {
+ iosub_str = "SG_IOSUBMIT(variable_blocking)";
+ if (!after1 && (clp->verbose > 1)) {
+ after1 = true;
+ pr2serr_lk("%s: %s\n", __func__, mrq_vb_s);
+ }
+ res = ioctl(fd, SG_IOSUBMIT, &ctl_v4);
+ } else if (clp->in_flags.mrq_svb || clp->out_flags.mrq_svb) {
+ iosub_str = "SG_IOSUBMIT(shared_variable_blocking)";
+ if (!after1 && (clp->verbose > 1)) {
+ after1 = true;
+ pr2serr_lk("%s: %s\n", __func__, mrq_svb_s);
+ }
+ res = ioctl(fd, SG_IOSUBMIT, &ctl_v4);
+ } else {
+ iosub_str = "SG_IO(ordered_blocking)";
+ if (!after1 && (clp->verbose > 1)) {
+ after1 = true;
+ pr2serr_lk("%s: %s\n", __func__, mrq_blk_s);
+ }
+ res = ioctl(fd, SG_IO, &ctl_v4);
+ }
+ }
+ if (res < 0) {
+ int err = errno;
+
+ if (E2BIG == err)
+ sg_take_snap(fd, id, true);
+ else if (EBUSY == err) {
+ ++num_ebusy;
+ std::this_thread::yield();/* allow another thread to progress */
+ goto try_again;
+ }
+ pr2serr_lk("%s: ioctl(%s, %s)-->%d, errno=%d: %s\n",
+ __func__, iosub_str, sg_flags_str(ctl_v4.flags, b_len, b),
+ res, err, strerror(err));
+ res = -1;
+ goto fini;
+ }
+ if (clp->verbose && vb_first_time.load()) {
+ pr2serr_lk("First controlling object output by ioctl(%s), flags: "
+ "%s\n", iosub_str, sg_flags_str(ctl_v4.flags, b_len, b));
+ vb_first_time.store(false);
+ } else if (clp->verbose > 4)
+ pr2serr_lk("%s: Controlling object output by ioctl(%s):\n",
+ __func__, iosub_str);
+ if (clp->verbose > 4) {
+ if (clp->verbose > 5)
+ hex2stderr_lk((const uint8_t *)&ctl_v4, sizeof(ctl_v4), 1);
+ v4hdr_out_lk("Controlling object after", &ctl_v4, id);
+ if (clp->verbose > 5) {
+ for (k = 0; k < nrq; ++k) {
+ pr2serr_lk("AFTER: def_arr[%d]:\n", k);
+ v4hdr_out_lk("normal v4 object", (a_v4p + k), id);
+ // hex2stderr_lk((const uint8_t *)(a_v4p + k), sizeof(*a_v4p),
+ // 1);
+ }
+ }
+ }
+ in_fin_blks = 0;
+ out_fin_blks = 0;
+ num_good = process_mrq_response(rep, &ctl_v4, a_v4p, nrq, in_fin_blks,
+ out_fin_blks);
+ if (clp->verbose > 2)
+ pr2serr_lk("%s: >>> num_good=%d, in_q/fin blks=%u/%u; out_q/fin "
+ "blks=%u/%u\n", __func__, num_good, rep->in_mrq_q_blks,
+ in_fin_blks, rep->out_mrq_q_blks, out_fin_blks);
+
+ if (num_good < 0)
+ res = -1;
+ else if (num_good < nrq) {
+ int resid_blks = rep->in_mrq_q_blks - in_fin_blks;
+
+ if (resid_blks > 0)
+ gcoll.in_rem_count += resid_blks;
+ resid_blks = rep->out_mrq_q_blks - out_fin_blks;
+ if (resid_blks > 0)
+ gcoll.out_rem_count += resid_blks;
+
+ res = -1;
+ }
+ rep->in_mrq_q_blks = 0;
+ rep->out_mrq_q_blks = 0;
+fini:
+ def_arr.first.clear();
+ def_arr.second.clear();
+ if (cmd_ap)
+ free(cmd_ap);
+ if (launch_mrq_abort) {
+ if (clp->verbose > 1)
+ pr2serr_lk("[%d] %s: About to join MRQ abort thread, "
+ "mrq_id=%d\n", id, __func__, mrq_pack_id);
+
+ void * vp; /* not used */
+ status = pthread_join(rep->mrq_abort_thread_id, &vp);
+ if (0 != status) err_exit(status, "pthread_join");
+ }
+ return res;
+}
+
+/* Returns 0 on success, 1 if ENOMEM error else -1 for other errors. */
+static int
+sg_start_io(Rq_elem * rep, mrq_arr_t & def_arr, int & pack_id,
+ struct sg_io_extra *xtrp)
+{
+ struct global_collection * clp = rep->clp;
+ bool wr = rep->wr;
+ bool fua = wr ? clp->out_flags.fua : clp->in_flags.fua;
+ bool dpo = wr ? clp->out_flags.dpo : clp->in_flags.dpo;
+ bool dio = wr ? clp->out_flags.dio : clp->in_flags.dio;
+ bool mmap = wr ? clp->out_flags.mmap : clp->in_flags.mmap;
+ bool noxfer = wr ? clp->out_flags.noxfer : clp->in_flags.noxfer;
+ bool v4 = wr ? clp->out_flags.v4 : clp->in_flags.v4;
+ bool qhead = wr ? clp->out_flags.qhead : clp->in_flags.qhead;
+ bool qtail = wr ? clp->out_flags.qtail : clp->in_flags.qtail;
+ bool polled = wr ? clp->out_flags.polled : clp->in_flags.polled;
+ bool mout_if = wr ? clp->out_flags.mout_if : clp->in_flags.mout_if;
+ bool prefetch = xtrp ? xtrp->prefetch : false;
+ bool is_wr2 = xtrp ? xtrp->is_wr2 : false;
+ int cdbsz = wr ? clp->cdbsz_out : clp->cdbsz_in;
+ int flags = 0;
+ int res, err, fd, b_len, nblks, blk_off;
+ int64_t blk = wr ? rep->oblk : rep->iblk;
+ struct sg_io_hdr * hp = &rep->io_hdr;
+ struct sg_io_v4 * h4p = &rep->io_hdr4[xtrp ? xtrp->hpv4_ind : 0];
+ const char * cp = "";
+ const char * crwp;
+ char b[80];
+
+ b_len = sizeof(b);
+ if (wr) {
+ fd = is_wr2 ? rep->out2fd : rep->outfd;
+ if (clp->verify) {
+ crwp = is_wr2 ? "verifying2" : "verifying";
+ if (prefetch)
+ crwp = is_wr2 ? "prefetch2" : "prefetch";
+ } else
+ crwp = is_wr2 ? "writing2" : "writing";
+ } else {
+ fd = rep->infd;
+ crwp = "reading";
+ }
+ if (qhead)
+ qtail = false; /* qhead takes precedence */
+
+ if (v4 && xtrp && xtrp->dout_is_split) {
+ res = sg_build_scsi_cdb(rep->cmd, cdbsz, xtrp->blks,
+ blk + (unsigned int)xtrp->blk_offset,
+ clp->verify, true, fua, dpo);
+ } else
+ res = sg_build_scsi_cdb(rep->cmd, cdbsz, rep->num_blks, blk,
+ wr ? clp->verify : false, wr, fua, dpo);
+ if (res) {
+ pr2serr_lk("%sbad cdb build, start_blk=%" PRId64 ", blocks=%d\n",
+ my_name, blk, rep->num_blks);
+ return -1;
+ }
+ if (prefetch) {
+ if (cdbsz == 10)
+ rep->cmd[0] = SGP_PRE_FETCH10;
+ else if (cdbsz == 16)
+ rep->cmd[0] = SGP_PRE_FETCH16;
+ else {
+ pr2serr_lk("%sbad PRE-FETCH build, start_blk=%" PRId64 ", "
+ "blocks=%d\n", my_name, blk, rep->num_blks);
+ return -1;
+ }
+ rep->cmd[1] = 0x2; /* set IMMED (no fua or dpo) */
+ }
+ if (mmap && (clp->noshare || (rep->outregfd >= 0)))
+ flags |= SG_FLAG_MMAP_IO;
+ if (noxfer)
+ flags |= SG_FLAG_NO_DXFER;
+ if (dio)
+ flags |= SG_FLAG_DIRECT_IO;
+ if (polled)
+ flags |= SGV4_FLAG_POLLED;
+ if (qhead)
+ flags |= SG_FLAG_Q_AT_HEAD;
+ if (qtail)
+ flags |= SG_FLAG_Q_AT_TAIL;
+ if (mout_if)
+ flags |= SGV4_FLAG_META_OUT_IF;
+ if (rep->has_share) {
+ flags |= SGV4_FLAG_SHARE;
+ if (wr)
+ flags |= SGV4_FLAG_NO_DXFER;
+ else if (rep->outregfd < 0)
+ flags |= SGV4_FLAG_NO_DXFER;
+
+ cp = (wr ? " write_side active" : " read_side active");
+ } else
+ cp = (wr ? " write-side not sharing" : " read_side not sharing");
+ if (rep->both_sg) {
+ if (wr)
+ pack_id = rep->rd_p_id + 1;
+ else {
+ pack_id = 2 * atomic_fetch_add(&mono_pack_id, 1);
+ rep->rd_p_id = pack_id;
+ }
+ } else
+ pack_id = atomic_fetch_add(&mono_pack_id, 1); /* fetch before */
+ rep->rq_id = pack_id;
+ nblks = rep->num_blks;
+ blk_off = 0;
+ if (clp->verbose && 0 == clp->nmrqs && vb_first_time.load()) {
+ vb_first_time.store(false);
+ pr2serr("First normal IO: %s, flags: %s\n", cp,
+ sg_flags_str(flags, b_len, b));
+ }
+ if (v4) {
+ memset(h4p, 0, sizeof(struct sg_io_v4));
+ if (clp->nmrqs > 0) {
+ if (rep->both_sg && (rep->outfd == fd))
+ flags |= SGV4_FLAG_DO_ON_OTHER;
+ }
+ if (xtrp && xtrp->dout_is_split && (nblks > 0)) {
+ if (1 == xtrp->hpv4_ind) {
+ flags |= SGV4_FLAG_DOUT_OFFSET;
+ blk_off = xtrp->blk_offset;
+ h4p->spare_in = clp->bs * blk_off;
+ }
+ nblks = xtrp->blks;
+ if ((0 == xtrp->hpv4_ind) && (nblks < rep->num_blks))
+ flags |= SGV4_FLAG_KEEP_SHARE;
+ }
+ if (clp->ofile2_given && wr && rep->has_share && ! is_wr2)
+ flags |= SGV4_FLAG_KEEP_SHARE; /* set on first write only */
+ else if (clp->fail_mask & 1)
+ flags |= SGV4_FLAG_KEEP_SHARE; /* troublemaking .... */
+ } else
+ memset(hp, 0, sizeof(struct sg_io_hdr));
+ if (clp->verbose > 3) {
+ bool lock = true;
+ char prefix[128];
+
+ if (4 == clp->verbose) {
+ snprintf(prefix, sizeof(prefix), "tid,rq_id=%d,%d: ", rep->id,
+ pack_id);
+ lock = false;
+ } else {
+ prefix[0] = '\0';
+ pr2serr_lk("%s tid,rq_id=%d,%d: SCSI %s%s %s, blk=%" PRId64
+ " num_blks=%d\n", __func__, rep->id, pack_id, crwp, cp,
+ sg_flags_str(flags, b_len, b), blk + blk_off, nblks);
+ }
+ lk_print_command_len(prefix, rep->cmd, cdbsz, lock);
+ }
+ if (v4)
+ goto do_v4; // <<<<<<<<<<<<<<< look further down
+
+ hp->interface_id = 'S';
+ hp->cmd_len = cdbsz;
+ hp->cmdp = rep->cmd;
+ hp->dxferp = get_buffp(rep);
+ hp->dxfer_len = clp->bs * rep->num_blks;
+ if (!wr)
+ hp->dxfer_direction = SG_DXFER_FROM_DEV;
+ else if (prefetch) {
+ hp->dxfer_direction = SG_DXFER_NONE;
+ hp->dxfer_len = 0;
+ hp->dxferp = NULL;
+ } else
+ hp->dxfer_direction = SG_DXFER_TO_DEV;
+ hp->mx_sb_len = sizeof(rep->sb);
+ hp->sbp = rep->sb;
+ hp->timeout = clp->cmd_timeout;
+ hp->usr_ptr = rep;
+ hp->pack_id = pack_id;
+ hp->flags = flags;
+
+ while (((res = write(fd, hp, sizeof(struct sg_io_hdr))) < 0) &&
+ ((EINTR == errno) || (EAGAIN == errno) || (EBUSY == errno))) {
+ if (EAGAIN == errno) {
+ ++num_start_eagain;
+#ifdef SGH_DD_SNAP_DEV
+ if (0 == (num_ebusy % 1000))
+ sg_take_snap(fd, rep->id, (clp->verbose > 2));
+#endif
+ } else if (EBUSY == errno) {
+ ++num_ebusy;
+#ifdef SGH_DD_SNAP_DEV
+ if (0 == (num_ebusy % 1000))
+ sg_take_snap(fd, rep->id, (clp->verbose > 2));
+#endif
+ }
+ std::this_thread::yield();/* another thread may be able to progress */
+ }
+ err = errno;
+ if (res < 0) {
+ if (ENOMEM == err)
+ return 1;
+ pr2serr_lk("%s tid=%d: %s %s write(2) failed: %s\n", __func__,
+ rep->id, cp, sg_flags_str(hp->flags, b_len, b),
+ strerror(err));
+ return -1;
+ }
+ return 0;
+
+do_v4:
+ h4p->guard = 'Q';
+ h4p->request_len = cdbsz;
+ h4p->request = (uint64_t)rep->cmd;
+ if (wr) {
+ if (prefetch) {
+ h4p->dout_xfer_len = 0; // din_xfer_len is also 0
+ h4p->dout_xferp = 0;
+ } else {
+ h4p->dout_xfer_len = clp->bs * nblks;
+ h4p->dout_xferp = (uint64_t)get_buffp(rep);
+ }
+ } else if (nblks > 0) {
+ h4p->din_xfer_len = clp->bs * nblks;
+ h4p->din_xferp = (uint64_t)get_buffp(rep);
+ }
+ h4p->max_response_len = sizeof(rep->sb);
+ h4p->response = (uint64_t)rep->sb;
+ h4p->timeout = clp->cmd_timeout;
+ h4p->usr_ptr = (uint64_t)rep;
+ h4p->request_extra = pack_id; /* this is the pack_id */
+ h4p->flags = flags;
+ if (clp->nmrqs > 0) {
+ big_cdb cdb_arr;
+ uint8_t * cmdp = &(cdb_arr[0]);
+
+ if (wr)
+ rep->out_mrq_q_blks += nblks;
+ else
+ rep->in_mrq_q_blks += nblks;
+ memcpy(cmdp, rep->cmd, cdbsz);
+ def_arr.first.push_back(*h4p);
+ def_arr.second.push_back(cdb_arr);
+ res = 0;
+ if ((int)def_arr.first.size() >= clp->nmrqs) {
+ res = sgh_do_deferred_mrq(rep, def_arr);
+ if (res)
+ pr2serr_lk("%s tid=%d: sgh_do_deferred_mrq failed\n",
+ __func__, rep->id);
+ }
+ return res;
+ }
+ while (((res = ioctl(fd, SG_IOSUBMIT, h4p)) < 0) &&
+ ((EINTR == errno) || (EAGAIN == errno) || (EBUSY == errno))) {
+ if (EAGAIN == errno) {
+ ++num_start_eagain;
+#ifdef SGH_DD_SNAP_DEV
+ if (0 == (num_ebusy % 1000))
+ sg_take_snap(fd, rep->id, (clp->verbose > 2));
+#endif
+ } else if (EBUSY == errno) {
+ ++num_ebusy;
+#ifdef SGH_DD_SNAP_DEV
+ if (0 == (num_ebusy % 1000))
+ sg_take_snap(fd, rep->id, (clp->verbose > 2));
+#endif
+ }
+ std::this_thread::yield();/* another thread may be able to progress */
+ }
+ err = errno;
+ if (res < 0) {
+ if (ENOMEM == err)
+ return 1;
+ if (E2BIG == err)
+ sg_take_snap(fd, rep->id, true);
+ pr2serr_lk("%s tid=%d: %s %s ioctl(2) failed: %s\n", __func__,
+ rep->id, cp, sg_flags_str(h4p->flags, b_len, b),
+ strerror(err));
+ // v4hdr_out_lk("leadin", h4p, rep->id);
+ return -1;
+ }
+ if ((clp->aen > 0) && (rep->rep_count > 0)) {
+ if (0 == (rep->rq_id % clp->aen)) {
+ struct timespec tspec = {0, 4000 /* 4 usecs */};
+
+ nanosleep(&tspec, NULL);
+#if 0
+ struct pollfd a_poll;
+
+ a_poll.fd = fd;
+ a_poll.events = POLL_IN;
+ a_poll.revents = 0;
+ res = poll(&a_poll, 1 /* element */, 1 /* millisecond */);
+ if (res < 0)
+ pr2serr_lk("%s: poll() failed: %s [%d]\n",
+ __func__, safe_strerror(errno), errno);
+ else if (0 == res) { /* timeout, cmd still inflight, so abort */
+ }
+#endif
+ ++num_abort_req;
+ res = ioctl(fd, SG_IOABORT, h4p);
+ if (res < 0) {
+ err = errno;
+ if (ENODATA == err) {
+ if (clp->verbose > 2)
+ pr2serr_lk("%s: ioctl(SG_IOABORT) no match on "
+ "pack_id=%d\n", __func__, pack_id);
+ } else
+ pr2serr_lk("%s: ioctl(SG_IOABORT) failed: %s [%d]\n",
+ __func__, safe_strerror(err), err);
+ } else {
+ ++num_abort_req_success;
+ if (clp->verbose > 2)
+ pr2serr_lk("%s: sent ioctl(SG_IOABORT) on rq_id=%d, "
+ "success\n", __func__, pack_id);
+ }
+ } /* else got response, too late for timeout, so skip */
+ }
+ return 0;
+}
+
+/* 0 -> successful, SG_LIB_CAT_UNIT_ATTENTION or SG_LIB_CAT_ABORTED_COMMAND
+ -> try again, SG_LIB_CAT_NOT_READY, SG_LIB_CAT_MEDIUM_HARD,
+ -1 other errors */
+static int
+sg_finish_io(bool wr, Rq_elem * rep, int pack_id, struct sg_io_extra *xtrp)
+{
+ struct global_collection * clp = rep->clp;
+ bool v4 = wr ? clp->out_flags.v4 : clp->in_flags.v4;
+ bool mout_if = wr ? clp->out_flags.mout_if : clp->in_flags.mout_if;
+ bool is_wr2 = xtrp ? xtrp->is_wr2 : false;
+ bool prefetch = xtrp ? xtrp->prefetch : false;
+ int res, fd;
+ int64_t blk = wr ? rep->oblk : rep->iblk;
+ struct sg_io_hdr io_hdr;
+ struct sg_io_hdr * hp;
+ struct sg_io_v4 * h4p;
+ const char *cp;
+
+ if (wr) {
+ fd = is_wr2 ? rep->out2fd : rep->outfd;
+ cp = is_wr2 ? "writing2" : "writing";
+ if (clp->verify) {
+ cp = is_wr2 ? "verifying2" : "verifying";
+ if (prefetch)
+ cp = is_wr2 ? "prefetch2" : "prefetch";
+ }
+ } else {
+ fd = rep->infd;
+ cp = "reading";
+ }
+ if (v4)
+ goto do_v4;
+ memset(&io_hdr, 0 , sizeof(struct sg_io_hdr));
+ /* FORCE_PACK_ID active set only read packet with matching pack_id */
+ io_hdr.interface_id = 'S';
+ io_hdr.dxfer_direction = wr ? SG_DXFER_TO_DEV : SG_DXFER_FROM_DEV;
+ io_hdr.pack_id = pack_id;
+
+ while (((res = read(fd, &io_hdr, sizeof(struct sg_io_hdr))) < 0) &&
+ ((EINTR == errno) || (EAGAIN == errno) || (EBUSY == errno))) {
+ if (EAGAIN == errno) {
+ ++num_fin_eagain;
+#ifdef SGH_DD_SNAP_DEV
+ if (0 == (num_ebusy % 1000))
+ sg_take_snap(fd, rep->id, (clp->verbose > 2));
+#endif
+ } else if (EBUSY == errno) {
+ ++num_ebusy;
+#ifdef SGH_DD_SNAP_DEV
+ if (0 == (num_ebusy % 1000))
+ sg_take_snap(fd, rep->id, (clp->verbose > 2));
+#endif
+ }
+ std::this_thread::yield();/* another thread may be able to progress */
+ }
+ if (res < 0) {
+ perror("finishing io [read(2)] on sg device, error");
+ return -1;
+ }
+ if (rep != (Rq_elem *)io_hdr.usr_ptr)
+ err_exit(0, "sg_finish_io: bad usr_ptr, request-response mismatch\n");
+ memcpy(&rep->io_hdr, &io_hdr, sizeof(struct sg_io_hdr));
+ hp = &rep->io_hdr;
+
+ res = sg_err_category3(hp);
+ switch (res) {
+ case SG_LIB_CAT_CLEAN:
+ case SG_LIB_CAT_CONDITION_MET:
+ break;
+ case SG_LIB_CAT_RECOVERED:
+ lk_chk_n_print3(cp, hp, false);
+ break;
+ case SG_LIB_CAT_ABORTED_COMMAND:
+ case SG_LIB_CAT_UNIT_ATTENTION:
+ if (clp->verbose > 3)
+ lk_chk_n_print3(cp, hp, false);
+ return res;
+ case SG_LIB_CAT_MISCOMPARE:
+ ++num_miscompare;
+ // fall through
+ case SG_LIB_CAT_NOT_READY:
+ default:
+ {
+ char ebuff[EBUFF_SZ];
+
+ snprintf(ebuff, EBUFF_SZ, "%s blk=%" PRId64, cp, blk);
+ lk_chk_n_print3(ebuff, hp, clp->verbose > 1);
+ return res;
+ }
+ }
+ if ((wr ? clp->out_flags.dio : clp->in_flags.dio) &&
+ (! (hp->info & SG_INFO_DIRECT_IO_MASK)))
+ rep->dio_incomplete_count = 1; /* count dios done as indirect IO */
+ else
+ rep->dio_incomplete_count = 0;
+ rep->resid = hp->resid;
+ if (clp->verbose > 3)
+ pr2serr_lk("%s: tid=%d: completed %s\n", __func__, rep->id, cp);
+ return 0;
+
+do_v4:
+ if (clp->nmrqs > 0) {
+ rep->resid = 0;
+ return 0;
+ }
+ h4p = &rep->io_hdr4[xtrp ? xtrp->hpv4_ind : 0];
+ h4p->request_extra = pack_id;
+ if (mout_if) {
+ h4p->info = 0;
+ h4p->din_resid = 0;
+ }
+ while (((res = ioctl(fd, SG_IORECEIVE, h4p)) < 0) &&
+ ((EINTR == errno) || (EAGAIN == errno) || (EBUSY == errno))) {
+ if (EAGAIN == errno) {
+ ++num_fin_eagain;
+#ifdef SGH_DD_SNAP_DEV
+ if (0 == (num_ebusy % 1000))
+ sg_take_snap(fd, rep->id, (clp->verbose > 2));
+#endif
+ } else if (EBUSY == errno) {
+ ++num_ebusy;
+#ifdef SGH_DD_SNAP_DEV
+ if (0 == (num_ebusy % 1000))
+ sg_take_snap(fd, rep->id, (clp->verbose > 2));
+#endif
+ }
+ std::this_thread::yield();/* another thread may be able to progress */
+ }
+ if (res < 0) {
+ perror("finishing io [SG_IORECEIVE] on sg device, error");
+ return -1;
+ }
+ if (mout_if && (0 == h4p->info) && (0 == h4p->din_resid))
+ goto all_good;
+ if (rep != (Rq_elem *)h4p->usr_ptr)
+ err_exit(0, "sg_finish_io: bad usr_ptr, request-response mismatch\n");
+ res = sg_err_category_new(h4p->device_status, h4p->transport_status,
+ h4p->driver_status,
+ (const uint8_t *)h4p->response,
+ h4p->response_len);
+ switch (res) {
+ case SG_LIB_CAT_CLEAN:
+ case SG_LIB_CAT_CONDITION_MET:
+ break;
+ case SG_LIB_CAT_RECOVERED:
+ lk_chk_n_print4(cp, h4p, false);
+ break;
+ case SG_LIB_CAT_ABORTED_COMMAND:
+ case SG_LIB_CAT_UNIT_ATTENTION:
+ if (clp->verbose > 3)
+ lk_chk_n_print4(cp, h4p, false);
+ return res;
+ case SG_LIB_CAT_MISCOMPARE:
+ ++num_miscompare;
+ // fall through
+ case SG_LIB_CAT_NOT_READY:
+ default:
+ {
+ char ebuff[EBUFF_SZ];
+
+ snprintf(ebuff, EBUFF_SZ, "%s rq_id=%d, blk=%" PRId64, cp,
+ pack_id, blk);
+ lk_chk_n_print4(ebuff, h4p, clp->verbose > 1);
+ if ((clp->verbose > 4) && h4p->info)
+ pr2serr_lk(" info=0x%x sg_info_check=%d direct=%d "
+ "detaching=%d aborted=%d\n", h4p->info,
+ !!(h4p->info & SG_INFO_CHECK),
+ !!(h4p->info & SG_INFO_DIRECT_IO),
+ !!(h4p->info & SG_INFO_DEVICE_DETACHING),
+ !!(h4p->info & SG_INFO_ABORTED));
+ return res;
+ }
+ }
+ if ((wr ? clp->out_flags.dio : clp->in_flags.dio) &&
+ ! (h4p->info & SG_INFO_DIRECT_IO))
+ rep->dio_incomplete_count = 1; /* count dios done as indirect IO */
+ else
+ rep->dio_incomplete_count = 0;
+ rep->resid = h4p->din_resid;
+ if (clp->verbose > 4) {
+ pr2serr_lk("%s: tid,rq_id=%d,%d: completed %s\n", __func__, rep->id,
+ pack_id, cp);
+ if ((clp->verbose > 4) && h4p->info)
+ pr2serr_lk(" info=0x%x sg_info_check=%d direct=%d "
+ "detaching=%d aborted=%d\n", h4p->info,
+ !!(h4p->info & SG_INFO_CHECK),
+ !!(h4p->info & SG_INFO_DIRECT_IO),
+ !!(h4p->info & SG_INFO_DEVICE_DETACHING),
+ !!(h4p->info & SG_INFO_ABORTED));
+ }
+all_good:
+ return 0;
+}
+
+/* Returns reserved_buffer_size/mmap_size if success, else 0 for failure */
+static int
+sg_prepare_resbuf(int fd, bool is_in, struct global_collection *clp,
+ uint8_t **mmpp)
+{
+ static bool done = false;
+ bool def_res = is_in ? clp->in_flags.defres : clp->out_flags.defres;
+ bool no_dur = is_in ? clp->in_flags.no_dur : clp->out_flags.no_dur;
+ bool masync = is_in ? clp->in_flags.masync : clp->out_flags.masync;
+ bool wq_excl = is_in ? clp->in_flags.wq_excl : clp->out_flags.wq_excl;
+ bool skip_thresh = is_in ? clp->in_flags.no_thresh :
+ clp->out_flags.no_thresh;
+ int res, t;
+ int num = 0;
+ uint8_t *mmp;
+ struct sg_extended_info sei {};
+ struct sg_extended_info * seip = &sei;
+
+ res = ioctl(fd, SG_GET_VERSION_NUM, &t);
+ if ((res < 0) || (t < 40000)) {
+ if (ioctl(fd, SG_GET_RESERVED_SIZE, &num) < 0) {
+ perror("SG_GET_RESERVED_SIZE ioctl failed");
+ return 0;
+ }
+ if (! done) {
+ done = true;
+ sg_version_lt_4 = true;
+ pr2serr_lk("%ssg driver prior to 4.0.00, reduced functionality\n",
+ my_name);
+ }
+ goto bypass;
+ }
+ if (! sg_version_ge_40045)
+ goto bypass;
+ if (clp->elem_sz >= 4096) {
+ seip->sei_rd_mask |= SG_SEIM_SGAT_ELEM_SZ;
+ res = ioctl(fd, SG_SET_GET_EXTENDED, seip);
+ if (res < 0)
+ pr2serr_lk("%s%s: SG_SET_GET_EXTENDED(SGAT_ELEM_SZ) rd "
+ "error: %s\n", my_name, __func__, strerror(errno));
+ if (clp->elem_sz != (int)seip->sgat_elem_sz) {
+ memset(seip, 0, sizeof(*seip));
+ seip->sei_wr_mask |= SG_SEIM_SGAT_ELEM_SZ;
+ seip->sgat_elem_sz = clp->elem_sz;
+ res = ioctl(fd, SG_SET_GET_EXTENDED, seip);
+ if (res < 0)
+ pr2serr_lk("%s%s: SG_SET_GET_EXTENDED(SGAT_ELEM_SZ) wr "
+ "error: %s\n", my_name, __func__, strerror(errno));
+ }
+ }
+ if (no_dur || masync || skip_thresh) {
+ memset(seip, 0, sizeof(*seip));
+ seip->sei_wr_mask |= SG_SEIM_CTL_FLAGS;
+ if (no_dur) {
+ seip->ctl_flags_wr_mask |= SG_CTL_FLAGM_NO_DURATION;
+ seip->ctl_flags |= SG_CTL_FLAGM_NO_DURATION;
+ }
+ if (masync) {
+ seip->ctl_flags_wr_mask |= SG_CTL_FLAGM_MORE_ASYNC;
+ seip->ctl_flags |= SG_CTL_FLAGM_MORE_ASYNC;
+ }
+ if (wq_excl) {
+ seip->ctl_flags_wr_mask |= SG_CTL_FLAGM_EXCL_WAITQ;
+ seip->ctl_flags |= SG_CTL_FLAGM_EXCL_WAITQ;
+ }
+ if (skip_thresh) {
+ seip->tot_fd_thresh = 0;
+ sei.sei_wr_mask |= SG_SEIM_TOT_FD_THRESH;
+ }
+ res = ioctl(fd, SG_SET_GET_EXTENDED, seip);
+ if (res < 0)
+ pr2serr_lk("%s%s: SG_SET_GET_EXTENDED(NO_DURATION) error: %s\n",
+ my_name, __func__, strerror(errno));
+ }
+bypass:
+ if (! def_res) {
+ num = clp->bs * clp->bpt;
+ res = ioctl(fd, SG_SET_RESERVED_SIZE, &num);
+ if (res < 0) {
+ perror("sgh_dd: SG_SET_RESERVED_SIZE error");
+ return 0;
+ } else {
+ int nn;
+
+ res = ioctl(fd, SG_GET_RESERVED_SIZE, &nn);
+ if (res < 0) {
+ perror("sgh_dd: SG_GET_RESERVED_SIZE error");
+ return 0;
+ }
+ if (nn < num) {
+ pr2serr_lk("%s: SG_GET_RESERVED_SIZE shows size truncated, "
+ "wanted %d got %d\n", __func__, num, nn);
+ return 0;
+ }
+ }
+ if (mmpp) {
+ mmp = (uint8_t *)mmap(NULL, num, PROT_READ | PROT_WRITE,
+ MAP_SHARED, fd, 0);
+ if (MAP_FAILED == mmp) {
+ int err = errno;
+
+ pr2serr_lk("%s%s: sz=%d, fd=%d, mmap() failed: %s\n",
+ my_name, __func__, num, fd, strerror(err));
+ return 0;
+ }
+ *mmpp = mmp;
+ }
+ }
+ t = 1;
+ res = ioctl(fd, SG_SET_FORCE_PACK_ID, &t);
+ if (res < 0)
+ perror("sgh_dd: SG_SET_FORCE_PACK_ID error");
+ if (clp->unit_nanosec && sg_version_ge_40045) {
+ memset(seip, 0, sizeof(*seip));
+ seip->sei_wr_mask |= SG_SEIM_CTL_FLAGS;
+ seip->ctl_flags_wr_mask |= SG_CTL_FLAGM_TIME_IN_NS;
+ seip->ctl_flags |= SG_CTL_FLAGM_TIME_IN_NS;
+ if (ioctl(fd, SG_SET_GET_EXTENDED, seip) < 0) {
+ pr2serr_lk("ioctl(EXTENDED(TIME_IN_NS)) failed, errno=%d %s\n",
+ errno, strerror(errno));
+ }
+ }
+ t = 1;
+ res = ioctl(fd, SG_SET_DEBUG, &t); /* more info in the kernel log */
+ if (res < 0)
+ perror("sgs_dd: SG_SET_DEBUG error");
+ return (res < 0) ? 0 : num;
+}
+
+static bool
+process_flags(const char * arg, struct flags_t * fp)
+{
+ char buff[256];
+ char * cp;
+ char * np;
+
+ strncpy(buff, arg, sizeof(buff));
+ buff[sizeof(buff) - 1] = '\0';
+ if ('\0' == buff[0]) {
+ pr2serr("no flag found\n");
+ return false;
+ }
+ cp = buff;
+ do {
+ np = strchr(cp, ',');
+ if (np)
+ *np++ = '\0';
+ if (0 == strcmp(cp, "00"))
+ fp->zero = true;
+ else if (0 == strcmp(cp, "append"))
+ fp->append = true;
+ else if (0 == strcmp(cp, "coe"))
+ fp->coe = true;
+ else if (0 == strcmp(cp, "defres"))
+ fp->defres = true;
+ else if (0 == strcmp(cp, "dio"))
+ fp->dio = true;
+ else if (0 == strcmp(cp, "direct"))
+ fp->direct = true;
+ else if (0 == strcmp(cp, "dpo"))
+ fp->dpo = true;
+ else if (0 == strcmp(cp, "dsync"))
+ fp->dsync = true;
+ else if (0 == strcmp(cp, "excl"))
+ fp->excl = true;
+ else if (0 == strcmp(cp, "ff"))
+ fp->ff = true;
+ else if (0 == strcmp(cp, "fua"))
+ fp->fua = true;
+ else if (0 == strcmp(cp, "hipri"))
+ fp->polled = true;
+ else if (0 == strcmp(cp, "masync"))
+ fp->masync = true;
+ else if (0 == strcmp(cp, "mmap"))
+ ++fp->mmap; /* mmap > 1 stops munmap() being called */
+ else if (0 == strcmp(cp, "mrq_imm"))
+ fp->mrq_immed = true;
+ else if (0 == strcmp(cp, "mrq_immed"))
+ fp->mrq_immed = true;
+ else if (0 == strcmp(cp, "mrq_svb"))
+ fp->mrq_svb = true;
+ else if (0 == strcmp(cp, "nodur"))
+ fp->no_dur = true;
+ else if (0 == strcmp(cp, "no_dur"))
+ fp->no_dur = true;
+ else if (0 == strcmp(cp, "nocreat"))
+ fp->nocreat = true;
+ else if (0 == strcmp(cp, "noshare"))
+ fp->noshare = true;
+ else if (0 == strcmp(cp, "no_share"))
+ fp->noshare = true;
+ else if (0 == strcmp(cp, "no_thresh"))
+ fp->no_thresh = true;
+ else if (0 == strcmp(cp, "no-thresh"))
+ fp->no_thresh = true;
+ else if (0 == strcmp(cp, "nothresh"))
+ fp->no_thresh = true;
+ else if (0 == strcmp(cp, "no_unshare"))
+ fp->no_unshare = true;
+ else if (0 == strcmp(cp, "no-unshare"))
+ fp->no_unshare = true;
+ else if (0 == strcmp(cp, "no_waitq"))
+ fp->no_waitq = true;
+ else if (0 == strcmp(cp, "no-waitq"))
+ fp->no_waitq = true;
+ else if (0 == strcmp(cp, "nowaitq"))
+ fp->no_waitq = true;
+ else if (0 == strcmp(cp, "noxfer"))
+ fp->noxfer = true;
+ else if (0 == strcmp(cp, "no_xfer"))
+ fp->noxfer = true;
+ else if (0 == strcmp(cp, "null"))
+ ;
+ else if (0 == strcmp(cp, "polled"))
+ fp->polled = true;
+ else if (0 == strcmp(cp, "qhead"))
+ fp->qhead = true;
+ else if (0 == strcmp(cp, "qtail"))
+ fp->qtail = true;
+ else if (0 == strcmp(cp, "random"))
+ fp->random = true;
+ else if ((0 == strcmp(cp, "mout_if")) || (0 == strcmp(cp, "mout-if")))
+ fp->mout_if = true;
+ else if (0 == strcmp(cp, "same_fds"))
+ fp->same_fds = true;
+ else if (0 == strcmp(cp, "swait"))
+ fp->swait = true;
+ else if (0 == strcmp(cp, "v3"))
+ fp->v3 = true;
+ else if (0 == strcmp(cp, "v4")) {
+ fp->v4 = true;
+ fp->v4_given = true;
+ } else if (0 == strcmp(cp, "wq_excl"))
+ fp->wq_excl = true;
+ else {
+ pr2serr("unrecognised flag: %s\n", cp);
+ return false;
+ }
+ cp = np;
+ } while (cp);
+ return true;
+}
+
+/* Returns the number of times 'ch' is found in string 's' given the
+ * string's length. */
+static int
+num_chs_in_str(const char * s, int slen, int ch)
+{
+ int res = 0;
+
+ while (--slen >= 0) {
+ if (ch == s[slen])
+ ++res;
+ }
+ return res;
+}
+
+static int
+sg_in_open(struct global_collection *clp, const char *inf, uint8_t **mmpp,
+ int * mmap_lenp)
+{
+ int fd, n;
+ int flags = O_RDWR;
+
+ if (clp->in_flags.direct)
+ flags |= O_DIRECT;
+ if (clp->in_flags.excl)
+ flags |= O_EXCL;
+ if (clp->in_flags.dsync)
+ flags |= O_SYNC;
+
+ if ((fd = open(inf, flags)) < 0) {
+ int err = errno;
+ char ebuff[EBUFF_SZ];
+
+ snprintf(ebuff, EBUFF_SZ, "%s: could not open %s for sg reading",
+ __func__, inf);
+ perror(ebuff);
+ return -sg_convert_errno(err);
+ }
+ n = sg_prepare_resbuf(fd, true, clp, mmpp);
+ if (n <= 0) {
+ close(fd);
+ return -SG_LIB_FILE_ERROR;
+ }
+ if (clp->noshare)
+ sg_noshare_enlarge(fd, clp->verbose > 3);
+ if (mmap_lenp)
+ *mmap_lenp = n;
+ return fd;
+}
+
+static int
+sg_out_open(struct global_collection *clp, const char *outf, uint8_t **mmpp,
+ int * mmap_lenp)
+{
+ int fd, n;
+ int flags = O_RDWR;
+
+ if (clp->out_flags.direct)
+ flags |= O_DIRECT;
+ if (clp->out_flags.excl)
+ flags |= O_EXCL;
+ if (clp->out_flags.dsync)
+ flags |= O_SYNC;
+
+ if ((fd = open(outf, flags)) < 0) {
+ int err = errno;
+ char ebuff[EBUFF_SZ];
+
+ snprintf(ebuff, EBUFF_SZ, "%s: could not open %s for sg %s",
+ __func__, outf, (clp->verify ? "verifying" : "writing"));
+ perror(ebuff);
+ return -sg_convert_errno(err);
+ }
+ n = sg_prepare_resbuf(fd, false, clp, mmpp);
+ if (n <= 0) {
+ close(fd);
+ return -SG_LIB_FILE_ERROR;
+ }
+ if (clp->noshare)
+ sg_noshare_enlarge(fd, clp->verbose > 3);
+ if (mmap_lenp)
+ *mmap_lenp = n;
+ return fd;
+}
+
+/* Process arguments given to 'conv=" option. Returns 0 on success,
+ * 1 on error. */
+static int
+process_conv(const char * arg, struct flags_t * ifp, struct flags_t * ofp)
+{
+ char buff[256];
+ char * cp;
+ char * np;
+
+ strncpy(buff, arg, sizeof(buff));
+ buff[sizeof(buff) - 1] = '\0';
+ if ('\0' == buff[0]) {
+ pr2serr("no conversions found\n");
+ return 1;
+ }
+ cp = buff;
+ do {
+ np = strchr(cp, ',');
+ if (np)
+ *np++ = '\0';
+ if (0 == strcmp(cp, "nocreat"))
+ ofp->nocreat = true;
+ else if (0 == strcmp(cp, "noerror"))
+ ifp->coe = true; /* will still fail on write error */
+ else if (0 == strcmp(cp, "notrunc"))
+ ; /* this is the default action of sg_dd so ignore */
+ else if (0 == strcmp(cp, "null"))
+ ;
+ else if (0 == strcmp(cp, "sync"))
+ ; /* dd(susv4): pad errored block(s) with zeros but sg_dd does
+ * that by default. Typical dd use: 'conv=noerror,sync' */
+ else {
+ pr2serr("unrecognised flag: %s\n", cp);
+ return 1;
+ }
+ cp = np;
+ } while (cp);
+ return 0;
+}
+
+#define STR_SZ 1024
+#define INOUTF_SZ 512
+
+static int
+parse_cmdline_sanity(int argc, char * argv[], struct global_collection * clp,
+ char * inf, char * outf, char * out2f, char * outregf)
+{
+ bool verbose_given = false;
+ bool version_given = false;
+ bool verify_given = false;
+ bool bpt_given = false;
+ int ibs = 0;
+ int obs = 0;
+ int k, keylen, n, res;
+ char str[STR_SZ];
+ char * key;
+ char * buf;
+ const char * cp;
+
+ for (k = 1; k < argc; k++) {
+ if (argv[k]) {
+ strncpy(str, argv[k], STR_SZ);
+ str[STR_SZ - 1] = '\0';
+ }
+ else
+ continue;
+ for (key = str, buf = key; *buf && *buf != '=';)
+ buf++;
+ if (*buf)
+ *buf++ = '\0';
+ keylen = strlen(key);
+ if (0 == strcmp(key, "ae")) {
+ clp->aen = sg_get_num(buf);
+ if (clp->aen < 0) {
+ pr2serr("%sbad AEN argument to 'ae=', want 0 or higher\n",
+ my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ cp = strchr(buf, ',');
+ if (cp) {
+ clp->m_aen = sg_get_num(cp + 1);
+ if (clp->m_aen < 0) {
+ pr2serr("%sbad MAEN argument to 'ae=', want 0 or "
+ "higher\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ clp->m_aen_given = true;
+ }
+ clp->aen_given = true;
+ } else if (0 == strcmp(key, "bpt")) {
+ clp->bpt = sg_get_num(buf);
+ if ((clp->bpt < 0) || (clp->bpt > MAX_BPT_VALUE)) {
+ pr2serr("%sbad argument to 'bpt='\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ bpt_given = true;
+ } else if (0 == strcmp(key, "bs")) {
+ clp->bs = sg_get_num(buf);
+ if ((clp->bs < 0) || (clp->bs > MAX_BPT_VALUE)) {
+ pr2serr("%sbad argument to 'bs='\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ } else if (0 == strcmp(key, "cdbsz")) {
+ clp->cdbsz_in = sg_get_num(buf);
+ if ((clp->cdbsz_in < 6) || (clp->cdbsz_in > 32)) {
+ pr2serr("%s'cdbsz' expects 6, 10, 12, 16 or 32\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ clp->cdbsz_out = clp->cdbsz_in;
+ clp->cdbsz_given = true;
+ } else if (0 == strcmp(key, "coe")) {
+ clp->in_flags.coe = !! sg_get_num(buf);
+ clp->out_flags.coe = clp->in_flags.coe;
+ } else if (0 == strcmp(key, "conv")) {
+ if (process_conv(buf, &clp->in_flags, &clp->out_flags)) {
+ pr2serr("%s: bad argument to 'conv='\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ } else if (0 == strcmp(key, "count")) {
+ if (0 != strcmp("-1", buf)) {
+ dd_count = sg_get_llnum(buf);
+ if ((dd_count < 0) || (dd_count > MAX_COUNT_SKIP_SEEK)) {
+ pr2serr("%sbad argument to 'count='\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ } /* treat 'count=-1' as calculate count (same as not given) */
+ } else if (0 == strcmp(key, "dio")) {
+ clp->in_flags.dio = !! sg_get_num(buf);
+ clp->out_flags.dio = clp->in_flags.dio;
+ } else if (0 == strcmp(key, "elemsz_kb")) {
+ n = sg_get_num(buf);
+ if (n < 1) {
+ pr2serr("elemsz_kb=EKB wants an integer > 0\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if (n & (n - 1)) {
+ pr2serr("elemsz_kb=EKB wants EKB to be power of 2\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ clp->elem_sz = n * 1024;
+ } else if ((0 == strcmp(key, "fail_mask")) ||
+ (0 == strcmp(key, "fail-mask"))) {
+ clp->fail_mask = sg_get_num(buf);
+ if (clp->fail_mask < 0) {
+ pr2serr("fail_mask: couldn't decode argument\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ } else if (0 == strcmp(key, "fua")) {
+ n = sg_get_num(buf);
+ if (n & 1)
+ clp->out_flags.fua = true;
+ if (n & 2)
+ clp->in_flags.fua = true;
+ } else if (0 == strcmp(key, "ibs")) {
+ ibs = sg_get_num(buf);
+ if ((ibs < 0) || (ibs > MAX_BPT_VALUE)) {
+ pr2serr("%sbad argument to 'ibs='\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ } else if (0 == strcmp(key, "if")) {
+ if ('\0' != inf[0]) {
+ pr2serr("Second 'if=' argument??\n");
+ return SG_LIB_SYNTAX_ERROR;
+ } else {
+ memcpy(inf, buf, INOUTF_SZ);
+ inf[INOUTF_SZ - 1] = '\0'; /* noisy compiler */
+ }
+ } else if (0 == strcmp(key, "iflag")) {
+ if (! process_flags(buf, &clp->in_flags)) {
+ pr2serr("%sbad argument to 'iflag='\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ } else if (0 == strcmp(key, "mrq")) {
+ if (isdigit(buf[0]))
+ cp = buf;
+ else {
+ if ('I' == isupper(buf[0]))
+ clp->is_mrq_i = true;
+ else if ('O' == isupper(buf[0]))
+ clp->is_mrq_o = true;
+ else {
+ pr2serr("%sonly mrq=i,NRQS or mrq=o,NRQS allowed here\n",
+ my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ cp = strchr(buf, ',');
+ ++cp;
+ }
+ clp->nmrqs = sg_get_num(cp);
+ if (clp->nmrqs < 0) {
+ pr2serr("%sbad argument to 'mrq='\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ cp = strchr(cp, ',');
+ if (cp && ('C' == toupper(cp[1])))
+ clp->mrq_cmds = true;
+ } else if (0 == strcmp(key, "noshare")) {
+ clp->noshare = !! sg_get_num(buf);
+ } else if (0 == strcmp(key, "obs")) {
+ obs = sg_get_num(buf);
+ if ((obs < 0) || (obs > MAX_BPT_VALUE)) {
+ pr2serr("%sbad argument to 'obs='\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ } else if (strcmp(key, "of2") == 0) {
+ if ('\0' != out2f[0]) {
+ pr2serr("Second OFILE2 argument??\n");
+ return SG_LIB_CONTRADICT;
+ } else {
+ memcpy(out2f, buf, INOUTF_SZ);
+ out2f[INOUTF_SZ - 1] = '\0'; /* noisy compiler */
+ }
+ } else if (strcmp(key, "ofreg") == 0) {
+ if ('\0' != outregf[0]) {
+ pr2serr("Second OFREG argument??\n");
+ return SG_LIB_CONTRADICT;
+ } else {
+ memcpy(outregf, buf, INOUTF_SZ);
+ outregf[INOUTF_SZ - 1] = '\0'; /* noisy compiler */
+ }
+ } else if (0 == strcmp(key, "ofsplit")) {
+ clp->ofsplit = sg_get_num(buf);
+ if (-1 == clp->ofsplit) {
+ pr2serr("%sbad argument to 'ofsplit='\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ } else if (strcmp(key, "of") == 0) {
+ if ('\0' != outf[0]) {
+ pr2serr("Second 'of=' argument??\n");
+ return SG_LIB_SYNTAX_ERROR;
+ } else {
+ memcpy(outf, buf, INOUTF_SZ);
+ outf[INOUTF_SZ - 1] = '\0'; /* noisy compiler */
+ }
+ } else if (0 == strcmp(key, "oflag")) {
+ if (! process_flags(buf, &clp->out_flags)) {
+ pr2serr("%sbad argument to 'oflag='\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ } else if (0 == strcmp(key, "sdt")) {
+ cp = strchr(buf, ',');
+ n = sg_get_num(buf);
+ if (n < 0) {
+ pr2serr("%sbad argument to 'sdt=CRT[,ICT]'\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ clp->sdt_crt = n;
+ if (cp) {
+ n = sg_get_num(cp + 1);
+ if (n < 0) {
+ pr2serr("%sbad 2nd argument to 'sdt=CRT,ICT'\n",
+ my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ clp->sdt_ict = n;
+ }
+ } else if (0 == strcmp(key, "seek")) {
+ clp->seek = sg_get_llnum(buf);
+ if (clp->seek < 0) {
+ pr2serr("%sbad argument to 'seek='\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ } else if (0 == strcmp(key, "skip")) {
+ clp->skip = sg_get_llnum(buf);
+ if (clp->skip < 0) {
+ pr2serr("%sbad argument to 'skip='\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ } else if (0 == strcmp(key, "sync"))
+ do_sync = !! sg_get_num(buf);
+ else if (0 == strcmp(key, "thr"))
+ num_threads = sg_get_num(buf);
+ else if (0 == strcmp(key, "time")) {
+ do_time = sg_get_num(buf);
+ if (do_time < 0) {
+ pr2serr("%sbad argument to 'time=0|1|2'\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ cp = strchr(buf, ',');
+ if (cp) {
+ n = sg_get_num(cp + 1);
+ if (n < 0) {
+ pr2serr("%sbad argument to 'time=0|1|2,TO'\n", my_name);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ clp->cmd_timeout = n ? (n * 1000) : DEF_TIMEOUT;
+ }
+ } else if (0 == strcmp(key, "unshare"))
+ clp->unshare = !! sg_get_num(buf); /* default: true */
+ else if (0 == strncmp(key, "verb", 4))
+ clp->verbose = sg_get_num(buf);
+ else if ((keylen > 1) && ('-' == key[0]) && ('-' != key[1])) {
+ res = 0;
+ n = num_chs_in_str(key + 1, keylen - 1, 'c');
+ clp->chkaddr += n;
+ res += n;
+ n = num_chs_in_str(key + 1, keylen - 1, 'd');
+ clp->dry_run += n;
+ res += n;
+ n = num_chs_in_str(key + 1, keylen - 1, 'h');
+ clp->help += n;
+ res += n;
+ n = num_chs_in_str(key + 1, keylen - 1, 'p');
+ if (n > 0)
+ clp->prefetch = true;
+ res += n;
+ n = num_chs_in_str(key + 1, keylen - 1, 'v');
+ if (n > 0)
+ verbose_given = true;
+ clp->verbose += n; /* -v ---> --verbose */
+ res += n;
+ n = num_chs_in_str(key + 1, keylen - 1, 'V');
+ if (n > 0)
+ version_given = true;
+ res += n;
+ n = num_chs_in_str(key + 1, keylen - 1, 'x');
+ if (n > 0)
+ verify_given = true;
+ res += n;
+
+ if (res < (keylen - 1)) {
+ pr2serr("Unrecognised short option in '%s', try '--help'\n",
+ key);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ } else if (0 == strncmp(key, "--chkaddr", 9))
+ ++clp->chkaddr;
+ else if ((0 == strncmp(key, "--dry-run", 9)) ||
+ (0 == strncmp(key, "--dry_run", 9)))
+ ++clp->dry_run;
+ else if ((0 == strncmp(key, "--help", 6)) ||
+ (0 == strcmp(key, "-?")))
+ ++clp->help;
+ else if ((0 == strncmp(key, "--prefetch", 10)) ||
+ (0 == strncmp(key, "--pre-fetch", 11)))
+ clp->prefetch = true;
+ else if (0 == strncmp(key, "--verb", 6)) {
+ verbose_given = true;
+ ++clp->verbose; /* --verbose */
+ } else if (0 == strncmp(key, "--veri", 6))
+ verify_given = true;
+ else if (0 == strncmp(key, "--vers", 6))
+ version_given = true;
+ else {
+ pr2serr("Unrecognized option '%s'\n", key);
+ pr2serr("For more information use '--help' or '-h'\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ }
+
+#ifdef DEBUG
+ pr2serr("In DEBUG mode, ");
+ if (verbose_given && version_given) {
+ pr2serr("but override: '-vV' given, zero verbose and continue\n");
+ verbose_given = false;
+ version_given = false;
+ clp->verbose = 0;
+ } else if (! verbose_given) {
+ pr2serr("set '-vv'\n");
+ clp->verbose = 2;
+ } else
+ pr2serr("keep verbose=%d\n", clp->verbose);
+#else
+ if (verbose_given && version_given)
+ pr2serr("Not in DEBUG mode, so '-vV' has no special action\n");
+#endif
+ if (version_given) {
+ pr2serr("%s%s\n", my_name, version_str);
+ return SG_LIB_OK_FALSE;
+ }
+ if (clp->help > 0) {
+ usage(clp->help);
+ return SG_LIB_OK_FALSE;
+ }
+ if (clp->bs <= 0) {
+ clp->bs = DEF_BLOCK_SIZE;
+ pr2serr("Assume default 'bs' ((logical) block size) of %d bytes\n",
+ clp->bs);
+ }
+ if (verify_given) {
+ pr2serr("Doing verify/cmp rather than copy\n");
+ clp->verify = true;
+ }
+ if ((ibs && (ibs != clp->bs)) || (obs && (obs != clp->bs))) {
+ pr2serr("If 'ibs' or 'obs' given must be same as 'bs'\n");
+ usage(0);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if ((clp->skip < 0) || (clp->seek < 0)) {
+ pr2serr("skip and seek cannot be negative\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if (clp->out_flags.append) {
+ if (clp->seek > 0) {
+ pr2serr("Can't use both append and seek switches\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if (verify_given) {
+ pr2serr("Can't use both append and verify switches\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ }
+ if (clp->bpt < 1) {
+ pr2serr("bpt must be greater than 0\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if (clp->in_flags.mmap && clp->out_flags.mmap) {
+ pr2serr("mmap flag on both IFILE and OFILE doesn't work\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if (! clp->noshare) {
+ if (clp->in_flags.noshare || clp->out_flags.noshare)
+ clp->noshare = true;
+ }
+ if (clp->unshare) {
+ if (clp->in_flags.no_unshare || clp->out_flags.no_unshare)
+ clp->unshare = false;
+ }
+ if (clp->out_flags.mmap && ! clp->noshare) {
+ pr2serr("oflag=mmap needs either noshare=1\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if ((clp->in_flags.mmap || clp->out_flags.mmap) &&
+ (clp->in_flags.same_fds || clp->out_flags.same_fds)) {
+ pr2serr("can't have both 'mmap' and 'same_fds' flags\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if ((! clp->noshare) && (clp->in_flags.dio || clp->out_flags.dio)) {
+ pr2serr("dio flag can only be used with noshare=1\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if (clp->nmrqs > 0) {
+ if (clp->in_flags.mrq_immed || clp->out_flags.mrq_immed)
+ clp->mrq_async = true;
+ }
+ /* defaulting transfer size to 128*2048 for CD/DVDs is too large
+ for the block layer in lk 2.6 and results in an EIO on the
+ SG_IO ioctl. So reduce it in that case. */
+ if ((clp->bs >= 2048) && (! bpt_given))
+ clp->bpt = DEF_BLOCKS_PER_2048TRANSFER;
+ if (clp->ofsplit >= clp->bpt) {
+ pr2serr("ofsplit when given must be less than BPT\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if ((num_threads < 1) || (num_threads > MAX_NUM_THREADS)) {
+ pr2serr("too few or too many threads requested\n");
+ usage(1);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if (clp->in_flags.swait || clp->out_flags.swait) {
+ if (clp->verbose)
+ pr2serr("the 'swait' flag is now ignored\n");
+ /* remnants ... */
+ if (clp->in_flags.swait && (! clp->out_flags.swait))
+ clp->out_flags.swait = true;
+ }
+ clp->unit_nanosec = (do_time > 1) || !!getenv("SG3_UTILS_LINUX_NANO");
+#if 0
+ if (clp->verbose) {
+ pr2serr("%sif=%s skip=%" PRId64 " of=%s seek=%" PRId64 " count=%"
+ PRId64, my_name, inf, clp->skip, outf, clp->seek, dd_count);
+ if (clp->nmrqs > 0)
+ pr2serr(" mrq=%d%s\n", clp->nmrqs, (clp->mrq_cmds ? ",C" : ""));
+ else
+ pr2serr("\n");
+ }
+#endif
+ return 0;
+}
+
+
+int
+main(int argc, char * argv[])
+{
+ char inf[INOUTF_SZ];
+ char outf[INOUTF_SZ];
+ char out2f[INOUTF_SZ];
+ char outregf[INOUTF_SZ];
+ int res, k, err;
+ int64_t in_num_sect = 0;
+ int64_t out_num_sect = 0;
+ int in_sect_sz, out_sect_sz, status, flags;
+ void * vp;
+ const char * ccp = NULL;
+ const char * cc2p;
+ struct global_collection * clp = &gcoll;
+ Thread_info thread_arr[MAX_NUM_THREADS] {};
+ char ebuff[EBUFF_SZ];
+#if SG_LIB_ANDROID
+ struct sigaction actions;
+
+ memset(&actions, 0, sizeof(actions));
+ sigemptyset(&actions.sa_mask);
+ actions.sa_flags = 0;
+ actions.sa_handler = thread_exit_handler;
+ sigaction(SIGUSR1, &actions, NULL);
+ sigaction(SIGUSR2, &actions, NULL);
+#endif
+ /* memset(clp, 0, sizeof(*clp)); */
+ clp->bpt = DEF_BLOCKS_PER_TRANSFER;
+ clp->cmd_timeout = DEF_TIMEOUT;
+ clp->in_type = FT_OTHER;
+ /* change dd's default: if of=OFILE not given, assume /dev/null */
+ clp->out_type = FT_DEV_NULL;
+ clp->out2_type = FT_DEV_NULL;
+ clp->cdbsz_in = DEF_SCSI_CDBSZ;
+ clp->cdbsz_out = DEF_SCSI_CDBSZ;
+ clp->sdt_ict = DEF_SDT_ICT_MS;
+ clp->sdt_crt = DEF_SDT_CRT_SEC;
+ clp->nmrqs = DEF_NUM_MRQS;
+ clp->unshare = true;
+ inf[0] = '\0';
+ outf[0] = '\0';
+ out2f[0] = '\0';
+ outregf[0] = '\0';
+ fetch_sg_version();
+ if (sg_version >= 40045)
+ sg_version_ge_40045 = true;
+
+ res = parse_cmdline_sanity(argc, argv, clp, inf, outf, out2f, outregf);
+ if (SG_LIB_OK_FALSE == res)
+ return 0;
+ if (res)
+ return res;
+ if (sg_version > 40000) {
+ if (! clp->in_flags.v3)
+ clp->in_flags.v4 = true;
+ if (! clp->out_flags.v3)
+ clp->out_flags.v4 = true;
+ }
+
+ install_handler(SIGINT, interrupt_handler);
+ install_handler(SIGQUIT, interrupt_handler);
+ install_handler(SIGPIPE, interrupt_handler);
+ install_handler(SIGUSR1, siginfo_handler);
+ install_handler(SIGUSR2, siginfo2_handler);
+
+ clp->infd = STDIN_FILENO;
+ clp->outfd = STDOUT_FILENO;
+ if (clp->in_flags.ff && clp->in_flags.zero) {
+ ccp = "<addr_as_data>";
+ cc2p = "addr_as_data";
+ } else if (clp->in_flags.ff) {
+ ccp = "<0xff bytes>";
+ cc2p = "ff";
+ } else if (clp->in_flags.random) {
+ ccp = "<random>";
+ cc2p = "random";
+ } else if (clp->in_flags.zero) {
+ ccp = "<zero bytes>";
+ cc2p = "00";
+ }
+ if (ccp) {
+ if (inf[0]) {
+ pr2serr("%siflag=%s and if=%s contradict\n", my_name, cc2p, inf);
+ return SG_LIB_CONTRADICT;
+ }
+ clp->in_type = FT_RANDOM_0_FF;
+ clp->infp = ccp;
+ clp->infd = -1;
+ } else if (inf[0] && ('-' != inf[0])) {
+ clp->in_type = dd_filetype(inf, clp->in_st_size);
+
+ if (FT_ERROR == clp->in_type) {
+ pr2serr("%sunable to access %s\n", my_name, inf);
+ return SG_LIB_FILE_ERROR;
+ } else if (FT_ST == clp->in_type) {
+ pr2serr("%sunable to use scsi tape device %s\n", my_name, inf);
+ return SG_LIB_FILE_ERROR;
+ } else if (FT_CHAR == clp->in_type) {
+ pr2serr("%sunable to use unknown char device %s\n", my_name, inf);
+ return SG_LIB_FILE_ERROR;
+ } else if (FT_SG == clp->in_type) {
+ clp->infd = sg_in_open(clp, inf, NULL, NULL);
+ if (clp->verbose > 2)
+ pr2serr("using sg v%c interface on %s\n",
+ (clp->in_flags.v4 ? '4' : '3'), inf);
+ if (clp->infd < 0)
+ return -clp->infd;
+ } else {
+ flags = O_RDONLY;
+ if (clp->in_flags.direct)
+ flags |= O_DIRECT;
+ if (clp->in_flags.excl)
+ flags |= O_EXCL;
+ if (clp->in_flags.dsync)
+ flags |= O_SYNC;
+
+ if ((clp->infd = open(inf, flags)) < 0) {
+ err = errno;
+ snprintf(ebuff, EBUFF_SZ, "%scould not open %s for reading",
+ my_name, inf);
+ perror(ebuff);
+ return sg_convert_errno(err);
+ } else if (clp->skip > 0) {
+ off64_t offset = clp->skip;
+
+ offset *= clp->bs; /* could exceed 32 here! */
+ if (lseek64(clp->infd, offset, SEEK_SET) < 0) {
+ err = errno;
+ snprintf(ebuff, EBUFF_SZ, "%scouldn't skip to required "
+ "position on %s", my_name, inf);
+ perror(ebuff);
+ return sg_convert_errno(err);
+ }
+ }
+ }
+ clp->infp = inf;
+ if ((clp->in_flags.v3 || clp->in_flags.v4_given) &&
+ (FT_SG != clp->in_type)) {
+ clp->in_flags.v3 = false;
+ clp->in_flags.v4 = false;
+ pr2serr("%siflag= v3 and v4 both ignored when IFILE is not sg "
+ "device\n", my_name);
+ }
+ }
+ if (clp->verbose && (clp->in_flags.no_waitq || clp->out_flags.no_waitq))
+ pr2serr("no_waitq: flag no longer does anything\n");
+ if (outf[0])
+ clp->ofile_given = true;
+ if (outf[0] && ('-' != outf[0])) {
+ clp->out_type = dd_filetype(outf, clp->out_st_size);
+
+ if ((FT_SG != clp->out_type) && clp->verify) {
+ pr2serr("%s --verify only supported by sg OFILEs\n", my_name);
+ return SG_LIB_FILE_ERROR;
+ } else if (FT_ST == clp->out_type) {
+ pr2serr("%sunable to use scsi tape device %s\n", my_name, outf);
+ return SG_LIB_FILE_ERROR;
+ } else if (FT_CHAR == clp->out_type) {
+ pr2serr("%sunable to use unknown char device %s\n", my_name, outf);
+ return SG_LIB_FILE_ERROR;
+ } else if (FT_SG == clp->out_type) {
+ clp->outfd = sg_out_open(clp, outf, NULL, NULL);
+ if (clp->verbose > 2)
+ pr2serr("using sg v%c interface on %s\n",
+ (clp->out_flags.v4 ? '4' : '3'), outf);
+ if (clp->outfd < 0)
+ return -clp->outfd;
+ } else if (FT_DEV_NULL == clp->out_type)
+ clp->outfd = -1; /* don't bother opening */
+ else {
+ flags = O_WRONLY;
+ if (! clp->out_flags.nocreat)
+ flags |= O_CREAT;
+ if (clp->out_flags.direct)
+ flags |= O_DIRECT;
+ if (clp->out_flags.excl)
+ flags |= O_EXCL;
+ if (clp->out_flags.dsync)
+ flags |= O_SYNC;
+ if (clp->out_flags.append)
+ flags |= O_APPEND;
+
+ if ((clp->outfd = open(outf, flags, 0666)) < 0) {
+ err = errno;
+ snprintf(ebuff, EBUFF_SZ, "%scould not open %s for "
+ "writing", my_name, outf);
+ perror(ebuff);
+ return sg_convert_errno(err);
+ }
+ if (clp->seek > 0) {
+ off64_t offset = clp->seek;
+
+ offset *= clp->bs; /* could exceed 32 bits here! */
+ if (lseek64(clp->outfd, offset, SEEK_SET) < 0) {
+ err = errno;
+ snprintf(ebuff, EBUFF_SZ, "%scouldn't seek to required "
+ "position on %s", my_name, outf);
+ perror(ebuff);
+ return sg_convert_errno(err);
+ }
+ }
+ }
+ clp->outfp = outf;
+ if ((clp->out_flags.v3 || clp->out_flags.v4_given) &&
+ (FT_SG != clp->out_type)) {
+ clp->out_flags.v3 = false;
+ clp->out_flags.v4 = false;
+ pr2serr("%soflag= v3 and v4 both ignored when OFILE is not sg "
+ "device\n", my_name);
+ }
+ }
+
+ if (out2f[0])
+ clp->ofile2_given = true;
+ if (out2f[0] && ('-' != out2f[0])) {
+ off_t out2_st_size;
+
+ clp->out2_type = dd_filetype(out2f, out2_st_size);
+ if (FT_ST == clp->out2_type) {
+ pr2serr("%sunable to use scsi tape device %s\n", my_name, out2f);
+ return SG_LIB_FILE_ERROR;
+ }
+ else if (FT_SG == clp->out2_type) {
+ clp->out2fd = sg_out_open(clp, out2f, NULL, NULL);
+ if (clp->out2fd < 0)
+ return -clp->out2fd;
+ }
+ else if (FT_DEV_NULL == clp->out2_type)
+ clp->out2fd = -1; /* don't bother opening */
+ else {
+ flags = O_WRONLY;
+ if (! clp->out_flags.nocreat)
+ flags |= O_CREAT;
+ if (clp->out_flags.direct)
+ flags |= O_DIRECT;
+ if (clp->out_flags.excl)
+ flags |= O_EXCL;
+ if (clp->out_flags.dsync)
+ flags |= O_SYNC;
+ if (clp->out_flags.append)
+ flags |= O_APPEND;
+
+ if ((clp->out2fd = open(out2f, flags, 0666)) < 0) {
+ err = errno;
+ snprintf(ebuff, EBUFF_SZ, "%scould not open %s for "
+ "writing", my_name, out2f);
+ perror(ebuff);
+ return sg_convert_errno(err);
+ }
+ if (clp->seek > 0) {
+ off64_t offset = clp->seek;
+
+ offset *= clp->bs; /* could exceed 32 bits here! */
+ if (lseek64(clp->out2fd, offset, SEEK_SET) < 0) {
+ err = errno;
+ snprintf(ebuff, EBUFF_SZ, "%scouldn't seek to required "
+ "position on %s", my_name, out2f);
+ perror(ebuff);
+ return sg_convert_errno(err);
+ }
+ }
+ }
+ clp->out2fp = out2f;
+ }
+ if ((FT_SG == clp->in_type ) && (FT_SG == clp->out_type)) {
+ if (clp->nmrqs > 0) {
+ if (clp->is_mrq_i == clp->is_mrq_o) {
+ if (clp->ofsplit > 0) {
+ if (0 != (clp->nmrqs % 3)) {
+ pr2serr("When both IFILE+OFILE sg devices and OSP>0, "
+ "mrq=NRQS must be divisible by 3\n");
+ pr2serr(" triple NRQS to avoid error\n");
+ clp->nmrqs *= 3;
+ }
+ } else if (0 != (clp->nmrqs % 2)) {
+ pr2serr("When both IFILE+OFILE sg devices (and OSP=0), "
+ "mrq=NRQS must be even\n");
+ pr2serr(" double NRQS to avoid error\n");
+ clp->nmrqs *= 2;
+ }
+ }
+ if (clp->is_mrq_i && clp->is_mrq_o)
+ ;
+ else if (clp->is_mrq_i || clp->is_mrq_o)
+ clp->unbalanced_mrq = true;
+ }
+ if (clp->in_flags.v4_given && (! clp->out_flags.v3)) {
+ if (! clp->out_flags.v4_given) {
+ clp->out_flags.v4 = true;
+ if (clp->verbose)
+ pr2serr("Changing OFILE from v3 to v4, use oflag=v3 to "
+ "force v3\n");
+ }
+ }
+ if (clp->out_flags.v4_given && (! clp->in_flags.v3)) {
+ if (! clp->in_flags.v4_given) {
+ clp->in_flags.v4 = true;
+ if (clp->verbose)
+ pr2serr("Changing IFILE from v3 to v4, use iflag=v3 to "
+ "force v3\n");
+ }
+ }
+#if 0
+ if (clp->mrq_async && !(clp->noshare)) {
+ pr2serr("With mrq_immed also need noshare on sg-->sg copy\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+#endif
+ } else if ((FT_SG == clp->in_type ) || (FT_SG == clp->out_type)) {
+ if (clp->nmrqs > 0)
+ clp->unbalanced_mrq = true;
+ }
+ if (outregf[0]) {
+ off_t outrf_st_size;
+ int ftyp = dd_filetype(outregf, outrf_st_size);
+
+ clp->outreg_type = ftyp;
+ if (! ((FT_OTHER == ftyp) || (FT_ERROR == ftyp) ||
+ (FT_DEV_NULL == ftyp))) {
+ pr2serr("File: %s can only be regular file or pipe (or "
+ "/dev/null)\n", outregf);
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if ((clp->outregfd = open(outregf, O_WRONLY | O_CREAT, 0666)) < 0) {
+ err = errno;
+ snprintf(ebuff, EBUFF_SZ, "could not open %s for writing",
+ outregf);
+ perror(ebuff);
+ return sg_convert_errno(err);
+ }
+ if (clp->verbose > 1)
+ pr2serr("ofreg=%s opened okay, fd=%d\n", outregf, clp->outregfd);
+ if (FT_ERROR == ftyp)
+ clp->outreg_type = FT_OTHER; /* regular file created */
+ } else
+ clp->outregfd = -1;
+
+ if ((STDIN_FILENO == clp->infd) && (STDOUT_FILENO == clp->outfd)) {
+ pr2serr("Won't default both IFILE to stdin _and_ OFILE to "
+ "/dev/null\n");
+ pr2serr("For more information use '--help' or '-h'\n");
+ return SG_LIB_SYNTAX_ERROR;
+ }
+ if (dd_count < 0) {
+ in_num_sect = -1;
+ if (FT_SG == clp->in_type) {
+ res = scsi_read_capacity(clp->infd, &in_num_sect, &in_sect_sz);
+ if (2 == res) {
+ pr2serr("Unit attention, media changed(in), continuing\n");
+ res = scsi_read_capacity(clp->infd, &in_num_sect,
+ &in_sect_sz);
+ }
+ if (0 != res) {
+ if (res == SG_LIB_CAT_INVALID_OP)
+ pr2serr("read capacity not supported on %s\n", inf);
+ else if (res == SG_LIB_CAT_NOT_READY)
+ pr2serr("read capacity failed, %s not ready\n", inf);
+ else
+ pr2serr("Unable to read capacity on %s\n", inf);
+ return SG_LIB_FILE_ERROR;
+ } else if (clp->bs != in_sect_sz) {
+ pr2serr(">> warning: logical block size on %s confusion: "
+ "bs=%d, device claims=%d\n", clp->infp, clp->bs,
+ in_sect_sz);
+ return SG_LIB_FILE_ERROR;
+ }
+ } else if (FT_BLOCK == clp->in_type) {
+ if (0 != read_blkdev_capacity(clp->infd, &in_num_sect,
+ &in_sect_sz)) {
+ pr2serr("Unable to read block capacity on %s\n", inf);
+ in_num_sect = -1;
+ }
+ if (clp->bs != in_sect_sz) {
+ pr2serr("logical block size on %s confusion; bs=%d, from "
+ "device=%d\n", inf, clp->bs, in_sect_sz);
+ in_num_sect = -1;
+ }
+ } else if (FT_OTHER == clp->in_type) {
+ in_num_sect = clp->in_st_size / clp->bs;
+ if (clp->in_st_size % clp->bs) {
+ ++in_num_sect;
+ pr2serr("Warning: the file size of %s is not a multiple of BS "
+ "[%d]\n", inf, clp->bs);
+ }
+ }
+ if (in_num_sect > clp->skip)
+ in_num_sect -= clp->skip;
+
+ out_num_sect = -1;
+ if (FT_SG == clp->out_type) {
+ res = scsi_read_capacity(clp->outfd, &out_num_sect, &out_sect_sz);
+ if (2 == res) {
+ pr2serr("Unit attention, media changed(out), continuing\n");
+ res = scsi_read_capacity(clp->outfd, &out_num_sect,
+ &out_sect_sz);
+ }
+ if (0 != res) {
+ if (res == SG_LIB_CAT_INVALID_OP)
+ pr2serr("read capacity not supported on %s\n", outf);
+ else if (res == SG_LIB_CAT_NOT_READY)
+ pr2serr("read capacity failed, %s not ready\n", outf);
+ else
+ pr2serr("Unable to read capacity on %s\n", outf);
+ out_num_sect = -1;
+ return SG_LIB_FILE_ERROR;
+ } else if (clp->bs != out_sect_sz) {
+ pr2serr(">> warning: logical block size on %s confusion: "
+ "bs=%d, device claims=%d\n", clp->outfp, clp->bs,
+ out_sect_sz);
+ return SG_LIB_FILE_ERROR;
+ }
+ } else if (FT_BLOCK == clp->out_type) {
+ if (0 != read_blkdev_capacity(clp->outfd, &out_num_sect,
+ &out_sect_sz)) {
+ pr2serr("Unable to read block capacity on %s\n", outf);
+ out_num_sect = -1;
+ }
+ if (clp->bs != out_sect_sz) {
+ pr2serr("logical block size on %s confusion: bs=%d, from "
+ "device=%d\n", outf, clp->bs, out_sect_sz);
+ out_num_sect = -1;
+ }
+ } else if (FT_OTHER == clp->out_type) {
+ out_num_sect = clp->out_st_size / clp->bs;
+ if (clp->out_st_size % clp->bs) {
+ ++out_num_sect;
+ pr2serr("Warning: the file size of %s is not a multiple of BS "
+ "[%d]\n", outf, clp->bs);
+ }
+ }
+ if (out_num_sect > clp->seek)
+ out_num_sect -= clp->seek;
+
+ if (in_num_sect > 0) {
+ if (out_num_sect > 0)
+ dd_count = (in_num_sect > out_num_sect) ? out_num_sect :
+ in_num_sect;
+ else
+ dd_count = in_num_sect;
+ }
+ else
+ dd_count = out_num_sect;
+ }
+ if (clp->verbose > 2)
+ pr2serr("Start of loop, count=%" PRId64 ", in_num_sect=%" PRId64
+ ", out_num_sect=%" PRId64 "\n", dd_count, in_num_sect,
+ out_num_sect);
+ if (dd_count < 0) {
+ pr2serr("Couldn't calculate count, please give one\n");
+ return SG_LIB_CAT_OTHER;
+ }
+ if (! clp->cdbsz_given) {
+ if ((FT_SG == clp->in_type) && (MAX_SCSI_CDBSZ != clp->cdbsz_in) &&
+ (((dd_count + clp->skip) > UINT_MAX) || (clp->bpt > USHRT_MAX))) {
+ pr2serr("Note: SCSI command size increased to 16 bytes (for "
+ "'if')\n");
+ clp->cdbsz_in = MAX_SCSI_CDBSZ;
+ }
+ if ((FT_SG == clp->out_type) && (MAX_SCSI_CDBSZ != clp->cdbsz_out) &&
+ (((dd_count + clp->seek) > UINT_MAX) || (clp->bpt > USHRT_MAX))) {
+ pr2serr("Note: SCSI command size increased to 16 bytes (for "
+ "'of')\n");
+ clp->cdbsz_out = MAX_SCSI_CDBSZ;
+ }
+ }
+
+ // clp->in_count = dd_count;
+ clp->in_rem_count = dd_count;
+ clp->out_count = dd_count;
+ clp->out_rem_count = dd_count;
+ clp->out_blk = clp->seek;
+ status = pthread_mutex_init(&clp->in_mutex, NULL);
+ if (0 != status) err_exit(status, "init in_mutex");
+ status = pthread_mutex_init(&clp->out_mutex, NULL);
+ if (0 != status) err_exit(status, "init out_mutex");
+ status = pthread_mutex_init(&clp->out2_mutex, NULL);
+ if (0 != status) err_exit(status, "init out2_mutex");
+ status = pthread_cond_init(&clp->out_sync_cv, NULL);
+ if (0 != status) err_exit(status, "init out_sync_cv");
+
+ if (clp->dry_run > 0) {
+ pr2serr("Due to --dry-run option, bypass copy/read\n");
+ goto fini;
+ }
+ if (! clp->ofile_given)
+ pr2serr("of=OFILE not given so only read from IFILE, to output to "
+ "stdout use 'of=-'\n");
+
+ sigemptyset(&signal_set);
+ sigaddset(&signal_set, SIGINT);
+ sigaddset(&signal_set, SIGUSR2);
+ status = pthread_sigmask(SIG_BLOCK, &signal_set, &orig_signal_set);
+ if (0 != status) err_exit(status, "pthread_sigmask");
+ status = pthread_create(&sig_listen_thread_id, NULL,
+ sig_listen_thread, (void *)clp);
+ if (0 != status) err_exit(status, "pthread_create, sig...");
+
+ if (do_time) {
+ start_tm.tv_sec = 0;
+ start_tm.tv_usec = 0;
+ gettimeofday(&start_tm, NULL);
+ }
+
+/* vvvvvvvvvvv Start worker threads vvvvvvvvvvvvvvvvvvvvvvvv */
+ if ((clp->out_rem_count.load() > 0) && (num_threads > 0)) {
+ Thread_info *tip = thread_arr + 0;
+
+ tip->gcp = clp;
+ tip->id = 0;
+ /* Run 1 work thread to shake down infant retryable stuff */
+ status = pthread_mutex_lock(&clp->out_mutex);
+ if (0 != status) err_exit(status, "lock out_mutex");
+ status = pthread_create(&tip->a_pthr, NULL, read_write_thread,
+ (void *)tip);
+ if (0 != status) err_exit(status, "pthread_create");
+
+ /* wait for any broadcast */
+ pthread_cleanup_push(cleanup_out, (void *)clp);
+ status = pthread_cond_wait(&clp->out_sync_cv, &clp->out_mutex);
+ if (0 != status) err_exit(status, "cond out_sync_cv");
+ pthread_cleanup_pop(0);
+ status = pthread_mutex_unlock(&clp->out_mutex);
+ if (0 != status) err_exit(status, "unlock out_mutex");
+
+ /* now start the rest of the threads */
+ for (k = 1; k < num_threads; ++k) {
+ tip = thread_arr + k;
+ tip->gcp = clp;
+ tip->id = k;
+ status = pthread_create(&tip->a_pthr, NULL, read_write_thread,
+ (void *)tip);
+ if (0 != status) err_exit(status, "pthread_create");
+ }
+
+ /* now wait for worker threads to finish */
+ for (k = 0; k < num_threads; ++k) {
+ tip = thread_arr + k;
+ status = pthread_join(tip->a_pthr, &vp);
+ if (0 != status) err_exit(status, "pthread_join");
+ if (clp->verbose > 2)
+ pr2serr_lk("%d <-- Worker thread terminated, vp=%s\n", k,
+ ((vp == clp) ? "clp" : "NULL (or !clp)"));
+ }
+ } /* started worker threads and here after they have all exited */
+
+ if (do_time && (start_tm.tv_sec || start_tm.tv_usec))
+ calc_duration_throughput(0);
+
+ shutting_down = true;
+ status = pthread_join(sig_listen_thread_id, &vp);
+ if (0 != status) err_exit(status, "pthread_join");
+#if 0
+ /* pthread_cancel() has issues and is not supported in Android */
+ status = pthread_kill(sig_listen_thread_id, SIGUSR2);
+ if (0 != status) err_exit(status, "pthread_kill");
+ std::this_thread::yield(); // not enough it seems
+ { /* allow time for SIGUSR2 signal to get through */
+ struct timespec tspec = {0, 400000}; /* 400 usecs */
+
+ nanosleep(&tspec, NULL);
+ }
+#endif
+
+ if (do_sync) {
+ if (FT_SG == clp->out_type) {
+ pr2serr_lk(">> Synchronizing cache on %s\n", outf);
+ res = sg_ll_sync_cache_10(clp->outfd, 0, 0, 0, 0, 0, false, 0);
+ if (SG_LIB_CAT_UNIT_ATTENTION == res) {
+ pr2serr_lk("Unit attention(out), continuing\n");
+ res = sg_ll_sync_cache_10(clp->outfd, 0, 0, 0, 0, 0, false,
+ 0);
+ }
+ if (0 != res)
+ pr2serr_lk("Unable to synchronize cache\n");
+ }
+ if (FT_SG == clp->out2_type) {
+ pr2serr_lk(">> Synchronizing cache on %s\n", out2f);
+ res = sg_ll_sync_cache_10(clp->out2fd, 0, 0, 0, 0, 0, false, 0);
+ if (SG_LIB_CAT_UNIT_ATTENTION == res) {
+ pr2serr_lk("Unit attention(out2), continuing\n");
+ res = sg_ll_sync_cache_10(clp->out2fd, 0, 0, 0, 0, 0, false,
+ 0);
+ }
+ if (0 != res)
+ pr2serr_lk("Unable to synchronize cache (of2)\n");
+ }
+ }
+
+fini:
+ if ((STDIN_FILENO != clp->infd) && (clp->infd >= 0))
+ close(clp->infd);
+ if ((STDOUT_FILENO != clp->outfd) && (FT_DEV_NULL != clp->out_type) &&
+ (clp->outfd >= 0))
+ close(clp->outfd);
+ if ((clp->out2fd >= 0) && (STDOUT_FILENO != clp->out2fd) &&
+ (FT_DEV_NULL != clp->out2_type))
+ close(clp->out2fd);
+ if ((clp->outregfd >= 0) && (STDOUT_FILENO != clp->outregfd) &&
+ (FT_DEV_NULL != clp->outreg_type))
+ close(clp->outregfd);
+ res = exit_status;
+ if ((0 != clp->out_count.load()) && (0 == clp->dry_run)) {
+ pr2serr(">>>> Some error occurred, remaining blocks=%" PRId64 "\n",
+ clp->out_count.load());
+ if (0 == res)
+ res = SG_LIB_CAT_OTHER;
+ }
+ print_stats("");
+ if (clp->dio_incomplete_count.load()) {
+ int fd;
+ char c;
+
+ pr2serr(">> Direct IO requested but incomplete %d times\n",
+ clp->dio_incomplete_count.load());
+ if ((fd = open(sg_allow_dio, O_RDONLY)) >= 0) {
+ if (1 == read(fd, &c, 1)) {
+ if ('0' == c)
+ pr2serr(">>> %s set to '0' but should be set to '1' for "
+ "direct IO\n", sg_allow_dio);
+ }
+ close(fd);
+ }
+ }
+ if (clp->sum_of_resids.load())
+ pr2serr(">> Non-zero sum of residual counts=%d\n",
+ clp->sum_of_resids.load());
+ if (clp->verbose && (num_start_eagain > 0))
+ pr2serr("Number of start EAGAINs: %d\n", num_start_eagain.load());
+ if (clp->verbose && (num_fin_eagain > 0))
+ pr2serr("Number of finish EAGAINs: %d\n", num_fin_eagain.load());
+ if (clp->verbose && (num_ebusy > 0))
+ pr2serr("Number of EBUSYs: %d\n", num_ebusy.load());
+ if (clp->verbose && clp->aen_given && (num_abort_req > 0)) {
+ pr2serr("Number of Aborts: %d\n", num_abort_req.load());
+ pr2serr("Number of successful Aborts: %d\n",
+ num_abort_req_success.load());
+ }
+ if (clp->verbose && clp->m_aen_given && (num_mrq_abort_req > 0)) {
+ pr2serr("Number of MRQ Aborts: %d\n", num_mrq_abort_req.load());
+ pr2serr("Number of successful MRQ Aborts: %d\n",
+ num_mrq_abort_req_success.load());
+ }
+ if (clp->verbose && (num_miscompare > 0))
+ pr2serr("Number of miscompare%s: %d\n",
+ (num_miscompare > 1) ? "s" : "", num_miscompare.load());
+ if (clp->verbose > 1) {
+ if (clp->verbose > 3)
+ pr2serr("Final pack_id=%d, mrq_id=%d\n", mono_pack_id.load(),
+ mono_mrq_id.load());
+ pr2serr("Number of SG_GET_NUM_WAITING calls=%ld\n",
+ num_waiting_calls.load());
+ }
+ if (clp->verify && (SG_LIB_CAT_MISCOMPARE == res))
+ pr2serr("Verify/compare failed due to miscompare\n");
+ return (res >= 0) ? res : SG_LIB_CAT_OTHER;
+}