forked from mirror/qemu
You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1050 lines
27 KiB
C
1050 lines
27 KiB
C
/*
|
|
* 9P network client for VirtIO 9P test cases (based on QTest)
|
|
*
|
|
* Copyright (c) 2014 SUSE LINUX Products GmbH
|
|
*
|
|
* This work is licensed under the terms of the GNU GPL, version 2 or later.
|
|
* See the COPYING file in the top-level directory.
|
|
*/
|
|
|
|
/*
|
|
* Not so fast! You might want to read the 9p developer docs first:
|
|
* https://wiki.qemu.org/Documentation/9p
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "virtio-9p-client.h"
|
|
|
|
#define QVIRTIO_9P_TIMEOUT_US (10 * 1000 * 1000)
|
|
static QGuestAllocator *alloc;
|
|
|
|
void v9fs_set_allocator(QGuestAllocator *t_alloc)
|
|
{
|
|
alloc = t_alloc;
|
|
}
|
|
|
|
/*
|
|
* Used to auto generate new fids. Start with arbitrary high value to avoid
|
|
* collision with hard coded fids in basic test code.
|
|
*/
|
|
static uint32_t fid_generator = 1000;
|
|
|
|
static uint32_t genfid(void)
|
|
{
|
|
return fid_generator++;
|
|
}
|
|
|
|
/**
|
|
* Splits the @a in string by @a delim into individual (non empty) strings
|
|
* and outputs them to @a out. The output array @a out is NULL terminated.
|
|
*
|
|
* Output array @a out must be freed by calling split_free().
|
|
*
|
|
* @returns number of individual elements in output array @a out (without the
|
|
* final NULL terminating element)
|
|
*/
|
|
static int split(const char *in, const char *delim, char ***out)
|
|
{
|
|
int n = 0, i = 0;
|
|
char *tmp, *p;
|
|
|
|
tmp = g_strdup(in);
|
|
for (p = strtok(tmp, delim); p != NULL; p = strtok(NULL, delim)) {
|
|
if (strlen(p) > 0) {
|
|
++n;
|
|
}
|
|
}
|
|
g_free(tmp);
|
|
|
|
*out = g_new0(char *, n + 1); /* last element NULL delimiter */
|
|
|
|
tmp = g_strdup(in);
|
|
for (p = strtok(tmp, delim); p != NULL; p = strtok(NULL, delim)) {
|
|
if (strlen(p) > 0) {
|
|
(*out)[i++] = g_strdup(p);
|
|
}
|
|
}
|
|
g_free(tmp);
|
|
|
|
return n;
|
|
}
|
|
|
|
static void split_free(char ***out)
|
|
{
|
|
int i;
|
|
if (!*out) {
|
|
return;
|
|
}
|
|
for (i = 0; (*out)[i]; ++i) {
|
|
g_free((*out)[i]);
|
|
}
|
|
g_free(*out);
|
|
*out = NULL;
|
|
}
|
|
|
|
void v9fs_memwrite(P9Req *req, const void *addr, size_t len)
|
|
{
|
|
qtest_memwrite(req->qts, req->t_msg + req->t_off, addr, len);
|
|
req->t_off += len;
|
|
}
|
|
|
|
void v9fs_memskip(P9Req *req, size_t len)
|
|
{
|
|
req->r_off += len;
|
|
}
|
|
|
|
void v9fs_memread(P9Req *req, void *addr, size_t len)
|
|
{
|
|
qtest_memread(req->qts, req->r_msg + req->r_off, addr, len);
|
|
req->r_off += len;
|
|
}
|
|
|
|
void v9fs_uint8_read(P9Req *req, uint8_t *val)
|
|
{
|
|
v9fs_memread(req, val, 1);
|
|
}
|
|
|
|
void v9fs_uint16_write(P9Req *req, uint16_t val)
|
|
{
|
|
uint16_t le_val = cpu_to_le16(val);
|
|
|
|
v9fs_memwrite(req, &le_val, 2);
|
|
}
|
|
|
|
void v9fs_uint16_read(P9Req *req, uint16_t *val)
|
|
{
|
|
v9fs_memread(req, val, 2);
|
|
le16_to_cpus(val);
|
|
}
|
|
|
|
void v9fs_uint32_write(P9Req *req, uint32_t val)
|
|
{
|
|
uint32_t le_val = cpu_to_le32(val);
|
|
|
|
v9fs_memwrite(req, &le_val, 4);
|
|
}
|
|
|
|
void v9fs_uint64_write(P9Req *req, uint64_t val)
|
|
{
|
|
uint64_t le_val = cpu_to_le64(val);
|
|
|
|
v9fs_memwrite(req, &le_val, 8);
|
|
}
|
|
|
|
void v9fs_uint32_read(P9Req *req, uint32_t *val)
|
|
{
|
|
v9fs_memread(req, val, 4);
|
|
le32_to_cpus(val);
|
|
}
|
|
|
|
void v9fs_uint64_read(P9Req *req, uint64_t *val)
|
|
{
|
|
v9fs_memread(req, val, 8);
|
|
le64_to_cpus(val);
|
|
}
|
|
|
|
/* len[2] string[len] */
|
|
uint16_t v9fs_string_size(const char *string)
|
|
{
|
|
size_t len = strlen(string);
|
|
|
|
g_assert_cmpint(len, <=, UINT16_MAX - 2);
|
|
|
|
return 2 + len;
|
|
}
|
|
|
|
void v9fs_string_write(P9Req *req, const char *string)
|
|
{
|
|
int len = strlen(string);
|
|
|
|
g_assert_cmpint(len, <=, UINT16_MAX);
|
|
|
|
v9fs_uint16_write(req, (uint16_t) len);
|
|
v9fs_memwrite(req, string, len);
|
|
}
|
|
|
|
void v9fs_string_read(P9Req *req, uint16_t *len, char **string)
|
|
{
|
|
uint16_t local_len;
|
|
|
|
v9fs_uint16_read(req, &local_len);
|
|
if (len) {
|
|
*len = local_len;
|
|
}
|
|
if (string) {
|
|
*string = g_malloc(local_len + 1);
|
|
v9fs_memread(req, *string, local_len);
|
|
(*string)[local_len] = 0;
|
|
} else {
|
|
v9fs_memskip(req, local_len);
|
|
}
|
|
}
|
|
|
|
typedef struct {
|
|
uint32_t size;
|
|
uint8_t id;
|
|
uint16_t tag;
|
|
} QEMU_PACKED P9Hdr;
|
|
|
|
P9Req *v9fs_req_init(QVirtio9P *v9p, uint32_t size, uint8_t id,
|
|
uint16_t tag)
|
|
{
|
|
P9Req *req = g_new0(P9Req, 1);
|
|
uint32_t total_size = 7; /* 9P header has well-known size of 7 bytes */
|
|
P9Hdr hdr = {
|
|
.id = id,
|
|
.tag = cpu_to_le16(tag)
|
|
};
|
|
|
|
g_assert_cmpint(total_size, <=, UINT32_MAX - size);
|
|
total_size += size;
|
|
hdr.size = cpu_to_le32(total_size);
|
|
|
|
g_assert_cmpint(total_size, <=, P9_MAX_SIZE);
|
|
|
|
req->qts = global_qtest;
|
|
req->v9p = v9p;
|
|
req->t_size = total_size;
|
|
req->t_msg = guest_alloc(alloc, req->t_size);
|
|
v9fs_memwrite(req, &hdr, 7);
|
|
req->tag = tag;
|
|
return req;
|
|
}
|
|
|
|
void v9fs_req_send(P9Req *req)
|
|
{
|
|
QVirtio9P *v9p = req->v9p;
|
|
|
|
req->r_msg = guest_alloc(alloc, P9_MAX_SIZE);
|
|
req->free_head = qvirtqueue_add(req->qts, v9p->vq, req->t_msg, req->t_size,
|
|
false, true);
|
|
qvirtqueue_add(req->qts, v9p->vq, req->r_msg, P9_MAX_SIZE, true, false);
|
|
qvirtqueue_kick(req->qts, v9p->vdev, v9p->vq, req->free_head);
|
|
req->t_off = 0;
|
|
}
|
|
|
|
static const char *rmessage_name(uint8_t id)
|
|
{
|
|
return
|
|
id == P9_RLERROR ? "RLERROR" :
|
|
id == P9_RVERSION ? "RVERSION" :
|
|
id == P9_RATTACH ? "RATTACH" :
|
|
id == P9_RWALK ? "RWALK" :
|
|
id == P9_RLOPEN ? "RLOPEN" :
|
|
id == P9_RWRITE ? "RWRITE" :
|
|
id == P9_RMKDIR ? "RMKDIR" :
|
|
id == P9_RLCREATE ? "RLCREATE" :
|
|
id == P9_RSYMLINK ? "RSYMLINK" :
|
|
id == P9_RLINK ? "RLINK" :
|
|
id == P9_RUNLINKAT ? "RUNLINKAT" :
|
|
id == P9_RFLUSH ? "RFLUSH" :
|
|
id == P9_RREADDIR ? "READDIR" :
|
|
"<unknown>";
|
|
}
|
|
|
|
void v9fs_req_wait_for_reply(P9Req *req, uint32_t *len)
|
|
{
|
|
QVirtio9P *v9p = req->v9p;
|
|
|
|
qvirtio_wait_used_elem(req->qts, v9p->vdev, v9p->vq, req->free_head, len,
|
|
QVIRTIO_9P_TIMEOUT_US);
|
|
}
|
|
|
|
void v9fs_req_recv(P9Req *req, uint8_t id)
|
|
{
|
|
P9Hdr hdr;
|
|
|
|
v9fs_memread(req, &hdr, 7);
|
|
hdr.size = ldl_le_p(&hdr.size);
|
|
hdr.tag = lduw_le_p(&hdr.tag);
|
|
|
|
g_assert_cmpint(hdr.size, >=, 7);
|
|
g_assert_cmpint(hdr.size, <=, P9_MAX_SIZE);
|
|
g_assert_cmpint(hdr.tag, ==, req->tag);
|
|
|
|
if (hdr.id != id) {
|
|
g_printerr("Received response %d (%s) instead of %d (%s)\n",
|
|
hdr.id, rmessage_name(hdr.id), id, rmessage_name(id));
|
|
|
|
if (hdr.id == P9_RLERROR) {
|
|
uint32_t err;
|
|
v9fs_uint32_read(req, &err);
|
|
g_printerr("Rlerror has errno %d (%s)\n", err, strerror(err));
|
|
}
|
|
}
|
|
g_assert_cmpint(hdr.id, ==, id);
|
|
}
|
|
|
|
void v9fs_req_free(P9Req *req)
|
|
{
|
|
guest_free(alloc, req->t_msg);
|
|
guest_free(alloc, req->r_msg);
|
|
g_free(req);
|
|
}
|
|
|
|
/* size[4] Rlerror tag[2] ecode[4] */
|
|
void v9fs_rlerror(P9Req *req, uint32_t *err)
|
|
{
|
|
v9fs_req_recv(req, P9_RLERROR);
|
|
v9fs_uint32_read(req, err);
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
/* size[4] Tversion tag[2] msize[4] version[s] */
|
|
TVersionRes v9fs_tversion(TVersionOpt opt)
|
|
{
|
|
P9Req *req;
|
|
uint32_t err;
|
|
uint32_t body_size = 4;
|
|
uint16_t string_size;
|
|
uint16_t server_len;
|
|
g_autofree char *server_version = NULL;
|
|
|
|
g_assert(opt.client);
|
|
|
|
if (!opt.msize) {
|
|
opt.msize = P9_MAX_SIZE;
|
|
}
|
|
|
|
if (!opt.tag) {
|
|
opt.tag = P9_NOTAG;
|
|
}
|
|
|
|
if (!opt.version) {
|
|
opt.version = "9P2000.L";
|
|
}
|
|
|
|
string_size = v9fs_string_size(opt.version);
|
|
g_assert_cmpint(body_size, <=, UINT32_MAX - string_size);
|
|
body_size += string_size;
|
|
req = v9fs_req_init(opt.client, body_size, P9_TVERSION, opt.tag);
|
|
|
|
v9fs_uint32_write(req, opt.msize);
|
|
v9fs_string_write(req, opt.version);
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_rversion(req, &server_len, &server_version);
|
|
g_assert_cmpmem(server_version, server_len,
|
|
opt.version, strlen(opt.version));
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
return (TVersionRes) {
|
|
.req = req,
|
|
};
|
|
}
|
|
|
|
/* size[4] Rversion tag[2] msize[4] version[s] */
|
|
void v9fs_rversion(P9Req *req, uint16_t *len, char **version)
|
|
{
|
|
uint32_t msize;
|
|
|
|
v9fs_req_recv(req, P9_RVERSION);
|
|
v9fs_uint32_read(req, &msize);
|
|
|
|
g_assert_cmpint(msize, ==, P9_MAX_SIZE);
|
|
|
|
if (len || version) {
|
|
v9fs_string_read(req, len, version);
|
|
}
|
|
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
/* size[4] Tattach tag[2] fid[4] afid[4] uname[s] aname[s] n_uname[4] */
|
|
TAttachRes v9fs_tattach(TAttachOpt opt)
|
|
{
|
|
uint32_t err;
|
|
const char *uname = ""; /* ignored by QEMU */
|
|
const char *aname = ""; /* ignored by QEMU */
|
|
|
|
g_assert(opt.client);
|
|
/* expecting either Rattach or Rlerror, but obviously not both */
|
|
g_assert(!opt.expectErr || !opt.rattach.qid);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_tversion((TVersionOpt) { .client = opt.client });
|
|
}
|
|
|
|
if (!opt.n_uname) {
|
|
opt.n_uname = getuid();
|
|
}
|
|
|
|
P9Req *req = v9fs_req_init(opt.client, 4 + 4 + 2 + 2 + 4, P9_TATTACH,
|
|
opt.tag);
|
|
|
|
v9fs_uint32_write(req, opt.fid);
|
|
v9fs_uint32_write(req, P9_NOFID);
|
|
v9fs_string_write(req, uname);
|
|
v9fs_string_write(req, aname);
|
|
v9fs_uint32_write(req, opt.n_uname);
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_rattach(req, opt.rattach.qid);
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
return (TAttachRes) {
|
|
.req = req,
|
|
};
|
|
}
|
|
|
|
/* size[4] Rattach tag[2] qid[13] */
|
|
void v9fs_rattach(P9Req *req, v9fs_qid *qid)
|
|
{
|
|
v9fs_req_recv(req, P9_RATTACH);
|
|
if (qid) {
|
|
v9fs_memread(req, qid, 13);
|
|
}
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
/* size[4] Twalk tag[2] fid[4] newfid[4] nwname[2] nwname*(wname[s]) */
|
|
TWalkRes v9fs_twalk(TWalkOpt opt)
|
|
{
|
|
P9Req *req;
|
|
int i;
|
|
uint32_t body_size = 4 + 4 + 2;
|
|
uint32_t err;
|
|
char **wnames = NULL;
|
|
|
|
g_assert(opt.client);
|
|
/* expecting either high- or low-level path, both not both */
|
|
g_assert(!opt.path || !(opt.nwname || opt.wnames));
|
|
/* expecting either Rwalk or Rlerror, but obviously not both */
|
|
g_assert(!opt.expectErr || !(opt.rwalk.nwqid || opt.rwalk.wqid));
|
|
|
|
if (!opt.newfid) {
|
|
opt.newfid = genfid();
|
|
}
|
|
|
|
if (opt.path) {
|
|
opt.nwname = split(opt.path, "/", &wnames);
|
|
opt.wnames = wnames;
|
|
}
|
|
|
|
for (i = 0; i < opt.nwname; i++) {
|
|
uint16_t wname_size = v9fs_string_size(opt.wnames[i]);
|
|
|
|
g_assert_cmpint(body_size, <=, UINT32_MAX - wname_size);
|
|
body_size += wname_size;
|
|
}
|
|
req = v9fs_req_init(opt.client, body_size, P9_TWALK, opt.tag);
|
|
v9fs_uint32_write(req, opt.fid);
|
|
v9fs_uint32_write(req, opt.newfid);
|
|
v9fs_uint16_write(req, opt.nwname);
|
|
for (i = 0; i < opt.nwname; i++) {
|
|
v9fs_string_write(req, opt.wnames[i]);
|
|
}
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_rwalk(req, opt.rwalk.nwqid, opt.rwalk.wqid);
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
split_free(&wnames);
|
|
|
|
return (TWalkRes) {
|
|
.newfid = opt.newfid,
|
|
.req = req,
|
|
};
|
|
}
|
|
|
|
/* size[4] Rwalk tag[2] nwqid[2] nwqid*(wqid[13]) */
|
|
void v9fs_rwalk(P9Req *req, uint16_t *nwqid, v9fs_qid **wqid)
|
|
{
|
|
uint16_t local_nwqid;
|
|
|
|
v9fs_req_recv(req, P9_RWALK);
|
|
v9fs_uint16_read(req, &local_nwqid);
|
|
if (nwqid) {
|
|
*nwqid = local_nwqid;
|
|
}
|
|
if (wqid) {
|
|
*wqid = g_malloc(local_nwqid * 13);
|
|
v9fs_memread(req, *wqid, local_nwqid * 13);
|
|
}
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
/* size[4] Tgetattr tag[2] fid[4] request_mask[8] */
|
|
TGetAttrRes v9fs_tgetattr(TGetAttrOpt opt)
|
|
{
|
|
P9Req *req;
|
|
uint32_t err;
|
|
|
|
g_assert(opt.client);
|
|
/* expecting either Rgetattr or Rlerror, but obviously not both */
|
|
g_assert(!opt.expectErr || !opt.rgetattr.attr);
|
|
|
|
if (!opt.request_mask) {
|
|
opt.request_mask = P9_GETATTR_ALL;
|
|
}
|
|
|
|
req = v9fs_req_init(opt.client, 4 + 8, P9_TGETATTR, opt.tag);
|
|
v9fs_uint32_write(req, opt.fid);
|
|
v9fs_uint64_write(req, opt.request_mask);
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_rgetattr(req, opt.rgetattr.attr);
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
return (TGetAttrRes) { .req = req };
|
|
}
|
|
|
|
/*
|
|
* size[4] Rgetattr tag[2] valid[8] qid[13] mode[4] uid[4] gid[4] nlink[8]
|
|
* rdev[8] size[8] blksize[8] blocks[8]
|
|
* atime_sec[8] atime_nsec[8] mtime_sec[8] mtime_nsec[8]
|
|
* ctime_sec[8] ctime_nsec[8] btime_sec[8] btime_nsec[8]
|
|
* gen[8] data_version[8]
|
|
*/
|
|
void v9fs_rgetattr(P9Req *req, v9fs_attr *attr)
|
|
{
|
|
v9fs_req_recv(req, P9_RGETATTR);
|
|
|
|
v9fs_uint64_read(req, &attr->valid);
|
|
v9fs_memread(req, &attr->qid, 13);
|
|
v9fs_uint32_read(req, &attr->mode);
|
|
v9fs_uint32_read(req, &attr->uid);
|
|
v9fs_uint32_read(req, &attr->gid);
|
|
v9fs_uint64_read(req, &attr->nlink);
|
|
v9fs_uint64_read(req, &attr->rdev);
|
|
v9fs_uint64_read(req, &attr->size);
|
|
v9fs_uint64_read(req, &attr->blksize);
|
|
v9fs_uint64_read(req, &attr->blocks);
|
|
v9fs_uint64_read(req, &attr->atime_sec);
|
|
v9fs_uint64_read(req, &attr->atime_nsec);
|
|
v9fs_uint64_read(req, &attr->mtime_sec);
|
|
v9fs_uint64_read(req, &attr->mtime_nsec);
|
|
v9fs_uint64_read(req, &attr->ctime_sec);
|
|
v9fs_uint64_read(req, &attr->ctime_nsec);
|
|
v9fs_uint64_read(req, &attr->btime_sec);
|
|
v9fs_uint64_read(req, &attr->btime_nsec);
|
|
v9fs_uint64_read(req, &attr->gen);
|
|
v9fs_uint64_read(req, &attr->data_version);
|
|
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
/* size[4] Treaddir tag[2] fid[4] offset[8] count[4] */
|
|
TReadDirRes v9fs_treaddir(TReadDirOpt opt)
|
|
{
|
|
P9Req *req;
|
|
uint32_t err;
|
|
|
|
g_assert(opt.client);
|
|
/* expecting either Rreaddir or Rlerror, but obviously not both */
|
|
g_assert(!opt.expectErr || !(opt.rreaddir.count ||
|
|
opt.rreaddir.nentries || opt.rreaddir.entries));
|
|
|
|
req = v9fs_req_init(opt.client, 4 + 8 + 4, P9_TREADDIR, opt.tag);
|
|
v9fs_uint32_write(req, opt.fid);
|
|
v9fs_uint64_write(req, opt.offset);
|
|
v9fs_uint32_write(req, opt.count);
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_rreaddir(req, opt.rreaddir.count, opt.rreaddir.nentries,
|
|
opt.rreaddir.entries);
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
return (TReadDirRes) { .req = req };
|
|
}
|
|
|
|
/* size[4] Rreaddir tag[2] count[4] data[count] */
|
|
void v9fs_rreaddir(P9Req *req, uint32_t *count, uint32_t *nentries,
|
|
struct V9fsDirent **entries)
|
|
{
|
|
uint32_t local_count;
|
|
struct V9fsDirent *e = NULL;
|
|
uint16_t slen;
|
|
uint32_t n = 0;
|
|
|
|
v9fs_req_recv(req, P9_RREADDIR);
|
|
v9fs_uint32_read(req, &local_count);
|
|
|
|
if (count) {
|
|
*count = local_count;
|
|
}
|
|
|
|
for (int32_t togo = (int32_t)local_count;
|
|
togo >= 13 + 8 + 1 + 2;
|
|
togo -= 13 + 8 + 1 + 2 + slen, ++n)
|
|
{
|
|
if (!e) {
|
|
e = g_new(struct V9fsDirent, 1);
|
|
if (entries) {
|
|
*entries = e;
|
|
}
|
|
} else {
|
|
e = e->next = g_new(struct V9fsDirent, 1);
|
|
}
|
|
e->next = NULL;
|
|
/* qid[13] offset[8] type[1] name[s] */
|
|
v9fs_memread(req, &e->qid, 13);
|
|
v9fs_uint64_read(req, &e->offset);
|
|
v9fs_uint8_read(req, &e->type);
|
|
v9fs_string_read(req, &slen, &e->name);
|
|
}
|
|
|
|
if (nentries) {
|
|
*nentries = n;
|
|
}
|
|
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
void v9fs_free_dirents(struct V9fsDirent *e)
|
|
{
|
|
struct V9fsDirent *next = NULL;
|
|
|
|
for (; e; e = next) {
|
|
next = e->next;
|
|
g_free(e->name);
|
|
g_free(e);
|
|
}
|
|
}
|
|
|
|
/* size[4] Tlopen tag[2] fid[4] flags[4] */
|
|
TLOpenRes v9fs_tlopen(TLOpenOpt opt)
|
|
{
|
|
P9Req *req;
|
|
uint32_t err;
|
|
|
|
g_assert(opt.client);
|
|
/* expecting either Rlopen or Rlerror, but obviously not both */
|
|
g_assert(!opt.expectErr || !(opt.rlopen.qid || opt.rlopen.iounit));
|
|
|
|
req = v9fs_req_init(opt.client, 4 + 4, P9_TLOPEN, opt.tag);
|
|
v9fs_uint32_write(req, opt.fid);
|
|
v9fs_uint32_write(req, opt.flags);
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_rlopen(req, opt.rlopen.qid, opt.rlopen.iounit);
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
return (TLOpenRes) { .req = req };
|
|
}
|
|
|
|
/* size[4] Rlopen tag[2] qid[13] iounit[4] */
|
|
void v9fs_rlopen(P9Req *req, v9fs_qid *qid, uint32_t *iounit)
|
|
{
|
|
v9fs_req_recv(req, P9_RLOPEN);
|
|
if (qid) {
|
|
v9fs_memread(req, qid, 13);
|
|
} else {
|
|
v9fs_memskip(req, 13);
|
|
}
|
|
if (iounit) {
|
|
v9fs_uint32_read(req, iounit);
|
|
}
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
/* size[4] Twrite tag[2] fid[4] offset[8] count[4] data[count] */
|
|
TWriteRes v9fs_twrite(TWriteOpt opt)
|
|
{
|
|
P9Req *req;
|
|
uint32_t err;
|
|
uint32_t body_size = 4 + 8 + 4;
|
|
uint32_t written = 0;
|
|
|
|
g_assert(opt.client);
|
|
|
|
g_assert_cmpint(body_size, <=, UINT32_MAX - opt.count);
|
|
body_size += opt.count;
|
|
req = v9fs_req_init(opt.client, body_size, P9_TWRITE, opt.tag);
|
|
v9fs_uint32_write(req, opt.fid);
|
|
v9fs_uint64_write(req, opt.offset);
|
|
v9fs_uint32_write(req, opt.count);
|
|
v9fs_memwrite(req, opt.data, opt.count);
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_rwrite(req, &written);
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
return (TWriteRes) {
|
|
.req = req,
|
|
.count = written
|
|
};
|
|
}
|
|
|
|
/* size[4] Rwrite tag[2] count[4] */
|
|
void v9fs_rwrite(P9Req *req, uint32_t *count)
|
|
{
|
|
v9fs_req_recv(req, P9_RWRITE);
|
|
if (count) {
|
|
v9fs_uint32_read(req, count);
|
|
}
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
/* size[4] Tflush tag[2] oldtag[2] */
|
|
TFlushRes v9fs_tflush(TFlushOpt opt)
|
|
{
|
|
P9Req *req;
|
|
uint32_t err;
|
|
|
|
g_assert(opt.client);
|
|
|
|
req = v9fs_req_init(opt.client, 2, P9_TFLUSH, opt.tag);
|
|
v9fs_uint32_write(req, opt.oldtag);
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_rflush(req);
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
return (TFlushRes) { .req = req };
|
|
}
|
|
|
|
/* size[4] Rflush tag[2] */
|
|
void v9fs_rflush(P9Req *req)
|
|
{
|
|
v9fs_req_recv(req, P9_RFLUSH);
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
/* size[4] Tmkdir tag[2] dfid[4] name[s] mode[4] gid[4] */
|
|
TMkdirRes v9fs_tmkdir(TMkdirOpt opt)
|
|
{
|
|
P9Req *req;
|
|
uint32_t err;
|
|
|
|
g_assert(opt.client);
|
|
/* expecting either hi-level atPath or low-level dfid, but not both */
|
|
g_assert(!opt.atPath || !opt.dfid);
|
|
/* expecting either Rmkdir or Rlerror, but obviously not both */
|
|
g_assert(!opt.expectErr || !opt.rmkdir.qid);
|
|
|
|
if (opt.atPath) {
|
|
opt.dfid = v9fs_twalk((TWalkOpt) { .client = opt.client,
|
|
.path = opt.atPath }).newfid;
|
|
}
|
|
|
|
if (!opt.mode) {
|
|
opt.mode = 0750;
|
|
}
|
|
|
|
uint32_t body_size = 4 + 4 + 4;
|
|
uint16_t string_size = v9fs_string_size(opt.name);
|
|
|
|
g_assert_cmpint(body_size, <=, UINT32_MAX - string_size);
|
|
body_size += string_size;
|
|
|
|
req = v9fs_req_init(opt.client, body_size, P9_TMKDIR, opt.tag);
|
|
v9fs_uint32_write(req, opt.dfid);
|
|
v9fs_string_write(req, opt.name);
|
|
v9fs_uint32_write(req, opt.mode);
|
|
v9fs_uint32_write(req, opt.gid);
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_rmkdir(req, opt.rmkdir.qid);
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
return (TMkdirRes) { .req = req };
|
|
}
|
|
|
|
/* size[4] Rmkdir tag[2] qid[13] */
|
|
void v9fs_rmkdir(P9Req *req, v9fs_qid *qid)
|
|
{
|
|
v9fs_req_recv(req, P9_RMKDIR);
|
|
if (qid) {
|
|
v9fs_memread(req, qid, 13);
|
|
} else {
|
|
v9fs_memskip(req, 13);
|
|
}
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
/* size[4] Tlcreate tag[2] fid[4] name[s] flags[4] mode[4] gid[4] */
|
|
TlcreateRes v9fs_tlcreate(TlcreateOpt opt)
|
|
{
|
|
P9Req *req;
|
|
uint32_t err;
|
|
|
|
g_assert(opt.client);
|
|
/* expecting either hi-level atPath or low-level fid, but not both */
|
|
g_assert(!opt.atPath || !opt.fid);
|
|
/* expecting either Rlcreate or Rlerror, but obviously not both */
|
|
g_assert(!opt.expectErr || !(opt.rlcreate.qid || opt.rlcreate.iounit));
|
|
|
|
if (opt.atPath) {
|
|
opt.fid = v9fs_twalk((TWalkOpt) { .client = opt.client,
|
|
.path = opt.atPath }).newfid;
|
|
}
|
|
|
|
if (!opt.mode) {
|
|
opt.mode = 0750;
|
|
}
|
|
|
|
uint32_t body_size = 4 + 4 + 4 + 4;
|
|
uint16_t string_size = v9fs_string_size(opt.name);
|
|
|
|
g_assert_cmpint(body_size, <=, UINT32_MAX - string_size);
|
|
body_size += string_size;
|
|
|
|
req = v9fs_req_init(opt.client, body_size, P9_TLCREATE, opt.tag);
|
|
v9fs_uint32_write(req, opt.fid);
|
|
v9fs_string_write(req, opt.name);
|
|
v9fs_uint32_write(req, opt.flags);
|
|
v9fs_uint32_write(req, opt.mode);
|
|
v9fs_uint32_write(req, opt.gid);
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_rlcreate(req, opt.rlcreate.qid, opt.rlcreate.iounit);
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
return (TlcreateRes) { .req = req };
|
|
}
|
|
|
|
/* size[4] Rlcreate tag[2] qid[13] iounit[4] */
|
|
void v9fs_rlcreate(P9Req *req, v9fs_qid *qid, uint32_t *iounit)
|
|
{
|
|
v9fs_req_recv(req, P9_RLCREATE);
|
|
if (qid) {
|
|
v9fs_memread(req, qid, 13);
|
|
} else {
|
|
v9fs_memskip(req, 13);
|
|
}
|
|
if (iounit) {
|
|
v9fs_uint32_read(req, iounit);
|
|
}
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
/* size[4] Tsymlink tag[2] fid[4] name[s] symtgt[s] gid[4] */
|
|
TsymlinkRes v9fs_tsymlink(TsymlinkOpt opt)
|
|
{
|
|
P9Req *req;
|
|
uint32_t err;
|
|
|
|
g_assert(opt.client);
|
|
/* expecting either hi-level atPath or low-level fid, but not both */
|
|
g_assert(!opt.atPath || !opt.fid);
|
|
/* expecting either Rsymlink or Rlerror, but obviously not both */
|
|
g_assert(!opt.expectErr || !opt.rsymlink.qid);
|
|
|
|
if (opt.atPath) {
|
|
opt.fid = v9fs_twalk((TWalkOpt) { .client = opt.client,
|
|
.path = opt.atPath }).newfid;
|
|
}
|
|
|
|
uint32_t body_size = 4 + 4;
|
|
uint16_t string_size = v9fs_string_size(opt.name) +
|
|
v9fs_string_size(opt.symtgt);
|
|
|
|
g_assert_cmpint(body_size, <=, UINT32_MAX - string_size);
|
|
body_size += string_size;
|
|
|
|
req = v9fs_req_init(opt.client, body_size, P9_TSYMLINK, opt.tag);
|
|
v9fs_uint32_write(req, opt.fid);
|
|
v9fs_string_write(req, opt.name);
|
|
v9fs_string_write(req, opt.symtgt);
|
|
v9fs_uint32_write(req, opt.gid);
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_rsymlink(req, opt.rsymlink.qid);
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
return (TsymlinkRes) { .req = req };
|
|
}
|
|
|
|
/* size[4] Rsymlink tag[2] qid[13] */
|
|
void v9fs_rsymlink(P9Req *req, v9fs_qid *qid)
|
|
{
|
|
v9fs_req_recv(req, P9_RSYMLINK);
|
|
if (qid) {
|
|
v9fs_memread(req, qid, 13);
|
|
} else {
|
|
v9fs_memskip(req, 13);
|
|
}
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
/* size[4] Tlink tag[2] dfid[4] fid[4] name[s] */
|
|
TlinkRes v9fs_tlink(TlinkOpt opt)
|
|
{
|
|
P9Req *req;
|
|
uint32_t err;
|
|
|
|
g_assert(opt.client);
|
|
/* expecting either hi-level atPath or low-level dfid, but not both */
|
|
g_assert(!opt.atPath || !opt.dfid);
|
|
/* expecting either hi-level toPath or low-level fid, but not both */
|
|
g_assert(!opt.toPath || !opt.fid);
|
|
|
|
if (opt.atPath) {
|
|
opt.dfid = v9fs_twalk((TWalkOpt) { .client = opt.client,
|
|
.path = opt.atPath }).newfid;
|
|
}
|
|
if (opt.toPath) {
|
|
opt.fid = v9fs_twalk((TWalkOpt) { .client = opt.client,
|
|
.path = opt.toPath }).newfid;
|
|
}
|
|
|
|
uint32_t body_size = 4 + 4;
|
|
uint16_t string_size = v9fs_string_size(opt.name);
|
|
|
|
g_assert_cmpint(body_size, <=, UINT32_MAX - string_size);
|
|
body_size += string_size;
|
|
|
|
req = v9fs_req_init(opt.client, body_size, P9_TLINK, opt.tag);
|
|
v9fs_uint32_write(req, opt.dfid);
|
|
v9fs_uint32_write(req, opt.fid);
|
|
v9fs_string_write(req, opt.name);
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_rlink(req);
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
return (TlinkRes) { .req = req };
|
|
}
|
|
|
|
/* size[4] Rlink tag[2] */
|
|
void v9fs_rlink(P9Req *req)
|
|
{
|
|
v9fs_req_recv(req, P9_RLINK);
|
|
v9fs_req_free(req);
|
|
}
|
|
|
|
/* size[4] Tunlinkat tag[2] dirfd[4] name[s] flags[4] */
|
|
TunlinkatRes v9fs_tunlinkat(TunlinkatOpt opt)
|
|
{
|
|
P9Req *req;
|
|
uint32_t err;
|
|
|
|
g_assert(opt.client);
|
|
/* expecting either hi-level atPath or low-level dirfd, but not both */
|
|
g_assert(!opt.atPath || !opt.dirfd);
|
|
|
|
if (opt.atPath) {
|
|
opt.dirfd = v9fs_twalk((TWalkOpt) { .client = opt.client,
|
|
.path = opt.atPath }).newfid;
|
|
}
|
|
|
|
uint32_t body_size = 4 + 4;
|
|
uint16_t string_size = v9fs_string_size(opt.name);
|
|
|
|
g_assert_cmpint(body_size, <=, UINT32_MAX - string_size);
|
|
body_size += string_size;
|
|
|
|
req = v9fs_req_init(opt.client, body_size, P9_TUNLINKAT, opt.tag);
|
|
v9fs_uint32_write(req, opt.dirfd);
|
|
v9fs_string_write(req, opt.name);
|
|
v9fs_uint32_write(req, opt.flags);
|
|
v9fs_req_send(req);
|
|
|
|
if (!opt.requestOnly) {
|
|
v9fs_req_wait_for_reply(req, NULL);
|
|
if (opt.expectErr) {
|
|
v9fs_rlerror(req, &err);
|
|
g_assert_cmpint(err, ==, opt.expectErr);
|
|
} else {
|
|
v9fs_runlinkat(req);
|
|
}
|
|
req = NULL; /* request was freed */
|
|
}
|
|
|
|
return (TunlinkatRes) { .req = req };
|
|
}
|
|
|
|
/* size[4] Runlinkat tag[2] */
|
|
void v9fs_runlinkat(P9Req *req)
|
|
{
|
|
v9fs_req_recv(req, P9_RUNLINKAT);
|
|
v9fs_req_free(req);
|
|
}
|