mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2026-01-11 17:10:13 +00:00
The read result collection for buffered reads seems to run ahead of the
completion of subrequests under some circumstances, as can be seen in the
following log snippet:
9p_client_res: client 18446612686390831168 response P9_TREAD tag 0 err 0
...
netfs_sreq: R=00001b55[1] DOWN TERM f=192 s=0 5fb2/5fb2 s=5 e=0
...
netfs_collect_folio: R=00001b55 ix=00004 r=4000-5000 t=4000/5fb2
netfs_folio: i=157f3 ix=00004-00004 read-done
netfs_folio: i=157f3 ix=00004-00004 read-unlock
netfs_collect_folio: R=00001b55 ix=00005 r=5000-5fb2 t=5000/5fb2
netfs_folio: i=157f3 ix=00005-00005 read-done
netfs_folio: i=157f3 ix=00005-00005 read-unlock
...
netfs_collect_stream: R=00001b55[0:] cto=5fb2 frn=ffffffff
netfs_collect_state: R=00001b55 col=5fb2 cln=6000 n=c
netfs_collect_stream: R=00001b55[0:] cto=5fb2 frn=ffffffff
netfs_collect_state: R=00001b55 col=5fb2 cln=6000 n=8
...
netfs_sreq: R=00001b55[2] ZERO SUBMT f=000 s=5fb2 0/4e s=0 e=0
netfs_sreq: R=00001b55[2] ZERO TERM f=102 s=5fb2 4e/4e s=5 e=0
The 'cto=5fb2' indicates the collected file pos we've collected results to
so far - but we still have 0x4e more bytes to go - so we shouldn't have
collected folio ix=00005 yet. The 'ZERO' subreq that clears the tail
happens after we unlock the folio, allowing the application to see the
uncleared tail through mmap.
The problem is that netfs_read_unlock_folios() will unlock a folio in which
the amount of read results collected hits EOF position - but the ZERO
subreq lies beyond that and so happens after.
Fix this by changing the end check to always be the end of the folio and
never the end of the file.
In the future, I should look at clearing to the end of the folio here rather
than adding a ZERO subreq to do this. On the other hand, the ZERO subreq can
run in parallel with an async READ subreq. Further, the ZERO subreq may still
be necessary to, say, handle extents in a ceph file that don't have any
backing store and are thus implicitly all zeros.
This can be reproduced by creating a file, the size of which doesn't align
to a page boundary, e.g. 24998 (0x5fb2) bytes and then doing something
like:
xfs_io -c "mmap -r 0 0x6000" -c "madvise -d 0 0x6000" \
-c "mread -v 0 0x6000" /xfstest.test/x
The last 0x4e bytes should all be 00, but if the tail hasn't been cleared
yet, you may see rubbish there. This can be reproduced with kafs by
modifying the kernel to disable the call to netfs_read_subreq_progress()
and to stop afs_issue_read() from doing the async call for NETFS_READAHEAD.
Reproduction can be made easier by inserting an mdelay(100) in
netfs_issue_read() for the ZERO-subreq case.
AFS and CIFS are normally unlikely to show this as they dispatch READ ops
asynchronously, which allows the ZERO-subreq to finish first. 9P's READ op is
completely synchronous, so the ZERO-subreq will always happen after. It isn't
seen all the time, though, because the collection may be done in a worker
thread.
Reported-by: Christian Schoenebeck <linux_oss@crudebyte.com>
Link: https://lore.kernel.org/r/8622834.T7Z3S40VBb@weasel/
Signed-off-by: David Howells <dhowells@redhat.com>
Link: https://patch.msgid.link/938162.1766233900@warthog.procyon.org.uk
Fixes: e2d46f2ec332 ("netfs: Change the read result collector to only use one work item")
Tested-by: Christian Schoenebeck <linux_oss@crudebyte.com>
Acked-by: Dominique Martinet <asmadeus@codewreck.org>
Suggested-by: Dominique Martinet <asmadeus@codewreck.org>
cc: Dominique Martinet <asmadeus@codewreck.org>
cc: Christian Schoenebeck <linux_oss@crudebyte.com>
cc: v9fs@lists.linux.dev
cc: netfs@lists.linux.dev
cc: linux-fsdevel@vger.kernel.org
Signed-off-by: Christian Brauner <brauner@kernel.org>
586 lines
18 KiB
C
586 lines
18 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/* Network filesystem read subrequest result collection, assessment and
|
|
* retrying.
|
|
*
|
|
* Copyright (C) 2024 Red Hat, Inc. All Rights Reserved.
|
|
* Written by David Howells (dhowells@redhat.com)
|
|
*/
|
|
|
|
#include <linux/export.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/pagemap.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/task_io_accounting_ops.h>
|
|
#include "internal.h"
|
|
|
|
/* Notes made in the collector */
|
|
#define HIT_PENDING 0x01 /* A front op was still pending */
|
|
#define MADE_PROGRESS 0x04 /* Made progress cleaning up a stream or the folio set */
|
|
#define BUFFERED 0x08 /* The pagecache needs cleaning up */
|
|
#define NEED_RETRY 0x10 /* A front op requests retrying */
|
|
#define COPY_TO_CACHE 0x40 /* Need to copy subrequest to cache */
|
|
#define ABANDON_SREQ 0x80 /* Need to abandon untransferred part of subrequest */
|
|
|
|
/*
|
|
* Clear the unread part of an I/O request.
|
|
*/
|
|
static void netfs_clear_unread(struct netfs_io_subrequest *subreq)
|
|
{
|
|
netfs_reset_iter(subreq);
|
|
WARN_ON_ONCE(subreq->len - subreq->transferred != iov_iter_count(&subreq->io_iter));
|
|
iov_iter_zero(iov_iter_count(&subreq->io_iter), &subreq->io_iter);
|
|
if (subreq->start + subreq->transferred >= subreq->rreq->i_size)
|
|
__set_bit(NETFS_SREQ_HIT_EOF, &subreq->flags);
|
|
}
|
|
|
|
/*
|
|
* Flush, mark and unlock a folio that's now completely read. If we want to
|
|
* cache the folio, we set the group to NETFS_FOLIO_COPY_TO_CACHE, mark it
|
|
* dirty and let writeback handle it.
|
|
*/
|
|
static void netfs_unlock_read_folio(struct netfs_io_request *rreq,
|
|
struct folio_queue *folioq,
|
|
int slot)
|
|
{
|
|
struct netfs_folio *finfo;
|
|
struct folio *folio = folioq_folio(folioq, slot);
|
|
|
|
if (unlikely(folio_pos(folio) < rreq->abandon_to)) {
|
|
trace_netfs_folio(folio, netfs_folio_trace_abandon);
|
|
goto just_unlock;
|
|
}
|
|
|
|
flush_dcache_folio(folio);
|
|
folio_mark_uptodate(folio);
|
|
|
|
if (!test_bit(NETFS_RREQ_USE_PGPRIV2, &rreq->flags)) {
|
|
finfo = netfs_folio_info(folio);
|
|
if (finfo) {
|
|
trace_netfs_folio(folio, netfs_folio_trace_filled_gaps);
|
|
if (finfo->netfs_group)
|
|
folio_change_private(folio, finfo->netfs_group);
|
|
else
|
|
folio_detach_private(folio);
|
|
kfree(finfo);
|
|
}
|
|
|
|
if (test_bit(NETFS_RREQ_FOLIO_COPY_TO_CACHE, &rreq->flags)) {
|
|
if (!WARN_ON_ONCE(folio_get_private(folio) != NULL)) {
|
|
trace_netfs_folio(folio, netfs_folio_trace_copy_to_cache);
|
|
folio_attach_private(folio, NETFS_FOLIO_COPY_TO_CACHE);
|
|
folio_mark_dirty(folio);
|
|
}
|
|
} else {
|
|
trace_netfs_folio(folio, netfs_folio_trace_read_done);
|
|
}
|
|
|
|
folioq_clear(folioq, slot);
|
|
} else {
|
|
// TODO: Use of PG_private_2 is deprecated.
|
|
if (test_bit(NETFS_RREQ_FOLIO_COPY_TO_CACHE, &rreq->flags))
|
|
netfs_pgpriv2_copy_to_cache(rreq, folio);
|
|
}
|
|
|
|
just_unlock:
|
|
if (folio->index == rreq->no_unlock_folio &&
|
|
test_bit(NETFS_RREQ_NO_UNLOCK_FOLIO, &rreq->flags)) {
|
|
_debug("no unlock");
|
|
} else {
|
|
trace_netfs_folio(folio, netfs_folio_trace_read_unlock);
|
|
folio_unlock(folio);
|
|
}
|
|
|
|
folioq_clear(folioq, slot);
|
|
}
|
|
|
|
/*
|
|
* Unlock any folios we've finished with.
|
|
*/
|
|
static void netfs_read_unlock_folios(struct netfs_io_request *rreq,
|
|
unsigned int *notes)
|
|
{
|
|
struct folio_queue *folioq = rreq->buffer.tail;
|
|
unsigned long long collected_to = rreq->collected_to;
|
|
unsigned int slot = rreq->buffer.first_tail_slot;
|
|
|
|
if (rreq->cleaned_to >= rreq->collected_to)
|
|
return;
|
|
|
|
// TODO: Begin decryption
|
|
|
|
if (slot >= folioq_nr_slots(folioq)) {
|
|
folioq = rolling_buffer_delete_spent(&rreq->buffer);
|
|
if (!folioq) {
|
|
rreq->front_folio_order = 0;
|
|
return;
|
|
}
|
|
slot = 0;
|
|
}
|
|
|
|
for (;;) {
|
|
struct folio *folio;
|
|
unsigned long long fpos, fend;
|
|
unsigned int order;
|
|
size_t fsize;
|
|
|
|
if (*notes & COPY_TO_CACHE)
|
|
set_bit(NETFS_RREQ_FOLIO_COPY_TO_CACHE, &rreq->flags);
|
|
|
|
folio = folioq_folio(folioq, slot);
|
|
if (WARN_ONCE(!folio_test_locked(folio),
|
|
"R=%08x: folio %lx is not locked\n",
|
|
rreq->debug_id, folio->index))
|
|
trace_netfs_folio(folio, netfs_folio_trace_not_locked);
|
|
|
|
order = folioq_folio_order(folioq, slot);
|
|
rreq->front_folio_order = order;
|
|
fsize = PAGE_SIZE << order;
|
|
fpos = folio_pos(folio);
|
|
fend = fpos + fsize;
|
|
|
|
trace_netfs_collect_folio(rreq, folio, fend, collected_to);
|
|
|
|
/* Unlock any folio we've transferred all of. */
|
|
if (collected_to < fend)
|
|
break;
|
|
|
|
netfs_unlock_read_folio(rreq, folioq, slot);
|
|
WRITE_ONCE(rreq->cleaned_to, fpos + fsize);
|
|
*notes |= MADE_PROGRESS;
|
|
|
|
clear_bit(NETFS_RREQ_FOLIO_COPY_TO_CACHE, &rreq->flags);
|
|
|
|
/* Clean up the head folioq. If we clear an entire folioq, then
|
|
* we can get rid of it provided it's not also the tail folioq
|
|
* being filled by the issuer.
|
|
*/
|
|
folioq_clear(folioq, slot);
|
|
slot++;
|
|
if (slot >= folioq_nr_slots(folioq)) {
|
|
folioq = rolling_buffer_delete_spent(&rreq->buffer);
|
|
if (!folioq)
|
|
goto done;
|
|
slot = 0;
|
|
trace_netfs_folioq(folioq, netfs_trace_folioq_read_progress);
|
|
}
|
|
|
|
if (fpos + fsize >= collected_to)
|
|
break;
|
|
}
|
|
|
|
rreq->buffer.tail = folioq;
|
|
done:
|
|
rreq->buffer.first_tail_slot = slot;
|
|
}
|
|
|
|
/*
|
|
* Collect and assess the results of various read subrequests. We may need to
|
|
* retry some of the results.
|
|
*
|
|
* Note that we have a sequence of subrequests, which may be drawing on
|
|
* different sources and may or may not be the same size or starting position
|
|
* and may not even correspond in boundary alignment.
|
|
*/
|
|
static void netfs_collect_read_results(struct netfs_io_request *rreq)
|
|
{
|
|
struct netfs_io_subrequest *front, *remove;
|
|
struct netfs_io_stream *stream = &rreq->io_streams[0];
|
|
unsigned int notes;
|
|
|
|
_enter("%llx-%llx", rreq->start, rreq->start + rreq->len);
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_collect);
|
|
trace_netfs_collect(rreq);
|
|
|
|
reassess:
|
|
if (rreq->origin == NETFS_READAHEAD ||
|
|
rreq->origin == NETFS_READPAGE ||
|
|
rreq->origin == NETFS_READ_FOR_WRITE)
|
|
notes = BUFFERED;
|
|
else
|
|
notes = 0;
|
|
|
|
/* Remove completed subrequests from the front of the stream and
|
|
* advance the completion point. We stop when we hit something that's
|
|
* in progress. The issuer thread may be adding stuff to the tail
|
|
* whilst we're doing this.
|
|
*/
|
|
front = READ_ONCE(stream->front);
|
|
while (front) {
|
|
size_t transferred;
|
|
|
|
trace_netfs_collect_sreq(rreq, front);
|
|
_debug("sreq [%x] %llx %zx/%zx",
|
|
front->debug_index, front->start, front->transferred, front->len);
|
|
|
|
if (stream->collected_to < front->start) {
|
|
trace_netfs_collect_gap(rreq, stream, front->start, 'F');
|
|
stream->collected_to = front->start;
|
|
}
|
|
|
|
if (netfs_check_subreq_in_progress(front))
|
|
notes |= HIT_PENDING;
|
|
smp_rmb(); /* Read counters after IN_PROGRESS flag. */
|
|
transferred = READ_ONCE(front->transferred);
|
|
|
|
/* If we can now collect the next folio, do so. We don't want
|
|
* to defer this as we have to decide whether we need to copy
|
|
* to the cache or not, and that may differ between adjacent
|
|
* subreqs.
|
|
*/
|
|
if (notes & BUFFERED) {
|
|
size_t fsize = PAGE_SIZE << rreq->front_folio_order;
|
|
|
|
/* Clear the tail of a short read. */
|
|
if (!(notes & HIT_PENDING) &&
|
|
front->error == 0 &&
|
|
transferred < front->len &&
|
|
(test_bit(NETFS_SREQ_HIT_EOF, &front->flags) ||
|
|
test_bit(NETFS_SREQ_CLEAR_TAIL, &front->flags))) {
|
|
netfs_clear_unread(front);
|
|
transferred = front->transferred = front->len;
|
|
trace_netfs_sreq(front, netfs_sreq_trace_clear);
|
|
}
|
|
|
|
stream->collected_to = front->start + transferred;
|
|
rreq->collected_to = stream->collected_to;
|
|
|
|
if (test_bit(NETFS_SREQ_COPY_TO_CACHE, &front->flags))
|
|
notes |= COPY_TO_CACHE;
|
|
|
|
if (test_bit(NETFS_SREQ_FAILED, &front->flags)) {
|
|
rreq->abandon_to = front->start + front->len;
|
|
front->transferred = front->len;
|
|
transferred = front->len;
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_set_abandon);
|
|
}
|
|
if (front->start + transferred >= rreq->cleaned_to + fsize ||
|
|
test_bit(NETFS_SREQ_HIT_EOF, &front->flags))
|
|
netfs_read_unlock_folios(rreq, ¬es);
|
|
} else {
|
|
stream->collected_to = front->start + transferred;
|
|
rreq->collected_to = stream->collected_to;
|
|
}
|
|
|
|
/* Stall if the front is still undergoing I/O. */
|
|
if (notes & HIT_PENDING)
|
|
break;
|
|
|
|
if (test_bit(NETFS_SREQ_FAILED, &front->flags)) {
|
|
if (!stream->failed) {
|
|
stream->error = front->error;
|
|
rreq->error = front->error;
|
|
set_bit(NETFS_RREQ_FAILED, &rreq->flags);
|
|
stream->failed = true;
|
|
}
|
|
notes |= MADE_PROGRESS | ABANDON_SREQ;
|
|
} else if (test_bit(NETFS_SREQ_NEED_RETRY, &front->flags)) {
|
|
stream->need_retry = true;
|
|
notes |= NEED_RETRY | MADE_PROGRESS;
|
|
break;
|
|
} else if (test_bit(NETFS_RREQ_SHORT_TRANSFER, &rreq->flags)) {
|
|
notes |= MADE_PROGRESS;
|
|
} else {
|
|
if (!stream->failed) {
|
|
stream->transferred += transferred;
|
|
stream->transferred_valid = true;
|
|
}
|
|
if (front->transferred < front->len)
|
|
set_bit(NETFS_RREQ_SHORT_TRANSFER, &rreq->flags);
|
|
notes |= MADE_PROGRESS;
|
|
}
|
|
|
|
/* Remove if completely consumed. */
|
|
stream->source = front->source;
|
|
spin_lock(&rreq->lock);
|
|
|
|
remove = front;
|
|
trace_netfs_sreq(front,
|
|
notes & ABANDON_SREQ ?
|
|
netfs_sreq_trace_abandoned : netfs_sreq_trace_consumed);
|
|
list_del_init(&front->rreq_link);
|
|
front = list_first_entry_or_null(&stream->subrequests,
|
|
struct netfs_io_subrequest, rreq_link);
|
|
stream->front = front;
|
|
spin_unlock(&rreq->lock);
|
|
netfs_put_subrequest(remove,
|
|
notes & ABANDON_SREQ ?
|
|
netfs_sreq_trace_put_abandon :
|
|
netfs_sreq_trace_put_done);
|
|
}
|
|
|
|
trace_netfs_collect_stream(rreq, stream);
|
|
trace_netfs_collect_state(rreq, rreq->collected_to, notes);
|
|
|
|
if (!(notes & BUFFERED))
|
|
rreq->cleaned_to = rreq->collected_to;
|
|
|
|
if (notes & NEED_RETRY)
|
|
goto need_retry;
|
|
if (notes & MADE_PROGRESS) {
|
|
netfs_wake_rreq_flag(rreq, NETFS_RREQ_PAUSE, netfs_rreq_trace_unpause);
|
|
//cond_resched();
|
|
goto reassess;
|
|
}
|
|
|
|
out:
|
|
_leave(" = %x", notes);
|
|
return;
|
|
|
|
need_retry:
|
|
/* Okay... We're going to have to retry parts of the stream. Note
|
|
* that any partially completed op will have had any wholly transferred
|
|
* folios removed from it.
|
|
*/
|
|
_debug("retry");
|
|
netfs_retry_reads(rreq);
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* Do page flushing and suchlike after DIO.
|
|
*/
|
|
static void netfs_rreq_assess_dio(struct netfs_io_request *rreq)
|
|
{
|
|
unsigned int i;
|
|
|
|
if (rreq->origin == NETFS_UNBUFFERED_READ ||
|
|
rreq->origin == NETFS_DIO_READ) {
|
|
for (i = 0; i < rreq->direct_bv_count; i++) {
|
|
flush_dcache_page(rreq->direct_bv[i].bv_page);
|
|
// TODO: cifs marks pages in the destination buffer
|
|
// dirty under some circumstances after a read. Do we
|
|
// need to do that too?
|
|
set_page_dirty(rreq->direct_bv[i].bv_page);
|
|
}
|
|
}
|
|
|
|
if (rreq->iocb) {
|
|
rreq->iocb->ki_pos += rreq->transferred;
|
|
if (rreq->iocb->ki_complete) {
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_ki_complete);
|
|
rreq->iocb->ki_complete(
|
|
rreq->iocb, rreq->error ? rreq->error : rreq->transferred);
|
|
}
|
|
}
|
|
if (rreq->netfs_ops->done)
|
|
rreq->netfs_ops->done(rreq);
|
|
if (rreq->origin == NETFS_UNBUFFERED_READ ||
|
|
rreq->origin == NETFS_DIO_READ)
|
|
inode_dio_end(rreq->inode);
|
|
}
|
|
|
|
/*
|
|
* Do processing after reading a monolithic single object.
|
|
*/
|
|
static void netfs_rreq_assess_single(struct netfs_io_request *rreq)
|
|
{
|
|
struct netfs_io_stream *stream = &rreq->io_streams[0];
|
|
|
|
if (!rreq->error && stream->source == NETFS_DOWNLOAD_FROM_SERVER &&
|
|
fscache_resources_valid(&rreq->cache_resources)) {
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_dirty);
|
|
netfs_single_mark_inode_dirty(rreq->inode);
|
|
}
|
|
|
|
if (rreq->iocb) {
|
|
rreq->iocb->ki_pos += rreq->transferred;
|
|
if (rreq->iocb->ki_complete) {
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_ki_complete);
|
|
rreq->iocb->ki_complete(
|
|
rreq->iocb, rreq->error ? rreq->error : rreq->transferred);
|
|
}
|
|
}
|
|
if (rreq->netfs_ops->done)
|
|
rreq->netfs_ops->done(rreq);
|
|
}
|
|
|
|
/*
|
|
* Perform the collection of subrequests and folios.
|
|
*
|
|
* Note that we're in normal kernel thread context at this point, possibly
|
|
* running on a workqueue.
|
|
*/
|
|
bool netfs_read_collection(struct netfs_io_request *rreq)
|
|
{
|
|
struct netfs_io_stream *stream = &rreq->io_streams[0];
|
|
|
|
netfs_collect_read_results(rreq);
|
|
|
|
/* We're done when the app thread has finished posting subreqs and the
|
|
* queue is empty.
|
|
*/
|
|
if (!test_bit(NETFS_RREQ_ALL_QUEUED, &rreq->flags))
|
|
return false;
|
|
smp_rmb(); /* Read ALL_QUEUED before subreq lists. */
|
|
|
|
if (!list_empty(&stream->subrequests))
|
|
return false;
|
|
|
|
/* Okay, declare that all I/O is complete. */
|
|
rreq->transferred = stream->transferred;
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_complete);
|
|
|
|
//netfs_rreq_is_still_valid(rreq);
|
|
|
|
switch (rreq->origin) {
|
|
case NETFS_UNBUFFERED_READ:
|
|
case NETFS_DIO_READ:
|
|
case NETFS_READ_GAPS:
|
|
netfs_rreq_assess_dio(rreq);
|
|
break;
|
|
case NETFS_READ_SINGLE:
|
|
netfs_rreq_assess_single(rreq);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
task_io_account_read(rreq->transferred);
|
|
|
|
netfs_wake_rreq_flag(rreq, NETFS_RREQ_IN_PROGRESS, netfs_rreq_trace_wake_ip);
|
|
/* As we cleared NETFS_RREQ_IN_PROGRESS, we acquired its ref. */
|
|
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_done);
|
|
netfs_clear_subrequests(rreq);
|
|
netfs_unlock_abandoned_read_pages(rreq);
|
|
if (unlikely(rreq->copy_to_cache))
|
|
netfs_pgpriv2_end_copy_to_cache(rreq);
|
|
return true;
|
|
}
|
|
|
|
void netfs_read_collection_worker(struct work_struct *work)
|
|
{
|
|
struct netfs_io_request *rreq = container_of(work, struct netfs_io_request, work);
|
|
|
|
netfs_see_request(rreq, netfs_rreq_trace_see_work);
|
|
if (netfs_check_rreq_in_progress(rreq)) {
|
|
if (netfs_read_collection(rreq))
|
|
/* Drop the ref from the IN_PROGRESS flag. */
|
|
netfs_put_request(rreq, netfs_rreq_trace_put_work_ip);
|
|
else
|
|
netfs_see_request(rreq, netfs_rreq_trace_see_work_complete);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* netfs_read_subreq_progress - Note progress of a read operation.
|
|
* @subreq: The read request that has terminated.
|
|
*
|
|
* This tells the read side of netfs lib that a contributory I/O operation has
|
|
* made some progress and that it may be possible to unlock some folios.
|
|
*
|
|
* Before calling, the filesystem should update subreq->transferred to track
|
|
* the amount of data copied into the output buffer.
|
|
*/
|
|
void netfs_read_subreq_progress(struct netfs_io_subrequest *subreq)
|
|
{
|
|
struct netfs_io_request *rreq = subreq->rreq;
|
|
struct netfs_io_stream *stream = &rreq->io_streams[0];
|
|
size_t fsize = PAGE_SIZE << rreq->front_folio_order;
|
|
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_progress);
|
|
|
|
/* If we are at the head of the queue, wake up the collector,
|
|
* getting a ref to it if we were the ones to do so.
|
|
*/
|
|
if (subreq->start + subreq->transferred > rreq->cleaned_to + fsize &&
|
|
(rreq->origin == NETFS_READAHEAD ||
|
|
rreq->origin == NETFS_READPAGE ||
|
|
rreq->origin == NETFS_READ_FOR_WRITE) &&
|
|
list_is_first(&subreq->rreq_link, &stream->subrequests)
|
|
) {
|
|
__set_bit(NETFS_SREQ_MADE_PROGRESS, &subreq->flags);
|
|
netfs_wake_collector(rreq);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(netfs_read_subreq_progress);
|
|
|
|
/**
|
|
* netfs_read_subreq_terminated - Note the termination of an I/O operation.
|
|
* @subreq: The I/O request that has terminated.
|
|
*
|
|
* This tells the read helper that a contributory I/O operation has terminated,
|
|
* one way or another, and that it should integrate the results.
|
|
*
|
|
* The caller indicates the outcome of the operation through @subreq->error,
|
|
* supplying 0 to indicate a successful or retryable transfer (if
|
|
* NETFS_SREQ_NEED_RETRY is set) or a negative error code. The helper will
|
|
* look after reissuing I/O operations as appropriate and writing downloaded
|
|
* data to the cache.
|
|
*
|
|
* Before calling, the filesystem should update subreq->transferred to track
|
|
* the amount of data copied into the output buffer.
|
|
*/
|
|
void netfs_read_subreq_terminated(struct netfs_io_subrequest *subreq)
|
|
{
|
|
struct netfs_io_request *rreq = subreq->rreq;
|
|
|
|
switch (subreq->source) {
|
|
case NETFS_READ_FROM_CACHE:
|
|
netfs_stat(&netfs_n_rh_read_done);
|
|
break;
|
|
case NETFS_DOWNLOAD_FROM_SERVER:
|
|
netfs_stat(&netfs_n_rh_download_done);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
/* Deal with retry requests, short reads and errors. If we retry
|
|
* but don't make progress, we abandon the attempt.
|
|
*/
|
|
if (!subreq->error && subreq->transferred < subreq->len) {
|
|
if (test_bit(NETFS_SREQ_HIT_EOF, &subreq->flags)) {
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_hit_eof);
|
|
} else if (test_bit(NETFS_SREQ_CLEAR_TAIL, &subreq->flags)) {
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_need_clear);
|
|
} else if (test_bit(NETFS_SREQ_NEED_RETRY, &subreq->flags)) {
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_need_retry);
|
|
} else if (test_bit(NETFS_SREQ_MADE_PROGRESS, &subreq->flags)) {
|
|
__set_bit(NETFS_SREQ_NEED_RETRY, &subreq->flags);
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_partial_read);
|
|
} else {
|
|
__set_bit(NETFS_SREQ_FAILED, &subreq->flags);
|
|
subreq->error = -ENODATA;
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_short);
|
|
}
|
|
}
|
|
|
|
if (unlikely(subreq->error < 0)) {
|
|
trace_netfs_failure(rreq, subreq, subreq->error, netfs_fail_read);
|
|
if (subreq->source == NETFS_READ_FROM_CACHE) {
|
|
netfs_stat(&netfs_n_rh_read_failed);
|
|
__set_bit(NETFS_SREQ_NEED_RETRY, &subreq->flags);
|
|
} else {
|
|
netfs_stat(&netfs_n_rh_download_failed);
|
|
__set_bit(NETFS_SREQ_FAILED, &subreq->flags);
|
|
}
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_set_pause);
|
|
set_bit(NETFS_RREQ_PAUSE, &rreq->flags);
|
|
}
|
|
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_terminated);
|
|
netfs_subreq_clear_in_progress(subreq);
|
|
netfs_put_subrequest(subreq, netfs_sreq_trace_put_terminated);
|
|
}
|
|
EXPORT_SYMBOL(netfs_read_subreq_terminated);
|
|
|
|
/*
|
|
* Handle termination of a read from the cache.
|
|
*/
|
|
void netfs_cache_read_terminated(void *priv, ssize_t transferred_or_error)
|
|
{
|
|
struct netfs_io_subrequest *subreq = priv;
|
|
|
|
if (transferred_or_error > 0) {
|
|
subreq->error = 0;
|
|
if (transferred_or_error > 0) {
|
|
subreq->transferred += transferred_or_error;
|
|
__set_bit(NETFS_SREQ_MADE_PROGRESS, &subreq->flags);
|
|
}
|
|
} else {
|
|
subreq->error = transferred_or_error;
|
|
}
|
|
netfs_read_subreq_terminated(subreq);
|
|
}
|