mirror of
https://git.FreeBSD.org/src.git
synced 2025-02-02 17:08:56 +00:00
When aio_read/write() is used on a raw device, physical buffers are
used for up to "vfs.aio.max_buf_aio" of the requests. If a request size is MAXPHYS, but the request base isn't page aligned, vmapbuf() will map the end of the user space buffer into the start of the kva allocated for the next physical buffer. Don't use a physical buffer in this case. (This change addresses problem report 25617.) When an aio_read/write() on a raw device has completed, timeout() is used to schedule a signal to the process. Thus, the reporting is delayed up to 10 ms (assuming hz is 100). The process might have terminated in the meantime, causing a trap 12 when attempting to deliver the signal. Thus, the timeout must be cancelled when removing the job. aio jobs in state JOBST_JOBQGLOBAL should be removed from the kaio_jobqueue list during process rundown. During process rundown, some aio jobs might move from one list to a different list that has already been "emptied", causing the rundown to be incomplete. Retry the rundown. A call to BUF_KERNPROC() is needed after obtaining a physical buffer to disassociate the lock from the running process since it can return to userland without releasing that lock. PR: 25617 Submitted by: tegge
This commit is contained in:
parent
26b104142d
commit
136446540a
Notes:
svn2git
2020-12-20 02:59:44 +00:00
svn path=/head/; revision=74089
@ -215,6 +215,7 @@ static void aio_physwakeup(struct buf *bp);
|
||||
static int aio_fphysio(struct proc *p, struct aiocblist *aiocbe);
|
||||
static int aio_qphysio(struct proc *p, struct aiocblist *iocb);
|
||||
static void aio_daemon(void *uproc);
|
||||
static void process_signal(void *aioj);
|
||||
|
||||
SYSINIT(aio, SI_SUB_VFS, SI_ORDER_ANY, aio_onceonly, NULL);
|
||||
|
||||
@ -298,11 +299,11 @@ aio_free_entry(struct aiocblist *aiocbe)
|
||||
if (ki == NULL)
|
||||
panic("aio_free_entry: missing p->p_aioinfo");
|
||||
|
||||
if (aiocbe->jobstate == JOBST_JOBRUNNING) {
|
||||
while (aiocbe->jobstate == JOBST_JOBRUNNING) {
|
||||
if (aiocbe->jobflags & AIOCBLIST_ASYNCFREE)
|
||||
return 0;
|
||||
aiocbe->jobflags |= AIOCBLIST_RUNDOWN;
|
||||
tsleep(aiocbe, PRIBIO|PCATCH, "jobwai", 0);
|
||||
tsleep(aiocbe, PRIBIO, "jobwai", 0);
|
||||
}
|
||||
aiocbe->jobflags &= ~AIOCBLIST_ASYNCFREE;
|
||||
|
||||
@ -353,9 +354,10 @@ aio_free_entry(struct aiocblist *aiocbe)
|
||||
} else if (aiocbe->jobstate == JOBST_JOBQPROC) {
|
||||
aiop = aiocbe->jobaioproc;
|
||||
TAILQ_REMOVE(&aiop->jobtorun, aiocbe, list);
|
||||
} else if (aiocbe->jobstate == JOBST_JOBQGLOBAL)
|
||||
} else if (aiocbe->jobstate == JOBST_JOBQGLOBAL) {
|
||||
TAILQ_REMOVE(&aio_jobs, aiocbe, list);
|
||||
else if (aiocbe->jobstate == JOBST_JOBFINISHED)
|
||||
TAILQ_REMOVE(&ki->kaio_jobqueue, aiocbe, plist);
|
||||
} else if (aiocbe->jobstate == JOBST_JOBFINISHED)
|
||||
TAILQ_REMOVE(&ki->kaio_jobdone, aiocbe, plist);
|
||||
else if (aiocbe->jobstate == JOBST_JOBBFINISHED) {
|
||||
s = splbio();
|
||||
@ -372,6 +374,7 @@ aio_free_entry(struct aiocblist *aiocbe)
|
||||
zfree(aiolio_zone, lj);
|
||||
}
|
||||
aiocbe->jobstate = JOBST_NULL;
|
||||
untimeout(process_signal, aiocbe, aiocbe->timeouthandle);
|
||||
zfree(aiocb_zone, aiocbe);
|
||||
return 0;
|
||||
}
|
||||
@ -481,6 +484,16 @@ restart4:
|
||||
}
|
||||
splx(s);
|
||||
|
||||
/*
|
||||
* If we've slept, jobs might have moved from one queue to another.
|
||||
* Retry rundown if we didn't manage to empty the queues.
|
||||
*/
|
||||
if (TAILQ_FIRST(&ki->kaio_jobdone) != NULL ||
|
||||
TAILQ_FIRST(&ki->kaio_jobqueue) != NULL ||
|
||||
TAILQ_FIRST(&ki->kaio_bufqueue) != NULL ||
|
||||
TAILQ_FIRST(&ki->kaio_bufdone) != NULL)
|
||||
goto restart1;
|
||||
|
||||
for (lj = TAILQ_FIRST(&ki->kaio_liojoblist); lj; lj = ljn) {
|
||||
ljn = TAILQ_NEXT(lj, lioj_list);
|
||||
if ((lj->lioj_buffer_count == 0) && (lj->lioj_queue_count ==
|
||||
@ -990,7 +1003,8 @@ aio_qphysio(struct proc *p, struct aiocblist *aiocbe)
|
||||
if (cb->aio_nbytes % vp->v_rdev->si_bsize_phys)
|
||||
return (-1);
|
||||
|
||||
if (cb->aio_nbytes > MAXPHYS)
|
||||
if (cb->aio_nbytes >
|
||||
MAXPHYS - (((vm_offset_t) cb->aio_buf) & PAGE_MASK))
|
||||
return (-1);
|
||||
|
||||
ki = p->p_aioinfo;
|
||||
@ -1005,6 +1019,7 @@ aio_qphysio(struct proc *p, struct aiocblist *aiocbe)
|
||||
|
||||
/* Create and build a buffer header for a transfer. */
|
||||
bp = (struct buf *)getpbuf(NULL);
|
||||
BUF_KERNPROC(bp);
|
||||
|
||||
/*
|
||||
* Get a copy of the kva from the physical buffer.
|
||||
@ -1209,6 +1224,7 @@ _aio_aqueue(struct proc *p, struct aiocb *job, struct aio_liojob *lj, int type)
|
||||
aiocbe = zalloc(aiocb_zone);
|
||||
aiocbe->inputcharge = 0;
|
||||
aiocbe->outputcharge = 0;
|
||||
callout_handle_init(&aiocbe->timeouthandle);
|
||||
SLIST_INIT(&aiocbe->klist);
|
||||
|
||||
suword(&job->_aiocb_private.status, -1);
|
||||
@ -2116,7 +2132,9 @@ aio_physwakeup(struct buf *bp)
|
||||
(LIOJ_SIGNAL|LIOJ_SIGNAL_POSTED)) ==
|
||||
LIOJ_SIGNAL) {
|
||||
lj->lioj_flags |= LIOJ_SIGNAL_POSTED;
|
||||
timeout(process_signal, aiocbe, 0);
|
||||
aiocbe->timeouthandle =
|
||||
timeout(process_signal,
|
||||
aiocbe, 0);
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -2137,7 +2155,8 @@ aio_physwakeup(struct buf *bp)
|
||||
}
|
||||
|
||||
if (aiocbe->uaiocb.aio_sigevent.sigev_notify == SIGEV_SIGNAL)
|
||||
timeout(process_signal, aiocbe, 0);
|
||||
aiocbe->timeouthandle =
|
||||
timeout(process_signal, aiocbe, 0);
|
||||
}
|
||||
}
|
||||
#endif /* VFS_AIO */
|
||||
|
@ -138,6 +138,7 @@ struct aiocblist {
|
||||
int jobflags;
|
||||
int jobstate;
|
||||
int inputcharge, outputcharge;
|
||||
struct callout_handle timeouthandle;
|
||||
struct buf *bp; /* Buffer pointer */
|
||||
struct proc *userproc; /* User process */
|
||||
struct file *fd_file; /* Pointer to file structure */
|
||||
|
Loading…
x
Reference in New Issue
Block a user