Current Path : /sys/amd64/compile/hs32/modules/usr/src/sys/modules/s3/@/amd64/compile/hs32/modules/usr/src/sys/modules/usb/usie/@/amd64/compile/hs32/modules/usr/src/sys/modules/oce/@/amd64/compile/hs32/modules/usr/src/sys/modules/procfs/@/geom/vinum/ |
FreeBSD hs32.drive.ne.jp 9.1-RELEASE FreeBSD 9.1-RELEASE #1: Wed Jan 14 12:18:08 JST 2015 root@hs32.drive.ne.jp:/sys/amd64/compile/hs32 amd64 |
Current File : //sys/amd64/compile/hs32/modules/usr/src/sys/modules/s3/@/amd64/compile/hs32/modules/usr/src/sys/modules/usb/usie/@/amd64/compile/hs32/modules/usr/src/sys/modules/oce/@/amd64/compile/hs32/modules/usr/src/sys/modules/procfs/@/geom/vinum/geom_vinum_volume.c |
/*- * Copyright (c) 2007 Lukas Ertl * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ #include <sys/cdefs.h> __FBSDID("$FreeBSD: release/9.1.0/sys/geom/vinum/geom_vinum_volume.c 190507 2009-03-28 17:20:08Z lulf $"); #include <sys/param.h> #include <sys/bio.h> #include <sys/lock.h> #include <sys/malloc.h> #include <sys/systm.h> #include <geom/geom.h> #include <geom/vinum/geom_vinum_var.h> #include <geom/vinum/geom_vinum.h> void gv_volume_flush(struct gv_volume *v) { struct gv_softc *sc; struct bio *bp; KASSERT(v != NULL, ("NULL v")); sc = v->vinumconf; KASSERT(sc != NULL, ("NULL sc")); bp = bioq_takefirst(v->wqueue); while (bp != NULL) { gv_volume_start(sc, bp); bp = bioq_takefirst(v->wqueue); } } void gv_volume_start(struct gv_softc *sc, struct bio *bp) { struct g_geom *gp; struct gv_volume *v; struct gv_plex *p, *lp; int numwrites; gp = sc->geom; v = bp->bio_to->private; if (v == NULL || v->state != GV_VOL_UP) { g_io_deliver(bp, ENXIO); return; } switch (bp->bio_cmd) { case BIO_READ: /* * Try to find a good plex where we can send the request to, * round-robin-style. The plex either has to be up, or it's a * degraded RAID5 plex. Check if we have delayed requests. Put * this request on the delayed queue if so. This makes sure that * we don't read old values. */ if (bioq_first(v->wqueue) != NULL) { bioq_insert_tail(v->wqueue, bp); break; } lp = v->last_read_plex; if (lp == NULL) lp = LIST_FIRST(&v->plexes); p = LIST_NEXT(lp, in_volume); if (p == NULL) p = LIST_FIRST(&v->plexes); do { if (p == NULL) { p = lp; break; } if ((p->state > GV_PLEX_DEGRADED) || (p->state >= GV_PLEX_DEGRADED && p->org == GV_PLEX_RAID5)) break; p = LIST_NEXT(p, in_volume); if (p == NULL) p = LIST_FIRST(&v->plexes); } while (p != lp); if ((p == NULL) || (p->org == GV_PLEX_RAID5 && p->state < GV_PLEX_DEGRADED) || (p->org != GV_PLEX_RAID5 && p->state <= GV_PLEX_DEGRADED)) { g_io_deliver(bp, ENXIO); return; } v->last_read_plex = p; /* Hand it down to the plex logic. */ gv_plex_start(p, bp); break; case BIO_WRITE: case BIO_DELETE: /* Delay write-requests if any plex is synchronizing. */ LIST_FOREACH(p, &v->plexes, in_volume) { if (p->flags & GV_PLEX_SYNCING) { bioq_insert_tail(v->wqueue, bp); return; } } numwrites = 0; /* Give the BIO to each plex of this volume. */ LIST_FOREACH(p, &v->plexes, in_volume) { if (p->state < GV_PLEX_DEGRADED) continue; gv_plex_start(p, bp); numwrites++; } if (numwrites == 0) g_io_deliver(bp, ENXIO); break; } } void gv_bio_done(struct gv_softc *sc, struct bio *bp) { struct gv_volume *v; struct gv_plex *p; struct gv_sd *s; s = bp->bio_caller1; KASSERT(s != NULL, ("gv_bio_done: NULL s")); p = s->plex_sc; KASSERT(p != NULL, ("gv_bio_done: NULL p")); v = p->vol_sc; KASSERT(v != NULL, ("gv_bio_done: NULL v")); switch (p->org) { case GV_PLEX_CONCAT: case GV_PLEX_STRIPED: gv_plex_normal_done(p, bp); break; case GV_PLEX_RAID5: gv_plex_raid5_done(p, bp); break; } }