NetBSD/libexec/lfs_cleanerd/lfs_cleanerd.c

1725 lines
45 KiB
C

/* $NetBSD: lfs_cleanerd.c,v 1.58 2016/03/18 10:10:21 mrg Exp $ */
/*-
* Copyright (c) 2005 The NetBSD Foundation, Inc.
* All rights reserved.
*
* This code is derived from software contributed to The NetBSD Foundation
* by Konrad E. Schroder <perseant@hhhh.org>.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/
/*
* The cleaner daemon for the NetBSD Log-structured File System.
* Only tested for use with version 2 LFSs.
*/
#include <sys/syslog.h>
#include <sys/param.h>
#include <sys/mount.h>
#include <sys/stat.h>
#include <ufs/lfs/lfs.h>
#include <assert.h>
#include <err.h>
#include <errno.h>
#include <fcntl.h>
#include <semaphore.h>
#include <stdbool.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <unistd.h>
#include <time.h>
#include <util.h>
#include "bufcache.h"
#include "vnode.h"
#include "lfs_user.h"
#include "fdfs.h"
#include "cleaner.h"
#include "kernelops.h"
#include "mount_lfs.h"
/*
* Global variables.
*/
/* XXX these top few should really be fs-specific */
int use_fs_idle; /* Use fs idle rather than cpu idle time */
int use_bytes; /* Use bytes written rather than segments cleaned */
double load_threshold; /* How idle is idle (CPU idle) */
int atatime; /* How many segments (bytes) to clean at a time */
int nfss; /* Number of filesystems monitored by this cleanerd */
struct clfs **fsp; /* Array of extended filesystem structures */
int segwait_timeout; /* Time to wait in lfs_segwait() */
int do_quit; /* Quit after one cleaning loop */
int do_coalesce; /* Coalesce filesystem */
int do_small; /* Use small writes through markv */
char *copylog_filename; /* File to use for fs debugging analysis */
int inval_segment; /* Segment to invalidate */
int stat_report; /* Report statistics for this period of cycles */
int debug; /* Turn on debugging */
struct cleaner_stats {
double util_tot;
double util_sos;
off_t bytes_read;
off_t bytes_written;
off_t segs_cleaned;
off_t segs_empty;
off_t segs_error;
} cleaner_stats;
extern u_int32_t cksum(void *, size_t);
extern u_int32_t lfs_sb_cksum(struct dlfs *);
extern u_int32_t lfs_cksum_part(void *, size_t, u_int32_t);
extern int ulfs_getlbns(struct lfs *, struct uvnode *, daddr_t, struct indir *, int *);
/* Ugh */
#define FSMNT_SIZE MAX(sizeof(((struct dlfs *)0)->dlfs_fsmnt), \
sizeof(((struct dlfs64 *)0)->dlfs_fsmnt))
/* Compat */
void pwarn(const char *unused, ...) { /* Does nothing */ };
/*
* Log a message if debugging is turned on.
*/
void
dlog(const char *fmt, ...)
{
va_list ap;
if (debug == 0)
return;
va_start(ap, fmt);
vsyslog(LOG_DEBUG, fmt, ap);
va_end(ap);
}
/*
* Remove the specified filesystem from the list, due to its having
* become unmounted or other error condition.
*/
void
handle_error(struct clfs **cfsp, int n)
{
syslog(LOG_NOTICE, "%s: detaching cleaner", lfs_sb_getfsmnt(cfsp[n]));
free(cfsp[n]);
if (n != nfss - 1)
cfsp[n] = cfsp[nfss - 1];
--nfss;
}
/*
* Reinitialize a filesystem if, e.g., its size changed.
*/
int
reinit_fs(struct clfs *fs)
{
char fsname[FSMNT_SIZE];
memcpy(fsname, lfs_sb_getfsmnt(fs), sizeof(fsname));
fsname[sizeof(fsname) - 1] = '\0';
kops.ko_close(fs->clfs_ifilefd);
kops.ko_close(fs->clfs_devfd);
fd_reclaim(fs->clfs_devvp);
fd_reclaim(fs->lfs_ivnode);
free(fs->clfs_dev);
free(fs->clfs_segtab);
free(fs->clfs_segtabp);
return init_fs(fs, fsname);
}
#ifdef REPAIR_ZERO_FINFO
/*
* Use fsck's lfs routines to load the Ifile from an unmounted fs.
* We interpret "fsname" as the name of the raw disk device.
*/
int
init_unmounted_fs(struct clfs *fs, char *fsname)
{
struct lfs *disc_fs;
int i;
fs->clfs_dev = fsname;
if ((fs->clfs_devfd = kops.ko_open(fs->clfs_dev, O_RDWR)) < 0) {
syslog(LOG_ERR, "couldn't open device %s read/write",
fs->clfs_dev);
return -1;
}
disc_fs = lfs_init(fs->clfs_devfd, 0, 0, 0, 0);
fs->lfs_dlfs = disc_fs->lfs_dlfs; /* Structure copy */
strncpy(fs->lfs_fsmnt, fsname, MNAMELEN);
fs->lfs_ivnode = (struct uvnode *)disc_fs->lfs_ivnode;
fs->clfs_devvp = fd_vget(fs->clfs_devfd, fs->lfs_fsize, fs->lfs_ssize,
atatime);
/* Allocate and clear segtab */
fs->clfs_segtab = (struct clfs_seguse *)malloc(lfs_sb_getnseg(fs) *
sizeof(*fs->clfs_segtab));
fs->clfs_segtabp = (struct clfs_seguse **)malloc(lfs_sb_getnseg(fs) *
sizeof(*fs->clfs_segtabp));
for (i = 0; i < lfs_sb_getnseg(fs); i++) {
fs->clfs_segtabp[i] = &(fs->clfs_segtab[i]);
fs->clfs_segtab[i].flags = 0x0;
}
syslog(LOG_NOTICE, "%s: unmounted cleaner starting", fsname);
return 0;
}
#endif
/*
* Set up the file descriptors, including the Ifile descriptor.
* If we can't get the Ifile, this is not an LFS (or the kernel is
* too old to support the fcntl).
* XXX Merge this and init_unmounted_fs, switching on whether
* XXX "fsname" is a dir or a char special device. Should
* XXX also be able to read unmounted devices out of fstab, the way
* XXX fsck does.
*/
int
init_fs(struct clfs *fs, char *fsname)
{
char mnttmp[FSMNT_SIZE];
struct statvfs sf;
int rootfd;
int i;
void *sbuf;
char *bn;
/*
* Get the raw device from the block device.
* XXX this is ugly. Is there a way to discover the raw device
* XXX for a given mount point?
*/
if (kops.ko_statvfs(fsname, &sf, ST_WAIT) < 0)
return -1;
fs->clfs_dev = malloc(strlen(sf.f_mntfromname) + 2);
if (fs->clfs_dev == NULL) {
syslog(LOG_ERR, "couldn't malloc device name string: %m");
return -1;
}
bn = strrchr(sf.f_mntfromname, '/');
bn = bn ? bn+1 : sf.f_mntfromname;
strlcpy(fs->clfs_dev, sf.f_mntfromname, bn - sf.f_mntfromname + 1);
strcat(fs->clfs_dev, "r");
strcat(fs->clfs_dev, bn);
if ((fs->clfs_devfd = kops.ko_open(fs->clfs_dev, O_RDONLY, 0)) < 0) {
syslog(LOG_ERR, "couldn't open device %s for reading",
fs->clfs_dev);
return -1;
}
/* Find the Ifile and open it */
if ((rootfd = kops.ko_open(fsname, O_RDONLY, 0)) < 0)
return -2;
if (kops.ko_fcntl(rootfd, LFCNIFILEFH, &fs->clfs_ifilefh) < 0)
return -3;
if ((fs->clfs_ifilefd = kops.ko_fhopen(&fs->clfs_ifilefh,
sizeof(fs->clfs_ifilefh), O_RDONLY)) < 0)
return -4;
kops.ko_close(rootfd);
sbuf = malloc(LFS_SBPAD);
if (sbuf == NULL) {
syslog(LOG_ERR, "couldn't malloc superblock buffer");
return -1;
}
/* Load in the superblock */
if (kops.ko_pread(fs->clfs_devfd, sbuf, LFS_SBPAD, LFS_LABELPAD) < 0) {
free(sbuf);
return -1;
}
__CTASSERT(sizeof(struct dlfs) == sizeof(struct dlfs64));
memcpy(&fs->lfs_dlfs_u, sbuf, sizeof(struct dlfs));
free(sbuf);
/* If it is not LFS, complain and exit! */
switch (fs->lfs_dlfs_u.u_32.dlfs_magic) {
case LFS_MAGIC:
fs->lfs_is64 = false;
fs->lfs_dobyteswap = false;
break;
case LFS_MAGIC_SWAPPED:
fs->lfs_is64 = false;
fs->lfs_dobyteswap = true;
break;
case LFS64_MAGIC:
fs->lfs_is64 = true;
fs->lfs_dobyteswap = false;
break;
case LFS64_MAGIC_SWAPPED:
fs->lfs_is64 = true;
fs->lfs_dobyteswap = true;
break;
default:
syslog(LOG_ERR, "%s: not LFS", fsname);
return -1;
}
/* XXX: can this ever need to be set? does the cleaner even care? */
fs->lfs_hasolddirfmt = 0;
/* If this is not a version 2 filesystem, complain and exit */
if (lfs_sb_getversion(fs) != 2) {
syslog(LOG_ERR, "%s: not a version 2 LFS", fsname);
return -1;
}
/* Assume fsname is the mounted name */
strncpy(mnttmp, fsname, sizeof(mnttmp));
mnttmp[sizeof(mnttmp) - 1] = '\0';
lfs_sb_setfsmnt(fs, mnttmp);
/* Set up vnodes for Ifile and raw device */
fs->lfs_ivnode = fd_vget(fs->clfs_ifilefd, lfs_sb_getbsize(fs), 0, 0);
fs->clfs_devvp = fd_vget(fs->clfs_devfd, lfs_sb_getfsize(fs), lfs_sb_getssize(fs),
atatime);
/* Allocate and clear segtab */
fs->clfs_segtab = (struct clfs_seguse *)malloc(lfs_sb_getnseg(fs) *
sizeof(*fs->clfs_segtab));
fs->clfs_segtabp = (struct clfs_seguse **)malloc(lfs_sb_getnseg(fs) *
sizeof(*fs->clfs_segtabp));
if (fs->clfs_segtab == NULL || fs->clfs_segtabp == NULL) {
syslog(LOG_ERR, "%s: couldn't malloc segment table: %m",
fs->clfs_dev);
return -1;
}
for (i = 0; i < lfs_sb_getnseg(fs); i++) {
fs->clfs_segtabp[i] = &(fs->clfs_segtab[i]);
fs->clfs_segtab[i].flags = 0x0;
}
syslog(LOG_NOTICE, "%s: attaching cleaner", fsname);
return 0;
}
/*
* Invalidate all the currently held Ifile blocks so they will be
* reread when we clean. Check the size while we're at it, and
* resize the buffer cache if necessary.
*/
void
reload_ifile(struct clfs *fs)
{
struct ubuf *bp;
struct stat st;
int ohashmax;
extern int hashmax;
while ((bp = LIST_FIRST(&fs->lfs_ivnode->v_dirtyblkhd)) != NULL) {
bremfree(bp);
buf_destroy(bp);
}
while ((bp = LIST_FIRST(&fs->lfs_ivnode->v_cleanblkhd)) != NULL) {
bremfree(bp);
buf_destroy(bp);
}
/* If Ifile is larger than buffer cache, rehash */
fstat(fs->clfs_ifilefd, &st);
if (st.st_size / lfs_sb_getbsize(fs) > hashmax) {
ohashmax = hashmax;
bufrehash(st.st_size / lfs_sb_getbsize(fs));
dlog("%s: resized buffer hash from %d to %d",
lfs_sb_getfsmnt(fs), ohashmax, hashmax);
}
}
/*
* Get IFILE entry for the given inode, store in ifpp. The buffer
* which contains that data is returned in bpp, and must be brelse()d
* by the caller.
*
* XXX this is cutpaste of LFS_IENTRY from lfs.h; unify the two.
*/
void
lfs_ientry(IFILE **ifpp, struct clfs *fs, ino_t ino, struct ubuf **bpp)
{
IFILE64 *ifp64;
IFILE32 *ifp32;
IFILE_V1 *ifp_v1;
int error;
error = bread(fs->lfs_ivnode,
ino / lfs_sb_getifpb(fs) + lfs_sb_getcleansz(fs) +
lfs_sb_getsegtabsz(fs), lfs_sb_getbsize(fs), 0, bpp);
if (error)
syslog(LOG_ERR, "%s: ientry failed for ino %d",
lfs_sb_getfsmnt(fs), (int)ino);
if (fs->lfs_is64) {
ifp64 = (IFILE64 *)(*bpp)->b_data;
ifp64 += ino % lfs_sb_getifpb(fs);
*ifpp = (IFILE *)ifp64;
} else if (lfs_sb_getversion(fs) > 1) {
ifp32 = (IFILE32 *)(*bpp)->b_data;
ifp32 += ino % lfs_sb_getifpb(fs);
*ifpp = (IFILE *)ifp32;
} else {
ifp_v1 = (IFILE_V1 *)(*bpp)->b_data;
ifp_v1 += ino % lfs_sb_getifpb(fs);
*ifpp = (IFILE *)ifp_v1;
}
return;
}
#ifdef TEST_PATTERN
/*
* Check ULFS_ROOTINO for file data. The assumption is that we are running
* the "twofiles" test with the rest of the filesystem empty. Files
* created by "twofiles" match the test pattern, but ULFS_ROOTINO and the
* executable itself (assumed to be inode 3) should not match.
*/
static void
check_test_pattern(BLOCK_INFO *bip)
{
int j;
unsigned char *cp = bip->bi_bp;
/* Check inode sanity */
if (bip->bi_lbn == LFS_UNUSED_LBN) {
assert(((struct ulfs1_dinode *)bip->bi_bp)->di_inumber ==
bip->bi_inode);
}
/* These can have the test pattern and it's all good */
if (bip->bi_inode > 3)
return;
for (j = 0; j < bip->bi_size; j++) {
if (cp[j] != (j & 0xff))
break;
}
assert(j < bip->bi_size);
}
#endif /* TEST_PATTERN */
/*
* Parse the partial segment at daddr, adding its information to
* bip. Return the address of the next partial segment to read.
*/
static daddr_t
parse_pseg(struct clfs *fs, daddr_t daddr, BLOCK_INFO **bipp, int *bic)
{
SEGSUM *ssp;
IFILE *ifp;
BLOCK_INFO *bip, *nbip;
daddr_t idaddr, odaddr;
FINFO *fip;
IINFO *iip;
struct ubuf *ifbp;
union lfs_dinode *dip;
u_int32_t ck, vers;
int fic, inoc, obic;
size_t sumstart;
int i;
char *cp;
odaddr = daddr;
obic = *bic;
bip = *bipp;
/*
* Retrieve the segment header, set up the SEGSUM pointer
* as well as the first FINFO and inode address pointer.
*/
cp = fd_ptrget(fs->clfs_devvp, daddr);
ssp = (SEGSUM *)cp;
iip = SEGSUM_IINFOSTART(fs, cp);
fip = SEGSUM_FINFOBASE(fs, cp);
/*
* Check segment header magic and checksum
*/
if (lfs_ss_getmagic(fs, ssp) != SS_MAGIC) {
syslog(LOG_WARNING, "%s: sumsum magic number bad at 0x%jx:"
" read 0x%x, expected 0x%x", lfs_sb_getfsmnt(fs),
(intmax_t)daddr, lfs_ss_getmagic(fs, ssp), SS_MAGIC);
return 0x0;
}
sumstart = lfs_ss_getsumstart(fs);
ck = cksum((char *)ssp + sumstart, lfs_sb_getsumsize(fs) - sumstart);
if (ck != lfs_ss_getsumsum(fs, ssp)) {
syslog(LOG_WARNING, "%s: sumsum checksum mismatch at 0x%jx:"
" read 0x%x, computed 0x%x", lfs_sb_getfsmnt(fs),
(intmax_t)daddr, lfs_ss_getsumsum(fs, ssp), ck);
return 0x0;
}
/* Initialize data sum */
ck = 0;
/* Point daddr at next block after segment summary */
++daddr;
/*
* Loop over file info and inode pointers. We always move daddr
* forward here because we are also computing the data checksum
* as we go.
*/
fic = inoc = 0;
while (fic < lfs_ss_getnfinfo(fs, ssp) || inoc < lfs_ss_getninos(fs, ssp)) {
/*
* We must have either a file block or an inode block.
* If we don't have either one, it's an error.
*/
if (fic >= lfs_ss_getnfinfo(fs, ssp) && lfs_ii_getblock(fs, iip) != daddr) {
syslog(LOG_WARNING, "%s: bad pseg at %jx (seg %d)",
lfs_sb_getfsmnt(fs), (intmax_t)odaddr, lfs_dtosn(fs, odaddr));
*bipp = bip;
return 0x0;
}
/*
* Note each inode from the inode blocks
*/
if (inoc < lfs_ss_getninos(fs, ssp) && lfs_ii_getblock(fs, iip) == daddr) {
cp = fd_ptrget(fs->clfs_devvp, daddr);
ck = lfs_cksum_part(cp, sizeof(u_int32_t), ck);
for (i = 0; i < lfs_sb_getinopb(fs); i++) {
dip = DINO_IN_BLOCK(fs, cp, i);
if (lfs_dino_getinumber(fs, dip) == 0)
break;
/*
* Check currency before adding it
*/
#ifndef REPAIR_ZERO_FINFO
lfs_ientry(&ifp, fs, lfs_dino_getinumber(fs, dip), &ifbp);
idaddr = lfs_if_getdaddr(fs, ifp);
brelse(ifbp, 0);
if (idaddr != daddr)
#endif
continue;
/*
* A current inode. Add it.
*/
++*bic;
nbip = (BLOCK_INFO *)realloc(bip, *bic *
sizeof(*bip));
if (nbip)
bip = nbip;
else {
--*bic;
*bipp = bip;
return 0x0;
}
bip[*bic - 1].bi_inode = lfs_dino_getinumber(fs, dip);
bip[*bic - 1].bi_lbn = LFS_UNUSED_LBN;
bip[*bic - 1].bi_daddr = daddr;
bip[*bic - 1].bi_segcreate = lfs_ss_getcreate(fs, ssp);
bip[*bic - 1].bi_version = lfs_dino_getgen(fs, dip);
bip[*bic - 1].bi_bp = dip;
bip[*bic - 1].bi_size = DINOSIZE(fs);
}
inoc += i;
daddr += lfs_btofsb(fs, lfs_sb_getibsize(fs));
iip = NEXTLOWER_IINFO(fs, iip);
continue;
}
/*
* Note each file block from the finfo blocks
*/
if (fic >= lfs_ss_getnfinfo(fs, ssp))
continue;
/* Count this finfo, whether or not we use it */
++fic;
/*
* If this finfo has nblocks==0, it was written wrong.
* Kernels with this problem always wrote this zero-sized
* finfo last, so just ignore it.
*/
if (lfs_fi_getnblocks(fs, fip) == 0) {
#ifdef REPAIR_ZERO_FINFO
struct ubuf *nbp;
SEGSUM *nssp;
syslog(LOG_WARNING, "fixing short FINFO at %jx (seg %d)",
(intmax_t)odaddr, lfs_dtosn(fs, odaddr));
bread(fs->clfs_devvp, odaddr, lfs_sb_getfsize(fs),
0, &nbp);
nssp = (SEGSUM *)nbp->b_data;
--nssp->ss_nfinfo;
nssp->ss_sumsum = cksum(&nssp->ss_datasum,
lfs_sb_getsumsize(fs) - sizeof(nssp->ss_sumsum));
bwrite(nbp);
#endif
syslog(LOG_WARNING, "zero-length FINFO at %jx (seg %d)",
(intmax_t)odaddr, lfs_dtosn(fs, odaddr));
continue;
}
/*
* Check currency before adding blocks
*/
#ifdef REPAIR_ZERO_FINFO
vers = -1;
#else
lfs_ientry(&ifp, fs, lfs_fi_getino(fs, fip), &ifbp);
vers = lfs_if_getversion(fs, ifp);
brelse(ifbp, 0);
#endif
if (vers != lfs_fi_getversion(fs, fip)) {
size_t size;
/* Read all the blocks from the data summary */
for (i = 0; i < lfs_fi_getnblocks(fs, fip); i++) {
size = (i == lfs_fi_getnblocks(fs, fip) - 1) ?
lfs_fi_getlastlength(fs, fip) : lfs_sb_getbsize(fs);
cp = fd_ptrget(fs->clfs_devvp, daddr);
ck = lfs_cksum_part(cp, sizeof(u_int32_t), ck);
daddr += lfs_btofsb(fs, size);
}
fip = NEXT_FINFO(fs, fip);
continue;
}
/* Add all the blocks from the finfos (current or not) */
nbip = (BLOCK_INFO *)realloc(bip, (*bic + lfs_fi_getnblocks(fs, fip)) *
sizeof(*bip));
if (nbip)
bip = nbip;
else {
*bipp = bip;
return 0x0;
}
for (i = 0; i < lfs_fi_getnblocks(fs, fip); i++) {
bip[*bic + i].bi_inode = lfs_fi_getino(fs, fip);
bip[*bic + i].bi_lbn = lfs_fi_getblock(fs, fip, i);
bip[*bic + i].bi_daddr = daddr;
bip[*bic + i].bi_segcreate = lfs_ss_getcreate(fs, ssp);
bip[*bic + i].bi_version = lfs_fi_getversion(fs, fip);
bip[*bic + i].bi_size = (i == lfs_fi_getnblocks(fs, fip) - 1) ?
lfs_fi_getlastlength(fs, fip) : lfs_sb_getbsize(fs);
cp = fd_ptrget(fs->clfs_devvp, daddr);
ck = lfs_cksum_part(cp, sizeof(u_int32_t), ck);
bip[*bic + i].bi_bp = cp;
daddr += lfs_btofsb(fs, bip[*bic + i].bi_size);
#ifdef TEST_PATTERN
check_test_pattern(bip + *bic + i); /* XXXDEBUG */
#endif
}
*bic += lfs_fi_getnblocks(fs, fip);
fip = NEXT_FINFO(fs, fip);
}
#ifndef REPAIR_ZERO_FINFO
if (lfs_ss_getdatasum(fs, ssp) != ck) {
syslog(LOG_WARNING, "%s: data checksum bad at 0x%jx:"
" read 0x%x, computed 0x%x", lfs_sb_getfsmnt(fs),
(intmax_t)odaddr,
lfs_ss_getdatasum(fs, ssp), ck);
*bic = obic;
return 0x0;
}
#endif
*bipp = bip;
return daddr;
}
static void
log_segment_read(struct clfs *fs, int sn)
{
FILE *fp;
char *cp;
/*
* Write the segment read, and its contents, into a log file in
* the current directory. We don't need to log the location of
* the segment, since that can be inferred from the segments up
* to this point (ss_nextseg field of the previously written segment).
*
* We can use this info later to reconstruct the filesystem at any
* given point in time for analysis, by replaying the log forward
* indexed by the segment serial numbers; but it is not suitable
* for everyday use since the copylog will be simply enormous.
*/
cp = fd_ptrget(fs->clfs_devvp, lfs_sntod(fs, sn));
fp = fopen(copylog_filename, "ab");
if (fp != NULL) {
if (fwrite(cp, (size_t)lfs_sb_getssize(fs), 1, fp) != 1) {
perror("writing segment to copy log");
}
}
fclose(fp);
}
/*
* Read a segment to populate the BLOCK_INFO structures.
* Return the number of partial segments read and parsed.
*/
int
load_segment(struct clfs *fs, int sn, BLOCK_INFO **bipp, int *bic)
{
daddr_t daddr;
int i, npseg;
daddr = lfs_sntod(fs, sn);
if (daddr < lfs_btofsb(fs, LFS_LABELPAD))
daddr = lfs_btofsb(fs, LFS_LABELPAD);
for (i = 0; i < LFS_MAXNUMSB; i++) {
if (lfs_sb_getsboff(fs, i) == daddr) {
daddr += lfs_btofsb(fs, LFS_SBPAD);
break;
}
}
/* Preload the segment buffer */
if (fd_preload(fs->clfs_devvp, lfs_sntod(fs, sn)) < 0)
return -1;
if (copylog_filename)
log_segment_read(fs, sn);
/* Note bytes read for stats */
cleaner_stats.segs_cleaned++;
cleaner_stats.bytes_read += lfs_sb_getssize(fs);
++fs->clfs_nactive;
npseg = 0;
while(lfs_dtosn(fs, daddr) == sn &&
lfs_dtosn(fs, daddr + lfs_btofsb(fs, lfs_sb_getbsize(fs))) == sn) {
daddr = parse_pseg(fs, daddr, bipp, bic);
if (daddr == 0x0) {
++cleaner_stats.segs_error;
break;
}
++npseg;
}
return npseg;
}
void
calc_cb(struct clfs *fs, int sn, struct clfs_seguse *t)
{
time_t now;
int64_t age, benefit, cost;
time(&now);
age = (now < t->lastmod ? 0 : now - t->lastmod);
/* Under no circumstances clean active or already-clean segments */
if ((t->flags & SEGUSE_ACTIVE) || !(t->flags & SEGUSE_DIRTY)) {
t->priority = 0;
return;
}
/*
* If the segment is empty, there is no reason to clean it.
* Clear its error condition, if any, since we are never going to
* try to parse this one.
*/
if (t->nbytes == 0) {
t->flags &= ~SEGUSE_ERROR; /* Strip error once empty */
t->priority = 0;
return;
}
if (t->flags & SEGUSE_ERROR) { /* No good if not already empty */
/* No benefit */
t->priority = 0;
return;
}
if (t->nbytes > lfs_sb_getssize(fs)) {
/* Another type of error */
syslog(LOG_WARNING, "segment %d: bad seguse count %d",
sn, t->nbytes);
t->flags |= SEGUSE_ERROR;
t->priority = 0;
return;
}
/*
* The non-degenerate case. Use Rosenblum's cost-benefit algorithm.
* Calculate the benefit from cleaning this segment (one segment,
* minus fragmentation, dirty blocks and a segment summary block)
* and weigh that against the cost (bytes read plus bytes written).
* We count the summary headers as "dirty" to avoid cleaning very
* old and very full segments.
*/
benefit = (int64_t)lfs_sb_getssize(fs) - t->nbytes -
(t->nsums + 1) * lfs_sb_getfsize(fs);
if (lfs_sb_getbsize(fs) > lfs_sb_getfsize(fs)) /* fragmentation */
benefit -= (lfs_sb_getbsize(fs) / 2);
if (benefit <= 0) {
t->priority = 0;
return;
}
cost = lfs_sb_getssize(fs) + t->nbytes;
t->priority = (256 * benefit * age) / cost;
return;
}
/*
* Comparator for BLOCK_INFO structures. Anything not in one of the segments
* we're looking at sorts higher; after that we sort first by inode number
* and then by block number (unsigned, i.e., negative sorts higher) *but*
* sort inodes before data blocks.
*/
static int
bi_comparator(const void *va, const void *vb)
{
const BLOCK_INFO *a, *b;
a = (const BLOCK_INFO *)va;
b = (const BLOCK_INFO *)vb;
/* Check for out-of-place block */
if (a->bi_segcreate == a->bi_daddr &&
b->bi_segcreate != b->bi_daddr)
return -1;
if (a->bi_segcreate != a->bi_daddr &&
b->bi_segcreate == b->bi_daddr)
return 1;
if (a->bi_size <= 0 && b->bi_size > 0)
return 1;
if (b->bi_size <= 0 && a->bi_size > 0)
return -1;
/* Check inode number */
if (a->bi_inode != b->bi_inode)
return a->bi_inode - b->bi_inode;
/* Check lbn */
if (a->bi_lbn == LFS_UNUSED_LBN) /* Inodes sort lower than blocks */
return -1;
if (b->bi_lbn == LFS_UNUSED_LBN)
return 1;
if ((u_int64_t)a->bi_lbn > (u_int64_t)b->bi_lbn)
return 1;
else
return -1;
return 0;
}
/*
* Comparator for sort_segments: cost-benefit equation.
*/
static int
cb_comparator(const void *va, const void *vb)
{
const struct clfs_seguse *a, *b;
a = *(const struct clfs_seguse * const *)va;
b = *(const struct clfs_seguse * const *)vb;
return a->priority > b->priority ? -1 : 1;
}
void
toss_old_blocks(struct clfs *fs, BLOCK_INFO **bipp, blkcnt_t *bic, int *sizep)
{
blkcnt_t i;
int r;
BLOCK_INFO *bip = *bipp;
struct lfs_fcntl_markv /* {
BLOCK_INFO *blkiov;
int blkcnt;
} */ lim;
if (bic == 0 || bip == NULL)
return;
/*
* Kludge: Store the disk address in segcreate so we know which
* ones to toss.
*/
for (i = 0; i < *bic; i++)
bip[i].bi_segcreate = bip[i].bi_daddr;
/*
* XXX: blkcnt_t is 64 bits, so *bic might overflow size_t
* (the argument type of heapsort's number argument) on a
* 32-bit platform. However, if so we won't have got this far
* because we'll have failed trying to allocate the array. So
* while *bic here might cause a 64->32 truncation, it's safe.
*/
/* Sort the blocks */
heapsort(bip, *bic, sizeof(BLOCK_INFO), bi_comparator);
/* Use bmapv to locate the blocks */
lim.blkiov = bip;
lim.blkcnt = *bic;
if ((r = kops.ko_fcntl(fs->clfs_ifilefd, LFCNBMAPV, &lim)) < 0) {
syslog(LOG_WARNING, "%s: bmapv returned %d (%m)",
lfs_sb_getfsmnt(fs), r);
return;
}
/* Toss blocks not in this segment */
heapsort(bip, *bic, sizeof(BLOCK_INFO), bi_comparator);
/* Get rid of stale blocks */
if (sizep)
*sizep = 0;
for (i = 0; i < *bic; i++) {
if (bip[i].bi_segcreate != bip[i].bi_daddr)
break;
if (sizep)
*sizep += bip[i].bi_size;
}
*bic = i; /* XXX should we shrink bip? */
*bipp = bip;
return;
}
/*
* Clean a segment and mark it invalid.
*/
int
invalidate_segment(struct clfs *fs, int sn)
{
BLOCK_INFO *bip;
int i, r, bic;
blkcnt_t widebic;
off_t nb;
double util;
struct lfs_fcntl_markv /* {
BLOCK_INFO *blkiov;
int blkcnt;
} */ lim;
dlog("%s: inval seg %d", lfs_sb_getfsmnt(fs), sn);
bip = NULL;
bic = 0;
fs->clfs_nactive = 0;
if (load_segment(fs, sn, &bip, &bic) <= 0)
return -1;
widebic = bic;
toss_old_blocks(fs, &bip, &widebic, NULL);
bic = widebic;
/* Record statistics */
for (i = nb = 0; i < bic; i++)
nb += bip[i].bi_size;
util = ((double)nb) / (fs->clfs_nactive * lfs_sb_getssize(fs));
cleaner_stats.util_tot += util;
cleaner_stats.util_sos += util * util;
cleaner_stats.bytes_written += nb;
/*
* Use markv to move the blocks.
*/
lim.blkiov = bip;
lim.blkcnt = bic;
if ((r = kops.ko_fcntl(fs->clfs_ifilefd, LFCNMARKV, &lim)) < 0) {
syslog(LOG_WARNING, "%s: markv returned %d (%m) "
"for seg %d", lfs_sb_getfsmnt(fs), r, sn);
return r;
}
/*
* Finally call invalidate to invalidate the segment.
*/
if ((r = kops.ko_fcntl(fs->clfs_ifilefd, LFCNINVAL, &sn)) < 0) {
syslog(LOG_WARNING, "%s: inval returned %d (%m) "
"for seg %d", lfs_sb_getfsmnt(fs), r, sn);
return r;
}
return 0;
}
/*
* Check to see if the given ino/lbn pair is represented in the BLOCK_INFO
* array we are sending to the kernel, or if the kernel will have to add it.
* The kernel will only add each such pair once, though, so keep track of
* previous requests in a separate "extra" BLOCK_INFO array. Returns 1
* if the block needs to be added, 0 if it is already represented.
*/
static int
check_or_add(ino_t ino, daddr_t lbn, BLOCK_INFO *bip, int bic, BLOCK_INFO **ebipp, int *ebicp)
{
BLOCK_INFO *t, *ebip = *ebipp;
int ebic = *ebicp;
int k;
for (k = 0; k < bic; k++) {
if (bip[k].bi_inode != ino)
break;
if (bip[k].bi_lbn == lbn) {
return 0;
}
}
/* Look on the list of extra blocks, too */
for (k = 0; k < ebic; k++) {
if (ebip[k].bi_inode == ino && ebip[k].bi_lbn == lbn) {
return 0;
}
}
++ebic;
t = realloc(ebip, ebic * sizeof(BLOCK_INFO));
if (t == NULL)
return 1; /* Note *ebicp is unchanged */
ebip = t;
ebip[ebic - 1].bi_inode = ino;
ebip[ebic - 1].bi_lbn = lbn;
*ebipp = ebip;
*ebicp = ebic;
return 1;
}
/*
* Look for indirect blocks we will have to write which are not
* contained in this collection of blocks. This constitutes
* a hidden cleaning cost, since we are unaware of it until we
* have already read the segments. Return the total cost, and fill
* in *ifc with the part of that cost due to rewriting the Ifile.
*/
static off_t
check_hidden_cost(struct clfs *fs, BLOCK_INFO *bip, int bic, off_t *ifc)
{
int start;
struct indir in[ULFS_NIADDR + 1];
int num;
int i, j, ebic;
BLOCK_INFO *ebip;
daddr_t lbn;
start = 0;
ebip = NULL;
ebic = 0;
for (i = 0; i < bic; i++) {
if (i == 0 || bip[i].bi_inode != bip[start].bi_inode) {
start = i;
/*
* Look for IFILE blocks, unless this is the Ifile.
*/
if (bip[i].bi_inode != LFS_IFILE_INUM) {
lbn = lfs_sb_getcleansz(fs) + bip[i].bi_inode /
lfs_sb_getifpb(fs);
*ifc += check_or_add(LFS_IFILE_INUM, lbn,
bip, bic, &ebip, &ebic);
}
}
if (bip[i].bi_lbn == LFS_UNUSED_LBN)
continue;
if (bip[i].bi_lbn < ULFS_NDADDR)
continue;
/* XXX the struct lfs cast is completely wrong/unsafe */
ulfs_getlbns((struct lfs *)fs, NULL, (daddr_t)bip[i].bi_lbn, in, &num);
for (j = 0; j < num; j++) {
check_or_add(bip[i].bi_inode, in[j].in_lbn,
bip + start, bic - start, &ebip, &ebic);
}
}
return ebic;
}
/*
* Select segments to clean, add blocks from these segments to a cleaning
* list, and send this list through lfs_markv() to move them to new
* locations on disk.
*/
static int
clean_fs(struct clfs *fs, const CLEANERINFO64 *cip)
{
int i, j, ngood, sn, bic, r, npos;
blkcnt_t widebic;
int bytes, totbytes;
struct ubuf *bp;
SEGUSE *sup;
static BLOCK_INFO *bip;
struct lfs_fcntl_markv /* {
BLOCK_INFO *blkiov;
int blkcnt;
} */ lim;
int mc;
BLOCK_INFO *mbip;
int inc;
off_t nb;
off_t goal;
off_t extra, if_extra;
double util;
/* Read the segment table into our private structure */
npos = 0;
for (i = 0; i < lfs_sb_getnseg(fs); i+= lfs_sb_getsepb(fs)) {
bread(fs->lfs_ivnode,
lfs_sb_getcleansz(fs) + i / lfs_sb_getsepb(fs),
lfs_sb_getbsize(fs), 0, &bp);
for (j = 0; j < lfs_sb_getsepb(fs) && i + j < lfs_sb_getnseg(fs); j++) {
sup = ((SEGUSE *)bp->b_data) + j;
fs->clfs_segtab[i + j].nbytes = sup->su_nbytes;
fs->clfs_segtab[i + j].nsums = sup->su_nsums;
fs->clfs_segtab[i + j].lastmod = sup->su_lastmod;
/* Keep error status but renew other flags */
fs->clfs_segtab[i + j].flags &= SEGUSE_ERROR;
fs->clfs_segtab[i + j].flags |= sup->su_flags;
/* Compute cost-benefit coefficient */
calc_cb(fs, i + j, fs->clfs_segtab + i + j);
if (fs->clfs_segtab[i + j].priority > 0)
++npos;
}
brelse(bp, 0);
}
/* Sort segments based on cleanliness, fulness, and condition */
heapsort(fs->clfs_segtabp, lfs_sb_getnseg(fs), sizeof(struct clfs_seguse *),
cb_comparator);
/* If no segment is cleanable, just return */
if (fs->clfs_segtabp[0]->priority == 0) {
dlog("%s: no segment cleanable", lfs_sb_getfsmnt(fs));
return 0;
}
/* Load some segments' blocks into bip */
bic = 0;
fs->clfs_nactive = 0;
ngood = 0;
if (use_bytes) {
/* Set attainable goal */
goal = lfs_sb_getssize(fs) * atatime;
if (goal > (cip->clean - 1) * lfs_sb_getssize(fs) / 2)
goal = MAX((cip->clean - 1) * lfs_sb_getssize(fs),
lfs_sb_getssize(fs)) / 2;
dlog("%s: cleaning with goal %" PRId64
" bytes (%d segs clean, %d cleanable)",
lfs_sb_getfsmnt(fs), goal, cip->clean, npos);
syslog(LOG_INFO, "%s: cleaning with goal %" PRId64
" bytes (%d segs clean, %d cleanable)",
lfs_sb_getfsmnt(fs), goal, cip->clean, npos);
totbytes = 0;
for (i = 0; i < lfs_sb_getnseg(fs) && totbytes < goal; i++) {
if (fs->clfs_segtabp[i]->priority == 0)
break;
/* Upper bound on number of segments at once */
if (ngood * lfs_sb_getssize(fs) > 4 * goal)
break;
sn = (fs->clfs_segtabp[i] - fs->clfs_segtab);
dlog("%s: add seg %d prio %" PRIu64
" containing %ld bytes",
lfs_sb_getfsmnt(fs), sn, fs->clfs_segtabp[i]->priority,
fs->clfs_segtabp[i]->nbytes);
if ((r = load_segment(fs, sn, &bip, &bic)) > 0) {
++ngood;
widebic = bic;
toss_old_blocks(fs, &bip, &widebic, &bytes);
bic = widebic;
totbytes += bytes;
} else if (r == 0)
fd_release(fs->clfs_devvp);
else
break;
}
} else {
/* Set attainable goal */
goal = atatime;
if (goal > cip->clean - 1)
goal = MAX(cip->clean - 1, 1);
dlog("%s: cleaning with goal %d segments (%d clean, %d cleanable)",
lfs_sb_getfsmnt(fs), (int)goal, cip->clean, npos);
for (i = 0; i < lfs_sb_getnseg(fs) && ngood < goal; i++) {
if (fs->clfs_segtabp[i]->priority == 0)
break;
sn = (fs->clfs_segtabp[i] - fs->clfs_segtab);
dlog("%s: add seg %d prio %" PRIu64,
lfs_sb_getfsmnt(fs), sn, fs->clfs_segtabp[i]->priority);
if ((r = load_segment(fs, sn, &bip, &bic)) > 0)
++ngood;
else if (r == 0)
fd_release(fs->clfs_devvp);
else
break;
}
widebic = bic;
toss_old_blocks(fs, &bip, &widebic, NULL);
bic = widebic;
}
/* If there is nothing to do, try again later. */
if (bic == 0) {
dlog("%s: no blocks to clean in %d cleanable segments",
lfs_sb_getfsmnt(fs), (int)ngood);
fd_release_all(fs->clfs_devvp);
return 0;
}
/* Record statistics */
for (i = nb = 0; i < bic; i++)
nb += bip[i].bi_size;
util = ((double)nb) / (fs->clfs_nactive * lfs_sb_getssize(fs));
cleaner_stats.util_tot += util;
cleaner_stats.util_sos += util * util;
cleaner_stats.bytes_written += nb;
/*
* Check out our blocks to see if there are hidden cleaning costs.
* If there are, we might be cleaning ourselves deeper into a hole
* rather than doing anything useful.
* XXX do something about this.
*/
if_extra = 0;
extra = lfs_sb_getbsize(fs) * (off_t)check_hidden_cost(fs, bip, bic, &if_extra);
if_extra *= lfs_sb_getbsize(fs);
/*
* Use markv to move the blocks.
*/
if (do_small)
inc = MAXPHYS / lfs_sb_getbsize(fs) - 1;
else
inc = LFS_MARKV_MAXBLKCNT / 2;
for (mc = 0, mbip = bip; mc < bic; mc += inc, mbip += inc) {
lim.blkiov = mbip;
lim.blkcnt = (bic - mc > inc ? inc : bic - mc);
#ifdef TEST_PATTERN
dlog("checking blocks %d-%d", mc, mc + lim.blkcnt - 1);
for (i = 0; i < lim.blkcnt; i++) {
check_test_pattern(mbip + i);
}
#endif /* TEST_PATTERN */
dlog("sending blocks %d-%d", mc, mc + lim.blkcnt - 1);
if ((r = kops.ko_fcntl(fs->clfs_ifilefd, LFCNMARKV, &lim))<0) {
int oerrno = errno;
syslog(LOG_WARNING, "%s: markv returned %d (errno %d, %m)",
lfs_sb_getfsmnt(fs), r, errno);
if (oerrno != EAGAIN && oerrno != ESHUTDOWN) {
syslog(LOG_DEBUG, "%s: errno %d, returning",
lfs_sb_getfsmnt(fs), oerrno);
fd_release_all(fs->clfs_devvp);
return r;
}
if (oerrno == ESHUTDOWN) {
syslog(LOG_NOTICE, "%s: filesystem unmounted",
lfs_sb_getfsmnt(fs));
fd_release_all(fs->clfs_devvp);
return r;
}
}
}
/*
* Report progress (or lack thereof)
*/
syslog(LOG_INFO, "%s: wrote %" PRId64 " dirty + %"
PRId64 " supporting indirect + %"
PRId64 " supporting Ifile = %"
PRId64 " bytes to clean %d segs (%" PRId64 "%% recovery)",
lfs_sb_getfsmnt(fs), (int64_t)nb, (int64_t)(extra - if_extra),
(int64_t)if_extra, (int64_t)(nb + extra), ngood,
(ngood ? (int64_t)(100 - (100 * (nb + extra)) /
(ngood * lfs_sb_getssize(fs))) :
(int64_t)0));
if (nb + extra >= ngood * lfs_sb_getssize(fs))
syslog(LOG_WARNING, "%s: cleaner not making forward progress",
lfs_sb_getfsmnt(fs));
/*
* Finally call reclaim to prompt cleaning of the segments.
*/
kops.ko_fcntl(fs->clfs_ifilefd, LFCNRECLAIM, NULL);
fd_release_all(fs->clfs_devvp);
return 0;
}
/*
* Read the cleanerinfo block and apply cleaning policy to determine whether
* the given filesystem needs to be cleaned. Returns 1 if it does, 0 if it
* does not, or -1 on error.
*/
static int
needs_cleaning(struct clfs *fs, CLEANERINFO64 *cip)
{
CLEANERINFO *cipu;
struct ubuf *bp;
struct stat st;
daddr_t fsb_per_seg, max_free_segs;
time_t now;
double loadavg;
/* If this fs is "on hold", don't clean it. */
if (fs->clfs_onhold) {
#if defined(__GNUC__) && \
(__GNUC__ > 4 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8)) && \
defined(__OPTIMIZE_SIZE__)
/*
* XXX: Work around apparent bug with GCC >= 4.8 and -Os: it
* claims that ci.clean is uninitialized in clean_fs (at one
* of the several uses of it, which is neither the first nor
* last use) -- this doesn't happen with plain -O2.
*
* Hopefully in the future further rearrangements will allow
* removing this hack.
*/
cip->clean = 0;
#endif
return 0;
}
/*
* Read the cleanerinfo block from the Ifile. We don't want
* the cached information, so invalidate the buffer before
* handing it back.
*/
if (bread(fs->lfs_ivnode, 0, lfs_sb_getbsize(fs), 0, &bp)) {
syslog(LOG_ERR, "%s: can't read inode", lfs_sb_getfsmnt(fs));
return -1;
}
cipu = (CLEANERINFO *)bp->b_data;
if (fs->lfs_is64) {
/* Structure copy */
*cip = cipu->u_64;
} else {
/* Copy the fields and promote to 64 bit */
cip->clean = cipu->u_32.clean;
cip->dirty = cipu->u_32.dirty;
cip->bfree = cipu->u_32.bfree;
cip->avail = cipu->u_32.avail;
cip->free_head = cipu->u_32.free_head;
cip->free_tail = cipu->u_32.free_tail;
cip->flags = cipu->u_32.flags;
}
brelse(bp, B_INVAL);
cleaner_stats.bytes_read += lfs_sb_getbsize(fs);
/*
* If the number of segments changed under us, reinit.
* We don't have to start over from scratch, however,
* since we don't hold any buffers.
*/
if (lfs_sb_getnseg(fs) != cip->clean + cip->dirty) {
if (reinit_fs(fs) < 0) {
/* The normal case for unmount */
syslog(LOG_NOTICE, "%s: filesystem unmounted", lfs_sb_getfsmnt(fs));
return -1;
}
syslog(LOG_NOTICE, "%s: nsegs changed", lfs_sb_getfsmnt(fs));
}
/* Compute theoretical "free segments" maximum based on usage */
fsb_per_seg = lfs_segtod(fs, 1);
max_free_segs = MAX(cip->bfree, 0) / fsb_per_seg + lfs_sb_getminfreeseg(fs);
dlog("%s: bfree = %d, avail = %d, clean = %d/%d",
lfs_sb_getfsmnt(fs), cip->bfree, cip->avail, cip->clean,
lfs_sb_getnseg(fs));
/* If the writer is waiting on us, clean it */
if (cip->clean <= lfs_sb_getminfreeseg(fs) ||
(cip->flags & LFS_CLEANER_MUST_CLEAN))
return 1;
/* If there are enough segments, don't clean it */
if (cip->bfree - cip->avail <= fsb_per_seg &&
cip->avail > fsb_per_seg)
return 0;
/* If we are in dire straits, clean it */
if (cip->bfree - cip->avail > fsb_per_seg &&
cip->avail <= fsb_per_seg)
return 1;
/* If under busy threshold, clean regardless of load */
if (cip->clean < max_free_segs * BUSY_LIM)
return 1;
/* Check busy status; clean if idle and under idle limit */
if (use_fs_idle) {
/* Filesystem idle */
time(&now);
if (fstat(fs->clfs_ifilefd, &st) < 0) {
syslog(LOG_ERR, "%s: failed to stat ifile",
lfs_sb_getfsmnt(fs));
return -1;
}
if (now - st.st_mtime > segwait_timeout &&
cip->clean < max_free_segs * IDLE_LIM)
return 1;
} else {
/* CPU idle - use one-minute load avg */
if (getloadavg(&loadavg, 1) == -1) {
syslog(LOG_ERR, "%s: failed to get load avg",
lfs_sb_getfsmnt(fs));
return -1;
}
if (loadavg < load_threshold &&
cip->clean < max_free_segs * IDLE_LIM)
return 1;
}
return 0;
}
/*
* Report statistics. If the signal was SIGUSR2, clear the statistics too.
* If the signal was SIGINT, exit.
*/
static void
sig_report(int sig)
{
double avg = 0.0, stddev;
avg = cleaner_stats.util_tot / MAX(cleaner_stats.segs_cleaned, 1.0);
stddev = cleaner_stats.util_sos / MAX(cleaner_stats.segs_cleaned -
avg * avg, 1.0);
syslog(LOG_INFO, "bytes read: %" PRId64, cleaner_stats.bytes_read);
syslog(LOG_INFO, "bytes written: %" PRId64, cleaner_stats.bytes_written);
syslog(LOG_INFO, "segments cleaned: %" PRId64, cleaner_stats.segs_cleaned);
#if 0
/* "Empty segments" is meaningless, since the kernel handles those */
syslog(LOG_INFO, "empty segments: %" PRId64, cleaner_stats.segs_empty);
#endif
syslog(LOG_INFO, "error segments: %" PRId64, cleaner_stats.segs_error);
syslog(LOG_INFO, "utilization total: %g", cleaner_stats.util_tot);
syslog(LOG_INFO, "utilization sos: %g", cleaner_stats.util_sos);
syslog(LOG_INFO, "utilization avg: %4.2f", avg);
syslog(LOG_INFO, "utilization sdev: %9.6f", stddev);
if (debug)
bufstats();
if (sig == SIGUSR2)
memset(&cleaner_stats, 0, sizeof(cleaner_stats));
if (sig == SIGINT)
exit(0);
}
static void
sig_exit(int sig)
{
exit(0);
}
static void
usage(void)
{
errx(1, "usage: lfs_cleanerd [-bcdfmqs] [-i segnum] [-l load] "
"[-n nsegs] [-r report_freq] [-t timeout] fs_name ...");
}
#ifndef LFS_CLEANER_AS_LIB
/*
* Main.
*/
int
main(int argc, char **argv)
{
return lfs_cleaner_main(argc, argv);
}
#endif
int
lfs_cleaner_main(int argc, char **argv)
{
int i, opt, error, r, loopcount, nodetach;
struct timeval tv;
#ifdef LFS_CLEANER_AS_LIB
sem_t *semaddr = NULL;
#endif
CLEANERINFO64 ci;
#ifndef USE_CLIENT_SERVER
char *cp, *pidname;
#endif
/*
* Set up defaults
*/
atatime = 1;
segwait_timeout = 300; /* Five minutes */
load_threshold = 0.2;
stat_report = 0;
inval_segment = -1;
copylog_filename = NULL;
nodetach = 0;
/*
* Parse command-line arguments
*/
while ((opt = getopt(argc, argv, "bC:cdDfi:l:mn:qr:sS:t:")) != -1) {
switch (opt) {
case 'b': /* Use bytes written, not segments read */
use_bytes = 1;
break;
case 'C': /* copy log */
copylog_filename = optarg;
break;
case 'c': /* Coalesce files */
do_coalesce++;
break;
case 'd': /* Debug mode. */
nodetach++;
debug++;
break;
case 'D': /* stay-on-foreground */
nodetach++;
break;
case 'f': /* Use fs idle time rather than cpu idle */
use_fs_idle = 1;
break;
case 'i': /* Invalidate this segment */
inval_segment = atoi(optarg);
break;
case 'l': /* Load below which to clean */
load_threshold = atof(optarg);
break;
case 'm': /* [compat only] */
break;
case 'n': /* How many segs to clean at once */
atatime = atoi(optarg);
break;
case 'q': /* Quit after one run */
do_quit = 1;
break;
case 'r': /* Report every stat_report segments */
stat_report = atoi(optarg);
break;
case 's': /* Small writes */
do_small = 1;
break;
#ifdef LFS_CLEANER_AS_LIB
case 'S': /* semaphore */
semaddr = (void*)(uintptr_t)strtoull(optarg,NULL,0);
break;
#endif
case 't': /* timeout */
segwait_timeout = atoi(optarg);
break;
default:
usage();
/* NOTREACHED */
}
}
argc -= optind;
argv += optind;
if (argc < 1)
usage();
if (inval_segment >= 0 && argc != 1) {
errx(1, "lfs_cleanerd: may only specify one filesystem when "
"using -i flag");
}
if (do_coalesce) {
errx(1, "lfs_cleanerd: -c disabled due to reports of file "
"corruption; you may re-enable it by rebuilding the "
"cleaner");
}
/*
* Set up daemon mode or foreground mode
*/
if (nodetach) {
openlog("lfs_cleanerd", LOG_NDELAY | LOG_PID | LOG_PERROR,
LOG_DAEMON);
signal(SIGINT, sig_report);
} else {
if (daemon(0, 0) == -1)
err(1, "lfs_cleanerd: couldn't become a daemon!");
openlog("lfs_cleanerd", LOG_NDELAY | LOG_PID, LOG_DAEMON);
signal(SIGINT, sig_exit);
}
/*
* Look for an already-running master daemon. If there is one,
* send it our filesystems to add to its list and exit.
* If there is none, become the master.
*/
#ifdef USE_CLIENT_SERVER
try_to_become_master(argc, argv);
#else
/* XXX think about this */
asprintf(&pidname, "lfs_cleanerd:m:%s", argv[0]);
if (pidname == NULL) {
syslog(LOG_ERR, "malloc failed: %m");
exit(1);
}
for (cp = pidname; cp != NULL; cp = strchr(cp, '/'))
*cp = '|';
pidfile(pidname);
#endif
/*
* Signals mean daemon should report its statistics
*/
memset(&cleaner_stats, 0, sizeof(cleaner_stats));
signal(SIGUSR1, sig_report);
signal(SIGUSR2, sig_report);
/*
* Start up buffer cache. We only use this for the Ifile,
* and we will resize it if necessary, so it can start small.
*/
bufinit(4);
#ifdef REPAIR_ZERO_FINFO
{
BLOCK_INFO *bip = NULL;
int bic = 0;
nfss = 1;
fsp = (struct clfs **)malloc(sizeof(*fsp));
fsp[0] = (struct clfs *)calloc(1, sizeof(**fsp));
if (init_unmounted_fs(fsp[0], argv[0]) < 0) {
err(1, "init_unmounted_fs");
}
dlog("Filesystem has %d segments", fsp[0]->lfs_nseg);
for (i = 0; i < fsp[0]->lfs_nseg; i++) {
load_segment(fsp[0], i, &bip, &bic);
bic = 0;
}
exit(0);
}
#endif
/*
* Initialize cleaning structures, open devices, etc.
*/
nfss = argc;
fsp = (struct clfs **)malloc(nfss * sizeof(*fsp));
if (fsp == NULL) {
syslog(LOG_ERR, "couldn't allocate fs table: %m");
exit(1);
}
for (i = 0; i < nfss; i++) {
fsp[i] = (struct clfs *)calloc(1, sizeof(**fsp));
if ((r = init_fs(fsp[i], argv[i])) < 0) {
syslog(LOG_ERR, "%s: couldn't init: error code %d",
argv[i], r);
handle_error(fsp, i);
--i; /* Do the new #i over again */
}
}
/*
* If asked to coalesce, do so and exit.
*/
if (do_coalesce) {
for (i = 0; i < nfss; i++)
clean_all_inodes(fsp[i]);
exit(0);
}
/*
* If asked to invalidate a segment, do that and exit.
*/
if (inval_segment >= 0) {
invalidate_segment(fsp[0], inval_segment);
exit(0);
}
/*
* Main cleaning loop.
*/
loopcount = 0;
#ifdef LFS_CLEANER_AS_LIB
if (semaddr)
sem_post(semaddr);
#endif
error = 0;
while (nfss > 0) {
int cleaned_one;
do {
#ifdef USE_CLIENT_SERVER
check_control_socket();
#endif
cleaned_one = 0;
for (i = 0; i < nfss; i++) {
if ((error = needs_cleaning(fsp[i], &ci)) < 0) {
syslog(LOG_DEBUG, "%s: needs_cleaning returned %d",
getprogname(), error);
handle_error(fsp, i);
continue;
}
if (error == 0) /* No need to clean */
continue;
reload_ifile(fsp[i]);
if ((error = clean_fs(fsp[i], &ci)) < 0) {
syslog(LOG_DEBUG, "%s: clean_fs returned %d",
getprogname(), error);
handle_error(fsp, i);
continue;
}
++cleaned_one;
}
++loopcount;
if (stat_report && loopcount % stat_report == 0)
sig_report(0);
if (do_quit)
exit(0);
} while(cleaned_one);
tv.tv_sec = segwait_timeout;
tv.tv_usec = 0;
/* XXX: why couldn't others work if fsp socket is shutdown? */
error = kops.ko_fcntl(fsp[0]->clfs_ifilefd,LFCNSEGWAITALL,&tv);
if (error) {
if (errno == ESHUTDOWN) {
for (i = 0; i < nfss; i++) {
syslog(LOG_INFO, "%s: shutdown",
getprogname());
handle_error(fsp, i);
assert(nfss == 0);
}
} else {
#ifdef LFS_CLEANER_AS_LIB
error = ESHUTDOWN;
break;
#else
err(1, "LFCNSEGWAITALL");
#endif
}
}
}
/* NOTREACHED */
return error;
}