aboutsummaryrefslogtreecommitdiff
path: root/sys/sys/buf.h
diff options
context:
space:
mode:
Diffstat (limited to 'sys/sys/buf.h')
-rw-r--r--sys/sys/buf.h374
1 files changed, 374 insertions, 0 deletions
diff --git a/sys/sys/buf.h b/sys/sys/buf.h
new file mode 100644
index 0000000..ea01ac0
--- /dev/null
+++ b/sys/sys/buf.h
@@ -0,0 +1,374 @@
+/* $OpenBSD: buf.h,v 1.112 2019/11/29 01:04:08 beck Exp $ */
+/* $NetBSD: buf.h,v 1.25 1997/04/09 21:12:17 mycroft Exp $ */
+
+/*
+ * Copyright (c) 1982, 1986, 1989, 1993
+ * The Regents of the University of California. All rights reserved.
+ * (c) UNIX System Laboratories, Inc.
+ * All or some portions of this file are derived from material licensed
+ * to the University of California by American Telephone and Telegraph
+ * Co. or Unix System Laboratories, Inc. and are reproduced herein with
+ * the permission of UNIX System Laboratories, Inc.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ * 3. Neither the name of the University nor the names of its contributors
+ * may be used to endorse or promote products derived from this software
+ * without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
+ * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
+ * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+ * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
+ * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
+ * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
+ * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ *
+ * @(#)buf.h 8.7 (Berkeley) 1/21/94
+ */
+
+#ifndef _SYS_BUF_H_
+#define _SYS_BUF_H_
+#include <sys/queue.h>
+#include <sys/tree.h>
+#include <sys/mutex.h>
+#include <uvm/uvm_extern.h>
+
+#define NOLIST ((struct buf *)0x87654321)
+
+struct buf;
+struct vnode;
+
+LIST_HEAD(bufhead, buf);
+
+/*
+ * To avoid including <ufs/ffs/softdep.h>
+ */
+
+LIST_HEAD(workhead, worklist);
+
+/*
+ * Buffer queues
+ */
+#define BUFQ_NSCAN_N 128
+#define BUFQ_FIFO 0
+#define BUFQ_NSCAN 1
+#define BUFQ_DEFAULT BUFQ_NSCAN
+#define BUFQ_HOWMANY 2
+
+/*
+ * Write limits for bufq - defines high and low water marks for how
+ * many kva slots are allowed to be consumed to parallelize writes from
+ * the buffer cache from any individual bufq.
+ */
+#define BUFQ_HI 128
+#define BUFQ_LOW 64
+
+struct bufq_impl;
+
+struct bufq {
+ SLIST_ENTRY(bufq) bufq_entries;
+ struct mutex bufq_mtx;
+ void *bufq_data;
+ u_int bufq_outstanding;
+ u_int bufq_hi;
+ u_int bufq_low;
+ int bufq_waiting;
+ int bufq_stop;
+ int bufq_type;
+ const struct bufq_impl *bufq_impl;
+};
+
+int bufq_init(struct bufq *, int);
+int bufq_switch(struct bufq *, int);
+void bufq_destroy(struct bufq *);
+
+void bufq_queue(struct bufq *, struct buf *);
+struct buf *bufq_dequeue(struct bufq *);
+void bufq_requeue(struct bufq *, struct buf *);
+int bufq_peek(struct bufq *);
+void bufq_drain(struct bufq *);
+
+void bufq_wait(struct bufq *);
+void bufq_done(struct bufq *, struct buf *);
+void bufq_quiesce(void);
+void bufq_restart(void);
+
+/* fifo */
+SIMPLEQ_HEAD(bufq_fifo_head, buf);
+struct bufq_fifo {
+ SIMPLEQ_ENTRY(buf) bqf_entries;
+};
+
+/* nscan */
+SIMPLEQ_HEAD(bufq_nscan_head, buf);
+struct bufq_nscan {
+ SIMPLEQ_ENTRY(buf) bqf_entries;
+};
+
+/* bufq link in struct buf */
+union bufq_data {
+ struct bufq_fifo bufq_data_fifo;
+ struct bufq_nscan bufq_data_nscan;
+};
+
+/*
+ * These are currently used only by the soft dependency code, hence
+ * are stored once in a global variable. If other subsystems wanted
+ * to use these hooks, a pointer to a set of bio_ops could be added
+ * to each buffer.
+ */
+extern struct bio_ops {
+ void (*io_start)(struct buf *);
+ void (*io_complete)(struct buf *);
+ void (*io_deallocate)(struct buf *);
+ void (*io_movedeps)(struct buf *, struct buf *);
+ int (*io_countdeps)(struct buf *, int, int);
+} bioops;
+
+/* The buffer header describes an I/O operation in the kernel. */
+struct buf {
+ RBT_ENTRY(buf) b_rbbufs; /* vnode "hash" tree */
+ LIST_ENTRY(buf) b_list; /* All allocated buffers. */
+ LIST_ENTRY(buf) b_vnbufs; /* Buffer's associated vnode. */
+ TAILQ_ENTRY(buf) b_freelist; /* Free list position if not active. */
+ int cache; /* which cache are we in */
+ struct proc *b_proc; /* Associated proc; NULL if kernel. */
+ volatile long b_flags; /* B_* flags. */
+ long b_bufsize; /* Allocated buffer size. */
+ long b_bcount; /* Valid bytes in buffer. */
+ size_t b_resid; /* Remaining I/O. */
+ int b_error; /* Errno value. */
+ dev_t b_dev; /* Device associated with buffer. */
+ caddr_t b_data; /* associated data */
+ void *b_saveaddr; /* Original b_data for physio. */
+
+ TAILQ_ENTRY(buf) b_valist; /* LRU of va to reuse. */
+
+ union bufq_data b_bufq;
+ struct bufq *b_bq; /* What bufq this buf is on */
+
+ struct uvm_object *b_pobj;
+ struct uvm_object b_uobj; /* Object containing the pages */
+ off_t b_poffs; /* Offset within object */
+
+ daddr_t b_lblkno; /* Logical block number. */
+ daddr_t b_blkno; /* Underlying physical block number. */
+ /* Function to call upon completion.
+ * Will be called at splbio(). */
+ void (*b_iodone)(struct buf *);
+ struct vnode *b_vp; /* Device vnode. */
+ int b_dirtyoff; /* Offset in buffer of dirty region. */
+ int b_dirtyend; /* Offset of end of dirty region. */
+ int b_validoff; /* Offset in buffer of valid region. */
+ int b_validend; /* Offset of end of valid region. */
+ struct workhead b_dep; /* List of filesystem dependencies. */
+};
+
+TAILQ_HEAD(bufqueue, buf);
+
+struct bufcache {
+ int64_t hotbufpages;
+ int64_t warmbufpages;
+ int64_t cachepages;
+ struct bufqueue hotqueue;
+ struct bufqueue coldqueue;
+ struct bufqueue warmqueue;
+};
+
+/* Device driver compatibility definitions. */
+#define b_active b_bcount /* Driver queue head: drive active. */
+
+/*
+ * These flags are kept in b_flags.
+ */
+#define B_WRITE 0x00000000 /* Write buffer (pseudo flag). */
+#define B_AGE 0x00000001 /* Move to age queue when I/O done. */
+#define B_NEEDCOMMIT 0x00000002 /* Needs committing to stable storage */
+#define B_ASYNC 0x00000004 /* Start I/O, do not wait. */
+#define B_BAD 0x00000008 /* Bad block revectoring in progress. */
+#define B_BUSY 0x00000010 /* I/O in progress. */
+#define B_CACHE 0x00000020 /* Bread found us in the cache. */
+#define B_CALL 0x00000040 /* Call b_iodone from biodone. */
+#define B_DELWRI 0x00000080 /* Delay I/O until buffer reused. */
+#define B_DONE 0x00000100 /* I/O completed. */
+#define B_EINTR 0x00000200 /* I/O was interrupted */
+#define B_ERROR 0x00000400 /* I/O error occurred. */
+#define B_INVAL 0x00000800 /* Does not contain valid info. */
+#define B_NOCACHE 0x00001000 /* Do not cache block after use. */
+#define B_PHYS 0x00002000 /* I/O to user memory. */
+#define B_RAW 0x00004000 /* Set by physio for raw transfers. */
+#define B_READ 0x00008000 /* Read buffer. */
+#define B_WANTED 0x00010000 /* Process wants this buffer. */
+#define B_WRITEINPROG 0x00020000 /* Write in progress. */
+#define B_XXX 0x00040000 /* Debugging flag. */
+#define B_DEFERRED 0x00080000 /* Skipped over for cleaning */
+#define B_SCANNED 0x00100000 /* Block already pushed during sync */
+#define B_PDAEMON 0x00200000 /* I/O started by pagedaemon */
+#define B_RELEASED 0x00400000 /* free this buffer after its kvm */
+#define B_WARM 0x00800000 /* buffer is or has been on the warm queue */
+#define B_COLD 0x01000000 /* buffer is on the cold queue */
+#define B_BC 0x02000000 /* buffer is managed by the cache */
+#define B_DMA 0x04000000 /* buffer is DMA reachable */
+
+#define B_BITS "\20\001AGE\002NEEDCOMMIT\003ASYNC\004BAD\005BUSY" \
+ "\006CACHE\007CALL\010DELWRI\011DONE\012EINTR\013ERROR" \
+ "\014INVAL\015NOCACHE\016PHYS\017RAW\020READ" \
+ "\021WANTED\022WRITEINPROG\023XXX(FORMAT)\024DEFERRED" \
+ "\025SCANNED\026DAEMON\027RELEASED\030WARM\031COLD\032BC\033DMA"
+
+/*
+ * Zero out the buffer's data area.
+ */
+#define clrbuf(bp) { \
+ bzero((bp)->b_data, (bp)->b_bcount); \
+ (bp)->b_resid = 0; \
+}
+
+
+/* Flags to low-level allocation routines. */
+#define B_CLRBUF 0x01 /* Request allocated buffer be cleared. */
+#define B_SYNC 0x02 /* Do all allocations synchronously. */
+
+struct cluster_info {
+ daddr_t ci_lastr; /* last read (read-ahead) */
+ daddr_t ci_lastw; /* last write (write cluster) */
+ daddr_t ci_cstart; /* start block of cluster */
+ daddr_t ci_lasta; /* last allocation */
+ int ci_clen; /* length of current cluster */
+ int ci_ralen; /* Read-ahead length */
+ daddr_t ci_maxra; /* last readahead block */
+};
+
+#ifdef _KERNEL
+__BEGIN_DECLS
+/* Kva slots (of size MAXPHYS) reserved for syncer and cleaner. */
+#define RESERVE_SLOTS 4
+/* Buffer cache pages reserved for syncer and cleaner. */
+#define RESERVE_PAGES (RESERVE_SLOTS * MAXPHYS / PAGE_SIZE)
+/* Minimum size of the buffer cache, in pages. */
+#define BCACHE_MIN (RESERVE_PAGES * 2)
+#define UNCLEAN_PAGES (bcstats.numbufpages - bcstats.numcleanpages)
+
+extern struct proc *cleanerproc;
+extern long bufpages; /* Max number of pages for buffers' data */
+extern struct pool bufpool;
+extern struct bufhead bufhead;
+
+void bawrite(struct buf *);
+void bdwrite(struct buf *);
+void biodone(struct buf *);
+int biowait(struct buf *);
+int bread(struct vnode *, daddr_t, int, struct buf **);
+int breadn(struct vnode *, daddr_t, int, daddr_t *, int *, int,
+ struct buf **);
+void brelse(struct buf *);
+#define bremfree bufcache_take
+void bufinit(void);
+void buf_dirty(struct buf *);
+void buf_undirty(struct buf *);
+void buf_adjcnt(struct buf *, long);
+int bwrite(struct buf *);
+struct buf *getblk(struct vnode *, daddr_t, int, int, uint64_t);
+struct buf *geteblk(size_t);
+struct buf *incore(struct vnode *, daddr_t);
+
+/*
+ * bufcache functions
+ */
+void bufcache_take(struct buf *);
+void bufcache_release(struct buf *);
+
+int buf_flip_high(struct buf *);
+void buf_flip_dma(struct buf *);
+struct buf *bufcache_getcleanbuf(int, int);
+struct buf *bufcache_getanycleanbuf(void);
+struct buf *bufcache_getdirtybuf(void);
+
+/*
+ * buf_kvm_init initializes the kvm handling for buffers.
+ * buf_acquire sets the B_BUSY flag and ensures that the buffer is
+ * mapped in the kvm.
+ * buf_release clears the B_BUSY flag and allows the buffer to become
+ * unmapped.
+ * buf_unmap is for internal use only. Unmaps the buffer from kvm.
+ */
+void buf_mem_init(vsize_t);
+void buf_acquire(struct buf *);
+void buf_acquire_unmapped(struct buf *);
+void buf_acquire_nomap(struct buf *);
+void buf_map(struct buf *);
+void buf_release(struct buf *);
+int buf_dealloc_mem(struct buf *);
+void buf_fix_mapping(struct buf *, vsize_t);
+void buf_alloc_pages(struct buf *, vsize_t);
+void buf_free_pages(struct buf *);
+
+void minphys(struct buf *bp);
+int physio(void (*strategy)(struct buf *), dev_t dev, int flags,
+ void (*minphys)(struct buf *), struct uio *uio);
+void brelvp(struct buf *);
+void reassignbuf(struct buf *);
+void bgetvp(struct vnode *, struct buf *);
+
+void buf_replacevnode(struct buf *, struct vnode *);
+void buf_daemon(void *);
+void buf_replacevnode(struct buf *, struct vnode *);
+int bread_cluster(struct vnode *, daddr_t, int, struct buf **);
+
+#ifdef DEBUG
+void buf_print(struct buf *);
+#endif
+
+static __inline void
+buf_start(struct buf *bp)
+{
+ if (bioops.io_start)
+ (*bioops.io_start)(bp);
+}
+
+static __inline void
+buf_complete(struct buf *bp)
+{
+ if (bioops.io_complete)
+ (*bioops.io_complete)(bp);
+}
+
+static __inline void
+buf_deallocate(struct buf *bp)
+{
+ if (bioops.io_deallocate)
+ (*bioops.io_deallocate)(bp);
+}
+
+static __inline void
+buf_movedeps(struct buf *bp, struct buf *bp2)
+{
+ if (bioops.io_movedeps)
+ (*bioops.io_movedeps)(bp, bp2);
+}
+
+static __inline int
+buf_countdeps(struct buf *bp, int i, int islocked)
+{
+ if (bioops.io_countdeps)
+ return ((*bioops.io_countdeps)(bp, i, islocked));
+ else
+ return (0);
+}
+
+void cluster_write(struct buf *, struct cluster_info *, u_quad_t);
+
+__END_DECLS
+#endif /* _KERNEL */
+#endif /* !_SYS_BUF_H_ */