2 * Copyright (c) 2008 The DragonFly Project. All rights reserved.
4 * This code is derived from software contributed to The DragonFly Project
5 * by Matthew Dillon <dillon@backplane.com>
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in
15 * the documentation and/or other materials provided with the
17 * 3. Neither the name of The DragonFly Project nor the names of its
18 * contributors may be used to endorse or promote products derived
19 * from this software without specific, prior written permission.
21 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
22 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
23 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
24 * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
25 * COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
26 * INCIDENTAL, SPECIAL, EXEMPLARY OR CONSEQUENTIAL DAMAGES (INCLUDING,
27 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
28 * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
29 * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
30 * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
31 * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
34 * $DragonFly: src/sys/vfs/hammer/hammer_mirror.c,v 1.17 2008/07/31 22:30:33 dillon Exp $
37 * HAMMER mirroring ioctls - serialize and deserialize modifications made
43 static int hammer_mirror_check(hammer_cursor_t cursor,
44 struct hammer_ioc_mrecord_rec *mrec);
45 static int hammer_mirror_update(hammer_cursor_t cursor,
46 struct hammer_ioc_mrecord_rec *mrec);
47 static int hammer_ioc_mirror_write_rec(hammer_cursor_t cursor,
48 struct hammer_ioc_mrecord_rec *mrec,
49 struct hammer_ioc_mirror_rw *mirror,
50 u_int32_t localization,
52 static int hammer_ioc_mirror_write_pass(hammer_cursor_t cursor,
53 struct hammer_ioc_mrecord_rec *mrec,
54 struct hammer_ioc_mirror_rw *mirror,
55 u_int32_t localization);
56 static int hammer_ioc_mirror_write_skip(hammer_cursor_t cursor,
57 struct hammer_ioc_mrecord_skip *mrec,
58 struct hammer_ioc_mirror_rw *mirror,
59 u_int32_t localization);
60 static int hammer_mirror_delete_to(hammer_cursor_t cursor,
61 struct hammer_ioc_mirror_rw *mirror);
62 static int hammer_mirror_nomirror(struct hammer_base_elm *base);
65 * All B-Tree records within the specified key range which also conform
66 * to the transaction id range are returned. Mirroring code keeps track
67 * of the last transaction id fully scanned and can efficiently pick up
68 * where it left off if interrupted.
70 * The PFS is identified in the mirror structure. The passed ip is just
71 * some directory in the overall HAMMER filesystem and has nothing to
75 hammer_ioc_mirror_read(hammer_transaction_t trans, hammer_inode_t ip,
76 struct hammer_ioc_mirror_rw *mirror)
78 struct hammer_cmirror cmirror;
79 struct hammer_cursor cursor;
80 union hammer_ioc_mrecord_any mrec;
81 hammer_btree_leaf_elm_t elm;
82 const int crc_start = HAMMER_MREC_CRCOFF;
89 u_int32_t localization;
92 localization = (u_int32_t)mirror->pfs_id << 16;
94 if ((mirror->key_beg.localization | mirror->key_end.localization) &
95 HAMMER_LOCALIZE_PSEUDOFS_MASK) {
98 if (hammer_btree_cmp(&mirror->key_beg, &mirror->key_end) > 0)
101 mirror->key_cur = mirror->key_beg;
102 mirror->key_cur.localization &= HAMMER_LOCALIZE_MASK;
103 mirror->key_cur.localization += localization;
104 bzero(&mrec, sizeof(mrec));
105 bzero(&cmirror, sizeof(cmirror));
108 * Make CRC errors non-fatal (at least on data), causing an EDOM
109 * error instead of EIO.
111 trans->flags |= HAMMER_TRANSF_CRCDOM;
114 error = hammer_init_cursor(trans, &cursor, NULL, NULL);
116 hammer_done_cursor(&cursor);
119 cursor.key_beg = mirror->key_cur;
120 cursor.key_end = mirror->key_end;
121 cursor.key_end.localization &= HAMMER_LOCALIZE_MASK;
122 cursor.key_end.localization += localization;
124 cursor.flags |= HAMMER_CURSOR_END_INCLUSIVE;
125 cursor.flags |= HAMMER_CURSOR_BACKEND;
128 * This flag filters the search to only return elements whos create
129 * or delete TID is >= mirror_tid. The B-Tree uses the mirror_tid
130 * field stored with internal and leaf nodes to shortcut the scan.
132 cursor.flags |= HAMMER_CURSOR_MIRROR_FILTERED;
133 cursor.cmirror = &cmirror;
134 cmirror.mirror_tid = mirror->tid_beg;
136 error = hammer_btree_first(&cursor);
139 * Yield to more important tasks
142 error = hammer_signal_check(trans->hmp);
148 * An internal node can be returned in mirror-filtered
149 * mode and indicates that the scan is returning a skip
150 * range in the cursor->cmirror structure.
152 uptr = (char *)mirror->ubuf + mirror->count;
153 if (cursor.node->ondisk->type == HAMMER_BTREE_TYPE_INTERNAL) {
157 mirror->key_cur = cmirror.skip_beg;
158 bytes = sizeof(mrec.skip);
159 if (mirror->count + HAMMER_HEAD_DOALIGN(bytes) >
167 mrec.head.signature = HAMMER_IOC_MIRROR_SIGNATURE;
168 mrec.head.type = HAMMER_MREC_TYPE_SKIP;
169 mrec.head.rec_size = bytes;
170 mrec.skip.skip_beg = cmirror.skip_beg;
171 mrec.skip.skip_end = cmirror.skip_end;
172 mrec.head.rec_crc = crc32(&mrec.head.rec_size,
174 error = copyout(&mrec, uptr, bytes);
180 * Leaf node. In full-history mode we could filter out
181 * elements modified outside the user-requested TID range.
183 * However, such elements must be returned so the writer
184 * can compare them against the target to determine what
185 * needs to be deleted on the target, particular for
186 * no-history mirrors.
188 KKASSERT(cursor.node->ondisk->type == HAMMER_BTREE_TYPE_LEAF);
189 elm = &cursor.node->ondisk->elms[cursor.index].leaf;
190 mirror->key_cur = elm->base;
193 * If the record was created after our end point we just
196 if (elm->base.create_tid > mirror->tid_end) {
204 * Determine if we should generate a PASS or a REC. PASS
205 * records are records without any data payload. Such
206 * records will be generated if the target is already expected
207 * to have the record, allowing it to delete the gaps.
209 * A PASS record is also used to perform deletions on the
212 * Such deletions are needed if the master or files on the
213 * master are no-history, or if the slave is so far behind
214 * the master has already been pruned.
216 if (elm->base.create_tid < mirror->tid_beg) {
217 bytes = sizeof(mrec.rec);
218 if (mirror->count + HAMMER_HEAD_DOALIGN(bytes) >
226 mrec.head.signature = HAMMER_IOC_MIRROR_SIGNATURE;
227 mrec.head.type = HAMMER_MREC_TYPE_PASS;
228 mrec.head.rec_size = bytes;
229 mrec.rec.leaf = *elm;
230 mrec.head.rec_crc = crc32(&mrec.head.rec_size,
232 error = copyout(&mrec, uptr, bytes);
239 * The core code exports the data to userland.
241 * CRC errors on data are reported but passed through,
242 * but the data must be washed by the user program.
244 * If userland just wants the btree records it can
245 * request that bulk data not be returned. This is
246 * use during mirror-stream histogram generation.
249 data_len = (elm->data_offset) ? elm->data_len : 0;
251 (mirror->head.flags & HAMMER_IOC_MIRROR_NODATA)) {
253 mrec_flags |= HAMMER_MRECF_NODATA;
256 error = hammer_btree_extract(&cursor,
257 HAMMER_CURSOR_GET_DATA);
261 mrec_flags |= HAMMER_MRECF_CRC_ERROR |
262 HAMMER_MRECF_DATA_CRC_BAD;
266 bytes = sizeof(mrec.rec) + data_len;
267 if (mirror->count + HAMMER_HEAD_DOALIGN(bytes) > mirror->size)
271 * Construct the record for userland and copyout.
273 * The user is asking for a snapshot, if the record was
274 * deleted beyond the user-requested ending tid, the record
275 * is not considered deleted from the point of view of
276 * userland and delete_tid is cleared.
278 mrec.head.signature = HAMMER_IOC_MIRROR_SIGNATURE;
279 mrec.head.type = HAMMER_MREC_TYPE_REC | mrec_flags;
280 mrec.head.rec_size = bytes;
281 mrec.rec.leaf = *elm;
283 if (elm->base.delete_tid > mirror->tid_end)
284 mrec.rec.leaf.base.delete_tid = 0;
285 rec_crc = crc32(&mrec.head.rec_size,
286 sizeof(mrec.rec) - crc_start);
288 rec_crc = crc32_ext(cursor.data, data_len, rec_crc);
289 mrec.head.rec_crc = rec_crc;
290 error = copyout(&mrec, uptr, sizeof(mrec.rec));
291 if (data_len && error == 0) {
292 error = copyout(cursor.data, uptr + sizeof(mrec.rec),
298 * eatdisk controls whether we skip the current cursor
299 * position on the next scan or not. If doing a SKIP
300 * the cursor is already positioned properly for the next
301 * scan and eatdisk will be 0.
305 mirror->count += HAMMER_HEAD_DOALIGN(bytes);
307 cursor.flags |= HAMMER_CURSOR_ATEDISK;
309 cursor.flags &= ~HAMMER_CURSOR_ATEDISK;
310 error = hammer_btree_iterate(&cursor);
313 if (error == ENOENT) {
314 mirror->key_cur = mirror->key_end;
317 hammer_done_cursor(&cursor);
318 if (error == EDEADLK)
320 if (error == EINTR) {
321 mirror->head.flags |= HAMMER_IOC_HEAD_INTR;
325 mirror->key_cur.localization &= HAMMER_LOCALIZE_MASK;
330 * Copy records from userland to the target mirror.
332 * The PFS is identified in the mirror structure. The passed ip is just
333 * some directory in the overall HAMMER filesystem and has nothing to
334 * do with the PFS. In fact, there might not even be a root directory for
338 hammer_ioc_mirror_write(hammer_transaction_t trans, hammer_inode_t ip,
339 struct hammer_ioc_mirror_rw *mirror)
341 union hammer_ioc_mrecord_any mrec;
342 struct hammer_cursor cursor;
343 u_int32_t localization;
344 int checkspace_count = 0;
350 localization = (u_int32_t)mirror->pfs_id << 16;
351 seq = trans->hmp->flusher.done;
354 * Validate the mirror structure and relocalize the tracking keys.
356 if (mirror->size < 0 || mirror->size > 0x70000000)
358 mirror->key_beg.localization &= HAMMER_LOCALIZE_MASK;
359 mirror->key_beg.localization += localization;
360 mirror->key_end.localization &= HAMMER_LOCALIZE_MASK;
361 mirror->key_end.localization += localization;
362 mirror->key_cur.localization &= HAMMER_LOCALIZE_MASK;
363 mirror->key_cur.localization += localization;
366 * Set up our tracking cursor for the loop. The tracking cursor
367 * is used to delete records that are no longer present on the
368 * master. The last handled record at key_cur must be skipped.
370 error = hammer_init_cursor(trans, &cursor, NULL, NULL);
372 cursor.key_beg = mirror->key_cur;
373 cursor.key_end = mirror->key_end;
374 cursor.flags |= HAMMER_CURSOR_BACKEND;
375 error = hammer_btree_first(&cursor);
377 cursor.flags |= HAMMER_CURSOR_ATEDISK;
382 * Loop until our input buffer has been exhausted.
385 mirror->count + sizeof(mrec.head) <= mirror->size) {
388 * Don't blow out the buffer cache. Leave room for frontend
391 * WARNING: See warnings in hammer_unlock_cursor() function.
393 while (hammer_flusher_meta_halflimit(trans->hmp) ||
394 hammer_flusher_undo_exhausted(trans, 2)) {
395 hammer_unlock_cursor(&cursor);
396 hammer_flusher_wait(trans->hmp, seq);
397 hammer_lock_cursor(&cursor);
398 seq = hammer_flusher_async_one(trans->hmp);
402 * If there is insufficient free space it may be due to
403 * reserved bigblocks, which flushing might fix.
405 if (hammer_checkspace(trans->hmp, HAMMER_CHKSPC_MIRROR)) {
406 if (++checkspace_count == 10) {
410 hammer_unlock_cursor(&cursor);
411 hammer_flusher_wait(trans->hmp, seq);
412 hammer_lock_cursor(&cursor);
413 seq = hammer_flusher_async(trans->hmp, NULL);
418 * Acquire and validate header
420 if ((bytes = mirror->size - mirror->count) > sizeof(mrec))
421 bytes = sizeof(mrec);
422 uptr = (char *)mirror->ubuf + mirror->count;
423 error = copyin(uptr, &mrec, bytes);
426 if (mrec.head.signature != HAMMER_IOC_MIRROR_SIGNATURE) {
430 if (mrec.head.rec_size < sizeof(mrec.head) ||
431 mrec.head.rec_size > sizeof(mrec) + HAMMER_XBUFSIZE ||
432 mirror->count + mrec.head.rec_size > mirror->size) {
437 switch(mrec.head.type & HAMMER_MRECF_TYPE_MASK) {
438 case HAMMER_MREC_TYPE_SKIP:
439 if (mrec.head.rec_size != sizeof(mrec.skip))
442 error = hammer_ioc_mirror_write_skip(&cursor, &mrec.skip, mirror, localization);
444 case HAMMER_MREC_TYPE_REC:
445 if (mrec.head.rec_size < sizeof(mrec.rec))
448 error = hammer_ioc_mirror_write_rec(&cursor, &mrec.rec, mirror, localization, uptr + sizeof(mrec.rec));
450 case HAMMER_MREC_TYPE_REC_NODATA:
451 case HAMMER_MREC_TYPE_REC_BADCRC:
453 * Records with bad data payloads are ignored XXX.
454 * Records with no data payload have to be skipped
455 * (they shouldn't have been written in the first
458 if (mrec.head.rec_size < sizeof(mrec.rec))
461 case HAMMER_MREC_TYPE_PASS:
462 if (mrec.head.rec_size != sizeof(mrec.rec))
465 error = hammer_ioc_mirror_write_pass(&cursor, &mrec.rec, mirror, localization);
473 * Retry the current record on deadlock, otherwise setup
476 if (error == EDEADLK) {
477 while (error == EDEADLK) {
478 hammer_sync_lock_sh(trans);
479 hammer_recover_cursor(&cursor);
480 error = hammer_cursor_upgrade(&cursor);
481 hammer_sync_unlock(trans);
484 if (error == EALREADY)
488 HAMMER_HEAD_DOALIGN(mrec.head.rec_size);
492 hammer_done_cursor(&cursor);
498 mirror->head.flags |= HAMMER_IOC_HEAD_ERROR;
499 mirror->head.error = error;
503 * ioctls don't update the RW data structure if an error is returned,
510 * Handle skip records.
512 * We must iterate from the last resolved record position at mirror->key_cur
513 * to skip_beg non-inclusive and delete any records encountered.
515 * mirror->key_cur must be carefully set when we succeed in processing
519 hammer_ioc_mirror_write_skip(hammer_cursor_t cursor,
520 struct hammer_ioc_mrecord_skip *mrec,
521 struct hammer_ioc_mirror_rw *mirror,
522 u_int32_t localization)
527 * Relocalize the skip range
529 mrec->skip_beg.localization &= HAMMER_LOCALIZE_MASK;
530 mrec->skip_beg.localization += localization;
531 mrec->skip_end.localization &= HAMMER_LOCALIZE_MASK;
532 mrec->skip_end.localization += localization;
535 * Iterate from current position to skip_beg, deleting any records
536 * we encounter. The record at skip_beg is not included (it is
539 cursor->key_end = mrec->skip_beg;
540 cursor->flags &= ~HAMMER_CURSOR_END_INCLUSIVE;
541 cursor->flags |= HAMMER_CURSOR_BACKEND;
542 error = hammer_mirror_delete_to(cursor, mirror);
545 * Now skip past the skip (which is the whole point point of
546 * having a skip record). The sender has not sent us any records
547 * for the skip area so we wouldn't know what to keep and what
550 * Clear ATEDISK because skip_end is non-inclusive, so we can't
551 * count an exact match if we happened to get one.
554 mirror->key_cur = mrec->skip_end;
555 cursor->key_beg = mrec->skip_end;
556 error = hammer_btree_lookup(cursor);
557 cursor->flags &= ~HAMMER_CURSOR_ATEDISK;
565 * Handle B-Tree records.
567 * We must iterate to mrec->base.key (non-inclusively), and then process
568 * the record. We are allowed to write a new record or delete an existing
569 * record, but cannot replace an existing record.
571 * mirror->key_cur must be carefully set when we succeed in processing
575 hammer_ioc_mirror_write_rec(hammer_cursor_t cursor,
576 struct hammer_ioc_mrecord_rec *mrec,
577 struct hammer_ioc_mirror_rw *mirror,
578 u_int32_t localization,
581 hammer_transaction_t trans;
585 trans = cursor->trans;
586 rec_crc = crc32(mrec, sizeof(*mrec));
588 if (mrec->leaf.data_len < 0 ||
589 mrec->leaf.data_len > HAMMER_XBUFSIZE ||
590 mrec->leaf.data_len + sizeof(*mrec) > mrec->head.rec_size) {
595 * Re-localize for target. relocalization of data is handled
596 * by hammer_mirror_write().
598 mrec->leaf.base.localization &= HAMMER_LOCALIZE_MASK;
599 mrec->leaf.base.localization += localization;
602 * Delete records through until we reach (non-inclusively) the
605 cursor->key_end = mrec->leaf.base;
606 cursor->flags &= ~HAMMER_CURSOR_END_INCLUSIVE;
607 cursor->flags |= HAMMER_CURSOR_BACKEND;
608 error = hammer_mirror_delete_to(cursor, mirror);
611 * Certain records are not part of the mirroring operation
613 if (error == 0 && hammer_mirror_nomirror(&mrec->leaf.base))
619 * If the record exists only the delete_tid may be updated.
621 * If the record does not exist we can create it only if the
622 * create_tid is not too old. If the create_tid is too old
623 * it may have already been destroyed on the slave from pruning.
625 * Note that mirror operations are effectively as-of operations
626 * and delete_tid can be 0 for mirroring purposes even if it is
627 * not actually 0 at the originator.
629 * These functions can return EDEADLK
632 cursor->key_beg = mrec->leaf.base;
633 cursor->flags |= HAMMER_CURSOR_BACKEND;
634 cursor->flags &= ~HAMMER_CURSOR_INSERT;
635 error = hammer_btree_lookup(cursor);
638 if (error == 0 && hammer_mirror_check(cursor, mrec)) {
639 error = hammer_mirror_update(cursor, mrec);
640 } else if (error == ENOENT) {
641 if (mrec->leaf.base.create_tid >= mirror->tid_beg) {
642 error = hammer_create_at_cursor(
644 uptr, HAMMER_CREATE_MODE_UMIRROR);
649 if (error == 0 || error == EALREADY)
650 mirror->key_cur = mrec->leaf.base;
655 * This works like write_rec but no write or update is necessary,
656 * and no data payload is included so we couldn't do a write even
659 * We must still iterate for deletions, and we can validate the
660 * record header which is a good way to test for corrupted mirror
663 * mirror->key_cur must be carefully set when we succeed in processing
668 hammer_ioc_mirror_write_pass(hammer_cursor_t cursor,
669 struct hammer_ioc_mrecord_rec *mrec,
670 struct hammer_ioc_mirror_rw *mirror,
671 u_int32_t localization)
673 hammer_transaction_t trans;
677 trans = cursor->trans;
678 rec_crc = crc32(mrec, sizeof(*mrec));
681 * Re-localize for target. Relocalization of data is handled
682 * by hammer_mirror_write().
684 mrec->leaf.base.localization &= HAMMER_LOCALIZE_MASK;
685 mrec->leaf.base.localization += localization;
688 * Delete records through until we reach (non-inclusively) the
691 cursor->key_end = mrec->leaf.base;
692 cursor->flags &= ~HAMMER_CURSOR_END_INCLUSIVE;
693 cursor->flags |= HAMMER_CURSOR_BACKEND;
694 error = hammer_mirror_delete_to(cursor, mirror);
697 * Certain records are not part of the mirroring operation
699 if (hammer_mirror_nomirror(&mrec->leaf.base))
703 * Locate the record and get past it by setting ATEDISK. Perform
704 * any necessary deletions. We have no data payload and cannot
705 * create a new record.
708 mirror->key_cur = mrec->leaf.base;
709 cursor->key_beg = mrec->leaf.base;
710 cursor->flags |= HAMMER_CURSOR_BACKEND;
711 cursor->flags &= ~HAMMER_CURSOR_INSERT;
712 error = hammer_btree_lookup(cursor);
714 if (hammer_mirror_check(cursor, mrec))
715 error = hammer_mirror_update(cursor, mrec);
716 cursor->flags |= HAMMER_CURSOR_ATEDISK;
718 cursor->flags &= ~HAMMER_CURSOR_ATEDISK;
727 * As part of the mirror write we iterate across swaths of records
728 * on the target which no longer exist on the source, and mark them
731 * The caller has indexed the cursor and set up key_end. We iterate
732 * through to key_end.
734 * There is an edge case where the master has deleted a record whos
735 * create_tid exactly matches our end_tid. We cannot delete this
736 * record on the slave yet because we cannot assign delete_tid == create_tid.
737 * The deletion should be picked up on the next sequence since in order
738 * to have been deleted on the master a transaction must have occured with
739 * a TID greater then the create_tid of the record.
741 * To support incremental re-mirroring, just for robustness, we do not
742 * touch any records created beyond (or equal to) mirror->tid_end.
746 hammer_mirror_delete_to(hammer_cursor_t cursor,
747 struct hammer_ioc_mirror_rw *mirror)
749 hammer_btree_leaf_elm_t elm;
752 error = hammer_btree_iterate(cursor);
754 elm = &cursor->node->ondisk->elms[cursor->index].leaf;
755 KKASSERT(elm->base.btype == HAMMER_BTREE_TYPE_RECORD);
756 cursor->flags |= HAMMER_CURSOR_ATEDISK;
759 * Certain records are not part of the mirroring operation
761 if (hammer_mirror_nomirror(&elm->base)) {
762 error = hammer_btree_iterate(cursor);
767 * Note: Must still delete records with create_tid < tid_beg,
768 * as record may have been pruned-away on source.
770 if (elm->base.delete_tid == 0 &&
771 elm->base.create_tid < mirror->tid_end) {
772 error = hammer_delete_at_cursor(cursor,
773 HAMMER_DELETE_ADJUST,
779 error = hammer_btree_iterate(cursor);
787 * Check whether an update is needed in the case where a match already
788 * exists on the target. The only type of update allowed in this case
789 * is an update of the delete_tid.
791 * Return non-zero if the update should proceed.
795 hammer_mirror_check(hammer_cursor_t cursor, struct hammer_ioc_mrecord_rec *mrec)
797 hammer_btree_leaf_elm_t leaf = cursor->leaf;
799 if (leaf->base.delete_tid != mrec->leaf.base.delete_tid) {
800 if (mrec->leaf.base.delete_tid != 0)
807 * Filter out records which are never mirrored, such as configuration space
808 * records (for hammer cleanup).
810 * NOTE: We currently allow HAMMER_RECTYPE_SNAPSHOT records to be mirrored.
814 hammer_mirror_nomirror(struct hammer_base_elm *base)
817 * Certain types of records are never updated when mirroring.
818 * Slaves have their own configuration space.
820 if (base->rec_type == HAMMER_RECTYPE_CONFIG)
827 * Update a record in-place. Only the delete_tid can change, and
828 * only from zero to non-zero.
832 hammer_mirror_update(hammer_cursor_t cursor,
833 struct hammer_ioc_mrecord_rec *mrec)
838 * This case shouldn't occur.
840 if (mrec->leaf.base.delete_tid == 0)
844 * Mark the record deleted on the mirror target.
846 error = hammer_delete_at_cursor(cursor, HAMMER_DELETE_ADJUST,
847 mrec->leaf.base.delete_tid,
848 mrec->leaf.delete_ts,
850 cursor->flags |= HAMMER_CURSOR_ATEDISK;
856 * MOVED TO HAMMER_OBJECT.C: hammer_create_at_cursor()
859 static int hammer_mirror_localize_data(hammer_data_ondisk_t data,
860 hammer_btree_leaf_elm_t leaf);
863 * Write out a new record.
867 hammer_mirror_write(hammer_cursor_t cursor,
868 struct hammer_ioc_mrecord_rec *mrec,
871 hammer_transaction_t trans;
872 hammer_buffer_t data_buffer;
873 hammer_off_t ndata_offset;
874 hammer_tid_t high_tid;
879 trans = cursor->trans;
883 * Get the sync lock so the whole mess is atomic
885 hammer_sync_lock_sh(trans);
888 * Allocate and adjust data
890 if (mrec->leaf.data_len && mrec->leaf.data_offset) {
891 ndata = hammer_alloc_data(trans, mrec->leaf.data_len,
892 mrec->leaf.base.rec_type,
893 &ndata_offset, &data_buffer,
897 mrec->leaf.data_offset = ndata_offset;
898 hammer_modify_buffer(trans, data_buffer, NULL, 0);
899 error = copyin(udata, ndata, mrec->leaf.data_len);
901 if (hammer_crc_test_leaf(ndata, &mrec->leaf) == 0) {
902 kprintf("data crc mismatch on pipe\n");
905 error = hammer_mirror_localize_data(
909 hammer_modify_buffer_done(data_buffer);
911 mrec->leaf.data_offset = 0;
919 * Do the insertion. This can fail with a EDEADLK or EALREADY
921 cursor->flags |= HAMMER_CURSOR_INSERT;
922 error = hammer_btree_lookup(cursor);
923 if (error != ENOENT) {
929 error = hammer_btree_insert(cursor, &mrec->leaf, &doprop);
932 * Cursor is left on the current element, we want to skip it now.
934 cursor->flags |= HAMMER_CURSOR_ATEDISK;
935 cursor->flags &= ~HAMMER_CURSOR_INSERT;
938 * Track a count of active inodes.
941 mrec->leaf.base.rec_type == HAMMER_RECTYPE_INODE &&
942 mrec->leaf.base.delete_tid == 0) {
943 hammer_modify_volume_field(trans,
946 ++trans->hmp->rootvol->ondisk->vol0_stat_inodes;
947 hammer_modify_volume_done(trans->rootvol);
951 * vol0_next_tid must track the highest TID stored in the filesystem.
952 * We do not need to generate undo for this update.
954 high_tid = mrec->leaf.base.create_tid;
955 if (high_tid < mrec->leaf.base.delete_tid)
956 high_tid = mrec->leaf.base.delete_tid;
957 if (trans->rootvol->ondisk->vol0_next_tid < high_tid) {
958 hammer_modify_volume(trans, trans->rootvol, NULL, 0);
959 trans->rootvol->ondisk->vol0_next_tid = high_tid;
960 hammer_modify_volume_done(trans->rootvol);
964 * WARNING! cursor's leaf pointer may have changed after
965 * do_propagation returns.
967 if (error == 0 && doprop)
968 hammer_btree_do_propagation(cursor, NULL, &mrec->leaf);
974 if (error && mrec->leaf.data_offset) {
975 hammer_blockmap_free(cursor->trans,
976 mrec->leaf.data_offset,
977 mrec->leaf.data_len);
979 hammer_sync_unlock(trans);
981 hammer_rel_buffer(data_buffer, 0);
986 * Localize the data payload. Directory entries may need their
987 * localization adjusted.
991 hammer_mirror_localize_data(hammer_data_ondisk_t data,
992 hammer_btree_leaf_elm_t leaf)
994 u_int32_t localization;
996 if (leaf->base.rec_type == HAMMER_RECTYPE_DIRENTRY) {
997 localization = leaf->base.localization &
998 HAMMER_LOCALIZE_PSEUDOFS_MASK;
999 if (data->entry.localization != localization) {
1000 data->entry.localization = localization;
1001 hammer_crc_set_leaf(data, leaf);