/linux-4.1.27/fs/xfs/ |
H A D | xfs_trans_ail.c | 97 * We need the AIL lock in order to get a coherent read of the lsn of the last 104 xfs_lsn_t lsn = 0; xfs_ail_min_lsn() local 110 lsn = lip->li_lsn; xfs_ail_min_lsn() 113 return lsn; xfs_ail_min_lsn() 117 * Return the maximum lsn held in the AIL, or zero if the AIL is empty. 123 xfs_lsn_t lsn = 0; xfs_ail_max_lsn() local 129 lsn = lip->li_lsn; xfs_ail_max_lsn() 132 return lsn; xfs_ail_max_lsn() 204 * Find the first item in the AIL with the given @lsn by searching in ascending 206 * ascending traversal. Pass a @lsn of zero to initialise the cursor to the 213 xfs_lsn_t lsn) xfs_trans_ail_cursor_first() 219 if (lsn == 0) { xfs_trans_ail_cursor_first() 225 if (XFS_LSN_CMP(lip->li_lsn, lsn) >= 0) xfs_trans_ail_cursor_first() 239 xfs_lsn_t lsn) __xfs_trans_ail_cursor_last() 244 if (XFS_LSN_CMP(lip->li_lsn, lsn) <= 0) __xfs_trans_ail_cursor_last() 251 * Find the last item in the AIL with the given @lsn by searching in descending 253 * item with the value of @lsn, then it sets the cursor to the last item with an 254 * LSN lower than @lsn. Returns NULL if the list is empty. 260 xfs_lsn_t lsn) xfs_trans_ail_cursor_last() 263 cur->item = __xfs_trans_ail_cursor_last(ailp, lsn); xfs_trans_ail_cursor_last() 278 xfs_lsn_t lsn) xfs_ail_splice() 291 lip = __xfs_trans_ail_cursor_last(ailp, lsn); xfs_ail_splice() 295 * in lsn order, and future items to be spliced in will xfs_ail_splice() 335 xfs_lsn_t lsn; xfsaild_push() local 376 lsn = lip->li_lsn; xfsaild_push() 391 ailp->xa_last_pushed_lsn = lsn; xfsaild_push() 410 ailp->xa_last_pushed_lsn = lsn; xfsaild_push() 452 lsn = lip->li_lsn; xfsaild_push() 460 if (!count || XFS_LSN_CMP(lsn, target) >= 0) { xfsaild_push() 643 xfs_lsn_t lsn) __releases(ailp->xa_lock) 657 if (XFS_LSN_CMP(lsn, lip->li_lsn) <= 0) 660 trace_xfs_ail_move(lip, lip->li_lsn, lsn); 666 trace_xfs_ail_insert(lip, 0, lsn); 668 lip->li_lsn = lsn; 673 xfs_ail_splice(ailp, cur, &tmp, lsn); 695 * flag from the item and reset the item's lsn to 0. If we remove the first 210 xfs_trans_ail_cursor_first( struct xfs_ail *ailp, struct xfs_ail_cursor *cur, xfs_lsn_t lsn) xfs_trans_ail_cursor_first() argument 237 __xfs_trans_ail_cursor_last( struct xfs_ail *ailp, xfs_lsn_t lsn) __xfs_trans_ail_cursor_last() argument 257 xfs_trans_ail_cursor_last( struct xfs_ail *ailp, struct xfs_ail_cursor *cur, xfs_lsn_t lsn) xfs_trans_ail_cursor_last() argument 274 xfs_ail_splice( struct xfs_ail *ailp, struct xfs_ail_cursor *cur, struct list_head *list, xfs_lsn_t lsn) xfs_ail_splice() argument
|
H A D | xfs_inode_item.h | 31 xfs_lsn_t ili_flush_lsn; /* lsn at last flush */ 32 xfs_lsn_t ili_last_lsn; /* lsn at last transaction */
|
H A D | xfs_trans_priv.h | 87 xfs_lsn_t lsn) __releases(ailp->xa_lock); 104 xfs_lsn_t lsn) __releases(ailp->xa_lock) 106 xfs_trans_ail_update_bulk(ailp, NULL, &lip, 1, lsn); 130 xfs_lsn_t lsn); 133 xfs_lsn_t lsn);
|
H A D | xfs_dquot_item.c | 112 xfs_lsn_t lsn) xfs_qm_dquot_logitem_committed() 118 return lsn; xfs_qm_dquot_logitem_committed() 223 * this needs to stamp an lsn into the dquot, I think. 230 xfs_lsn_t lsn) xfs_qm_dquot_logitem_committing() 347 * so simply return the lsn at which it's been logged. 352 xfs_lsn_t lsn) xfs_qm_qoff_logitem_committed() 354 return lsn; xfs_qm_qoff_logitem_committed() 360 xfs_lsn_t lsn) xfs_qm_qoffend_logitem_committed() 388 * one that does nothing and one that stamps in the lsn that matters 110 xfs_qm_dquot_logitem_committed( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_qm_dquot_logitem_committed() argument 228 xfs_qm_dquot_logitem_committing( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_qm_dquot_logitem_committing() argument 350 xfs_qm_qoff_logitem_committed( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_qm_qoff_logitem_committed() argument 358 xfs_qm_qoffend_logitem_committed( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_qm_qoffend_logitem_committed() argument
|
H A D | xfs_dquot_item.h | 29 xfs_lsn_t qli_flush_lsn; /* lsn at last flush */
|
H A D | xfs_export.c | 231 xfs_lsn_t lsn = 0; xfs_fs_nfs_commit_metadata() local 235 lsn = ip->i_itemp->ili_last_lsn; xfs_fs_nfs_commit_metadata() 238 if (!lsn) xfs_fs_nfs_commit_metadata() 240 return _xfs_log_force_lsn(mp, lsn, XFS_LOG_SYNC, NULL); xfs_fs_nfs_commit_metadata()
|
H A D | xfs_extfree_item.c | 180 * the lsn at which it's been logged. 185 xfs_lsn_t lsn) xfs_efi_item_committed() 187 return lsn; xfs_efi_item_committed() 200 xfs_lsn_t lsn) xfs_efi_item_committing() 436 xfs_lsn_t lsn) xfs_efd_item_committed() 461 xfs_lsn_t lsn) xfs_efd_item_committing() 183 xfs_efi_item_committed( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_efi_item_committed() argument 198 xfs_efi_item_committing( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_efi_item_committing() argument 434 xfs_efd_item_committed( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_efd_item_committed() argument 459 xfs_efd_item_committing( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_efd_item_committing() argument
|
H A D | xfs_icreate_item.c | 107 xfs_lsn_t lsn) xfs_icreate_item_committed() 129 xfs_lsn_t lsn) xfs_icreate_item_committing() 105 xfs_icreate_item_committed( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_icreate_item_committed() argument 127 xfs_icreate_item_committing( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_icreate_item_committing() argument
|
H A D | xfs_inode_item.c | 494 * at the given lsn. Since we always re-log all dirty data in an inode, the 496 * simply return the given lsn. 516 xfs_lsn_t lsn) xfs_inode_item_committed() 525 return lsn; xfs_inode_item_committed() 535 xfs_lsn_t lsn) xfs_inode_item_committing() 537 INODE_ITEM(lip)->ili_last_lsn = lsn; xfs_inode_item_committing() 653 * if the ili_logged flag is set and the inode's lsn has not xfs_iflush_done() 654 * changed. First we check the lsn outside xfs_iflush_done() 514 xfs_inode_item_committed( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_inode_item_committed() argument 533 xfs_inode_item_committing( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_inode_item_committing() argument
|
H A D | xfs_trace.h | 989 __field(xfs_lsn_t, lsn) 996 __entry->lsn = lip->li_lsn; 998 TP_printk("dev %d:%d lip 0x%p lsn %d/%d type %s flags %s", 1001 CYCLE_LSN(__entry->lsn), BLOCK_LSN(__entry->lsn), 1007 TP_PROTO(struct xfs_mount *mp, xfs_lsn_t lsn), 1008 TP_ARGS(mp, lsn), 1011 __field(xfs_lsn_t, lsn) 1015 __entry->lsn = lsn; 1017 TP_printk("dev %d:%d lsn 0x%llx", 1019 __entry->lsn) 1050 TP_printk("dev %d:%d lip 0x%p old lsn %d/%d new lsn %d/%d type %s flags %s", 1082 TP_printk("dev %d:%d new tail lsn %d/%d, old lsn %d/%d, last sync %d/%d", 1425 __field(xfs_lsn_t, lsn) 1430 __entry->lsn = trans->t_commit_lsn; 1435 __entry->lsn)
|
H A D | xfs_log_recover.c | 875 * a sync lsn embedded in it. LSNs hold block numbers, so it is easy 880 * lsn. The entire log record does not need to be valid. We only care 997 * Save the current tail lsn to use to pass to xlog_find_tail() 999 * unmount record if there is one, so we pass the lsn of the xlog_find_tail() 1056 * We use the lsn from before modifying it so that we'll never xlog_find_tail() 1315 * The tail lsn is passed in rather than taken from 1839 xfs_lsn_t lsn = -1; xlog_recover_get_buf_lsn() local 1855 lsn = be64_to_cpu(btb->bb_u.s.bb_lsn); xlog_recover_get_buf_lsn() 1863 lsn = be64_to_cpu(btb->bb_u.l.bb_lsn); xlog_recover_get_buf_lsn() 1868 lsn = be64_to_cpu(((struct xfs_agf *)blk)->agf_lsn); xlog_recover_get_buf_lsn() 1872 lsn = be64_to_cpu(((struct xfs_agfl *)blk)->agfl_lsn); xlog_recover_get_buf_lsn() 1876 lsn = be64_to_cpu(((struct xfs_agi *)blk)->agi_lsn); xlog_recover_get_buf_lsn() 1880 lsn = be64_to_cpu(((struct xfs_dsymlink_hdr *)blk)->sl_lsn); xlog_recover_get_buf_lsn() 1886 lsn = be64_to_cpu(((struct xfs_dir3_blk_hdr *)blk)->lsn); xlog_recover_get_buf_lsn() 1899 lsn = be64_to_cpu(((struct xfs_dsb *)blk)->sb_lsn); xlog_recover_get_buf_lsn() 1906 if (lsn != (xfs_lsn_t)-1) { xlog_recover_get_buf_lsn() 1909 return lsn; xlog_recover_get_buf_lsn() 1917 lsn = be64_to_cpu(((struct xfs_da3_blkinfo *)blk)->lsn); xlog_recover_get_buf_lsn() 1924 if (lsn != (xfs_lsn_t)-1) { xlog_recover_get_buf_lsn() 1927 return lsn; xlog_recover_get_buf_lsn() 2323 xfs_lsn_t lsn; xlog_recover_buffer_pass2() local 2353 * the lsn of the transaction we are replaying. xlog_recover_buffer_pass2() 2370 lsn = xlog_recover_get_buf_lsn(mp, bp); xlog_recover_buffer_pass2() 2371 if (lsn && lsn != -1 && XFS_LSN_CMP(lsn, current_lsn) >= 0) { xlog_recover_buffer_pass2() 2582 * than the lsn of the transaction we are replaying. Note: we still xlog_recover_inode_pass2() 2588 xfs_lsn_t lsn = be64_to_cpu(dip->di_lsn); xlog_recover_inode_pass2() local 2590 if (lsn && lsn != -1 && XFS_LSN_CMP(lsn, current_lsn) >= 0) { xlog_recover_inode_pass2() 2897 * than the lsn of the transaction we are replaying. xlog_recover_dquot_pass2() 2901 xfs_lsn_t lsn = be64_to_cpu(dqb->dd_lsn); xlog_recover_dquot_pass2() local 2903 if (lsn && lsn != -1 && XFS_LSN_CMP(lsn, current_lsn) >= 0) { xlog_recover_dquot_pass2() 2935 xfs_lsn_t lsn) xlog_recover_efi_pass2() 2956 xfs_trans_ail_update(log->l_ailp, &efip->efi_item, lsn); xlog_recover_efi_pass2() 4444 * lsn of the last known good LR on disk. If there are extent frees xlog_do_recover() 2932 xlog_recover_efi_pass2( struct xlog *log, struct xlog_recover_item *item, xfs_lsn_t lsn) xlog_recover_efi_pass2() argument
|
H A D | xfs_log.c | 519 xfs_lsn_t lsn = 0; xfs_log_done() local 527 (xlog_commit_record(log, ticket, iclog, &lsn)))) { xfs_log_done() 528 lsn = (xfs_lsn_t) -1; xfs_log_done() 556 return lsn; xfs_log_done() 787 xfs_lsn_t lsn; xfs_log_unmount_write() local 835 error = xlog_write(log, &vec, tic, &lsn, xfs_log_unmount_write() 1501 * ticket. Return the lsn of the commit record. 1532 * log space. This code pushes on the lsn which would supposedly free up 1534 * pushes on an lsn which is further along in the log once we reach the high 1576 * Don't pass in an lsn greater than the lsn of the last xlog_grant_push_ail() 1577 * log record known to be on disk. Use a snapshot of the last sync lsn xlog_grant_push_ail() 2270 * B. If this is first write, save away start lsn 2356 /* start_lsn is the first lsn written to. That's all we need. */ xlog_write() 2565 xfs_lsn_t lowest_lsn, lsn; xlog_get_lowest_lsn() local 2571 lsn = be64_to_cpu(lsn_log->ic_header.h_lsn); xlog_get_lowest_lsn() 2572 if ((lsn && !lowest_lsn) || xlog_get_lowest_lsn() 2573 (XFS_LSN_CMP(lsn, lowest_lsn) < 0)) { xlog_get_lowest_lsn() 2574 lowest_lsn = lsn; xlog_get_lowest_lsn() 2908 /* On the 1st write to an iclog, figure out lsn. This works xlog_state_get_iclog_space() 3193 xfs_lsn_t lsn; _xfs_log_force() local 3237 lsn = be64_to_cpu(iclog->ic_header.h_lsn); _xfs_log_force() 3247 if (be64_to_cpu(iclog->ic_header.h_lsn) == lsn && _xfs_log_force() 3265 * which would give it another lsn. If we have a new lsn, just _xfs_log_force() 3320 * Find in-core log with lsn. 3327 * to force a given lsn to disk will wait on a the sv attached to the 3335 xfs_lsn_t lsn, _xfs_log_force_lsn() 3343 ASSERT(lsn != 0); _xfs_log_force_lsn() 3347 lsn = xlog_cil_force_lsn(log, lsn); _xfs_log_force_lsn() 3348 if (lsn == NULLCOMMITLSN) _xfs_log_force_lsn() 3360 if (be64_to_cpu(iclog->ic_header.h_lsn) != lsn) { _xfs_log_force_lsn() 3455 xfs_lsn_t lsn, xfs_log_force_lsn() 3460 trace_xfs_log_force(mp, lsn); xfs_log_force_lsn() 3461 error = _xfs_log_force_lsn(mp, lsn, flags, NULL); xfs_log_force_lsn() 3333 _xfs_log_force_lsn( struct xfs_mount *mp, xfs_lsn_t lsn, uint flags, int *log_flushed) _xfs_log_force_lsn() argument 3453 xfs_log_force_lsn( xfs_mount_t *mp, xfs_lsn_t lsn, uint flags) xfs_log_force_lsn() argument
|
H A D | xfs_log.h | 148 xfs_lsn_t lsn, 152 xfs_lsn_t lsn,
|
H A D | xfs_log_priv.h | 252 xfs_lsn_t commit_lsn; /* chkpt commit record lsn */ 354 * The reservation head lsn is not made up of a cycle number and block number. 400 /* lsn of last LR on disk */ 402 /* lsn of 1st LR with unflushed * buffers */ 470 xlog_crack_atomic_lsn(atomic64_t *lsn, uint *cycle, uint *block) xlog_crack_atomic_lsn() argument 472 xfs_lsn_t val = atomic64_read(lsn); xlog_crack_atomic_lsn() 482 xlog_assign_atomic_lsn(atomic64_t *lsn, uint cycle, uint block) xlog_assign_atomic_lsn() argument 484 atomic64_set(lsn, xlog_assign_lsn(cycle, block)); xlog_assign_atomic_lsn()
|
H A D | xfs_file.c | 181 xfs_lsn_t lsn = 0; xfs_dir_fsync() local 187 lsn = ip->i_itemp->ili_last_lsn; xfs_dir_fsync() 190 if (!lsn) xfs_dir_fsync() 192 return _xfs_log_force_lsn(mp, lsn, XFS_LOG_SYNC, NULL); xfs_dir_fsync() 207 xfs_lsn_t lsn = 0; xfs_file_fsync() local 242 lsn = ip->i_itemp->ili_last_lsn; xfs_file_fsync() 246 if (lsn) xfs_file_fsync() 247 error = _xfs_log_force_lsn(mp, lsn, XFS_LOG_SYNC, &log_flushed); xfs_file_fsync()
|
H A D | xfs_dquot.c | 900 * Thus, we only bother if the dquot's lsn has xfs_qm_dqflush_done() 901 * not changed. First we check the lsn outside the lock xfs_qm_dqflush_done() 1001 * Clear the dirty field and remember the flush lsn for later use. xfs_qm_dqflush() 1009 * copy the lsn into the on-disk dquot now while we have the in memory xfs_qm_dqflush()
|
H A D | xfs_trans.h | 39 xfs_lsn_t li_lsn; /* last on-disk lsn */
|
H A D | xfs_buf_item.c | 666 * write of it completed at the given lsn. 669 * those cases we simply return the given lsn. 678 * by returning the original lsn of that transaction here rather than 684 xfs_lsn_t lsn) xfs_buf_item_committed() 692 return lsn; xfs_buf_item_committed() 682 xfs_buf_item_committed( struct xfs_log_item *lip, xfs_lsn_t lsn) xfs_buf_item_committed() argument
|
H A D | xfs_log_cil.c | 526 * during log forces to extract the commit lsn of the sequence that xlog_cil_push() 572 * record lsn then we can move the tail beyond the grant write head. xlog_cil_push() 826 * We return the current commit lsn to allow the callers to determine if a
|
H A D | xfs_trans.c | 851 * Push into AIL immediately, rechecking the lsn once xfs_trans_committed_bulk()
|
H A D | xfs_inode.c | 3584 * moved in the AIL from xfs_iflush_done(). In order to read the lsn we xfs_iflush_int() 3603 /* update the lsn in the on disk inode if required */ xfs_iflush_int()
|
/linux-4.1.27/fs/jfs/ |
H A D | jfs_logmgr.c | 58 * lsn - 64-bit monotonically increasing integer vs 233 * RETURN: lsn - offset to the next log record to write (end-of-log); 241 int lsn; lmLog() local 261 * initialize/update page/transaction recovery lsn lmLog() 263 lsn = log->lsn; lmLog() 268 * initialize page lsn if first log write of the page lmLog() 270 if (mp->lsn == 0) { lmLog() 272 mp->lsn = lsn; lmLog() 280 * initialize/update lsn of tblock of the page lmLog() 282 * transaction inherits oldest lsn of pages associated lmLog() 288 * allocation map pages inherit transaction lsn at lmLog() 295 * initialize transaction lsn: lmLog() 297 if (tblk->lsn == 0) { lmLog() 298 /* inherit lsn of its first page logged */ lmLog() 299 tblk->lsn = mp->lsn; lmLog() 306 * update transaction lsn: lmLog() 309 /* inherit oldest/smallest lsn of page */ lmLog() 310 logdiff(diffp, mp->lsn, log); lmLog() 311 logdiff(difft, tblk->lsn, log); lmLog() 313 /* update tblock lsn with page lsn */ lmLog() 314 tblk->lsn = mp->lsn; lmLog() 327 lsn = lmWriteRecord(log, tblk, lrd, tlck); lmLog() 332 logdiff(diffp, lsn, log); lmLog() 334 lsn = lmLogSync(log, 0); lmLog() 336 /* update end-of-log lsn */ lmLog() 337 log->lsn = lsn; lmLog() 342 return lsn; lmLog() 360 int lsn = 0; /* end-of-log address */ lmWriteRecord() local 510 lsn = (log->page << L2LOGPSIZE) + dstoffset; lmWriteRecord() 513 tblk->clsn = lsn; lmWriteRecord() 546 return lsn; lmWriteRecord() 558 return lsn; lmWriteRecord() 950 int lsn; lmLogSync() local 970 log->sync = log->lsn; lmLogSync() 974 log->sync = lp->lsn; lmLogSync() 990 lsn = lmWriteRecord(log, NULL, &lrd, NULL); lmLogSync() 994 lsn = log->lsn; lmLogSync() 1001 logdiff(written, lsn, log); lmLogSync() 1026 log->syncpt = log->sync = lsn; lmLogSync() 1040 jfs_info("log barrier on: lsn=0x%x syncpt=0x%x", lsn, lmLogSync() 1048 return lsn; lmLogSync() 1270 int lsn = 0; lmLogInit() local 1363 jfs_info("lmLogInit: lsn:0x%x page:%d eor:%d:%d", lmLogInit() 1390 lsn = lmWriteRecord(log, NULL, &lrd, NULL); lmLogInit() 1412 log->lsn = lsn; lmLogInit() 1413 log->syncpt = lsn; lmLogInit() 1417 jfs_info("lmLogInit: lsn:0x%x syncpt:0x%x sync:0x%x", lmLogInit() 1418 log->lsn, log->syncpt, log->sync); lmLogInit() 1423 log->clsn = lsn; lmLogInit() 1658 int lsn; lmLogShutdown() local 1678 lsn = lmWriteRecord(log, NULL, &lrd, NULL); lmLogShutdown() 1696 logsuper->end = cpu_to_le32(lsn); lmLogShutdown() 1700 jfs_info("lmLogShutdown: lsn:0x%x page:%d eor:%d", lmLogShutdown() 1701 lsn, log->page, log->eor); lmLogShutdown() 2252 /* update committed lsn */ lbmIODone()
|
H A D | jfs_metapage.c | 192 mp->lsn = 0; alloc_metapage() 298 * acquiring logsync lock, recheck lsn remove_from_logsync() 304 if (mp->lsn) { remove_from_logsync() 306 mp->lsn = 0; remove_from_logsync() 322 if (mp->lsn) last_write_complete() 559 if (mp->lsn) metapage_releasepage() 760 } else if (mp->lsn) /* discard_metapage doesn't remove it */ release_metapage() 802 if (mp->lsn) __invalidate_metapages()
|
H A D | jfs_metapage.h | 29 int lsn; member in struct:metapage 116 * This serializes access to mp->lsn when metapages are added to logsynclist
|
H A D | jfs_txnmgr.h | 43 s32 lsn; /* recovery lsn */ member in struct:tblock 57 s32 clsn; /* commit lsn */
|
H A D | jfs_logmgr.h | 405 int lsn; /* 4: end-of-log */ member in struct:jfs_log 482 s32 lsn; /* log sequence number */ member in struct:logsyncblk 495 /* compute the difference in bytes of lsn from sync point */ 496 #define logdiff(diff, lsn, log)\ 498 diff = (lsn) - (log)->syncpt;\
|
H A D | jfs_txnmgr.c | 431 tblk->next = tblk->last = tblk->xflag = tblk->flag = tblk->lsn = 0; txBegin() 559 jfs_info("log barrier off: 0x%x", log->lsn); txEnd() 720 if ((tid == 0) && mp->lsn) txLock() 978 * (allocation map pages inherited lsn of tblk and txUnlock() 981 if (tblk->lsn) { txUnlock() 2314 * update lsn of the pmap page; txUpdateMap() 2398 * update lsn of the pmap page; txUpdateMap() 2437 * lsn - log sequence number; 2662 * reset lsn of page to avoid logwarap: txAbort() 2669 if (mp->xflag & COMMIT_PAGE && mp->lsn) txAbort()
|
H A D | jfs_dmap.c | 435 int lsn, difft, diffp; dbUpdatePMap() local 447 /* compute delta of transaction lsn from log syncpt */ dbUpdatePMap() 448 lsn = tblk->lsn; dbUpdatePMap() 450 logdiff(difft, lsn, log); dbUpdatePMap() 534 * update dmap lsn dbUpdatePMap() 542 if (mp->lsn != 0) { dbUpdatePMap() 543 /* inherit older/smaller lsn */ dbUpdatePMap() 544 logdiff(diffp, mp->lsn, log); dbUpdatePMap() 546 mp->lsn = lsn; dbUpdatePMap() 559 mp->lsn = lsn; dbUpdatePMap()
|
H A D | jfs_imap.c | 2747 int lsn, difft, diffp; diUpdatePMap() local 2818 * update iag lsn diUpdatePMap() 2820 lsn = tblk->lsn; diUpdatePMap() 2823 if (mp->lsn != 0) { diUpdatePMap() 2824 /* inherit older/smaller lsn */ diUpdatePMap() 2825 logdiff(difft, lsn, log); diUpdatePMap() 2826 logdiff(diffp, mp->lsn, log); diUpdatePMap() 2828 mp->lsn = lsn; diUpdatePMap() 2840 mp->lsn = lsn; diUpdatePMap()
|
/linux-4.1.27/arch/s390/include/asm/ |
H A D | chpid.h | 13 u8 lsn; member in struct:channel_path_desc
|
/linux-4.1.27/fs/xfs/libxfs/ |
H A D | xfs_log_recover.h | 51 xfs_lsn_t r_lsn; /* xact lsn */
|
H A D | xfs_log_format.h | 65 /* get lsn fields */ 66 #define CYCLE_LSN(lsn) ((uint)((lsn)>>32)) 67 #define BLOCK_LSN(lsn) ((uint)(lsn)) 70 #define CYCLE_LSN_DISK(lsn) (((__be32 *)&(lsn))[0]) 160 __be64 h_lsn; /* lsn of this LR : 8 */ 161 __be64 h_tail_lsn; /* lsn of 1st LR w/ buffers not committed: 8 */
|
H A D | xfs_da_format.h | 62 __be64 lsn; /* sequence number of last write */ member in struct:xfs_da3_blkinfo 359 __be64 lsn; /* sequence number of last write */ member in struct:xfs_dir3_blk_hdr
|
H A D | xfs_dir2_block.c | 117 hdr3->lsn = cpu_to_be64(bip->bli_item.li_lsn); xfs_dir3_block_write_verify()
|
H A D | xfs_dir2_data.c | 299 hdr3->lsn = cpu_to_be64(bip->bli_item.li_lsn); xfs_dir3_data_write_verify()
|
H A D | xfs_format.h | 1380 * The blkno, crc, lsn, owner and uuid fields are only available in filesystems
|
H A D | xfs_dir2_leaf.c | 211 hdr3->info.lsn = cpu_to_be64(bip->bli_item.li_lsn); __write_verify()
|
H A D | xfs_attr_leaf.c | 300 hdr3->info.lsn = cpu_to_be64(bip->bli_item.li_lsn); xfs_attr3_leaf_write_verify()
|
H A D | xfs_da_btree.c | 195 hdr3->info.lsn = cpu_to_be64(bip->bli_item.li_lsn); xfs_da3_node_write_verify()
|
H A D | xfs_dir2_node.c | 144 hdr3->lsn = cpu_to_be64(bip->bli_item.li_lsn); xfs_dir3_free_write_verify()
|
/linux-4.1.27/drivers/hwmon/ |
H A D | sch56xx-common.c | 239 int msb, lsn; sch56xx_read_virtual_reg12() local 246 lsn = sch56xx_read_virtual_reg(addr, lsn_reg); sch56xx_read_virtual_reg12() 247 if (lsn < 0) sch56xx_read_virtual_reg12() 248 return lsn; sch56xx_read_virtual_reg12() 251 return (msb << 4) | (lsn >> 4); sch56xx_read_virtual_reg12() 253 return (msb << 4) | (lsn & 0x0f); sch56xx_read_virtual_reg12()
|
/linux-4.1.27/fs/ntfs/ |
H A D | logfile.c | 328 * @lsn: [OUT] set to the current logfile lsn on success 341 * Simillarly, if @lsn is not NULL, on success *@lsn will be set to the current 342 * logfile lsn according to this restart page. On failure, *@lsn is undefined. 351 LSN *lsn) ntfs_check_and_load_restart_page() 451 if (lsn) { ntfs_check_and_load_restart_page() 453 *lsn = sle64_to_cpu(ra->current_lsn); ntfs_check_and_load_restart_page() 455 *lsn = sle64_to_cpu(rp->chkdsk_lsn); ntfs_check_and_load_restart_page() 349 ntfs_check_and_load_restart_page(struct inode *vi, RESTART_PAGE_HEADER *rp, s64 pos, RESTART_PAGE_HEADER **wrp, LSN *lsn) ntfs_check_and_load_restart_page() argument
|
H A D | layout.h | 333 /* 8*/ le64 lsn; /* $LogFile sequence number for this record. member in struct:__anon11148 402 /* 8*/ le64 lsn; /* $LogFile sequence number for this record. member in struct:__anon11149 2058 /* 8*/ sle64 lsn; /* $LogFile sequence number of the last member in struct:__anon11208
|
H A D | mft.c | 2053 m->lsn = 0; ntfs_mft_record_layout()
|
/linux-4.1.27/drivers/s390/cio/ |
H A D | chsc.h | 26 u8 lsn; member in struct:channel_path_desc_fmt1
|