[ This code is above my pay grade. The cache is refcounted so it might not be freed. But hopefully the patch is recent enough that the details are still top of the head for you to review it quickly? Hopefully, no big deal if it's a false positive. These reports are one time only. -dan ] Hello Jan Kara, The patch 132991ed2882: "ext4: fix race when reusing xattr blocks" from Jul 12, 2022, leads to the following Smatch static checker warning: fs/ext4/xattr.c:2038 ext4_xattr_block_set() warn: 'ea_block_cache' was already freed. fs/ext4/xattr.c 1850 static int 1851 ext4_xattr_block_set(handle_t *handle, struct inode *inode, 1852 struct ext4_xattr_info *i, 1853 struct ext4_xattr_block_find *bs) 1854 { 1855 struct super_block *sb = inode->i_sb; 1856 struct buffer_head *new_bh = NULL; 1857 struct ext4_xattr_search s_copy = bs->s; 1858 struct ext4_xattr_search *s = &s_copy; 1859 struct mb_cache_entry *ce = NULL; 1860 int error = 0; 1861 struct mb_cache *ea_block_cache = EA_BLOCK_CACHE(inode); 1862 struct inode *ea_inode = NULL, *tmp_inode; 1863 size_t old_ea_inode_quota = 0; 1864 unsigned int ea_ino; 1865 1866 1867 #define header(x) ((struct ext4_xattr_header *)(x)) 1868 1869 if (s->base) { 1870 int offset = (char *)s->here - bs->bh->b_data; 1871 1872 BUFFER_TRACE(bs->bh, "get_write_access"); 1873 error = ext4_journal_get_write_access(handle, sb, bs->bh, 1874 EXT4_JTR_NONE); 1875 if (error) 1876 goto cleanup; 1877 lock_buffer(bs->bh); 1878 1879 if (header(s->base)->h_refcount == cpu_to_le32(1)) { 1880 __u32 hash = le32_to_cpu(BHDR(bs->bh)->h_hash); 1881 1882 /* 1883 * This must happen under buffer lock for 1884 * ext4_xattr_block_set() to reliably detect modified 1885 * block 1886 */ 1887 if (ea_block_cache) { 1888 struct mb_cache_entry *oe; 1889 1890 oe = mb_cache_entry_delete_or_get(ea_block_cache, 1891 hash, bs->bh->b_blocknr); "ea_block_cache" potentially freed here? 1892 if (oe) { 1893 /* 1894 * Xattr block is getting reused. Leave 1895 * it alone. 1896 */ 1897 mb_cache_entry_put(ea_block_cache, oe); Also here? 1898 goto clone_block; 1899 } 1900 } 1901 ea_bdebug(bs->bh, "modifying in-place"); 1902 error = ext4_xattr_set_entry(i, s, handle, inode, 1903 true /* is_block */); 1904 ext4_xattr_block_csum_set(inode, bs->bh); 1905 unlock_buffer(bs->bh); 1906 if (error == -EFSCORRUPTED) 1907 goto bad_block; 1908 if (!error) 1909 error = ext4_handle_dirty_metadata(handle, 1910 inode, 1911 bs->bh); 1912 if (error) 1913 goto cleanup; 1914 goto inserted; 1915 } 1916 clone_block: 1917 unlock_buffer(bs->bh); 1918 ea_bdebug(bs->bh, "cloning"); 1919 s->base = kmemdup(BHDR(bs->bh), bs->bh->b_size, GFP_NOFS); 1920 error = -ENOMEM; 1921 if (s->base == NULL) 1922 goto cleanup; 1923 s->first = ENTRY(header(s->base)+1); 1924 header(s->base)->h_refcount = cpu_to_le32(1); 1925 s->here = ENTRY(s->base + offset); 1926 s->end = s->base + bs->bh->b_size; 1927 1928 /* 1929 * If existing entry points to an xattr inode, we need 1930 * to prevent ext4_xattr_set_entry() from decrementing 1931 * ref count on it because the reference belongs to the 1932 * original block. In this case, make the entry look 1933 * like it has an empty value. 1934 */ 1935 if (!s->not_found && s->here->e_value_inum) { 1936 ea_ino = le32_to_cpu(s->here->e_value_inum); 1937 error = ext4_xattr_inode_iget(inode, ea_ino, 1938 le32_to_cpu(s->here->e_hash), 1939 &tmp_inode); 1940 if (error) 1941 goto cleanup; 1942 1943 if (!ext4_test_inode_state(tmp_inode, 1944 EXT4_STATE_LUSTRE_EA_INODE)) { 1945 /* 1946 * Defer quota free call for previous 1947 * inode until success is guaranteed. 1948 */ 1949 old_ea_inode_quota = le32_to_cpu( 1950 s->here->e_value_size); 1951 } 1952 iput(tmp_inode); 1953 1954 s->here->e_value_inum = 0; 1955 s->here->e_value_size = 0; 1956 } 1957 } else { 1958 /* Allocate a buffer where we construct the new block. */ 1959 s->base = kzalloc(sb->s_blocksize, GFP_NOFS); 1960 error = -ENOMEM; 1961 if (s->base == NULL) 1962 goto cleanup; 1963 header(s->base)->h_magic = cpu_to_le32(EXT4_XATTR_MAGIC); 1964 header(s->base)->h_blocks = cpu_to_le32(1); 1965 header(s->base)->h_refcount = cpu_to_le32(1); 1966 s->first = ENTRY(header(s->base)+1); 1967 s->here = ENTRY(header(s->base)+1); 1968 s->end = s->base + sb->s_blocksize; 1969 } 1970 1971 error = ext4_xattr_set_entry(i, s, handle, inode, true /* is_block */); 1972 if (error == -EFSCORRUPTED) 1973 goto bad_block; 1974 if (error) 1975 goto cleanup; 1976 1977 if (i->value && s->here->e_value_inum) { 1978 /* 1979 * A ref count on ea_inode has been taken as part of the call to 1980 * ext4_xattr_set_entry() above. We would like to drop this 1981 * extra ref but we have to wait until the xattr block is 1982 * initialized and has its own ref count on the ea_inode. 1983 */ 1984 ea_ino = le32_to_cpu(s->here->e_value_inum); 1985 error = ext4_xattr_inode_iget(inode, ea_ino, 1986 le32_to_cpu(s->here->e_hash), 1987 &ea_inode); 1988 if (error) { 1989 ea_inode = NULL; 1990 goto cleanup; 1991 } 1992 } 1993 1994 inserted: 1995 if (!IS_LAST_ENTRY(s->first)) { 1996 new_bh = ext4_xattr_block_cache_find(inode, header(s->base), 1997 &ce); 1998 if (new_bh) { 1999 /* We found an identical block in the cache. */ 2000 if (new_bh == bs->bh) 2001 ea_bdebug(new_bh, "keeping"); 2002 else { 2003 u32 ref; 2004 2005 WARN_ON_ONCE(dquot_initialize_needed(inode)); 2006 2007 /* The old block is released after updating 2008 the inode. */ 2009 error = dquot_alloc_block(inode, 2010 EXT4_C2B(EXT4_SB(sb), 1)); 2011 if (error) 2012 goto cleanup; 2013 BUFFER_TRACE(new_bh, "get_write_access"); 2014 error = ext4_journal_get_write_access( 2015 handle, sb, new_bh, 2016 EXT4_JTR_NONE); 2017 if (error) 2018 goto cleanup_dquot; 2019 lock_buffer(new_bh); 2020 /* 2021 * We have to be careful about races with 2022 * adding references to xattr block. Once we 2023 * hold buffer lock xattr block's state is 2024 * stable so we can check the additional 2025 * reference fits. 2026 */ 2027 ref = le32_to_cpu(BHDR(new_bh)->h_refcount) + 1; 2028 if (ref > EXT4_XATTR_REFCOUNT_MAX) { 2029 /* 2030 * Undo everything and check mbcache 2031 * again. 2032 */ 2033 unlock_buffer(new_bh); 2034 dquot_free_block(inode, 2035 EXT4_C2B(EXT4_SB(sb), 2036 1)); 2037 brelse(new_bh); --> 2038 mb_cache_entry_put(ea_block_cache, ce); Warning. 2039 ce = NULL; 2040 new_bh = NULL; 2041 goto inserted; 2042 } 2043 BHDR(new_bh)->h_refcount = cpu_to_le32(ref); 2044 if (ref == EXT4_XATTR_REFCOUNT_MAX) 2045 ce->e_reusable = 0; 2046 ea_bdebug(new_bh, "reusing; refcount now=%d", 2047 ref); 2048 ext4_xattr_block_csum_set(inode, new_bh); 2049 unlock_buffer(new_bh); 2050 error = ext4_handle_dirty_metadata(handle, 2051 inode, 2052 new_bh); 2053 if (error) 2054 goto cleanup_dquot; 2055 } 2056 mb_cache_entry_touch(ea_block_cache, ce); 2057 mb_cache_entry_put(ea_block_cache, ce); 2058 ce = NULL; 2059 } else if (bs->bh && s->base == bs->bh->b_data) { 2060 /* We were modifying this block in-place. */ 2061 ea_bdebug(bs->bh, "keeping this block"); 2062 ext4_xattr_block_cache_insert(ea_block_cache, bs->bh); 2063 new_bh = bs->bh; 2064 get_bh(new_bh); 2065 } else { 2066 /* We need to allocate a new block */ 2067 ext4_fsblk_t goal, block; 2068 2069 WARN_ON_ONCE(dquot_initialize_needed(inode)); 2070 2071 goal = ext4_group_first_block_no(sb, 2072 EXT4_I(inode)->i_block_group); 2073 2074 /* non-extent files can't have physical blocks past 2^32 */ 2075 if (!(ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS))) 2076 goal = goal & EXT4_MAX_BLOCK_FILE_PHYS; 2077 2078 block = ext4_new_meta_blocks(handle, inode, goal, 0, 2079 NULL, &error); 2080 if (error) 2081 goto cleanup; 2082 2083 if (!(ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS))) 2084 BUG_ON(block > EXT4_MAX_BLOCK_FILE_PHYS); 2085 2086 ea_idebug(inode, "creating block %llu", 2087 (unsigned long long)block); 2088 2089 new_bh = sb_getblk(sb, block); 2090 if (unlikely(!new_bh)) { 2091 error = -ENOMEM; 2092 getblk_failed: 2093 ext4_free_blocks(handle, inode, NULL, block, 1, 2094 EXT4_FREE_BLOCKS_METADATA); 2095 goto cleanup; 2096 } 2097 error = ext4_xattr_inode_inc_ref_all(handle, inode, 2098 ENTRY(header(s->base)+1)); 2099 if (error) 2100 goto getblk_failed; 2101 if (ea_inode) { 2102 /* Drop the extra ref on ea_inode. */ 2103 error = ext4_xattr_inode_dec_ref(handle, 2104 ea_inode); 2105 if (error) 2106 ext4_warning_inode(ea_inode, 2107 "dec ref error=%d", 2108 error); 2109 iput(ea_inode); 2110 ea_inode = NULL; 2111 } 2112 2113 lock_buffer(new_bh); 2114 error = ext4_journal_get_create_access(handle, sb, 2115 new_bh, EXT4_JTR_NONE); 2116 if (error) { 2117 unlock_buffer(new_bh); 2118 error = -EIO; 2119 goto getblk_failed; 2120 } 2121 memcpy(new_bh->b_data, s->base, new_bh->b_size); 2122 ext4_xattr_block_csum_set(inode, new_bh); 2123 set_buffer_uptodate(new_bh); 2124 unlock_buffer(new_bh); 2125 ext4_xattr_block_cache_insert(ea_block_cache, new_bh); 2126 error = ext4_handle_dirty_metadata(handle, inode, 2127 new_bh); 2128 if (error) 2129 goto cleanup; 2130 } 2131 } 2132 2133 if (old_ea_inode_quota) 2134 ext4_xattr_inode_free_quota(inode, NULL, old_ea_inode_quota); 2135 2136 /* Update the inode. */ 2137 EXT4_I(inode)->i_file_acl = new_bh ? new_bh->b_blocknr : 0; 2138 2139 /* Drop the previous xattr block. */ 2140 if (bs->bh && bs->bh != new_bh) { 2141 struct ext4_xattr_inode_array *ea_inode_array = NULL; 2142 2143 ext4_xattr_release_block(handle, inode, bs->bh, 2144 &ea_inode_array, 2145 0 /* extra_credits */); 2146 ext4_xattr_inode_array_free(ea_inode_array); 2147 } 2148 error = 0; 2149 2150 cleanup: 2151 if (ea_inode) { 2152 int error2; 2153 2154 error2 = ext4_xattr_inode_dec_ref(handle, ea_inode); 2155 if (error2) 2156 ext4_warning_inode(ea_inode, "dec ref error=%d", 2157 error2); 2158 2159 /* If there was an error, revert the quota charge. */ 2160 if (error) 2161 ext4_xattr_inode_free_quota(inode, ea_inode, 2162 i_size_read(ea_inode)); 2163 iput(ea_inode); 2164 } 2165 if (ce) 2166 mb_cache_entry_put(ea_block_cache, ce); 2167 brelse(new_bh); 2168 if (!(bs->bh && s->base == bs->bh->b_data)) 2169 kfree(s->base); 2170 2171 return error; 2172 2173 cleanup_dquot: 2174 dquot_free_block(inode, EXT4_C2B(EXT4_SB(sb), 1)); 2175 goto cleanup; 2176 2177 bad_block: 2178 EXT4_ERROR_INODE(inode, "bad block %llu", 2179 EXT4_I(inode)->i_file_acl); 2180 goto cleanup; 2181 2182 #undef header 2183 } regards, dan carpenter