Displaying 8 results from an estimated 8 matches for "unlock_new_inode".
2009 Feb 20
2
[PATCH 1/1] OCFS2: add error check for ocfs2_read_locked_inode() call
...&& inode->i_state & I_NEW) {
mlog(0, "Inode was not in inode cache, reading it.\n");
- ocfs2_read_locked_inode(inode, &args);
+ status = ocfs2_read_locked_inode(inode, &args);
+ if (status) {
+ iput(inode);
+ inode = ERR_PTR(status);
+ goto bail;
+ }
unlock_new_inode(inode);
}
if (inode == NULL) {
2009 Feb 26
1
[PATCH 0/7] OCFS2 locking fixes and lockdep annotations
Hi,
the first four patches in this series fix locking problems in OCFS2 quota code (three of
them can lead to potential deadlocks). The fifth patch reorders ip_alloc_sem for directories
to be acquired before localalloc locks. Mark would you please merge these?
The last two patches implement lockdep annotations for OCFS2 cluster locks. We annotate all
the cluster locks except for special ones
2009 Jun 04
2
[PATCH 0/2] OCFS2 lockdep support
Hi,
here comes the next version of OCFS2 lockdep support. I've dropped patches
with fixes from the series since they were already merged.
As Joel suggested, I've simplified the main patch a bit so that we don't
have ifdefs around lock declarations and there are also a few other minor
improvements.
Honza
2011 Sep 05
1
Error on creating snapshots (btrfs: could not do orphan cleanup -116)
This happens on a freshly created btrfs filesystem in a raid10 (4x1TB)
configuration with three
subvolumes and 1.5 TB data.
When I try to snapshot one of the subvolumes (with 100 GB of data), it
says that the snapshot
creation failed and I get the following error message:
btrfs: could not do orphan cleanup -116
After the failure:
- The snapshot exists in `btrfs subvolume list''
- The
2009 Jun 02
10
[PATCH 0/7] [RESEND] Fix some deadlocks in quota code and implement lockdep for cluster locks
Hi,
I'm resending this patch series. It's rediffed against linux-next branch of
Joel's git tree. The first four patches are obvious fixes of deadlocks in quota
code and should go in as soon as possible. The other three patches implement
lockdep support for OCFS2 cluster locks. So you can have a look whether the
code make sence to you and possibly merge them. They should be NOP when
2008 Jun 24
1
[RFC][PATCH] btrfs orphan code
...;
+
+ if (!inode) {
+ inode = ERR_PTR(-EACCES);
+ goto out;
+ }
+
+ /* should technically always be true */
+ if (inode->i_state & I_NEW) {
+ BTRFS_I(inode)->root = root;
+ memcpy(&BTRFS_I(inode)->location, &location, sizeof(location));
+ btrfs_read_locked_inode(inode);
+ unlock_new_inode(inode);
+ }
+out:
+ return inode;
}
static int merge_extent_mapping(struct extent_map_tree *em_tree,
@@ -2737,6 +2995,7 @@ struct inode *btrfs_alloc_inode(struct s
ei->ordered_trans = 0;
ei->i_acl = BTRFS_ACL_NOT_CACHED;
ei->i_default_acl = BTRFS_ACL_NOT_CACHED;
+ INIT_LIST_HEAD(...
2020 Oct 17
10
[RFC] treewide: cleanup unreachable breaks
...fs/inode.c
@@ -161,11 +161,10 @@ struct inode *efs_iget(struct super_block *super, unsigned long ino)
init_special_inode(inode, inode->i_mode, device);
break;
default:
pr_warn("unsupported inode mode %o\n", inode->i_mode);
goto read_inode_error;
- break;
}
unlock_new_inode(inode);
return inode;
diff --git a/fs/ocfs2/cluster/tcp.c b/fs/ocfs2/cluster/tcp.c
index 79a231719460..3bd8119bed5e 100644
--- a/fs/ocfs2/cluster/tcp.c
+++ b/fs/ocfs2/cluster/tcp.c
@@ -1196,11 +1196,10 @@ static int o2net_process_message(struct o2net_sock_container *sc,
break;
de...
2020 Oct 17
10
[RFC] treewide: cleanup unreachable breaks
...fs/inode.c
@@ -161,11 +161,10 @@ struct inode *efs_iget(struct super_block *super, unsigned long ino)
init_special_inode(inode, inode->i_mode, device);
break;
default:
pr_warn("unsupported inode mode %o\n", inode->i_mode);
goto read_inode_error;
- break;
}
unlock_new_inode(inode);
return inode;
diff --git a/fs/ocfs2/cluster/tcp.c b/fs/ocfs2/cluster/tcp.c
index 79a231719460..3bd8119bed5e 100644
--- a/fs/ocfs2/cluster/tcp.c
+++ b/fs/ocfs2/cluster/tcp.c
@@ -1196,11 +1196,10 @@ static int o2net_process_message(struct o2net_sock_container *sc,
break;
de...