From: Sage Weil Date: Tue, 23 Mar 2010 19:13:09 +0000 (-0700) Subject: todo X-Git-Tag: v0.20~203 X-Git-Url: http://git-server-git.apps.pok.os.sepia.ceph.com/?a=commitdiff_plain;h=d77185eb3484ca5eba67f6f0822061db02863e07;p=ceph.git todo --- diff --git a/src/TODO b/src/TODO index 5db30e3163d6..29b0dd72cfa7 100644 --- a/src/TODO +++ b/src/TODO @@ -60,10 +60,126 @@ filestore - need an osdmap cache layer? bugs +- mds states + - closing -> opening transition - mds prepare_force_open_sessions, then import aborts.. session is still OPENING but no client_session is sent... - rm -r failure (on kernel tree) - dbench 1, restart mds (may take a few times), dbench will error out. +- kclient lockdep warning +[ 1615.328733] ======================================================= +[ 1615.331050] [ INFO: possible circular locking dependency detected ] +[ 1615.331050] 2.6.34-rc2 #22 +[ 1615.331050] ------------------------------------------------------- +[ 1615.331050] fixdep/3263 is trying to acquire lock: +[ 1615.331050] (&osdc->request_mutex){+.+...}, at: [] ceph_osdc_start_request+0x4d/0x278 [ceph] +[ 1615.331050] +[ 1615.331050] but task is already holding lock: +[ 1615.331050] (&mm->mmap_sem){++++++}, at: [] do_page_fault+0x104/0x278 +[ 1615.331050] +[ 1615.331050] which lock already depends on the new lock. +[ 1615.331050] +[ 1615.331050] +[ 1615.331050] the existing dependency chain (in reverse order) is: +[ 1615.331050] +[ 1615.331050] -> #3 (&mm->mmap_sem){++++++}: +[ 1615.331050] [] validate_chain+0xa4d/0xd28 +[ 1615.331050] [] __lock_acquire+0x7d1/0x84e +[ 1615.331050] [] lock_acquire+0x88/0xa5 +[ 1615.331050] [] might_fault+0x90/0xb3 +[ 1615.331050] [] memcpy_fromiovecend+0x54/0x8e +[ 1615.331050] [] ip_generic_getfrag+0x2a/0x8f +[ 1615.331050] [] ip_append_data+0x5f6/0x971 +[ 1615.331050] [] udp_sendmsg+0x4e8/0x603 +[ 1615.331050] [] inet_sendmsg+0x46/0x53 +[ 1615.331050] [] sock_sendmsg+0xd4/0xf5 +[ 1615.331050] [] sys_sendto+0xdf/0x107 +[ 1615.331050] [] system_call_fastpath+0x16/0x1b +[ 1615.331050] +[ 1615.331050] -> #2 (sk_lock-AF_INET){+.+.+.}: +[ 1615.331050] [] validate_chain+0xa4d/0xd28 +[ 1615.331050] [] __lock_acquire+0x7d1/0x84e +[ 1615.331050] [] lock_acquire+0x88/0xa5 +[ 1615.331050] [] lock_sock_nested+0xeb/0xff +[ 1615.331050] [] inet_stream_connect+0x2b/0x25c +[ 1615.331050] [] try_write+0x26e/0x102c [ceph] +[ 1615.331050] [] con_work+0x126/0x6bc [ceph] +[ 1615.529553] [] worker_thread+0x1e8/0x2fa +[ 1615.529553] [] kthread+0x7d/0x85 +[ 1615.529553] [] kernel_thread_helper+0x4/0x10 +[ 1615.529553] +[ 1615.529553] -> #1 (&con->mutex){+.+.+.}: +[ 1615.529553] [] validate_chain+0xa4d/0xd28 +[ 1615.529553] [] __lock_acquire+0x7d1/0x84e +[ 1615.529553] [] lock_acquire+0x88/0xa5 +[ 1615.529553] [] mutex_lock_nested+0x62/0x32c +[ 1615.529553] [] ceph_con_send+0xb3/0x244 [ceph] +[ 1615.529553] [] __send_request+0x108/0x196 [ceph] +[ 1615.529553] [] ceph_osdc_start_request+0x175/0x278 [ceph] +[ 1615.529553] [] ceph_writepages_start+0xb23/0x112a [ceph] +[ 1615.529553] [] do_writepages+0x1f/0x28 +[ 1615.529553] [] writeback_single_inode+0xb6/0x1f5 +[ 1615.529553] [] writeback_inodes_wb+0x2d1/0x378 +[ 1615.529553] [] wb_writeback+0x166/0x1e0 +[ 1615.529553] [] wb_do_writeback+0x83/0x1d3 +[ 1615.529553] [] bdi_writeback_task+0x2e/0x9b +[ 1615.529553] [] bdi_start_fn+0x71/0xd2 +[ 1615.529553] [] kthread+0x7d/0x85 +[ 1615.529553] [] kernel_thread_helper+0x4/0x10 +[ 1615.529553] +[ 1615.529553] -> #0 (&osdc->request_mutex){+.+...}: +[ 1615.529553] [] validate_chain+0x739/0xd28 +[ 1615.529553] [] __lock_acquire+0x7d1/0x84e +[ 1615.529553] [] lock_acquire+0x88/0xa5 +[ 1615.529553] [] mutex_lock_nested+0x62/0x32c +[ 1615.529553] [] ceph_osdc_start_request+0x4d/0x278 [ceph] +[ 1615.529553] [] ceph_osdc_readpages+0x123/0x222 [ceph] +[ 1615.529553] [] ceph_readpages+0x193/0x456 [ceph] +[ 1615.529553] [] __do_page_cache_readahead+0x17d/0x1f5 +[ 1615.529553] [] ra_submit+0x1c/0x20 +[ 1615.529553] [] ondemand_readahead+0x264/0x277 +[ 1615.529553] [] page_cache_sync_readahead+0x33/0x35 +[ 1615.529553] [] filemap_fault+0x143/0x31f +[ 1615.529553] [] __do_fault+0x50/0x415 +[ 1615.529553] [] handle_mm_fault+0x334/0x6a6 +[ 1615.529553] [] do_page_fault+0x1f3/0x278 +[ 1615.529553] [] page_fault+0x1f/0x30 +[ 1615.529553] +[ 1615.529553] other info that might help us debug this: +[ 1615.529553] +[ 1615.529553] 1 lock held by fixdep/3263: +[ 1615.529553] #0: (&mm->mmap_sem){++++++}, at: [] do_page_fault+0x104/0x278 +[ 1615.529553] +[ 1615.529553] stack backtrace: +[ 1615.529553] Pid: 3263, comm: fixdep Not tainted 2.6.34-rc2 #22 +[ 1615.529553] Call Trace: +[ 1615.529553] [] print_circular_bug+0xb3/0xc1 +[ 1615.529553] [] validate_chain+0x739/0xd28 +[ 1615.529553] [] ? native_sched_clock+0x37/0x71 +[ 1615.824177] [] __lock_acquire+0x7d1/0x84e +[ 1615.824177] [] lock_acquire+0x88/0xa5 +[ 1615.824177] [] ? ceph_osdc_start_request+0x4d/0x278 [ceph] +[ 1615.824177] [] ? ceph_osdc_start_request+0x4d/0x278 [ceph] +[ 1615.824177] [] mutex_lock_nested+0x62/0x32c +[ 1615.824177] [] ? ceph_osdc_start_request+0x4d/0x278 [ceph] +[ 1615.824177] [] ceph_osdc_start_request+0x4d/0x278 [ceph] +[ 1615.824177] [] ceph_osdc_readpages+0x123/0x222 [ceph] +[ 1615.824177] [] ceph_readpages+0x193/0x456 [ceph] +[ 1615.824177] [] ? native_sched_clock+0x37/0x71 +[ 1615.824177] [] ? get_lock_stats+0x19/0x4c +[ 1615.824177] [] __do_page_cache_readahead+0x17d/0x1f5 +[ 1615.824177] [] ? __do_page_cache_readahead+0x7c/0x1f5 +[ 1615.824177] [] ? find_get_page+0xd9/0x12d +[ 1615.824177] [] ra_submit+0x1c/0x20 +[ 1615.916887] [] ondemand_readahead+0x264/0x277 +[ 1615.916887] [] page_cache_sync_readahead+0x33/0x35 +[ 1615.931403] [] filemap_fault+0x143/0x31f +[ 1615.931403] [] __do_fault+0x50/0x415 +[ 1615.931403] [] ? __lock_acquire+0x7eb/0x84e +[ 1615.946963] [] handle_mm_fault+0x334/0x6a6 +[ 1615.946963] [] do_page_fault+0x1f3/0x278 +[ 1615.946963] [] page_fault+0x1f/0x30 + - kclient: moonbeamer gets this with iozone -a... [17608.696906] ------------[ cut here ]------------ [17608.701761] WARNING: at lib/kref.c:43 kref_get+0x23/0x2a()