zoukankan      html  css  js  c++  java
  • FUSE引起的SIGBUS问题分析报告

    【问题描述】

    经常在AssetManager中扫描apk的时候出现SIGBUS错误:

    pid: 870, tid: 16106, name: Thread-6536  >>> com.xxx.guardprovider <<<
    signal 7 (SIGBUS), code 0 (SI_USER), fault addr 0x9ed30253
        r0 00009a98  r1 b6206e40  r2 0000000c  r3 00000000
        r4 00000000  r5 9fd3f088  r6 0004f4e8  r7 00000000
        r8 00000000  r9 9f9e3800  sl 9fd3f088  fp 9ed30253
        ip b6206e58  sp 9fd3f000  lr b6bee519  pc b61fd6a8  cpsr 20000010
    
    backtrace:
        #00 pc 0000c6a8  /system/lib/libz.so (inflate+944)
        #01 pc 0001c515  /system/lib/libandroidfw.so (_Z15inflateToBufferI12BufferReaderEbRT_Pvll+164)
        #02 pc 0001c597  /system/lib/libandroidfw.so (_ZN7android8ZipUtils15inflateToBufferEPvS1_ll+14)
        #03 pc 00010701  /system/lib/libandroidfw.so (_ZN7android16_CompressedAsset9getBufferEb+38)
        #04 pc 00011911  /system/lib/libandroidfw.so (_ZN7android12AssetManager9SharedZip21setResourceTableAssetEPNS_5AssetE+36)
        #05 pc 000122f9  /system/lib/libandroidfw.so (_ZN7android12AssetManager6ZipSet24setZipResourceTableAssetERKNS_7String8EPNS_5AssetE+28)
        #06 pc 000132cf  /system/lib/libandroidfw.so (_ZNK7android12AssetManager20appendPathToResTableERKNS0_10asset_pathEPj+194)
        #07 pc 0001375f  /system/lib/libandroidfw.so (_ZNK7android12AssetManager11getResTableEb+98)
        #08 pc 00087009  /system/lib/libandroid_runtime.so
    
    memory near fp:
        9ed30230 b5ae1300 31f68d46 04f4e83c 135bbc00  ....F..1<.....[.
        9ed30240 00000e00 73657200 6372756f 612e7365  .....resources.a
        9ed30250 24637372 55bc65dc e030f6d5 410e8275  rsc$.e.U..0.u..A
        9ed30260 544441ba 151410b0 ba524c15 3bbbba53  .ADT.....LR.S..;
        9ed30270 a1eddd0f bbbbbbbb d0e1bbbb dcffe75d  ............]...
        9ed30280 fcf0fdf7 bd9f65ee cc739cd6 fb9cc631  .....e....s.1...
    
    --->9ed30000-9ed7ffff r--    47d000     50000  /storage/emulated/0/UCDownloads/com.qihoo.appstore_300030395.apk

    奇怪的是fatal addr指向的地址实际上是可读的,因为debuggerd能读出来。

    该问题还有一个特点是每次都是读取sdcard内容时出问题。

    【问题分析】

    显然这个问题和sdcard相关,也就是fuse文件系统相关。

    在出问题的地址是sdcard文件映射到内存上去的,这种情况下只有page_fault的时候fuse文件系统才会工作。

    这个问题看起来是fuse的page_fault处理出了问题。

    static const struct vm_operations_struct fuse_file_vm_ops = {
        .close        = fuse_vma_close,
        .fault        = filemap_fault,
        .map_pages    = filemap_map_pages,
        .page_mkwrite    = fuse_page_mkwrite,
        .remap_pages    = generic_file_remap_pages,
    };

    在filemap_fault()中有可能返回SIGBUS的地方都添加日志:

    int filemap_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
    {
        int error;
        struct file *file = vma->vm_file;
        struct address_space *mapping = file->f_mapping;
        struct file_ra_state *ra = &file->f_ra;
        struct inode *inode = mapping->host;
        pgoff_t offset = vmf->pgoff;
        struct page *page;
        loff_t size;
        int ret = 0;
    
        size = round_up(i_size_read(inode), PAGE_CACHE_SIZE);
        if (offset >= size >> PAGE_CACHE_SHIFT)
            printk("filemap_fault :pid=%lld, tgid=%lld size=%lld, offset=%lld
    ",
                          (long long)current->pid, (long long)current->tgid, (long long)size, (long long)offset);
            return VM_FAULT_SIGBUS;
    
        /*
         * Do we have something in the page cache already?
         */
        page = find_get_page(mapping, offset);
        if (likely(page) && !(vmf->flags & FAULT_FLAG_TRIED)) {
            /*
             * We found the page, so try async readahead before
             * waiting for the lock.
             */
            do_async_mmap_readahead(vma, ra, file, page, offset);
        } else if (!page) {
            /* No page in the page cache at all */
            do_sync_mmap_readahead(vma, ra, file, offset);
            count_vm_event(PGMAJFAULT);
            mem_cgroup_count_vm_event(vma->vm_mm, PGMAJFAULT);
            ret = VM_FAULT_MAJOR;
    retry_find:
            page = find_get_page(mapping, offset);
            if (!page)
                goto no_cached_page;
        }
    
        if (!lock_page_or_retry(page, vma->vm_mm, vmf->flags)) {
            page_cache_release(page);
            return ret | VM_FAULT_RETRY;
        }
    
        /* Did it get truncated? */
        if (unlikely(page->mapping != mapping)) {
            unlock_page(page);
            put_page(page);
            goto retry_find;
        }
        VM_BUG_ON_PAGE(page->index != offset, page);
    
        /*
         * We have a locked page in the page cache, now we need to check
         * that it's up-to-date. If not, it is going to be due to an error.
         */
        if (unlikely(!PageUptodate(page)))
            goto page_not_uptodate;
    
        /*
         * Found the page and have a reference on it.
         * We must recheck i_size under page lock.
         */
        size = round_up(i_size_read(inode), PAGE_CACHE_SIZE);
        if (unlikely(offset >= size >> PAGE_CACHE_SHIFT)) {
            unlock_page(page);
            page_cache_release(page);
            printk("filemap_fault retry_find:pid=%lld, tgid=%lld size=%lld, offset=%lld
    ",
                          (long long)current->pid, (long long)current->tgid, (long long)size, (long long)offset);
    
            return VM_FAULT_SIGBUS;
        }
    
        vmf->page = page;
        return ret | VM_FAULT_LOCKED;
    
    no_cached_page:
        /*
         * We're only likely to ever get here if MADV_RANDOM is in
         * effect.
         */
        error = page_cache_read(file, offset);
    
        /*
         * The page we want has now been added to the page cache.
         * In the unlikely event that someone removed it in the
         * meantime, we'll just come back here and read it again.
         */
        if (error >= 0)
            goto retry_find;
    
        /*
         * An error return from page_cache_read can result if the
         * system is low on memory, or a problem occurs while trying
         * to schedule I/O.
         */
        if (error == -ENOMEM)
            return VM_FAULT_OOM;
    printk(
    "filemap_fault no_cached_page:pid=%lld, tgid=%lld size=%lld, offset=%lld, error=%d ", (long long)current->pid, (long long)current->tgid, (long long)size, (long long)offset, error); return VM_FAULT_SIGBUS; page_not_uptodate: /* * Umm, take care of errors if the page isn't up-to-date. * Try to re-read it _once_. We do this synchronously, * because there really aren't any performance issues here * and we need to check for errors. */ ClearPageError(page); error = mapping->a_ops->readpage(file, page); if (!error) { wait_on_page_locked(page); if (!PageUptodate(page)) error = -EIO; } page_cache_release(page); if (!error || error == AOP_TRUNCATED_PAGE) goto retry_find; /* Things didn't work out. Return zero to tell the mm layer so. */ shrink_readahead_size_eio(file, ra); printk("filemap_fault page_not_uptodate:pid=%lld, tgid=%lld size=%lld, offset=%lld, error=%d ", (long long)current->pid, (long long)current->tgid, (long long)size, (long long)offset, error); return VM_FAULT_SIGBUS;

    复现问题后,发现每次都是mapping->a_ops->readpage()返回了-4, 而这个函数正是fuse_readpage()函数。

    #define EINTR 4 /* Interrupted system call */
    
    static int fuse_readpage(struct file *file, struct page *page)
    {
        struct inode *inode = page->mapping->host;
        int err;
    
        err = -EIO;
        if (is_bad_inode(inode))
            goto out;
    
        err = fuse_do_readpage(file, page);
        fuse_invalidate_atime(inode);
     out:
        unlock_page(page);
        return err;
    }
    
    static int fuse_do_readpage(struct file *file, struct page *page)
    {
        struct fuse_io_priv io = { .async = 0, .file = file };
        struct inode *inode = page->mapping->host;
        struct fuse_conn *fc = get_fuse_conn(inode);
        struct fuse_req *req;
        size_t num_read;
        loff_t pos = page_offset(page);
        size_t count = PAGE_CACHE_SIZE;
        u64 attr_ver;
        int err;
    
        /*
         * Page writeback can extend beyond the lifetime of the
         * page-cache page, so make sure we read a properly synced
         * page.
         */
        fuse_wait_on_page_writeback(inode, page->index);
    
        req = fuse_get_req(fc, 1);
        if (IS_ERR(req))
            return PTR_ERR(req);
    
        attr_ver = fuse_get_attr_version(fc);
    
        req->out.page_zeroing = 1;
        req->out.argpages = 1;
        req->num_pages = 1;
        req->pages[0] = page;
        req->page_descs[0].length = count;
        num_read = fuse_send_read(req, &io, pos, count, NULL);
        err = req->out.h.error;
    
        if (!err) {
            /*
             * Short read means EOF.  If file size is larger, truncate it
             */
            if (num_read < count)
                fuse_short_read(req, inode, attr_ver);
    
            SetPageUptodate(page);
        }
    
        fuse_put_request(fc, req);
    
        return err;
    }
    
    static size_t fuse_send_read(struct fuse_req *req, struct fuse_io_priv *io,
                     loff_t pos, size_t count, fl_owner_t owner)
    {
        struct file *file = io->file;
        struct fuse_file *ff = file->private_data;
        struct fuse_conn *fc = ff->fc;
    
        fuse_read_fill(req, file, pos, count, FUSE_READ);
        if (owner != NULL) {
            struct fuse_read_in *inarg = &req->misc.read.in;
    
            inarg->read_flags |= FUSE_READ_LOCKOWNER;
            inarg->lock_owner = fuse_lock_owner_id(fc, owner);
        }
    
        if (io->async)
            return fuse_async_req_send(fc, req, count, io);
    
        fuse_request_send(fc, req);
        return req->out.args[0].size;
    }
    
    void fuse_request_send(struct fuse_conn *fc, struct fuse_req *req)
    {
        req->isreply = 1;
        __fuse_request_send(fc, req);
    }
    
    static void __fuse_request_send(struct fuse_conn *fc, struct fuse_req *req)
    {
        BUG_ON(req->background);
        spin_lock(&fc->lock);
        if (!fc->connected)
            req->out.h.error = -ENOTCONN;
        else if (fc->conn_error)
            req->out.h.error = -ECONNREFUSED;
        else {
            req->in.h.unique = fuse_get_unique(fc);
            queue_request(fc, req);
            /* acquire extra reference, since request is still needed
               after request_end() */
            __fuse_get_request(req);
    
            request_wait_answer(fc, req);
        }
        spin_unlock(&fc->lock);
    }
    
    static void request_wait_answer(struct fuse_conn *fc, struct fuse_req *req)
    __releases(fc->lock)
    __acquires(fc->lock)
    {
        if (!fc->no_interrupt) {
            /* Any signal may interrupt this */
            wait_answer_interruptible(fc, req);
    
            if (req->aborted)
                goto aborted;
            if (req->state == FUSE_REQ_FINISHED)
                return;
    
            req->interrupted = 1;
            if (req->state == FUSE_REQ_SENT)
                queue_interrupt(fc, req);
        }
    
        if (!req->force) {
            sigset_t oldset;
    
            /* Only fatal signals may interrupt this */
            block_sigs(&oldset);
            wait_answer_interruptible(fc, req);
            restore_sigs(&oldset);
    
            if (req->aborted)
                goto aborted;
            if (req->state == FUSE_REQ_FINISHED)
                return;
    
            /* Request is not yet in userspace, bail out */
            if (req->state == FUSE_REQ_PENDING) {
                list_del(&req->list);
                __fuse_put_request(req);
                req->out.h.error = -EINTR;
                return;
            }
        }
    
        /*
         * Either request is already in userspace, or it was forced.
         * Wait it out.
         */
        spin_unlock(&fc->lock);
        wait_event(req->waitq, req->state == FUSE_REQ_FINISHED);
        spin_lock(&fc->lock);
    
        if (!req->aborted)
            return;
    
     aborted:
        BUG_ON(req->state != FUSE_REQ_FINISHED);
        if (req->locked) {
            /* This is uninterruptible sleep, because data is
               being copied to/from the buffers of req.  During
               locked state, there mustn't be any filesystem
               operation (e.g. page fault), since that could lead
               to deadlock */
            spin_unlock(&fc->lock);
            wait_event(req->waitq, !req->locked);
            spin_lock(&fc->lock);
        }
    }

    添加log,最终确认是下面这个wait_answer_interruptible()被打断了:

            block_sigs(&oldset);
            wait_answer_interruptible(fc, req);
            restore_sigs(&oldset);

    其中:

    static void block_sigs(sigset_t *oldset)
    {
        sigset_t mask;
    
        siginitsetinv(&mask, sigmask(SIGKILL));
        sigprocmask(SIG_BLOCK, &mask, oldset);
    }
    
    static void restore_sigs(sigset_t *oldset)
    {
        sigprocmask(SIG_SETMASK, oldset, NULL);
    }
    
    static void wait_answer_interruptible(struct fuse_conn *fc,
                          struct fuse_req *req)
    __releases(fc->lock)
    __acquires(fc->lock)
    {
        if (signal_pending(current))
            return;
    
        spin_unlock(&fc->lock);
        wait_event_interruptible(req->waitq, req->state == FUSE_REQ_FINISHED);
        spin_lock(&fc->lock);
    }

    代码原意是:

    1、屏蔽除SIGKILL以外的所有信号

    2、如果有pending signal(这里应该指的是SIGKILL)则返回

    3、进入interruptible sleep状态

    也就是说除了SIGKILL信号,其他信号都不可能打断这个流程。

    但打印日志后发现出问题的时候sigset_t中并没有任何信号!

    查看代码发现,signal pending并不是检查sigset_t,它只是判断thread_info中的flag:

    static inline int signal_pending(struct task_struct *p)
    {
        return unlikely(test_tsk_thread_flag(p,TIF_SIGPENDING));
    }

    可能是有个路径设置了TIF_SIGPENDING导致的。

    仔细排查日志,发现每次出现问题时,都有如下日志:

    [21814.557813] PM: suspend entry 2016-11-17 02:07:33.679008730 UTC
    [21814.577530] PM: Syncing filesystems ...
    [21814.577533] done.
    [21814.608656] Freezing user space processes ...

    系统在进入休眠状态,代码如下:

    int freeze_processes(void)
    {
        ...
        printk("Freezing user space processes ... ");
        pm_freezing = true;
        oom_kills_saved = oom_kills_count();
        error = try_to_freeze_tasks(true);
        ...
        return error;
    }
    
    static int try_to_freeze_tasks(bool user_only)
    {
        ...
        while (true) {
            todo = 0;
            read_lock(&tasklist_lock);
            for_each_process_thread(g, p) {
                if (p == current || !freeze_task(p))
                    continue;
    
                if (!freezer_should_skip(p))
                    todo++;
            }
            ...
        }
        ...
        return todo ? -EBUSY : 0;
    }
    
    bool freeze_task(struct task_struct *p)
    {
        ...
    
        if (!(p->flags & PF_KTHREAD))
            fake_signal_wake_up(p);
        else
            wake_up_state(p, TASK_INTERRUPTIBLE);
    
        spin_unlock_irqrestore(&freezer_lock, flags);
        return true;
    }
    
    static void fake_signal_wake_up(struct task_struct *p)
    {
        unsigned long flags;
    
        if (lock_task_sighand(p, &flags)) {
            signal_wake_up(p, 0);
            unlock_task_sighand(p, &flags);
        }
    }
    
    static inline void signal_wake_up(struct task_struct *t, bool resume)
    {
        signal_wake_up_state(t, resume ? TASK_WAKEKILL : 0);
    }
    
    void signal_wake_up_state(struct task_struct *t, unsigned int state)
    {
        set_tsk_thread_flag(t, TIF_SIGPENDING);
    
        if (!wake_up_state(t, state | TASK_INTERRUPTIBLE))
            kick_process(t);
    }

    原来就是被freeze流程给打断了。

    所以wait_answer_interruptible()确实可以被freeze打断。

    【解决方案】

            block_sigs(&oldset);
            wait_answer_interruptible(fc, req);
            restore_sigs(&oldset);

    改成

            wait_answer_killable(fc, req);

    其中:

    @fs/fuse/dev.c
    static void wait_answer_killable(struct fuse_conn *fc, struct fuse_req *req)
    __releases(fc->lock)
    __acquires(fc->lock)
    {
            if (fatal_signal_pending(current))
                    return;
    
            spin_unlock(&fc->lock);
            wait_event_killable(req->waitq, req->state == FUSE_REQ_FINISHED);
            spin_lock(&fc->lock);
    }
    
    @include/linux/wait.h
    #define __wait_event_killable_exclusive(wq, condition, ret)            
    do {                                                                   
           DEFINE_WAIT(__wait);                                            
                                                                           
           for (;;) {                                                      
                   prepare_to_wait_exclusive(&wq, &__wait, TASK_KILLABLE); 
                   if (condition)                                          
                           break;                                          
                   if (!fatal_signal_pending(current)) {                   
                           schedule();                                     
                           continue;                                       
                   }                                                       
                   ret = -ERESTARTSYS;                                     
                   break;                                                  
           }                                                               
           finish_wait(&wq, &__wait);                                      
    } while (0)
    
    
    #define wait_event_killable_exclusive(wq, condition)                   
    ({                                                                     
           int __ret = 0;                                                  
           if (!(condition))                                               
                   __wait_event_killable_exclusive(wq, condition, __ret);  
           __ret;                                                          
    })
  • 相关阅读:
    PHP 5 echo 和 print 语句
    MySQL存储过程-遍历游标的例子
    bzoj2554: Color
    win10 uwp 入门
    win10 uwp 入门
    win10 uwp 自定义控件 SplitViewItem
    win10 uwp 自定义控件 SplitViewItem
    win10 uwp ContentDialog 点确定不关闭
    win10 uwp ContentDialog 点确定不关闭
    win10 uwp smms图床
  • 原文地址:https://www.cnblogs.com/YYPapa/p/6862956.html
Copyright © 2011-2022 走看看