-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Referee reports # 1, CMAME Sep. 2016 (decision: reject) #19
Comments
comment # 2: |
comment # 3: |
comment # 4:
The last iteration before convergence can make εk as large as 1. The residual of the last iteration rk should be smaller than the desired tolerance η, thus the first element in the min function is larger than 1, so εk is 1, which means the allowed pertubation approaches its maximum (εk = 1) at convergence. |
comment # 5: |
comment # 6: |
comment # 7:
The FMM error bound in eqn.18 is a function of the order of expansion p. Since θMAC = 0.5, we replace a/r with 0.5 (for the worst case) in the eqn.18. By equating the FMM error bound with the allowed perturbation εk at k-th iteration in eqn.17, we can obtain the relationship between the required pk and εk at k-th iteration as shown in eqn.19. εk can be calculated by using eqn.17 given the residual at the previous step rk-1 and the desired tolerance η. |
comment # 8:
I agree it is always better to have both qualitative and quantitative results. The fitted exponent (or the fitted slope in this log-log plot) is 1.06, showing a linear scaling with respect to N. |
comment # 9:
The acceleration compared with the traditional BEM is twofold, and it comes from: 1). using FMM-accelerated mat-vec and 2). introducing relaxation in GMRES (using a decreasing p in FMM). The first improvement is very common in Fast BEM applications (see Nishimura's review article on fast multipole accelerated boundary integral equation methods). They showed that using FMM-accelerated mat-vec would not blow up the number of iteration, and that is why we did not compare with the exact mat-vec. Moreover, our main contribution is to relax further (gradually decrease p) in GMRES solver. Therefore, in the result section, we measure the speedups based on FMM-accelerated BEM without relaxation, rather than vanilla BEM. |
comment # 10:
The fixed-p residual history curve is very close to the relaxed-p curve (the solid line on figure 6), since relaxation will not greatly affect the number of iterations to converge. Table 1 and Table 2 show that both cases take the same number of iterations to converge. |
comment # 11:
We first clarify that the near-field and far-field loads are unbalanced in each iteration, but well-balanced overall. The reason for not adding the time breakdown plot in section 3.1 is that the trend is very similar to Figure 12 in section 3.2. Since we follow the same mechanism to reduce p for each case, we will always get a bloated far-field in the first several iterations and a smaller far-field in the later iterations. Therefore, we only show this trend once to avoid redundancy. For the same reason, we did not add a residual history & required-p plot (like Figure 6) in StokesBEM section. |
comment # 12:
The schedule for varying p comes from eqn.17 ~ eqn.19 (see comment # 7 above). For the relaxed cases in a Stokes flow problem, there are two p s to distinguish between: pinitial (the high p used in the first iteration) and pmin (the minimum value of p allowed in the relaxed solver). As the iteration proceeds, the required-p for each iteration decreases from pinitial to pmin. Table 5 is a parameter study to determine pmin. In Figure 11, we use pinitial=16 and pmin=5. |
comment # 13:
First, this test is aimed to prove that the relaxation will always offer a decent speedup no matter what the user-defined tolerance is. Second, from Figure 14 we do see that the speedup due to relaxation is greater at the middle tolerances than at the extreme tolerances. The desired tolerance η's effect on the speedup is tricky and indirect. η first determines allowed perturbation (eqn.17), then effects the required-p at each iteration. The combination of these required-p s together determines the optimal ncrit, resulting in the best possible runtime. Therefore, I would say that the shape of this curve is problem-dependent. |
comment # 14
The last "peak" number of iterations at N=8192 is 39. The two cases that follow with larger N indeed take fewer iterations, but not by much: 35 and 31, respectively. We don't know "why," but we don't think it is important. In practice, one should probably use tighter solver tolerances when refining the surface mesh. However, for presentation purposes, we generally want to change one parameter at a time. |
Revised paper submitted to CMAME on 15 March 2016 — Decision: Reject
Editor comments
Referee # 1 comments
Summary of the problem being solved. This paper considers boundary element method (BEM) formulations of certain problems, namely, Laplace problems and Stokes flow problems that arise in blood flow simulation. Computationally, this formulation requires solving a large dense linear system, most typically solved by Krylov subspace-based iterative methods, of which this paper considers GMRES. The usual performance bottleneck of a Krylov subspace solver is a matrix-vector product ("matvec"). In the target applications, the matrix is special: its entries are K(x, y) evaluated at pairs of points (x, y), where K(x, y) decays rapidly with the distance between x and y. In this case, one may implement fast but approximate matrix-vector products. A classical approach for the target problems is the fast multipole method (FMM) of Greengard & Rokhlin (mid-80s). The FMM has the feature that one may control tradeoff performance and accuracy via a parameter, the so-called$p$ and roughly corresponding to the order of truncation in a series expansion. (Higher values of $p$ are more accurate but slower.)
degree of the multipole expansion, denoted by
With this background, the specific problem this paper tackles is to study the interaction between$p$ and the outer GMRES solver.
Technical contribution. The paper develops an integrated GMRES+FMM method for its target problems that exploits a clever and nontrivial observation from the theory of inexact Krylov methods: as the iterations proceed, the matvec can actually become progressively less accurate without harming the convergence rate of the solver. The idea of the paper is to decrease$p$ as the iterations proceed.
From this idea, the paper studies several model problems to show that it can work and characterizes the time and accuracy tradeoffs, with some interesting FMM-specific observations along the way.
Summary evaluation and recommendation. Though there is a lot of "setup," the idea is both clever and simple, meaning it is likely to work and have an impact. The results are a good first "proof-of-concept" of the feasibility of the proposed technique. The suggested edits that appear below are all minor. As such, I recommend the paper be accepted with just minor revisions.
One major strength of the paper, beyond the idea and its demonstration, is the effort the authors made toward making the results reproducible, including an open-source code repository with sufficient infrastructure to recompute the results and even make the plots. Cool!
Detailed comments for the authors. My remaining issues, detailed below, are fairly minor and mostly a matter of editing or questions to consider clarifying in a revision.
The text was updated successfully, but these errors were encountered: