多强调你的方法的特点,与众不同之处,让审稿人一定能看到。
第一句:预热一下 |
---|
We now present our xx framework for xx. |
In this section, we provide a more detailed description of the presented approach, xxx. |
Given xx, we aim to achieve xxx (你文章是干什么的) |
Our goal is to ... |
This section introduces the overall working mechanism and specific technical implementations of the proposed xx. |
如果你的方法由多个部分构成的话 |
---|
Our xx process is split into two parts: a xx branch, which xxx, and a xx branch that xxx. |
章节结构安排 |
---|
In the following, we first introduce xx in Sec. 3.1, before proceeding to xx in Sec 3.2. |
The organization of this section is as follows: We introduce xx in Sec. 3.1, including xx. In Sec. 3.2, we first describe xx. Then we present xx. |
To begin with, we review the basic idea and pipeline of xxx. |
To facilitate understanding, we begin with an overview of our framework xx |
插补一些有用的连词:Thereafter
最后一句:介绍核心的流程图 |
---|
Fig. x gives an overview of the training pipeline. |
Our model is illustrated in Fig. x. |
In essence, our method extends, ..., as demonstrated in Fig. 1 |
Fig.1 depicts xx |
Fig.1 shows the framework of the proposed method. |
schematic illustrating |
比较具体,这里就不再过多介绍。
训练细节 |
---|
We inherit the training procedure from xx with minimal changes. |
The optimization is performed by Adam with learning rate of 0.002 and betas of 0 and 0.99 |
Further details can be found in the source code. |
有的时候会为了简单起见,省略一些符号标记 |
---|
for the sake of simplicity, we |
For ease of notation, we will use latent codes c to denote the concatenation of all latent variables ci. |
对公式的描述 |
---|
where λ is a positive constant trading off the importance of the first and second terms of the loss |
一些不错的说辞 |
---|
we want to avoid investing effort into a component that is not a part of the actual solution. |
just show many examples
$$
f_{j}^{\prime}=\left[\nabla_{\mathbf{x}} f_{j}\left(\mathbf{x}_{0}\right)\right]^{t} \mathbf{d}
$$
where $\nabla_{\mathbf{x}}$ is the symbol for the gradient w.r.t. $\mathbf{x}$ and the superscript $^{t}$ stands for transposition. One seeks for a vector $\mathbf{d}$ such that the scalar product of any objective gradient $\nabla_{x} f_{j}\left(\mathbf{x}{0}\right)$ with the vector d remains strictly positive $f{j}^{\prime} > 0$
we use xx to denote xx