|
a |
|
b/docs/_modules/pathflowai/schedulers.html |
|
|
1 |
|
|
|
2 |
|
|
|
3 |
<!DOCTYPE html> |
|
|
4 |
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]--> |
|
|
5 |
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]--> |
|
|
6 |
<head> |
|
|
7 |
<meta charset="utf-8"> |
|
|
8 |
|
|
|
9 |
<meta name="viewport" content="width=device-width, initial-scale=1.0"> |
|
|
10 |
|
|
|
11 |
<title>pathflowai.schedulers — PathFlowAI 0.1 documentation</title> |
|
|
12 |
|
|
|
13 |
|
|
|
14 |
|
|
|
15 |
|
|
|
16 |
|
|
|
17 |
|
|
|
18 |
|
|
|
19 |
|
|
|
20 |
<script type="text/javascript" src="../../_static/js/modernizr.min.js"></script> |
|
|
21 |
|
|
|
22 |
|
|
|
23 |
<script type="text/javascript" id="documentation_options" data-url_root="../../" src="../../_static/documentation_options.js"></script> |
|
|
24 |
<script type="text/javascript" src="../../_static/jquery.js"></script> |
|
|
25 |
<script type="text/javascript" src="../../_static/underscore.js"></script> |
|
|
26 |
<script type="text/javascript" src="../../_static/doctools.js"></script> |
|
|
27 |
<script type="text/javascript" src="../../_static/language_data.js"></script> |
|
|
28 |
<script async="async" type="text/javascript" src="https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.5/latest.js?config=TeX-AMS-MML_HTMLorMML"></script> |
|
|
29 |
|
|
|
30 |
<script type="text/javascript" src="../../_static/js/theme.js"></script> |
|
|
31 |
|
|
|
32 |
|
|
|
33 |
|
|
|
34 |
|
|
|
35 |
<link rel="stylesheet" href="../../_static/css/theme.css" type="text/css" /> |
|
|
36 |
<link rel="stylesheet" href="../../_static/pygments.css" type="text/css" /> |
|
|
37 |
<link rel="index" title="Index" href="../../genindex.html" /> |
|
|
38 |
<link rel="search" title="Search" href="../../search.html" /> |
|
|
39 |
</head> |
|
|
40 |
|
|
|
41 |
<body class="wy-body-for-nav"> |
|
|
42 |
|
|
|
43 |
|
|
|
44 |
<div class="wy-grid-for-nav"> |
|
|
45 |
|
|
|
46 |
<nav data-toggle="wy-nav-shift" class="wy-nav-side"> |
|
|
47 |
<div class="wy-side-scroll"> |
|
|
48 |
<div class="wy-side-nav-search" > |
|
|
49 |
|
|
|
50 |
|
|
|
51 |
|
|
|
52 |
<a href="../../index.html" class="icon icon-home"> PathFlowAI |
|
|
53 |
|
|
|
54 |
|
|
|
55 |
|
|
|
56 |
</a> |
|
|
57 |
|
|
|
58 |
|
|
|
59 |
|
|
|
60 |
|
|
|
61 |
|
|
|
62 |
|
|
|
63 |
|
|
|
64 |
<div role="search"> |
|
|
65 |
<form id="rtd-search-form" class="wy-form" action="../../search.html" method="get"> |
|
|
66 |
<input type="text" name="q" placeholder="Search docs" /> |
|
|
67 |
<input type="hidden" name="check_keywords" value="yes" /> |
|
|
68 |
<input type="hidden" name="area" value="default" /> |
|
|
69 |
</form> |
|
|
70 |
</div> |
|
|
71 |
|
|
|
72 |
|
|
|
73 |
</div> |
|
|
74 |
|
|
|
75 |
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation"> |
|
|
76 |
|
|
|
77 |
|
|
|
78 |
|
|
|
79 |
|
|
|
80 |
|
|
|
81 |
|
|
|
82 |
<!-- Local TOC --> |
|
|
83 |
<div class="local-toc"></div> |
|
|
84 |
|
|
|
85 |
|
|
|
86 |
</div> |
|
|
87 |
</div> |
|
|
88 |
</nav> |
|
|
89 |
|
|
|
90 |
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap"> |
|
|
91 |
|
|
|
92 |
|
|
|
93 |
<nav class="wy-nav-top" aria-label="top navigation"> |
|
|
94 |
|
|
|
95 |
<i data-toggle="wy-nav-top" class="fa fa-bars"></i> |
|
|
96 |
<a href="../../index.html">PathFlowAI</a> |
|
|
97 |
|
|
|
98 |
</nav> |
|
|
99 |
|
|
|
100 |
|
|
|
101 |
<div class="wy-nav-content"> |
|
|
102 |
|
|
|
103 |
<div class="rst-content"> |
|
|
104 |
|
|
|
105 |
|
|
|
106 |
|
|
|
107 |
|
|
|
108 |
|
|
|
109 |
|
|
|
110 |
|
|
|
111 |
|
|
|
112 |
|
|
|
113 |
|
|
|
114 |
|
|
|
115 |
|
|
|
116 |
|
|
|
117 |
|
|
|
118 |
|
|
|
119 |
|
|
|
120 |
|
|
|
121 |
<div role="navigation" aria-label="breadcrumbs navigation"> |
|
|
122 |
|
|
|
123 |
<ul class="wy-breadcrumbs"> |
|
|
124 |
|
|
|
125 |
<li><a href="../../index.html">Docs</a> »</li> |
|
|
126 |
|
|
|
127 |
<li><a href="../index.html">Module code</a> »</li> |
|
|
128 |
|
|
|
129 |
<li>pathflowai.schedulers</li> |
|
|
130 |
|
|
|
131 |
|
|
|
132 |
<li class="wy-breadcrumbs-aside"> |
|
|
133 |
|
|
|
134 |
</li> |
|
|
135 |
|
|
|
136 |
</ul> |
|
|
137 |
|
|
|
138 |
|
|
|
139 |
<hr/> |
|
|
140 |
</div> |
|
|
141 |
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article"> |
|
|
142 |
<div itemprop="articleBody"> |
|
|
143 |
|
|
|
144 |
<h1>Source code for pathflowai.schedulers</h1><div class="highlight"><pre> |
|
|
145 |
<span></span><span class="sd">"""</span> |
|
|
146 |
<span class="sd">schedulers.py</span> |
|
|
147 |
<span class="sd">=======================</span> |
|
|
148 |
<span class="sd">Modulates the learning rate during the training process.</span> |
|
|
149 |
<span class="sd">"""</span> |
|
|
150 |
<span class="kn">import</span> <span class="nn">torch</span> |
|
|
151 |
<span class="kn">import</span> <span class="nn">math</span> |
|
|
152 |
<span class="kn">from</span> <span class="nn">torch.optim.lr_scheduler</span> <span class="k">import</span> <span class="n">ExponentialLR</span> |
|
|
153 |
|
|
|
154 |
<div class="viewcode-block" id="CosineAnnealingWithRestartsLR"><a class="viewcode-back" href="../../index.html#pathflowai.schedulers.CosineAnnealingWithRestartsLR">[docs]</a><span class="k">class</span> <span class="nc">CosineAnnealingWithRestartsLR</span><span class="p">(</span><span class="n">torch</span><span class="o">.</span><span class="n">optim</span><span class="o">.</span><span class="n">lr_scheduler</span><span class="o">.</span><span class="n">_LRScheduler</span><span class="p">):</span> |
|
|
155 |
<span class="sa">r</span><span class="sd">"""Set the learning rate of each parameter group using a cosine annealing</span> |
|
|
156 |
<span class="sd"> schedule, where :math:`\eta_{max}` is set to the initial lr and</span> |
|
|
157 |
<span class="sd"> :math:`T_{cur}` is the number of epochs since the last restart in SGDR:</span> |
|
|
158 |
<span class="sd"> .. math::</span> |
|
|
159 |
<span class="sd"> \eta_t = \eta_{min} + \frac{1}{2}(\eta_{max} - \eta_{min})(1 +</span> |
|
|
160 |
<span class="sd"> \cos(\frac{T_{cur}}{T_{max}}\pi))</span> |
|
|
161 |
<span class="sd"> When last_epoch=-1, sets initial lr as lr.</span> |
|
|
162 |
<span class="sd"> It has been proposed in</span> |
|
|
163 |
<span class="sd"> `SGDR: Stochastic Gradient Descent with Warm Restarts`_. This implements</span> |
|
|
164 |
<span class="sd"> the cosine annealing part of SGDR, the restarts and number of iterations multiplier.</span> |
|
|
165 |
<span class="sd"> Args:</span> |
|
|
166 |
<span class="sd"> optimizer (Optimizer): Wrapped optimizer.</span> |
|
|
167 |
<span class="sd"> T_max (int): Maximum number of iterations.</span> |
|
|
168 |
<span class="sd"> T_mult (float): Multiply T_max by this number after each restart. Default: 1.</span> |
|
|
169 |
<span class="sd"> eta_min (float): Minimum learning rate. Default: 0.</span> |
|
|
170 |
<span class="sd"> last_epoch (int): The index of last epoch. Default: -1.</span> |
|
|
171 |
<span class="sd"> .. _SGDR\: Stochastic Gradient Descent with Warm Restarts:</span> |
|
|
172 |
<span class="sd"> https://arxiv.org/abs/1608.03983</span> |
|
|
173 |
<span class="sd"> """</span> |
|
|
174 |
<span class="k">def</span> <span class="nf">__init__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">optimizer</span><span class="p">,</span> <span class="n">T_max</span><span class="p">,</span> <span class="n">eta_min</span><span class="o">=</span><span class="mi">0</span><span class="p">,</span> <span class="n">last_epoch</span><span class="o">=-</span><span class="mi">1</span><span class="p">,</span> <span class="n">T_mult</span><span class="o">=</span><span class="mf">1.</span><span class="p">,</span> <span class="n">alpha_decay</span><span class="o">=</span><span class="mf">1.0</span><span class="p">):</span> |
|
|
175 |
<span class="bp">self</span><span class="o">.</span><span class="n">T_max</span> <span class="o">=</span> <span class="n">T_max</span> |
|
|
176 |
<span class="bp">self</span><span class="o">.</span><span class="n">T_mult</span> <span class="o">=</span> <span class="n">T_mult</span> |
|
|
177 |
<span class="bp">self</span><span class="o">.</span><span class="n">restart_every</span> <span class="o">=</span> <span class="n">T_max</span> |
|
|
178 |
<span class="bp">self</span><span class="o">.</span><span class="n">eta_min</span> <span class="o">=</span> <span class="n">eta_min</span> |
|
|
179 |
<span class="bp">self</span><span class="o">.</span><span class="n">restarts</span> <span class="o">=</span> <span class="mi">0</span> |
|
|
180 |
<span class="bp">self</span><span class="o">.</span><span class="n">restarted_at</span> <span class="o">=</span> <span class="mi">0</span> |
|
|
181 |
<span class="bp">self</span><span class="o">.</span><span class="n">alpha</span> <span class="o">=</span> <span class="n">alpha_decay</span> |
|
|
182 |
<span class="nb">super</span><span class="p">()</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="n">optimizer</span><span class="p">,</span> <span class="n">last_epoch</span><span class="p">)</span> |
|
|
183 |
|
|
|
184 |
<span class="k">def</span> <span class="nf">restart</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span> |
|
|
185 |
<span class="bp">self</span><span class="o">.</span><span class="n">restarts</span> <span class="o">+=</span> <span class="mi">1</span> |
|
|
186 |
<span class="bp">self</span><span class="o">.</span><span class="n">restart_every</span> <span class="o">=</span> <span class="nb">int</span><span class="p">(</span><span class="nb">round</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">restart_every</span> <span class="o">*</span> <span class="bp">self</span><span class="o">.</span><span class="n">T_mult</span><span class="p">))</span> |
|
|
187 |
<span class="bp">self</span><span class="o">.</span><span class="n">restarted_at</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">last_epoch</span> |
|
|
188 |
|
|
|
189 |
<span class="k">def</span> <span class="nf">cosine</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">base_lr</span><span class="p">):</span> |
|
|
190 |
<span class="k">return</span> <span class="bp">self</span><span class="o">.</span><span class="n">eta_min</span> <span class="o">+</span> <span class="bp">self</span><span class="o">.</span><span class="n">alpha</span><span class="o">**</span><span class="bp">self</span><span class="o">.</span><span class="n">restarts</span> <span class="o">*</span> <span class="p">(</span><span class="n">base_lr</span> <span class="o">-</span> <span class="bp">self</span><span class="o">.</span><span class="n">eta_min</span><span class="p">)</span> <span class="o">*</span> <span class="p">(</span><span class="mi">1</span> <span class="o">+</span> <span class="n">math</span><span class="o">.</span><span class="n">cos</span><span class="p">(</span><span class="n">math</span><span class="o">.</span><span class="n">pi</span> <span class="o">*</span> <span class="bp">self</span><span class="o">.</span><span class="n">step_n</span> <span class="o">/</span> <span class="bp">self</span><span class="o">.</span><span class="n">restart_every</span><span class="p">))</span> <span class="o">/</span> <span class="mi">2</span> |
|
|
191 |
|
|
|
192 |
<span class="nd">@property</span> |
|
|
193 |
<span class="k">def</span> <span class="nf">step_n</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span> |
|
|
194 |
<span class="k">return</span> <span class="bp">self</span><span class="o">.</span><span class="n">last_epoch</span> <span class="o">-</span> <span class="bp">self</span><span class="o">.</span><span class="n">restarted_at</span> |
|
|
195 |
|
|
|
196 |
<span class="k">def</span> <span class="nf">get_lr</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span> |
|
|
197 |
<span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">step_n</span> <span class="o">>=</span> <span class="bp">self</span><span class="o">.</span><span class="n">restart_every</span><span class="p">:</span> |
|
|
198 |
<span class="bp">self</span><span class="o">.</span><span class="n">restart</span><span class="p">()</span> |
|
|
199 |
<span class="k">return</span> <span class="p">[</span><span class="bp">self</span><span class="o">.</span><span class="n">cosine</span><span class="p">(</span><span class="n">base_lr</span><span class="p">)</span> <span class="k">for</span> <span class="n">base_lr</span> <span class="ow">in</span> <span class="bp">self</span><span class="o">.</span><span class="n">base_lrs</span><span class="p">]</span></div> |
|
|
200 |
|
|
|
201 |
<div class="viewcode-block" id="Scheduler"><a class="viewcode-back" href="../../index.html#pathflowai.schedulers.Scheduler">[docs]</a><span class="k">class</span> <span class="nc">Scheduler</span><span class="p">:</span> |
|
|
202 |
<span class="sd">"""Scheduler class that modulates learning rate of torch optimizers over epochs.</span> |
|
|
203 |
|
|
|
204 |
<span class="sd"> Parameters</span> |
|
|
205 |
<span class="sd"> ----------</span> |
|
|
206 |
<span class="sd"> optimizer : type</span> |
|
|
207 |
<span class="sd"> torch.Optimizer object</span> |
|
|
208 |
<span class="sd"> opts : type</span> |
|
|
209 |
<span class="sd"> Options of setting the learning rate scheduler, see default.</span> |
|
|
210 |
|
|
|
211 |
<span class="sd"> Attributes</span> |
|
|
212 |
<span class="sd"> ----------</span> |
|
|
213 |
<span class="sd"> schedulers : type</span> |
|
|
214 |
<span class="sd"> Different types of schedulers to choose from.</span> |
|
|
215 |
<span class="sd"> scheduler_step_fn : type</span> |
|
|
216 |
<span class="sd"> How scheduler updates learning rate.</span> |
|
|
217 |
<span class="sd"> initial_lr : type</span> |
|
|
218 |
<span class="sd"> Initial set learning rate.</span> |
|
|
219 |
<span class="sd"> scheduler_choice : type</span> |
|
|
220 |
<span class="sd"> What scheduler type was chosen.</span> |
|
|
221 |
<span class="sd"> scheduler : type</span> |
|
|
222 |
<span class="sd"> Scheduler object chosen that will more directly update optimizer LR.</span> |
|
|
223 |
|
|
|
224 |
<span class="sd"> """</span> |
|
|
225 |
<span class="k">def</span> <span class="nf">__init__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">optimizer</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span> <span class="n">opts</span><span class="o">=</span><span class="nb">dict</span><span class="p">(</span><span class="n">scheduler</span><span class="o">=</span><span class="s1">'null'</span><span class="p">,</span><span class="n">lr_scheduler_decay</span><span class="o">=</span><span class="mf">0.5</span><span class="p">,</span><span class="n">T_max</span><span class="o">=</span><span class="mi">10</span><span class="p">,</span><span class="n">eta_min</span><span class="o">=</span><span class="mf">5e-8</span><span class="p">,</span><span class="n">T_mult</span><span class="o">=</span><span class="mi">2</span><span class="p">)):</span> |
|
|
226 |
<span class="bp">self</span><span class="o">.</span><span class="n">schedulers</span> <span class="o">=</span> <span class="p">{</span><span class="s1">'exp'</span><span class="p">:(</span><span class="k">lambda</span> <span class="n">optimizer</span><span class="p">:</span> <span class="n">ExponentialLR</span><span class="p">(</span><span class="n">optimizer</span><span class="p">,</span> <span class="n">opts</span><span class="p">[</span><span class="s2">"lr_scheduler_decay"</span><span class="p">])),</span> |
|
|
227 |
<span class="s1">'null'</span><span class="p">:(</span><span class="k">lambda</span> <span class="n">optimizer</span><span class="p">:</span> <span class="kc">None</span><span class="p">),</span> |
|
|
228 |
<span class="s1">'warm_restarts'</span><span class="p">:(</span><span class="k">lambda</span> <span class="n">optimizer</span><span class="p">:</span> <span class="n">CosineAnnealingWithRestartsLR</span><span class="p">(</span><span class="n">optimizer</span><span class="p">,</span> <span class="n">T_max</span><span class="o">=</span><span class="n">opts</span><span class="p">[</span><span class="s1">'T_max'</span><span class="p">],</span> <span class="n">eta_min</span><span class="o">=</span><span class="n">opts</span><span class="p">[</span><span class="s1">'eta_min'</span><span class="p">],</span> <span class="n">last_epoch</span><span class="o">=-</span><span class="mi">1</span><span class="p">,</span> <span class="n">T_mult</span><span class="o">=</span><span class="n">opts</span><span class="p">[</span><span class="s1">'T_mult'</span><span class="p">]))}</span> |
|
|
229 |
<span class="bp">self</span><span class="o">.</span><span class="n">scheduler_step_fn</span> <span class="o">=</span> <span class="p">{</span><span class="s1">'exp'</span><span class="p">:(</span><span class="k">lambda</span> <span class="n">scheduler</span><span class="p">:</span> <span class="n">scheduler</span><span class="o">.</span><span class="n">step</span><span class="p">()),</span> |
|
|
230 |
<span class="s1">'warm_restarts'</span><span class="p">:(</span><span class="k">lambda</span> <span class="n">scheduler</span><span class="p">:</span> <span class="n">scheduler</span><span class="o">.</span><span class="n">step</span><span class="p">()),</span> |
|
|
231 |
<span class="s1">'null'</span><span class="p">:(</span><span class="k">lambda</span> <span class="n">scheduler</span><span class="p">:</span> <span class="kc">None</span><span class="p">)}</span> |
|
|
232 |
<span class="bp">self</span><span class="o">.</span><span class="n">initial_lr</span> <span class="o">=</span> <span class="n">optimizer</span><span class="o">.</span><span class="n">param_groups</span><span class="p">[</span><span class="mi">0</span><span class="p">][</span><span class="s1">'lr'</span><span class="p">]</span> |
|
|
233 |
<span class="bp">self</span><span class="o">.</span><span class="n">scheduler_choice</span> <span class="o">=</span> <span class="n">opts</span><span class="p">[</span><span class="s1">'scheduler'</span><span class="p">]</span> |
|
|
234 |
<span class="bp">self</span><span class="o">.</span><span class="n">scheduler</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">schedulers</span><span class="p">[</span><span class="bp">self</span><span class="o">.</span><span class="n">scheduler_choice</span><span class="p">](</span><span class="n">optimizer</span><span class="p">)</span> <span class="k">if</span> <span class="n">optimizer</span> <span class="ow">is</span> <span class="ow">not</span> <span class="kc">None</span> <span class="k">else</span> <span class="kc">None</span> |
|
|
235 |
|
|
|
236 |
<div class="viewcode-block" id="Scheduler.step"><a class="viewcode-back" href="../../index.html#pathflowai.schedulers.Scheduler.step">[docs]</a> <span class="k">def</span> <span class="nf">step</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span> |
|
|
237 |
<span class="sd">"""Update optimizer learning rate"""</span> |
|
|
238 |
<span class="bp">self</span><span class="o">.</span><span class="n">scheduler_step_fn</span><span class="p">[</span><span class="bp">self</span><span class="o">.</span><span class="n">scheduler_choice</span><span class="p">](</span><span class="bp">self</span><span class="o">.</span><span class="n">scheduler</span><span class="p">)</span></div> |
|
|
239 |
|
|
|
240 |
<div class="viewcode-block" id="Scheduler.get_lr"><a class="viewcode-back" href="../../index.html#pathflowai.schedulers.Scheduler.get_lr">[docs]</a> <span class="k">def</span> <span class="nf">get_lr</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span> |
|
|
241 |
<span class="sd">"""Return current learning rate.</span> |
|
|
242 |
|
|
|
243 |
<span class="sd"> Returns</span> |
|
|
244 |
<span class="sd"> -------</span> |
|
|
245 |
<span class="sd"> float</span> |
|
|
246 |
<span class="sd"> Current learning rate.</span> |
|
|
247 |
|
|
|
248 |
<span class="sd"> """</span> |
|
|
249 |
<span class="n">lr</span> <span class="o">=</span> <span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">initial_lr</span> <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">scheduler_choice</span> <span class="o">==</span> <span class="s1">'null'</span> <span class="k">else</span> <span class="bp">self</span><span class="o">.</span><span class="n">scheduler</span><span class="o">.</span><span class="n">optimizer</span><span class="o">.</span><span class="n">param_groups</span><span class="p">[</span><span class="mi">0</span><span class="p">][</span><span class="s1">'lr'</span><span class="p">])</span> |
|
|
250 |
<span class="k">return</span> <span class="n">lr</span></div></div> |
|
|
251 |
</pre></div> |
|
|
252 |
|
|
|
253 |
</div> |
|
|
254 |
|
|
|
255 |
</div> |
|
|
256 |
<footer> |
|
|
257 |
|
|
|
258 |
|
|
|
259 |
<hr/> |
|
|
260 |
|
|
|
261 |
<div role="contentinfo"> |
|
|
262 |
<p> |
|
|
263 |
© Copyright 2019, Joshua Levy |
|
|
264 |
|
|
|
265 |
</p> |
|
|
266 |
</div> |
|
|
267 |
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/rtfd/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>. |
|
|
268 |
|
|
|
269 |
</footer> |
|
|
270 |
|
|
|
271 |
</div> |
|
|
272 |
</div> |
|
|
273 |
|
|
|
274 |
</section> |
|
|
275 |
|
|
|
276 |
</div> |
|
|
277 |
|
|
|
278 |
|
|
|
279 |
|
|
|
280 |
<script type="text/javascript"> |
|
|
281 |
jQuery(function () { |
|
|
282 |
SphinxRtdTheme.Navigation.enable(true); |
|
|
283 |
}); |
|
|
284 |
</script> |
|
|
285 |
|
|
|
286 |
|
|
|
287 |
|
|
|
288 |
|
|
|
289 |
|
|
|
290 |
|
|
|
291 |
</body> |
|
|
292 |
</html> |