@@ -161,24 +161,24 @@ function docTokenFunction (token) {
161
161
const namespaceRegex = / \: | \. /
162
162
let toSplitWords = token . toString ( )
163
163
164
- if ( arityRegex . test ( toSplitWords ) ) {
164
+ if ( arityRegex . test ( toSplitWords ) ) {
165
165
const withoutArity = token
166
166
. toString ( )
167
- . replace ( arityRegex , "" )
167
+ . replace ( arityRegex , '' )
168
168
169
169
// This token represents ExDoc.Markdown.to_ast
170
170
tokens . push ( token . clone ( ) . update ( ( ) => withoutArity ) )
171
171
172
172
// And now we get each part as token: ExDoc, Markdown, and to_ast
173
- let parts = withoutArity . split ( namespaceRegex )
173
+ const parts = withoutArity . split ( namespaceRegex )
174
174
175
- if ( parts . length > 1 ) {
176
- for ( let part of parts ) {
175
+ if ( parts . length > 1 ) {
176
+ for ( const part of parts ) {
177
177
tokens . push ( token . clone ( ) . update ( ( ) => part ) )
178
178
}
179
179
180
180
// Let's also add to_ast/2
181
- let lastWithArity = token . toString ( ) . split ( namespaceRegex )
181
+ const lastWithArity = token . toString ( ) . split ( namespaceRegex )
182
182
tokens . push ( token . clone ( ) . update ( ( ) => lastWithArity [ lastWithArity . length - 1 ] ) )
183
183
}
184
184
@@ -187,10 +187,10 @@ function docTokenFunction (token) {
187
187
188
188
// Now split the function name (or the token, if that's all we had),
189
189
// on _ or - (but we keep the original)
190
- let words = toSplitWords . split ( / \_ | \- / )
190
+ const words = toSplitWords . split ( / \_ | \- / )
191
191
192
- if ( words . length > 1 ) {
193
- for ( let word of words ) {
192
+ if ( words . length > 1 ) {
193
+ for ( const word of words ) {
194
194
tokens . push ( token . clone ( ) . update ( ( ) => word ) )
195
195
}
196
196
}
0 commit comments