2018-10-10 22:03:03 +00:00
|
|
|
|
---
|
|
|
|
|
id: 594faaab4e2a8626833e9c3d
|
2020-12-16 07:37:30 +00:00
|
|
|
|
title: 使用转义标记字符串
|
2018-10-10 22:03:03 +00:00
|
|
|
|
challengeType: 5
|
|
|
|
|
videoUrl: ''
|
2021-01-13 02:31:00 +00:00
|
|
|
|
dashedName: tokenize-a-string-with-escaping
|
2018-10-10 22:03:03 +00:00
|
|
|
|
---
|
|
|
|
|
|
2020-12-16 07:37:30 +00:00
|
|
|
|
# --description--
|
2018-10-10 22:03:03 +00:00
|
|
|
|
|
2020-12-16 07:37:30 +00:00
|
|
|
|
<p>编写一个函数或程序,可以在分隔符的每个非转义事件中拆分字符串。 </p><p>它应该接受三个输入参数: </p> <b>字符串</b> <b>分隔符字符</b> <b>转义字符</b> <p>它应该输出一个字符串列表。 </p><p>拆分规则: </p>由分隔符分隔的字段将成为输出列表的元素。应保留空字段,即使在开始和结束时也是如此。 <p>转义规则: </p> “Escaped”意味着出现一个尚未自行转义的转义字符。当转义字符位于没有特殊含义的字符之前时,它仍然被视为转义符(但不会做任何特殊操作)。用于转义某些内容的每次出现的转义字符都不应成为输出的一部分。 <p>证明您的函数满足以下测试用例:给定字符串</p><pre>一个^ | UNO || 3 ^^^^ |四^^^ | ^夸| </pre>和使用<pre> | </pre>作为分隔符和<pre> ^ </pre>作为转义字符,您的函数应输出以下数组: <p></p><pre> ['one | uno',“,'three ^^','four ^ | quatro',”]
|
|
|
|
|
</pre>
|
2018-10-10 22:03:03 +00:00
|
|
|
|
|
2020-12-16 07:37:30 +00:00
|
|
|
|
# --hints--
|
2018-10-10 22:03:03 +00:00
|
|
|
|
|
2020-12-16 07:37:30 +00:00
|
|
|
|
`tokenize`是一个函数。
|
2018-10-10 22:03:03 +00:00
|
|
|
|
|
2020-12-16 07:37:30 +00:00
|
|
|
|
```js
|
|
|
|
|
assert(typeof tokenize === 'function');
|
2018-10-10 22:03:03 +00:00
|
|
|
|
```
|
|
|
|
|
|
2020-12-16 07:37:30 +00:00
|
|
|
|
`tokenize`应该返回一个数组。
|
2018-10-10 22:03:03 +00:00
|
|
|
|
|
|
|
|
|
```js
|
2020-12-16 07:37:30 +00:00
|
|
|
|
assert(typeof tokenize('a', 'b', 'c') === 'object');
|
2018-10-10 22:03:03 +00:00
|
|
|
|
```
|
|
|
|
|
|
2020-12-16 07:37:30 +00:00
|
|
|
|
`tokenize("one^|uno||three^^^^|four^^^|^cuatro|", "|", "^")`应返回[“one | uno”,“”,“three ^^” ,“四个^ | cuatro”,“”]“)
|
2018-10-10 22:03:03 +00:00
|
|
|
|
|
|
|
|
|
```js
|
2020-12-16 07:37:30 +00:00
|
|
|
|
assert.deepEqual(tokenize(testStr1, '|', '^'), res1);
|
2018-10-10 22:03:03 +00:00
|
|
|
|
```
|
|
|
|
|
|
2020-12-16 07:37:30 +00:00
|
|
|
|
`tokenize("a@&bcd&ef&&@@hi", "&", "@")`应返回`["a&bcd", "ef", "", "@hi"]`
|
2018-10-10 22:03:03 +00:00
|
|
|
|
|
|
|
|
|
```js
|
2020-12-16 07:37:30 +00:00
|
|
|
|
assert.deepEqual(tokenize(testStr2, '&', '@'), res2);
|
2018-10-10 22:03:03 +00:00
|
|
|
|
```
|
2020-08-13 15:24:35 +00:00
|
|
|
|
|
2021-01-13 02:31:00 +00:00
|
|
|
|
# --seed--
|
|
|
|
|
|
|
|
|
|
## --after-user-code--
|
|
|
|
|
|
|
|
|
|
```js
|
|
|
|
|
const testStr1 = 'one^|uno||three^^^^|four^^^|^cuatro|';
|
|
|
|
|
const res1 = ['one|uno', '', 'three^^', 'four^|cuatro', ''];
|
|
|
|
|
|
|
|
|
|
// TODO add more tests
|
|
|
|
|
const testStr2 = 'a@&bcd&ef&&@@hi';
|
|
|
|
|
const res2 = ['a&bcd', 'ef', '', '@hi'];
|
|
|
|
|
```
|
|
|
|
|
|
|
|
|
|
## --seed-contents--
|
|
|
|
|
|
|
|
|
|
```js
|
|
|
|
|
function tokenize(str, sep, esc) {
|
|
|
|
|
return true;
|
|
|
|
|
}
|
|
|
|
|
```
|
|
|
|
|
|
2020-12-16 07:37:30 +00:00
|
|
|
|
# --solutions--
|
|
|
|
|
|
2021-01-13 02:31:00 +00:00
|
|
|
|
```js
|
|
|
|
|
// tokenize :: String -> Character -> Character -> [String]
|
|
|
|
|
function tokenize(str, charDelim, charEsc) {
|
|
|
|
|
const dctParse = str.split('')
|
|
|
|
|
.reduce((a, x) => {
|
|
|
|
|
const blnEsc = a.esc;
|
|
|
|
|
const blnBreak = !blnEsc && x === charDelim;
|
|
|
|
|
const blnEscChar = !blnEsc && x === charEsc;
|
|
|
|
|
|
|
|
|
|
return {
|
|
|
|
|
esc: blnEscChar,
|
|
|
|
|
token: blnBreak ? '' : (
|
|
|
|
|
a.token + (blnEscChar ? '' : x)
|
|
|
|
|
),
|
|
|
|
|
list: a.list.concat(blnBreak ? a.token : [])
|
|
|
|
|
};
|
|
|
|
|
}, {
|
|
|
|
|
esc: false,
|
|
|
|
|
token: '',
|
|
|
|
|
list: []
|
|
|
|
|
});
|
|
|
|
|
|
|
|
|
|
return dctParse.list.concat(
|
|
|
|
|
dctParse.token
|
|
|
|
|
);
|
|
|
|
|
}
|
|
|
|
|
```
|