Overview
| Comment: | [fx] lexicographer: count paragraphs analyzed |
|---|---|
| Downloads: | Tarball | ZIP archive | SQL archive |
| Timelines: | family | ancestors | descendants | both | fx | webext2 |
| Files: | files | file ages | folders |
| SHA3-256: |
eef32ad83f8e38eba2e0b95c15e1a52b |
| User & Date: | olr on 2017-08-15 08:36:29 |
| Other Links: | branch diff | manifest | tags |
Context
|
2017-08-15
| ||
| 08:41 | [fx] movable panel height depending on window height check-in: 75f1b73645 user: olr tags: fx, webext2 | |
| 08:36 | [fx] lexicographer: count paragraphs analyzed check-in: eef32ad83f user: olr tags: fx, webext2 | |
| 08:12 | [core][js][py] text.getParagraph(): end of line handling check-in: ec919db910 user: olr tags: core, webext2 | |
Changes
Modified gc_lang/fr/webext/content_scripts/content_panels.css from [8844c08781] to [87d3446a4a].
| ︙ | ︙ | |||
113 114 115 116 117 118 119 |
.grammalecte_lxg_list_of_tokens {
margin: 10px 0;
padding: 10px;
background-color: hsla(0, 0%, 96%, 1);
border-radius: 2px;
}
| > > > > > > > > | | 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 |
.grammalecte_lxg_list_of_tokens {
margin: 10px 0;
padding: 10px;
background-color: hsla(0, 0%, 96%, 1);
border-radius: 2px;
}
.grammalecte_lxg_list_of_tokens .num {
float: right;
margin: -12px 0 5px 10px;
padding: 5px 10px;
font-weight: bold;
border-radius: 0 0 4px 4px;
background-color: hsl(0, 50%, 50%);
color: hsl(0, 10%, 96%);
}
.grammalecte_token {
padding: 4px 0;
}
.grammalecte_token .separator {
margin: 20px 0;
padding: 5px 5px;
background-color: hsla(0, 0%, 75%, 1);
|
| ︙ | ︙ |
Modified gc_lang/fr/webext/content_scripts/lxg_content.js from [196128af47] to [7b81590ebf].
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 |
// JavaScript
"use strict";
const oLxgPanelContent = {
_xContentNode: createNode("div", {id: "grammalecte_lxg_panel_content"}),
getNode: function () {
return this._xContentNode;
},
clear: function () {
while (this._xContentNode.firstChild) {
this._xContentNode.removeChild(this._xContentNode.firstChild);
}
},
addSeparator: function (sText) {
if (this._xContentNode.textContent !== "") {
this._xContentNode.appendChild(createNode("div", {className: "grammalecte_lxg_separator", textContent: sText}));
}
},
addMessage: function (sClass, sText) {
this._xContentNode.appendChild(createNode("div", {className: sClass, textContent: sText}));
},
addListOfTokens: function (lTokens) {
try {
if (lTokens) {
let xNodeDiv = createNode("div", {className: "grammalecte_lxg_list_of_tokens"});
for (let oToken of lTokens) {
xNodeDiv.appendChild(this._createTokenNode(oToken));
}
this._xContentNode.appendChild(xNodeDiv);
}
}
catch (e) {
| > > > > > | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 |
// JavaScript
"use strict";
const oLxgPanelContent = {
_xContentNode: createNode("div", {id: "grammalecte_lxg_panel_content"}),
_nCount: 0,
getNode: function () {
return this._xContentNode;
},
clear: function () {
this._nCount = 0;
while (this._xContentNode.firstChild) {
this._xContentNode.removeChild(this._xContentNode.firstChild);
}
},
addSeparator: function (sText) {
if (this._xContentNode.textContent !== "") {
this._xContentNode.appendChild(createNode("div", {className: "grammalecte_lxg_separator", textContent: sText}));
}
},
addMessage: function (sClass, sText) {
this._xContentNode.appendChild(createNode("div", {className: sClass, textContent: sText}));
},
addListOfTokens: function (lTokens) {
try {
if (lTokens) {
this._nCount += 1;
let xNodeDiv = createNode("div", {className: "grammalecte_lxg_list_of_tokens"});
xNodeDiv.appendChild(createNode("div", {className: "num", textContent: this._nCount}));
for (let oToken of lTokens) {
xNodeDiv.appendChild(this._createTokenNode(oToken));
}
this._xContentNode.appendChild(xNodeDiv);
}
}
catch (e) {
|
| ︙ | ︙ |
Modified gc_lang/fr/webext/gce_worker.js from [64558676c7] to [6354516b05].
| ︙ | ︙ | |||
247 248 249 250 251 252 253 |
// Lexicographer
function getListOfTokens (sText, dInfo={}) {
try {
for (let sParagraph of text.getParagraph(sText)) {
| > | | | | | | | | | > | 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 |
// Lexicographer
function getListOfTokens (sText, dInfo={}) {
try {
for (let sParagraph of text.getParagraph(sText)) {
if (sParagraph.trim() !== "") {
let aElem = [];
let aRes = null;
for (let oToken of oTokenizer.genTokens(sParagraph)) {
aRes = oLxg.getInfoForToken(oToken);
if (aRes) {
aElem.push(aRes);
}
}
postMessage(createResponse("getListOfTokens", aElem, dInfo, false));
}
}
postMessage(createResponse("getListOfTokens", null, dInfo, true));
}
catch (e) {
helpers.logerror(e);
postMessage(createResponse("getListOfTokens", createErrorResult(e, "no tokens"), dInfo, true, true));
}
|
| ︙ | ︙ |