file_name
large_stringlengths
4
140
prefix
large_stringlengths
0
12.1k
suffix
large_stringlengths
0
12k
middle
large_stringlengths
0
7.51k
fim_type
large_stringclasses
4 values
tae-ui.js
var originalStyles = []; // Used to store the tagged paragraphs CSSstyles var simplifyBoxIdSuffix = '-simp-text-paragraph'; // Component-related methods and behaviour function initComponent(parameters) { primaryColor = parameters.primaryColor; secondaryColor = parameters.secondaryColor; elementsToEnhanceClassName = parameters.elementsToEnhanceClassName; simplifyBoxTitle = parameters.simplifyBoxTitle; simplifyBoxClassName = parameters.simplifyBoxClassName; wordPropertiesClassName = parameters.wordPropertiesClassName; synonymLabel = parameters.synonymLabel || 'Synonyms'; definitionLabel = parameters.definitionLabel || 'Definitions'; wikipediaLabel = parameters.wikipediaLabel || 'Wikipedia'; emptyText = parameters.emptyText || 'no simplification found for the text'; taeCORE.getInstance().init({ endpoint: parameters.endpoint, language: parameters.language }); } function enableComponentFeatures() { if (featureEnabled) return; featureEnabled = true; // Gets the tagged paragraphs the first time if (paragraphs.length === 0) { paragraphs = document.getElementsByClassName(elementsToEnhanceClassName); } // Add special format and add a couple of attributes to the paragraphs var paragrapId = 1; var paragraphName = ''; for (var i = 0, len = paragraphs.length; i < len; i++) { if (paragraphs[i].className.indexOf(elementsToEnhanceClassName + "-active") < 0) paragraphs[i].className += ' '+elementsToEnhanceClassName + "-active"; paragraphName = "taeParagraph" + paragrapId; // Store original style // originalStyles[i] = paragraphs[i].style; // paragraphs[i].style.position = 'relative'; // paragraphs[i].style.borderLeft = "12px solid " + primaryColor; // paragraphs[i].style.borderRadius = "16px"; // // paragraphs[i].style.padding = '0px 0px 0px 8px'; // paragraphs[i].style.margin = '0px 0px 8px 0px'; paragraphs[i].setAttribute("id", paragraphName); paragraphs[i].setAttribute("onclick", "taeUI.getInstance()." + "paragraphEvent('" + paragraphName + "');"); var loadingImage = document.createElement("img"); loadingImage.setAttribute("src", "img/loader.gif"); loadingImage.setAttribute("id", "loading_"+paragraphName); loadingImage.style.display = "none"; paragraphs[i].appendChild(loadingImage); paragrapId++; } } function disableComponentFeatures() { if (!featureEnabled) return; featureEnabled = false; // Remove Question Boxes var questionsBoxes = document.getElementsByClassName(simplifyBoxClassName); for (var i = questionsBoxes.length - 1; i >= 0; i--) { questionsBoxes[i].parentNode.removeChild(questionsBoxes[i]); } // Reformat the paragraphs with the original style for (var i = 0, len = paragraphs.length; i < len; i++) { // Restore the original style paragraphs[i].style = originalStyles[i]; paragraphs[i].className = paragraphs[i].className.replace(elementsToEnhanceClassName + "-active", ""); // Remove the onclick event to enhance the paragraph paragraphs[i].removeAttribute("onclick"); } } // It uses the log component to register the produced events var logger = function(event, details) { var nop = function(){}; if (logCORE != null) return logCORE.getInstance().taeLogger; else return {logParagraph: nop, logPhrase: nop, logWord: nop, logFreetext: nop}; } // If the Component feature is enabled it calls to the TAE engine instance to // get the simplifications related to the paragraph passed as parameter // - paragraphID: the id of the paragraph which has produced the event function paragraphEvent(paragraphID) { if (!featureEnabled) return; var currentParagraph = document.getElementById(paragraphID + simplifyBoxIdSuffix); if ( currentParagraph === null) { logger().logParagraph(simpaticoEservice, paragraphID); currentParagraph = document.getElementById(paragraphID); var text = currentParagraph.textContent ? currentParagraph.textContent : currentParagraph.innerText;//IE uses innerText taeCORE.getInstance().simplifyText(paragraphID, text, showSimplificationBox); } else { hideSimplificationBox(paragraphID); } } // It creates the HTML content of a complex word // Used by createSimplifiedTextHTML(...) // - item: the object wich contains the description passed as parameter function createSimplifiedWordLabel(item) { return '<span class="simp-word" ' + 'onclick="taeUI.getInstance().wordEvent(event, this)">' + item.originalValue + '</span>'; } // It creates the HTML content of a simplified paraghaph // Used by getSimplifiedText(...) // - originalText: the original text contained in a paragraph // - simplifications: A list of simplified words of the text function createSimplifiedTextHTML(originalText, simplifications) { Array.prototype.keySort = function(key, desc){ this.sort(function(a, b) { var result = desc ? (a[key] < b[key]) : (a[key] > b[key]); return result ? 1 : -1; }); return this; } simplifications.keySort('start'); if (simplifications.length == 0) { var result = emptyText;//'No hay palabras que necesiten ser simplificadas'; }else{ var result = originalText; var item = ''; // for each simplified word add an element containing it for (var i = simplifications.length -1; i >= 0; i--)
} return result; } // Method used to cancel the propagation of the events // - event: the event to cancel function cancelEventPropagation(event) { event = event || window.event // cross-browser event if (event.stopPropagation) { event.stopPropagation(); // W3C standard variant } else { event.cancelBubble = true; // IE variant } } // Function called when an user clicks on a difficult word // It manages the event and shows the synonyms and definition of the // selected word calling to showWordProperties(...) // - event: the click event. It is cancelled // - wordHTMLelement: the element that contains the word function wordEvent(event, wordHTMLelement) { cancelEventPropagation(event); showWordProperties(wordHTMLelement); } // Function called when an user clicks on a highlighted word // It shows synonyms and the definition of the word contailed by the // HTML element passed as parameter // - wordHTMLelement: the element that contains the word function showWordProperties(wordHTMLelement) { var simplifiedBoxNode = document.getElementById(wordHTMLelement .parentNode .parentNode .parentNode.id); var paragraphId = simplifiedBoxNode.parentNode.id; var currentBox = simplifiedBoxNode .getElementsByClassName(wordPropertiesClassName)[0]; // If the currentBox is not created, create and attach it if (currentBox == null) { currentBox = document.createElement('li'); currentBox.className = wordPropertiesClassName; currentBox.setAttribute("onclick", "taeUI.getInstance()." + "wordPropertiesEvent(event,'" + paragraphId + "');"); simplifiedBoxNode.getElementsByTagName('ul')[0].appendChild(currentBox); } // Get the synonyms and definition var definition = taeCORE.getInstance() .termDefinition(paragraphId, wordHTMLelement.innerHTML); var synonyms = taeCORE.getInstance() .termSynonyms(paragraphId, wordHTMLelement.innerHTML); var wiki = taeCORE.getInstance() .termWikipedia(paragraphId, wordHTMLelement.innerHTML); // Update the content currentBox.innerHTML = '<b>' + wordHTMLelement.innerText + '</b></br>'; if (definition != null) // If the word has definition show it currentBox.innerHTML += '<i>' + definitionLabel + ':' + '</i>' + definition + '</br>'; if (synonyms != null) // If the word has synonyms show them currentBox.innerHTML += '<i>' + synonymLabel +':' + '</i>' + synonyms; if (wiki != null) // If the word has a wikipedia link currentBox.innerHTML += '<br/><i>' + wikipediaLabel +':' + '</i><a target="_blank" href="'+wiki+'">' + wiki + '<a/>'; logger().logWord(simpaticoEservice, wordHTMLelement.innerHTML); } // Function called when an user clicks on a WordProperties box // It hides the selected WordProperties box // - event: the click event. It is cancelled // - paragraphID: the id paragraph that contains the WordProperties box function hideWordProperties(event, paragraphID)
{ item = simplifications[i]; console.log(item); result = result.substring(0, item.start) + createSimplifiedWordLabel(item) + result.substring(item.end, result.length); }
conditional_block
tae-ui.js
// Component-related variables var primaryColor = ''; var secondaryColor = ''; var elementsToEnhanceClassName = ''; var simplifyBoxTitle = ''; var simplifyBoxClassName = ''; var wordPropertiesClassName = ''; var synonymLabel = ''; var definitionLabel = ''; var emptyText = ''; // Internal usage variables var paragraphs = []; // Used to store all the tagged paragraphs var originalStyles = []; // Used to store the tagged paragraphs CSSstyles var simplifyBoxIdSuffix = '-simp-text-paragraph'; // Component-related methods and behaviour function initComponent(parameters) { primaryColor = parameters.primaryColor; secondaryColor = parameters.secondaryColor; elementsToEnhanceClassName = parameters.elementsToEnhanceClassName; simplifyBoxTitle = parameters.simplifyBoxTitle; simplifyBoxClassName = parameters.simplifyBoxClassName; wordPropertiesClassName = parameters.wordPropertiesClassName; synonymLabel = parameters.synonymLabel || 'Synonyms'; definitionLabel = parameters.definitionLabel || 'Definitions'; wikipediaLabel = parameters.wikipediaLabel || 'Wikipedia'; emptyText = parameters.emptyText || 'no simplification found for the text'; taeCORE.getInstance().init({ endpoint: parameters.endpoint, language: parameters.language }); } function enableComponentFeatures() { if (featureEnabled) return; featureEnabled = true; // Gets the tagged paragraphs the first time if (paragraphs.length === 0) { paragraphs = document.getElementsByClassName(elementsToEnhanceClassName); } // Add special format and add a couple of attributes to the paragraphs var paragrapId = 1; var paragraphName = ''; for (var i = 0, len = paragraphs.length; i < len; i++) { if (paragraphs[i].className.indexOf(elementsToEnhanceClassName + "-active") < 0) paragraphs[i].className += ' '+elementsToEnhanceClassName + "-active"; paragraphName = "taeParagraph" + paragrapId; // Store original style // originalStyles[i] = paragraphs[i].style; // paragraphs[i].style.position = 'relative'; // paragraphs[i].style.borderLeft = "12px solid " + primaryColor; // paragraphs[i].style.borderRadius = "16px"; // // paragraphs[i].style.padding = '0px 0px 0px 8px'; // paragraphs[i].style.margin = '0px 0px 8px 0px'; paragraphs[i].setAttribute("id", paragraphName); paragraphs[i].setAttribute("onclick", "taeUI.getInstance()." + "paragraphEvent('" + paragraphName + "');"); var loadingImage = document.createElement("img"); loadingImage.setAttribute("src", "img/loader.gif"); loadingImage.setAttribute("id", "loading_"+paragraphName); loadingImage.style.display = "none"; paragraphs[i].appendChild(loadingImage); paragrapId++; } } function disableComponentFeatures() { if (!featureEnabled) return; featureEnabled = false; // Remove Question Boxes var questionsBoxes = document.getElementsByClassName(simplifyBoxClassName); for (var i = questionsBoxes.length - 1; i >= 0; i--) { questionsBoxes[i].parentNode.removeChild(questionsBoxes[i]); } // Reformat the paragraphs with the original style for (var i = 0, len = paragraphs.length; i < len; i++) { // Restore the original style paragraphs[i].style = originalStyles[i]; paragraphs[i].className = paragraphs[i].className.replace(elementsToEnhanceClassName + "-active", ""); // Remove the onclick event to enhance the paragraph paragraphs[i].removeAttribute("onclick"); } } // It uses the log component to register the produced events var logger = function(event, details) { var nop = function(){}; if (logCORE != null) return logCORE.getInstance().taeLogger; else return {logParagraph: nop, logPhrase: nop, logWord: nop, logFreetext: nop}; } // If the Component feature is enabled it calls to the TAE engine instance to // get the simplifications related to the paragraph passed as parameter // - paragraphID: the id of the paragraph which has produced the event function paragraphEvent(paragraphID) { if (!featureEnabled) return; var currentParagraph = document.getElementById(paragraphID + simplifyBoxIdSuffix); if ( currentParagraph === null) { logger().logParagraph(simpaticoEservice, paragraphID); currentParagraph = document.getElementById(paragraphID); var text = currentParagraph.textContent ? currentParagraph.textContent : currentParagraph.innerText;//IE uses innerText taeCORE.getInstance().simplifyText(paragraphID, text, showSimplificationBox); } else { hideSimplificationBox(paragraphID); } } // It creates the HTML content of a complex word // Used by createSimplifiedTextHTML(...) // - item: the object wich contains the description passed as parameter function createSimplifiedWordLabel(item) { return '<span class="simp-word" ' + 'onclick="taeUI.getInstance().wordEvent(event, this)">' + item.originalValue + '</span>'; } // It creates the HTML content of a simplified paraghaph // Used by getSimplifiedText(...) // - originalText: the original text contained in a paragraph // - simplifications: A list of simplified words of the text function createSimplifiedTextHTML(originalText, simplifications) { Array.prototype.keySort = function(key, desc){ this.sort(function(a, b) { var result = desc ? (a[key] < b[key]) : (a[key] > b[key]); return result ? 1 : -1; }); return this; } simplifications.keySort('start'); if (simplifications.length == 0) { var result = emptyText;//'No hay palabras que necesiten ser simplificadas'; }else{ var result = originalText; var item = ''; // for each simplified word add an element containing it for (var i = simplifications.length -1; i >= 0; i--) { item = simplifications[i]; console.log(item); result = result.substring(0, item.start) + createSimplifiedWordLabel(item) + result.substring(item.end, result.length); } } return result; } // Method used to cancel the propagation of the events // - event: the event to cancel function cancelEventPropagation(event) { event = event || window.event // cross-browser event if (event.stopPropagation) { event.stopPropagation(); // W3C standard variant } else { event.cancelBubble = true; // IE variant } } // Function called when an user clicks on a difficult word // It manages the event and shows the synonyms and definition of the // selected word calling to showWordProperties(...) // - event: the click event. It is cancelled // - wordHTMLelement: the element that contains the word function wordEvent(event, wordHTMLelement) { cancelEventPropagation(event); showWordProperties(wordHTMLelement); } // Function called when an user clicks on a highlighted word // It shows synonyms and the definition of the word contailed by the // HTML element passed as parameter // - wordHTMLelement: the element that contains the word function showWordProperties(wordHTMLelement) { var simplifiedBoxNode = document.getElementById(wordHTMLelement .parentNode .parentNode .parentNode.id); var paragraphId = simplifiedBoxNode.parentNode.id; var currentBox = simplifiedBoxNode .getElementsByClassName(wordPropertiesClassName)[0]; // If the currentBox is not created, create and attach it if (currentBox == null) { currentBox = document.createElement('li'); currentBox.className = wordPropertiesClassName; currentBox.setAttribute("onclick", "taeUI.getInstance()." + "wordPropertiesEvent(event,'" + paragraphId + "');"); simplifiedBoxNode.getElementsByTagName('ul')[0].appendChild(currentBox); } // Get the synonyms and definition var definition = taeCORE.getInstance() .termDefinition(paragraphId, wordHTMLelement.innerHTML); var synonyms = taeCORE.getInstance() .termSynonyms(paragraphId, wordHTMLelement.innerHTML); var wiki = taeCORE.getInstance() .termWikipedia(paragraphId, wordHTMLelement.innerHTML); // Update the content currentBox.innerHTML = '<b>' + wordHTMLelement.innerText + '</b></br>'; if (definition != null) // If the word has definition show it currentBox.innerHTML += '<i>' + definitionLabel + ':' + '</i>' + definition + '</br>'; if (synonyms != null) // If the word has synonyms show them currentBox.innerHTML += '<i>' + synonymLabel +':' + '</i>' + synonyms; if (wiki != null) // If the word has a wikipedia link
var taeUI = (function () { var instance; // Singleton Instance of the UI component var featureEnabled = false; function Singleton () {
random_line_split
tae-ui.js
'Synonyms'; definitionLabel = parameters.definitionLabel || 'Definitions'; wikipediaLabel = parameters.wikipediaLabel || 'Wikipedia'; emptyText = parameters.emptyText || 'no simplification found for the text'; taeCORE.getInstance().init({ endpoint: parameters.endpoint, language: parameters.language }); } function enableComponentFeatures() { if (featureEnabled) return; featureEnabled = true; // Gets the tagged paragraphs the first time if (paragraphs.length === 0) { paragraphs = document.getElementsByClassName(elementsToEnhanceClassName); } // Add special format and add a couple of attributes to the paragraphs var paragrapId = 1; var paragraphName = ''; for (var i = 0, len = paragraphs.length; i < len; i++) { if (paragraphs[i].className.indexOf(elementsToEnhanceClassName + "-active") < 0) paragraphs[i].className += ' '+elementsToEnhanceClassName + "-active"; paragraphName = "taeParagraph" + paragrapId; // Store original style // originalStyles[i] = paragraphs[i].style; // paragraphs[i].style.position = 'relative'; // paragraphs[i].style.borderLeft = "12px solid " + primaryColor; // paragraphs[i].style.borderRadius = "16px"; // // paragraphs[i].style.padding = '0px 0px 0px 8px'; // paragraphs[i].style.margin = '0px 0px 8px 0px'; paragraphs[i].setAttribute("id", paragraphName); paragraphs[i].setAttribute("onclick", "taeUI.getInstance()." + "paragraphEvent('" + paragraphName + "');"); var loadingImage = document.createElement("img"); loadingImage.setAttribute("src", "img/loader.gif"); loadingImage.setAttribute("id", "loading_"+paragraphName); loadingImage.style.display = "none"; paragraphs[i].appendChild(loadingImage); paragrapId++; } } function disableComponentFeatures() { if (!featureEnabled) return; featureEnabled = false; // Remove Question Boxes var questionsBoxes = document.getElementsByClassName(simplifyBoxClassName); for (var i = questionsBoxes.length - 1; i >= 0; i--) { questionsBoxes[i].parentNode.removeChild(questionsBoxes[i]); } // Reformat the paragraphs with the original style for (var i = 0, len = paragraphs.length; i < len; i++) { // Restore the original style paragraphs[i].style = originalStyles[i]; paragraphs[i].className = paragraphs[i].className.replace(elementsToEnhanceClassName + "-active", ""); // Remove the onclick event to enhance the paragraph paragraphs[i].removeAttribute("onclick"); } } // It uses the log component to register the produced events var logger = function(event, details) { var nop = function(){}; if (logCORE != null) return logCORE.getInstance().taeLogger; else return {logParagraph: nop, logPhrase: nop, logWord: nop, logFreetext: nop}; } // If the Component feature is enabled it calls to the TAE engine instance to // get the simplifications related to the paragraph passed as parameter // - paragraphID: the id of the paragraph which has produced the event function paragraphEvent(paragraphID) { if (!featureEnabled) return; var currentParagraph = document.getElementById(paragraphID + simplifyBoxIdSuffix); if ( currentParagraph === null) { logger().logParagraph(simpaticoEservice, paragraphID); currentParagraph = document.getElementById(paragraphID); var text = currentParagraph.textContent ? currentParagraph.textContent : currentParagraph.innerText;//IE uses innerText taeCORE.getInstance().simplifyText(paragraphID, text, showSimplificationBox); } else { hideSimplificationBox(paragraphID); } } // It creates the HTML content of a complex word // Used by createSimplifiedTextHTML(...) // - item: the object wich contains the description passed as parameter function createSimplifiedWordLabel(item) { return '<span class="simp-word" ' + 'onclick="taeUI.getInstance().wordEvent(event, this)">' + item.originalValue + '</span>'; } // It creates the HTML content of a simplified paraghaph // Used by getSimplifiedText(...) // - originalText: the original text contained in a paragraph // - simplifications: A list of simplified words of the text function createSimplifiedTextHTML(originalText, simplifications) { Array.prototype.keySort = function(key, desc){ this.sort(function(a, b) { var result = desc ? (a[key] < b[key]) : (a[key] > b[key]); return result ? 1 : -1; }); return this; } simplifications.keySort('start'); if (simplifications.length == 0) { var result = emptyText;//'No hay palabras que necesiten ser simplificadas'; }else{ var result = originalText; var item = ''; // for each simplified word add an element containing it for (var i = simplifications.length -1; i >= 0; i--) { item = simplifications[i]; console.log(item); result = result.substring(0, item.start) + createSimplifiedWordLabel(item) + result.substring(item.end, result.length); } } return result; } // Method used to cancel the propagation of the events // - event: the event to cancel function cancelEventPropagation(event) { event = event || window.event // cross-browser event if (event.stopPropagation) { event.stopPropagation(); // W3C standard variant } else { event.cancelBubble = true; // IE variant } } // Function called when an user clicks on a difficult word // It manages the event and shows the synonyms and definition of the // selected word calling to showWordProperties(...) // - event: the click event. It is cancelled // - wordHTMLelement: the element that contains the word function wordEvent(event, wordHTMLelement) { cancelEventPropagation(event); showWordProperties(wordHTMLelement); } // Function called when an user clicks on a highlighted word // It shows synonyms and the definition of the word contailed by the // HTML element passed as parameter // - wordHTMLelement: the element that contains the word function showWordProperties(wordHTMLelement) { var simplifiedBoxNode = document.getElementById(wordHTMLelement .parentNode .parentNode .parentNode.id); var paragraphId = simplifiedBoxNode.parentNode.id; var currentBox = simplifiedBoxNode .getElementsByClassName(wordPropertiesClassName)[0]; // If the currentBox is not created, create and attach it if (currentBox == null) { currentBox = document.createElement('li'); currentBox.className = wordPropertiesClassName; currentBox.setAttribute("onclick", "taeUI.getInstance()." + "wordPropertiesEvent(event,'" + paragraphId + "');"); simplifiedBoxNode.getElementsByTagName('ul')[0].appendChild(currentBox); } // Get the synonyms and definition var definition = taeCORE.getInstance() .termDefinition(paragraphId, wordHTMLelement.innerHTML); var synonyms = taeCORE.getInstance() .termSynonyms(paragraphId, wordHTMLelement.innerHTML); var wiki = taeCORE.getInstance() .termWikipedia(paragraphId, wordHTMLelement.innerHTML); // Update the content currentBox.innerHTML = '<b>' + wordHTMLelement.innerText + '</b></br>'; if (definition != null) // If the word has definition show it currentBox.innerHTML += '<i>' + definitionLabel + ':' + '</i>' + definition + '</br>'; if (synonyms != null) // If the word has synonyms show them currentBox.innerHTML += '<i>' + synonymLabel +':' + '</i>' + synonyms; if (wiki != null) // If the word has a wikipedia link currentBox.innerHTML += '<br/><i>' + wikipediaLabel +':' + '</i><a target="_blank" href="'+wiki+'">' + wiki + '<a/>'; logger().logWord(simpaticoEservice, wordHTMLelement.innerHTML); } // Function called when an user clicks on a WordProperties box // It hides the selected WordProperties box // - event: the click event. It is cancelled // - paragraphID: the id paragraph that contains the WordProperties box function hideWordProperties(event, paragraphID) { cancelEventPropagation(event); var paragraphNode = document.getElementById(paragraphID); var currentBox = paragraphNode .getElementsByClassName(wordPropertiesClassName)[0]; if (currentBox != null) { currentBox.parentNode.removeChild(currentBox); } } // Draw the simplification box // - paragraphID: the id of the paragraph // - originalText: the original text contained in the paragraph // - response: the JSON Object of the questions related to the paragraph function
showSimplificationBox
identifier_name
tae-ui.js
var originalStyles = []; // Used to store the tagged paragraphs CSSstyles var simplifyBoxIdSuffix = '-simp-text-paragraph'; // Component-related methods and behaviour function initComponent(parameters) { primaryColor = parameters.primaryColor; secondaryColor = parameters.secondaryColor; elementsToEnhanceClassName = parameters.elementsToEnhanceClassName; simplifyBoxTitle = parameters.simplifyBoxTitle; simplifyBoxClassName = parameters.simplifyBoxClassName; wordPropertiesClassName = parameters.wordPropertiesClassName; synonymLabel = parameters.synonymLabel || 'Synonyms'; definitionLabel = parameters.definitionLabel || 'Definitions'; wikipediaLabel = parameters.wikipediaLabel || 'Wikipedia'; emptyText = parameters.emptyText || 'no simplification found for the text'; taeCORE.getInstance().init({ endpoint: parameters.endpoint, language: parameters.language }); } function enableComponentFeatures()
// // paragraphs[i].style.padding = '0px 0px 0px 8px'; // paragraphs[i].style.margin = '0px 0px 8px 0px'; paragraphs[i].setAttribute("id", paragraphName); paragraphs[i].setAttribute("onclick", "taeUI.getInstance()." + "paragraphEvent('" + paragraphName + "');"); var loadingImage = document.createElement("img"); loadingImage.setAttribute("src", "img/loader.gif"); loadingImage.setAttribute("id", "loading_"+paragraphName); loadingImage.style.display = "none"; paragraphs[i].appendChild(loadingImage); paragrapId++; } } function disableComponentFeatures() { if (!featureEnabled) return; featureEnabled = false; // Remove Question Boxes var questionsBoxes = document.getElementsByClassName(simplifyBoxClassName); for (var i = questionsBoxes.length - 1; i >= 0; i--) { questionsBoxes[i].parentNode.removeChild(questionsBoxes[i]); } // Reformat the paragraphs with the original style for (var i = 0, len = paragraphs.length; i < len; i++) { // Restore the original style paragraphs[i].style = originalStyles[i]; paragraphs[i].className = paragraphs[i].className.replace(elementsToEnhanceClassName + "-active", ""); // Remove the onclick event to enhance the paragraph paragraphs[i].removeAttribute("onclick"); } } // It uses the log component to register the produced events var logger = function(event, details) { var nop = function(){}; if (logCORE != null) return logCORE.getInstance().taeLogger; else return {logParagraph: nop, logPhrase: nop, logWord: nop, logFreetext: nop}; } // If the Component feature is enabled it calls to the TAE engine instance to // get the simplifications related to the paragraph passed as parameter // - paragraphID: the id of the paragraph which has produced the event function paragraphEvent(paragraphID) { if (!featureEnabled) return; var currentParagraph = document.getElementById(paragraphID + simplifyBoxIdSuffix); if ( currentParagraph === null) { logger().logParagraph(simpaticoEservice, paragraphID); currentParagraph = document.getElementById(paragraphID); var text = currentParagraph.textContent ? currentParagraph.textContent : currentParagraph.innerText;//IE uses innerText taeCORE.getInstance().simplifyText(paragraphID, text, showSimplificationBox); } else { hideSimplificationBox(paragraphID); } } // It creates the HTML content of a complex word // Used by createSimplifiedTextHTML(...) // - item: the object wich contains the description passed as parameter function createSimplifiedWordLabel(item) { return '<span class="simp-word" ' + 'onclick="taeUI.getInstance().wordEvent(event, this)">' + item.originalValue + '</span>'; } // It creates the HTML content of a simplified paraghaph // Used by getSimplifiedText(...) // - originalText: the original text contained in a paragraph // - simplifications: A list of simplified words of the text function createSimplifiedTextHTML(originalText, simplifications) { Array.prototype.keySort = function(key, desc){ this.sort(function(a, b) { var result = desc ? (a[key] < b[key]) : (a[key] > b[key]); return result ? 1 : -1; }); return this; } simplifications.keySort('start'); if (simplifications.length == 0) { var result = emptyText;//'No hay palabras que necesiten ser simplificadas'; }else{ var result = originalText; var item = ''; // for each simplified word add an element containing it for (var i = simplifications.length -1; i >= 0; i--) { item = simplifications[i]; console.log(item); result = result.substring(0, item.start) + createSimplifiedWordLabel(item) + result.substring(item.end, result.length); } } return result; } // Method used to cancel the propagation of the events // - event: the event to cancel function cancelEventPropagation(event) { event = event || window.event // cross-browser event if (event.stopPropagation) { event.stopPropagation(); // W3C standard variant } else { event.cancelBubble = true; // IE variant } } // Function called when an user clicks on a difficult word // It manages the event and shows the synonyms and definition of the // selected word calling to showWordProperties(...) // - event: the click event. It is cancelled // - wordHTMLelement: the element that contains the word function wordEvent(event, wordHTMLelement) { cancelEventPropagation(event); showWordProperties(wordHTMLelement); } // Function called when an user clicks on a highlighted word // It shows synonyms and the definition of the word contailed by the // HTML element passed as parameter // - wordHTMLelement: the element that contains the word function showWordProperties(wordHTMLelement) { var simplifiedBoxNode = document.getElementById(wordHTMLelement .parentNode .parentNode .parentNode.id); var paragraphId = simplifiedBoxNode.parentNode.id; var currentBox = simplifiedBoxNode .getElementsByClassName(wordPropertiesClassName)[0]; // If the currentBox is not created, create and attach it if (currentBox == null) { currentBox = document.createElement('li'); currentBox.className = wordPropertiesClassName; currentBox.setAttribute("onclick", "taeUI.getInstance()." + "wordPropertiesEvent(event,'" + paragraphId + "');"); simplifiedBoxNode.getElementsByTagName('ul')[0].appendChild(currentBox); } // Get the synonyms and definition var definition = taeCORE.getInstance() .termDefinition(paragraphId, wordHTMLelement.innerHTML); var synonyms = taeCORE.getInstance() .termSynonyms(paragraphId, wordHTMLelement.innerHTML); var wiki = taeCORE.getInstance() .termWikipedia(paragraphId, wordHTMLelement.innerHTML); // Update the content currentBox.innerHTML = '<b>' + wordHTMLelement.innerText + '</b></br>'; if (definition != null) // If the word has definition show it currentBox.innerHTML += '<i>' + definitionLabel + ':' + '</i>' + definition + '</br>'; if (synonyms != null) // If the word has synonyms show them currentBox.innerHTML += '<i>' + synonymLabel +':' + '</i>' + synonyms; if (wiki != null) // If the word has a wikipedia link currentBox.innerHTML += '<br/><i>' + wikipediaLabel +':' + '</i><a target="_blank" href="'+wiki+'">' + wiki + '<a/>'; logger().logWord(simpaticoEservice, wordHTMLelement.innerHTML); } // Function called when an user clicks on a WordProperties box // It hides the selected WordProperties box // - event: the click event. It is cancelled // - paragraphID: the id paragraph that contains the WordProperties box function hideWordProperties(event, paragraphID)
{ if (featureEnabled) return; featureEnabled = true; // Gets the tagged paragraphs the first time if (paragraphs.length === 0) { paragraphs = document.getElementsByClassName(elementsToEnhanceClassName); } // Add special format and add a couple of attributes to the paragraphs var paragrapId = 1; var paragraphName = ''; for (var i = 0, len = paragraphs.length; i < len; i++) { if (paragraphs[i].className.indexOf(elementsToEnhanceClassName + "-active") < 0) paragraphs[i].className += ' '+elementsToEnhanceClassName + "-active"; paragraphName = "taeParagraph" + paragrapId; // Store original style // originalStyles[i] = paragraphs[i].style; // paragraphs[i].style.position = 'relative'; // paragraphs[i].style.borderLeft = "12px solid " + primaryColor; // paragraphs[i].style.borderRadius = "16px";
identifier_body
data_utils.py
.long) else: all_label_ids = torch.tensor([f.label_id for f in features], dtype=torch.float) all_seq_lengths = torch.tensor([f.seq_length for f in features], dtype=torch.long) all_input_ids = torch.tensor([f.input_ids for f in features], dtype=torch.long) all_input_mask = torch.tensor([f.input_mask for f in features], dtype=torch.long) all_segment_ids = torch.tensor([f.segment_ids for f in features], dtype=torch.long) s_ids = [f.guid for f in features] tensor_data = TensorDataset(all_input_ids, all_input_mask, all_segment_ids, all_label_ids, all_seq_lengths) return tensor_data, s_ids def convert_examples_to_features(examples, label_list, max_seq_length, tokenizer, output_mode, is_master=True): """Loads a data file into a list of `InputBatch`s.""" label_map = {label: i for i, label in enumerate(label_list)} features = [] for (ex_index, example) in enumerate(examples): if ex_index % 10000 == 0 and is_master: print("Writing example %d of %d" % (ex_index, len(examples))) tokens_a = tokenizer.tokenize(example.text_a) tokens_b = None if example.text_b: tokens_b = tokenizer.tokenize(example.text_b) _truncate_seq_pair(tokens_a, tokens_b, max_seq_length - 3) else: if len(tokens_a) > max_seq_length - 2: tokens_a = tokens_a[:(max_seq_length - 2)] tokens = ["[CLS]"] + tokens_a + ["[SEP]"] segment_ids = [0] * len(tokens) if tokens_b: tokens += tokens_b + ["[SEP]"] segment_ids += [1] * (len(tokens_b) + 1) input_ids = tokenizer.convert_tokens_to_ids(tokens) input_mask = [1] * len(input_ids) seq_length = len(input_ids) padding = [0] * (max_seq_length - len(input_ids)) input_ids += padding input_mask += padding segment_ids += padding assert len(input_ids) == max_seq_length assert len(input_mask) == max_seq_length assert len(segment_ids) == max_seq_length if output_mode == "classification": label_id = label_map[example.label] elif output_mode == "regression": label_id = float(example.label) else: raise KeyError(output_mode) if ex_index == 0 and is_master: print("*** Example ***") print("guid: %s" % (example.guid)) print("tokens: %s" % " ".join([str(x) for x in tokens])) print("input_ids: %s" % " ".join([str(x) for x in input_ids])) print("input_mask: %s" % " ".join([str(x) for x in input_mask])) print("segment_ids: %s" % " ".join([str(x) for x in segment_ids])) print("label: {}".format(example.label)) print("label_id: {}".format(label_id)) features.append( InputFeatures( input_ids=input_ids, input_mask=input_mask, segment_ids=segment_ids, label_id=label_id, seq_length=seq_length, guid=example.guid)) return features def _truncate_seq_pair(tokens_a, tokens_b, max_length): """Truncates a sequence pair in place to the maximum length.""" while True: total_length = len(tokens_a) + len(tokens_b) if total_length <= max_length: break if len(tokens_a) > len(tokens_b): tokens_a.pop() else: tokens_b.pop() def load_glue_dataset(config): from bert_fineturn.data_processor.glue import glue_processors as processors from bert_fineturn.data_processor.glue import glue_output_modes as output_modes from transformers import BertConfig, BertTokenizer task_name = config.datasets config.is_master = True config.multi_gpu = False processor = processors[task_name.lower()]() output_mode = output_modes[task_name.lower()] label_list = processor.get_labels() if output_mode == 'classification': n_classes = len(label_list) else: n_classes = 1 sids = dict() tokenizer = BertTokenizer.from_pretrained('teacher_utils/bert_base_uncased', do_lower_case=True) train_examples = processor.get_train_examples(config.data_src_path) train_features = convert_examples_to_features(train_examples, label_list, config.max_seq_length, tokenizer, output_mode, config.is_master) train_data, train_sids = get_tensor_data(output_mode, train_features) eval_examples = processor.get_dev_examples(config.data_src_path) eval_features = convert_examples_to_features(eval_examples, label_list, config.max_seq_length, tokenizer, output_mode, config.is_master) eval_data, eval_sids = get_tensor_data(output_mode, eval_features) test_examples = processor.get_test_examples(config.data_src_path) test_features = convert_examples_to_features(test_examples, label_list, config.max_seq_length, tokenizer, output_mode, config.is_master) test_data, test_sids = get_tensor_data(output_mode, test_features) train_eval_data, _ = get_tensor_data(output_mode, eval_features) if not config.multi_gpu: train_sampler = RandomSampler(train_data) train_eval_sampler = RandomSampler(train_eval_data) else: train_sampler = DistributedSampler(train_data) train_eval_sampler = DistributedSampler(train_eval_data) eval_sampler = SequentialSampler(eval_data) test_sampler = SequentialSampler(test_data) train_dataloader = DataLoader(train_data, sampler=train_sampler, batch_size=config.batch_size) eval_dataloader = DataLoader(eval_data, sampler=eval_sampler, batch_size=config.batch_size) train_eval_dataloader = DataLoader(train_eval_data, sampler=train_eval_sampler, batch_size=config.batch_size) test_dataloader = DataLoader(test_data, sampler=test_sampler, batch_size=config.batch_size) bert_config = BertConfig.from_pretrained("teacher_utils/bert_base_uncased/config.json") config.bert_config = bert_config sids = {"train": train_sids, "test": test_sids, "dev": eval_sids} return train_dataloader, train_eval_dataloader, eval_dataloader, test_dataloader, output_mode, n_classes, config, sids from torch.utils.data.sampler import Sampler class OrderdedSampler(Sampler): def __init__(self, dataset, order): self._dataset = dataset self._train_data_list = order self._train_data_list def __len__(self): return len(self._dataset) def __iter__(self): random.shuffle(self._train_data_list) for index in self._train_data_list: yield self._dataset[index] def check_data_vaild(data1, data2): # data1, data2 = next(iter(data1)), next(iter(data2)) def pad_replace(x): x = np.array(x) pad_mask = np.array([not(i == '[PAD]' or i == "<pad>") for i in x]) new_x = x[pad_mask].tolist() + [f'[PAD] * { - sum(pad_mask - 1)}'] return new_x def
(x): t = sum(x) new_x = f"1 * {t}, 0 * {len(x) - t}" return new_x with open('/data/lxk/NLP/github/darts-KD/data/MRPC-nas/embedding/vocab.txt') as f: vocab1 = {i:x.strip() for i, x in enumerate(f.readlines())} with open('/data/lxk/NLP/github/darts-KD/teacher_utils/teacher_model/MRPC/vocab.txt') as f: vocab2 = {i:x.strip() for i, x in enumerate(f.readlines())} sent_words = torch.split(data1[0], 1, dim=1) sent_words = [torch.squeeze(x, dim=1) for x in sent_words] mask = [x.ne(0) for x in sent_words] if len(mask) > 1: mask = torch.logical_or(mask[0], mask[1]) else: mask = mask[0] print("SENT1:", pad_replace([vocab1[x.item()] for x in data1[0][0][0]])) if data1[0].shape[1] == 2: print("SENT2:", pad_replace([vocab1[x.item()] for x in data1[0][0][1]])) print("MASK:", mask_replace(mask[0])) print("LABEL:", data1[2][0].item()) input_ids, input_mask, segment_ids, label_ids, seq_lengths = data2 print("TEACHER SENT:", pad_replace([vocab2[x.item()] for x in input_ids[0]])) print("TEACHER MASK", mask_replace(input_mask[0])) print("TEACHER LABEL", label_ids[0].item()) class RandomSamplerByOrder(Sampler): r"""Samples elements randomly. If without replacement, then sample from a shuffled dataset. If with replacement, then user can specify :attr:`num_samples` to draw. Arguments: data_source (Dataset): dataset to sample from replacement (bool): samples
mask_replace
identifier_name
data_utils.py
.long) else: all_label_ids = torch.tensor([f.label_id for f in features], dtype=torch.float) all_seq_lengths = torch.tensor([f.seq_length for f in features], dtype=torch.long) all_input_ids = torch.tensor([f.input_ids for f in features], dtype=torch.long) all_input_mask = torch.tensor([f.input_mask for f in features], dtype=torch.long) all_segment_ids = torch.tensor([f.segment_ids for f in features], dtype=torch.long) s_ids = [f.guid for f in features] tensor_data = TensorDataset(all_input_ids, all_input_mask, all_segment_ids, all_label_ids, all_seq_lengths) return tensor_data, s_ids def convert_examples_to_features(examples, label_list, max_seq_length, tokenizer, output_mode, is_master=True): """Loads a data file into a list of `InputBatch`s.""" label_map = {label: i for i, label in enumerate(label_list)} features = [] for (ex_index, example) in enumerate(examples): if ex_index % 10000 == 0 and is_master: print("Writing example %d of %d" % (ex_index, len(examples))) tokens_a = tokenizer.tokenize(example.text_a) tokens_b = None if example.text_b: tokens_b = tokenizer.tokenize(example.text_b) _truncate_seq_pair(tokens_a, tokens_b, max_seq_length - 3) else: if len(tokens_a) > max_seq_length - 2: tokens_a = tokens_a[:(max_seq_length - 2)] tokens = ["[CLS]"] + tokens_a + ["[SEP]"] segment_ids = [0] * len(tokens) if tokens_b: tokens += tokens_b + ["[SEP]"] segment_ids += [1] * (len(tokens_b) + 1) input_ids = tokenizer.convert_tokens_to_ids(tokens) input_mask = [1] * len(input_ids) seq_length = len(input_ids)
input_mask += padding segment_ids += padding assert len(input_ids) == max_seq_length assert len(input_mask) == max_seq_length assert len(segment_ids) == max_seq_length if output_mode == "classification": label_id = label_map[example.label] elif output_mode == "regression": label_id = float(example.label) else: raise KeyError(output_mode) if ex_index == 0 and is_master: print("*** Example ***") print("guid: %s" % (example.guid)) print("tokens: %s" % " ".join([str(x) for x in tokens])) print("input_ids: %s" % " ".join([str(x) for x in input_ids])) print("input_mask: %s" % " ".join([str(x) for x in input_mask])) print("segment_ids: %s" % " ".join([str(x) for x in segment_ids])) print("label: {}".format(example.label)) print("label_id: {}".format(label_id)) features.append( InputFeatures( input_ids=input_ids, input_mask=input_mask, segment_ids=segment_ids, label_id=label_id, seq_length=seq_length, guid=example.guid)) return features def _truncate_seq_pair(tokens_a, tokens_b, max_length): """Truncates a sequence pair in place to the maximum length.""" while True: total_length = len(tokens_a) + len(tokens_b) if total_length <= max_length: break if len(tokens_a) > len(tokens_b): tokens_a.pop() else: tokens_b.pop() def load_glue_dataset(config): from bert_fineturn.data_processor.glue import glue_processors as processors from bert_fineturn.data_processor.glue import glue_output_modes as output_modes from transformers import BertConfig, BertTokenizer task_name = config.datasets config.is_master = True config.multi_gpu = False processor = processors[task_name.lower()]() output_mode = output_modes[task_name.lower()] label_list = processor.get_labels() if output_mode == 'classification': n_classes = len(label_list) else: n_classes = 1 sids = dict() tokenizer = BertTokenizer.from_pretrained('teacher_utils/bert_base_uncased', do_lower_case=True) train_examples = processor.get_train_examples(config.data_src_path) train_features = convert_examples_to_features(train_examples, label_list, config.max_seq_length, tokenizer, output_mode, config.is_master) train_data, train_sids = get_tensor_data(output_mode, train_features) eval_examples = processor.get_dev_examples(config.data_src_path) eval_features = convert_examples_to_features(eval_examples, label_list, config.max_seq_length, tokenizer, output_mode, config.is_master) eval_data, eval_sids = get_tensor_data(output_mode, eval_features) test_examples = processor.get_test_examples(config.data_src_path) test_features = convert_examples_to_features(test_examples, label_list, config.max_seq_length, tokenizer, output_mode, config.is_master) test_data, test_sids = get_tensor_data(output_mode, test_features) train_eval_data, _ = get_tensor_data(output_mode, eval_features) if not config.multi_gpu: train_sampler = RandomSampler(train_data) train_eval_sampler = RandomSampler(train_eval_data) else: train_sampler = DistributedSampler(train_data) train_eval_sampler = DistributedSampler(train_eval_data) eval_sampler = SequentialSampler(eval_data) test_sampler = SequentialSampler(test_data) train_dataloader = DataLoader(train_data, sampler=train_sampler, batch_size=config.batch_size) eval_dataloader = DataLoader(eval_data, sampler=eval_sampler, batch_size=config.batch_size) train_eval_dataloader = DataLoader(train_eval_data, sampler=train_eval_sampler, batch_size=config.batch_size) test_dataloader = DataLoader(test_data, sampler=test_sampler, batch_size=config.batch_size) bert_config = BertConfig.from_pretrained("teacher_utils/bert_base_uncased/config.json") config.bert_config = bert_config sids = {"train": train_sids, "test": test_sids, "dev": eval_sids} return train_dataloader, train_eval_dataloader, eval_dataloader, test_dataloader, output_mode, n_classes, config, sids from torch.utils.data.sampler import Sampler class OrderdedSampler(Sampler): def __init__(self, dataset, order): self._dataset = dataset self._train_data_list = order self._train_data_list def __len__(self): return len(self._dataset) def __iter__(self): random.shuffle(self._train_data_list) for index in self._train_data_list: yield self._dataset[index] def check_data_vaild(data1, data2): # data1, data2 = next(iter(data1)), next(iter(data2)) def pad_replace(x): x = np.array(x) pad_mask = np.array([not(i == '[PAD]' or i == "<pad>") for i in x]) new_x = x[pad_mask].tolist() + [f'[PAD] * { - sum(pad_mask - 1)}'] return new_x def mask_replace(x): t = sum(x) new_x = f"1 * {t}, 0 * {len(x) - t}" return new_x with open('/data/lxk/NLP/github/darts-KD/data/MRPC-nas/embedding/vocab.txt') as f: vocab1 = {i:x.strip() for i, x in enumerate(f.readlines())} with open('/data/lxk/NLP/github/darts-KD/teacher_utils/teacher_model/MRPC/vocab.txt') as f: vocab2 = {i:x.strip() for i, x in enumerate(f.readlines())} sent_words = torch.split(data1[0], 1, dim=1) sent_words = [torch.squeeze(x, dim=1) for x in sent_words] mask = [x.ne(0) for x in sent_words] if len(mask) > 1: mask = torch.logical_or(mask[0], mask[1]) else: mask = mask[0] print("SENT1:", pad_replace([vocab1[x.item()] for x in data1[0][0][0]])) if data1[0].shape[1] == 2: print("SENT2:", pad_replace([vocab1[x.item()] for x in data1[0][0][1]])) print("MASK:", mask_replace(mask[0])) print("LABEL:", data1[2][0].item()) input_ids, input_mask, segment_ids, label_ids, seq_lengths = data2 print("TEACHER SENT:", pad_replace([vocab2[x.item()] for x in input_ids[0]])) print("TEACHER MASK", mask_replace(input_mask[0])) print("TEACHER LABEL", label_ids[0].item()) class RandomSamplerByOrder(Sampler): r"""Samples elements randomly. If without replacement, then sample from a shuffled dataset. If with replacement, then user can specify :attr:`num_samples` to draw. Arguments: data_source (Dataset): dataset to sample from replacement (bool): samples are
padding = [0] * (max_seq_length - len(input_ids)) input_ids += padding
random_line_split
data_utils.py
def get_tensor_data(output_mode, features, ): if output_mode == "classification": all_label_ids = torch.tensor([f.label_id for f in features], dtype=torch.long) else: all_label_ids = torch.tensor([f.label_id for f in features], dtype=torch.float) all_seq_lengths = torch.tensor([f.seq_length for f in features], dtype=torch.long) all_input_ids = torch.tensor([f.input_ids for f in features], dtype=torch.long) all_input_mask = torch.tensor([f.input_mask for f in features], dtype=torch.long) all_segment_ids = torch.tensor([f.segment_ids for f in features], dtype=torch.long) s_ids = [f.guid for f in features] tensor_data = TensorDataset(all_input_ids, all_input_mask, all_segment_ids, all_label_ids, all_seq_lengths) return tensor_data, s_ids def convert_examples_to_features(examples, label_list, max_seq_length, tokenizer, output_mode, is_master=True): """Loads a data file into a list of `InputBatch`s.""" label_map = {label: i for i, label in enumerate(label_list)} features = [] for (ex_index, example) in enumerate(examples): if ex_index % 10000 == 0 and is_master: print("Writing example %d of %d" % (ex_index, len(examples))) tokens_a = tokenizer.tokenize(example.text_a) tokens_b = None if example.text_b: tokens_b = tokenizer.tokenize(example.text_b) _truncate_seq_pair(tokens_a, tokens_b, max_seq_length - 3) else: if len(tokens_a) > max_seq_length - 2: tokens_a = tokens_a[:(max_seq_length - 2)] tokens = ["[CLS]"] + tokens_a + ["[SEP]"] segment_ids = [0] * len(tokens) if tokens_b: tokens += tokens_b + ["[SEP]"] segment_ids += [1] * (len(tokens_b) + 1) input_ids = tokenizer.convert_tokens_to_ids(tokens) input_mask = [1] * len(input_ids) seq_length = len(input_ids) padding = [0] * (max_seq_length - len(input_ids)) input_ids += padding input_mask += padding segment_ids += padding assert len(input_ids) == max_seq_length assert len(input_mask) == max_seq_length assert len(segment_ids) == max_seq_length if output_mode == "classification": label_id = label_map[example.label] elif output_mode == "regression": label_id = float(example.label) else: raise KeyError(output_mode) if ex_index == 0 and is_master: print("*** Example ***") print("guid: %s" % (example.guid)) print("tokens: %s" % " ".join([str(x) for x in tokens])) print("input_ids: %s" % " ".join([str(x) for x in input_ids])) print("input_mask: %s" % " ".join([str(x) for x in input_mask])) print("segment_ids: %s" % " ".join([str(x) for x in segment_ids])) print("label: {}".format(example.label)) print("label_id: {}".format(label_id)) features.append( InputFeatures( input_ids=input_ids, input_mask=input_mask, segment_ids=segment_ids, label_id=label_id, seq_length=seq_length, guid=example.guid)) return features def _truncate_seq_pair(tokens_a, tokens_b, max_length): """Truncates a sequence pair in place to the maximum length.""" while True: total_length = len(tokens_a) + len(tokens_b) if total_length <= max_length: break if len(tokens_a) > len(tokens_b): tokens_a.pop() else: tokens_b.pop() def load_glue_dataset(config): from bert_fineturn.data_processor.glue import glue_processors as processors from bert_fineturn.data_processor.glue import glue_output_modes as output_modes from transformers import BertConfig, BertTokenizer task_name = config.datasets config.is_master = True config.multi_gpu = False processor = processors[task_name.lower()]() output_mode = output_modes[task_name.lower()] label_list = processor.get_labels() if output_mode == 'classification': n_classes = len(label_list) else: n_classes = 1 sids = dict() tokenizer = BertTokenizer.from_pretrained('teacher_utils/bert_base_uncased', do_lower_case=True) train_examples = processor.get_train_examples(config.data_src_path) train_features = convert_examples_to_features(train_examples, label_list, config.max_seq_length, tokenizer, output_mode, config.is_master) train_data, train_sids = get_tensor_data(output_mode, train_features) eval_examples = processor.get_dev_examples(config.data_src_path) eval_features = convert_examples_to_features(eval_examples, label_list, config.max_seq_length, tokenizer, output_mode, config.is_master) eval_data, eval_sids = get_tensor_data(output_mode, eval_features) test_examples = processor.get_test_examples(config.data_src_path) test_features = convert_examples_to_features(test_examples, label_list, config.max_seq_length, tokenizer, output_mode, config.is_master) test_data, test_sids = get_tensor_data(output_mode, test_features) train_eval_data, _ = get_tensor_data(output_mode, eval_features) if not config.multi_gpu: train_sampler = RandomSampler(train_data) train_eval_sampler = RandomSampler(train_eval_data) else: train_sampler = DistributedSampler(train_data) train_eval_sampler = DistributedSampler(train_eval_data) eval_sampler = SequentialSampler(eval_data) test_sampler = SequentialSampler(test_data) train_dataloader = DataLoader(train_data, sampler=train_sampler, batch_size=config.batch_size) eval_dataloader = DataLoader(eval_data, sampler=eval_sampler, batch_size=config.batch_size) train_eval_dataloader = DataLoader(train_eval_data, sampler=train_eval_sampler, batch_size=config.batch_size) test_dataloader = DataLoader(test_data, sampler=test_sampler, batch_size=config.batch_size) bert_config = BertConfig.from_pretrained("teacher_utils/bert_base_uncased/config.json") config.bert_config = bert_config sids = {"train": train_sids, "test": test_sids, "dev": eval_sids} return train_dataloader, train_eval_dataloader, eval_dataloader, test_dataloader, output_mode, n_classes, config, sids from torch.utils.data.sampler import Sampler class OrderdedSampler(Sampler): def __init__(self, dataset, order): self._dataset = dataset self._train_data_list = order self._train_data_list def __len__(self): return len(self._dataset) def __iter__(self): random.shuffle(self._train_data_list) for index in self._train_data_list: yield self._dataset[index] def check_data_vaild(data1, data2): # data1, data2 = next(iter(data1)), next(iter(data2)) def pad_replace(x): x = np.array(x) pad_mask = np.array([not(i == '[PAD]' or i == "<pad>") for i in x]) new_x = x[pad_mask].tolist() + [f'[PAD] * { - sum(pad_mask - 1)}'] return new_x def mask_replace(x): t = sum(x) new_x = f"1 * {t}, 0 * {len(x) - t}" return new_x with open('/data/lxk/NLP/github/darts-KD/data/MRPC-nas/embedding/vocab.txt') as f: vocab1 = {i:x.strip() for i, x in enumerate(f.readlines())} with open('/data/lxk/NLP/github/darts-KD/teacher_utils/teacher_model/MRPC/vocab.txt') as f: vocab2 = {i:x.strip() for i, x in enumerate(f.readlines())} sent_words = torch.split(data1[0], 1, dim=1) sent_words = [torch.squeeze(x, dim=1) for x in sent_words] mask = [x.ne(0) for x in sent_words] if len(mask) > 1: mask = torch.logical_or(mask[0], mask[1]) else: mask = mask[0] print("SENT1:", pad_replace([vocab1[x.item()] for x in data1[0][0][0]])) if data1[0].shape[1] == 2: print("SENT2:", pad_replace([vocab1[x.item()] for x in data1[0][0][1]])) print("MASK:", mask_replace(mask[0])) print("LABEL:", data1[2][0].item()) input_ids, input_mask, segment_ids, label_ids, seq_lengths = data2
"""A single set of features of data.""" def __init__(self, input_ids, input_mask, segment_ids, label_id, seq_length=None, guid=None): self.input_ids = input_ids self.input_mask = input_mask self.segment_ids = segment_ids self.seq_length = seq_length self.label_id = label_id self.guid = guid
identifier_body
data_utils.py
.long) else: all_label_ids = torch.tensor([f.label_id for f in features], dtype=torch.float) all_seq_lengths = torch.tensor([f.seq_length for f in features], dtype=torch.long) all_input_ids = torch.tensor([f.input_ids for f in features], dtype=torch.long) all_input_mask = torch.tensor([f.input_mask for f in features], dtype=torch.long) all_segment_ids = torch.tensor([f.segment_ids for f in features], dtype=torch.long) s_ids = [f.guid for f in features] tensor_data = TensorDataset(all_input_ids, all_input_mask, all_segment_ids, all_label_ids, all_seq_lengths) return tensor_data, s_ids def convert_examples_to_features(examples, label_list, max_seq_length, tokenizer, output_mode, is_master=True): """Loads a data file into a list of `InputBatch`s.""" label_map = {label: i for i, label in enumerate(label_list)} features = [] for (ex_index, example) in enumerate(examples): if ex_index % 10000 == 0 and is_master: print("Writing example %d of %d" % (ex_index, len(examples))) tokens_a = tokenizer.tokenize(example.text_a) tokens_b = None if example.text_b: tokens_b = tokenizer.tokenize(example.text_b) _truncate_seq_pair(tokens_a, tokens_b, max_seq_length - 3) else: if len(tokens_a) > max_seq_length - 2: tokens_a = tokens_a[:(max_seq_length - 2)] tokens = ["[CLS]"] + tokens_a + ["[SEP]"] segment_ids = [0] * len(tokens) if tokens_b: tokens += tokens_b + ["[SEP]"] segment_ids += [1] * (len(tokens_b) + 1) input_ids = tokenizer.convert_tokens_to_ids(tokens) input_mask = [1] * len(input_ids) seq_length = len(input_ids) padding = [0] * (max_seq_length - len(input_ids)) input_ids += padding input_mask += padding segment_ids += padding assert len(input_ids) == max_seq_length assert len(input_mask) == max_seq_length assert len(segment_ids) == max_seq_length if output_mode == "classification": label_id = label_map[example.label] elif output_mode == "regression": label_id = float(example.label) else: raise KeyError(output_mode) if ex_index == 0 and is_master: print("*** Example ***") print("guid: %s" % (example.guid)) print("tokens: %s" % " ".join([str(x) for x in tokens])) print("input_ids: %s" % " ".join([str(x) for x in input_ids])) print("input_mask: %s" % " ".join([str(x) for x in input_mask])) print("segment_ids: %s" % " ".join([str(x) for x in segment_ids])) print("label: {}".format(example.label)) print("label_id: {}".format(label_id)) features.append( InputFeatures( input_ids=input_ids, input_mask=input_mask, segment_ids=segment_ids, label_id=label_id, seq_length=seq_length, guid=example.guid)) return features def _truncate_seq_pair(tokens_a, tokens_b, max_length): """Truncates a sequence pair in place to the maximum length.""" while True: total_length = len(tokens_a) + len(tokens_b) if total_length <= max_length: break if len(tokens_a) > len(tokens_b): tokens_a.pop() else: tokens_b.pop() def load_glue_dataset(config): from bert_fineturn.data_processor.glue import glue_processors as processors from bert_fineturn.data_processor.glue import glue_output_modes as output_modes from transformers import BertConfig, BertTokenizer task_name = config.datasets config.is_master = True config.multi_gpu = False processor = processors[task_name.lower()]() output_mode = output_modes[task_name.lower()] label_list = processor.get_labels() if output_mode == 'classification':
else: n_classes = 1 sids = dict() tokenizer = BertTokenizer.from_pretrained('teacher_utils/bert_base_uncased', do_lower_case=True) train_examples = processor.get_train_examples(config.data_src_path) train_features = convert_examples_to_features(train_examples, label_list, config.max_seq_length, tokenizer, output_mode, config.is_master) train_data, train_sids = get_tensor_data(output_mode, train_features) eval_examples = processor.get_dev_examples(config.data_src_path) eval_features = convert_examples_to_features(eval_examples, label_list, config.max_seq_length, tokenizer, output_mode, config.is_master) eval_data, eval_sids = get_tensor_data(output_mode, eval_features) test_examples = processor.get_test_examples(config.data_src_path) test_features = convert_examples_to_features(test_examples, label_list, config.max_seq_length, tokenizer, output_mode, config.is_master) test_data, test_sids = get_tensor_data(output_mode, test_features) train_eval_data, _ = get_tensor_data(output_mode, eval_features) if not config.multi_gpu: train_sampler = RandomSampler(train_data) train_eval_sampler = RandomSampler(train_eval_data) else: train_sampler = DistributedSampler(train_data) train_eval_sampler = DistributedSampler(train_eval_data) eval_sampler = SequentialSampler(eval_data) test_sampler = SequentialSampler(test_data) train_dataloader = DataLoader(train_data, sampler=train_sampler, batch_size=config.batch_size) eval_dataloader = DataLoader(eval_data, sampler=eval_sampler, batch_size=config.batch_size) train_eval_dataloader = DataLoader(train_eval_data, sampler=train_eval_sampler, batch_size=config.batch_size) test_dataloader = DataLoader(test_data, sampler=test_sampler, batch_size=config.batch_size) bert_config = BertConfig.from_pretrained("teacher_utils/bert_base_uncased/config.json") config.bert_config = bert_config sids = {"train": train_sids, "test": test_sids, "dev": eval_sids} return train_dataloader, train_eval_dataloader, eval_dataloader, test_dataloader, output_mode, n_classes, config, sids from torch.utils.data.sampler import Sampler class OrderdedSampler(Sampler): def __init__(self, dataset, order): self._dataset = dataset self._train_data_list = order self._train_data_list def __len__(self): return len(self._dataset) def __iter__(self): random.shuffle(self._train_data_list) for index in self._train_data_list: yield self._dataset[index] def check_data_vaild(data1, data2): # data1, data2 = next(iter(data1)), next(iter(data2)) def pad_replace(x): x = np.array(x) pad_mask = np.array([not(i == '[PAD]' or i == "<pad>") for i in x]) new_x = x[pad_mask].tolist() + [f'[PAD] * { - sum(pad_mask - 1)}'] return new_x def mask_replace(x): t = sum(x) new_x = f"1 * {t}, 0 * {len(x) - t}" return new_x with open('/data/lxk/NLP/github/darts-KD/data/MRPC-nas/embedding/vocab.txt') as f: vocab1 = {i:x.strip() for i, x in enumerate(f.readlines())} with open('/data/lxk/NLP/github/darts-KD/teacher_utils/teacher_model/MRPC/vocab.txt') as f: vocab2 = {i:x.strip() for i, x in enumerate(f.readlines())} sent_words = torch.split(data1[0], 1, dim=1) sent_words = [torch.squeeze(x, dim=1) for x in sent_words] mask = [x.ne(0) for x in sent_words] if len(mask) > 1: mask = torch.logical_or(mask[0], mask[1]) else: mask = mask[0] print("SENT1:", pad_replace([vocab1[x.item()] for x in data1[0][0][0]])) if data1[0].shape[1] == 2: print("SENT2:", pad_replace([vocab1[x.item()] for x in data1[0][0][1]])) print("MASK:", mask_replace(mask[0])) print("LABEL:", data1[2][0].item()) input_ids, input_mask, segment_ids, label_ids, seq_lengths = data2 print("TEACHER SENT:", pad_replace([vocab2[x.item()] for x in input_ids[0]])) print("TEACHER MASK", mask_replace(input_mask[0])) print("TEACHER LABEL", label_ids[0].item()) class RandomSamplerByOrder(Sampler): r"""Samples elements randomly. If without replacement, then sample from a shuffled dataset. If with replacement, then user can specify :attr:`num_samples` to draw. Arguments: data_source (Dataset): dataset to sample from replacement (bool): samples
n_classes = len(label_list)
conditional_block
d.rs
::*; use crate::util::codejam::run_cases; impl Tile { fn to_char(self) -> char { match self { Empty => '.', Building => '#', Soldier => 'S', Turret => 'T', } } } impl From<char> for Tile { fn from(item: char) -> Self { match item { '.' => Empty, '#' => Building, 'S' => Soldier, 'T' => Turret, _ => panic!("Character not recognized: {}", item), } } } impl Display for Tile { fn fmt(&self, f: &mut Formatter) -> fmt::Result { write!(f, "{}", self.to_char()) } } impl Default for Tile { fn default() -> Tile { Empty } } //problem specific code fn reachable(grid: &Grid<Tile>, location: &GridCoord) -> HashSet<GridRowColVec> { let mut r = HashSet::new(); //debug!("\nTracing {} starting at {}", location, direction); for direction in DIRECTIONS.iter() { let mut loc: GridRowColVec = location.convert(); for _ in 0..=grid.R + grid.C { loc += direction; if let Some(tile) = grid.get_value(&loc) { match *tile { Building => { break; } _ => { r.insert(loc.clone()); } }; } else { break; } } } r } /* impl<L, R> FromIterator<(L, R)> for BiMap<L, R> { fn from_iter<I: IntoIterator<Item = (L, R)>>(iter: I) -> Self { let mut c = BiMap::new(); for i in iter { c.insert(i.0, i.1); } c } }*/ fn solve<'a>(case_no: u32, grid: &mut Grid<Tile>, M_soldier_limit: usize) -> String { debug!( "Solving case {}\nM={}\n{}\n", case_no, M_soldier_limit, grid ); //original solider & turret index to location map let S_map = grid .filter_by_val(&Soldier) .enumerate() .collect::<BiMap<_, _>>(); let turret_locations = grid.filter_by_val(&Turret).collect::<Vec<_>>(); //precalucate what squares a turret can reach let turret_reachable_squares_list = turret_locations .iter() .map(|t_loc| reachable(&grid, &t_loc)) .collect::<Vec<_>>(); let T_map = turret_locations .into_iter() .enumerate() .collect::<BiMap<_, _>>(); let S = grid.filter_by_val(&Soldier).count(); let T = grid.filter_by_val(&Turret).count(); //Construct the initial Graph let G_edges = build_graph( &grid, false, M_soldier_limit, &S_map, &T_map, &turret_reachable_squares_list, ); let mut G = FlowGraph::new(2 + S + T, 4); for uv in G_edges { G.add_edge(uv.0, uv.1, 1, 1); } let source = S + T; let sink = S + T + 1; let vertex_to_string = |v: usize| match v { s if s < S => format!("Soldier #{} ({:?})", s + 1, *S_map.get_by_left(&s).unwrap()), t if t >= S && t < S + T => format!( "Turret #{} ({:?})", t - S + 1, *T_map.get_by_left(&(t - S)).unwrap() ), v if v == sink => "Sink".to_string(), _source => "Source".to_string(), }; //BFS for each soldier //will be in left to right order, then top down order //Now find max matching of G (G has an edge from soldier s to turret t if and only if soldier s can destroy turret t after all other turrets have been destroyed) for s in 0..S { G.add_edge(source, s, 1, 1); } for t in S..S + T { G.add_edge(t, sink, 1, 1); } let (R, flow) = G.dinic(source, sink); let mut ans = format!("Case #{}: {}\n", case_no, R); //Compute initial matching let mut M = flow .iter() .enumerate() .filter(|&(_e, f)| *f > 0) //map to u->v .map(|(e, _f)| (G.graph.endp[e ^ 1], G.graph.endp[e])) //leave out source and sink nodes .filter(|&(u, v)| u != source && v != sink) .collect::<Vec<_>>(); debug!( "Edges in M initial matching=\n{}\n", M.iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); let mut r = R; while r > 0 { //Let us define the graph G' with the same nodes as G, but an edge between soldier s and turret t only exists in G' if s can destroy t with the other turrets active let Gprime = build_graph( &grid, true, M_soldier_limit, &S_map, &T_map, &turret_reachable_squares_list, ); //Now build graph H let mut H = Graph::new(S + T, 4); let soldiers_in_m = M.iter().map(|&(s, _t)| s).collect::<Vec<_>>(); for &(s, t) in Gprime.iter() { if soldiers_in_m.contains(&s) { H.add_edge(s, t); } } for &(s, t) in M.iter() { H.add_edge(t, s); } debug!( "Current matching M =\n{}\n", M.iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); debug!( "Edges in G'=\n{}\n", Gprime .iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); debug!( "Edges in H=\n{}\n", H.edges() .map(|(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); let turrets_in_M = M.iter().map(|&(_s, t)| t).collect::<Vec<_>>(); //find an edge (s,t') where t' is not in m let st_prime = Gprime.iter().find(|&(_s, t)| !turrets_in_M.contains(t)); if st_prime.is_some() { let &(s, t) = st_prime.unwrap(); debug!("Found (s,t') s={} t'={}", s, t - S); ans += &format!("{} {}\n", s + 1, t - S + 1); grid[S_map.get_by_left(&s).unwrap()] = Empty; grid[T_map.get_by_left(&(t - S)).unwrap()] = Empty; r -= 1; //Also remove from current matching let to_remove = M .iter() .position(|&(s_in_m, _t)| s_in_m == s) .expect("Soldier should be in mapping"); M.remove(to_remove); continue; } //Now we need to find a cycle //Start at a soldier in H let soldier_in_h = H.edges().filter(|&(u, _v)| u <= S).next().unwrap().0; let mut cycle_edges = VecDeque::new(); let mut edge = ( soldier_in_h, H.adj_list_with_edges(soldier_in_h).next().unwrap().1, ); let mut visited = BitVec::from_elem(H.num_v(), false);
cycle_edges.push_back(edge); debug!( "pushed Edge {:?} ", format!("{}->{}", vertex_to_string(edge.0), vertex_to_string(edge.1)) ); //adj list returns an (internal edge index, next vertex) edge = (edge.1, H.adj_list_with_edges(edge.1).next().unwrap().1); debug!("Edge {:?} ", edge); } //cut to the actual cycle found let cycle_end = cycle_edges.back().unwrap().1; let cycle_start = cycle_edges .iter() .position(|
while !visited[edge.0] { visited.set(edge.0, true);
random_line_split
d.rs
::*; use crate::util::codejam::run_cases; impl Tile { fn to_char(self) -> char { match self { Empty => '.', Building => '#', Soldier => 'S', Turret => 'T', } } } impl From<char> for Tile { fn from(item: char) -> Self { match item { '.' => Empty, '#' => Building, 'S' => Soldier, 'T' => Turret, _ => panic!("Character not recognized: {}", item), } } } impl Display for Tile { fn fmt(&self, f: &mut Formatter) -> fmt::Result { write!(f, "{}", self.to_char()) } } impl Default for Tile { fn default() -> Tile { Empty } } //problem specific code fn reachable(grid: &Grid<Tile>, location: &GridCoord) -> HashSet<GridRowColVec>
break; } } } r } /* impl<L, R> FromIterator<(L, R)> for BiMap<L, R> { fn from_iter<I: IntoIterator<Item = (L, R)>>(iter: I) -> Self { let mut c = BiMap::new(); for i in iter { c.insert(i.0, i.1); } c } }*/ fn solve<'a>(case_no: u32, grid: &mut Grid<Tile>, M_soldier_limit: usize) -> String { debug!( "Solving case {}\nM={}\n{}\n", case_no, M_soldier_limit, grid ); //original solider & turret index to location map let S_map = grid .filter_by_val(&Soldier) .enumerate() .collect::<BiMap<_, _>>(); let turret_locations = grid.filter_by_val(&Turret).collect::<Vec<_>>(); //precalucate what squares a turret can reach let turret_reachable_squares_list = turret_locations .iter() .map(|t_loc| reachable(&grid, &t_loc)) .collect::<Vec<_>>(); let T_map = turret_locations .into_iter() .enumerate() .collect::<BiMap<_, _>>(); let S = grid.filter_by_val(&Soldier).count(); let T = grid.filter_by_val(&Turret).count(); //Construct the initial Graph let G_edges = build_graph( &grid, false, M_soldier_limit, &S_map, &T_map, &turret_reachable_squares_list, ); let mut G = FlowGraph::new(2 + S + T, 4); for uv in G_edges { G.add_edge(uv.0, uv.1, 1, 1); } let source = S + T; let sink = S + T + 1; let vertex_to_string = |v: usize| match v { s if s < S => format!("Soldier #{} ({:?})", s + 1, *S_map.get_by_left(&s).unwrap()), t if t >= S && t < S + T => format!( "Turret #{} ({:?})", t - S + 1, *T_map.get_by_left(&(t - S)).unwrap() ), v if v == sink => "Sink".to_string(), _source => "Source".to_string(), }; //BFS for each soldier //will be in left to right order, then top down order //Now find max matching of G (G has an edge from soldier s to turret t if and only if soldier s can destroy turret t after all other turrets have been destroyed) for s in 0..S { G.add_edge(source, s, 1, 1); } for t in S..S + T { G.add_edge(t, sink, 1, 1); } let (R, flow) = G.dinic(source, sink); let mut ans = format!("Case #{}: {}\n", case_no, R); //Compute initial matching let mut M = flow .iter() .enumerate() .filter(|&(_e, f)| *f > 0) //map to u->v .map(|(e, _f)| (G.graph.endp[e ^ 1], G.graph.endp[e])) //leave out source and sink nodes .filter(|&(u, v)| u != source && v != sink) .collect::<Vec<_>>(); debug!( "Edges in M initial matching=\n{}\n", M.iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); let mut r = R; while r > 0 { //Let us define the graph G' with the same nodes as G, but an edge between soldier s and turret t only exists in G' if s can destroy t with the other turrets active let Gprime = build_graph( &grid, true, M_soldier_limit, &S_map, &T_map, &turret_reachable_squares_list, ); //Now build graph H let mut H = Graph::new(S + T, 4); let soldiers_in_m = M.iter().map(|&(s, _t)| s).collect::<Vec<_>>(); for &(s, t) in Gprime.iter() { if soldiers_in_m.contains(&s) { H.add_edge(s, t); } } for &(s, t) in M.iter() { H.add_edge(t, s); } debug!( "Current matching M =\n{}\n", M.iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); debug!( "Edges in G'=\n{}\n", Gprime .iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); debug!( "Edges in H=\n{}\n", H.edges() .map(|(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); let turrets_in_M = M.iter().map(|&(_s, t)| t).collect::<Vec<_>>(); //find an edge (s,t') where t' is not in m let st_prime = Gprime.iter().find(|&(_s, t)| !turrets_in_M.contains(t)); if st_prime.is_some() { let &(s, t) = st_prime.unwrap(); debug!("Found (s,t') s={} t'={}", s, t - S); ans += &format!("{} {}\n", s + 1, t - S + 1); grid[S_map.get_by_left(&s).unwrap()] = Empty; grid[T_map.get_by_left(&(t - S)).unwrap()] = Empty; r -= 1; //Also remove from current matching let to_remove = M .iter() .position(|&(s_in_m, _t)| s_in_m == s) .expect("Soldier should be in mapping"); M.remove(to_remove); continue; } //Now we need to find a cycle //Start at a soldier in H let soldier_in_h = H.edges().filter(|&(u, _v)| u <= S).next().unwrap().0; let mut cycle_edges = VecDeque::new(); let mut edge = ( soldier_in_h, H.adj_list_with_edges(soldier_in_h).next().unwrap().1, ); let mut visited = BitVec::from_elem(H.num_v(), false); while !visited[edge.0] { visited.set(edge.0, true); cycle_edges.push_back(edge); debug!( "pushed Edge {:?} ", format!("{}->{}", vertex_to_string(edge.0), vertex_to_string(edge.1)) ); //adj list returns an (internal edge index, next vertex) edge = (edge.1, H.adj_list_with_edges(edge.1).next().unwrap().1); debug!("Edge {:?} ", edge); } //cut to the actual cycle found let cycle_end = cycle_edges.back().unwrap().1; let cycle_start = cycle_edges .iter() .position(|
{ let mut r = HashSet::new(); //debug!("\nTracing {} starting at {}", location, direction); for direction in DIRECTIONS.iter() { let mut loc: GridRowColVec = location.convert(); for _ in 0..=grid.R + grid.C { loc += direction; if let Some(tile) = grid.get_value(&loc) { match *tile { Building => { break; } _ => { r.insert(loc.clone()); } }; } else {
identifier_body
d.rs
.iter() { if soldiers_in_m.contains(&s) { H.add_edge(s, t); } } for &(s, t) in M.iter() { H.add_edge(t, s); } debug!( "Current matching M =\n{}\n", M.iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); debug!( "Edges in G'=\n{}\n", Gprime .iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); debug!( "Edges in H=\n{}\n", H.edges() .map(|(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); let turrets_in_M = M.iter().map(|&(_s, t)| t).collect::<Vec<_>>(); //find an edge (s,t') where t' is not in m let st_prime = Gprime.iter().find(|&(_s, t)| !turrets_in_M.contains(t)); if st_prime.is_some() { let &(s, t) = st_prime.unwrap(); debug!("Found (s,t') s={} t'={}", s, t - S); ans += &format!("{} {}\n", s + 1, t - S + 1); grid[S_map.get_by_left(&s).unwrap()] = Empty; grid[T_map.get_by_left(&(t - S)).unwrap()] = Empty; r -= 1; //Also remove from current matching let to_remove = M .iter() .position(|&(s_in_m, _t)| s_in_m == s) .expect("Soldier should be in mapping"); M.remove(to_remove); continue; } //Now we need to find a cycle //Start at a soldier in H let soldier_in_h = H.edges().filter(|&(u, _v)| u <= S).next().unwrap().0; let mut cycle_edges = VecDeque::new(); let mut edge = ( soldier_in_h, H.adj_list_with_edges(soldier_in_h).next().unwrap().1, ); let mut visited = BitVec::from_elem(H.num_v(), false); while !visited[edge.0] { visited.set(edge.0, true); cycle_edges.push_back(edge); debug!( "pushed Edge {:?} ", format!("{}->{}", vertex_to_string(edge.0), vertex_to_string(edge.1)) ); //adj list returns an (internal edge index, next vertex) edge = (edge.1, H.adj_list_with_edges(edge.1).next().unwrap().1); debug!("Edge {:?} ", edge); } //cut to the actual cycle found let cycle_end = cycle_edges.back().unwrap().1; let cycle_start = cycle_edges .iter() .position(|&(u, _v)| u == cycle_end) .unwrap(); cycle_edges.drain(0..cycle_start); debug!( "Cycle C =\n{}\n", cycle_edges .iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); //Consider a new matching M' of G consisting of the edges of M whose reverse is not in C, p // lus the edges in C whose reverse is not in M. That is, M' is M but exchanging the edges // present in C in some direction. M' in this case is also a matching of G of the same size as M //because it is a cycle, we know we have new edges from G' to replace the ones removed from M let mut M_new: Vec<(usize, usize)> = Vec::new(); M_new.extend(M.iter().filter(|&&(u, v)| !cycle_edges.contains(&(v, u)))); M_new.extend(cycle_edges.iter().filter(|&&(u, v)| !M.contains(&(v, u)))); debug!( "New matching M =\n{}\n", M_new .iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); //Find all edges from G' which are actions we can take let st_actions = M_new .iter() .filter(|&uv| Gprime.contains(uv)) .collect::<Vec<_>>(); for &&(s, t) in st_actions.iter() { debug!("Taking actions from g' s {} t {}", s + 1, t + 1 - S); ans += &format!("{} {}\n", s + 1, t - S + 1); grid[S_map.get_by_left(&s).unwrap()] = Empty; grid[T_map.get_by_left(&(t - S)).unwrap()] = Empty; r -= 1; } M = M_new; } ans } fn build_graph( grid: &Grid<Tile>, is_g_prime: bool, M: usize, s_mapping: &BiMap<usize, GridCoord>, t_mapping: &BiMap<usize, GridCoord>, turret_reachable_squares_list: &Vec<HashSet<GridRowColVec>>, ) -> IndexSet<(usize, usize)> { let mut G: IndexSet<(usize, usize)> = IndexSet::new(); let turret_locations = grid.filter_by_val(&Turret).collect::<HashSet<_>>(); /* for (turret_index, turret_squares) in turret_squares_list.iter().enumerate() { debug!("Turret {} can see {:?}", turret_index, turret_squares); } */ let soldier_locations = grid.filter_by_val(&Soldier).collect::<Vec<_>>(); let S = soldier_locations.len(); let T = turret_reachable_squares_list.len(); for (_soldier_index, soldier_loc) in soldier_locations.iter().enumerate() { //debug!("BFS search on soldier {} @ {}", soldier_index, soldier_loc); //Node is location, distance, seen_turret let mut queue: VecDeque<(GridRowColVec, usize, bool)> = VecDeque::new(); let mut visited = BitVec::from_elem(grid.C * grid.R, false); queue.push_back((soldier_loc.convert(), 0, false)); visited.set(soldier_loc.data[0] * grid.C + soldier_loc.data[1], true); while !queue.is_empty() { let (loc, dist, seen_turret) = queue.pop_front().unwrap(); let visible_turrets = turret_reachable_squares_list .iter() .enumerate() .filter(|(turret_index, turret_squares)| { turret_locations.contains(t_mapping.get_by_left(turret_index).unwrap()) && turret_squares.contains(&loc) }) .map(|(turret_index, _)| turret_index); let mut turret_visible = false; for turret_index in visible_turrets { turret_visible = true; if !is_g_prime || (!seen_turret && is_g_prime) { let s_vertex = *s_mapping.get_by_right(soldier_loc).unwrap(); //The turret index is already using the original grids index /*debug!("Found s{} t{} mapped to soldier {} => {} at loc {}", soldier_index, turret_index, s_vertex, t_vertex, loc);*/ G.insert((s_vertex, s_mapping.len() + turret_index)); } } //no need to queue once we have been shot by a turret if is_g_prime && turret_visible { continue; } /* debug!( "Viewing {} dist {} seen turret? {} turret visible? {}", loc, dist, seen_turret, turret_visible );*/ for dir in DIRECTIONS.iter() { let new_loc = loc.clone() + dir; if let Some(tile) = grid.get_value(&new_loc) { if *tile == Building { continue; } let newLocIndex = (new_loc.data[0] * grid.C as i64 + new_loc.data[1]) as usize; if visited[newLocIndex] { continue; } visited.set(newLocIndex, true); let new_dist = dist + 1; if new_dist > M { continue; } let new_seen_turret = seen_turret || turret_visible; queue.push_back((new_loc, new_dist, new_seen_turret)); } } } } debug!("Built graph from\n{}\n S={} T={}", grid, S, T); G } impl Display for Grid<Tile> { fn fmt(&self, f: &mut Formatter) -> fmt::Result { for r in 0..self.R { for c in 0..self.C { if let Err(err) = write!(f, "{}", self[(r, c)])
{ return Err(err); }
conditional_block
d.rs
::*; use crate::util::codejam::run_cases; impl Tile { fn to_char(self) -> char { match self { Empty => '.', Building => '#', Soldier => 'S', Turret => 'T', } } } impl From<char> for Tile { fn from(item: char) -> Self { match item { '.' => Empty, '#' => Building, 'S' => Soldier, 'T' => Turret, _ => panic!("Character not recognized: {}", item), } } } impl Display for Tile { fn fmt(&self, f: &mut Formatter) -> fmt::Result { write!(f, "{}", self.to_char()) } } impl Default for Tile { fn default() -> Tile { Empty } } //problem specific code fn reachable(grid: &Grid<Tile>, location: &GridCoord) -> HashSet<GridRowColVec> { let mut r = HashSet::new(); //debug!("\nTracing {} starting at {}", location, direction); for direction in DIRECTIONS.iter() { let mut loc: GridRowColVec = location.convert(); for _ in 0..=grid.R + grid.C { loc += direction; if let Some(tile) = grid.get_value(&loc) { match *tile { Building => { break; } _ => { r.insert(loc.clone()); } }; } else { break; } } } r } /* impl<L, R> FromIterator<(L, R)> for BiMap<L, R> { fn from_iter<I: IntoIterator<Item = (L, R)>>(iter: I) -> Self { let mut c = BiMap::new(); for i in iter { c.insert(i.0, i.1); } c } }*/ fn
<'a>(case_no: u32, grid: &mut Grid<Tile>, M_soldier_limit: usize) -> String { debug!( "Solving case {}\nM={}\n{}\n", case_no, M_soldier_limit, grid ); //original solider & turret index to location map let S_map = grid .filter_by_val(&Soldier) .enumerate() .collect::<BiMap<_, _>>(); let turret_locations = grid.filter_by_val(&Turret).collect::<Vec<_>>(); //precalucate what squares a turret can reach let turret_reachable_squares_list = turret_locations .iter() .map(|t_loc| reachable(&grid, &t_loc)) .collect::<Vec<_>>(); let T_map = turret_locations .into_iter() .enumerate() .collect::<BiMap<_, _>>(); let S = grid.filter_by_val(&Soldier).count(); let T = grid.filter_by_val(&Turret).count(); //Construct the initial Graph let G_edges = build_graph( &grid, false, M_soldier_limit, &S_map, &T_map, &turret_reachable_squares_list, ); let mut G = FlowGraph::new(2 + S + T, 4); for uv in G_edges { G.add_edge(uv.0, uv.1, 1, 1); } let source = S + T; let sink = S + T + 1; let vertex_to_string = |v: usize| match v { s if s < S => format!("Soldier #{} ({:?})", s + 1, *S_map.get_by_left(&s).unwrap()), t if t >= S && t < S + T => format!( "Turret #{} ({:?})", t - S + 1, *T_map.get_by_left(&(t - S)).unwrap() ), v if v == sink => "Sink".to_string(), _source => "Source".to_string(), }; //BFS for each soldier //will be in left to right order, then top down order //Now find max matching of G (G has an edge from soldier s to turret t if and only if soldier s can destroy turret t after all other turrets have been destroyed) for s in 0..S { G.add_edge(source, s, 1, 1); } for t in S..S + T { G.add_edge(t, sink, 1, 1); } let (R, flow) = G.dinic(source, sink); let mut ans = format!("Case #{}: {}\n", case_no, R); //Compute initial matching let mut M = flow .iter() .enumerate() .filter(|&(_e, f)| *f > 0) //map to u->v .map(|(e, _f)| (G.graph.endp[e ^ 1], G.graph.endp[e])) //leave out source and sink nodes .filter(|&(u, v)| u != source && v != sink) .collect::<Vec<_>>(); debug!( "Edges in M initial matching=\n{}\n", M.iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); let mut r = R; while r > 0 { //Let us define the graph G' with the same nodes as G, but an edge between soldier s and turret t only exists in G' if s can destroy t with the other turrets active let Gprime = build_graph( &grid, true, M_soldier_limit, &S_map, &T_map, &turret_reachable_squares_list, ); //Now build graph H let mut H = Graph::new(S + T, 4); let soldiers_in_m = M.iter().map(|&(s, _t)| s).collect::<Vec<_>>(); for &(s, t) in Gprime.iter() { if soldiers_in_m.contains(&s) { H.add_edge(s, t); } } for &(s, t) in M.iter() { H.add_edge(t, s); } debug!( "Current matching M =\n{}\n", M.iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); debug!( "Edges in G'=\n{}\n", Gprime .iter() .map(|&(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); debug!( "Edges in H=\n{}\n", H.edges() .map(|(u, v)| format!("{}->{}", vertex_to_string(u), vertex_to_string(v))) .collect::<Vec<_>>() .join("\n") ); let turrets_in_M = M.iter().map(|&(_s, t)| t).collect::<Vec<_>>(); //find an edge (s,t') where t' is not in m let st_prime = Gprime.iter().find(|&(_s, t)| !turrets_in_M.contains(t)); if st_prime.is_some() { let &(s, t) = st_prime.unwrap(); debug!("Found (s,t') s={} t'={}", s, t - S); ans += &format!("{} {}\n", s + 1, t - S + 1); grid[S_map.get_by_left(&s).unwrap()] = Empty; grid[T_map.get_by_left(&(t - S)).unwrap()] = Empty; r -= 1; //Also remove from current matching let to_remove = M .iter() .position(|&(s_in_m, _t)| s_in_m == s) .expect("Soldier should be in mapping"); M.remove(to_remove); continue; } //Now we need to find a cycle //Start at a soldier in H let soldier_in_h = H.edges().filter(|&(u, _v)| u <= S).next().unwrap().0; let mut cycle_edges = VecDeque::new(); let mut edge = ( soldier_in_h, H.adj_list_with_edges(soldier_in_h).next().unwrap().1, ); let mut visited = BitVec::from_elem(H.num_v(), false); while !visited[edge.0] { visited.set(edge.0, true); cycle_edges.push_back(edge); debug!( "pushed Edge {:?} ", format!("{}->{}", vertex_to_string(edge.0), vertex_to_string(edge.1)) ); //adj list returns an (internal edge index, next vertex) edge = (edge.1, H.adj_list_with_edges(edge.1).next().unwrap().1); debug!("Edge {:?} ", edge); } //cut to the actual cycle found let cycle_end = cycle_edges.back().unwrap().1; let cycle_start = cycle_edges .iter() .position
solve
identifier_name
resource_fusion_sec_azure.go
.3" ) var fusionSECAzureTemplateTags = []string{ "Microsoft.Network/loadBalancers", "Microsoft.ManagedIdentity/userAssignedIdentities", } var fusionSECAzureParams = []interface{}{ "fusionSECName", "location", "loadBalancerNetworkRg", "loadBalancerNetworkName", "loadBalancerSubnet", } var renamedFusionSECAzureParams = map[string]string{} var fusionSECAzureTFOutputs = []string{ "applicationName", "managedResourceGroupName", "hmvip0", "hmvip1", "loadBalancerFullIdentityId", } func
() *schema.Resource { return &schema.Resource{ CreateContext: resourceFusionSECAzureCreate, ReadContext: resourceFusionSECAzureRead, UpdateContext: resourceFusionSECAzureUpdate, DeleteContext: resourceFusionSECAzureDelete, Schema: map[string]*schema.Schema{ "resource_group_name": { Type: schema.TypeString, Required: true, ForceNew: true, ValidateFunc: validateAzureResourceGroupName, }, "location": { Type: schema.TypeString, Required: true, }, // parameters "fusion_sec_name": { Description: "The name of the Fusion Storage Endpoint Collection (SEC). 0-59 alphanumeric characters only.", Type: schema.TypeString, Required: true, ValidateFunc: validateAzureManagedApplicationName, }, "load_balancer_network_rg": { Type: schema.TypeString, Required: true, }, "load_balancer_network_name": { Type: schema.TypeString, Required: true, }, "load_balancer_subnet": { Type: schema.TypeString, Required: true, }, "jit_approval_group_object_ids": { Description: "This is a list of Azure group object IDs for people who are allowed to approve JIT requests", Required: true, Type: schema.TypeList, Elem: &schema.Schema{ Type: schema.TypeString, ValidateFunc: validation.IsUUID, }, }, "plan": { Type: schema.TypeList, Optional: true, MaxItems: 1, Elem: &schema.Resource{ Schema: map[string]*schema.Schema{ "name": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, "product": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, "publisher": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, "version": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, }, }, }, "tags": { Type: schema.TypeMap, Optional: true, Elem: &schema.Schema{ Type: schema.TypeString, ValidateFunc: validation.StringIsNotEmpty, }, }, // Outputs "application_name": { Type: schema.TypeString, Computed: true, }, "managed_resource_group_name": { Type: schema.TypeString, Computed: true, }, "hmvip0": { Type: schema.TypeString, Computed: true, }, "hmvip1": { Type: schema.TypeString, Computed: true, }, "load_balancer_full_identity_id": { Type: schema.TypeString, Computed: true, }, }, Timeouts: &schema.ResourceTimeout{ Create: schema.DefaultTimeout(30 * time.Minute), Read: schema.DefaultTimeout(5 * time.Minute), Delete: schema.DefaultTimeout(30 * time.Minute), }, } } func resourceFusionSECAzureCreate(ctx context.Context, d *schema.ResourceData, m interface{}) (returnedDiags diag.Diagnostics) { tflog.Trace(ctx, "resourceFusionSECAzurereate") azureClient, diags := m.(*CbsService).azureClientService(ctx) if diags.HasError() { return diags } name := d.Get("fusion_sec_name").(string) managedResourceGroup := toAzureManagedResourceGroup(name) resourceGroupName := d.Get("resource_group_name").(string) if d.IsNewResource() { existing, err := azureClient.AppsGet(ctx, resourceGroupName, name) if err != nil { if !responseWasNotFound(existing.Response) { return diag.Errorf("failed to check for presence of existing Managed Application Name %q (Resource Group %q): %+v", name, resourceGroupName, err) } } if existing.ID != nil && *existing.ID != "" { return diag.Errorf( "A resource with the name %q, Resource Group %q and ID %q already exists - to be managed via Terraform this resource needs to be imported into the State.", name, resourceGroupName, *existing.ID, ) } } parameters := managedapplications.Application{ Location: to.StringPtr(d.Get("location").(string)), } targetResourceGroupId := fmt.Sprintf("/subscriptions/%s/resourceGroups/%s", azureClient.SubscriptionID(), managedResourceGroup) parameters.ApplicationProperties = &managedapplications.ApplicationProperties{ ManagedResourceGroupID: to.StringPtr(targetResourceGroupId), } parameters.Kind = to.StringPtr("MarketPlace") if v, ok1 := d.GetOk("plan"); ok1 && len(v.([]interface{})) > 0 { parameters.Plan = expandPlan(v.([]interface{})) } else { parameters.Plan = &managedapplications.Plan{ Name: to.StringPtr(defaultFusionSECPlanName), Product: to.StringPtr(defaultFusionSECPlanPublisher), Publisher: to.StringPtr(defaultFusionSECPlanPublisher), Version: to.StringPtr(defaultFusionSECPlanVersion), } } parameters.Parameters = make(map[string]interface{}) setAppParameter := func(key string, value interface{}) { (parameters.Parameters.(map[string]interface{}))[key] = map[string]interface{}{"value": value} } for _, value := range fusionSECAzureParams { valueStr := value.(string) setAppParameter(valueStr, d.Get(templateToTFParam(valueStr, renamedFusionSECAzureParams))) } returnedDiags = setAzureJitAccessPolicy(&parameters, d) if v, ok := d.GetOk("tags"); ok { tags := v.(map[string]interface{}) tagsMap := make(map[string]interface{}) for _, tag := range fusionSECAzureTemplateTags { tagsMap[tag] = tags } setAppParameter("tagsByResource", tagsMap) } // Error out now, before we create resources if returnedDiags.HasError() { return returnedDiags } tflog.Trace(ctx, "resourceFusionSECAzureCreate AppsCreateOrUpdate") err := azureClient.AppsCreateOrUpdate(ctx, resourceGroupName, name, parameters) defer func() { if returnedDiags.HasError() { if err = azureClient.AppsDelete(ctx, resourceGroupName, name); err != nil { tflog.Error( ctx, fmt.Sprintf( "failed to delete Managed Application %q (Resource Group %q) after failed CreateOrUpdate operation: %+v", name, resourceGroupName, err, ), ) } } }() if err != nil { return diag.FromErr(err) } resp, err := azureClient.AppsGet(ctx, resourceGroupName, name) if err != nil { return diag.Errorf("failed to retrieve Managed Application %q (Resource Group %q): %+v", name, resourceGroupName, err) } if resp.ID == nil || *resp.ID == "" { return diag.Errorf("cannot read Managed Application %q (Resource Group %q) ID", name, resourceGroupName) } d.SetId(*resp.ID) diags = resourceFusionSECAzureRead(ctx, d, m) if diags.HasError() { returnedDiags = append(returnedDiags, diags...) } return returnedDiags } func resourceFusionSECAzureRead(ctx context.Context, d *schema.ResourceData, m interface{}) diag.Diagnostics { tflog.Trace(ctx, "resourceFusionSECAzureRead") azureClient, diags := m.(*CbsService).azureClientService(ctx) if diags.HasError() { return diags } v, ok := d.GetOk("fusion_sec_name") if !ok { log.Printf("[WARN] No Managed Application found with Id %q, removing from state", d.Id()) d.SetId("") return nil } appName := v.(string) managedResourceGroup := toAzureManagedResourceGroup(appName) resourceGroup := d.Get("resource_group_name").(string) resp, err := azureClient.AppsGet(ctx, resourceGroup, appName) if
resourceFusionSECAzure
identifier_name
resource_fusion_sec_azure.go
3" ) var fusionSECAzureTemplateTags = []string{ "Microsoft.Network/loadBalancers", "Microsoft.ManagedIdentity/userAssignedIdentities", } var fusionSECAzureParams = []interface{}{ "fusionSECName", "location", "loadBalancerNetworkRg", "loadBalancerNetworkName", "loadBalancerSubnet", } var renamedFusionSECAzureParams = map[string]string{} var fusionSECAzureTFOutputs = []string{ "applicationName", "managedResourceGroupName", "hmvip0", "hmvip1", "loadBalancerFullIdentityId", } func resourceFusionSECAzure() *schema.Resource { return &schema.Resource{ CreateContext: resourceFusionSECAzureCreate, ReadContext: resourceFusionSECAzureRead, UpdateContext: resourceFusionSECAzureUpdate, DeleteContext: resourceFusionSECAzureDelete, Schema: map[string]*schema.Schema{ "resource_group_name": { Type: schema.TypeString, Required: true, ForceNew: true, ValidateFunc: validateAzureResourceGroupName, }, "location": { Type: schema.TypeString, Required: true, }, // parameters "fusion_sec_name": { Description: "The name of the Fusion Storage Endpoint Collection (SEC). 0-59 alphanumeric characters only.", Type: schema.TypeString, Required: true, ValidateFunc: validateAzureManagedApplicationName, }, "load_balancer_network_rg": { Type: schema.TypeString, Required: true, }, "load_balancer_network_name": { Type: schema.TypeString, Required: true, }, "load_balancer_subnet": { Type: schema.TypeString, Required: true, }, "jit_approval_group_object_ids": { Description: "This is a list of Azure group object IDs for people who are allowed to approve JIT requests", Required: true, Type: schema.TypeList, Elem: &schema.Schema{ Type: schema.TypeString, ValidateFunc: validation.IsUUID, }, }, "plan": { Type: schema.TypeList, Optional: true, MaxItems: 1, Elem: &schema.Resource{ Schema: map[string]*schema.Schema{ "name": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, "product": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, "publisher": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, "version": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, }, }, }, "tags": { Type: schema.TypeMap, Optional: true, Elem: &schema.Schema{ Type: schema.TypeString, ValidateFunc: validation.StringIsNotEmpty, }, }, // Outputs "application_name": { Type: schema.TypeString, Computed: true, }, "managed_resource_group_name": { Type: schema.TypeString, Computed: true, }, "hmvip0": { Type: schema.TypeString, Computed: true, }, "hmvip1": { Type: schema.TypeString, Computed: true, }, "load_balancer_full_identity_id": { Type: schema.TypeString, Computed: true, }, }, Timeouts: &schema.ResourceTimeout{ Create: schema.DefaultTimeout(30 * time.Minute), Read: schema.DefaultTimeout(5 * time.Minute), Delete: schema.DefaultTimeout(30 * time.Minute), }, } } func resourceFusionSECAzureCreate(ctx context.Context, d *schema.ResourceData, m interface{}) (returnedDiags diag.Diagnostics) { tflog.Trace(ctx, "resourceFusionSECAzurereate") azureClient, diags := m.(*CbsService).azureClientService(ctx) if diags.HasError() { return diags } name := d.Get("fusion_sec_name").(string) managedResourceGroup := toAzureManagedResourceGroup(name) resourceGroupName := d.Get("resource_group_name").(string) if d.IsNewResource() { existing, err := azureClient.AppsGet(ctx, resourceGroupName, name) if err != nil { if !responseWasNotFound(existing.Response) { return diag.Errorf("failed to check for presence of existing Managed Application Name %q (Resource Group %q): %+v", name, resourceGroupName, err) } } if existing.ID != nil && *existing.ID != "" { return diag.Errorf( "A resource with the name %q, Resource Group %q and ID %q already exists - to be managed via Terraform this resource needs to be imported into the State.", name, resourceGroupName, *existing.ID, ) } } parameters := managedapplications.Application{ Location: to.StringPtr(d.Get("location").(string)), } targetResourceGroupId := fmt.Sprintf("/subscriptions/%s/resourceGroups/%s", azureClient.SubscriptionID(), managedResourceGroup) parameters.ApplicationProperties = &managedapplications.ApplicationProperties{ ManagedResourceGroupID: to.StringPtr(targetResourceGroupId), } parameters.Kind = to.StringPtr("MarketPlace") if v, ok1 := d.GetOk("plan"); ok1 && len(v.([]interface{})) > 0 { parameters.Plan = expandPlan(v.([]interface{})) } else { parameters.Plan = &managedapplications.Plan{ Name: to.StringPtr(defaultFusionSECPlanName), Product: to.StringPtr(defaultFusionSECPlanPublisher), Publisher: to.StringPtr(defaultFusionSECPlanPublisher), Version: to.StringPtr(defaultFusionSECPlanVersion), } } parameters.Parameters = make(map[string]interface{}) setAppParameter := func(key string, value interface{}) { (parameters.Parameters.(map[string]interface{}))[key] = map[string]interface{}{"value": value} } for _, value := range fusionSECAzureParams
returnedDiags = setAzureJitAccessPolicy(&parameters, d) if v, ok := d.GetOk("tags"); ok { tags := v.(map[string]interface{}) tagsMap := make(map[string]interface{}) for _, tag := range fusionSECAzureTemplateTags { tagsMap[tag] = tags } setAppParameter("tagsByResource", tagsMap) } // Error out now, before we create resources if returnedDiags.HasError() { return returnedDiags } tflog.Trace(ctx, "resourceFusionSECAzureCreate AppsCreateOrUpdate") err := azureClient.AppsCreateOrUpdate(ctx, resourceGroupName, name, parameters) defer func() { if returnedDiags.HasError() { if err = azureClient.AppsDelete(ctx, resourceGroupName, name); err != nil { tflog.Error( ctx, fmt.Sprintf( "failed to delete Managed Application %q (Resource Group %q) after failed CreateOrUpdate operation: %+v", name, resourceGroupName, err, ), ) } } }() if err != nil { return diag.FromErr(err) } resp, err := azureClient.AppsGet(ctx, resourceGroupName, name) if err != nil { return diag.Errorf("failed to retrieve Managed Application %q (Resource Group %q): %+v", name, resourceGroupName, err) } if resp.ID == nil || *resp.ID == "" { return diag.Errorf("cannot read Managed Application %q (Resource Group %q) ID", name, resourceGroupName) } d.SetId(*resp.ID) diags = resourceFusionSECAzureRead(ctx, d, m) if diags.HasError() { returnedDiags = append(returnedDiags, diags...) } return returnedDiags } func resourceFusionSECAzureRead(ctx context.Context, d *schema.ResourceData, m interface{}) diag.Diagnostics { tflog.Trace(ctx, "resourceFusionSECAzureRead") azureClient, diags := m.(*CbsService).azureClientService(ctx) if diags.HasError() { return diags } v, ok := d.GetOk("fusion_sec_name") if !ok { log.Printf("[WARN] No Managed Application found with Id %q, removing from state", d.Id()) d.SetId("") return nil } appName := v.(string) managedResourceGroup := toAzureManagedResourceGroup(appName) resourceGroup := d.Get("resource_group_name").(string) resp, err := azureClient.AppsGet(ctx, resourceGroup, appName) if
{ valueStr := value.(string) setAppParameter(valueStr, d.Get(templateToTFParam(valueStr, renamedFusionSECAzureParams))) }
conditional_block
resource_fusion_sec_azure.go
.3" ) var fusionSECAzureTemplateTags = []string{ "Microsoft.Network/loadBalancers", "Microsoft.ManagedIdentity/userAssignedIdentities", } var fusionSECAzureParams = []interface{}{ "fusionSECName", "location", "loadBalancerNetworkRg", "loadBalancerNetworkName", "loadBalancerSubnet", } var renamedFusionSECAzureParams = map[string]string{} var fusionSECAzureTFOutputs = []string{ "applicationName", "managedResourceGroupName", "hmvip0", "hmvip1", "loadBalancerFullIdentityId", } func resourceFusionSECAzure() *schema.Resource
"fusion_sec_name": { Description: "The name of the Fusion Storage Endpoint Collection (SEC). 0-59 alphanumeric characters only.", Type: schema.TypeString, Required: true, ValidateFunc: validateAzureManagedApplicationName, }, "load_balancer_network_rg": { Type: schema.TypeString, Required: true, }, "load_balancer_network_name": { Type: schema.TypeString, Required: true, }, "load_balancer_subnet": { Type: schema.TypeString, Required: true, }, "jit_approval_group_object_ids": { Description: "This is a list of Azure group object IDs for people who are allowed to approve JIT requests", Required: true, Type: schema.TypeList, Elem: &schema.Schema{ Type: schema.TypeString, ValidateFunc: validation.IsUUID, }, }, "plan": { Type: schema.TypeList, Optional: true, MaxItems: 1, Elem: &schema.Resource{ Schema: map[string]*schema.Schema{ "name": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, "product": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, "publisher": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, "version": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, }, }, }, "tags": { Type: schema.TypeMap, Optional: true, Elem: &schema.Schema{ Type: schema.TypeString, ValidateFunc: validation.StringIsNotEmpty, }, }, // Outputs "application_name": { Type: schema.TypeString, Computed: true, }, "managed_resource_group_name": { Type: schema.TypeString, Computed: true, }, "hmvip0": { Type: schema.TypeString, Computed: true, }, "hmvip1": { Type: schema.TypeString, Computed: true, }, "load_balancer_full_identity_id": { Type: schema.TypeString, Computed: true, }, }, Timeouts: &schema.ResourceTimeout{ Create: schema.DefaultTimeout(30 * time.Minute), Read: schema.DefaultTimeout(5 * time.Minute), Delete: schema.DefaultTimeout(30 * time.Minute), }, } } func resourceFusionSECAzureCreate(ctx context.Context, d *schema.ResourceData, m interface{}) (returnedDiags diag.Diagnostics) { tflog.Trace(ctx, "resourceFusionSECAzurereate") azureClient, diags := m.(*CbsService).azureClientService(ctx) if diags.HasError() { return diags } name := d.Get("fusion_sec_name").(string) managedResourceGroup := toAzureManagedResourceGroup(name) resourceGroupName := d.Get("resource_group_name").(string) if d.IsNewResource() { existing, err := azureClient.AppsGet(ctx, resourceGroupName, name) if err != nil { if !responseWasNotFound(existing.Response) { return diag.Errorf("failed to check for presence of existing Managed Application Name %q (Resource Group %q): %+v", name, resourceGroupName, err) } } if existing.ID != nil && *existing.ID != "" { return diag.Errorf( "A resource with the name %q, Resource Group %q and ID %q already exists - to be managed via Terraform this resource needs to be imported into the State.", name, resourceGroupName, *existing.ID, ) } } parameters := managedapplications.Application{ Location: to.StringPtr(d.Get("location").(string)), } targetResourceGroupId := fmt.Sprintf("/subscriptions/%s/resourceGroups/%s", azureClient.SubscriptionID(), managedResourceGroup) parameters.ApplicationProperties = &managedapplications.ApplicationProperties{ ManagedResourceGroupID: to.StringPtr(targetResourceGroupId), } parameters.Kind = to.StringPtr("MarketPlace") if v, ok1 := d.GetOk("plan"); ok1 && len(v.([]interface{})) > 0 { parameters.Plan = expandPlan(v.([]interface{})) } else { parameters.Plan = &managedapplications.Plan{ Name: to.StringPtr(defaultFusionSECPlanName), Product: to.StringPtr(defaultFusionSECPlanPublisher), Publisher: to.StringPtr(defaultFusionSECPlanPublisher), Version: to.StringPtr(defaultFusionSECPlanVersion), } } parameters.Parameters = make(map[string]interface{}) setAppParameter := func(key string, value interface{}) { (parameters.Parameters.(map[string]interface{}))[key] = map[string]interface{}{"value": value} } for _, value := range fusionSECAzureParams { valueStr := value.(string) setAppParameter(valueStr, d.Get(templateToTFParam(valueStr, renamedFusionSECAzureParams))) } returnedDiags = setAzureJitAccessPolicy(&parameters, d) if v, ok := d.GetOk("tags"); ok { tags := v.(map[string]interface{}) tagsMap := make(map[string]interface{}) for _, tag := range fusionSECAzureTemplateTags { tagsMap[tag] = tags } setAppParameter("tagsByResource", tagsMap) } // Error out now, before we create resources if returnedDiags.HasError() { return returnedDiags } tflog.Trace(ctx, "resourceFusionSECAzureCreate AppsCreateOrUpdate") err := azureClient.AppsCreateOrUpdate(ctx, resourceGroupName, name, parameters) defer func() { if returnedDiags.HasError() { if err = azureClient.AppsDelete(ctx, resourceGroupName, name); err != nil { tflog.Error( ctx, fmt.Sprintf( "failed to delete Managed Application %q (Resource Group %q) after failed CreateOrUpdate operation: %+v", name, resourceGroupName, err, ), ) } } }() if err != nil { return diag.FromErr(err) } resp, err := azureClient.AppsGet(ctx, resourceGroupName, name) if err != nil { return diag.Errorf("failed to retrieve Managed Application %q (Resource Group %q): %+v", name, resourceGroupName, err) } if resp.ID == nil || *resp.ID == "" { return diag.Errorf("cannot read Managed Application %q (Resource Group %q) ID", name, resourceGroupName) } d.SetId(*resp.ID) diags = resourceFusionSECAzureRead(ctx, d, m) if diags.HasError() { returnedDiags = append(returnedDiags, diags...) } return returnedDiags } func resourceFusionSECAzureRead(ctx context.Context, d *schema.ResourceData, m interface{}) diag.Diagnostics { tflog.Trace(ctx, "resourceFusionSECAzureRead") azureClient, diags := m.(*CbsService).azureClientService(ctx) if diags.HasError() { return diags } v, ok := d.GetOk("fusion_sec_name") if !ok { log.Printf("[WARN] No Managed Application found with Id %q, removing from state", d.Id()) d.SetId("") return nil } appName := v.(string) managedResourceGroup := toAzureManagedResourceGroup(appName) resourceGroup := d.Get("resource_group_name").(string) resp, err := azureClient.AppsGet(ctx, resourceGroup, appName) if err
{ return &schema.Resource{ CreateContext: resourceFusionSECAzureCreate, ReadContext: resourceFusionSECAzureRead, UpdateContext: resourceFusionSECAzureUpdate, DeleteContext: resourceFusionSECAzureDelete, Schema: map[string]*schema.Schema{ "resource_group_name": { Type: schema.TypeString, Required: true, ForceNew: true, ValidateFunc: validateAzureResourceGroupName, }, "location": { Type: schema.TypeString, Required: true, }, // parameters
identifier_body
resource_fusion_sec_azure.go
.3" ) var fusionSECAzureTemplateTags = []string{ "Microsoft.Network/loadBalancers", "Microsoft.ManagedIdentity/userAssignedIdentities", } var fusionSECAzureParams = []interface{}{ "fusionSECName", "location", "loadBalancerNetworkRg", "loadBalancerNetworkName", "loadBalancerSubnet", } var renamedFusionSECAzureParams = map[string]string{} var fusionSECAzureTFOutputs = []string{ "applicationName", "managedResourceGroupName", "hmvip0", "hmvip1", "loadBalancerFullIdentityId", } func resourceFusionSECAzure() *schema.Resource { return &schema.Resource{ CreateContext: resourceFusionSECAzureCreate, ReadContext: resourceFusionSECAzureRead, UpdateContext: resourceFusionSECAzureUpdate, DeleteContext: resourceFusionSECAzureDelete, Schema: map[string]*schema.Schema{ "resource_group_name": { Type: schema.TypeString, Required: true, ForceNew: true, ValidateFunc: validateAzureResourceGroupName, }, "location": { Type: schema.TypeString, Required: true, }, // parameters "fusion_sec_name": { Description: "The name of the Fusion Storage Endpoint Collection (SEC). 0-59 alphanumeric characters only.", Type: schema.TypeString, Required: true, ValidateFunc: validateAzureManagedApplicationName, }, "load_balancer_network_rg": { Type: schema.TypeString, Required: true, }, "load_balancer_network_name": { Type: schema.TypeString, Required: true, }, "load_balancer_subnet": {
Description: "This is a list of Azure group object IDs for people who are allowed to approve JIT requests", Required: true, Type: schema.TypeList, Elem: &schema.Schema{ Type: schema.TypeString, ValidateFunc: validation.IsUUID, }, }, "plan": { Type: schema.TypeList, Optional: true, MaxItems: 1, Elem: &schema.Resource{ Schema: map[string]*schema.Schema{ "name": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, "product": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, "publisher": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, "version": { Type: schema.TypeString, Required: true, ValidateFunc: validation.StringIsNotEmpty, }, }, }, }, "tags": { Type: schema.TypeMap, Optional: true, Elem: &schema.Schema{ Type: schema.TypeString, ValidateFunc: validation.StringIsNotEmpty, }, }, // Outputs "application_name": { Type: schema.TypeString, Computed: true, }, "managed_resource_group_name": { Type: schema.TypeString, Computed: true, }, "hmvip0": { Type: schema.TypeString, Computed: true, }, "hmvip1": { Type: schema.TypeString, Computed: true, }, "load_balancer_full_identity_id": { Type: schema.TypeString, Computed: true, }, }, Timeouts: &schema.ResourceTimeout{ Create: schema.DefaultTimeout(30 * time.Minute), Read: schema.DefaultTimeout(5 * time.Minute), Delete: schema.DefaultTimeout(30 * time.Minute), }, } } func resourceFusionSECAzureCreate(ctx context.Context, d *schema.ResourceData, m interface{}) (returnedDiags diag.Diagnostics) { tflog.Trace(ctx, "resourceFusionSECAzurereate") azureClient, diags := m.(*CbsService).azureClientService(ctx) if diags.HasError() { return diags } name := d.Get("fusion_sec_name").(string) managedResourceGroup := toAzureManagedResourceGroup(name) resourceGroupName := d.Get("resource_group_name").(string) if d.IsNewResource() { existing, err := azureClient.AppsGet(ctx, resourceGroupName, name) if err != nil { if !responseWasNotFound(existing.Response) { return diag.Errorf("failed to check for presence of existing Managed Application Name %q (Resource Group %q): %+v", name, resourceGroupName, err) } } if existing.ID != nil && *existing.ID != "" { return diag.Errorf( "A resource with the name %q, Resource Group %q and ID %q already exists - to be managed via Terraform this resource needs to be imported into the State.", name, resourceGroupName, *existing.ID, ) } } parameters := managedapplications.Application{ Location: to.StringPtr(d.Get("location").(string)), } targetResourceGroupId := fmt.Sprintf("/subscriptions/%s/resourceGroups/%s", azureClient.SubscriptionID(), managedResourceGroup) parameters.ApplicationProperties = &managedapplications.ApplicationProperties{ ManagedResourceGroupID: to.StringPtr(targetResourceGroupId), } parameters.Kind = to.StringPtr("MarketPlace") if v, ok1 := d.GetOk("plan"); ok1 && len(v.([]interface{})) > 0 { parameters.Plan = expandPlan(v.([]interface{})) } else { parameters.Plan = &managedapplications.Plan{ Name: to.StringPtr(defaultFusionSECPlanName), Product: to.StringPtr(defaultFusionSECPlanPublisher), Publisher: to.StringPtr(defaultFusionSECPlanPublisher), Version: to.StringPtr(defaultFusionSECPlanVersion), } } parameters.Parameters = make(map[string]interface{}) setAppParameter := func(key string, value interface{}) { (parameters.Parameters.(map[string]interface{}))[key] = map[string]interface{}{"value": value} } for _, value := range fusionSECAzureParams { valueStr := value.(string) setAppParameter(valueStr, d.Get(templateToTFParam(valueStr, renamedFusionSECAzureParams))) } returnedDiags = setAzureJitAccessPolicy(&parameters, d) if v, ok := d.GetOk("tags"); ok { tags := v.(map[string]interface{}) tagsMap := make(map[string]interface{}) for _, tag := range fusionSECAzureTemplateTags { tagsMap[tag] = tags } setAppParameter("tagsByResource", tagsMap) } // Error out now, before we create resources if returnedDiags.HasError() { return returnedDiags } tflog.Trace(ctx, "resourceFusionSECAzureCreate AppsCreateOrUpdate") err := azureClient.AppsCreateOrUpdate(ctx, resourceGroupName, name, parameters) defer func() { if returnedDiags.HasError() { if err = azureClient.AppsDelete(ctx, resourceGroupName, name); err != nil { tflog.Error( ctx, fmt.Sprintf( "failed to delete Managed Application %q (Resource Group %q) after failed CreateOrUpdate operation: %+v", name, resourceGroupName, err, ), ) } } }() if err != nil { return diag.FromErr(err) } resp, err := azureClient.AppsGet(ctx, resourceGroupName, name) if err != nil { return diag.Errorf("failed to retrieve Managed Application %q (Resource Group %q): %+v", name, resourceGroupName, err) } if resp.ID == nil || *resp.ID == "" { return diag.Errorf("cannot read Managed Application %q (Resource Group %q) ID", name, resourceGroupName) } d.SetId(*resp.ID) diags = resourceFusionSECAzureRead(ctx, d, m) if diags.HasError() { returnedDiags = append(returnedDiags, diags...) } return returnedDiags } func resourceFusionSECAzureRead(ctx context.Context, d *schema.ResourceData, m interface{}) diag.Diagnostics { tflog.Trace(ctx, "resourceFusionSECAzureRead") azureClient, diags := m.(*CbsService).azureClientService(ctx) if diags.HasError() { return diags } v, ok := d.GetOk("fusion_sec_name") if !ok { log.Printf("[WARN] No Managed Application found with Id %q, removing from state", d.Id()) d.SetId("") return nil } appName := v.(string) managedResourceGroup := toAzureManagedResourceGroup(appName) resourceGroup := d.Get("resource_group_name").(string) resp, err := azureClient.AppsGet(ctx, resourceGroup, appName) if err
Type: schema.TypeString, Required: true, }, "jit_approval_group_object_ids": {
random_line_split
iter.go
: invalid whence") } if abs < 0 { return 0, fmt.Errorf("norm: negative position") } if int(abs) >= i.rb.nsrc { i.setDone() return int64(i.p), nil } i.p = int(abs) i.multiSeg = nil i.next = i.rb.f.nextMain i.info = i.rb.f.info(i.rb.src, i.p) i.rb.ss.first(i.info) return abs, nil } // returnSlice returns a slice of the underlying input type as a byte slice. // If the underlying is of type []byte, it will simply return a slice. // If the underlying is of type string, it will copy the slice to the buffer // and return that. func (i *Iter) returnSlice(a, b int) []byte { if i.rb.src.bytes == nil { return i.buf[:copy(i.buf[:], i.rb.src.str[a:b])] } return i.rb.src.bytes[a:b] } // Pos returns the byte position at which the next call to Next will commence processing. func (i *Iter) Pos() int { return i.p } func (i *Iter) setDone() { i.next = nextDone i.p = i.rb.nsrc } // Done returns true if there is no more input to process. func (i *Iter) Done() bool { return i.p >= i.rb.nsrc } // Next returns f(i.input[i.Pos():n]), where n is a boundary of i.input. // For any input a and b for which f(a) == f(b), subsequent calls // to Next will return the same segments. // Modifying runes are grouped together with the preceding starter, if such a starter exists. // Although not guaranteed, n will typically be the smallest possible n. func (i *Iter) Next() []byte { return i.next(i) } func nextASCIIBytes(i *Iter) []byte { p := i.p + 1 if p >= i.rb.nsrc { p0 := i.p i.setDone() return i.rb.src.bytes[p0:p] } if i.rb.src.bytes[p] < utf8.RuneSelf { p0 := i.p i.p = p return i.rb.src.bytes[p0:p] } i.info = i.rb.f.info(i.rb.src, i.p) i.next = i.rb.f.nextMain return i.next(i) } func nextASCIIString(i *Iter) []byte { p := i.p + 1 if p >= i.rb.nsrc { i.buf[0] = i.rb.src.str[i.p] i.setDone() return i.buf[:1] } if i.rb.src.str[p] < utf8.RuneSelf { i.buf[0] = i.rb.src.str[i.p] i.p = p return i.buf[:1] } i.info = i.rb.f.info(i.rb.src, i.p) i.next = i.rb.f.nextMain return i.next(i) } func nextHangul(i *Iter) []byte { p := i.p next := p + hangulUTF8Size if next >= i.rb.nsrc { i.setDone() } else if i.rb.src.hangul(next) == 0 { i.rb.ss.next(i.info) i.info = i.rb.f.info(i.rb.src, i.p) i.next = i.rb.f.nextMain return i.next(i) } i.p = next return i.buf[:decomposeHangul(i.buf[:], i.rb.src.hangul(p))] } func nextDone(i *Iter) []byte { return nil } // nextMulti is used for iterating over multi-segment decompositions // for decomposing normal forms. func nextMulti(i *Iter) []byte { j := 0 d := i.multiSeg // skip first rune for j = 1; j < len(d) && !utf8.RuneStart(d[j]); j++ { } for j < len(d) { info := i.rb.f.info(input{bytes: d}, j) if info.BoundaryBefore() { i.multiSeg = d[j:] return d[:j] } j += int(info.size) } // treat last segment as normal decomposition i.next = i.rb.f.nextMain return i.next(i) } // nextMultiNorm is used for iterating over multi-segment decompositions // for composing normal forms. func nextMultiNorm(i *Iter) []byte { j := 0 d := i.multiSeg for j < len(d) { info := i.rb.f.info(input{bytes: d}, j) if info.BoundaryBefore() { i.rb.compose() seg := i.buf[:i.rb.flushCopy(i.buf[:])] i.rb.insertUnsafe(input{bytes: d}, j, info) i.multiSeg = d[j+int(info.size):] return seg } i.rb.insertUnsafe(input{bytes: d}, j, info) j += int(info.size) } i.multiSeg = nil i.next = nextComposed return doNormComposed(i) } // nextDecomposed is the implementation of Next for forms NFD and NFKD. func nextDecomposed(i *Iter) (next []byte) { outp := 0 inCopyStart, outCopyStart := i.p, 0 for { if sz := int(i.info.size); sz <= 1 { i.rb.ss = 0 p := i.p i.p++ // ASCII or illegal byte. Either way, advance by 1. if i.p >= i.rb.nsrc { i.setDone() return i.returnSlice(p, i.p) } else if i.rb.src._byte(i.p) < utf8.RuneSelf { i.next = i.asciiF return i.returnSlice(p, i.p) } outp++ } else if d := i.info.Decomposition(); d != nil { // Note: If leading CCC != 0, then len(d) == 2 and last is also non-zero. // Case 1: there is a leftover to copy. In this case the decomposition // must begin with a modifier and should always be appended. // Case 2: no leftover. Simply return d if followed by a ccc == 0 value. p := outp + len(d) if outp > 0 { i.rb.src.copySlice(i.buf[outCopyStart:], inCopyStart, i.p) // TODO: this condition should not be possible, but we leave it // in for defensive purposes. if p > len(i.buf) { return i.buf[:outp] } } else if i.info.multiSegment() { // outp must be 0 as multi-segment decompositions always // start a new segment. if i.multiSeg == nil { i.multiSeg = d i.next = nextMulti return nextMulti(i) } // We are in the last segment. Treat as normal decomposition. d = i.multiSeg i.multiSeg = nil p = len(d) } prevCC := i.info.tccc if i.p += sz; i.p >= i.rb.nsrc { i.setDone() i.info = Properties{} // Force BoundaryBefore to succeed. } else { i.info = i.rb.f.info(i.rb.src, i.p) } switch i.rb.ss.next(i.info) { case ssOverflow: i.next = nextCGJDecompose fallthrough case ssStarter: if outp > 0 { copy(i.buf[outp:], d) return i.buf[:p] } return d } copy(i.buf[outp:], d) outp = p inCopyStart, outCopyStart = i.p, outp if i.info.ccc < prevCC { goto doNorm } continue } else if r := i.rb.src.hangul(i.p); r != 0 { outp = decomposeHangul(i.buf[:], r) i.p += hangulUTF8Size inCopyStart, outCopyStart = i.p, outp if i.p >= i.rb.nsrc { i.setDone() break } else if i.rb.src.hangul(i.p) != 0 { i.next = nextHangul return i.buf[:outp] } } else { p := outp + sz if p > len(i.buf) { break } outp = p i.p += sz } if i.p >= i.rb.nsrc { i.setDone() break } prevCC := i.info.tccc i.info = i.rb.f.info(i.rb.src, i.p) if v := i.rb.ss.next(i.info); v == ssStarter { break } else if v == ssOverflow { i.next = nextCGJDecompose break } if i.info.ccc < prevCC { goto doNorm } } if outCopyStart == 0
{ return i.returnSlice(inCopyStart, i.p) }
conditional_block
iter.go
(i *Iter) Seek(offset int64, whence int) (int64, error) { var abs int64 switch whence { case 0: abs = offset case 1: abs = int64(i.p) + offset case 2: abs = int64(i.rb.nsrc) + offset default: return 0, fmt.Errorf("norm: invalid whence") } if abs < 0 { return 0, fmt.Errorf("norm: negative position") } if int(abs) >= i.rb.nsrc { i.setDone() return int64(i.p), nil } i.p = int(abs) i.multiSeg = nil i.next = i.rb.f.nextMain i.info = i.rb.f.info(i.rb.src, i.p) i.rb.ss.first(i.info) return abs, nil } // returnSlice returns a slice of the underlying input type as a byte slice. // If the underlying is of type []byte, it will simply return a slice. // If the underlying is of type string, it will copy the slice to the buffer // and return that. func (i *Iter) returnSlice(a, b int) []byte { if i.rb.src.bytes == nil { return i.buf[:copy(i.buf[:], i.rb.src.str[a:b])] } return i.rb.src.bytes[a:b] } // Pos returns the byte position at which the next call to Next will commence processing. func (i *Iter) Pos() int { return i.p } func (i *Iter)
() { i.next = nextDone i.p = i.rb.nsrc } // Done returns true if there is no more input to process. func (i *Iter) Done() bool { return i.p >= i.rb.nsrc } // Next returns f(i.input[i.Pos():n]), where n is a boundary of i.input. // For any input a and b for which f(a) == f(b), subsequent calls // to Next will return the same segments. // Modifying runes are grouped together with the preceding starter, if such a starter exists. // Although not guaranteed, n will typically be the smallest possible n. func (i *Iter) Next() []byte { return i.next(i) } func nextASCIIBytes(i *Iter) []byte { p := i.p + 1 if p >= i.rb.nsrc { p0 := i.p i.setDone() return i.rb.src.bytes[p0:p] } if i.rb.src.bytes[p] < utf8.RuneSelf { p0 := i.p i.p = p return i.rb.src.bytes[p0:p] } i.info = i.rb.f.info(i.rb.src, i.p) i.next = i.rb.f.nextMain return i.next(i) } func nextASCIIString(i *Iter) []byte { p := i.p + 1 if p >= i.rb.nsrc { i.buf[0] = i.rb.src.str[i.p] i.setDone() return i.buf[:1] } if i.rb.src.str[p] < utf8.RuneSelf { i.buf[0] = i.rb.src.str[i.p] i.p = p return i.buf[:1] } i.info = i.rb.f.info(i.rb.src, i.p) i.next = i.rb.f.nextMain return i.next(i) } func nextHangul(i *Iter) []byte { p := i.p next := p + hangulUTF8Size if next >= i.rb.nsrc { i.setDone() } else if i.rb.src.hangul(next) == 0 { i.rb.ss.next(i.info) i.info = i.rb.f.info(i.rb.src, i.p) i.next = i.rb.f.nextMain return i.next(i) } i.p = next return i.buf[:decomposeHangul(i.buf[:], i.rb.src.hangul(p))] } func nextDone(i *Iter) []byte { return nil } // nextMulti is used for iterating over multi-segment decompositions // for decomposing normal forms. func nextMulti(i *Iter) []byte { j := 0 d := i.multiSeg // skip first rune for j = 1; j < len(d) && !utf8.RuneStart(d[j]); j++ { } for j < len(d) { info := i.rb.f.info(input{bytes: d}, j) if info.BoundaryBefore() { i.multiSeg = d[j:] return d[:j] } j += int(info.size) } // treat last segment as normal decomposition i.next = i.rb.f.nextMain return i.next(i) } // nextMultiNorm is used for iterating over multi-segment decompositions // for composing normal forms. func nextMultiNorm(i *Iter) []byte { j := 0 d := i.multiSeg for j < len(d) { info := i.rb.f.info(input{bytes: d}, j) if info.BoundaryBefore() { i.rb.compose() seg := i.buf[:i.rb.flushCopy(i.buf[:])] i.rb.insertUnsafe(input{bytes: d}, j, info) i.multiSeg = d[j+int(info.size):] return seg } i.rb.insertUnsafe(input{bytes: d}, j, info) j += int(info.size) } i.multiSeg = nil i.next = nextComposed return doNormComposed(i) } // nextDecomposed is the implementation of Next for forms NFD and NFKD. func nextDecomposed(i *Iter) (next []byte) { outp := 0 inCopyStart, outCopyStart := i.p, 0 for { if sz := int(i.info.size); sz <= 1 { i.rb.ss = 0 p := i.p i.p++ // ASCII or illegal byte. Either way, advance by 1. if i.p >= i.rb.nsrc { i.setDone() return i.returnSlice(p, i.p) } else if i.rb.src._byte(i.p) < utf8.RuneSelf { i.next = i.asciiF return i.returnSlice(p, i.p) } outp++ } else if d := i.info.Decomposition(); d != nil { // Note: If leading CCC != 0, then len(d) == 2 and last is also non-zero. // Case 1: there is a leftover to copy. In this case the decomposition // must begin with a modifier and should always be appended. // Case 2: no leftover. Simply return d if followed by a ccc == 0 value. p := outp + len(d) if outp > 0 { i.rb.src.copySlice(i.buf[outCopyStart:], inCopyStart, i.p) // TODO: this condition should not be possible, but we leave it // in for defensive purposes. if p > len(i.buf) { return i.buf[:outp] } } else if i.info.multiSegment() { // outp must be 0 as multi-segment decompositions always // start a new segment. if i.multiSeg == nil { i.multiSeg = d i.next = nextMulti return nextMulti(i) } // We are in the last segment. Treat as normal decomposition. d = i.multiSeg i.multiSeg = nil p = len(d) } prevCC := i.info.tccc if i.p += sz; i.p >= i.rb.nsrc { i.setDone() i.info = Properties{} // Force BoundaryBefore to succeed. } else { i.info = i.rb.f.info(i.rb.src, i.p) } switch i.rb.ss.next(i.info) { case ssOverflow: i.next = nextCGJDecompose fallthrough case ssStarter: if outp > 0 { copy(i.buf[outp:], d) return i.buf[:p] } return d } copy(i.buf[outp:], d) outp = p inCopyStart, outCopyStart = i.p, outp if i.info.ccc < prevCC { goto doNorm } continue } else if r := i.rb.src.hangul(i.p); r != 0 { outp = decomposeHangul(i.buf[:], r) i.p += hangulUTF8Size inCopyStart, outCopyStart = i.p, outp if i.p >= i.rb.nsrc { i.setDone() break } else if i.rb.src.hangul(i.p) != 0 { i.next = nextHangul return i.buf[:outp] } } else { p := outp + sz if p > len(i.buf) { break } outp = p i.p += sz } if i.p >= i.rb.nsrc { i.setDone() break } prevCC := i.info.tccc i.info = i.rb.f.info(i.rb.src, i.p)
setDone
identifier_name
iter.go
(i *Iter) Seek(offset int64, whence int) (int64, error) { var abs int64 switch whence { case 0: abs = offset case 1: abs = int64(i.p) + offset case 2: abs = int64(i.rb.nsrc) + offset default: return 0, fmt.Errorf("norm: invalid whence") } if abs < 0 { return 0, fmt.Errorf("norm: negative position") } if int(abs) >= i.rb.nsrc { i.setDone() return int64(i.p), nil } i.p = int(abs) i.multiSeg = nil i.next = i.rb.f.nextMain i.info = i.rb.f.info(i.rb.src, i.p) i.rb.ss.first(i.info) return abs, nil } // returnSlice returns a slice of the underlying input type as a byte slice. // If the underlying is of type []byte, it will simply return a slice. // If the underlying is of type string, it will copy the slice to the buffer // and return that. func (i *Iter) returnSlice(a, b int) []byte { if i.rb.src.bytes == nil { return i.buf[:copy(i.buf[:], i.rb.src.str[a:b])] } return i.rb.src.bytes[a:b] } // Pos returns the byte position at which the next call to Next will commence processing. func (i *Iter) Pos() int { return i.p } func (i *Iter) setDone() { i.next = nextDone i.p = i.rb.nsrc } // Done returns true if there is no more input to process. func (i *Iter) Done() bool { return i.p >= i.rb.nsrc } // Next returns f(i.input[i.Pos():n]), where n is a boundary of i.input. // For any input a and b for which f(a) == f(b), subsequent calls // to Next will return the same segments. // Modifying runes are grouped together with the preceding starter, if such a starter exists. // Although not guaranteed, n will typically be the smallest possible n. func (i *Iter) Next() []byte { return i.next(i) } func nextASCIIBytes(i *Iter) []byte { p := i.p + 1 if p >= i.rb.nsrc { p0 := i.p i.setDone() return i.rb.src.bytes[p0:p] } if i.rb.src.bytes[p] < utf8.RuneSelf { p0 := i.p i.p = p return i.rb.src.bytes[p0:p] } i.info = i.rb.f.info(i.rb.src, i.p) i.next = i.rb.f.nextMain return i.next(i) } func nextASCIIString(i *Iter) []byte { p := i.p + 1 if p >= i.rb.nsrc { i.buf[0] = i.rb.src.str[i.p] i.setDone() return i.buf[:1] } if i.rb.src.str[p] < utf8.RuneSelf { i.buf[0] = i.rb.src.str[i.p] i.p = p return i.buf[:1] } i.info = i.rb.f.info(i.rb.src, i.p) i.next = i.rb.f.nextMain return i.next(i) } func nextHangul(i *Iter) []byte { p := i.p next := p + hangulUTF8Size if next >= i.rb.nsrc { i.setDone() } else if i.rb.src.hangul(next) == 0 { i.rb.ss.next(i.info) i.info = i.rb.f.info(i.rb.src, i.p) i.next = i.rb.f.nextMain return i.next(i) } i.p = next return i.buf[:decomposeHangul(i.buf[:], i.rb.src.hangul(p))] } func nextDone(i *Iter) []byte { return nil } // nextMulti is used for iterating over multi-segment decompositions // for decomposing normal forms. func nextMulti(i *Iter) []byte { j := 0 d := i.multiSeg // skip first rune for j = 1; j < len(d) && !utf8.RuneStart(d[j]); j++ { } for j < len(d) { info := i.rb.f.info(input{bytes: d}, j) if info.BoundaryBefore() { i.multiSeg = d[j:] return d[:j] } j += int(info.size) } // treat last segment as normal decomposition i.next = i.rb.f.nextMain return i.next(i) } // nextMultiNorm is used for iterating over multi-segment decompositions // for composing normal forms. func nextMultiNorm(i *Iter) []byte { j := 0 d := i.multiSeg for j < len(d) { info := i.rb.f.info(input{bytes: d}, j) if info.BoundaryBefore() { i.rb.compose() seg := i.buf[:i.rb.flushCopy(i.buf[:])] i.rb.insertUnsafe(input{bytes: d}, j, info) i.multiSeg = d[j+int(info.size):] return seg } i.rb.insertUnsafe(input{bytes: d}, j, info) j += int(info.size) } i.multiSeg = nil i.next = nextComposed return doNormComposed(i) } // nextDecomposed is the implementation of Next for forms NFD and NFKD. func nextDecomposed(i *Iter) (next []byte) { outp := 0 inCopyStart, outCopyStart := i.p, 0 for { if sz := int(i.info.size); sz <= 1 { i.rb.ss = 0 p := i.p i.p++ // ASCII or illegal byte. Either way, advance by 1.
if i.p >= i.rb.nsrc { i.setDone() return i.returnSlice(p, i.p) } else if i.rb.src._byte(i.p) < utf8.RuneSelf { i.next = i.asciiF return i.returnSlice(p, i.p) } outp++ } else if d := i.info.Decomposition(); d != nil { // Note: If leading CCC != 0, then len(d) == 2 and last is also non-zero. // Case 1: there is a leftover to copy. In this case the decomposition // must begin with a modifier and should always be appended. // Case 2: no leftover. Simply return d if followed by a ccc == 0 value. p := outp + len(d) if outp > 0 { i.rb.src.copySlice(i.buf[outCopyStart:], inCopyStart, i.p) // TODO: this condition should not be possible, but we leave it // in for defensive purposes. if p > len(i.buf) { return i.buf[:outp] } } else if i.info.multiSegment() { // outp must be 0 as multi-segment decompositions always // start a new segment. if i.multiSeg == nil { i.multiSeg = d i.next = nextMulti return nextMulti(i) } // We are in the last segment. Treat as normal decomposition. d = i.multiSeg i.multiSeg = nil p = len(d) } prevCC := i.info.tccc if i.p += sz; i.p >= i.rb.nsrc { i.setDone() i.info = Properties{} // Force BoundaryBefore to succeed. } else { i.info = i.rb.f.info(i.rb.src, i.p) } switch i.rb.ss.next(i.info) { case ssOverflow: i.next = nextCGJDecompose fallthrough case ssStarter: if outp > 0 { copy(i.buf[outp:], d) return i.buf[:p] } return d } copy(i.buf[outp:], d) outp = p inCopyStart, outCopyStart = i.p, outp if i.info.ccc < prevCC { goto doNorm } continue } else if r := i.rb.src.hangul(i.p); r != 0 { outp = decomposeHangul(i.buf[:], r) i.p += hangulUTF8Size inCopyStart, outCopyStart = i.p, outp if i.p >= i.rb.nsrc { i.setDone() break } else if i.rb.src.hangul(i.p) != 0 { i.next = nextHangul return i.buf[:outp] } } else { p := outp + sz if p > len(i.buf) { break } outp = p i.p += sz } if i.p >= i.rb.nsrc { i.setDone() break } prevCC := i.info.tccc i.info = i.rb.f.info(i.rb.src, i.p) if
random_line_split
iter.go
(i *Iter) Seek(offset int64, whence int) (int64, error) { var abs int64 switch whence { case 0: abs = offset case 1: abs = int64(i.p) + offset case 2: abs = int64(i.rb.nsrc) + offset default: return 0, fmt.Errorf("norm: invalid whence") } if abs < 0 { return 0, fmt.Errorf("norm: negative position") } if int(abs) >= i.rb.nsrc { i.setDone() return int64(i.p), nil } i.p = int(abs) i.multiSeg = nil i.next = i.rb.f.nextMain i.info = i.rb.f.info(i.rb.src, i.p) i.rb.ss.first(i.info) return abs, nil } // returnSlice returns a slice of the underlying input type as a byte slice. // If the underlying is of type []byte, it will simply return a slice. // If the underlying is of type string, it will copy the slice to the buffer // and return that. func (i *Iter) returnSlice(a, b int) []byte { if i.rb.src.bytes == nil { return i.buf[:copy(i.buf[:], i.rb.src.str[a:b])] } return i.rb.src.bytes[a:b] } // Pos returns the byte position at which the next call to Next will commence processing. func (i *Iter) Pos() int { return i.p } func (i *Iter) setDone() { i.next = nextDone i.p = i.rb.nsrc } // Done returns true if there is no more input to process. func (i *Iter) Done() bool { return i.p >= i.rb.nsrc } // Next returns f(i.input[i.Pos():n]), where n is a boundary of i.input. // For any input a and b for which f(a) == f(b), subsequent calls // to Next will return the same segments. // Modifying runes are grouped together with the preceding starter, if such a starter exists. // Although not guaranteed, n will typically be the smallest possible n. func (i *Iter) Next() []byte { return i.next(i) } func nextASCIIBytes(i *Iter) []byte { p := i.p + 1 if p >= i.rb.nsrc { p0 := i.p i.setDone() return i.rb.src.bytes[p0:p] } if i.rb.src.bytes[p] < utf8.RuneSelf { p0 := i.p i.p = p return i.rb.src.bytes[p0:p] } i.info = i.rb.f.info(i.rb.src, i.p) i.next = i.rb.f.nextMain return i.next(i) } func nextASCIIString(i *Iter) []byte { p := i.p + 1 if p >= i.rb.nsrc { i.buf[0] = i.rb.src.str[i.p] i.setDone() return i.buf[:1] } if i.rb.src.str[p] < utf8.RuneSelf { i.buf[0] = i.rb.src.str[i.p] i.p = p return i.buf[:1] } i.info = i.rb.f.info(i.rb.src, i.p) i.next = i.rb.f.nextMain return i.next(i) } func nextHangul(i *Iter) []byte { p := i.p next := p + hangulUTF8Size if next >= i.rb.nsrc { i.setDone() } else if i.rb.src.hangul(next) == 0 { i.rb.ss.next(i.info) i.info = i.rb.f.info(i.rb.src, i.p) i.next = i.rb.f.nextMain return i.next(i) } i.p = next return i.buf[:decomposeHangul(i.buf[:], i.rb.src.hangul(p))] } func nextDone(i *Iter) []byte { return nil } // nextMulti is used for iterating over multi-segment decompositions // for decomposing normal forms. func nextMulti(i *Iter) []byte
// nextMultiNorm is used for iterating over multi-segment decompositions // for composing normal forms. func nextMultiNorm(i *Iter) []byte { j := 0 d := i.multiSeg for j < len(d) { info := i.rb.f.info(input{bytes: d}, j) if info.BoundaryBefore() { i.rb.compose() seg := i.buf[:i.rb.flushCopy(i.buf[:])] i.rb.insertUnsafe(input{bytes: d}, j, info) i.multiSeg = d[j+int(info.size):] return seg } i.rb.insertUnsafe(input{bytes: d}, j, info) j += int(info.size) } i.multiSeg = nil i.next = nextComposed return doNormComposed(i) } // nextDecomposed is the implementation of Next for forms NFD and NFKD. func nextDecomposed(i *Iter) (next []byte) { outp := 0 inCopyStart, outCopyStart := i.p, 0 for { if sz := int(i.info.size); sz <= 1 { i.rb.ss = 0 p := i.p i.p++ // ASCII or illegal byte. Either way, advance by 1. if i.p >= i.rb.nsrc { i.setDone() return i.returnSlice(p, i.p) } else if i.rb.src._byte(i.p) < utf8.RuneSelf { i.next = i.asciiF return i.returnSlice(p, i.p) } outp++ } else if d := i.info.Decomposition(); d != nil { // Note: If leading CCC != 0, then len(d) == 2 and last is also non-zero. // Case 1: there is a leftover to copy. In this case the decomposition // must begin with a modifier and should always be appended. // Case 2: no leftover. Simply return d if followed by a ccc == 0 value. p := outp + len(d) if outp > 0 { i.rb.src.copySlice(i.buf[outCopyStart:], inCopyStart, i.p) // TODO: this condition should not be possible, but we leave it // in for defensive purposes. if p > len(i.buf) { return i.buf[:outp] } } else if i.info.multiSegment() { // outp must be 0 as multi-segment decompositions always // start a new segment. if i.multiSeg == nil { i.multiSeg = d i.next = nextMulti return nextMulti(i) } // We are in the last segment. Treat as normal decomposition. d = i.multiSeg i.multiSeg = nil p = len(d) } prevCC := i.info.tccc if i.p += sz; i.p >= i.rb.nsrc { i.setDone() i.info = Properties{} // Force BoundaryBefore to succeed. } else { i.info = i.rb.f.info(i.rb.src, i.p) } switch i.rb.ss.next(i.info) { case ssOverflow: i.next = nextCGJDecompose fallthrough case ssStarter: if outp > 0 { copy(i.buf[outp:], d) return i.buf[:p] } return d } copy(i.buf[outp:], d) outp = p inCopyStart, outCopyStart = i.p, outp if i.info.ccc < prevCC { goto doNorm } continue } else if r := i.rb.src.hangul(i.p); r != 0 { outp = decomposeHangul(i.buf[:], r) i.p += hangulUTF8Size inCopyStart, outCopyStart = i.p, outp if i.p >= i.rb.nsrc { i.setDone() break } else if i.rb.src.hangul(i.p) != 0 { i.next = nextHangul return i.buf[:outp] } } else { p := outp + sz if p > len(i.buf) { break } outp = p i.p += sz } if i.p >= i.rb.nsrc { i.setDone() break } prevCC := i.info.tccc i.info = i.rb.f.info(i.rb.src, i.p)
{ j := 0 d := i.multiSeg // skip first rune for j = 1; j < len(d) && !utf8.RuneStart(d[j]); j++ { } for j < len(d) { info := i.rb.f.info(input{bytes: d}, j) if info.BoundaryBefore() { i.multiSeg = d[j:] return d[:j] } j += int(info.size) } // treat last segment as normal decomposition i.next = i.rb.f.nextMain return i.next(i) }
identifier_body
consulta-ine.page.ts
/app/herramientas/imagen'; import { Camera, CameraOptions } from '@ionic-native/camera/ngx'; import { LoadingService } from 'src/app/services/loading.service'; import { Cliente } from '../../tipo-identificacion/consulta-similitud-confirmacion/model/Cliente.model'; import { JsonPersonalData } from 'src/app/services/actividades/model/json-personal-data.model'; import { JsonOperationData } from 'src/app/services/actividades/model/json-operation-data.model'; import { JsonInnerData } from 'src/app/services/actividades/model/json-inner-data.model'; import { ActivatedRoute } from '@angular/router'; @Component({ selector: 'app-consulta-ine', templateUrl: './consulta-ine.page.html', styleUrls: ['./consulta-ine.page.scss'], }) export class ConsultaInePage implements OnInit { cardValid: boolean; browser: any; esCargando: boolean; capturasINE: string; frontImg: string; isenabled: boolean; isValidoSpinnerFront: boolean; frontImg2: any; secuenciaId: number; client: ClientData; pData: JsonPersonalData; validINE: string; constructor( private iab: InAppBrowser, private alertCtrl: AlertController, private navCtrl: NavController, private documentosService: DocumentosService, private oauth: OauthService, private login: LoginService, private loading: LoadingService, private activityService: ActivitiesService, private saveS: GuardarStorageService, private activitiesService: ActivitiesService, private route: ActivatedRoute, public camera: Camera) { this.route.queryParams.subscribe(params => { if (params) { this.pData = JSON.parse(params.client); this.pData.observations = this.validINE; } }); this.cardValid = true; this.validINE = 'Identificación válida'; this.secuenciaId = 0; if (this.saveS.getTipoFlujo() === "alhajas") { this.secuenciaId = 4; } else { this.secuenciaId = 9; } } ngOnInit() { } getImagenFront() { this.isValidoSpinnerFront = true; const options: CameraOptions = { quality: 70, destinationType: this.camera.DestinationType.DATA_URL, encodingType: this.camera.EncodingType.JPEG, mediaType: this.camera.MediaType.PICTURE }; this.camera.getPicture(options).then((imageData) => { this.frontImg = 'data:image/jpeg;base64,' + imageData; this.frontImg = imageData; this.saveS.guardarStorageImagenF(this.frontImg); if (this.frontImg) { // enable the button this.isenabled = true; this.isValidoSpinnerFront = false; } else { // disable the button this.isenabled = false; this.isValidoSpinnerFront = false; } }, (err) => { // Handle error console.log('Camera issue:' + err, ); }); this.isValidoSpinnerFront = false; } openB() { this.browser = this.iab.create('https://listanominal.ine.mx', '_system'); } openGallery() { const cameraOptions = { sourceType: this.camera.PictureSourceType.PHOTOLIBRARY, destinationType: this.camera.DestinationType.DATA_URL, encodingType: this.camera.EncodingType.JPEG, quality: 70, targetWidth: 1000, targetHeight: 1000, correctOrientation: true }; /* this.camera.getPicture(cameraOptions) .then(file_uri => this.frontImg = file_uri, err => console.log(err)); */ this.camera.getPicture(cameraOptions).then((imageData) => { this.frontImg = 'data:image/jpeg;base64,' + imageData; this.frontImg2 = imageData; this.saveS.guardarStorageImagenF(this.frontImg); if (this.frontImg) { // enable the button this.isenabled = true; this.isValidoSpinnerFront = false; } else { // disable the button this.isenabled = false; this.isValidoSpinnerFront = false; } }, (err) => { // Handle error console.log('Camera issue:' + err); }); this.isValidoSpinnerFront = false; } goCallBack() { console.log('Typescript callback has been called'); this.browser.hide(); } ch
event): void { let mensajeError = ''; const file: File = $event.target.files[0]; console.log('changeListenerF'); if ((file.type !== 'image/jpeg' && file.type !== 'image/png') || (file.size > 1000000)) { mensajeError = 'Formato y/o tamaño de imagen incorrecto'; } else { const myReader: FileReader = new FileReader(); myReader.onloadend = (e) => { this.frontImg = myReader.result.toString(); console.log('frontImg'); console.log(this.frontImg); // this.saveS.guardarStorageImagenF(this.capturasINE); // this.imgCapturada.emit(capturas); }; myReader.readAsDataURL(file); } if (this.frontImg) { // enable the button this.isenabled = true; this.isValidoSpinnerFront = false; } else { // disable the button this.isenabled = false; this.isValidoSpinnerFront = false; } } goCargarDocumento() { try { const imagen = new Imagen(); const blobAnverso = imagen.convertirImagenEnBlob(this.frontImg); console.log('blobAnverso'); console.log(blobAnverso); if (blobAnverso) { this.cargarDocumento(blobAnverso, this.saveS.getBearerToken()); } else { alert('Imagen Invalida'); } } catch (error) { console.log(error); } } cargarDocumento(fileAnverso: any, bearerToken: string) { this.loading.present('Cargando...'); // this.actualizarActivity('EN PROCESO'); this.esCargando = true; const date = new Date(); const dataFile1 = new DataFile( 'bid:Anverso', 'Nombre', 'Primer apellido', 'Segundo apellido', '123549', date.toISOString(), 'RES_BURO', '123123'); const jsonRequest = new JsonRequest('IDOFA', this.saveS.getOperationID(), 'OK', '', ''); this.documentosService.cargarDocumento(jsonRequest, dataFile1, fileAnverso, bearerToken).subscribe( (respuesta: any) => { console.log('cargarDocumento respuesta', respuesta); if (respuesta['resultOK']) { // this.actualizarActivity('FINALIZADO'); this.esCargando = false; this.loading.dismiss(); // alert('Archivo guardado con éxito'); // this.navCtrl.navigateRoot('info-grales'); this.pData.observations=this.validINE; this.guardarDatos(this.pData); } else { // alert(respuesta['message']); this.loading.dismiss(); this.esCargando = false; } }, (error: HttpErrorResponse) => { this.loading.dismiss(); console.log(error); switch (error['status']) { case 401: alert('Es necesario iniciar session, nuemente para continuar'); this.navCtrl.navigateRoot('login'); break; case 404: alert('Es necesario iniciar session, nuemente para continuar'); this.navCtrl.navigateRoot('login'); break; case 500: alert('Por favor, reintentar para continuar'); this.cargarDocumento(fileAnverso, bearerToken); break; case 501: alert('Por favor, reintentar para continuar'); this.cargarDocumento(fileAnverso, bearerToken); break; default: alert('Es necesario iniciar session, nuemente para continuar'); this.navCtrl.navigateRoot('login'); break; } // alert('Hubo un error al enviar los datos, intenta de nuevo'); }); } actualizarActivity(estatus: string) { const productId = 1; const jsonData = new JsonData( productId, this.saveS.getSystemCode(), estatus, '1', '', this.secuenciaId, 1, this.saveS.getPersonId()); // {'id': this.saveS.getPersonId(), 'observations': this.validINE.toString()} console.log('jsonData validINE :'); console.log(jsonData); const jsonMetaData = new JsonMetadata(0, '', 0, 0, 1, 1); /* const jsonData = { 'data': { 'productId': 1, 'code': '', 'activityStatus': 'FINALIZADO', 'activityValue': null, 'data': '{"personal_data":{"id":' + this.saveS.getPersonId() + ',"observations":"' + this.validINE + '}}', 'secuence': 17, 'workflowId': 1, 'personId': this.saveS.getPersonId() }, 'metadata': { 'accuracy': 0, 'deviceInfo': '', 'latut
angeListenerINE($
identifier_name
consulta-ine.page.ts
/app/herramientas/imagen'; import { Camera, CameraOptions } from '@ionic-native/camera/ngx'; import { LoadingService } from 'src/app/services/loading.service'; import { Cliente } from '../../tipo-identificacion/consulta-similitud-confirmacion/model/Cliente.model'; import { JsonPersonalData } from 'src/app/services/actividades/model/json-personal-data.model'; import { JsonOperationData } from 'src/app/services/actividades/model/json-operation-data.model'; import { JsonInnerData } from 'src/app/services/actividades/model/json-inner-data.model'; import { ActivatedRoute } from '@angular/router'; @Component({ selector: 'app-consulta-ine', templateUrl: './consulta-ine.page.html', styleUrls: ['./consulta-ine.page.scss'], }) export class ConsultaInePage implements OnInit { cardValid: boolean; browser: any; esCargando: boolean; capturasINE: string; frontImg: string; isenabled: boolean; isValidoSpinnerFront: boolean; frontImg2: any; secuenciaId: number; client: ClientData; pData: JsonPersonalData; validINE: string; constructor( private iab: InAppBrowser, private alertCtrl: AlertController, private navCtrl: NavController, private documentosService: DocumentosService, private oauth: OauthService, private login: LoginService, private loading: LoadingService, private activityService: ActivitiesService, private saveS: GuardarStorageService, private activitiesService: ActivitiesService, private route: ActivatedRoute, public camera: Camera) { this.route.queryParams.subscribe(params => { if (params) { this.pData = JSON.parse(params.client); this.pData.observations = this.validINE; } }); this.cardValid = true; this.validINE = 'Identificación válida'; this.secuenciaId = 0; if (this.saveS.getTipoFlujo() === "alhajas") { this.secuenciaId = 4; } else { this.secuenciaId = 9; } } ngOnInit() { } getImagenFront() { this.isValidoSpinnerFront = true; const options: CameraOptions = { quality: 70, destinationType: this.camera.DestinationType.DATA_URL, encodingType: this.camera.EncodingType.JPEG, mediaType: this.camera.MediaType.PICTURE }; this.camera.getPicture(options).then((imageData) => { this.frontImg = 'data:image/jpeg;base64,' + imageData; this.frontImg = imageData; this.saveS.guardarStorageImagenF(this.frontImg); if (this.frontImg) { // enable the button this.isenabled = true; this.isValidoSpinnerFront = false; } else { // disable the button this.isenabled = false; this.isValidoSpinnerFront = false; } }, (err) => { // Handle error console.log('Camera issue:' + err, ); }); this.isValidoSpinnerFront = false; } openB() { this.browser = this.iab.create('https://listanominal.ine.mx', '_system'); } openGallery() { const cameraOptions = { sourceType: this.camera.PictureSourceType.PHOTOLIBRARY, destinationType: this.camera.DestinationType.DATA_URL, encodingType: this.camera.EncodingType.JPEG, quality: 70, targetWidth: 1000, targetHeight: 1000,
/* this.camera.getPicture(cameraOptions) .then(file_uri => this.frontImg = file_uri, err => console.log(err)); */ this.camera.getPicture(cameraOptions).then((imageData) => { this.frontImg = 'data:image/jpeg;base64,' + imageData; this.frontImg2 = imageData; this.saveS.guardarStorageImagenF(this.frontImg); if (this.frontImg) { // enable the button this.isenabled = true; this.isValidoSpinnerFront = false; } else { // disable the button this.isenabled = false; this.isValidoSpinnerFront = false; } }, (err) => { // Handle error console.log('Camera issue:' + err); }); this.isValidoSpinnerFront = false; } goCallBack() { console.log('Typescript callback has been called'); this.browser.hide(); } changeListenerINE($event): void { let mensajeError = ''; const file: File = $event.target.files[0]; console.log('changeListenerF'); if ((file.type !== 'image/jpeg' && file.type !== 'image/png') || (file.size > 1000000)) { mensajeError = 'Formato y/o tamaño de imagen incorrecto'; } else { const myReader: FileReader = new FileReader(); myReader.onloadend = (e) => { this.frontImg = myReader.result.toString(); console.log('frontImg'); console.log(this.frontImg); // this.saveS.guardarStorageImagenF(this.capturasINE); // this.imgCapturada.emit(capturas); }; myReader.readAsDataURL(file); } if (this.frontImg) { // enable the button this.isenabled = true; this.isValidoSpinnerFront = false; } else { // disable the button this.isenabled = false; this.isValidoSpinnerFront = false; } } goCargarDocumento() { try { const imagen = new Imagen(); const blobAnverso = imagen.convertirImagenEnBlob(this.frontImg); console.log('blobAnverso'); console.log(blobAnverso); if (blobAnverso) { this.cargarDocumento(blobAnverso, this.saveS.getBearerToken()); } else { alert('Imagen Invalida'); } } catch (error) { console.log(error); } } cargarDocumento(fileAnverso: any, bearerToken: string) { this.loading.present('Cargando...'); // this.actualizarActivity('EN PROCESO'); this.esCargando = true; const date = new Date(); const dataFile1 = new DataFile( 'bid:Anverso', 'Nombre', 'Primer apellido', 'Segundo apellido', '123549', date.toISOString(), 'RES_BURO', '123123'); const jsonRequest = new JsonRequest('IDOFA', this.saveS.getOperationID(), 'OK', '', ''); this.documentosService.cargarDocumento(jsonRequest, dataFile1, fileAnverso, bearerToken).subscribe( (respuesta: any) => { console.log('cargarDocumento respuesta', respuesta); if (respuesta['resultOK']) { // this.actualizarActivity('FINALIZADO'); this.esCargando = false; this.loading.dismiss(); // alert('Archivo guardado con éxito'); // this.navCtrl.navigateRoot('info-grales'); this.pData.observations=this.validINE; this.guardarDatos(this.pData); } else { // alert(respuesta['message']); this.loading.dismiss(); this.esCargando = false; } }, (error: HttpErrorResponse) => { this.loading.dismiss(); console.log(error); switch (error['status']) { case 401: alert('Es necesario iniciar session, nuemente para continuar'); this.navCtrl.navigateRoot('login'); break; case 404: alert('Es necesario iniciar session, nuemente para continuar'); this.navCtrl.navigateRoot('login'); break; case 500: alert('Por favor, reintentar para continuar'); this.cargarDocumento(fileAnverso, bearerToken); break; case 501: alert('Por favor, reintentar para continuar'); this.cargarDocumento(fileAnverso, bearerToken); break; default: alert('Es necesario iniciar session, nuemente para continuar'); this.navCtrl.navigateRoot('login'); break; } // alert('Hubo un error al enviar los datos, intenta de nuevo'); }); } actualizarActivity(estatus: string) { const productId = 1; const jsonData = new JsonData( productId, this.saveS.getSystemCode(), estatus, '1', '', this.secuenciaId, 1, this.saveS.getPersonId()); // {'id': this.saveS.getPersonId(), 'observations': this.validINE.toString()} console.log('jsonData validINE :'); console.log(jsonData); const jsonMetaData = new JsonMetadata(0, '', 0, 0, 1, 1); /* const jsonData = { 'data': { 'productId': 1, 'code': '', 'activityStatus': 'FINALIZADO', 'activityValue': null, 'data': '{"personal_data":{"id":' + this.saveS.getPersonId() + ',"observations":"' + this.validINE + '}}', 'secuence': 17, 'workflowId': 1, 'personId': this.saveS.getPersonId() }, 'metadata': { 'accuracy': 0, 'deviceInfo': '', 'latutide
correctOrientation: true };
random_line_split
consulta-ine.page.ts
/app/herramientas/imagen'; import { Camera, CameraOptions } from '@ionic-native/camera/ngx'; import { LoadingService } from 'src/app/services/loading.service'; import { Cliente } from '../../tipo-identificacion/consulta-similitud-confirmacion/model/Cliente.model'; import { JsonPersonalData } from 'src/app/services/actividades/model/json-personal-data.model'; import { JsonOperationData } from 'src/app/services/actividades/model/json-operation-data.model'; import { JsonInnerData } from 'src/app/services/actividades/model/json-inner-data.model'; import { ActivatedRoute } from '@angular/router'; @Component({ selector: 'app-consulta-ine', templateUrl: './consulta-ine.page.html', styleUrls: ['./consulta-ine.page.scss'], }) export class ConsultaInePage implements OnInit { cardValid: boolean; browser: any; esCargando: boolean; capturasINE: string; frontImg: string; isenabled: boolean; isValidoSpinnerFront: boolean; frontImg2: any; secuenciaId: number; client: ClientData; pData: JsonPersonalData; validINE: string; constructor( private iab: InAppBrowser, private alertCtrl: AlertController, private navCtrl: NavController, private documentosService: DocumentosService, private oauth: OauthService, private login: LoginService, private loading: LoadingService, private activityService: ActivitiesService, private saveS: GuardarStorageService, private activitiesService: ActivitiesService, private route: ActivatedRoute, public camera: Camera) { this.route.queryParams.subscribe(params => { if (params) { this.pData = JSON.parse(params.client); this.pData.observations = this.validINE; } }); this.cardValid = true; this.validINE = 'Identificación válida'; this.secuenciaId = 0; if (this.saveS.getTipoFlujo() === "alhajas") { this.secuenciaId = 4; } else { this.secuenciaId = 9; } } ngOnInit() { } getImagenFront() { this.isValidoSpinnerFront = true; const options: CameraOptions = { quality: 70, destinationType: this.camera.DestinationType.DATA_URL, encodingType: this.camera.EncodingType.JPEG, mediaType: this.camera.MediaType.PICTURE }; this.camera.getPicture(options).then((imageData) => { this.frontImg = 'data:image/jpeg;base64,' + imageData; this.frontImg = imageData; this.saveS.guardarStorageImagenF(this.frontImg); if (this.frontImg) { // enable the button this.isenabled = true; this.isValidoSpinnerFront = false; } else { // disable the button this.isenabled = false; this.isValidoSpinnerFront = false; } }, (err) => { // Handle error console.log('Camera issue:' + err, ); }); this.isValidoSpinnerFront = false; } openB() { this.browser = this.iab.create('https://listanominal.ine.mx', '_system'); } openGallery() { const cameraOptions = { sourceType: this.camera.PictureSourceType.PHOTOLIBRARY, destinationType: this.camera.DestinationType.DATA_URL, encodingType: this.camera.EncodingType.JPEG, quality: 70, targetWidth: 1000, targetHeight: 1000, correctOrientation: true }; /* this.camera.getPicture(cameraOptions) .then(file_uri => this.frontImg = file_uri, err => console.log(err)); */ this.camera.getPicture(cameraOptions).then((imageData) => { this.frontImg = 'data:image/jpeg;base64,' + imageData; this.frontImg2 = imageData; this.saveS.guardarStorageImagenF(this.frontImg); if (this.frontImg) { // enable the button this.isenabled = true; this.isValidoSpinnerFront = false; } else { // disable the button this.isenabled = false; this.isValidoSpinnerFront = false; } }, (err) => { // Handle error console.log('Camera issue:' + err); }); this.isValidoSpinnerFront = false; } goCallBack() { console.log('Typescript callback has been called'); this.browser.hide(); } changeListenerINE($event): void { let mensajeError = ''; const file: File = $event.target.files[0]; console.log('changeListenerF'); if ((file.type !== 'image/jpeg' && file.type !== 'image/png') || (file.size > 1000000)) { mensajeError = 'Formato y/o tamaño de imagen incorrecto'; } else { const myReader: FileReader = new FileReader(); myReader.onloadend = (e) => { this.frontImg = myReader.result.toString(); console.log('frontImg'); console.log(this.frontImg); // this.saveS.guardarStorageImagenF(this.capturasINE); // this.imgCapturada.emit(capturas); }; myReader.readAsDataURL(file); } if (this.frontImg) { // enable the button this.isenabled = true; this.isValidoSpinnerFront = false; } else {
goCargarDocumento() { try { const imagen = new Imagen(); const blobAnverso = imagen.convertirImagenEnBlob(this.frontImg); console.log('blobAnverso'); console.log(blobAnverso); if (blobAnverso) { this.cargarDocumento(blobAnverso, this.saveS.getBearerToken()); } else { alert('Imagen Invalida'); } } catch (error) { console.log(error); } } cargarDocumento(fileAnverso: any, bearerToken: string) { this.loading.present('Cargando...'); // this.actualizarActivity('EN PROCESO'); this.esCargando = true; const date = new Date(); const dataFile1 = new DataFile( 'bid:Anverso', 'Nombre', 'Primer apellido', 'Segundo apellido', '123549', date.toISOString(), 'RES_BURO', '123123'); const jsonRequest = new JsonRequest('IDOFA', this.saveS.getOperationID(), 'OK', '', ''); this.documentosService.cargarDocumento(jsonRequest, dataFile1, fileAnverso, bearerToken).subscribe( (respuesta: any) => { console.log('cargarDocumento respuesta', respuesta); if (respuesta['resultOK']) { // this.actualizarActivity('FINALIZADO'); this.esCargando = false; this.loading.dismiss(); // alert('Archivo guardado con éxito'); // this.navCtrl.navigateRoot('info-grales'); this.pData.observations=this.validINE; this.guardarDatos(this.pData); } else { // alert(respuesta['message']); this.loading.dismiss(); this.esCargando = false; } }, (error: HttpErrorResponse) => { this.loading.dismiss(); console.log(error); switch (error['status']) { case 401: alert('Es necesario iniciar session, nuemente para continuar'); this.navCtrl.navigateRoot('login'); break; case 404: alert('Es necesario iniciar session, nuemente para continuar'); this.navCtrl.navigateRoot('login'); break; case 500: alert('Por favor, reintentar para continuar'); this.cargarDocumento(fileAnverso, bearerToken); break; case 501: alert('Por favor, reintentar para continuar'); this.cargarDocumento(fileAnverso, bearerToken); break; default: alert('Es necesario iniciar session, nuemente para continuar'); this.navCtrl.navigateRoot('login'); break; } // alert('Hubo un error al enviar los datos, intenta de nuevo'); }); } actualizarActivity(estatus: string) { const productId = 1; const jsonData = new JsonData( productId, this.saveS.getSystemCode(), estatus, '1', '', this.secuenciaId, 1, this.saveS.getPersonId()); // {'id': this.saveS.getPersonId(), 'observations': this.validINE.toString()} console.log('jsonData validINE :'); console.log(jsonData); const jsonMetaData = new JsonMetadata(0, '', 0, 0, 1, 1); /* const jsonData = { 'data': { 'productId': 1, 'code': '', 'activityStatus': 'FINALIZADO', 'activityValue': null, 'data': '{"personal_data":{"id":' + this.saveS.getPersonId() + ',"observations":"' + this.validINE + '}}', 'secuence': 17, 'workflowId': 1, 'personId': this.saveS.getPersonId() }, 'metadata': { 'accuracy': 0, 'deviceInfo': '', 'lat
// disable the button this.isenabled = false; this.isValidoSpinnerFront = false; } }
conditional_block
consulta-ine.page.ts
/app/herramientas/imagen'; import { Camera, CameraOptions } from '@ionic-native/camera/ngx'; import { LoadingService } from 'src/app/services/loading.service'; import { Cliente } from '../../tipo-identificacion/consulta-similitud-confirmacion/model/Cliente.model'; import { JsonPersonalData } from 'src/app/services/actividades/model/json-personal-data.model'; import { JsonOperationData } from 'src/app/services/actividades/model/json-operation-data.model'; import { JsonInnerData } from 'src/app/services/actividades/model/json-inner-data.model'; import { ActivatedRoute } from '@angular/router'; @Component({ selector: 'app-consulta-ine', templateUrl: './consulta-ine.page.html', styleUrls: ['./consulta-ine.page.scss'], }) export class ConsultaInePage implements OnInit { cardValid: boolean; browser: any; esCargando: boolean; capturasINE: string; frontImg: string; isenabled: boolean; isValidoSpinnerFront: boolean; frontImg2: any; secuenciaId: number; client: ClientData; pData: JsonPersonalData; validINE: string; constructor( private iab: InAppBrowser, private alertCtrl: AlertController, private navCtrl: NavController, private documentosService: DocumentosService, private oauth: OauthService, private login: LoginService, private loading: LoadingService, private activityService: ActivitiesService, private saveS: GuardarStorageService, private activitiesService: ActivitiesService, private route: ActivatedRoute, public camera: Camera) { this.route.queryParams.subscribe(params => { if (params) { this.pData = JSON.parse(params.client); this.pData.observations = this.validINE; } }); this.cardValid = true; this.validINE = 'Identificación válida'; this.secuenciaId = 0; if (this.saveS.getTipoFlujo() === "alhajas") { this.secuenciaId = 4; } else { this.secuenciaId = 9; } } ngOnInit() {
getImagenFront() { this.isValidoSpinnerFront = true; const options: CameraOptions = { quality: 70, destinationType: this.camera.DestinationType.DATA_URL, encodingType: this.camera.EncodingType.JPEG, mediaType: this.camera.MediaType.PICTURE }; this.camera.getPicture(options).then((imageData) => { this.frontImg = 'data:image/jpeg;base64,' + imageData; this.frontImg = imageData; this.saveS.guardarStorageImagenF(this.frontImg); if (this.frontImg) { // enable the button this.isenabled = true; this.isValidoSpinnerFront = false; } else { // disable the button this.isenabled = false; this.isValidoSpinnerFront = false; } }, (err) => { // Handle error console.log('Camera issue:' + err, ); }); this.isValidoSpinnerFront = false; } openB() { this.browser = this.iab.create('https://listanominal.ine.mx', '_system'); } openGallery() { const cameraOptions = { sourceType: this.camera.PictureSourceType.PHOTOLIBRARY, destinationType: this.camera.DestinationType.DATA_URL, encodingType: this.camera.EncodingType.JPEG, quality: 70, targetWidth: 1000, targetHeight: 1000, correctOrientation: true }; /* this.camera.getPicture(cameraOptions) .then(file_uri => this.frontImg = file_uri, err => console.log(err)); */ this.camera.getPicture(cameraOptions).then((imageData) => { this.frontImg = 'data:image/jpeg;base64,' + imageData; this.frontImg2 = imageData; this.saveS.guardarStorageImagenF(this.frontImg); if (this.frontImg) { // enable the button this.isenabled = true; this.isValidoSpinnerFront = false; } else { // disable the button this.isenabled = false; this.isValidoSpinnerFront = false; } }, (err) => { // Handle error console.log('Camera issue:' + err); }); this.isValidoSpinnerFront = false; } goCallBack() { console.log('Typescript callback has been called'); this.browser.hide(); } changeListenerINE($event): void { let mensajeError = ''; const file: File = $event.target.files[0]; console.log('changeListenerF'); if ((file.type !== 'image/jpeg' && file.type !== 'image/png') || (file.size > 1000000)) { mensajeError = 'Formato y/o tamaño de imagen incorrecto'; } else { const myReader: FileReader = new FileReader(); myReader.onloadend = (e) => { this.frontImg = myReader.result.toString(); console.log('frontImg'); console.log(this.frontImg); // this.saveS.guardarStorageImagenF(this.capturasINE); // this.imgCapturada.emit(capturas); }; myReader.readAsDataURL(file); } if (this.frontImg) { // enable the button this.isenabled = true; this.isValidoSpinnerFront = false; } else { // disable the button this.isenabled = false; this.isValidoSpinnerFront = false; } } goCargarDocumento() { try { const imagen = new Imagen(); const blobAnverso = imagen.convertirImagenEnBlob(this.frontImg); console.log('blobAnverso'); console.log(blobAnverso); if (blobAnverso) { this.cargarDocumento(blobAnverso, this.saveS.getBearerToken()); } else { alert('Imagen Invalida'); } } catch (error) { console.log(error); } } cargarDocumento(fileAnverso: any, bearerToken: string) { this.loading.present('Cargando...'); // this.actualizarActivity('EN PROCESO'); this.esCargando = true; const date = new Date(); const dataFile1 = new DataFile( 'bid:Anverso', 'Nombre', 'Primer apellido', 'Segundo apellido', '123549', date.toISOString(), 'RES_BURO', '123123'); const jsonRequest = new JsonRequest('IDOFA', this.saveS.getOperationID(), 'OK', '', ''); this.documentosService.cargarDocumento(jsonRequest, dataFile1, fileAnverso, bearerToken).subscribe( (respuesta: any) => { console.log('cargarDocumento respuesta', respuesta); if (respuesta['resultOK']) { // this.actualizarActivity('FINALIZADO'); this.esCargando = false; this.loading.dismiss(); // alert('Archivo guardado con éxito'); // this.navCtrl.navigateRoot('info-grales'); this.pData.observations=this.validINE; this.guardarDatos(this.pData); } else { // alert(respuesta['message']); this.loading.dismiss(); this.esCargando = false; } }, (error: HttpErrorResponse) => { this.loading.dismiss(); console.log(error); switch (error['status']) { case 401: alert('Es necesario iniciar session, nuemente para continuar'); this.navCtrl.navigateRoot('login'); break; case 404: alert('Es necesario iniciar session, nuemente para continuar'); this.navCtrl.navigateRoot('login'); break; case 500: alert('Por favor, reintentar para continuar'); this.cargarDocumento(fileAnverso, bearerToken); break; case 501: alert('Por favor, reintentar para continuar'); this.cargarDocumento(fileAnverso, bearerToken); break; default: alert('Es necesario iniciar session, nuemente para continuar'); this.navCtrl.navigateRoot('login'); break; } // alert('Hubo un error al enviar los datos, intenta de nuevo'); }); } actualizarActivity(estatus: string) { const productId = 1; const jsonData = new JsonData( productId, this.saveS.getSystemCode(), estatus, '1', '', this.secuenciaId, 1, this.saveS.getPersonId()); // {'id': this.saveS.getPersonId(), 'observations': this.validINE.toString()} console.log('jsonData validINE :'); console.log(jsonData); const jsonMetaData = new JsonMetadata(0, '', 0, 0, 1, 1); /* const jsonData = { 'data': { 'productId': 1, 'code': '', 'activityStatus': 'FINALIZADO', 'activityValue': null, 'data': '{"personal_data":{"id":' + this.saveS.getPersonId() + ',"observations":"' + this.validINE + '}}', 'secuence': 17, 'workflowId': 1, 'personId': this.saveS.getPersonId() }, 'metadata': { 'accuracy': 0, 'deviceInfo': '', 'latutide
}
identifier_body
simple_http.rs
4) Parse into socket address. // At this point we either have <host_name> or <host_name_>:<port> // `std::net::ToSocketAddrs` requires `&str` to have <host_name_>:<port> format. let mut addr = match after_auth.to_socket_addrs() { Ok(addr) => addr, Err(_) => { // Invalid socket address. Try to add port. format!("{}:{}", after_auth, fallback_port).to_socket_addrs()? } }; match addr.next() { Some(a) => Ok((a, path.to_owned())), None => Err(Error::url(url, "invalid hostname: error extracting socket address")), } } impl Transport for SimpleHttpTransport { fn send_request(&self, req: Request) -> Result<Response, crate::Error> { Ok(self.request(req)?) } fn send_batch(&self, reqs: &[Request]) -> Result<Vec<Response>, crate::Error> { Ok(self.request(reqs)?) } fn fmt_target(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "http://{}:{}{}", self.addr.ip(), self.addr.port(), self.path) } } /// Builder for simple bitcoind [`SimpleHttpTransport`]. #[derive(Clone, Debug)] pub struct Builder { tp: SimpleHttpTransport, } impl Builder { /// Constructs a new [`Builder`] with default configuration. pub fn new() -> Builder { Builder { tp: SimpleHttpTransport::new(), } } /// Sets the timeout after which requests will abort if they aren't finished. pub fn timeout(mut self, timeout: Duration) -> Self { self.tp.timeout = timeout; self } /// Sets the URL of the server to the transport. pub fn url(mut self, url: &str) -> Result<Self, Error> { self.tp.set_url(url)?; Ok(self) } /// Adds authentication information to the transport. pub fn auth<S: AsRef<str>>(mut self, user: S, pass: Option<S>) -> Self { let mut auth = user.as_ref().to_owned(); auth.push(':'); if let Some(ref pass) = pass { auth.push_str(pass.as_ref()); } self.tp.basic_auth = Some(format!("Basic {}", &base64::encode(auth.as_bytes()))); self } /// Adds authentication information to the transport using a cookie string ('user:pass'). pub fn cookie_auth<S: AsRef<str>>(mut self, cookie: S) -> Self { self.tp.basic_auth = Some(format!("Basic {}", &base64::encode(cookie.as_ref().as_bytes()))); self } /// Adds proxy address to the transport for SOCKS5 proxy. #[cfg(feature = "proxy")] pub fn proxy_addr<S: AsRef<str>>(mut self, proxy_addr: S) -> Result<Self, Error> { // We don't expect path in proxy address. self.tp.proxy_addr = check_url(proxy_addr.as_ref())?.0; Ok(self) } /// Adds optional proxy authentication as ('username', 'password'). #[cfg(feature = "proxy")] pub fn proxy_auth<S: AsRef<str>>(mut self, user: S, pass: S) -> Self { self.tp.proxy_auth = Some((user, pass)).map(|(u, p)| (u.as_ref().to_string(), p.as_ref().to_string())); self } /// Builds the final [`SimpleHttpTransport`]. pub fn build(self) -> SimpleHttpTransport { self.tp } } impl Default for Builder { fn default() -> Self { Builder::new() } } impl crate::Client { /// Creates a new JSON-RPC client using a bare-minimum HTTP transport. pub fn simple_http( url: &str, user: Option<String>, pass: Option<String>, ) -> Result<crate::Client, Error> { let mut builder = Builder::new().url(url)?; if let Some(user) = user { builder = builder.auth(user, pass); } Ok(crate::Client::with_transport(builder.build())) } /// Creates a new JSON_RPC client using a HTTP-Socks5 proxy transport. #[cfg(feature = "proxy")] pub fn http_proxy( url: &str, user: Option<String>, pass: Option<String>, proxy_addr: &str, proxy_auth: Option<(&str, &str)>, ) -> Result<crate::Client, Error> { let mut builder = Builder::new().url(url)?; if let Some(user) = user { builder = builder.auth(user, pass); } builder = builder.proxy_addr(proxy_addr)?; if let Some((user, pass)) = proxy_auth { builder = builder.proxy_auth(user, pass); } let tp = builder.build(); Ok(crate::Client::with_transport(tp)) } } /// Error that can happen when sending requests. #[derive(Debug)] pub enum Error { /// An invalid URL was passed. InvalidUrl { /// The URL passed. url: String, /// The reason the URL is invalid. reason: &'static str, }, /// An error occurred on the socket layer. SocketError(io::Error), /// The HTTP response was too short to even fit a HTTP 1.1 header. HttpResponseTooShort { /// The total length of the response. actual: usize, /// Minimum length we can parse. needed: usize, }, /// The HTTP response started with a HTTP/1.1 line which was not ASCII. HttpResponseNonAsciiHello(Vec<u8>), /// The HTTP response did not start with HTTP/1.1 HttpResponseBadHello { /// Actual HTTP-whatever string. actual: String, /// The hello string of the HTTP version we support. expected: String, }, /// Could not parse the status value as a number. HttpResponseBadStatus(String, num::ParseIntError), /// Could not parse the status value as a number. HttpResponseBadContentLength(String, num::ParseIntError), /// The indicated content-length header exceeded our maximum. HttpResponseContentLengthTooLarge { /// The length indicated in the content-length header. length: u64, /// Our hard maximum on number of bytes we'll try to read. max: u64, }, /// Unexpected HTTP error code (non-200). HttpErrorCode(u16), /// Received EOF before getting as many bytes as were indicated by the content-length header. IncompleteResponse { /// The content-length header. content_length: u64, /// The number of bytes we actually read. n_read: u64, }, /// JSON parsing error. Json(serde_json::Error), } impl Error { /// Utility method to create [`Error::InvalidUrl`] variants. fn url<U: Into<String>>(url: U, reason: &'static str) -> Error { Error::InvalidUrl { url: url.into(), reason, } } } impl fmt::Display for Error { fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> { use Error::*; match *self { InvalidUrl { ref url, ref reason, } => write!(f, "invalid URL '{}': {}", url, reason), SocketError(ref e) => write!(f, "Couldn't connect to host: {}", e), HttpResponseTooShort { ref actual, ref needed, } => { write!(f, "HTTP response too short: length {}, needed {}.", actual, needed) } HttpResponseNonAsciiHello(ref bytes) => { write!(f, "HTTP response started with non-ASCII {:?}", bytes) } HttpResponseBadHello { ref actual, ref expected, } => { write!(f, "HTTP response started with `{}`; expected `{}`.", actual, expected) } HttpResponseBadStatus(ref status, ref err) => { write!(f, "HTTP response had bad status code `{}`: {}.", status, err) } HttpResponseBadContentLength(ref len, ref err) => { write!(f, "HTTP response had bad content length `{}`: {}.", len, err) } HttpResponseContentLengthTooLarge { length, max, } => { write!(f, "HTTP response content length {} exceeds our max {}.", length, max) } HttpErrorCode(c) => write!(f, "unexpected HTTP code: {}", c), IncompleteResponse { content_length, n_read, } => { write!( f, "read {} bytes but HTTP response content-length header was {}.", n_read, content_length ) } Json(ref e) => write!(f, "JSON error: {}", e), } } } impl error::Error for Error { fn source(&self) -> Option<&(dyn error::Error + 'static)>
{ use self::Error::*; match *self { InvalidUrl { .. } | HttpResponseTooShort { .. } | HttpResponseNonAsciiHello(..) | HttpResponseBadHello { .. } | HttpResponseBadStatus(..) | HttpResponseBadContentLength(..) | HttpResponseContentLengthTooLarge { .. } | HttpErrorCode(_)
identifier_body
simple_http.rs
sock.get_mut().write_all(request_bytes.as_slice()).is_ok() && sock.get_mut().flush().is_ok(); // This indicates the socket is broken so let's retry the send once with a fresh socket if !write_success { *sock.get_mut() = self.fresh_socket()?; sock.get_mut().write_all(request_bytes.as_slice())?; sock.get_mut().flush()?; } // Parse first HTTP response header line let mut header_buf = String::new(); let read_success = sock.read_line(&mut header_buf).is_ok(); // This is another possible indication that the socket is broken so let's retry the send once // with a fresh socket IF the write attempt has not already experienced a failure if (!read_success || header_buf.is_empty()) && write_success { *sock.get_mut() = self.fresh_socket()?; sock.get_mut().write_all(request_bytes.as_slice())?; sock.get_mut().flush()?; sock.read_line(&mut header_buf)?; } if header_buf.len() < 12 { return Err(Error::HttpResponseTooShort { actual: header_buf.len(), needed: 12, }); } if !header_buf.as_bytes()[..12].is_ascii() { return Err(Error::HttpResponseNonAsciiHello(header_buf.as_bytes()[..12].to_vec())); } if !header_buf.starts_with("HTTP/1.1 ") { return Err(Error::HttpResponseBadHello { actual: header_buf[0..9].into(), expected: "HTTP/1.1 ".into(), }); } let response_code = match header_buf[9..12].parse::<u16>() { Ok(n) => n, Err(e) => return Err(Error::HttpResponseBadStatus(header_buf[9..12].into(), e)), }; // Parse response header fields let mut content_length = None; loop { header_buf.clear(); sock.read_line(&mut header_buf)?; if header_buf == "\r\n" { break; } header_buf.make_ascii_lowercase(); const CONTENT_LENGTH: &str = "content-length: "; if let Some(s) = header_buf.strip_prefix(CONTENT_LENGTH) { content_length = Some( s.trim() .parse::<u64>() .map_err(|e| Error::HttpResponseBadContentLength(s.into(), e))?, ); } } if response_code == 401 { // There is no body in a 401 response, so don't try to read it return Err(Error::HttpErrorCode(response_code)); } // Read up to `content_length` bytes. Note that if there is no content-length // header, we will assume an effectively infinite content length, i.e. we will // just keep reading from the socket until it is closed. let mut reader = match content_length { None => sock.take(FINAL_RESP_ALLOC), Some(n) if n > FINAL_RESP_ALLOC => { return Err(Error::HttpResponseContentLengthTooLarge { length: n, max: FINAL_RESP_ALLOC, }); } Some(n) => sock.take(n), }; // Attempt to parse the response. Don't check the HTTP error code until // after parsing, since Bitcoin Core will often return a descriptive JSON // error structure which is more useful than the error code. match serde_json::from_reader(&mut reader) { Ok(s) => { if content_length.is_some() { reader.bytes().count(); // consume any trailing bytes } Ok(s) } Err(e) => { // If the response was not 200, assume the parse failed because of that if response_code != 200 { Err(Error::HttpErrorCode(response_code)) } else { // If it was 200 then probably it was legitimately a parse error Err(e.into()) } } } } } /// Does some very basic manual URL parsing because the uri/url crates /// all have unicode-normalization as a dependency and that's broken. fn check_url(url: &str) -> Result<(SocketAddr, String), Error> { // The fallback port in case no port was provided. // This changes when the http or https scheme was provided. let mut fallback_port = DEFAULT_PORT; // We need to get the hostname and the port. // (1) Split scheme let after_scheme = { let mut split = url.splitn(2, "://"); let s = split.next().unwrap(); match split.next() { None => s, // no scheme present Some(after) => { // Check if the scheme is http or https. if s == "http" { fallback_port = 80; } else if s == "https" { fallback_port = 443; } else { return Err(Error::url(url, "scheme should be http or https")); } after } } }; // (2) split off path let (before_path, path) = { if let Some(slash) = after_scheme.find('/') { (&after_scheme[0..slash], &after_scheme[slash..]) } else { (after_scheme, "/") } }; // (3) split off auth part let after_auth = { let mut split = before_path.splitn(2, '@'); let s = split.next().unwrap(); split.next().unwrap_or(s) }; // (4) Parse into socket address. // At this point we either have <host_name> or <host_name_>:<port> // `std::net::ToSocketAddrs` requires `&str` to have <host_name_>:<port> format. let mut addr = match after_auth.to_socket_addrs() { Ok(addr) => addr, Err(_) => { // Invalid socket address. Try to add port. format!("{}:{}", after_auth, fallback_port).to_socket_addrs()? } }; match addr.next() { Some(a) => Ok((a, path.to_owned())), None => Err(Error::url(url, "invalid hostname: error extracting socket address")), } } impl Transport for SimpleHttpTransport { fn send_request(&self, req: Request) -> Result<Response, crate::Error> { Ok(self.request(req)?) } fn send_batch(&self, reqs: &[Request]) -> Result<Vec<Response>, crate::Error> { Ok(self.request(reqs)?) } fn fmt_target(&self, f: &mut fmt::Formatter) -> fmt::Result { write!(f, "http://{}:{}{}", self.addr.ip(), self.addr.port(), self.path) } } /// Builder for simple bitcoind [`SimpleHttpTransport`]. #[derive(Clone, Debug)] pub struct Builder { tp: SimpleHttpTransport, } impl Builder { /// Constructs a new [`Builder`] with default configuration. pub fn new() -> Builder { Builder { tp: SimpleHttpTransport::new(), } } /// Sets the timeout after which requests will abort if they aren't finished. pub fn timeout(mut self, timeout: Duration) -> Self { self.tp.timeout = timeout; self } /// Sets the URL of the server to the transport. pub fn url(mut self, url: &str) -> Result<Self, Error> { self.tp.set_url(url)?; Ok(self) } /// Adds authentication information to the transport. pub fn auth<S: AsRef<str>>(mut self, user: S, pass: Option<S>) -> Self { let mut auth = user.as_ref().to_owned(); auth.push(':'); if let Some(ref pass) = pass { auth.push_str(pass.as_ref()); } self.tp.basic_auth = Some(format!("Basic {}", &base64::encode(auth.as_bytes()))); self } /// Adds authentication information to the transport using a cookie string ('user:pass'). pub fn cookie_auth<S: AsRef<str>>(mut self, cookie: S) -> Self { self.tp.basic_auth = Some(format!("Basic {}", &base64::encode(cookie.as_ref().as_bytes()))); self } /// Adds proxy address to the transport for SOCKS5 proxy. #[cfg(feature = "proxy")] pub fn proxy_addr<S: AsRef<str>>(mut self, proxy_addr: S) -> Result<Self, Error> { // We don't expect path in proxy address. self.tp.proxy_addr = check_url(proxy_addr.as_ref())?.0; Ok(self) } /// Adds optional proxy authentication as ('username', 'password'). #[cfg(feature = "proxy")] pub fn proxy_auth<S: AsRef<str>>(mut self, user: S, pass: S) -> Self { self.tp.proxy_auth = Some((user, pass)).map(|(u, p)| (u.as_ref().to_string(), p.as_ref().to_string())); self } /// Builds the final [`SimpleHttpTransport`]. pub fn
(self) -> SimpleHttpTransport { self.tp } } impl Default for Builder { fn default() -> Self { Builder::new() } } impl crate::Client { /// Creates a new JSON-RPC client using a bare-minimum HTTP transport. pub fn simple_http( url
build
identifier_name
simple_http.rs
we can parse. needed: usize, }, /// The HTTP response started with a HTTP/1.1 line which was not ASCII. HttpResponseNonAsciiHello(Vec<u8>), /// The HTTP response did not start with HTTP/1.1 HttpResponseBadHello { /// Actual HTTP-whatever string. actual: String, /// The hello string of the HTTP version we support. expected: String, }, /// Could not parse the status value as a number. HttpResponseBadStatus(String, num::ParseIntError), /// Could not parse the status value as a number. HttpResponseBadContentLength(String, num::ParseIntError), /// The indicated content-length header exceeded our maximum. HttpResponseContentLengthTooLarge { /// The length indicated in the content-length header. length: u64, /// Our hard maximum on number of bytes we'll try to read. max: u64, }, /// Unexpected HTTP error code (non-200). HttpErrorCode(u16), /// Received EOF before getting as many bytes as were indicated by the content-length header. IncompleteResponse { /// The content-length header. content_length: u64, /// The number of bytes we actually read. n_read: u64, }, /// JSON parsing error. Json(serde_json::Error), } impl Error { /// Utility method to create [`Error::InvalidUrl`] variants. fn url<U: Into<String>>(url: U, reason: &'static str) -> Error { Error::InvalidUrl { url: url.into(), reason, } } } impl fmt::Display for Error { fn fmt(&self, f: &mut fmt::Formatter) -> Result<(), fmt::Error> { use Error::*; match *self { InvalidUrl { ref url, ref reason, } => write!(f, "invalid URL '{}': {}", url, reason), SocketError(ref e) => write!(f, "Couldn't connect to host: {}", e), HttpResponseTooShort { ref actual, ref needed, } => { write!(f, "HTTP response too short: length {}, needed {}.", actual, needed) } HttpResponseNonAsciiHello(ref bytes) => { write!(f, "HTTP response started with non-ASCII {:?}", bytes) } HttpResponseBadHello { ref actual, ref expected, } => { write!(f, "HTTP response started with `{}`; expected `{}`.", actual, expected) } HttpResponseBadStatus(ref status, ref err) => { write!(f, "HTTP response had bad status code `{}`: {}.", status, err) } HttpResponseBadContentLength(ref len, ref err) => { write!(f, "HTTP response had bad content length `{}`: {}.", len, err) } HttpResponseContentLengthTooLarge { length, max, } => { write!(f, "HTTP response content length {} exceeds our max {}.", length, max) } HttpErrorCode(c) => write!(f, "unexpected HTTP code: {}", c), IncompleteResponse { content_length, n_read, } => { write!( f, "read {} bytes but HTTP response content-length header was {}.", n_read, content_length ) } Json(ref e) => write!(f, "JSON error: {}", e), } } } impl error::Error for Error { fn source(&self) -> Option<&(dyn error::Error + 'static)> { use self::Error::*; match *self { InvalidUrl { .. } | HttpResponseTooShort { .. } | HttpResponseNonAsciiHello(..) | HttpResponseBadHello { .. } | HttpResponseBadStatus(..) | HttpResponseBadContentLength(..) | HttpResponseContentLengthTooLarge { .. } | HttpErrorCode(_) | IncompleteResponse { .. } => None, SocketError(ref e) => Some(e), Json(ref e) => Some(e), } } } impl From<io::Error> for Error { fn from(e: io::Error) -> Self { Error::SocketError(e) } } impl From<serde_json::Error> for Error { fn from(e: serde_json::Error) -> Self { Error::Json(e) } } impl From<Error> for crate::Error { fn from(e: Error) -> crate::Error { match e { Error::Json(e) => crate::Error::Json(e), e => crate::Error::Transport(Box::new(e)), } } } /// Global mutex used by the fuzzing harness to inject data into the read end of the TCP stream. #[cfg(jsonrpc_fuzz)] pub static FUZZ_TCP_SOCK: Mutex<Option<io::Cursor<Vec<u8>>>> = Mutex::new(None); #[cfg(jsonrpc_fuzz)] #[derive(Clone, Debug)] struct TcpStream; #[cfg(jsonrpc_fuzz)] mod impls { use super::*; impl Read for TcpStream { fn read(&mut self, buf: &mut [u8]) -> io::Result<usize> { match *FUZZ_TCP_SOCK.lock().unwrap() { Some(ref mut cursor) => io::Read::read(cursor, buf), None => Ok(0), } } } impl Write for TcpStream { fn write(&mut self, buf: &[u8]) -> io::Result<usize> { io::sink().write(buf) } fn flush(&mut self) -> io::Result<()> { Ok(()) } } impl TcpStream { pub fn connect_timeout(_: &SocketAddr, _: Duration) -> io::Result<Self> { Ok(TcpStream) } pub fn set_read_timeout(&self, _: Option<Duration>) -> io::Result<()> { Ok(()) } pub fn set_write_timeout(&self, _: Option<Duration>) -> io::Result<()> { Ok(()) } } } #[cfg(test)] mod tests { use std::net; #[cfg(feature = "proxy")] use std::str::FromStr; use super::*; use crate::Client; #[test] fn test_urls() { let addr: net::SocketAddr = ("localhost", 22).to_socket_addrs().unwrap().next().unwrap(); let urls = [ "localhost:22", "http://localhost:22/", "https://localhost:22/walletname/stuff?it=working", "http://me:weak@localhost:22/wallet", ]; for u in &urls { let tp = Builder::new().url(u).unwrap().build(); assert_eq!(tp.addr, addr); } // Default port and 80 and 443 fill-in. let addr: net::SocketAddr = ("localhost", 80).to_socket_addrs().unwrap().next().unwrap(); let tp = Builder::new().url("http://localhost/").unwrap().build(); assert_eq!(tp.addr, addr); let addr: net::SocketAddr = ("localhost", 443).to_socket_addrs().unwrap().next().unwrap(); let tp = Builder::new().url("https://localhost/").unwrap().build(); assert_eq!(tp.addr, addr); let addr: net::SocketAddr = ("localhost", super::DEFAULT_PORT).to_socket_addrs().unwrap().next().unwrap(); let tp = Builder::new().url("localhost").unwrap().build(); assert_eq!(tp.addr, addr); let valid_urls = [ "localhost", "127.0.0.1:8080", "http://127.0.0.1:8080/", "http://127.0.0.1:8080/rpc/test", "https://127.0.0.1/rpc/test", "http://[2001:0db8:85a3:0000:0000:8a2e:0370:7334]:8300", "http://[2001:0db8:85a3:0000:0000:8a2e:0370:7334]", ]; for u in &valid_urls { let (addr, path) = check_url(u).unwrap(); let builder = Builder::new().url(u).unwrap_or_else(|_| panic!("error for: {}", u)); assert_eq!(builder.tp.addr, addr); assert_eq!(builder.tp.path, path); assert_eq!(builder.tp.timeout, DEFAULT_TIMEOUT); assert_eq!(builder.tp.basic_auth, None); #[cfg(feature = "proxy")] assert_eq!(builder.tp.proxy_addr, SocketAddr::from_str("127.0.0.1:9050").unwrap()); } let invalid_urls = [ "127.0.0.1.0:8080", "httpx://127.0.0.1:8080/", "ftp://127.0.0.1:8080/rpc/test", "http://127.0.0./rpc/test",
random_line_split
DataManager.py
j]) nodename = self.id_nodename_map[pred_ent] new_out.append(nodename) elif self.index2word[each] in entity_attr_list: attr_name = self.index2word[each] cnt = 0 suc_flag = False for idx, prob in sorted(enumerate(pred_ents[i][j]), key=lambda i: i[1], reverse=True): if suc_flag or cnt >= topK: break nodename = self.id_nodename_map[idx] if nodename not in self.nodename_attr_map: cnt += 1 continue for attr, val in self.nodename_attr_map[nodename].items(): if attr in attr_name: new_out.append(val) suc_flag = True break cnt += 1 if not suc_flag: new_out.append("<UNK>") else: new_out.append(self.index2word[each]) """ if each == self.word2index["<$>"]: pred_ent = np.argmax(pred_ents[i][j]) nodename = self.id_nodename_map[pred_ent] nodename_wordids = [self.word2index[x] for x in nodename.split()] new_out += nodename_wordids else: new_out.append(each) """ outputs[i] = new_out return outputs def get_nodes_rep(self, node_id_map, node_info_map, max_len=50): nodes_rep = [] nodes_rep_map = [] for name, id_ in sorted(node_id_map.items(), key=lambda i: i[1]): if name == "none" and id_ == 0: nodes_rep.append([PAD] * max_len) nodes_rep_map.append({"words": ["none"], "idx": [0]}) continue # the attributes used to build relationship # attributes as nodes: {"pricerange", "area", "food"} # attributes only as relation: {"internet", "parking", "stars", "attraction_type", "hotel_type"} # only user node name as node's feature name = name.split("::")[-1] node_desc = [name] nodes_rep_idx = [PAD] * max_len nodes_rep_idx[0] = self.word2index[name] nodes_rep_word = [name] """ for attr, val in node_info_map.items(): #if attr in {"address", "area", "pricerange", "introduction", "food", "stars"} or "type" in attr: if attr == "introduction": node_desc.append(val) node_desc = " ".join(node_desc) nodes_rep_idx = [] nodes_rep_word = [] for each_word in node_desc.split(): for word in re.split(r'[\[\](::)_]', each_word): if word == "": continue else: if word not in self.word2index: continue else: word_idx = self.word2index[word] nodes_rep_idx.append(word_idx) nodes_rep_word.append(word) len_ = len(nodes_rep_idx) if len_ >= max_len: nodes_rep_idx = nodes_rep_idx[0:max_len] nodes_rep_word = nodes_rep_word[0:max_len] else: nodes_rep_idx += [PAD] * (max_len - len_) """ nodes_rep.append(nodes_rep_idx) nodes_rep_map.append({"words": nodes_rep_word, "idx": nodes_rep_idx}) json.dump(nodes_rep_map, open("nodes_rep_words_idx.json", "w")) json.dump(self.word2index, open("word2index.json", "w")) #exit() return nodes_rep def get_node_id_map(self): data = json.load(open("./data/entity_list_simple.json")) node_id_map = {} id_nodename_map = {} for i, node in enumerate(data): node_id_map[node] = i + 1 tmp = node.split("::") #node_name = " ".join(tmp[1].split("_")) node_name = tmp[1] id_nodename_map[i+1] = node_name node_id_map["none"] = 0 id_nodename_map[0] = "" return node_id_map, id_nodename_map def get_adj_mat(self, input_file, item_id_map): adj = json.load(open(input_file)) new_adj = {} for i, neibors in adj.items(): i_idx = item_id_map[i] new_adj[i_idx] = [] for j in neibors: j_idx = item_id_map[j] new_adj[i_idx].append(j_idx) new_adj = nx.adjacency_matrix(nx.from_dict_of_lists(new_adj)) new_adj = self.normalize_adj(new_adj + sp.eye(new_adj.shape[0])) new_adj = torch.FloatTensor(np.array(new_adj.todense())) return new_adj def normalize_adj(self, mx): """Row-normalize sparse matrix""" rowsum = np.array(mx.sum(1)) r_inv_sqrt = np.power(rowsum, -0.5).flatten() r_inv_sqrt[np.isinf(r_inv_sqrt)] = 0. r_mat_inv_sqrt = sp.diags(r_inv_sqrt) return mx.dot(r_mat_inv_sqrt).transpose().dot(r_mat_inv_sqrt).tocoo() def create_dataset(self, name, batch_size): datas = self.data[name] src_seq_lens = [] src_seqs, trg_seqs = [], [] trg_stops, src_tfs = [], [] trg_ents, trg_ents_mask, trg_seqs_ori = [], [], [] nonstop_voc_size = len(self.index2nonstop) for item in datas: src_len, src, trg, ents, ents_mask, trg_ori = item tensor_src_len, tensor_src, tensor_trg = torch.LongTensor(src_len), \ torch.LongTensor(src), torch.LongTensor(trg) src_seq_lens.append(tensor_src_len) src_seqs.append(tensor_src) trg_seqs.append(tensor_trg) trg_stop = torch.zeros_like(tensor_trg) for i, index in enumerate(trg): if index in self.stop_words_index: trg_stop[i] = 1 trg_stops.append(trg_stop) src_tf = torch.zeros(nonstop_voc_size) for j, uttr in enumerate(src): for i, index in enumerate(uttr): if i == src_len[j]: break if index not in self.stop_words_index: src_tf[self.index2nonstop[index]] += 1 if src_tf.sum().item() > 0: src_tf /= src_tf.sum() src_tfs.append(src_tf) trg_ents.append(torch.LongTensor(ents)) trg_ents_mask.append(torch.LongTensor(ents_mask)) trg_seqs_ori.append(trg_ori) print(len(trg_stops), len(trg_seqs), len(trg_ents), len(trg_seqs_ori)) dataset = Dataset(src_seq_lens, src_seqs, trg_seqs, trg_stops, src_tfs, trg_ents, trg_ents_mask, trg_seqs_ori) dataloader = data.DataLoader(dataset, batch_size, True, num_workers=0, collate_fn=pad_packed_collate) return dataloader def compute_stopword(self, y): res = torch.zeros_like(y).to(device=device) for i, row in enumerate(y): words_index = row.tolist() res[i] = torch.LongTensor([int(index in self.stop_words_index) for index in words_index]) return res def interpret(self, preds, refs, lens, f): i = random.randrange(0, len(lens)) l = max(lens) for j in range(l): word = self.index2word[preds[i][j].item()] print(word, end=' ') f.write('{0} '.format(word)) if word == '<EOS>': break print() f.write('\n') l = lens[i] for j in range(l): word = self.index2word[refs[i][j].item()] print(word, end=' ') f.write('{0} '.format(word)) print() f.write('\n') class Dataset(data.Dataset):
def __init__(self, src_seq_lens, src_seqs, trg_seqs, trg_stops, src_tfs, trg_ents, trg_ents_mask, trg_seqs_ori): self.src_seq_lens = src_seq_lens self.src_seqs = src_seqs self.trg_seqs = trg_seqs self.trg_stops = trg_stops self.src_tfs = src_tfs self.num_total_seqs = len(src_seqs) self.trg_ents = trg_ents self.trg_ents_mask = trg_ents_mask self.trg_seqs_ori = trg_seqs_ori def __getitem__(self, index): src_seq_len = self.src_seq_lens[index] src_seq = self.src_seqs[index] trg_seq = self.trg_seqs[index] trg_stop = self.trg_stops[index] src_tf = self.src_tfs[index] trg_ent = self.trg_ents[index] trg_ent_mask = self.trg_ents_mask[index] trg_seq_ori = self.trg_seqs_ori[index]
identifier_body
DataManager.py
self.word2index[key] = i + 4 #PAD,UNK,GO,EOS if value >= stopword_freq_lb: stopwords_self.add(key) # to add all entity name into vocab entity_list = json.load(open("./data/entity_list_simple.json")) start_idx = len(self.word2index) for entity_name in entity_list: entity_name = entity_name.split("::")[-1] if entity_name not in self.word2index: self.word2index[entity_name] = start_idx start_idx += 1 self.stop_words_index = set([PAD, UNK, GO, EOS]) #self.stop_words_index |= set([self.word2index[word] for word in STOP_WORDS # if word in self.word2index]) # here we add all words into stopword list self.stop_words_index |= set([self.word2index[word] for word in stopwords_self]) self.index2word = dict((v, k) for k, v in self.word2index.items()) #load word vector if no_pretrain_word2vec: self.vector = None else: self.vector = 0.1 * np.random.rand(len(self.word2index), dim) with open("{0}/vector.txt".format(path)) as fl: for line in fl: vec = line.strip().split() word = vec[0].lower() vec = list(map(float, vec[1:])) if word in self.word2index: self.vector[self.word2index[word]] = np.asarray(vec) self.vector = torch.Tensor(self.vector) # compute tf len_voc = len(self.word2index.values()) self.index2nonstop = {} cnt = 0 for i in range(len_voc): if i not in self.stop_words_index: self.index2nonstop[i] = cnt cnt += 1 # for graph initialization self.node_id_map, self.id_nodename_map = self.get_node_id_map() self.node_info_map, self.nodename_attr_map = self.get_node_info() self.adj = self.get_adj_mat("./data/adj_simple.json", self.node_id_map) self.nodes_rep = self.get_nodes_rep(self.node_id_map, self.node_info_map) self.n_entity = len(self.node_id_map) #get index self.data = {} for name in ["train", "valid", "test"]: self.data[name] = [] for number, item in enumerate(self.text[name]): len_u = len(item[0]) indices = [[], [[] for _ in range(len_u)], [], [], [], []] #src_len, src, trg, trg_entities, trg_entities_mask indices[0] = [u.count(' ')+1 for u in item[0]] # on purpose max_u_len = max(indices[0]) # history for i in range(len_u): words = item[0][i].split() indices[1][i] = [self.word2index[word] if word in self.word2index else UNK for word in words] + [PAD] * (max_u_len - len(words)) # answer words = item[1].split() #print("item1:: ", len(words)) indices[2] = [self.word2index[word] if word in self.word2index else UNK for word in words] indices[2].append(EOS) # answer entity entities = item[2].split() #print("item2 entities:: ", len(entities)) indices[3] = [self.node_id_map[entity_name] for entity_name in entities] indices[3].append(0) indices[4] = [] for x in indices[3]: if x == 0: indices[4].append(0) else: indices[4].append(1) # ansuer original sentence words = item[3].split() indices[5] = words indices[5].append("<EOS>") if len(indices[2]) != len(indices[3]): print(number, len(indices[2]), len(indices[3])) print(item[1]) print(item[2]) exit() self.data[name].append(indices) def get_node_info(self): node_info_map = json.load(open("./data/entity_info.json")) nodename_attr_map = {} for node, res in node_info_map.items(): node_name = node.split("::")[-1] nodename_attr_map[node_name] = res return node_info_map, nodename_attr_map def post_process(self, outputs, pred_ents, topK=1): outputs = outputs.cpu().numpy().tolist() pred_ents = pred_ents.cpu().numpy() entity_attr_list = { "[attraction_address]", "[restaurant_address]", "[attraction_phone]", "[restaurant_phone]", "[hotel_address]", "[restaurant_postcode]", "[attraction_postcode]", "[hotel_phone]", "[hotel_postcode]", "[hospital_phone]" } lens_new = [] for i, out in enumerate(outputs): new_out = [] for j, each in enumerate(out): if self.index2word[each] == "<$>": pred_ent = np.argmax(pred_ents[i][j]) nodename = self.id_nodename_map[pred_ent] new_out.append(nodename) elif self.index2word[each] in entity_attr_list: attr_name = self.index2word[each] cnt = 0 suc_flag = False for idx, prob in sorted(enumerate(pred_ents[i][j]), key=lambda i: i[1], reverse=True): if suc_flag or cnt >= topK: break nodename = self.id_nodename_map[idx] if nodename not in self.nodename_attr_map: cnt += 1 continue for attr, val in self.nodename_attr_map[nodename].items(): if attr in attr_name: new_out.append(val) suc_flag = True break cnt += 1 if not suc_flag: new_out.append("<UNK>") else: new_out.append(self.index2word[each]) """ if each == self.word2index["<$>"]: pred_ent = np.argmax(pred_ents[i][j]) nodename = self.id_nodename_map[pred_ent] nodename_wordids = [self.word2index[x] for x in nodename.split()] new_out += nodename_wordids else: new_out.append(each) """ outputs[i] = new_out return outputs def get_nodes_rep(self, node_id_map, node_info_map, max_len=50): nodes_rep = [] nodes_rep_map = [] for name, id_ in sorted(node_id_map.items(), key=lambda i: i[1]): if name == "none" and id_ == 0: nodes_rep.append([PAD] * max_len) nodes_rep_map.append({"words": ["none"], "idx": [0]}) continue # the attributes used to build relationship # attributes as nodes: {"pricerange", "area", "food"} # attributes only as relation: {"internet", "parking", "stars", "attraction_type", "hotel_type"} # only user node name as node's feature name = name.split("::")[-1] node_desc = [name] nodes_rep_idx = [PAD] * max_len nodes_rep_idx[0] = self.word2index[name] nodes_rep_word = [name] """ for attr, val in node_info_map.items(): #if attr in {"address", "area", "pricerange", "introduction", "food", "stars"} or "type" in attr: if attr == "introduction": node_desc.append(val) node_desc = " ".join(node_desc) nodes_rep_idx = [] nodes_rep_word = [] for each_word in node_desc.split(): for word in re.split(r'[\[\](::)_]', each_word): if word == "": continue else: if word not in self.word2index: continue else: word_idx = self.word2index[word] nodes_rep_idx.append(word_idx) nodes_rep_word.append(word) len_ = len(nodes_rep_idx) if len_ >= max_len: nodes_rep_idx = nodes_rep_idx[0:max_len] nodes_rep_word = nodes_rep_word[0:max_len] else: nodes_rep_idx += [PAD] * (max_len - len_) """ nodes_rep.append(nodes_rep_idx) nodes_rep_map.append({"words": nodes_rep_word, "idx": nodes_rep_idx}) json.dump(nodes_rep_map, open("nodes_rep_words_idx.json", "w")) json.dump(self.word2index, open("word2index.json", "w")) #exit() return nodes_rep def
self.word2index = {'<PAD>':0, '<UNK>':1, '<GO>':2, '<EOS>':3} stopwords_self = set() for i, (key, value) in enumerate(wordssorted): if value <= 5: break
random_line_split
DataManager.py
else: wordscount[word] = 1 wordssorted = sorted(wordscount.items(), key = lambda d: (d[1],d[0]), reverse=True) output = open("word_cnt_stat.txt", "w") for i, (key, value) in enumerate(wordssorted): output.write(str(value) + ":" + str(key) + "\n") self.word2index = {'<PAD>':0, '<UNK>':1, '<GO>':2, '<EOS>':3} stopwords_self = set() for i, (key, value) in enumerate(wordssorted): if value <= 5: break self.word2index[key] = i + 4 #PAD,UNK,GO,EOS if value >= stopword_freq_lb: stopwords_self.add(key) # to add all entity name into vocab entity_list = json.load(open("./data/entity_list_simple.json")) start_idx = len(self.word2index) for entity_name in entity_list: entity_name = entity_name.split("::")[-1] if entity_name not in self.word2index: self.word2index[entity_name] = start_idx start_idx += 1 self.stop_words_index = set([PAD, UNK, GO, EOS]) #self.stop_words_index |= set([self.word2index[word] for word in STOP_WORDS # if word in self.word2index]) # here we add all words into stopword list self.stop_words_index |= set([self.word2index[word] for word in stopwords_self]) self.index2word = dict((v, k) for k, v in self.word2index.items()) #load word vector if no_pretrain_word2vec: self.vector = None else: self.vector = 0.1 * np.random.rand(len(self.word2index), dim) with open("{0}/vector.txt".format(path)) as fl: for line in fl: vec = line.strip().split() word = vec[0].lower() vec = list(map(float, vec[1:])) if word in self.word2index: self.vector[self.word2index[word]] = np.asarray(vec) self.vector = torch.Tensor(self.vector) # compute tf len_voc = len(self.word2index.values()) self.index2nonstop = {} cnt = 0 for i in range(len_voc): if i not in self.stop_words_index: self.index2nonstop[i] = cnt cnt += 1 # for graph initialization self.node_id_map, self.id_nodename_map = self.get_node_id_map() self.node_info_map, self.nodename_attr_map = self.get_node_info() self.adj = self.get_adj_mat("./data/adj_simple.json", self.node_id_map) self.nodes_rep = self.get_nodes_rep(self.node_id_map, self.node_info_map) self.n_entity = len(self.node_id_map) #get index self.data = {} for name in ["train", "valid", "test"]: self.data[name] = [] for number, item in enumerate(self.text[name]): len_u = len(item[0]) indices = [[], [[] for _ in range(len_u)], [], [], [], []] #src_len, src, trg, trg_entities, trg_entities_mask indices[0] = [u.count(' ')+1 for u in item[0]] # on purpose max_u_len = max(indices[0]) # history for i in range(len_u): words = item[0][i].split() indices[1][i] = [self.word2index[word] if word in self.word2index else UNK for word in words] + [PAD] * (max_u_len - len(words)) # answer words = item[1].split() #print("item1:: ", len(words)) indices[2] = [self.word2index[word] if word in self.word2index else UNK for word in words] indices[2].append(EOS) # answer entity entities = item[2].split() #print("item2 entities:: ", len(entities)) indices[3] = [self.node_id_map[entity_name] for entity_name in entities] indices[3].append(0) indices[4] = [] for x in indices[3]: if x == 0: indices[4].append(0) else: indices[4].append(1) # ansuer original sentence words = item[3].split() indices[5] = words indices[5].append("<EOS>") if len(indices[2]) != len(indices[3]): print(number, len(indices[2]), len(indices[3])) print(item[1]) print(item[2]) exit() self.data[name].append(indices) def get_node_info(self): node_info_map = json.load(open("./data/entity_info.json")) nodename_attr_map = {} for node, res in node_info_map.items(): node_name = node.split("::")[-1] nodename_attr_map[node_name] = res return node_info_map, nodename_attr_map def post_process(self, outputs, pred_ents, topK=1): outputs = outputs.cpu().numpy().tolist() pred_ents = pred_ents.cpu().numpy() entity_attr_list = { "[attraction_address]", "[restaurant_address]", "[attraction_phone]", "[restaurant_phone]", "[hotel_address]", "[restaurant_postcode]", "[attraction_postcode]", "[hotel_phone]", "[hotel_postcode]", "[hospital_phone]" } lens_new = [] for i, out in enumerate(outputs): new_out = [] for j, each in enumerate(out): if self.index2word[each] == "<$>": pred_ent = np.argmax(pred_ents[i][j]) nodename = self.id_nodename_map[pred_ent] new_out.append(nodename) elif self.index2word[each] in entity_attr_list: attr_name = self.index2word[each] cnt = 0 suc_flag = False for idx, prob in sorted(enumerate(pred_ents[i][j]), key=lambda i: i[1], reverse=True): if suc_flag or cnt >= topK: break nodename = self.id_nodename_map[idx] if nodename not in self.nodename_attr_map: cnt += 1 continue for attr, val in self.nodename_attr_map[nodename].items(): if attr in attr_name: new_out.append(val) suc_flag = True break cnt += 1 if not suc_flag: new_out.append("<UNK>") else: new_out.append(self.index2word[each]) """ if each == self.word2index["<$>"]: pred_ent = np.argmax(pred_ents[i][j]) nodename = self.id_nodename_map[pred_ent] nodename_wordids = [self.word2index[x] for x in nodename.split()] new_out += nodename_wordids else: new_out.append(each) """ outputs[i] = new_out return outputs def get_nodes_rep(self, node_id_map, node_info_map, max_len=50): nodes_rep = [] nodes_rep_map = [] for name, id_ in sorted(node_id_map.items(), key=lambda i: i[1]): if name == "none" and id_ == 0: nodes_rep.append([PAD] * max_len) nodes_rep_map.append({"words": ["none"], "idx": [0]}) continue # the attributes used to build relationship # attributes as nodes: {"pricerange", "area", "food"} # attributes only as relation: {"internet", "parking", "stars", "attraction_type", "hotel_type"} # only user node name as node's feature name = name.split("::")[-1] node_desc = [name] nodes_rep_idx = [PAD] * max_len nodes_rep_idx[0] = self.word2index[name] nodes_rep_word = [name] """ for attr, val in node_info_map.items(): #if attr in {"address", "area", "pricerange", "introduction", "food", "stars"} or "type" in attr: if attr == "introduction": node_desc.append(val) node_desc = " ".join(node_desc) nodes_rep_idx = [] nodes_rep_word = [] for each_word in node_desc.split(): for word in re.split(r'[\[\](::)_]', each_word): if word == "": continue else: if word not in self.word2index: continue else: word_idx = self.word2index[word] nodes_rep_idx.append(word_idx) nodes_rep_word.append(word) len_ = len(nodes_rep_idx) if len_ >= max_len: nodes_rep_idx = nodes_rep_idx[0:max_len] nodes_rep_word = nodes_rep_word[0:max_len]
wordscount[word] += 1
conditional_block
DataManager.py
cnt += 1 continue for attr, val in self.nodename_attr_map[nodename].items(): if attr in attr_name: new_out.append(val) suc_flag = True break cnt += 1 if not suc_flag: new_out.append("<UNK>") else: new_out.append(self.index2word[each]) """ if each == self.word2index["<$>"]: pred_ent = np.argmax(pred_ents[i][j]) nodename = self.id_nodename_map[pred_ent] nodename_wordids = [self.word2index[x] for x in nodename.split()] new_out += nodename_wordids else: new_out.append(each) """ outputs[i] = new_out return outputs def get_nodes_rep(self, node_id_map, node_info_map, max_len=50): nodes_rep = [] nodes_rep_map = [] for name, id_ in sorted(node_id_map.items(), key=lambda i: i[1]): if name == "none" and id_ == 0: nodes_rep.append([PAD] * max_len) nodes_rep_map.append({"words": ["none"], "idx": [0]}) continue # the attributes used to build relationship # attributes as nodes: {"pricerange", "area", "food"} # attributes only as relation: {"internet", "parking", "stars", "attraction_type", "hotel_type"} # only user node name as node's feature name = name.split("::")[-1] node_desc = [name] nodes_rep_idx = [PAD] * max_len nodes_rep_idx[0] = self.word2index[name] nodes_rep_word = [name] """ for attr, val in node_info_map.items(): #if attr in {"address", "area", "pricerange", "introduction", "food", "stars"} or "type" in attr: if attr == "introduction": node_desc.append(val) node_desc = " ".join(node_desc) nodes_rep_idx = [] nodes_rep_word = [] for each_word in node_desc.split(): for word in re.split(r'[\[\](::)_]', each_word): if word == "": continue else: if word not in self.word2index: continue else: word_idx = self.word2index[word] nodes_rep_idx.append(word_idx) nodes_rep_word.append(word) len_ = len(nodes_rep_idx) if len_ >= max_len: nodes_rep_idx = nodes_rep_idx[0:max_len] nodes_rep_word = nodes_rep_word[0:max_len] else: nodes_rep_idx += [PAD] * (max_len - len_) """ nodes_rep.append(nodes_rep_idx) nodes_rep_map.append({"words": nodes_rep_word, "idx": nodes_rep_idx}) json.dump(nodes_rep_map, open("nodes_rep_words_idx.json", "w")) json.dump(self.word2index, open("word2index.json", "w")) #exit() return nodes_rep def get_node_id_map(self): data = json.load(open("./data/entity_list_simple.json")) node_id_map = {} id_nodename_map = {} for i, node in enumerate(data): node_id_map[node] = i + 1 tmp = node.split("::") #node_name = " ".join(tmp[1].split("_")) node_name = tmp[1] id_nodename_map[i+1] = node_name node_id_map["none"] = 0 id_nodename_map[0] = "" return node_id_map, id_nodename_map def get_adj_mat(self, input_file, item_id_map): adj = json.load(open(input_file)) new_adj = {} for i, neibors in adj.items(): i_idx = item_id_map[i] new_adj[i_idx] = [] for j in neibors: j_idx = item_id_map[j] new_adj[i_idx].append(j_idx) new_adj = nx.adjacency_matrix(nx.from_dict_of_lists(new_adj)) new_adj = self.normalize_adj(new_adj + sp.eye(new_adj.shape[0])) new_adj = torch.FloatTensor(np.array(new_adj.todense())) return new_adj def normalize_adj(self, mx): """Row-normalize sparse matrix""" rowsum = np.array(mx.sum(1)) r_inv_sqrt = np.power(rowsum, -0.5).flatten() r_inv_sqrt[np.isinf(r_inv_sqrt)] = 0. r_mat_inv_sqrt = sp.diags(r_inv_sqrt) return mx.dot(r_mat_inv_sqrt).transpose().dot(r_mat_inv_sqrt).tocoo() def create_dataset(self, name, batch_size): datas = self.data[name] src_seq_lens = [] src_seqs, trg_seqs = [], [] trg_stops, src_tfs = [], [] trg_ents, trg_ents_mask, trg_seqs_ori = [], [], [] nonstop_voc_size = len(self.index2nonstop) for item in datas: src_len, src, trg, ents, ents_mask, trg_ori = item tensor_src_len, tensor_src, tensor_trg = torch.LongTensor(src_len), \ torch.LongTensor(src), torch.LongTensor(trg) src_seq_lens.append(tensor_src_len) src_seqs.append(tensor_src) trg_seqs.append(tensor_trg) trg_stop = torch.zeros_like(tensor_trg) for i, index in enumerate(trg): if index in self.stop_words_index: trg_stop[i] = 1 trg_stops.append(trg_stop) src_tf = torch.zeros(nonstop_voc_size) for j, uttr in enumerate(src): for i, index in enumerate(uttr): if i == src_len[j]: break if index not in self.stop_words_index: src_tf[self.index2nonstop[index]] += 1 if src_tf.sum().item() > 0: src_tf /= src_tf.sum() src_tfs.append(src_tf) trg_ents.append(torch.LongTensor(ents)) trg_ents_mask.append(torch.LongTensor(ents_mask)) trg_seqs_ori.append(trg_ori) print(len(trg_stops), len(trg_seqs), len(trg_ents), len(trg_seqs_ori)) dataset = Dataset(src_seq_lens, src_seqs, trg_seqs, trg_stops, src_tfs, trg_ents, trg_ents_mask, trg_seqs_ori) dataloader = data.DataLoader(dataset, batch_size, True, num_workers=0, collate_fn=pad_packed_collate) return dataloader def compute_stopword(self, y): res = torch.zeros_like(y).to(device=device) for i, row in enumerate(y): words_index = row.tolist() res[i] = torch.LongTensor([int(index in self.stop_words_index) for index in words_index]) return res def interpret(self, preds, refs, lens, f): i = random.randrange(0, len(lens)) l = max(lens) for j in range(l): word = self.index2word[preds[i][j].item()] print(word, end=' ') f.write('{0} '.format(word)) if word == '<EOS>': break print() f.write('\n') l = lens[i] for j in range(l): word = self.index2word[refs[i][j].item()] print(word, end=' ') f.write('{0} '.format(word)) print() f.write('\n') class Dataset(data.Dataset): def __init__(self, src_seq_lens, src_seqs, trg_seqs, trg_stops, src_tfs, trg_ents, trg_ents_mask, trg_seqs_ori): self.src_seq_lens = src_seq_lens self.src_seqs = src_seqs self.trg_seqs = trg_seqs self.trg_stops = trg_stops self.src_tfs = src_tfs self.num_total_seqs = len(src_seqs) self.trg_ents = trg_ents self.trg_ents_mask = trg_ents_mask self.trg_seqs_ori = trg_seqs_ori def __getitem__(self, index): src_seq_len = self.src_seq_lens[index] src_seq = self.src_seqs[index] trg_seq = self.trg_seqs[index] trg_stop = self.trg_stops[index] src_tf = self.src_tfs[index] trg_ent = self.trg_ents[index] trg_ent_mask = self.trg_ents_mask[index] trg_seq_ori = self.trg_seqs_ori[index] return src_seq_len, src_seq, trg_seq, trg_stop, src_tf, trg_ent, trg_ent_mask, trg_seq_ori def __len__(self): return self.num_total_seqs def pad_packed_collate(batch_data): def merge(sequences): lengths = [len(seq) for seq in sequences] padded_seqs = torch.zeros(len(sequences), max(lengths)).long() for i, seq in enumerate(sequences): end = lengths[i] padded_seqs[i, :end] = seq return padded_seqs, lengths def
hierarchical_merge
identifier_name
compression_utils.py
stochastic rounding rounding. If `None`, defaults to `l2_norm(x)`. beta: A constant in [0, 1) controlling the concentration inequality for the probabilistic norm bound after rounding. Returns: The rounded tensor. """ def post_rounding_l2_norm_bound(x, l2_norm_bound, beta): """Computes the L2 norm bound of a vector after rounding (Thm. 1, Eq. 2).""" beta = tf.cast(beta, x.dtype) dim = tf.cast(tf.size(x), x.dtype) if l2_norm_bound is None: x_norm = tf.norm(x, ord=2) else: x_norm = tf.cast(l2_norm_bound, x.dtype) # We consider 2 (scaled) norm bounds and take the min (Proposition 22). bound1 = x_norm + tf.sqrt(dim) squared_bound2 = tf.square(x_norm) + 0.25 * dim squared_bound2 += ( tf.sqrt(2.0 * tf.math.log(1.0 / beta)) * (x_norm + 0.5 * tf.sqrt(dim))) bound2 = tf.sqrt(squared_bound2) # bound2 is inf if beta = 0, in which case we fall back to bound1. return tf.minimum(bound1, bound2) conditional = tf.cast(conditional, tf.bool) l2_norm_threshold = post_rounding_l2_norm_bound(x, l2_norm_bound, beta) floored_x = tf.floor(x) decimal_x = x - floored_x def round_fn(repeat, _): # 1. Try stochastic rounding on input (ignore previous iterations' outputs). uniform = tf.random.uniform(tf.shape(x), dtype=x.dtype, minval=0, maxval=1) bernoulli = uniform < decimal_x rounded_x = floored_x + tf.cast(bernoulli, x.dtype) # 2. Try again if the rounded vector has excessive L2 norm. rounded_l2_norm = tf.norm(rounded_x, ord=2) repeat = tf.logical_and(conditional, tf.greater(rounded_l2_norm, l2_norm_threshold)) return [repeat, rounded_x] repeat = tf.constant(True) _, result_x = tf.while_loop( cond=lambda r, _: r, body=round_fn, loop_vars=[repeat, x]) return result_x def scaled_quantization(x, scale, stochastic, conditional, l2_norm_bound, beta=DEFAULT_BETA): """Scales the tensors and rounds to integers.""" scale = tf.cast(scale, x.dtype) l2_norm_bound = tf.cast(l2_norm_bound, x.dtype) scaled_x = x * scale scaled_bound = l2_norm_bound * scale quantized_x = tf.cond( tf.cast(stochastic, tf.bool), lambda: stochastic_rounding(scaled_x, conditional, scaled_bound, beta), lambda: tf.round(scaled_x)) return quantized_x def inverse_scaled_quantization(x, scale): """Restores the value range of `x` from `scaled_quantization`.""" return x / tf.cast(scale, x.dtype) def flatten_concat(structure): """Flattens each tensor in the structure and concats them as a vector. Each tensor within the structure should have rank >= 1 (i.e. no scalars). Args: structure: The input structure of tensors. Returns: The flattened and concatenated component tensors as a tf.Tensor with shape (d,) where `d` is the total number of elements in the structure. """ flattened_as_list = [] for x in tf.nest.flatten(structure): with tf.control_dependencies([tf.debugging.assert_rank_at_least(x, 1)]): flattened_as_list.append(tf.reshape(x, [-1])) return tf.concat(flattened_as_list, axis=0) def inverse_flatten_concat(flat_vector, original_structure): """Applies the inverse of `flatten_concat` given the original structure.""" location, split_tensors = 0, [] for orig_t in tf.nest.flatten(original_structure): length = tf.size(orig_t) split_vector = tf.slice(flat_vector, [location], [length]) split_tensors.append(tf.reshape(split_vector, orig_t.shape)) location += length return tf.nest.pack_sequence_as(original_structure, split_tensors) def sample_rademacher(shape, dtype, seed_pair): """Sample uniform random +1/-1 values with specified shape/dtype/seed_pair.""" rand_uniform = tf.random.stateless_uniform(shape=shape, seed=seed_pair) return tf.cast(tf.sign(rand_uniform - 0.5), dtype) def pad_zeros(x): """Pads a vector with shape (d,) with zeros to the next power of two.""" dim = tf.shape(x)[0] log2_dim = tf.math.log(tf.cast(dim, tf.float32)) / tf.math.log(2.0) pad_dim = tf.pow(2, tf.cast(tf.math.ceil(log2_dim), tf.int32)) with tf.control_dependencies([tf.debugging.assert_rank(x, 1)]): return tf.pad(x, [[0, tf.maximum(0, pad_dim - dim)]]) def randomized_hadamard_transform(x, seed_pair, repeat=1): """Applies randomized Hadamard transform to a vector with the given seed. Args: x: The input vector. seed_pair: The seed pair for generating randomness. repeat: Number of times to repeat the randomized Hadamard transform. Returns: The transformed vector. """ def
(repeat_index, x): # All sources of randomness depend on the input seed. cur_seed = seed_pair + repeat_index # Randomly flip signs. signs = sample_rademacher(tf.shape(x), dtype=x.dtype, seed_pair=cur_seed) rademacher_x = signs * x # Apply Hadamard (+ expand/squeeze dims). encoded_x = tf.squeeze( fast_walsh_hadamard_transform(tf.expand_dims(rademacher_x, axis=0)), axis=0) return encoded_x tf.debugging.assert_type(x, tf.float32) padded_x = pad_zeros(x) # Hadamard transform requires vectors with 2^n dims. i, result_x = tf.constant(0), padded_x cond_fn = lambda i, _: tf.less(i, repeat) body_fn = lambda i, x: [tf.add(i, 1), apply_transform(i, x)] _, result_x = tf.while_loop(cond_fn, body_fn, [i, result_x]) return result_x def inverse_randomized_hadamard_transform(x, original_dim, seed_pair, repeat=1): """Applies inverse of `randomized_hadamard_transform` with the given seed. Args: x: The transformed vector. original_dim: The dimension of the original vector. seed_pair: The same seed pair used in the forward transform. repeat: Number of times the randomized Hadamard transform was applied. Returns: The original vector. """ def inverse_transform(repeat_index, x): # All sources of randomness depend on the input seed. cur_seed = seed_pair + repeat_index # Apply Hadamard. unrotated_x = fast_walsh_hadamard_transform(tf.expand_dims(x, axis=0)) unrotated_x = tf.squeeze(unrotated_x, axis=0) # Unflip signs. signs = sample_rademacher( tf.shape(unrotated_x), dtype=x.dtype, seed_pair=cur_seed) decoded_x = signs * unrotated_x return decoded_x # Repeat inverse transforms (with reversed indices). tf.debugging.assert_type(x, tf.float32) i, result_x = tf.constant(repeat - 1), x cond_fn = lambda i, _: tf.greater_equal(i, 0) body_fn = lambda i, x: [tf.subtract(i, 1), inverse_transform(i, x)] _, result_x = tf.while_loop(cond_fn, body_fn, [i, result_x]) # Unpad zeros from forward transform. return result_x[:original_dim] def fast_walsh_hadamard_transform(x): """Applies the fast Walsh-Hadamard transform to a set of vectors. This method uses a composition of existing TensorFlow operations to implement the transform. This function is forked from https://github.com/tensorflow/model-optimization. Args: x: A `Tensor`. Must be of shape `[a, b]`, where `a` can be anything (not necessarily known), and `b` must be a power of two, not required to be statically known. Returns: A `Tensor` of shape `[a, b]`, where `[i, :]` is the product `x[i, :]*H`, where `H` is the Hadamard matrix. Raises: ValueError: If the input is not rank 2 `Tensor`, and if the second dimension is statically known and is not a power of two. OpError: If the second dimension is not statically known and is not a power of two. Note that in graph execution, this error is not raised during the execution of the Python function, but during execution of
apply_transform
identifier_name
compression_utils.py
]) return result_x def scaled_quantization(x, scale, stochastic, conditional, l2_norm_bound, beta=DEFAULT_BETA): """Scales the tensors and rounds to integers.""" scale = tf.cast(scale, x.dtype) l2_norm_bound = tf.cast(l2_norm_bound, x.dtype) scaled_x = x * scale scaled_bound = l2_norm_bound * scale quantized_x = tf.cond( tf.cast(stochastic, tf.bool), lambda: stochastic_rounding(scaled_x, conditional, scaled_bound, beta), lambda: tf.round(scaled_x)) return quantized_x def inverse_scaled_quantization(x, scale): """Restores the value range of `x` from `scaled_quantization`.""" return x / tf.cast(scale, x.dtype) def flatten_concat(structure): """Flattens each tensor in the structure and concats them as a vector. Each tensor within the structure should have rank >= 1 (i.e. no scalars). Args: structure: The input structure of tensors. Returns: The flattened and concatenated component tensors as a tf.Tensor with shape (d,) where `d` is the total number of elements in the structure. """ flattened_as_list = [] for x in tf.nest.flatten(structure): with tf.control_dependencies([tf.debugging.assert_rank_at_least(x, 1)]): flattened_as_list.append(tf.reshape(x, [-1])) return tf.concat(flattened_as_list, axis=0) def inverse_flatten_concat(flat_vector, original_structure): """Applies the inverse of `flatten_concat` given the original structure.""" location, split_tensors = 0, [] for orig_t in tf.nest.flatten(original_structure): length = tf.size(orig_t) split_vector = tf.slice(flat_vector, [location], [length]) split_tensors.append(tf.reshape(split_vector, orig_t.shape)) location += length return tf.nest.pack_sequence_as(original_structure, split_tensors) def sample_rademacher(shape, dtype, seed_pair): """Sample uniform random +1/-1 values with specified shape/dtype/seed_pair.""" rand_uniform = tf.random.stateless_uniform(shape=shape, seed=seed_pair) return tf.cast(tf.sign(rand_uniform - 0.5), dtype) def pad_zeros(x): """Pads a vector with shape (d,) with zeros to the next power of two.""" dim = tf.shape(x)[0] log2_dim = tf.math.log(tf.cast(dim, tf.float32)) / tf.math.log(2.0) pad_dim = tf.pow(2, tf.cast(tf.math.ceil(log2_dim), tf.int32)) with tf.control_dependencies([tf.debugging.assert_rank(x, 1)]): return tf.pad(x, [[0, tf.maximum(0, pad_dim - dim)]]) def randomized_hadamard_transform(x, seed_pair, repeat=1): """Applies randomized Hadamard transform to a vector with the given seed. Args: x: The input vector. seed_pair: The seed pair for generating randomness. repeat: Number of times to repeat the randomized Hadamard transform. Returns: The transformed vector. """ def apply_transform(repeat_index, x): # All sources of randomness depend on the input seed. cur_seed = seed_pair + repeat_index # Randomly flip signs. signs = sample_rademacher(tf.shape(x), dtype=x.dtype, seed_pair=cur_seed) rademacher_x = signs * x # Apply Hadamard (+ expand/squeeze dims). encoded_x = tf.squeeze( fast_walsh_hadamard_transform(tf.expand_dims(rademacher_x, axis=0)), axis=0) return encoded_x tf.debugging.assert_type(x, tf.float32) padded_x = pad_zeros(x) # Hadamard transform requires vectors with 2^n dims. i, result_x = tf.constant(0), padded_x cond_fn = lambda i, _: tf.less(i, repeat) body_fn = lambda i, x: [tf.add(i, 1), apply_transform(i, x)] _, result_x = tf.while_loop(cond_fn, body_fn, [i, result_x]) return result_x def inverse_randomized_hadamard_transform(x, original_dim, seed_pair, repeat=1): """Applies inverse of `randomized_hadamard_transform` with the given seed. Args: x: The transformed vector. original_dim: The dimension of the original vector. seed_pair: The same seed pair used in the forward transform. repeat: Number of times the randomized Hadamard transform was applied. Returns: The original vector. """ def inverse_transform(repeat_index, x): # All sources of randomness depend on the input seed. cur_seed = seed_pair + repeat_index # Apply Hadamard. unrotated_x = fast_walsh_hadamard_transform(tf.expand_dims(x, axis=0)) unrotated_x = tf.squeeze(unrotated_x, axis=0) # Unflip signs. signs = sample_rademacher( tf.shape(unrotated_x), dtype=x.dtype, seed_pair=cur_seed) decoded_x = signs * unrotated_x return decoded_x # Repeat inverse transforms (with reversed indices). tf.debugging.assert_type(x, tf.float32) i, result_x = tf.constant(repeat - 1), x cond_fn = lambda i, _: tf.greater_equal(i, 0) body_fn = lambda i, x: [tf.subtract(i, 1), inverse_transform(i, x)] _, result_x = tf.while_loop(cond_fn, body_fn, [i, result_x]) # Unpad zeros from forward transform. return result_x[:original_dim] def fast_walsh_hadamard_transform(x): """Applies the fast Walsh-Hadamard transform to a set of vectors. This method uses a composition of existing TensorFlow operations to implement the transform. This function is forked from https://github.com/tensorflow/model-optimization. Args: x: A `Tensor`. Must be of shape `[a, b]`, where `a` can be anything (not necessarily known), and `b` must be a power of two, not required to be statically known. Returns: A `Tensor` of shape `[a, b]`, where `[i, :]` is the product `x[i, :]*H`, where `H` is the Hadamard matrix. Raises: ValueError: If the input is not rank 2 `Tensor`, and if the second dimension is statically known and is not a power of two. OpError: If the second dimension is not statically known and is not a power of two. Note that in graph execution, this error is not raised during the execution of the Python function, but during execution of the resulting computation. """ with tf.compat.v1.name_scope(None, 'fast_walsh_hadamard_transform'): # Validate input. x = tf.convert_to_tensor(x) if x.shape.ndims != 2: raise ValueError('Number of dimensions of x must be 2. Shape of x: %s' % x.shape) original_x_shape = x.shape.as_list() dim = x.shape.as_list()[-1] if dim is None: # dim is not statically known. dim = tf.shape(x)[-1] log2 = tf.cast( tf.math.round( tf.math.log(tf.cast(dim, tf.float32)) / tf.math.log(2.)), tf.int32) with tf.control_dependencies([ tf.compat.v1.assert_equal( dim, tf.math.pow(2, log2), message='The dimension of x must be a power of two.' 'Provided dimension is: %s' % dim) ]): x = tf.identity(x) else: # dim is statically known. if not (dim and ((dim & (dim - 1)) == 0)): raise ValueError('The dimension of x must be a power of two. ' 'Provided dimension is: %s' % dim) log2 = int(np.ceil(np.log2(dim))) if dim == 1: # Equivalent to identity. return tf.identity(x) h_core = tf.constant([[1., 1.], [1., -1.]], dtype=x.dtype, name='hadamard_weights_2x2') permutation = tf.constant([0, 2, 1], name='hadamard_permutation') # A step of the fast Walsh-Hadamard algorithm. def _hadamard_step(x, dim): """A single step in the fast Walsh-Hadamard transform.""" x_shape = x.shape.as_list() x = tf.reshape(x, [-1, 2]) # Reshape so that we have a matrix. x = tf.matmul(x, h_core) # Multiply. x = tf.reshape(x, [-1, dim // 2, 2]) # Reshape to rank-3. x = tf.transpose(x, perm=permutation) # Swap last two dimensions.
x.set_shape(x_shape) # Failed shape inference in tf.while_loop. return x
random_line_split
compression_utils.py
stochastic rounding rounding. If `None`, defaults to `l2_norm(x)`. beta: A constant in [0, 1) controlling the concentration inequality for the probabilistic norm bound after rounding. Returns: The rounded tensor. """ def post_rounding_l2_norm_bound(x, l2_norm_bound, beta): """Computes the L2 norm bound of a vector after rounding (Thm. 1, Eq. 2).""" beta = tf.cast(beta, x.dtype) dim = tf.cast(tf.size(x), x.dtype) if l2_norm_bound is None: x_norm = tf.norm(x, ord=2) else: x_norm = tf.cast(l2_norm_bound, x.dtype) # We consider 2 (scaled) norm bounds and take the min (Proposition 22). bound1 = x_norm + tf.sqrt(dim) squared_bound2 = tf.square(x_norm) + 0.25 * dim squared_bound2 += ( tf.sqrt(2.0 * tf.math.log(1.0 / beta)) * (x_norm + 0.5 * tf.sqrt(dim))) bound2 = tf.sqrt(squared_bound2) # bound2 is inf if beta = 0, in which case we fall back to bound1. return tf.minimum(bound1, bound2) conditional = tf.cast(conditional, tf.bool) l2_norm_threshold = post_rounding_l2_norm_bound(x, l2_norm_bound, beta) floored_x = tf.floor(x) decimal_x = x - floored_x def round_fn(repeat, _): # 1. Try stochastic rounding on input (ignore previous iterations' outputs). uniform = tf.random.uniform(tf.shape(x), dtype=x.dtype, minval=0, maxval=1) bernoulli = uniform < decimal_x rounded_x = floored_x + tf.cast(bernoulli, x.dtype) # 2. Try again if the rounded vector has excessive L2 norm. rounded_l2_norm = tf.norm(rounded_x, ord=2) repeat = tf.logical_and(conditional, tf.greater(rounded_l2_norm, l2_norm_threshold)) return [repeat, rounded_x] repeat = tf.constant(True) _, result_x = tf.while_loop( cond=lambda r, _: r, body=round_fn, loop_vars=[repeat, x]) return result_x def scaled_quantization(x, scale, stochastic, conditional, l2_norm_bound, beta=DEFAULT_BETA): """Scales the tensors and rounds to integers.""" scale = tf.cast(scale, x.dtype) l2_norm_bound = tf.cast(l2_norm_bound, x.dtype) scaled_x = x * scale scaled_bound = l2_norm_bound * scale quantized_x = tf.cond( tf.cast(stochastic, tf.bool), lambda: stochastic_rounding(scaled_x, conditional, scaled_bound, beta), lambda: tf.round(scaled_x)) return quantized_x def inverse_scaled_quantization(x, scale): """Restores the value range of `x` from `scaled_quantization`.""" return x / tf.cast(scale, x.dtype) def flatten_concat(structure): """Flattens each tensor in the structure and concats them as a vector. Each tensor within the structure should have rank >= 1 (i.e. no scalars). Args: structure: The input structure of tensors. Returns: The flattened and concatenated component tensors as a tf.Tensor with shape (d,) where `d` is the total number of elements in the structure. """ flattened_as_list = [] for x in tf.nest.flatten(structure): with tf.control_dependencies([tf.debugging.assert_rank_at_least(x, 1)]): flattened_as_list.append(tf.reshape(x, [-1])) return tf.concat(flattened_as_list, axis=0) def inverse_flatten_concat(flat_vector, original_structure): """Applies the inverse of `flatten_concat` given the original structure.""" location, split_tensors = 0, [] for orig_t in tf.nest.flatten(original_structure): length = tf.size(orig_t) split_vector = tf.slice(flat_vector, [location], [length]) split_tensors.append(tf.reshape(split_vector, orig_t.shape)) location += length return tf.nest.pack_sequence_as(original_structure, split_tensors) def sample_rademacher(shape, dtype, seed_pair): """Sample uniform random +1/-1 values with specified shape/dtype/seed_pair.""" rand_uniform = tf.random.stateless_uniform(shape=shape, seed=seed_pair) return tf.cast(tf.sign(rand_uniform - 0.5), dtype) def pad_zeros(x): """Pads a vector with shape (d,) with zeros to the next power of two.""" dim = tf.shape(x)[0] log2_dim = tf.math.log(tf.cast(dim, tf.float32)) / tf.math.log(2.0) pad_dim = tf.pow(2, tf.cast(tf.math.ceil(log2_dim), tf.int32)) with tf.control_dependencies([tf.debugging.assert_rank(x, 1)]): return tf.pad(x, [[0, tf.maximum(0, pad_dim - dim)]]) def randomized_hadamard_transform(x, seed_pair, repeat=1):
axis=0) return encoded_x tf.debugging.assert_type(x, tf.float32) padded_x = pad_zeros(x) # Hadamard transform requires vectors with 2^n dims. i, result_x = tf.constant(0), padded_x cond_fn = lambda i, _: tf.less(i, repeat) body_fn = lambda i, x: [tf.add(i, 1), apply_transform(i, x)] _, result_x = tf.while_loop(cond_fn, body_fn, [i, result_x]) return result_x def inverse_randomized_hadamard_transform(x, original_dim, seed_pair, repeat=1): """Applies inverse of `randomized_hadamard_transform` with the given seed. Args: x: The transformed vector. original_dim: The dimension of the original vector. seed_pair: The same seed pair used in the forward transform. repeat: Number of times the randomized Hadamard transform was applied. Returns: The original vector. """ def inverse_transform(repeat_index, x): # All sources of randomness depend on the input seed. cur_seed = seed_pair + repeat_index # Apply Hadamard. unrotated_x = fast_walsh_hadamard_transform(tf.expand_dims(x, axis=0)) unrotated_x = tf.squeeze(unrotated_x, axis=0) # Unflip signs. signs = sample_rademacher( tf.shape(unrotated_x), dtype=x.dtype, seed_pair=cur_seed) decoded_x = signs * unrotated_x return decoded_x # Repeat inverse transforms (with reversed indices). tf.debugging.assert_type(x, tf.float32) i, result_x = tf.constant(repeat - 1), x cond_fn = lambda i, _: tf.greater_equal(i, 0) body_fn = lambda i, x: [tf.subtract(i, 1), inverse_transform(i, x)] _, result_x = tf.while_loop(cond_fn, body_fn, [i, result_x]) # Unpad zeros from forward transform. return result_x[:original_dim] def fast_walsh_hadamard_transform(x): """Applies the fast Walsh-Hadamard transform to a set of vectors. This method uses a composition of existing TensorFlow operations to implement the transform. This function is forked from https://github.com/tensorflow/model-optimization. Args: x: A `Tensor`. Must be of shape `[a, b]`, where `a` can be anything (not necessarily known), and `b` must be a power of two, not required to be statically known. Returns: A `Tensor` of shape `[a, b]`, where `[i, :]` is the product `x[i, :]*H`, where `H` is the Hadamard matrix. Raises: ValueError: If the input is not rank 2 `Tensor`, and if the second dimension is statically known and is not a power of two. OpError: If the second dimension is not statically known and is not a power of two. Note that in graph execution, this error is not raised during the execution of the Python function, but during execution of the
"""Applies randomized Hadamard transform to a vector with the given seed. Args: x: The input vector. seed_pair: The seed pair for generating randomness. repeat: Number of times to repeat the randomized Hadamard transform. Returns: The transformed vector. """ def apply_transform(repeat_index, x): # All sources of randomness depend on the input seed. cur_seed = seed_pair + repeat_index # Randomly flip signs. signs = sample_rademacher(tf.shape(x), dtype=x.dtype, seed_pair=cur_seed) rademacher_x = signs * x # Apply Hadamard (+ expand/squeeze dims). encoded_x = tf.squeeze( fast_walsh_hadamard_transform(tf.expand_dims(rademacher_x, axis=0)),
identifier_body
compression_utils.py
the L2 norm bound of a vector after rounding (Thm. 1, Eq. 2).""" beta = tf.cast(beta, x.dtype) dim = tf.cast(tf.size(x), x.dtype) if l2_norm_bound is None: x_norm = tf.norm(x, ord=2) else: x_norm = tf.cast(l2_norm_bound, x.dtype) # We consider 2 (scaled) norm bounds and take the min (Proposition 22). bound1 = x_norm + tf.sqrt(dim) squared_bound2 = tf.square(x_norm) + 0.25 * dim squared_bound2 += ( tf.sqrt(2.0 * tf.math.log(1.0 / beta)) * (x_norm + 0.5 * tf.sqrt(dim))) bound2 = tf.sqrt(squared_bound2) # bound2 is inf if beta = 0, in which case we fall back to bound1. return tf.minimum(bound1, bound2) conditional = tf.cast(conditional, tf.bool) l2_norm_threshold = post_rounding_l2_norm_bound(x, l2_norm_bound, beta) floored_x = tf.floor(x) decimal_x = x - floored_x def round_fn(repeat, _): # 1. Try stochastic rounding on input (ignore previous iterations' outputs). uniform = tf.random.uniform(tf.shape(x), dtype=x.dtype, minval=0, maxval=1) bernoulli = uniform < decimal_x rounded_x = floored_x + tf.cast(bernoulli, x.dtype) # 2. Try again if the rounded vector has excessive L2 norm. rounded_l2_norm = tf.norm(rounded_x, ord=2) repeat = tf.logical_and(conditional, tf.greater(rounded_l2_norm, l2_norm_threshold)) return [repeat, rounded_x] repeat = tf.constant(True) _, result_x = tf.while_loop( cond=lambda r, _: r, body=round_fn, loop_vars=[repeat, x]) return result_x def scaled_quantization(x, scale, stochastic, conditional, l2_norm_bound, beta=DEFAULT_BETA): """Scales the tensors and rounds to integers.""" scale = tf.cast(scale, x.dtype) l2_norm_bound = tf.cast(l2_norm_bound, x.dtype) scaled_x = x * scale scaled_bound = l2_norm_bound * scale quantized_x = tf.cond( tf.cast(stochastic, tf.bool), lambda: stochastic_rounding(scaled_x, conditional, scaled_bound, beta), lambda: tf.round(scaled_x)) return quantized_x def inverse_scaled_quantization(x, scale): """Restores the value range of `x` from `scaled_quantization`.""" return x / tf.cast(scale, x.dtype) def flatten_concat(structure): """Flattens each tensor in the structure and concats them as a vector. Each tensor within the structure should have rank >= 1 (i.e. no scalars). Args: structure: The input structure of tensors. Returns: The flattened and concatenated component tensors as a tf.Tensor with shape (d,) where `d` is the total number of elements in the structure. """ flattened_as_list = [] for x in tf.nest.flatten(structure): with tf.control_dependencies([tf.debugging.assert_rank_at_least(x, 1)]): flattened_as_list.append(tf.reshape(x, [-1])) return tf.concat(flattened_as_list, axis=0) def inverse_flatten_concat(flat_vector, original_structure): """Applies the inverse of `flatten_concat` given the original structure.""" location, split_tensors = 0, [] for orig_t in tf.nest.flatten(original_structure): length = tf.size(orig_t) split_vector = tf.slice(flat_vector, [location], [length]) split_tensors.append(tf.reshape(split_vector, orig_t.shape)) location += length return tf.nest.pack_sequence_as(original_structure, split_tensors) def sample_rademacher(shape, dtype, seed_pair): """Sample uniform random +1/-1 values with specified shape/dtype/seed_pair.""" rand_uniform = tf.random.stateless_uniform(shape=shape, seed=seed_pair) return tf.cast(tf.sign(rand_uniform - 0.5), dtype) def pad_zeros(x): """Pads a vector with shape (d,) with zeros to the next power of two.""" dim = tf.shape(x)[0] log2_dim = tf.math.log(tf.cast(dim, tf.float32)) / tf.math.log(2.0) pad_dim = tf.pow(2, tf.cast(tf.math.ceil(log2_dim), tf.int32)) with tf.control_dependencies([tf.debugging.assert_rank(x, 1)]): return tf.pad(x, [[0, tf.maximum(0, pad_dim - dim)]]) def randomized_hadamard_transform(x, seed_pair, repeat=1): """Applies randomized Hadamard transform to a vector with the given seed. Args: x: The input vector. seed_pair: The seed pair for generating randomness. repeat: Number of times to repeat the randomized Hadamard transform. Returns: The transformed vector. """ def apply_transform(repeat_index, x): # All sources of randomness depend on the input seed. cur_seed = seed_pair + repeat_index # Randomly flip signs. signs = sample_rademacher(tf.shape(x), dtype=x.dtype, seed_pair=cur_seed) rademacher_x = signs * x # Apply Hadamard (+ expand/squeeze dims). encoded_x = tf.squeeze( fast_walsh_hadamard_transform(tf.expand_dims(rademacher_x, axis=0)), axis=0) return encoded_x tf.debugging.assert_type(x, tf.float32) padded_x = pad_zeros(x) # Hadamard transform requires vectors with 2^n dims. i, result_x = tf.constant(0), padded_x cond_fn = lambda i, _: tf.less(i, repeat) body_fn = lambda i, x: [tf.add(i, 1), apply_transform(i, x)] _, result_x = tf.while_loop(cond_fn, body_fn, [i, result_x]) return result_x def inverse_randomized_hadamard_transform(x, original_dim, seed_pair, repeat=1): """Applies inverse of `randomized_hadamard_transform` with the given seed. Args: x: The transformed vector. original_dim: The dimension of the original vector. seed_pair: The same seed pair used in the forward transform. repeat: Number of times the randomized Hadamard transform was applied. Returns: The original vector. """ def inverse_transform(repeat_index, x): # All sources of randomness depend on the input seed. cur_seed = seed_pair + repeat_index # Apply Hadamard. unrotated_x = fast_walsh_hadamard_transform(tf.expand_dims(x, axis=0)) unrotated_x = tf.squeeze(unrotated_x, axis=0) # Unflip signs. signs = sample_rademacher( tf.shape(unrotated_x), dtype=x.dtype, seed_pair=cur_seed) decoded_x = signs * unrotated_x return decoded_x # Repeat inverse transforms (with reversed indices). tf.debugging.assert_type(x, tf.float32) i, result_x = tf.constant(repeat - 1), x cond_fn = lambda i, _: tf.greater_equal(i, 0) body_fn = lambda i, x: [tf.subtract(i, 1), inverse_transform(i, x)] _, result_x = tf.while_loop(cond_fn, body_fn, [i, result_x]) # Unpad zeros from forward transform. return result_x[:original_dim] def fast_walsh_hadamard_transform(x): """Applies the fast Walsh-Hadamard transform to a set of vectors. This method uses a composition of existing TensorFlow operations to implement the transform. This function is forked from https://github.com/tensorflow/model-optimization. Args: x: A `Tensor`. Must be of shape `[a, b]`, where `a` can be anything (not necessarily known), and `b` must be a power of two, not required to be statically known. Returns: A `Tensor` of shape `[a, b]`, where `[i, :]` is the product `x[i, :]*H`, where `H` is the Hadamard matrix. Raises: ValueError: If the input is not rank 2 `Tensor`, and if the second dimension is statically known and is not a power of two. OpError: If the second dimension is not statically known and is not a power of two. Note that in graph execution, this error is not raised during the execution of the Python function, but during execution of the resulting computation. """ with tf.compat.v1.name_scope(None, 'fast_walsh_hadamard_transform'): # Validate input. x = tf.convert_to_tensor(x) if x.shape.ndims != 2:
raise ValueError('Number of dimensions of x must be 2. Shape of x: %s' % x.shape)
conditional_block
import_openapi.go
chemas/"):] } return oasSchema.Ref //? } return "" } func convertOasType(name string, oasSchema *Schema) (sadl.TypeSpec, error) { var err error var ts sadl.TypeSpec if oasSchema.Example != nil { ex := &sadl.ExampleDef{ Target: name, Example: oasSchema.Example, } examples = append(examples, ex) } switch oasSchema.Type { case "boolean": ts.Type = "Bool" case "string": if oasSchema.Enum != nil { //OAS defines element *descriptions* as the values, not symbolic identifiers. //so we look for the case where all values look like identifiers, and call that an enum. Else a strings with accepted "values" //perhaps the spirit of JSON Schema enums are just values, not what I think of as "enums", i.e. "a set of named values", per wikipedia. //still, with symbolic values, perhaps the intent is to use proper enums, if only JSON Schema had them. isEnum := EnumTypes var values []string for _, val := range oasSchema.Enum { if s, ok := val.(string); ok { values = append(values, s) if !sadl.IsSymbol(s) { isEnum = false } } else { return ts, fmt.Errorf("Error in OAS source: string enum value is not a string: %v", val) } } if isEnum { ts.Type = "Enum" for _, sym := range values { el := &sadl.EnumElementDef{ Symbol: sym, } ts.Elements = append(ts.Elements, el) } } else { ts.Type = "String" ts.Values = values } } else { ts.Type = "String" } if ts.Type == "String" { if oasSchema.Format == "uuid" { ts.Type = "UUID" } else if oasSchema.Format == "date-time" { ts.Type = "Timestamp" } else { ts.Pattern = oasSchema.Pattern if oasSchema.MinLength > 0 { tmpMin := int64(oasSchema.MinLength) ts.MinSize = &tmpMin } if oasSchema.MaxLength != nil { tmpMax := int64(*oasSchema.MaxLength) ts.MaxSize = &tmpMax } if oasSchema.Format != "" { fmt.Println("NYI: String 'format':", oasSchema.Format) } } } case "array": ts.Type = "Array" if oasSchema.Items != nil { if oasSchema.Items.Ref != "" { ts.Items = oasTypeRef(oasSchema.Items) } else { its, err := convertOasType(name+".Items", oasSchema.Items) if err == nil { ts.Items = its.Type } } } //minsize, maxsize //comment case "number": ts.Type = "Decimal" if oasSchema.Min != nil { ts.Min = sadl.NewDecimal(*oasSchema.Min) } if oasSchema.Max != nil { ts.Max = sadl.NewDecimal(*oasSchema.Max) } case "integer": switch oasSchema.Format { case "int8": ts.Type = "Int8" case "int16": ts.Type = "Int16" case "int32": ts.Type = "Int32" case "int64": ts.Type = "Int64" default: ts.Type = "Int64" } if oasSchema.Min != nil { ts.Min = sadl.NewDecimal(*oasSchema.Min) } if oasSchema.Max != nil { ts.Max = sadl.NewDecimal(*oasSchema.Max) } case "", "object": ts.Type = "Struct" if oasSchema.Properties != nil { req := oasSchema.Required for fname, fschema := range oasSchema.Properties { fd := &sadl.StructFieldDef{ Name: fname, Comment: fschema.Description, } if containsString(req, fname) { fd.Required = true } fd.Type = oasTypeRef(fschema) if fd.Type == "" { fd.TypeSpec, err = convertOasType(name+"."+fname, fschema) } ts.Fields = append(ts.Fields, fd) } } default: fmt.Printf("oas type is %q\n", oasSchema.Type) panic("oas type not handled") } return ts, err } func containsString(lst []string, val string) bool { for _, s := range lst { if s == val { return true } } return false } func capitalize(s string) string { return strings.ToUpper(s[0:1]) + s[1:] } func uncapitalize(s string) string { return strings.ToLower(s[0:1]) + s[1:] } func makeIdentifier(text string) string { reg, _ := regexp.Compile("[^a-zA-Z_][^a-zA-Z_0-9]*") return reg.ReplaceAllString(text, "") } func convertOasPath(path string, op *Operation, method string) (*sadl.HttpDef, error) { hact := &sadl.HttpDef{ Name: op.OperationId, Path: path, Method: method, Comment: op.Summary, } if len(op.Tags) > 0 { hact.Annotations = make(map[string]string, 0) //note: first tag is used as the "resource" name in SADL. tmp := "" rez := "" for _, tag := range op.Tags { if rez == "" { rez = tag } else if tmp == "" { tmp = tag } else { tmp = tmp + "," + tag } } hact.Resource = rez if len(tmp) > 0 { hact.Annotations["x_tags"] = tmp } } var queries []string for _, param := range op.Parameters { name := makeIdentifier(param.Name) spec := &sadl.HttpParamSpec{ StructFieldDef: sadl.StructFieldDef{ Name: name, Comment: param.Description, Required: param.Required, }, } switch param.In { case "query": spec.Query = param.Name queries = append(queries, param.Name+"={"+name+"}") case "path": spec.Path = true if strings.Index(path, "{"+name+"}") < 0 { fmt.Println("WARNING: path param is not in path template:", path, name) panic("here") } case "header": spec.Header = param.Name case "cookie": return nil, fmt.Errorf("Cookie params NYI: %v", sadl.AsString(param)) } spec.Type = oasTypeRef(param.Schema) if spec.Type == "" { if param.Schema != nil { spec.Type = sadlPrimitiveType(param.Schema.Type) } if spec.Type == "Array" { if param.Schema.Items == nil { spec.Items = "Any" } else { schref := param.Schema.Items switch schref.Type { case "string": spec.Items = "String" default: spec.Items = "Any" } } } if spec.Type == "Struct" { panic("Whoops, that can't be right") } if param.Schema != nil && param.Schema.Enum != nil { for _, val := range param.Schema.Enum { if s, ok := val.(string); ok { spec.Values = append(spec.Values, s) } else { return nil, fmt.Errorf("String enum values are not strings: %v", param.Schema.Enum) } } } } else { } hact.Inputs = append(hact.Inputs, spec) } if len(queries) > 0 { hact.Path = hact.Path + "?" + strings.Join(queries, "&") } if hact.Method == "POST" || hact.Method == "PUT" || hact.Method == "PATCH"
{ if op.RequestBody != nil { for contentType, mediadef := range op.RequestBody.Content { if contentType == "application/json" { //hack bodyType := oasTypeRef(mediadef.Schema) if bodyType != "" { spec := &sadl.HttpParamSpec{ StructFieldDef: sadl.StructFieldDef{ TypeSpec: sadl.TypeSpec{ Type: bodyType, }, Comment: op.RequestBody.Description, Name: "body", Required: op.RequestBody.Required, }, } hact.Inputs = append(hact.Inputs, spec) } } }
conditional_block
import_openapi.go
.Name+"={"+name+"}") case "path": spec.Path = true if strings.Index(path, "{"+name+"}") < 0 { fmt.Println("WARNING: path param is not in path template:", path, name) panic("here") } case "header": spec.Header = param.Name case "cookie": return nil, fmt.Errorf("Cookie params NYI: %v", sadl.AsString(param)) } spec.Type = oasTypeRef(param.Schema) if spec.Type == "" { if param.Schema != nil { spec.Type = sadlPrimitiveType(param.Schema.Type) } if spec.Type == "Array" { if param.Schema.Items == nil { spec.Items = "Any" } else { schref := param.Schema.Items switch schref.Type { case "string": spec.Items = "String" default: spec.Items = "Any" } } } if spec.Type == "Struct" { panic("Whoops, that can't be right") } if param.Schema != nil && param.Schema.Enum != nil { for _, val := range param.Schema.Enum { if s, ok := val.(string); ok { spec.Values = append(spec.Values, s) } else { return nil, fmt.Errorf("String enum values are not strings: %v", param.Schema.Enum) } } } } else { } hact.Inputs = append(hact.Inputs, spec) } if len(queries) > 0 { hact.Path = hact.Path + "?" + strings.Join(queries, "&") } if hact.Method == "POST" || hact.Method == "PUT" || hact.Method == "PATCH" { if op.RequestBody != nil { for contentType, mediadef := range op.RequestBody.Content { if contentType == "application/json" { //hack bodyType := oasTypeRef(mediadef.Schema) if bodyType != "" { spec := &sadl.HttpParamSpec{ StructFieldDef: sadl.StructFieldDef{ TypeSpec: sadl.TypeSpec{ Type: bodyType, }, Comment: op.RequestBody.Description, Name: "body", Required: op.RequestBody.Required, }, } hact.Inputs = append(hact.Inputs, spec) } } } } } //expected: if 200 is in the list, use that //else: if 201 is in the list, use that //else: ? find a likely candidate. var expectedStatus string = "default" for status, _ := range op.Responses { if strings.HasPrefix(status, "2") || strings.HasPrefix(status, "3") { expectedStatus = status break } } // if expectedStatus == "default" { // expectedStatus = "200" //? // } if expectedStatus != "" { eparam := op.Responses[expectedStatus] if eparam == nil { return nil, fmt.Errorf("no response entity type provided for operation %q", op.OperationId) } var err error code := 200 if expectedStatus != "default" && strings.Index(expectedStatus, "X") < 0 { code, err = strconv.Atoi(expectedStatus) if err != nil { return nil, err } } ex := &sadl.HttpExpectedSpec{ Status: int32(code), Comment: eparam.Description, } for header, def := range eparam.Headers { param := &sadl.HttpParamSpec{} param.Header = header param.Comment = def.Description s := param.Header //most app-defined headers start with "x-" or "X-". Strip that off for a more reasonable variable name. if strings.HasPrefix(param.Header, "x-") || strings.HasPrefix(param.Header, "X-") { s = s[2:] } param.Name = makeIdentifier(s) schref := def.Schema if schref != nil { if schref.Ref != "" { param.Type = oasTypeRef(schref) } else { param.TypeSpec, err = convertOasType(hact.Name+".Expected."+param.Name, schref) //fix: example } ex.Outputs = append(ex.Outputs, param) } } for contentType, mediadef := range eparam.Content { if contentType == "application/json" { //hack result := &sadl.HttpParamSpec{} result.Name = "body" schref := mediadef.Schema if schref != nil { if schref.Ref != "" { result.Type = oasTypeRef(schref) } else { result.TypeSpec, err = convertOasType(hact.Name+".Expected.payload", schref) //fix: example } ex.Outputs = append(ex.Outputs, result) } else { fmt.Println("HTTP Action has no expected result type:", sadl.Pretty(eparam)) } } } hact.Expected = ex } for status, param := range op.Responses { if status != expectedStatus { //the status can be "default", or "4XX" (where 'X' is a wildcard) or "404". If the latter, it takes precedence. //for SADL, not specifying the response is a bug. So "default" will be turned into "500". The wildcards if status == "default" { status = "0" } else if strings.Index(status, "X") >= 0 { panic("wildcard response codes not supported") } code, err := strconv.Atoi(status) if err != nil { return nil, fmt.Errorf("Invalid status code: %q", status) } ex := &sadl.HttpExceptionSpec{ Status: int32(code), Comment: param.Description, } //FIXME: sadl should allow response headers for exceptions, also. for contentType, mediadef := range param.Content { if contentType == "application/json" { //hack schref := mediadef.Schema if schref != nil { if schref.Ref != "" { ex.Type = oasTypeRef(schref) } else { panic("inline response types not yet supported") } break } } } hact.Exceptions = append(hact.Exceptions, ex) } } //tags: add `x-tags="one,two"` annotation return hact, nil } func getPathOperation(oasPathItem *PathItem, method string) *Operation { switch method { case "GET": return oasPathItem.Get case "PUT": // fmt.Println("xxxxxxxxxxxxxxxx----!!!!", method, oasPathItem.OperationId) // panic("here") return oasPathItem.Put case "POST": return oasPathItem.Post case "DELETE": return oasPathItem.Delete case "HEAD": return oasPathItem.Head /* case "PATCH": return oasPathItem.Patch case "OPTIONS": return oasPathItem.Options case "TRACE": return oasPathItem.Trace case "CONNECT": return oasPathItem.Connect */ } return nil } func guessOperationName(op *Operation, method string) string { defaultStatus := guessDefaultResponseCode(op) switch method { case "GET": resp := op.Responses[defaultStatus] if resp == nil { resp = op.Responses["default"] } for contentType, mediadef := range resp.Content { if contentType == "application/json" { schref := mediadef.Schema if schref != nil { if schref.Ref != "" { entityType := oasTypeRef(schref) return entityType } else { entityType := sadlPrimitiveType(schref.Type) if entityType == "Array" { itemType := schref.Items if itemType.Ref != "" { itemTypeName := oasTypeRef(itemType) entityType = "ArrayOf" + itemTypeName } } return entityType } } else { fmt.Println("HTTP Action has no expected result type:", sadl.Pretty(resp)) } } } } return "" } func sadlPrimitiveType(name string) string { switch name { case "string": return "String" case "number": return "Decimal" case "integer": return "Int32" case "array": return "Array" case "object": return "Struct" case "boolean": return "Bool" default: fmt.Println("sadlPrimitiveType for:", name) panic("what?") } } func findTypeDef(schema *sadl.Schema, name string) *sadl.TypeDef { for _, td := range schema.Types { if td.Name == name { return td
}
random_line_split
import_openapi.go
(paths []string, conf *sadl.Data) (*sadl.Model, error) { if len(paths) != 1 { return nil, fmt.Errorf("Cannot merge multiple OpenAPI files") } path := paths[0] name := path n := strings.LastIndex(name, "/") // format := "" if n >= 0 { name = name[n+1:] } n = strings.LastIndex(name, ".") if n >= 0 { // format = name[n+1:] name = name[:n] name = strings.Replace(name, ".", "_", -1) } oas3, err := Load(path) if err != nil { return nil, err } model, err := oas3.ToSadl(name) if err != nil { return nil, fmt.Errorf("Cannot convert to SADL: %v\n", err) } //err = model.ConvertInlineEnums() return model, err } /* func DetermineVersion(data []byte, format string) (string, error) { var raw map[string]interface{} var err error switch format { case "json": err = json.Unmarshal(data, &raw) case "yaml": err = yaml.Unmarshal(data, &raw) default: err = fmt.Errorf("Unsupported file format: %q. Only \"json\" and \"yaml\" are supported.", format) } if err != nil { return "", err } if v, ok := raw["openapi"]; ok { if s, ok := v.(string); ok { return s, nil } } if v, ok := raw["swagger"]; ok { if s, ok := v.(string); ok { return s, nil } } return "", fmt.Errorf("Cannot find an 'openapi' in the specified %s file to determine the version", format) } */ /* func xParse(data []byte, format string) (*Model, error) { version, err := DetermineVersion(data, format) if err != nil { return nil, err } oas := &Oas{ source: version, } if strings.HasPrefix(version, "3.") { oas.V3, err = oas3.Parse(data, format) return oas, err } else if strings.HasPrefix(version, "2.") { v2, err := oas2.Parse(data, format) if err == nil { oas.V3, err = oas2.ConvertToV3(v2) } return oas, err } return nil, fmt.Errorf("Unsupported version of OpenAPI Spec: %s", version) } */ var examples []*sadl.ExampleDef var methods = []string{"GET", "PUT", "POST", "DELETE", "HEAD"} //to do: "PATCH", "OPTIONS", "TRACE" func (model *Model) ToSadl(name string) (*sadl.Model, error) { annotations := make(map[string]string, 0) examples = nil annotations["x_openapi_version"] = model.OpenAPI comment := model.Info.Description if model.Info.Title != "" { if sadl.IsSymbol(model.Info.Title) { name = model.Info.Title } else { comment = model.Info.Title + " - " + comment } } schema := &sadl.Schema{ Name: name, Comment: comment, Version: model.Info.Version, } for name, oasSchema := range model.Components.Schemas { name = validSadlName(name, oasSchema) if name == "" { continue } var ts sadl.TypeSpec var err error comment := "" tname := oasTypeRef(oasSchema) if tname != "" { if oasDef, ok := model.Components.Schemas[tname]; ok { ts, err = convertOasType(tname, oasDef) //doesn't handle N levels } else { panic("hmm") } } else { ts, err = convertOasType(name, oasSchema) comment = oasSchema.Description } if err != nil { return nil, err } td := &sadl.TypeDef{ TypeSpec: ts, Name: name, Comment: comment, //annotations } schema.Types = append(schema.Types, td) } httpBindings := true for tmpl, path := range model.Paths { path2 := *path for _, method := range methods { op := getPathOperation(&path2, method) if op != nil { if strings.HasPrefix(tmpl, "x-") { continue } if httpBindings { hact, err := convertOasPath(tmpl, op, method) if err != nil { return nil, err } schema.Http = append(schema.Http, hact) } } } } for _, server := range model.Servers { annotations["x_server"] = server.URL } if model.Info.License != nil { if model.Info.License.Name != "" { annotations["x_license_name"] = model.Info.License.Name } if model.Info.License.URL != "" { // schema.Annotations["x_license_url"] = oas.V3.Info.License.URL annotations["x_license_url"] = model.Info.License.URL } } if len(annotations) > 0 { schema.Annotations = annotations } schema.Examples = examples return sadl.NewModel(schema) } func validSadlName(name string, oasSchema *Schema) string { if name == "Timestamp" { if oasSchema.Type == "string" { return "" } } else if name == "Decimal" { return "" } return name } func oasTypeRef(oasSchema *Schema) string { if oasSchema != nil && oasSchema.Ref != "" { if strings.HasPrefix(oasSchema.Ref, "#/components/schemas/") { return oasSchema.Ref[len("#/components/schemas/"):] } return oasSchema.Ref //? } return "" } func convertOasType(name string, oasSchema *Schema) (sadl.TypeSpec, error) { var err error var ts sadl.TypeSpec if oasSchema.Example != nil { ex := &sadl.ExampleDef{ Target: name, Example: oasSchema.Example, } examples = append(examples, ex) } switch oasSchema.Type { case "boolean": ts.Type = "Bool" case "string": if oasSchema.Enum != nil { //OAS defines element *descriptions* as the values, not symbolic identifiers. //so we look for the case where all values look like identifiers, and call that an enum. Else a strings with accepted "values" //perhaps the spirit of JSON Schema enums are just values, not what I think of as "enums", i.e. "a set of named values", per wikipedia. //still, with symbolic values, perhaps the intent is to use proper enums, if only JSON Schema had them. isEnum := EnumTypes var values []string for _, val := range oasSchema.Enum { if s, ok := val.(string); ok { values = append(values, s) if !sadl.IsSymbol(s) { isEnum = false } } else { return ts, fmt.Errorf("Error in OAS source: string enum value is not a string: %v", val) } } if isEnum { ts.Type = "Enum" for _, sym := range values { el := &sadl.EnumElementDef{ Symbol: sym, } ts.Elements = append(ts.Elements, el) } } else { ts.Type = "String" ts.Values = values } } else { ts.Type = "String" } if ts.Type == "String" { if oasSchema.Format == "uuid" { ts.Type = "UUID" } else if oasSchema.Format == "date-time" { ts.Type = "Timestamp" } else { ts.Pattern = oasSchema.Pattern if oasSchema.MinLength > 0 { tmpMin := int64(oasSchema.MinLength) ts.MinSize = &tmpMin } if oasSchema.MaxLength != nil { tmpMax := int64(*oasSchema.MaxLength) ts.MaxSize = &tmpMax } if oasSchema.Format != "" { fmt.Println("NYI: String 'format':", oasSchema.Format) } } } case "array": ts.Type = "Array" if oasSchema.Items != nil { if oasSchema.Items.Ref != "" { ts.Items = oasTypeRef(oasSchema.Items) } else { its, err := convertOasType(name+".Items", oasSchema.Items) if err == nil { ts.Items = its.Type } } } //minsize, maxsize
Import
identifier_name
import_openapi.go
} model, err := oas3.ToSadl(name) if err != nil { return nil, fmt.Errorf("Cannot convert to SADL: %v\n", err) } //err = model.ConvertInlineEnums() return model, err } /* func DetermineVersion(data []byte, format string) (string, error) { var raw map[string]interface{} var err error switch format { case "json": err = json.Unmarshal(data, &raw) case "yaml": err = yaml.Unmarshal(data, &raw) default: err = fmt.Errorf("Unsupported file format: %q. Only \"json\" and \"yaml\" are supported.", format) } if err != nil { return "", err } if v, ok := raw["openapi"]; ok { if s, ok := v.(string); ok { return s, nil } } if v, ok := raw["swagger"]; ok { if s, ok := v.(string); ok { return s, nil } } return "", fmt.Errorf("Cannot find an 'openapi' in the specified %s file to determine the version", format) } */ /* func xParse(data []byte, format string) (*Model, error) { version, err := DetermineVersion(data, format) if err != nil { return nil, err } oas := &Oas{ source: version, } if strings.HasPrefix(version, "3.") { oas.V3, err = oas3.Parse(data, format) return oas, err } else if strings.HasPrefix(version, "2.") { v2, err := oas2.Parse(data, format) if err == nil { oas.V3, err = oas2.ConvertToV3(v2) } return oas, err } return nil, fmt.Errorf("Unsupported version of OpenAPI Spec: %s", version) } */ var examples []*sadl.ExampleDef var methods = []string{"GET", "PUT", "POST", "DELETE", "HEAD"} //to do: "PATCH", "OPTIONS", "TRACE" func (model *Model) ToSadl(name string) (*sadl.Model, error) { annotations := make(map[string]string, 0) examples = nil annotations["x_openapi_version"] = model.OpenAPI comment := model.Info.Description if model.Info.Title != "" { if sadl.IsSymbol(model.Info.Title) { name = model.Info.Title } else { comment = model.Info.Title + " - " + comment } } schema := &sadl.Schema{ Name: name, Comment: comment, Version: model.Info.Version, } for name, oasSchema := range model.Components.Schemas { name = validSadlName(name, oasSchema) if name == "" { continue } var ts sadl.TypeSpec var err error comment := "" tname := oasTypeRef(oasSchema) if tname != "" { if oasDef, ok := model.Components.Schemas[tname]; ok { ts, err = convertOasType(tname, oasDef) //doesn't handle N levels } else { panic("hmm") } } else { ts, err = convertOasType(name, oasSchema) comment = oasSchema.Description } if err != nil { return nil, err } td := &sadl.TypeDef{ TypeSpec: ts, Name: name, Comment: comment, //annotations } schema.Types = append(schema.Types, td) } httpBindings := true for tmpl, path := range model.Paths { path2 := *path for _, method := range methods { op := getPathOperation(&path2, method) if op != nil { if strings.HasPrefix(tmpl, "x-") { continue } if httpBindings { hact, err := convertOasPath(tmpl, op, method) if err != nil { return nil, err } schema.Http = append(schema.Http, hact) } } } } for _, server := range model.Servers { annotations["x_server"] = server.URL } if model.Info.License != nil { if model.Info.License.Name != "" { annotations["x_license_name"] = model.Info.License.Name } if model.Info.License.URL != "" { // schema.Annotations["x_license_url"] = oas.V3.Info.License.URL annotations["x_license_url"] = model.Info.License.URL } } if len(annotations) > 0 { schema.Annotations = annotations } schema.Examples = examples return sadl.NewModel(schema) } func validSadlName(name string, oasSchema *Schema) string { if name == "Timestamp" { if oasSchema.Type == "string" { return "" } } else if name == "Decimal" { return "" } return name } func oasTypeRef(oasSchema *Schema) string { if oasSchema != nil && oasSchema.Ref != "" { if strings.HasPrefix(oasSchema.Ref, "#/components/schemas/") { return oasSchema.Ref[len("#/components/schemas/"):] } return oasSchema.Ref //? } return "" } func convertOasType(name string, oasSchema *Schema) (sadl.TypeSpec, error) { var err error var ts sadl.TypeSpec if oasSchema.Example != nil { ex := &sadl.ExampleDef{ Target: name, Example: oasSchema.Example, } examples = append(examples, ex) } switch oasSchema.Type { case "boolean": ts.Type = "Bool" case "string": if oasSchema.Enum != nil { //OAS defines element *descriptions* as the values, not symbolic identifiers. //so we look for the case where all values look like identifiers, and call that an enum. Else a strings with accepted "values" //perhaps the spirit of JSON Schema enums are just values, not what I think of as "enums", i.e. "a set of named values", per wikipedia. //still, with symbolic values, perhaps the intent is to use proper enums, if only JSON Schema had them. isEnum := EnumTypes var values []string for _, val := range oasSchema.Enum { if s, ok := val.(string); ok { values = append(values, s) if !sadl.IsSymbol(s) { isEnum = false } } else { return ts, fmt.Errorf("Error in OAS source: string enum value is not a string: %v", val) } } if isEnum { ts.Type = "Enum" for _, sym := range values { el := &sadl.EnumElementDef{ Symbol: sym, } ts.Elements = append(ts.Elements, el) } } else { ts.Type = "String" ts.Values = values } } else { ts.Type = "String" } if ts.Type == "String" { if oasSchema.Format == "uuid" { ts.Type = "UUID" } else if oasSchema.Format == "date-time" { ts.Type = "Timestamp" } else { ts.Pattern = oasSchema.Pattern if oasSchema.MinLength > 0 { tmpMin := int64(oasSchema.MinLength) ts.MinSize = &tmpMin } if oasSchema.MaxLength != nil { tmpMax := int64(*oasSchema.MaxLength) ts.MaxSize = &tmpMax } if oasSchema.Format != "" { fmt.Println("NYI: String 'format':", oasSchema.Format) } } } case "array": ts.Type = "Array" if oasSchema.Items != nil { if oasSchema.Items.Ref != "" { ts.Items = oasTypeRef(oasSchema.Items) } else { its, err := convertOasType(name+".Items", oasSchema.Items) if err == nil { ts.Items = its.Type } } } //minsize, maxsize //comment case "number": ts.Type = "Decimal" if oasSchema
{ if len(paths) != 1 { return nil, fmt.Errorf("Cannot merge multiple OpenAPI files") } path := paths[0] name := path n := strings.LastIndex(name, "/") // format := "" if n >= 0 { name = name[n+1:] } n = strings.LastIndex(name, ".") if n >= 0 { // format = name[n+1:] name = name[:n] name = strings.Replace(name, ".", "_", -1) } oas3, err := Load(path) if err != nil { return nil, err
identifier_body
ser.rs
} impl<'a> Drop for Serializer<'a> { fn drop(&mut self) { // Drop layers in reverse order. while !self.stack.is_empty() { self.stack.pop(); } } } #[allow(nonstandard_style)] struct write_u64 { major: u8, v: u64, } impl write_u64 { fn into(self, out: &'_ mut (dyn io::Write)) -> io::Result<()> { let Self { major, v: value } = self; let mask = major << 5; macro_rules! with_uNs {( $($uN:ident)<* ) => ({ mod c { $( pub mod $uN { pub const MAX: u64 = ::core::$uN::MAX as _; } )* pub mod u8 { pub const MAX: u64 = ::core::u8::MAX as _; } } const SMALL_U8_MAX: u64 = 0x17; #[allow(nonstandard_style)] enum MaskFor { u8 = (SMALL_U8_MAX + 1) as _, $($uN),* } match value { 0 ..= SMALL_U8_MAX => out.write_all(&[mask | (value as u8)]), 0 ..= c::u8::MAX => out.write_all(&[ mask | (MaskFor::u8 as u8), value as u8, ]), $( 0 ..= c::$uN::MAX => { let value = value as $uN; let ref mut buf = [0; 1 + ::core::mem::size_of::<$uN>()]; buf[0] = mask | (MaskFor::$uN as u8); buf[1 ..].copy_from_slice(&value.to_be_bytes()); out.write_all(buf) }, )* _ => unreachable!(), } })} with_uNs!(u16 < u32 < u64) } } /// Serialize any serializable type as a CBOR byte sequence into a /// [`Write`][io::Write]able sink. /// /// Returns: /// - `Ok(())` on success. /// - `Err(Some(io_error))` on I/O failure. /// - `Err(None)` on serialization error (unrepresentable integer). pub fn to_writer<'value>( out: &'_ mut dyn io::Write, value: &'value dyn Serialize, ) -> Result<(), Option<io::Error>> { // Borrow-checker-friendly "closure" #[cfg_attr(rustfmt, rustfmt::skip)] macro_rules! write { ($bytes:expr) => ({ out.write_all($bytes).map_err(Some) })} // Use a manual stack to avoid (stack-allocated) recursion. let mut stack: Vec<Layer<'value>> = vec![Layer::Single(value)]; // where: enum Layer<'value> { Seq(Box<dyn Seq<'value> + 'value>), Map(Box<dyn Map<'value> + 'value>), Single(&'value dyn Serialize), } while let Some(last) = stack.last_mut() { let view: ValueView<'value> = match last { &mut Layer::Single(value) => { let view = value.view(); drop(stack.pop()); view } Layer::Seq(seq) => { match seq.next() { Some(value) => stack.push(Layer::Single(value)), None => drop(stack.pop()), } continue; } Layer::Map(map) => { match map.next() { Some((key, value)) => { stack.push(Layer::Single(value)); stack.push(Layer::Single(key)); } None => drop(stack.pop()), } continue; } }; match view { ValueView::Null => write!(&[0xf6])?, ValueView::Bool(b) => write!(&[0xf4 | (b as u8)])?, ValueView::Str(s) => { write_u64 { major: 3, v: s.len() as u64, } .into(out)?; write!(s.as_bytes())?; } ValueView::Bytes(bs) => { write_u64 { major: 2, v: bs.len() as u64, } .into(out)?; write!(&*bs)?; } ValueView::Int(i) => { const MIN: i128 = -(1_i128 << 64); const MAX: i128 = ::core::u64::MAX as _; match i { MIN..=-1 => write_u64 { major: 1, v: (-(i + 1)) as u64, } .into(out)?, 0..=MAX => write_u64 { major: 0, v: i as u64, } .into(out)?, _ => err!("Cannot serialize integer {:?} as CBOR: out of range", i), } } ValueView::F64(f) if f.is_infinite() => write!(if f.is_sign_positive() { &[0xf9, 0x7c, 0x00] } else { &[0xf9, 0xfc, 0x00] })?, ValueView::F64(f) if f.is_nan() => { write!(&[0xf9, 0x7e, 0x00])?; } ValueView::F64(f) => { // Finite float. let f_16; let f_32; match () { _case if { f_16 = ::half::f16::from_f64(f); f64::from(f_16) == f } => { let ref mut buf = [0xf9, 0, 0]; buf[1..].copy_from_slice(&f_16.to_bits().to_be_bytes()); write!(buf)?; } _case if { f_32 = f as f32; f64::from(f_32) == f } => { let ref mut buf = [0xfa, 0, 0, 0, 0]; buf[1..].copy_from_slice(&f_32.to_bits().to_be_bytes()); write!(buf)?; } _default => { let ref mut buf = [0xfb, 0, 0, 0, 0, 0, 0, 0, 0]; buf[1..].copy_from_slice(&f.to_bits().to_be_bytes()); write!(buf)?; } } } ValueView::Seq(mut seq) => { let count = seq.remaining(); write_u64 { major: 4, v: count as _, } .into(out)?; stack.push(Layer::Seq(seq)); } ValueView::Map(mut map) => { let count = map.remaining(); write_u64 { major: 5, v: count as _, } .into(out)?; stack.push(Layer::Map(map)); } } } Ok(()) } #[cfg(test)] mod tests { //! Most of these tests have been taken from //! https://github.com/pyfisch/cbor/blob/a218403a52e60c991313f429e4acc05cce81ce25/tests/ser.rs use super::*; use crate::{ cbor::{value::*, *}, Serialize, }; #[cfg_attr(rustfmt, rustfmt::skip)] macro_rules! assert_eq_hex {( $left:expr, $right:expr $(,)? ) => ( match (&$left[..], &$right[..]) { (ref left, ref right) => { if <[u8] as ::core::cmp::PartialEq>::ne(left, right) { panic!( "assertion failed: (`{}` == `{}`)\n{}]", stringify!($left), stringify!($right), (0..left.len().max(right.len())) .map(|i| match (left.get(i), right.get(i)) { (Some(l), Some(r)) => format!( " {:01}|{:02x} – {:01}|{:02x},\n", l >> 5, l & 0x1f, r >> 5, r & 0x1f ), (Some(l), _) => format!(" {:01}|{:02x} - ____,\n", l >> 5, l & 0x1f), (_, Some(r)) => format!("____ - {:01}|{:02x},\n", r >> 5, r & 0x1f), _ => unreachable!(), })
enum Layer<'a> { Seq(Box<dyn Seq<'a> + 'a>), Map(Box<dyn Map<'a> + 'a>),
random_line_split
ser.rs
(self, out: &'_ mut (dyn io::Write)) -> io::Result<()> { let Self { major, v: value } = self; let mask = major << 5; macro_rules! with_uNs {( $($uN:ident)<* ) => ({ mod c { $( pub mod $uN { pub const MAX: u64 = ::core::$uN::MAX as _; } )* pub mod u8 { pub const MAX: u64 = ::core::u8::MAX as _; } } const SMALL_U8_MAX: u64 = 0x17; #[allow(nonstandard_style)] enum MaskFor { u8 = (SMALL_U8_MAX + 1) as _, $($uN),* } match value { 0 ..= SMALL_U8_MAX => out.write_all(&[mask | (value as u8)]), 0 ..= c::u8::MAX => out.write_all(&[ mask | (MaskFor::u8 as u8), value as u8, ]), $( 0 ..= c::$uN::MAX => { let value = value as $uN; let ref mut buf = [0; 1 + ::core::mem::size_of::<$uN>()]; buf[0] = mask | (MaskFor::$uN as u8); buf[1 ..].copy_from_slice(&value.to_be_bytes()); out.write_all(buf) }, )* _ => unreachable!(), } })} with_uNs!(u16 < u32 < u64) } } /// Serialize any serializable type as a CBOR byte sequence into a /// [`Write`][io::Write]able sink. /// /// Returns: /// - `Ok(())` on success. /// - `Err(Some(io_error))` on I/O failure. /// - `Err(None)` on serialization error (unrepresentable integer). pub fn to_writer<'value>( out: &'_ mut dyn io::Write, value: &'value dyn Serialize, ) -> Result<(), Option<io::Error>> { // Borrow-checker-friendly "closure" #[cfg_attr(rustfmt, rustfmt::skip)] macro_rules! write { ($bytes:expr) => ({ out.write_all($bytes).map_err(Some) })} // Use a manual stack to avoid (stack-allocated) recursion. let mut stack: Vec<Layer<'value>> = vec![Layer::Single(value)]; // where: enum Layer<'value> { Seq(Box<dyn Seq<'value> + 'value>), Map(Box<dyn Map<'value> + 'value>), Single(&'value dyn Serialize), } while let Some(last) = stack.last_mut() { let view: ValueView<'value> = match last { &mut Layer::Single(value) => { let view = value.view(); drop(stack.pop()); view } Layer::Seq(seq) => { match seq.next() { Some(value) => stack.push(Layer::Single(value)), None => drop(stack.pop()), } continue; } Layer::Map(map) => { match map.next() { Some((key, value)) => { stack.push(Layer::Single(value)); stack.push(Layer::Single(key)); } None => drop(stack.pop()), } continue; } }; match view { ValueView::Null => write!(&[0xf6])?, ValueView::Bool(b) => write!(&[0xf4 | (b as u8)])?, ValueView::Str(s) => { write_u64 { major: 3, v: s.len() as u64, } .into(out)?; write!(s.as_bytes())?; } ValueView::Bytes(bs) => { write_u64 { major: 2, v: bs.len() as u64, } .into(out)?; write!(&*bs)?; } ValueView::Int(i) => { const MIN: i128 = -(1_i128 << 64); const MAX: i128 = ::core::u64::MAX as _; match i { MIN..=-1 => write_u64 { major: 1, v: (-(i + 1)) as u64, } .into(out)?, 0..=MAX => write_u64 { major: 0, v: i as u64, } .into(out)?, _ => err!("Cannot serialize integer {:?} as CBOR: out of range", i), } } ValueView::F64(f) if f.is_infinite() => write!(if f.is_sign_positive() { &[0xf9, 0x7c, 0x00] } else { &[0xf9, 0xfc, 0x00] })?, ValueView::F64(f) if f.is_nan() => { write!(&[0xf9, 0x7e, 0x00])?; } ValueView::F64(f) => { // Finite float. let f_16; let f_32; match () { _case if { f_16 = ::half::f16::from_f64(f); f64::from(f_16) == f } => { let ref mut buf = [0xf9, 0, 0]; buf[1..].copy_from_slice(&f_16.to_bits().to_be_bytes()); write!(buf)?; } _case if { f_32 = f as f32; f64::from(f_32) == f } => { let ref mut buf = [0xfa, 0, 0, 0, 0]; buf[1..].copy_from_slice(&f_32.to_bits().to_be_bytes()); write!(buf)?; } _default => { let ref mut buf = [0xfb, 0, 0, 0, 0, 0, 0, 0, 0]; buf[1..].copy_from_slice(&f.to_bits().to_be_bytes()); write!(buf)?; } } } ValueView::Seq(mut seq) => { let count = seq.remaining(); write_u64 { major: 4, v: count as _, } .into(out)?; stack.push(Layer::Seq(seq)); } ValueView::Map(mut map) => { let count = map.remaining(); write_u64 { major: 5, v: count as _, } .into(out)?; stack.push(Layer::Map(map)); } } } Ok(()) } #[cfg(test)] mod tests { //! Most of these tests have been taken from //! https://github.com/pyfisch/cbor/blob/a218403a52e60c991313f429e4acc05cce81ce25/tests/ser.rs use super::*; use crate::{ cbor::{value::*, *}, Serialize, }; #[cfg_attr(rustfmt, rustfmt::skip)] macro_rules! assert_eq_hex {( $left:expr, $right:expr $(,)? ) => ( match (&$left[..], &$right[..]) { (ref left, ref right) => { if <[u8] as ::core::cmp::PartialEq>::ne(left, right) { panic!( "assertion failed: (`{}` == `{}`)\n{}]", stringify!($left), stringify!($right), (0..left.len().max(right.len())) .map(|i| match (left.get(i), right.get(i)) { (Some(l), Some(r)) => format!( " {:01}|{:02x} – {:01}|{:02x},\n", l >> 5, l & 0x1f, r >> 5, r & 0x1f ), (Some(l), _) => format!(" {:01}|{:02x} - ____,\n", l >> 5, l & 0x1f), (_, Some(r)) => format!("____ - {:01}|{:02x},\n", r >> 5, r & 0x1f), _ => unreachable!(), }) .collect::<String>(), ); } } } )} #[test] fn test_str() { serialize_and_compare("foobar", b"ffoobar"); } #[test] fn test_list() { serialize_and_compare(&[1, 2, 3][..], b"\x83\x01\x02\x03"); } #[test] fn test_float() { serialize_and_compare(12.3f64, b"\xfb@(\x
into
identifier_name
ser.rs
, 32, 246, 129, 33, 246, 130, 0, 0, 246, 130, 0, 32, 246 ], vec ); let test_object = from_slice(&vec[..]).unwrap(); assert_eq!(object, test_object); } #[test] fn test_object_object_keys() { use ::std::iter::FromIterator; let mut object = BTreeMap::new(); let keys = vec![ vec!["a"], vec!["b"], vec!["c"], vec!["d"], vec!["aa"], vec!["a", "aa"], ] .into_iter() .map(|v| BTreeMap::from_iter(v.into_iter().map(|s| (s.to_owned(), ())))); for key in keys { object.insert(key, ()); } let vec = to_vec(&to_value(&object).unwrap()).unwrap(); assert_eq_hex!( vec![ 166, 161, 97, 97, 246, 246, 161, 97, 98, 246, 246, 161, 97, 99, 246, 246, 161, 97, 100, 246, 246, 161, 98, 97, 97, 246, 246, 162, 97, 97, 246, 98, 97, 97, 246, 246 ], vec ); let test_object = from_slice(&vec[..]).unwrap(); assert_eq!(object, test_object); } #[test] fn test_float() { let vec = to_vec(&12.3f64).unwrap(); assert_eq_hex!(vec, b"\xfb@(\x99\x99\x99\x99\x99\x9a"); } #[test] fn test_f32() { let vec = to_vec(&4000.5f32).unwrap(); assert_eq_hex!(vec, b"\xfa\x45\x7a\x08\x00"); } #[test] fn test_infinity() { let vec = to_vec(&::std::f64::INFINITY).unwrap(); assert_eq_hex!(vec, b"\xf9|\x00"); } #[test] fn test_neg_infinity() { let vec = to_vec(&::std::f64::NEG_INFINITY).unwrap(); assert_eq_hex!(vec, b"\xf9\xfc\x00"); } #[test] fn test_nan() { let vec = to_vec(&::std::f32::NAN).unwrap(); assert_eq_hex!(vec, b"\xf9\x7e\x00"); } #[test] fn test_integer() { // u8 let vec = to_vec(&24).unwrap(); assert_eq_hex!(vec, b"\x18\x18"); // i8 let vec = to_vec(&-5).unwrap(); assert_eq_hex!(vec, b"\x24"); // i16 let vec = to_vec(&-300).unwrap(); assert_eq_hex!(vec, b"\x39\x01\x2b"); // i32 let vec = to_vec(&-23567997).unwrap(); assert_eq_hex!(vec, b"\x3a\x01\x67\x9e\x7c"); // u64 let vec = to_vec(&::std::u64::MAX).unwrap(); assert_eq_hex!(vec, b"\x1b\xff\xff\xff\xff\xff\xff\xff\xff"); } // #[test] // fn test_self_describing() { // let mut vec = Vec::new(); // { // let mut serializer = ser::Serializer::new(&mut vec); // serializer.self_describe().unwrap(); // serializer.serialize_u64(9).unwrap(); // } // assert_eq_hex!(vec, b"\xd9\xd9\xf7\x09"); // } // #[test] // fn test_ip_addr() { // use ::std::net::Ipv4Addr; // let addr = Ipv4Addr::new(8, 8, 8, 8); // let vec = to_vec(&addr).unwrap(); // println!("{:?}", vec); // assert_eq_hex!(vec.len(), 5); // let test_addr: Ipv4Addr = from_slice(&vec).unwrap(); // assert_eq_hex!(addr, test_addr); // } /// Test all of CBOR's fixed-length byte string types #[test] fn test_byte_string() { // Very short byte strings have 1-byte headers let short = vec![0_u8, 1, 2, 255]; let short_s = to_vec(&short).unwrap(); assert_eq_hex!(&short_s[..], [0x44, 0, 1, 2, 255]); // byte strings > 23 bytes have 2-byte headers let medium = vec![ 0u8, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 255, ]; let medium_s = to_vec(&medium).unwrap(); assert_eq_hex!( &medium_s[..], [ 0x58, 24, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 255 ] ); // byte strings ≥ 256 bytes have 3-byte headers let long_vec = (0..256).map(|i| (i & 0xFF) as u8).collect::<Vec<_>>(); let long_s = to_vec(&long_vec).unwrap(); assert_eq_hex!(&long_s[0..3], [0x59, 1, 0]); assert_eq_hex!(&long_s[3..], &long_vec[..]); // byte strings ≥ 2^16 bytes have 5-byte headers let very_long_vec = (0..65536).map(|i| (i & 0xFF) as u8).collect::<Vec<_>>(); let very_long_s = to_vec(&very_long_vec).unwrap(); assert_eq_hex!(&very_long_s[0..5], [0x5a, 0, 1, 0, 0]); assert_eq_hex!(&very_long_s[5..], &very_long_vec[..]); // byte strings ≥ 2^32 bytes have 9-byte headers, // but they take too much RAM to test in most CI setups, such as Travis. // Confident on our implementation of the serialization code (which // only copies the byte slice contents provided the writer allows it), // we `unsafe`-ly fake a gigantic slice by using a writer // that will saturate right after the header has been written. #[cfg(all(not(miri), target_pointer_width = "64"))] #[cfg_attr(rustfmt, rustfmt::skip)] unsafe { let fake_huge_byte_seq: &'_ [u8] = ::core::slice::from_raw_parts( 0x1 as _, 0x00_00_00_01_de_ad_be_ef, ); let mut _9 = [0_u8; 9]; let _ = to_writer(&mut &mut _9[..], &fake_huge_byte_seq); assert_eq_hex!( &_9[..], [ 0x5b, 0x00, 0x00, 0x00, 0x01, 0xde, 0xad, 0xbe, 0xef, ], ); } } #[test] fn test_half() {
let vec = to_vec(&42.5f32).unwrap(); assert_eq_hex!(vec, b"\xF9\x51\x50"); assert_eq!(from_slice::<f32>(&vec[..]).unwrap(), 42.5f32); } } }
identifier_body
result-info-plot-scatter-exec-time.py
default=" execution time (s)", dest="axis_label_suffix", ) parser.add_argument("--axis-label-colour", type=str, default="black", dest="axis_label_colour", ) parser.add_argument("--annotate", default=False, action='store_true', ) parser.add_argument("--annotate-use-legacy-values", default=False, action='store_true', ) parser.add_argument("--output", default=None, type=argparse.FileType('wb'), ) parser.add_argument("--error-bars", default=False, action='store_true', ) parser.add_argument("--annotate-timeout-point", dest='annotate_timeout_point', default=False, action='store_true', ) parser.add_argument("--require-time-abs-diff", dest="require_time_abs_diff", default=0.0, type=float ) parser.add_argument('--true-type-fonts', default=False, action='store_true' ) pargs = parser.parse_args(args) DriverUtil.handleLoggerArgs(pargs, parser) _logger = logging.getLogger(__name__) if pargs.max_exec_time is None: _logger.error('--max-exec-time must be specified') return 1 if pargs.true_type_fonts: smtrunner.util.set_true_type_font() index_to_raw_result_infos = [] index_to_file_name = [] for index, result_infos_file in enumerate([pargs.first_result_info, pargs.second_result_info]): try: _logger.info('Loading "{}"'.format(result_infos_file.name)) result_infos = ResultInfo.loadRawResultInfos(result_infos_file) index_to_raw_result_infos.append(result_infos) index_to_file_name.append(result_infos_file.name) except ResultInfo.ResultInfoValidationError as e: _logger.error('Validation error:\n{}'.format(e)) return 1 _logger.info('Loading done') result_infos = None # Perform grouping by benchmark name key_to_results_infos, rejected_result_infos = ResultInfoUtil.group_result_infos_by( index_to_raw_result_infos) if len(rejected_result_infos) > 0: _logger.warning('There were rejected result infos') num_merge_failures = 0 for index, l in enumerate(rejected_result_infos): _logger.warning('Index {} had {} rejections'.format(index, len(l))) num_merge_failures += len(l) if num_merge_failures > 0: if pargs.allow_merge_failures: _logger.warning('Merge failures being allowed') else: _logger.error('Merge failures are not allowed') return 1 # Generate scatter points x_scatter_points = [] x_scatter_errors = [[], [] ] y_scatter_points = [] y_scatter_errors = [[], []] count_dual_timeout = 0 count_x_lt_y_not_dt = 0 count_x_gt_y_not_dt = 0 count_x_eq_y_not_dt = 0 # New counting vars bounds_incomparable_keys = set() x_gt_y_keys = set() x_lt_y_keys = set() x_eq_y_keys = set() x_eq_y_and_is_timeout_keys = set() for key, raw_result_info_list in sorted(key_to_results_infos.items(), key=lambda kv:kv[0]): _logger.info('Ranking on "{}" : '.format(key)) indices_to_use = [] # Compute indices to use modified_raw_result_info_list = [ ] # Handle "unknown" # Only compare results that gave sat/unsat for index, ri in enumerate(raw_result_info_list): if isinstance(ri['event_tag'], str): # single result event_tag = ri['event_tag'] else:
# Event must be sat or timeout _logger.info('index {} is {}'.format(index, event_tag)) if event_tag not in { 'sat', 'timeout', 'soft_timeout'}: # Skip this. We can't do a meaningful comparison here continue indices_to_use.append(index) # Normalise timeouts to have fixed values for the time. if event_tag in {'timeout', 'soft_timeout'}: modified_ri = analysis.get_result_with_modified_time( ri, pargs.max_exec_time) _logger.debug('modified_ri: {}'.format( pprint.pformat(modified_ri))) _logger.debug( 'Treating index {} for {} due to unknown as having max-time'.format( index, key)) modified_raw_result_info_list.append(modified_ri) else: modified_raw_result_info_list.append(ri) _logger.debug('used indices_to_use: {}'.format(indices_to_use)) if len(indices_to_use) != 2: # Skip this one. One of the result infos can't be compared # against. continue assert len(indices_to_use) == 2 # Get execution times index_to_execution_time_bounds = analysis.get_index_to_execution_time_bounds( modified_raw_result_info_list, indices_to_use, pargs.max_exec_time, analysis.get_arithmetic_mean_and_99_confidence_intervals, ['dsoes_wallclock', 'wallclock']) assert isinstance(index_to_execution_time_bounds, list) x_scatter_point_bounds = index_to_execution_time_bounds[0] y_scatter_point_bounds = index_to_execution_time_bounds[1] x_scatter_point = x_scatter_point_bounds[1] # mean y_scatter_point = y_scatter_point_bounds[1] # mean x_scatter_lower_error = x_scatter_point_bounds[1] - x_scatter_point_bounds[0] assert x_scatter_lower_error >= 0 x_scatter_higher_error = x_scatter_point_bounds[2] - x_scatter_point_bounds[1] assert x_scatter_higher_error >= 0 y_scatter_lower_error = y_scatter_point_bounds[1] - y_scatter_point_bounds[0] assert y_scatter_lower_error >= 0 y_scatter_higher_error = y_scatter_point_bounds[2] - y_scatter_point_bounds[1] assert y_scatter_higher_error >= 0 x_scatter_points.append(x_scatter_point) y_scatter_points.append(y_scatter_point) # Error bar points #x_scatter_errors.append((x_scatter_lower_error, x_scatter_higher_error)) x_scatter_errors[0].append(x_scatter_lower_error) x_scatter_errors[1].append(x_scatter_higher_error) #y_scatter_errors.append((y_scatter_lower_error, y_scatter_higher_error)) y_scatter_errors[0].append(y_scatter_lower_error) y_scatter_errors[1].append(y_scatter_higher_error) # LEGACY: Now do some counting if x_scatter_point == y_scatter_point: if x_scatter_point == pargs.max_exec_time: assert x_scatter_lower_error == 0 assert x_scatter_higher_error == 0 assert y_scatter_lower_error == 0 assert y_scatter_higher_error == 0 count_dual_timeout += 1 else: _logger.info('Found count_x_eq_y_not_dt: x: {}, key: {}'.format( x_scatter_point, key)) count_x_eq_y_not_dt += 1 elif x_scatter_point > y_scatter_point: count_x_gt_y_not_dt += 1 else: assert x_scatter_point < y_scatter_point count_x_lt_y_not_dt += 1 # SMARTER counting: uses error bounds if analysis.bounds_overlap(x_scatter_point_bounds, y_scatter_point_bounds): # Bounds overlap, we can't compare the execution times in a meaningful way bounds_incomparable_keys.add(key) # However if both are timeouts we can note this if x_scatter_point == pargs.max_exec_time: x_eq_y_and_is_timeout_keys.add(key) else: # Compare the means if x_scatter_point > y_scatter_point and abs(x_scatter_point - y_scatter_point) > pargs.require_time_abs_diff: x_gt_y_keys.add(key) elif x_scatter_point < y_scatter_point and abs(x_scatter_point - y_scatter_point) > pargs.require_time_abs_diff: x_lt_y_keys.add(key) else: if pargs.require_time_abs_diff == 0.0: assert x_scatter_point == y_scatter_point x_eq_y_keys.add(key) # Report counts print("# of points : {}".format(len(x_scatter_points))) print("LEGACY: count_dual_timeout: {}".format(count_dual_timeout)) print("LEGACY: count_x_eq_y_not_dt: {}".format(count_x_eq_y_not_dt)) print("LEGACY: count_x_gt_y_not_dt: {}".format(count_x_gt_y_not_dt)) print("LEGACY: count_x_lt_y_not_dt: {}".format(count_x_lt_y_not_dt)) print("") print("# x > y and no bound overlap: {}".format(len(x_gt_y_keys))) print("# x < y and no bound overlap: {}".
assert isinstance(ri['event_tag'], list) event_tag, _ = event_analysis.merge_aggregate_events( ri['event_tag'])
conditional_block
result-info-plot-scatter-exec-time.py
def main(args): global _logger global _fail_count parser = argparse.ArgumentParser(description=__doc__) DriverUtil.parserAddLoggerArg(parser) parser.add_argument('first_result_info', type=argparse.FileType('r')) parser.add_argument('second_result_info', type=argparse.FileType('r')) parser.add_argument('--base', type=str, default="") parser.add_argument('--point-size', type=float, default=25.0, dest='point_size') parser.add_argument('--allow-merge-failures', dest='allow_merge_failures', default=False, action='store_true', ) parser.add_argument('--max-exec-time', default=None, type=float, dest='max_exec_time', ) parser.add_argument('--title', default="{num_keys} benchmarks, {num_points} jointly SAT or timeout" ) parser.add_argument("--xlabel", type=str, default=None, ) parser.add_argument("--ylabel", type=str, default=None, ) parser.add_argument("--axis-label-suffix", type=str, default=" execution time (s)", dest="axis_label_suffix", ) parser.add_argument("--axis-label-colour", type=str, default="black", dest="axis_label_colour", ) parser.add_argument("--annotate", default=False, action='store_true', ) parser.add_argument("--annotate-use-legacy-values", default=False, action='store_true', ) parser.add_argument("--output", default=None, type=argparse.FileType('wb'), ) parser.add_argument("--error-bars", default=False, action='store_true', ) parser.add_argument("--annotate-timeout-point", dest='annotate_timeout_point', default=False, action='store_true', ) parser.add_argument("--require-time-abs-diff", dest="require_time_abs_diff", default=0.0, type=float ) parser.add_argument('--true-type-fonts', default=False, action='store_true' ) pargs = parser.parse_args(args) DriverUtil.handleLoggerArgs(pargs, parser) _logger = logging.getLogger(__name__) if pargs.max_exec_time is None: _logger.error('--max-exec-time must be specified') return 1 if pargs.true_type_fonts: smtrunner.util.set_true_type_font() index_to_raw_result_infos = [] index_to_file_name = [] for index, result_infos_file in enumerate([pargs.first_result_info, pargs.second_result_info]): try: _logger.info('Loading "{}"'.format(result_infos_file.name)) result_infos = ResultInfo.loadRawResultInfos(result_infos_file) index_to_raw_result_infos.append(result_infos) index_to_file_name.append(result_infos_file.name) except ResultInfo.ResultInfoValidationError as e: _logger.error('Validation error:\n{}'.format(e)) return 1 _logger.info('Loading done') result_infos = None # Perform grouping by benchmark name key_to_results_infos, rejected_result_infos = ResultInfoUtil.group_result_infos_by( index_to_raw_result_infos) if len(rejected_result_infos) > 0: _logger.warning('There were rejected result infos') num_merge_failures = 0 for index, l in enumerate(rejected_result_infos): _logger.warning('Index {} had {} rejections'.format(index, len(l))) num_merge_failures += len(l) if num_merge_failures > 0: if pargs.allow_merge_failures: _logger.warning('Merge failures being allowed') else: _logger.error('Merge failures are not allowed') return 1 # Generate scatter points x_scatter_points = [] x_scatter_errors = [[], [] ] y_scatter_points = [] y_scatter_errors = [[], []] count_dual_timeout = 0 count_x_lt_y_not_dt = 0 count_x_gt_y_not_dt = 0 count_x_eq_y_not_dt = 0 # New counting vars bounds_incomparable_keys = set() x_gt_y_keys = set() x_lt_y_keys = set() x_eq_y_keys = set() x_eq_y_and_is_timeout_keys = set() for key, raw_result_info_list in sorted(key_to_results_infos.items(), key=lambda kv:kv[0]): _logger.info('Ranking on "{}" : '.format(key)) indices_to_use = [] # Compute indices to use modified_raw_result_info_list = [ ] # Handle "unknown" # Only compare results that gave sat/unsat for index, ri in enumerate(raw_result_info_list): if isinstance(ri['event_tag'], str): # single result event_tag = ri['event_tag'] else: assert isinstance(ri['event_tag'], list) event_tag, _ = event_analysis.merge_aggregate_events( ri['event_tag']) # Event must be sat or timeout _logger.info('index {} is {}'.format(index, event_tag)) if event_tag not in { 'sat', 'timeout', 'soft_timeout'}: # Skip this. We can't do a meaningful comparison here continue indices_to_use.append(index) # Normalise timeouts to have fixed values for the time. if event_tag in {'timeout', 'soft_timeout'}: modified_ri = analysis.get_result_with_modified_time( ri, pargs.max_exec_time) _logger.debug('modified_ri: {}'.format( pprint.pformat(modified_ri))) _logger.debug( 'Treating index {} for {} due to unknown as having max-time'.format( index, key)) modified_raw_result_info_list.append(modified_ri) else: modified_raw_result_info_list.append(ri) _logger.debug('used indices_to_use: {}'.format(indices_to_use)) if len(indices_to_use) != 2: # Skip this one. One of the result infos can't be compared # against. continue assert len(indices_to_use) == 2 # Get execution times index_to_execution_time_bounds = analysis.get_index_to_execution_time_bounds( modified_raw_result_info_list, indices_to_use, pargs.max_exec_time, analysis.get_arithmetic_mean_and_99_confidence_intervals, ['dsoes_wallclock', 'wallclock']) assert isinstance(index_to_execution_time_bounds, list) x_scatter_point_bounds = index_to_execution_time_bounds[0] y_scatter_point_bounds = index_to_execution_time_bounds[1] x_scatter_point = x_scatter_point_bounds[1] # mean y_scatter_point = y_scatter_point_bounds[1] # mean x_scatter_lower_error = x_scatter_point_bounds[1] - x_scatter_point_bounds[0] assert x_scatter_lower_error >= 0 x_scatter_higher_error = x_scatter_point_bounds[2] - x_scatter_point_bounds[1] assert x_scatter_higher_error >= 0 y_scatter_lower_error = y_scatter_point_bounds[1] - y_scatter_point_bounds[0] assert y_scatter_lower_error >= 0 y_scatter_higher_error = y_scatter_point_bounds[2] - y_scatter_point_bounds[1] assert y_scatter_higher_error >= 0 x_scatter_points.append(x_scatter_point) y_scatter_points.append(y_scatter_point) # Error bar points #x_scatter_errors.append((x_scatter_lower_error, x_scatter_higher_error)) x_scatter_errors[0].append(x_scatter_lower_error) x_scatter_errors[1].append(x_scatter_higher_error) #y_scatter_errors.append((y_scatter_lower_error, y_scatter_higher_error)) y_scatter_errors[0].append(y_scatter_lower_error) y_scatter_errors[1].append(y_scatter_higher_error) # LEGACY: Now do some counting if x_scatter_point == y_scatter_point: if x_scatter_point == pargs.max_exec_time: assert x_scatter_lower_error == 0 assert x_scatter_higher_error == 0 assert y_scatter_lower_error == 0 assert y_scatter_higher_error == 0 count_dual_timeout += 1 else: _logger.info('Found count_x_eq_y_not_dt: x: {}, key: {}'.format( x_scatter_point, key)) count_x_eq_y_not_dt += 1 elif x_scatter_point > y_scatter_point: count_x_gt_y_not_dt += 1 else: assert x_scatter_point < y_scatter_point count_x_lt_y_not_dt += 1 # SMARTER counting: uses error bounds if analysis.bounds_overlap(x_scatter_point_bounds, y_scatter_point_bounds): # Bounds overlap, we can't compare the execution times in a meaningful way bounds_incomparable_keys.add(key) # However if both are timeouts we can note this if x_scatter_point == pargs.max_exec_time: x_eq_y_and_is_timeout_keys.add(key
if prefix == "": return path if path.startswith(prefix): return path[len(prefix):]
identifier_body
result-info-plot-scatter-exec-time.py
default=False, action='store_true', ) parser.add_argument("--output", default=None, type=argparse.FileType('wb'), ) parser.add_argument("--error-bars", default=False, action='store_true', ) parser.add_argument("--annotate-timeout-point", dest='annotate_timeout_point', default=False, action='store_true', ) parser.add_argument("--require-time-abs-diff", dest="require_time_abs_diff", default=0.0, type=float ) parser.add_argument('--true-type-fonts', default=False, action='store_true' ) pargs = parser.parse_args(args) DriverUtil.handleLoggerArgs(pargs, parser) _logger = logging.getLogger(__name__) if pargs.max_exec_time is None: _logger.error('--max-exec-time must be specified') return 1 if pargs.true_type_fonts: smtrunner.util.set_true_type_font() index_to_raw_result_infos = [] index_to_file_name = [] for index, result_infos_file in enumerate([pargs.first_result_info, pargs.second_result_info]): try: _logger.info('Loading "{}"'.format(result_infos_file.name)) result_infos = ResultInfo.loadRawResultInfos(result_infos_file) index_to_raw_result_infos.append(result_infos) index_to_file_name.append(result_infos_file.name) except ResultInfo.ResultInfoValidationError as e: _logger.error('Validation error:\n{}'.format(e)) return 1 _logger.info('Loading done') result_infos = None # Perform grouping by benchmark name key_to_results_infos, rejected_result_infos = ResultInfoUtil.group_result_infos_by( index_to_raw_result_infos) if len(rejected_result_infos) > 0: _logger.warning('There were rejected result infos') num_merge_failures = 0 for index, l in enumerate(rejected_result_infos): _logger.warning('Index {} had {} rejections'.format(index, len(l))) num_merge_failures += len(l) if num_merge_failures > 0: if pargs.allow_merge_failures: _logger.warning('Merge failures being allowed') else: _logger.error('Merge failures are not allowed') return 1 # Generate scatter points x_scatter_points = [] x_scatter_errors = [[], [] ] y_scatter_points = [] y_scatter_errors = [[], []] count_dual_timeout = 0 count_x_lt_y_not_dt = 0 count_x_gt_y_not_dt = 0 count_x_eq_y_not_dt = 0 # New counting vars bounds_incomparable_keys = set() x_gt_y_keys = set() x_lt_y_keys = set() x_eq_y_keys = set() x_eq_y_and_is_timeout_keys = set() for key, raw_result_info_list in sorted(key_to_results_infos.items(), key=lambda kv:kv[0]): _logger.info('Ranking on "{}" : '.format(key)) indices_to_use = [] # Compute indices to use modified_raw_result_info_list = [ ] # Handle "unknown" # Only compare results that gave sat/unsat for index, ri in enumerate(raw_result_info_list): if isinstance(ri['event_tag'], str): # single result event_tag = ri['event_tag'] else: assert isinstance(ri['event_tag'], list) event_tag, _ = event_analysis.merge_aggregate_events( ri['event_tag']) # Event must be sat or timeout _logger.info('index {} is {}'.format(index, event_tag)) if event_tag not in { 'sat', 'timeout', 'soft_timeout'}: # Skip this. We can't do a meaningful comparison here continue indices_to_use.append(index) # Normalise timeouts to have fixed values for the time. if event_tag in {'timeout', 'soft_timeout'}: modified_ri = analysis.get_result_with_modified_time( ri, pargs.max_exec_time) _logger.debug('modified_ri: {}'.format( pprint.pformat(modified_ri))) _logger.debug( 'Treating index {} for {} due to unknown as having max-time'.format( index, key)) modified_raw_result_info_list.append(modified_ri) else: modified_raw_result_info_list.append(ri) _logger.debug('used indices_to_use: {}'.format(indices_to_use)) if len(indices_to_use) != 2: # Skip this one. One of the result infos can't be compared # against. continue assert len(indices_to_use) == 2 # Get execution times index_to_execution_time_bounds = analysis.get_index_to_execution_time_bounds( modified_raw_result_info_list, indices_to_use, pargs.max_exec_time, analysis.get_arithmetic_mean_and_99_confidence_intervals, ['dsoes_wallclock', 'wallclock']) assert isinstance(index_to_execution_time_bounds, list) x_scatter_point_bounds = index_to_execution_time_bounds[0] y_scatter_point_bounds = index_to_execution_time_bounds[1] x_scatter_point = x_scatter_point_bounds[1] # mean y_scatter_point = y_scatter_point_bounds[1] # mean x_scatter_lower_error = x_scatter_point_bounds[1] - x_scatter_point_bounds[0] assert x_scatter_lower_error >= 0 x_scatter_higher_error = x_scatter_point_bounds[2] - x_scatter_point_bounds[1] assert x_scatter_higher_error >= 0 y_scatter_lower_error = y_scatter_point_bounds[1] - y_scatter_point_bounds[0] assert y_scatter_lower_error >= 0 y_scatter_higher_error = y_scatter_point_bounds[2] - y_scatter_point_bounds[1] assert y_scatter_higher_error >= 0 x_scatter_points.append(x_scatter_point) y_scatter_points.append(y_scatter_point) # Error bar points #x_scatter_errors.append((x_scatter_lower_error, x_scatter_higher_error)) x_scatter_errors[0].append(x_scatter_lower_error) x_scatter_errors[1].append(x_scatter_higher_error) #y_scatter_errors.append((y_scatter_lower_error, y_scatter_higher_error)) y_scatter_errors[0].append(y_scatter_lower_error) y_scatter_errors[1].append(y_scatter_higher_error) # LEGACY: Now do some counting if x_scatter_point == y_scatter_point: if x_scatter_point == pargs.max_exec_time: assert x_scatter_lower_error == 0 assert x_scatter_higher_error == 0 assert y_scatter_lower_error == 0 assert y_scatter_higher_error == 0 count_dual_timeout += 1 else: _logger.info('Found count_x_eq_y_not_dt: x: {}, key: {}'.format( x_scatter_point, key)) count_x_eq_y_not_dt += 1 elif x_scatter_point > y_scatter_point: count_x_gt_y_not_dt += 1 else: assert x_scatter_point < y_scatter_point count_x_lt_y_not_dt += 1 # SMARTER counting: uses error bounds if analysis.bounds_overlap(x_scatter_point_bounds, y_scatter_point_bounds): # Bounds overlap, we can't compare the execution times in a meaningful way bounds_incomparable_keys.add(key) # However if both are timeouts we can note this if x_scatter_point == pargs.max_exec_time: x_eq_y_and_is_timeout_keys.add(key) else: # Compare the means if x_scatter_point > y_scatter_point and abs(x_scatter_point - y_scatter_point) > pargs.require_time_abs_diff: x_gt_y_keys.add(key) elif x_scatter_point < y_scatter_point and abs(x_scatter_point - y_scatter_point) > pargs.require_time_abs_diff: x_lt_y_keys.add(key) else: if pargs.require_time_abs_diff == 0.0: assert x_scatter_point == y_scatter_point x_eq_y_keys.add(key) # Report counts print("# of points : {}".format(len(x_scatter_points))) print("LEGACY: count_dual_timeout: {}".format(count_dual_timeout)) print("LEGACY: count_x_eq_y_not_dt: {}".format(count_x_eq_y_not_dt)) print("LEGACY: count_x_gt_y_not_dt: {}".format(count_x_gt_y_not_dt)) print("LEGACY: count_x_lt_y_not_dt: {}".format(count_x_lt_y_not_dt)) print("") print("# x > y and no bound overlap: {}".format(len(x_gt_y_keys))) print("# x < y and no bound overlap: {}".format(len(x_lt_y_keys))) print("# x = y and no bound overlap: {}".format(len(x_eq_y_keys)))
print("# incomparable: {}".format(len(bounds_incomparable_keys))) print("# of x = y and is timeout: {}".format(len(x_eq_y_and_is_timeout_keys))) # Now plot extend = 100
random_line_split
result-info-plot-scatter-exec-time.py
(prefix, path): if prefix == "": return path if path.startswith(prefix): return path[len(prefix):] def main(args): global _logger global _fail_count parser = argparse.ArgumentParser(description=__doc__) DriverUtil.parserAddLoggerArg(parser) parser.add_argument('first_result_info', type=argparse.FileType('r')) parser.add_argument('second_result_info', type=argparse.FileType('r')) parser.add_argument('--base', type=str, default="") parser.add_argument('--point-size', type=float, default=25.0, dest='point_size') parser.add_argument('--allow-merge-failures', dest='allow_merge_failures', default=False, action='store_true', ) parser.add_argument('--max-exec-time', default=None, type=float, dest='max_exec_time', ) parser.add_argument('--title', default="{num_keys} benchmarks, {num_points} jointly SAT or timeout" ) parser.add_argument("--xlabel", type=str, default=None, ) parser.add_argument("--ylabel", type=str, default=None, ) parser.add_argument("--axis-label-suffix", type=str, default=" execution time (s)", dest="axis_label_suffix", ) parser.add_argument("--axis-label-colour", type=str, default="black", dest="axis_label_colour", ) parser.add_argument("--annotate", default=False, action='store_true', ) parser.add_argument("--annotate-use-legacy-values", default=False, action='store_true', ) parser.add_argument("--output", default=None, type=argparse.FileType('wb'), ) parser.add_argument("--error-bars", default=False, action='store_true', ) parser.add_argument("--annotate-timeout-point", dest='annotate_timeout_point', default=False, action='store_true', ) parser.add_argument("--require-time-abs-diff", dest="require_time_abs_diff", default=0.0, type=float ) parser.add_argument('--true-type-fonts', default=False, action='store_true' ) pargs = parser.parse_args(args) DriverUtil.handleLoggerArgs(pargs, parser) _logger = logging.getLogger(__name__) if pargs.max_exec_time is None: _logger.error('--max-exec-time must be specified') return 1 if pargs.true_type_fonts: smtrunner.util.set_true_type_font() index_to_raw_result_infos = [] index_to_file_name = [] for index, result_infos_file in enumerate([pargs.first_result_info, pargs.second_result_info]): try: _logger.info('Loading "{}"'.format(result_infos_file.name)) result_infos = ResultInfo.loadRawResultInfos(result_infos_file) index_to_raw_result_infos.append(result_infos) index_to_file_name.append(result_infos_file.name) except ResultInfo.ResultInfoValidationError as e: _logger.error('Validation error:\n{}'.format(e)) return 1 _logger.info('Loading done') result_infos = None # Perform grouping by benchmark name key_to_results_infos, rejected_result_infos = ResultInfoUtil.group_result_infos_by( index_to_raw_result_infos) if len(rejected_result_infos) > 0: _logger.warning('There were rejected result infos') num_merge_failures = 0 for index, l in enumerate(rejected_result_infos): _logger.warning('Index {} had {} rejections'.format(index, len(l))) num_merge_failures += len(l) if num_merge_failures > 0: if pargs.allow_merge_failures: _logger.warning('Merge failures being allowed') else: _logger.error('Merge failures are not allowed') return 1 # Generate scatter points x_scatter_points = [] x_scatter_errors = [[], [] ] y_scatter_points = [] y_scatter_errors = [[], []] count_dual_timeout = 0 count_x_lt_y_not_dt = 0 count_x_gt_y_not_dt = 0 count_x_eq_y_not_dt = 0 # New counting vars bounds_incomparable_keys = set() x_gt_y_keys = set() x_lt_y_keys = set() x_eq_y_keys = set() x_eq_y_and_is_timeout_keys = set() for key, raw_result_info_list in sorted(key_to_results_infos.items(), key=lambda kv:kv[0]): _logger.info('Ranking on "{}" : '.format(key)) indices_to_use = [] # Compute indices to use modified_raw_result_info_list = [ ] # Handle "unknown" # Only compare results that gave sat/unsat for index, ri in enumerate(raw_result_info_list): if isinstance(ri['event_tag'], str): # single result event_tag = ri['event_tag'] else: assert isinstance(ri['event_tag'], list) event_tag, _ = event_analysis.merge_aggregate_events( ri['event_tag']) # Event must be sat or timeout _logger.info('index {} is {}'.format(index, event_tag)) if event_tag not in { 'sat', 'timeout', 'soft_timeout'}: # Skip this. We can't do a meaningful comparison here continue indices_to_use.append(index) # Normalise timeouts to have fixed values for the time. if event_tag in {'timeout', 'soft_timeout'}: modified_ri = analysis.get_result_with_modified_time( ri, pargs.max_exec_time) _logger.debug('modified_ri: {}'.format( pprint.pformat(modified_ri))) _logger.debug( 'Treating index {} for {} due to unknown as having max-time'.format( index, key)) modified_raw_result_info_list.append(modified_ri) else: modified_raw_result_info_list.append(ri) _logger.debug('used indices_to_use: {}'.format(indices_to_use)) if len(indices_to_use) != 2: # Skip this one. One of the result infos can't be compared # against. continue assert len(indices_to_use) == 2 # Get execution times index_to_execution_time_bounds = analysis.get_index_to_execution_time_bounds( modified_raw_result_info_list, indices_to_use, pargs.max_exec_time, analysis.get_arithmetic_mean_and_99_confidence_intervals, ['dsoes_wallclock', 'wallclock']) assert isinstance(index_to_execution_time_bounds, list) x_scatter_point_bounds = index_to_execution_time_bounds[0] y_scatter_point_bounds = index_to_execution_time_bounds[1] x_scatter_point = x_scatter_point_bounds[1] # mean y_scatter_point = y_scatter_point_bounds[1] # mean x_scatter_lower_error = x_scatter_point_bounds[1] - x_scatter_point_bounds[0] assert x_scatter_lower_error >= 0 x_scatter_higher_error = x_scatter_point_bounds[2] - x_scatter_point_bounds[1] assert x_scatter_higher_error >= 0 y_scatter_lower_error = y_scatter_point_bounds[1] - y_scatter_point_bounds[0] assert y_scatter_lower_error >= 0 y_scatter_higher_error = y_scatter_point_bounds[2] - y_scatter_point_bounds[1] assert y_scatter_higher_error >= 0 x_scatter_points.append(x_scatter_point) y_scatter_points.append(y_scatter_point) # Error bar points #x_scatter_errors.append((x_scatter_lower_error, x_scatter_higher_error)) x_scatter_errors[0].append(x_scatter_lower_error) x_scatter_errors[1].append(x_scatter_higher_error) #y_scatter_errors.append((y_scatter_lower_error, y_scatter_higher_error)) y_scatter_errors[0].append(y_scatter_lower_error) y_scatter_errors[1].append(y_scatter_higher_error) # LEGACY: Now do some counting if x_scatter_point == y_scatter_point: if x_scatter_point == pargs.max_exec_time: assert x_scatter_lower_error == 0 assert x_scatter_higher_error == 0 assert y_scatter_lower_error == 0 assert y_scatter_higher_error == 0 count_dual_timeout += 1 else: _logger.info('Found count_x_eq_y_not_dt: x: {}, key: {}'.format( x_scatter_point, key)) count_x_eq_y_not_dt += 1 elif x_scatter_point > y_scatter_point: count_x_gt_y_not_dt += 1 else: assert x_scatter_point < y_scatter_point count_x_lt_y_not_dt += 1 # SMARTER counting: uses error bounds if analysis.bounds_overlap(x_scatter_point_bounds, y_scatter_point_bounds): # Bounds overlap, we can't compare the execution times in a meaningful way bounds_incomparable_keys.add(key) # However if both are timeouts we can note this if x_scatter_point == pargs.max_exec_time: x_eq_y_and
strip
identifier_name
api.js
Descrizione" + SPACE + "unità" + SPACE + "di" + SPACE + "stima", jsonId:"Descrizione_unita_di_stima", skip:1}, {text:"Valore" + SPACE + "di" + SPACE + "mercato" + SPACE + "del" + SPACE + "lotto", jsonId:"Valore_di_mercato_del_lotto", skip:2, number:true}, {text:"Foglio" + SPACE + "di" + SPACE + "mappa", jsonId:"Foglio", skip:1}, {text:"Totale" + SPACE + "superficie" + SPACE + "principale", jsonId:"Totale_superficie_principale", skip:2, number:true}, {text:"Impianto" + SPACE + "elettrico", jsonId:"Impianto_elettrico_anni", skip:2}, {text:"Impianto" + SPACE + "idraulico", jsonId:"Impianto_idraulico_anni", skip:2}, {text:"Tipologia" + SPACE + "edilizia" , jsonId:"Tipologia_edilizia", skip:1}, {text:"COMMERCIALE", jsonId:"SUPERFICIE_COMMERCIALE_MQ", skip:2, number:true}, {text:"Particella", jsonId:"Particella", skip:7}, {text:"Categoria", jsonId:"Categoria", skip:7}, {text:"Consistenza", jsonId:"Consistenza", skip:7, number:true}, {text:"RC", jsonId:"RC", skip:7, number:true}, {text:"Anno" + SPACE + "di" + SPACE + "costruzione", jsonId:"Anno_di_costruzione", skip:1} ]; var superficiLines = [ {text:"Descrizione", jsonId:"Descrizione", skip:1}, {text:"Misura (mq)", jsonId:"Misura_mq", skip:1, number:true}, {text:"Rapporto mercantile", jsonId:"Rapporto_mercantile", skip:1, number:true}, {text:"Sup Rap", jsonId:"Sup_Rap", skip:1, number:true} ]; var parseJsonPdf = function(arr){ var str = ''; var jsonString = '{'; for (var i = 0; i < arr.length; i ++) { for (var j = 0; j < arr[i].Texts.length; j ++) { str = decodeURIComponent(arr[i].Texts[j].R[0].T); if (str.lastIndexOf('COLLEGATE', 0) === 0){ jsonString = jsonString + '\"SUPERFICI_SECONDARIE_ANNESSE_E_COLLEGATE\":\n['; j = j + 4; for (var h = 1; h < 20; h ++){ j = j + 1 str = decodeURIComponent(arr[i].Texts[j].R[0].T); if (str.lastIndexOf(h.toString() + '.', 0) === 0){ if (h != 1){ jsonString += ','; } jsonString = jsonString + '{'; for (var w = 0; w < superficiLines.length; w ++){ j = j + 1; str = decodeURIComponent(arr[i].Texts[j].R[0].T); if (superficiLines[w].number){ str = str.replace('.','').replace(',','.'); } jsonString = jsonString + '\"' + superficiLines[w].jsonId + '\":\"' + str + '\"'; if (w != superficiLines.length - 1) { jsonString += ','; } jsonString += '\n'; } jsonString += '}\n'; //jsonString = jsonString + '},\n'; } else { jsonString += '],\n'; h = 20; } } } else if (str.lastIndexOf(SPACE + 'UNITA\'' + SPACE + 'IMMOBILIARE', 0 ) === 0){ j = j + 1 str = decodeURIComponent(arr[i].Texts[j].R[0].T); if (str.lastIndexOf('UBICAZIONE', 0) === 0){ for (var w = 0; w < ubicazioneLines.length; w ++){ j = j + 1; str = decodeURIComponent(arr[i].Texts[j].R[0].T); if (str.lastIndexOf(ubicazioneLines[w].text, 0) === 0){ j = j + 1; str = decodeURIComponent(arr[i].Texts[j].R[0].T); jsonString += '\"' + ubicazioneLines[w].jsonId + '\":\"' + str + '\",\n'; } } } } else { for (var w = 0; w < relevantLines.length; w ++){ if (str.lastIndexOf(relevantLines[w].text, 0) === 0){ j = j + relevantLines[w].skip; var text = decodeURIComponent(arr[i].Texts[j].R[0].T); //trasforma un numero in formato italiano in un numero in formato json standard //e.g. 1,23 -> 1.23 if (relevantLines[w].number){ text = text.replace('.','').replace(',','.'); } else if (relevantLines[w].date){ //trasforma una data in formato italiano dd/mm/yyyy in una data in formato json standard text = moment(text, 'DD/MM/YYYY').format('MM/DD/YYYY'); } jsonString += '\"' + relevantLines[w].jsonId + '\":\"' + text + '\"'; j = j - relevantLines[w].skip; if (w != ubicazioneLines.length - 1) { jsonString += ','; } jsonString += '\n'; } } } } } jsonString += '}'; return jsonString; } var salvaPerizia = function (json){ console.log("salvo la perizia"); var uploadedPerizia = new Perizia(json); uploadedPerizia.save(function(err) { if (err) { //console.log(err.code); if (err.code === 11000){//duplicate key console.log(err.message); return; } else { console.log(err); } } else { console.log('La perizia CRIF ' + json.Nome_File + ' è stata salvata in mongo'); } }); } router.post('/upload', function(req, res){ upload(req, res, function(err){ if (err){ res.json({error_code:1, err_desc:err}); return; } res.json({error_code:0, err_desc:null}); var pdfParser = new PDFParser(this, 1); pdfParser.on("pdfParser_dataError", errData => console.error(errData.parserError) ); pdfParser.on("pdfParser_dataReady", pdfData => { var jsonPdf = JSON.parse(JSON.stringify(pdfData, null, '\t')).formImage.Pages; var jsonString = parseJsonPdf(jsonPdf); fs.writeFile(uploadDestination + path.sep + req.file.filename + '.txt', jsonString, function(errf) { if (errf) { return console.log(errf); } jsonPdf = null; }); var jsonMongo = JSON.parse(jsonString); jsonMongo.Nome_File = req.file.filename; //salvaPerizia(jsonMongo); //usa Nominatim per ottenere longitudine e latitudine associate al luogo della perizia //quindi salva la perizia in mongo var indirizzo = jsonMongo.Indirizzo + " " + jsonMongo.N_civico + ", " + jsonMongo.Comune + ", " + jsonMongo.Provincia + ", " + jsonMongo.CAP; nominatim.search({q: indirizzo}, function (err, res, data) { console.log(indirizzo); if (err) { throw err; } var loc = new Array(); if (data != undefined && data[0] != undefined){ loc = [data[0].lon, data[0].lat]; jsonMongo.loc = loc; salvaPerizia(jsonMongo); } else {
var indirizzo2 = jsonMongo.Indirizzo + " " + jsonMongo.N_civico + ", " + jsonMongo.Comune + ", " + jsonMongo.Provincia; console.log(indirizzo2); nominatim.search({q: indirizzo2}, function (err2, res2, data2) { if (err2) { throw err2; } if (data2 != undefined && data2[0] != undefined) { loc = [data2[0].lon, data2[0].lat]; jsonMongo.loc = loc; salvaPerizia(jsonMongo); } else { salvaPerizia(jsonMongo); } }); }
conditional_block
api.js
:"Provincia", skip:1}, {text:"CAP", jsonId:"CAP", skip:1}, {text:"Indirizzo", jsonId:"Indirizzo", skip:1}, {text:"N." + SPACE + "civico", jsonId:"N_civico", skip:1}, {text:"Interno", jsonId:"Interno", skip:1}, {text:"Scala", jsonId:"Scala", skip:1}, {text:"Piano", jsonId:"Piano", skip:1} ]; var relevantLines = [ {text:"Codice" + SPACE + "CRIF", jsonId:"_id", skip:1}, {text:"Data" + SPACE + "Evasione" + SPACE + "Perizia", jsonId:"Data_Evasione_Perizia", skip:1, date: true}, {text:"Descrizione" + SPACE + "unità" + SPACE + "di" + SPACE + "stima", jsonId:"Descrizione_unita_di_stima", skip:1}, {text:"Valore" + SPACE + "di" + SPACE + "mercato" + SPACE + "del" + SPACE + "lotto", jsonId:"Valore_di_mercato_del_lotto", skip:2, number:true}, {text:"Foglio" + SPACE + "di" + SPACE + "mappa", jsonId:"Foglio", skip:1}, {text:"Totale" + SPACE + "superficie" + SPACE + "principale", jsonId:"Totale_superficie_principale", skip:2, number:true}, {text:"Impianto" + SPACE + "elettrico", jsonId:"Impianto_elettrico_anni", skip:2}, {text:"Impianto" + SPACE + "idraulico", jsonId:"Impianto_idraulico_anni", skip:2}, {text:"Tipologia" + SPACE + "edilizia" , jsonId:"Tipologia_edilizia", skip:1}, {text:"COMMERCIALE", jsonId:"SUPERFICIE_COMMERCIALE_MQ", skip:2, number:true}, {text:"Particella", jsonId:"Particella", skip:7}, {text:"Categoria", jsonId:"Categoria", skip:7}, {text:"Consistenza", jsonId:"Consistenza", skip:7, number:true}, {text:"RC", jsonId:"RC", skip:7, number:true}, {text:"Anno" + SPACE + "di" + SPACE + "costruzione", jsonId:"Anno_di_costruzione", skip:1} ]; var superficiLines = [ {text:"Descrizione", jsonId:"Descrizione", skip:1}, {text:"Misura (mq)", jsonId:"Misura_mq", skip:1, number:true}, {text:"Rapporto mercantile", jsonId:"Rapporto_mercantile", skip:1, number:true}, {text:"Sup Rap", jsonId:"Sup_Rap", skip:1, number:true} ]; var parseJsonPdf = function(arr){ var str = ''; var jsonString = '{'; for (var i = 0; i < arr.length; i ++) { for (var j = 0; j < arr[i].Texts.length; j ++) { str = decodeURIComponent(arr[i].Texts[j].R[0].T); if (str.lastIndexOf('COLLEGATE', 0) === 0){ jsonString = jsonString + '\"SUPERFICI_SECONDARIE_ANNESSE_E_COLLEGATE\":\n['; j = j + 4; for (var h = 1; h < 20; h ++){ j = j + 1 str = decodeURIComponent(arr[i].Texts[j].R[0].T); if (str.lastIndexOf(h.toString() + '.', 0) === 0){ if (h != 1){ jsonString += ','; } jsonString = jsonString + '{'; for (var w = 0; w < superficiLines.length; w ++){ j = j + 1; str = decodeURIComponent(arr[i].Texts[j].R[0].T); if (superficiLines[w].number){ str = str.replace('.','').replace(',','.'); } jsonString = jsonString + '\"' + superficiLines[w].jsonId + '\":\"' + str + '\"'; if (w != superficiLines.length - 1) { jsonString += ','; } jsonString += '\n'; } jsonString += '}\n'; //jsonString = jsonString + '},\n'; } else { jsonString += '],\n'; h = 20; } } } else if (str.lastIndexOf(SPACE + 'UNITA\'' + SPACE + 'IMMOBILIARE', 0 ) === 0){ j = j + 1 str = decodeURIComponent(arr[i].Texts[j].R[0].T); if (str.lastIndexOf('UBICAZIONE', 0) === 0){ for (var w = 0; w < ubicazioneLines.length; w ++){ j = j + 1; str = decodeURIComponent(arr[i].Texts[j].R[0].T); if (str.lastIndexOf(ubicazioneLines[w].text, 0) === 0){ j = j + 1; str = decodeURIComponent(arr[i].Texts[j].R[0].T); jsonString += '\"' + ubicazioneLines[w].jsonId + '\":\"' + str + '\",\n'; } } } } else { for (var w = 0; w < relevantLines.length; w ++){ if (str.lastIndexOf(relevantLines[w].text, 0) === 0){ j = j + relevantLines[w].skip; var text = decodeURIComponent(arr[i].Texts[j].R[0].T); //trasforma un numero in formato italiano in un numero in formato json standard //e.g. 1,23 -> 1.23 if (relevantLines[w].number){ text = text.replace('.','').replace(',','.'); } else if (relevantLines[w].date){ //trasforma una data in formato italiano dd/mm/yyyy in una data in formato json standard text = moment(text, 'DD/MM/YYYY').format('MM/DD/YYYY'); } jsonString += '\"' + relevantLines[w].jsonId + '\":\"' + text + '\"'; j = j - relevantLines[w].skip; if (w != ubicazioneLines.length - 1) { jsonString += ','; } jsonString += '\n'; } } } } } jsonString += '}'; return jsonString; } var salvaPerizia = function (json){ console.log("salvo la perizia"); var uploadedPerizia = new Perizia(json); uploadedPerizia.save(function(err) { if (err) { //console.log(err.code); if (err.code === 11000){//duplicate key console.log(err.message); return;
console.log('La perizia CRIF ' + json.Nome_File + ' è stata salvata in mongo'); } }); } router.post('/upload', function(req, res){ upload(req, res, function(err){ if (err){ res.json({error_code:1, err_desc:err}); return; } res.json({error_code:0, err_desc:null}); var pdfParser = new PDFParser(this, 1); pdfParser.on("pdfParser_dataError", errData => console.error(errData.parserError) ); pdfParser.on("pdfParser_dataReady", pdfData => { var jsonPdf = JSON.parse(JSON.stringify(pdfData, null, '\t')).formImage.Pages; var jsonString = parseJsonPdf(jsonPdf); fs.writeFile(uploadDestination + path.sep + req.file.filename + '.txt', jsonString, function(errf) { if (errf) { return console.log(errf); } jsonPdf = null; }); var jsonMongo = JSON.parse(jsonString); jsonMongo.Nome_File = req.file.filename; //salvaPerizia(jsonMongo); //usa Nominatim per ottenere longitudine e latitudine associate al luogo della perizia //quindi salva la perizia in mongo var indirizzo = jsonMongo.Indirizzo + " " + jsonMongo.N_civico + ", " + jsonMongo.Comune + ", " + jsonMongo.Provincia + ", " + jsonMongo.CAP; nominatim.search({q: indirizzo}, function (err, res, data) { console.log(indirizzo); if (err) { throw err; } var loc = new Array(); if (data != undefined && data[0] != undefined){ loc = [data[
} else { console.log(err); } } else {
random_line_split
apply.rs
/// * `values` - Vec with paths fn random(values: Vec<path::PathBuf>) -> Result<path::PathBuf> { let chosen = values.choose(&mut rand::thread_rng()).ok_or_else(|| { anyhow!( "Scheme not found. Check if it exists, or run update schemes if you didn't already." ) })?; Ok(chosen.to_path_buf()) } /// Runs hook commands /// /// * `command` - Command string to execute /// * `verbose` - Should we be verbose? fn run_hook(command: Option<String>, shell: &str, verbose: bool) -> Result<()> { if let Some(command) = command { let full_command = shell.replace("{}", &command); if verbose { println!("running {}", full_command); } let command_vec = shell_words::split(&full_command)?; if command_vec.len() == 1
else { process::Command::new(&command_vec[0]) .args(&command_vec[1..]) .stdout(process::Stdio::null()) .stderr(process::Stdio::null()) .status() .with_context(|| format!("Couldn't run hook '{}'", full_command))?; } } Ok(()) } /// Replace with delimiter lines /// /// In a string, removes everything from one line to another, and puts the built template in place /// /// * `file_content` - String with lines to be replaced /// * `start` - Where to start replacing /// * `end` - Where to stop replacing /// * `built_template` - Built template to be injected fn replace_delimiter( file_content: &str, start: &str, end: &str, built_template: &str, ) -> Result<String> { let mut changed_content = String::new(); let mut found_start = false; let mut found_end = false; let mut appended = false; for line in file_content.lines() { if found_start && !found_end { if !appended { changed_content.push_str(&built_template); appended = true; } if line.trim().to_lowercase().eq(&end) { changed_content.push_str(&format!("{}\n", line)); found_end = true; } } else { changed_content.push_str(&format!("{}\n", line)); if line.trim().to_lowercase().eq(&start) { found_start = true } } } if !found_start { Err(anyhow!("Couldn't find starting string.")) } else if !found_end { Err(anyhow!("Couldn't find ending string.")) } else { Ok(changed_content) } } /// Apply function /// /// * `patterns` - Which patterns the user specified /// * `base_dir` - Flavours base directory /// * `config_path` - Flavours configuration path /// * `light` - Don't run hooks marked as non-lightweight /// * `from_stdin` - Read scheme from stdin? /// * `verbose` - Should we be verbose? pub fn apply( patterns: Vec<&str>, base_dir: &path::Path, config_path: &path::Path, light_mode: bool, from_stdin: bool, verbose: bool, ) -> Result<()> { let (scheme_contents, scheme_slug) = if from_stdin { let mut buffer = String::new(); let stdin = io::stdin(); let mut handle = stdin.lock(); handle.read_to_string(&mut buffer)?; (buffer, String::from("generated")) } else { //Find schemes that match given patterns let mut schemes = Vec::new(); for pattern in patterns { let found_schemes = find(pattern, &base_dir.join("base16").join("schemes"))?; for found_scheme in found_schemes { schemes.push(found_scheme); } } //Sort and remove duplicates schemes.sort(); schemes.dedup(); //Get random scheme let scheme_file = random(schemes)?; let scheme_slug: String = scheme_file .file_stem() .ok_or_else(|| anyhow!("Couldn't get scheme name."))? .to_str() .ok_or_else(|| anyhow!("Couldn't convert scheme file name."))? .into(); //Read chosen scheme ( fs::read_to_string(&scheme_file) .with_context(|| format!("Couldn't read scheme file at {:?}.", scheme_file))?, scheme_slug, ) }; let scheme = Scheme::from_str(&scheme_contents, &scheme_slug)?; if verbose { println!( "Using scheme: {} ({}), by {}", scheme.name, scheme.slug, scheme.author ); println!(); } //Check if config file exists if !config_path.exists() { eprintln!("Config {:?} doesn't exist, creating", config_path); let default_content = match fs::read_to_string(path::Path::new("/etc/flavours.conf")) { Ok(content) => content, Err(_) => String::from(""), }; let config_path_parent = config_path .parent() .with_context(|| format!("Couldn't get parent directory of {:?}", config_path))?; fs::create_dir_all(config_path_parent).with_context(|| { format!( "Couldn't create configuration file parent directory {:?}", config_path_parent ) })?; fs::write(config_path, default_content) .with_context(|| format!("Couldn't create configuration file at {:?}", config_path))?; } let config_contents = fs::read_to_string(config_path) .with_context(|| format!("Couldn't read configuration file {:?}.", config_path))?; let config = Config::from_str(&config_contents)?; // If shell is present, check if it contains the placeholder let shell = config.shell.unwrap_or_else(|| "sh -c '{}'".into()); if !shell.contains("{}") { return Err(anyhow!("The configured shell does not contain the required command placeholder '{}'. Check the default file or github for config examples.")); } let mut hooks = Vec::new(); //Iterate configurated entries (templates) let items_legacy = config.item.unwrap_or_default(); let mut items = config.items.unwrap_or_default(); items.extend(items_legacy.into_iter()) ; if items.is_empty() { return Err(anyhow!("Couldn't get items from config file. Check the default file or github for config examples.")); } for item in items.iter() { //Template name let template = &item.template; //Subtemplate name let subtemplate = match &item.subtemplate { Some(value) => String::from(value), None => String::from("default"), }; //Is the hook lightweight? let light = match &item.light { Some(value) => *value, None => true, }; //Rewrite or replace let rewrite = match &item.rewrite { Some(value) => *value, None => false, }; //Replace start delimiter let start = match &item.start { Some(value) => String::from(value), None => String::from("# Start flavours"), } .trim() .to_lowercase(); //Replace end delimiter let end = match &item.end { Some(value) => String::from(value), None => String::from("# End flavours"), } .trim() .to_lowercase(); //(sub)template file path let subtemplate_file = &base_dir .join("base16") .join("templates") .join(&template) .join("templates") .join(format!("{}.mustache", subtemplate)); //Template content let template_content = fs::read_to_string(subtemplate_file) .with_context(||format!("Couldn't read template {}/{} at {:?}. Check if the correct template/subtemplate was specified, and run the update templates command if you didn't already.", template, subtemplate, subtemplate_file))?; //Template with correct colors let built_template = build_template(template_content, &scheme) .context("Couldn't replace placeholders. Check if all colors on the specified scheme file are valid (don't include a leading '#').")?; //File to write let file = shellexpand::full(&item.file)?.to_string(); //Rewrite file with built template if rewrite { std::path::Path::new(&file).parent().and_then(|p| fs::create_dir_all(p).ok()); fs::write(&file, built_template) .with_context(|| format!("Couldn't write to file {:?}.", file))?; if verbose { println!("Wrote {}/{} on: {:?}", template, subtemplate, file) } } else { //Or replace with delimiters let file_content = fs::read_to_string(&file)?; match replace_delimiter(&file_content, &start, &end, &built_template) { Ok(content) => fs::write(&file, content) .with_context(|| format!("Couldn't write to file {:?}", file))?, Err(error) =>
{ process::Command::new(&command_vec[0]) .stdout(process::Stdio::null()) .stderr(process::Stdio::null()) .status() .with_context(|| format!("Couldn't run hook '{}'", full_command))?; }
conditional_block
apply.rs
/// * `values` - Vec with paths fn random(values: Vec<path::PathBuf>) -> Result<path::PathBuf> { let chosen = values.choose(&mut rand::thread_rng()).ok_or_else(|| { anyhow!( "Scheme not found. Check if it exists, or run update schemes if you didn't already." ) })?; Ok(chosen.to_path_buf()) } /// Runs hook commands /// /// * `command` - Command string to execute /// * `verbose` - Should we be verbose? fn run_hook(command: Option<String>, shell: &str, verbose: bool) -> Result<()> { if let Some(command) = command { let full_command = shell.replace("{}", &command); if verbose { println!("running {}", full_command); } let command_vec = shell_words::split(&full_command)?; if command_vec.len() == 1 { process::Command::new(&command_vec[0]) .stdout(process::Stdio::null()) .stderr(process::Stdio::null()) .status() .with_context(|| format!("Couldn't run hook '{}'", full_command))?; } else { process::Command::new(&command_vec[0]) .args(&command_vec[1..]) .stdout(process::Stdio::null()) .stderr(process::Stdio::null()) .status() .with_context(|| format!("Couldn't run hook '{}'", full_command))?; } } Ok(()) } /// Replace with delimiter lines /// /// In a string, removes everything from one line to another, and puts the built template in place /// /// * `file_content` - String with lines to be replaced /// * `start` - Where to start replacing /// * `end` - Where to stop replacing /// * `built_template` - Built template to be injected fn
( file_content: &str, start: &str, end: &str, built_template: &str, ) -> Result<String> { let mut changed_content = String::new(); let mut found_start = false; let mut found_end = false; let mut appended = false; for line in file_content.lines() { if found_start && !found_end { if !appended { changed_content.push_str(&built_template); appended = true; } if line.trim().to_lowercase().eq(&end) { changed_content.push_str(&format!("{}\n", line)); found_end = true; } } else { changed_content.push_str(&format!("{}\n", line)); if line.trim().to_lowercase().eq(&start) { found_start = true } } } if !found_start { Err(anyhow!("Couldn't find starting string.")) } else if !found_end { Err(anyhow!("Couldn't find ending string.")) } else { Ok(changed_content) } } /// Apply function /// /// * `patterns` - Which patterns the user specified /// * `base_dir` - Flavours base directory /// * `config_path` - Flavours configuration path /// * `light` - Don't run hooks marked as non-lightweight /// * `from_stdin` - Read scheme from stdin? /// * `verbose` - Should we be verbose? pub fn apply( patterns: Vec<&str>, base_dir: &path::Path, config_path: &path::Path, light_mode: bool, from_stdin: bool, verbose: bool, ) -> Result<()> { let (scheme_contents, scheme_slug) = if from_stdin { let mut buffer = String::new(); let stdin = io::stdin(); let mut handle = stdin.lock(); handle.read_to_string(&mut buffer)?; (buffer, String::from("generated")) } else { //Find schemes that match given patterns let mut schemes = Vec::new(); for pattern in patterns { let found_schemes = find(pattern, &base_dir.join("base16").join("schemes"))?; for found_scheme in found_schemes { schemes.push(found_scheme); } } //Sort and remove duplicates schemes.sort(); schemes.dedup(); //Get random scheme let scheme_file = random(schemes)?; let scheme_slug: String = scheme_file .file_stem() .ok_or_else(|| anyhow!("Couldn't get scheme name."))? .to_str() .ok_or_else(|| anyhow!("Couldn't convert scheme file name."))? .into(); //Read chosen scheme ( fs::read_to_string(&scheme_file) .with_context(|| format!("Couldn't read scheme file at {:?}.", scheme_file))?, scheme_slug, ) }; let scheme = Scheme::from_str(&scheme_contents, &scheme_slug)?; if verbose { println!( "Using scheme: {} ({}), by {}", scheme.name, scheme.slug, scheme.author ); println!(); } //Check if config file exists if !config_path.exists() { eprintln!("Config {:?} doesn't exist, creating", config_path); let default_content = match fs::read_to_string(path::Path::new("/etc/flavours.conf")) { Ok(content) => content, Err(_) => String::from(""), }; let config_path_parent = config_path .parent() .with_context(|| format!("Couldn't get parent directory of {:?}", config_path))?; fs::create_dir_all(config_path_parent).with_context(|| { format!( "Couldn't create configuration file parent directory {:?}", config_path_parent ) })?; fs::write(config_path, default_content) .with_context(|| format!("Couldn't create configuration file at {:?}", config_path))?; } let config_contents = fs::read_to_string(config_path) .with_context(|| format!("Couldn't read configuration file {:?}.", config_path))?; let config = Config::from_str(&config_contents)?; // If shell is present, check if it contains the placeholder let shell = config.shell.unwrap_or_else(|| "sh -c '{}'".into()); if !shell.contains("{}") { return Err(anyhow!("The configured shell does not contain the required command placeholder '{}'. Check the default file or github for config examples.")); } let mut hooks = Vec::new(); //Iterate configurated entries (templates) let items_legacy = config.item.unwrap_or_default(); let mut items = config.items.unwrap_or_default(); items.extend(items_legacy.into_iter()) ; if items.is_empty() { return Err(anyhow!("Couldn't get items from config file. Check the default file or github for config examples.")); } for item in items.iter() { //Template name let template = &item.template; //Subtemplate name let subtemplate = match &item.subtemplate { Some(value) => String::from(value), None => String::from("default"), }; //Is the hook lightweight? let light = match &item.light { Some(value) => *value, None => true, }; //Rewrite or replace let rewrite = match &item.rewrite { Some(value) => *value, None => false, }; //Replace start delimiter let start = match &item.start { Some(value) => String::from(value), None => String::from("# Start flavours"), } .trim() .to_lowercase(); //Replace end delimiter let end = match &item.end { Some(value) => String::from(value), None => String::from("# End flavours"), } .trim() .to_lowercase(); //(sub)template file path let subtemplate_file = &base_dir .join("base16") .join("templates") .join(&template) .join("templates") .join(format!("{}.mustache", subtemplate)); //Template content let template_content = fs::read_to_string(subtemplate_file) .with_context(||format!("Couldn't read template {}/{} at {:?}. Check if the correct template/subtemplate was specified, and run the update templates command if you didn't already.", template, subtemplate, subtemplate_file))?; //Template with correct colors let built_template = build_template(template_content, &scheme) .context("Couldn't replace placeholders. Check if all colors on the specified scheme file are valid (don't include a leading '#').")?; //File to write let file = shellexpand::full(&item.file)?.to_string(); //Rewrite file with built template if rewrite { std::path::Path::new(&file).parent().and_then(|p| fs::create_dir_all(p).ok()); fs::write(&file, built_template) .with_context(|| format!("Couldn't write to file {:?}.", file))?; if verbose { println!("Wrote {}/{} on: {:?}", template, subtemplate, file) } } else { //Or replace with delimiters let file_content = fs::read_to_string(&file)?; match replace_delimiter(&file_content, &start, &end, &built_template) { Ok(content) => fs::write(&file, content) .with_context(|| format!("Couldn't write to file {:?}", file))?, Err(error) =>
replace_delimiter
identifier_name
apply.rs
/// * `values` - Vec with paths fn random(values: Vec<path::PathBuf>) -> Result<path::PathBuf> { let chosen = values.choose(&mut rand::thread_rng()).ok_or_else(|| { anyhow!( "Scheme not found. Check if it exists, or run update schemes if you didn't already." ) })?; Ok(chosen.to_path_buf()) } /// Runs hook commands /// /// * `command` - Command string to execute /// * `verbose` - Should we be verbose? fn run_hook(command: Option<String>, shell: &str, verbose: bool) -> Result<()> { if let Some(command) = command { let full_command = shell.replace("{}", &command); if verbose { println!("running {}", full_command); } let command_vec = shell_words::split(&full_command)?; if command_vec.len() == 1 { process::Command::new(&command_vec[0]) .stdout(process::Stdio::null()) .stderr(process::Stdio::null()) .status() .with_context(|| format!("Couldn't run hook '{}'", full_command))?; } else { process::Command::new(&command_vec[0]) .args(&command_vec[1..]) .stdout(process::Stdio::null()) .stderr(process::Stdio::null()) .status() .with_context(|| format!("Couldn't run hook '{}'", full_command))?; } } Ok(()) } /// Replace with delimiter lines /// /// In a string, removes everything from one line to another, and puts the built template in place /// /// * `file_content` - String with lines to be replaced /// * `start` - Where to start replacing /// * `end` - Where to stop replacing /// * `built_template` - Built template to be injected fn replace_delimiter( file_content: &str, start: &str, end: &str, built_template: &str, ) -> Result<String> { let mut changed_content = String::new(); let mut found_start = false; let mut found_end = false; let mut appended = false; for line in file_content.lines() { if found_start && !found_end { if !appended { changed_content.push_str(&built_template); appended = true; } if line.trim().to_lowercase().eq(&end) { changed_content.push_str(&format!("{}\n", line)); found_end = true; } } else { changed_content.push_str(&format!("{}\n", line)); if line.trim().to_lowercase().eq(&start) { found_start = true } } } if !found_start { Err(anyhow!("Couldn't find starting string.")) } else if !found_end { Err(anyhow!("Couldn't find ending string.")) } else { Ok(changed_content) } } /// Apply function /// /// * `patterns` - Which patterns the user specified /// * `base_dir` - Flavours base directory /// * `config_path` - Flavours configuration path /// * `light` - Don't run hooks marked as non-lightweight /// * `from_stdin` - Read scheme from stdin? /// * `verbose` - Should we be verbose? pub fn apply( patterns: Vec<&str>, base_dir: &path::Path, config_path: &path::Path, light_mode: bool, from_stdin: bool, verbose: bool, ) -> Result<()> { let (scheme_contents, scheme_slug) = if from_stdin { let mut buffer = String::new(); let stdin = io::stdin(); let mut handle = stdin.lock(); handle.read_to_string(&mut buffer)?; (buffer, String::from("generated")) } else { //Find schemes that match given patterns let mut schemes = Vec::new(); for pattern in patterns { let found_schemes = find(pattern, &base_dir.join("base16").join("schemes"))?; for found_scheme in found_schemes { schemes.push(found_scheme); } } //Sort and remove duplicates schemes.sort(); schemes.dedup(); //Get random scheme let scheme_file = random(schemes)?; let scheme_slug: String = scheme_file .file_stem() .ok_or_else(|| anyhow!("Couldn't get scheme name."))? .to_str() .ok_or_else(|| anyhow!("Couldn't convert scheme file name."))? .into(); //Read chosen scheme ( fs::read_to_string(&scheme_file) .with_context(|| format!("Couldn't read scheme file at {:?}.", scheme_file))?, scheme_slug, ) }; let scheme = Scheme::from_str(&scheme_contents, &scheme_slug)?; if verbose { println!( "Using scheme: {} ({}), by {}", scheme.name, scheme.slug, scheme.author );
//Check if config file exists if !config_path.exists() { eprintln!("Config {:?} doesn't exist, creating", config_path); let default_content = match fs::read_to_string(path::Path::new("/etc/flavours.conf")) { Ok(content) => content, Err(_) => String::from(""), }; let config_path_parent = config_path .parent() .with_context(|| format!("Couldn't get parent directory of {:?}", config_path))?; fs::create_dir_all(config_path_parent).with_context(|| { format!( "Couldn't create configuration file parent directory {:?}", config_path_parent ) })?; fs::write(config_path, default_content) .with_context(|| format!("Couldn't create configuration file at {:?}", config_path))?; } let config_contents = fs::read_to_string(config_path) .with_context(|| format!("Couldn't read configuration file {:?}.", config_path))?; let config = Config::from_str(&config_contents)?; // If shell is present, check if it contains the placeholder let shell = config.shell.unwrap_or_else(|| "sh -c '{}'".into()); if !shell.contains("{}") { return Err(anyhow!("The configured shell does not contain the required command placeholder '{}'. Check the default file or github for config examples.")); } let mut hooks = Vec::new(); //Iterate configurated entries (templates) let items_legacy = config.item.unwrap_or_default(); let mut items = config.items.unwrap_or_default(); items.extend(items_legacy.into_iter()) ; if items.is_empty() { return Err(anyhow!("Couldn't get items from config file. Check the default file or github for config examples.")); } for item in items.iter() { //Template name let template = &item.template; //Subtemplate name let subtemplate = match &item.subtemplate { Some(value) => String::from(value), None => String::from("default"), }; //Is the hook lightweight? let light = match &item.light { Some(value) => *value, None => true, }; //Rewrite or replace let rewrite = match &item.rewrite { Some(value) => *value, None => false, }; //Replace start delimiter let start = match &item.start { Some(value) => String::from(value), None => String::from("# Start flavours"), } .trim() .to_lowercase(); //Replace end delimiter let end = match &item.end { Some(value) => String::from(value), None => String::from("# End flavours"), } .trim() .to_lowercase(); //(sub)template file path let subtemplate_file = &base_dir .join("base16") .join("templates") .join(&template) .join("templates") .join(format!("{}.mustache", subtemplate)); //Template content let template_content = fs::read_to_string(subtemplate_file) .with_context(||format!("Couldn't read template {}/{} at {:?}. Check if the correct template/subtemplate was specified, and run the update templates command if you didn't already.", template, subtemplate, subtemplate_file))?; //Template with correct colors let built_template = build_template(template_content, &scheme) .context("Couldn't replace placeholders. Check if all colors on the specified scheme file are valid (don't include a leading '#').")?; //File to write let file = shellexpand::full(&item.file)?.to_string(); //Rewrite file with built template if rewrite { std::path::Path::new(&file).parent().and_then(|p| fs::create_dir_all(p).ok()); fs::write(&file, built_template) .with_context(|| format!("Couldn't write to file {:?}.", file))?; if verbose { println!("Wrote {}/{} on: {:?}", template, subtemplate, file) } } else { //Or replace with delimiters let file_content = fs::read_to_string(&file)?; match replace_delimiter(&file_content, &start, &end, &built_template) { Ok(content) => fs::write(&file, content) .with_context(|| format!("Couldn't write to file {:?}", file))?, Err(error) => e
println!(); }
random_line_split
apply.rs
/// * `values` - Vec with paths fn random(values: Vec<path::PathBuf>) -> Result<path::PathBuf> { let chosen = values.choose(&mut rand::thread_rng()).ok_or_else(|| { anyhow!( "Scheme not found. Check if it exists, or run update schemes if you didn't already." ) })?; Ok(chosen.to_path_buf()) } /// Runs hook commands /// /// * `command` - Command string to execute /// * `verbose` - Should we be verbose? fn run_hook(command: Option<String>, shell: &str, verbose: bool) -> Result<()>
.with_context(|| format!("Couldn't run hook '{}'", full_command))?; } } Ok(()) } /// Replace with delimiter lines /// /// In a string, removes everything from one line to another, and puts the built template in place /// /// * `file_content` - String with lines to be replaced /// * `start` - Where to start replacing /// * `end` - Where to stop replacing /// * `built_template` - Built template to be injected fn replace_delimiter( file_content: &str, start: &str, end: &str, built_template: &str, ) -> Result<String> { let mut changed_content = String::new(); let mut found_start = false; let mut found_end = false; let mut appended = false; for line in file_content.lines() { if found_start && !found_end { if !appended { changed_content.push_str(&built_template); appended = true; } if line.trim().to_lowercase().eq(&end) { changed_content.push_str(&format!("{}\n", line)); found_end = true; } } else { changed_content.push_str(&format!("{}\n", line)); if line.trim().to_lowercase().eq(&start) { found_start = true } } } if !found_start { Err(anyhow!("Couldn't find starting string.")) } else if !found_end { Err(anyhow!("Couldn't find ending string.")) } else { Ok(changed_content) } } /// Apply function /// /// * `patterns` - Which patterns the user specified /// * `base_dir` - Flavours base directory /// * `config_path` - Flavours configuration path /// * `light` - Don't run hooks marked as non-lightweight /// * `from_stdin` - Read scheme from stdin? /// * `verbose` - Should we be verbose? pub fn apply( patterns: Vec<&str>, base_dir: &path::Path, config_path: &path::Path, light_mode: bool, from_stdin: bool, verbose: bool, ) -> Result<()> { let (scheme_contents, scheme_slug) = if from_stdin { let mut buffer = String::new(); let stdin = io::stdin(); let mut handle = stdin.lock(); handle.read_to_string(&mut buffer)?; (buffer, String::from("generated")) } else { //Find schemes that match given patterns let mut schemes = Vec::new(); for pattern in patterns { let found_schemes = find(pattern, &base_dir.join("base16").join("schemes"))?; for found_scheme in found_schemes { schemes.push(found_scheme); } } //Sort and remove duplicates schemes.sort(); schemes.dedup(); //Get random scheme let scheme_file = random(schemes)?; let scheme_slug: String = scheme_file .file_stem() .ok_or_else(|| anyhow!("Couldn't get scheme name."))? .to_str() .ok_or_else(|| anyhow!("Couldn't convert scheme file name."))? .into(); //Read chosen scheme ( fs::read_to_string(&scheme_file) .with_context(|| format!("Couldn't read scheme file at {:?}.", scheme_file))?, scheme_slug, ) }; let scheme = Scheme::from_str(&scheme_contents, &scheme_slug)?; if verbose { println!( "Using scheme: {} ({}), by {}", scheme.name, scheme.slug, scheme.author ); println!(); } //Check if config file exists if !config_path.exists() { eprintln!("Config {:?} doesn't exist, creating", config_path); let default_content = match fs::read_to_string(path::Path::new("/etc/flavours.conf")) { Ok(content) => content, Err(_) => String::from(""), }; let config_path_parent = config_path .parent() .with_context(|| format!("Couldn't get parent directory of {:?}", config_path))?; fs::create_dir_all(config_path_parent).with_context(|| { format!( "Couldn't create configuration file parent directory {:?}", config_path_parent ) })?; fs::write(config_path, default_content) .with_context(|| format!("Couldn't create configuration file at {:?}", config_path))?; } let config_contents = fs::read_to_string(config_path) .with_context(|| format!("Couldn't read configuration file {:?}.", config_path))?; let config = Config::from_str(&config_contents)?; // If shell is present, check if it contains the placeholder let shell = config.shell.unwrap_or_else(|| "sh -c '{}'".into()); if !shell.contains("{}") { return Err(anyhow!("The configured shell does not contain the required command placeholder '{}'. Check the default file or github for config examples.")); } let mut hooks = Vec::new(); //Iterate configurated entries (templates) let items_legacy = config.item.unwrap_or_default(); let mut items = config.items.unwrap_or_default(); items.extend(items_legacy.into_iter()) ; if items.is_empty() { return Err(anyhow!("Couldn't get items from config file. Check the default file or github for config examples.")); } for item in items.iter() { //Template name let template = &item.template; //Subtemplate name let subtemplate = match &item.subtemplate { Some(value) => String::from(value), None => String::from("default"), }; //Is the hook lightweight? let light = match &item.light { Some(value) => *value, None => true, }; //Rewrite or replace let rewrite = match &item.rewrite { Some(value) => *value, None => false, }; //Replace start delimiter let start = match &item.start { Some(value) => String::from(value), None => String::from("# Start flavours"), } .trim() .to_lowercase(); //Replace end delimiter let end = match &item.end { Some(value) => String::from(value), None => String::from("# End flavours"), } .trim() .to_lowercase(); //(sub)template file path let subtemplate_file = &base_dir .join("base16") .join("templates") .join(&template) .join("templates") .join(format!("{}.mustache", subtemplate)); //Template content let template_content = fs::read_to_string(subtemplate_file) .with_context(||format!("Couldn't read template {}/{} at {:?}. Check if the correct template/subtemplate was specified, and run the update templates command if you didn't already.", template, subtemplate, subtemplate_file))?; //Template with correct colors let built_template = build_template(template_content, &scheme) .context("Couldn't replace placeholders. Check if all colors on the specified scheme file are valid (don't include a leading '#').")?; //File to write let file = shellexpand::full(&item.file)?.to_string(); //Rewrite file with built template if rewrite { std::path::Path::new(&file).parent().and_then(|p| fs::create_dir_all(p).ok()); fs::write(&file, built_template) .with_context(|| format!("Couldn't write to file {:?}.", file))?; if verbose { println!("Wrote {}/{} on: {:?}", template, subtemplate, file) } } else { //Or replace with delimiters let file_content = fs::read_to_string(&file)?; match replace_delimiter(&file_content, &start, &end, &built_template) { Ok(content) => fs::write(&file, content) .with_context(|| format!("Couldn't write to file {:?}", file))?, Err(error) =>
{ if let Some(command) = command { let full_command = shell.replace("{}", &command); if verbose { println!("running {}", full_command); } let command_vec = shell_words::split(&full_command)?; if command_vec.len() == 1 { process::Command::new(&command_vec[0]) .stdout(process::Stdio::null()) .stderr(process::Stdio::null()) .status() .with_context(|| format!("Couldn't run hook '{}'", full_command))?; } else { process::Command::new(&command_vec[0]) .args(&command_vec[1..]) .stdout(process::Stdio::null()) .stderr(process::Stdio::null()) .status()
identifier_body
constrained_attack.py
as DataFrame Autoencoder output vector Returns -------- recontrution vector sorted descending and dropped """ # print(temp) temp = temp.sort_values(by=row_index, axis=1, ascending=False) i = 0 for col in temp.columns: if temp.loc[row_index, col] < theta: break i = i + 1 temp = temp.drop(columns=temp.columns[i:43]) return(temp.copy()) def scale_input_and_detect_single(index, X): """ given a row of the dataset we transform it with the scaler and we see if it is detected by the model. Parameters ---------- index : int row index X : pandas DataFrame Dataframe Containing one row of sensor readings Returns -------- bool detection outcome float average recontrution error pandas dataframe reconstruction error vector for the considered sensor readings """ X_transformed = pd.DataFrame( index=[index], columns=xset, data=scaler.transform(X)) Yhat, error, temp, _ = autoencoder.detect( X_transformed, theta=theta, window=1, average=True) return Yhat, error, temp def scale_input_and_detect(index, X): """ given a row of the dataset we transform it with the scaler and we see if it is detected by the model. Parameters ---------- index : int row index X : pandas DataFrame Dataframe Containing one row of sensor readings Returns -------- bool detection outcome float average recontrution error pandas dataframe reconstruction error vector for the considered sensor readings """ X_transformed = pd.DataFrame( columns=xset, data=scaler.transform(X), index=X.index) _, error, _, _ = autoencoder.detect( X_transformed, theta=theta, window=1, average=True) error_df = pd.DataFrame({'error': error}) X = pd.concat([X, error_df], axis=1) X = X.iloc[X['error'].idxmin()] # print(X) error = X['error'] X = X.drop('error') X = pd.DataFrame([X]) # print(X) return X, error def compute_mutation_factor(att_data, newBest): """ compute how many columns values have been changed at the end of the transformation. Parameters ---------- att_data : pandas DataFrame original sensor readings newBest : pandas DataFrame concealed sensor readings """ X2 = pd.DataFrame(index=att_data.index, columns=xset, data=att_data[xset]) frames = [X2, newBest] merge = pd.concat(frames) merge.loc['Diff'] = merge.iloc[0] - merge.iloc[1] changed_columns[row_index] = merge.loc['Diff'].astype(bool).sum() print('changed tuples: ' + str(len(changed_columns))) def
(row_index, prev_col_name, changed_variables, max_concealable_variables): """ select the sensor value to be manipulated depending on the constrints Parameters ---------- row_index : int prev_col_name : string changed_variables : list variables that can be manipulated max_concealable_variables : int number of variables that can manipulated Returns ---------- string the column that will be manipulated in the current iteration """ if(prev_col_name == None): return changed_variables[row_index][0] return changed_variables[row_index][(changed_variables[row_index].index(prev_col_name)+1) % max_concealable_variables] # this is the main algorithm, it actually transforms the input row trying to change its label. # second attempt, updates after 5 changes on the same variable the ranking and optimizes this variable def change_vector_label(row_index, att_data, solutions_found, changed_variables, variables): """ this is the main algorithm, it actually transforms the input row trying to change its predicted label. updates after 5 changes on the same variable the ranking and optimizes the new ranked 1 variable Parameters ---------- row_index : int att_data : pandas DataFrame original data to be concealed solutions_found : int counter of found solution Returns -------- pandas dataframe solution found int updated counter of solutions """ original_vector = att_data.copy() changes = 0 found_solution = 0 _, error, temp = scale_input_and_detect_single(row_index, att_data) previous_best_error = error[row_index] temp = sort_temp_and_drop(row_index, temp) prev_col_name = None num_changes_without_optimizations = 0 last_optimization = 0 newBest = att_data.copy() optimized = False changed_variables[row_index] = variables[max_concealable_variables] while changes < budget and (changes - last_optimization) < patience and not(found_solution): col_name = choose_column(row_index, temp, prev_col_name, num_changes_without_optimizations, changed_variables, max_concealable_variables) prev_col_name = col_name if debug: print('______________________________') print(col_name) print('______________________________') values = np.arange( normal_op_ranges[col_name]['min'], normal_op_ranges[col_name]['max']+0.1, normal_op_ranges[col_name]['step']) # print(values) att_data = att_data.append( [att_data] * (len(values)), ignore_index=True) att_data = att_data[:-1] # delete eccessive lenght # substitute column values usign normal operations att_data[col_name] = values att_data, error = scale_input_and_detect(row_index, att_data) if error < previous_best_error: if debug: print(error, previous_best_error) previous_best_error = error newBest = att_data.copy() last_optimization = changes num_changes_without_optimizations = 0 optimized = True try: if not(col_name) in changed_variables[row_index]: changed_variables[row_index].append(col_name) except: changed_variables[row_index] = [col_name] else: optimized = False if error < theta: solutions_found = solutions_found + 1 found_solution = 1 print('Found solution number: ' + str(solutions_found)) if optimized == False: num_changes_without_optimizations = num_changes_without_optimizations + 1 att_data = newBest.copy() _, error, temp = scale_input_and_detect_single( row_index, att_data) temp = sort_temp_and_drop(row_index, temp) changes = changes + 1 if debug: print(temp) print('--__--__--') print(changes) print('--__--__--') compute_mutation_factor(original_vector, att_data.copy()) return newBest.copy(), solutions_found """ Select wich dataset are you considering (we are not allowed to publish WADI data, please request them itrust Singapore website) """ dataset = 'BATADAL' #'WADI' data_folder = '../../Data/'+dataset if dataset == 'BATADAL': attack_ids = range(1,15) att_data = pd.read_csv(data_folder+'/attack_1_from_test_dataset.csv') xset = [col for col in att_data.columns if col not in [ 'Unnamed: 0', 'DATETIME', 'ATT_FLAG']] budget = 200 patience = 15 if dataset == 'WADI': attack_ids = [1, 2, 3, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15] att_data = pd.read_csv(data_folder+'/attack_1_from_test_dataset.csv') xset = [col for col in att_data.columns if col not in [ 'Row', 'DATETIME','ATT_FLAG', '2_MV_001_STATUS', '2_LT_001_PV', '2_MV_002_STATUS']] budget = 300 patience = 40 yset = ['ATT_FLAG'] autoencoder = load_AEED("../../Attacked_Model/"+dataset+"/autoencoder.json", "../../Attacked_Model/"+dataset+"/autoencoder.h5") scaler = pickle.load(open("../../Attacked_Model/"+dataset+"/scaler.p", "rb")) with open("../../Attacked_Model/"+dataset+"/theta") as f: theta = float(f.read()) normal_op_ranges = pickle.load(open('dict_'+dataset+'.p', 'rb')) for att_number in attack_ids: variables = {} f = open("./constraints/"+dataset+"/constraint_variables_attack_"+str(att_number)+".txt", 'r').read() variables = eval(f) for max_concealable_variables in [2, 3, 4, 5, 6, 7, 8, 9, 10, 15, 20]: debug = False changed_columns = {} changed_variables = {} print('ATT NUMBER: '+str(att_number)) att_data = pd.read_csv( data_folder+'/attack_'+str(att_number)+'_from_test
choose_column
identifier_name
constrained_attack.py
as DataFrame Autoencoder output vector Returns -------- recontrution vector sorted descending and dropped """ # print(temp) temp = temp.sort_values(by=row_index, axis=1, ascending=False) i = 0 for col in temp.columns: if temp.loc[row_index, col] < theta: break i = i + 1 temp = temp.drop(columns=temp.columns[i:43]) return(temp.copy()) def scale_input_and_detect_single(index, X): """ given a row of the dataset we transform it with the scaler and we see if it is detected by the model. Parameters ---------- index : int row index X : pandas DataFrame Dataframe Containing one row of sensor readings Returns -------- bool detection outcome float average recontrution error pandas dataframe reconstruction error vector for the considered sensor readings """ X_transformed = pd.DataFrame( index=[index], columns=xset, data=scaler.transform(X)) Yhat, error, temp, _ = autoencoder.detect( X_transformed, theta=theta, window=1, average=True) return Yhat, error, temp def scale_input_and_detect(index, X): """ given a row of the dataset we transform it with the scaler and we see if it is detected by the model. Parameters ---------- index : int row index X : pandas DataFrame Dataframe Containing one row of sensor readings Returns -------- bool detection outcome float average recontrution error pandas dataframe reconstruction error vector for the considered sensor readings """ X_transformed = pd.DataFrame( columns=xset, data=scaler.transform(X), index=X.index) _, error, _, _ = autoencoder.detect( X_transformed, theta=theta, window=1, average=True) error_df = pd.DataFrame({'error': error}) X = pd.concat([X, error_df], axis=1) X = X.iloc[X['error'].idxmin()] # print(X) error = X['error'] X = X.drop('error') X = pd.DataFrame([X]) # print(X) return X, error def compute_mutation_factor(att_data, newBest): """ compute how many columns values have been changed at the end of the transformation. Parameters ---------- att_data : pandas DataFrame original sensor readings newBest : pandas DataFrame concealed sensor readings """ X2 = pd.DataFrame(index=att_data.index, columns=xset, data=att_data[xset]) frames = [X2, newBest] merge = pd.concat(frames) merge.loc['Diff'] = merge.iloc[0] - merge.iloc[1] changed_columns[row_index] = merge.loc['Diff'].astype(bool).sum() print('changed tuples: ' + str(len(changed_columns))) def choose_column(row_index, prev_col_name, changed_variables, max_concealable_variables): """ select the sensor value to be manipulated depending on the constrints Parameters ---------- row_index : int prev_col_name : string changed_variables : list variables that can be manipulated max_concealable_variables : int number of variables that can manipulated Returns ---------- string the column that will be manipulated in the current iteration """ if(prev_col_name == None): return changed_variables[row_index][0] return changed_variables[row_index][(changed_variables[row_index].index(prev_col_name)+1) % max_concealable_variables] # this is the main algorithm, it actually transforms the input row trying to change its label. # second attempt, updates after 5 changes on the same variable the ranking and optimizes this variable def change_vector_label(row_index, att_data, solutions_found, changed_variables, variables): """ this is the main algorithm, it actually transforms the input row trying to change its predicted label. updates after 5 changes on the same variable the ranking and optimizes the new ranked 1 variable Parameters ---------- row_index : int att_data : pandas DataFrame original data to be concealed solutions_found : int counter of found solution Returns -------- pandas dataframe solution found int updated counter of solutions """ original_vector = att_data.copy() changes = 0 found_solution = 0 _, error, temp = scale_input_and_detect_single(row_index, att_data) previous_best_error = error[row_index] temp = sort_temp_and_drop(row_index, temp) prev_col_name = None num_changes_without_optimizations = 0 last_optimization = 0 newBest = att_data.copy() optimized = False changed_variables[row_index] = variables[max_concealable_variables] while changes < budget and (changes - last_optimization) < patience and not(found_solution): col_name = choose_column(row_index, temp, prev_col_name, num_changes_without_optimizations, changed_variables, max_concealable_variables) prev_col_name = col_name if debug: print('______________________________') print(col_name) print('______________________________') values = np.arange( normal_op_ranges[col_name]['min'], normal_op_ranges[col_name]['max']+0.1, normal_op_ranges[col_name]['step']) # print(values) att_data = att_data.append( [att_data] * (len(values)), ignore_index=True) att_data = att_data[:-1] # delete eccessive lenght # substitute column values usign normal operations att_data[col_name] = values att_data, error = scale_input_and_detect(row_index, att_data) if error < previous_best_error: if debug:
previous_best_error = error newBest = att_data.copy() last_optimization = changes num_changes_without_optimizations = 0 optimized = True try: if not(col_name) in changed_variables[row_index]: changed_variables[row_index].append(col_name) except: changed_variables[row_index] = [col_name] else: optimized = False if error < theta: solutions_found = solutions_found + 1 found_solution = 1 print('Found solution number: ' + str(solutions_found)) if optimized == False: num_changes_without_optimizations = num_changes_without_optimizations + 1 att_data = newBest.copy() _, error, temp = scale_input_and_detect_single( row_index, att_data) temp = sort_temp_and_drop(row_index, temp) changes = changes + 1 if debug: print(temp) print('--__--__--') print(changes) print('--__--__--') compute_mutation_factor(original_vector, att_data.copy()) return newBest.copy(), solutions_found """ Select wich dataset are you considering (we are not allowed to publish WADI data, please request them itrust Singapore website) """ dataset = 'BATADAL' #'WADI' data_folder = '../../Data/'+dataset if dataset == 'BATADAL': attack_ids = range(1,15) att_data = pd.read_csv(data_folder+'/attack_1_from_test_dataset.csv') xset = [col for col in att_data.columns if col not in [ 'Unnamed: 0', 'DATETIME', 'ATT_FLAG']] budget = 200 patience = 15 if dataset == 'WADI': attack_ids = [1, 2, 3, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15] att_data = pd.read_csv(data_folder+'/attack_1_from_test_dataset.csv') xset = [col for col in att_data.columns if col not in [ 'Row', 'DATETIME','ATT_FLAG', '2_MV_001_STATUS', '2_LT_001_PV', '2_MV_002_STATUS']] budget = 300 patience = 40 yset = ['ATT_FLAG'] autoencoder = load_AEED("../../Attacked_Model/"+dataset+"/autoencoder.json", "../../Attacked_Model/"+dataset+"/autoencoder.h5") scaler = pickle.load(open("../../Attacked_Model/"+dataset+"/scaler.p", "rb")) with open("../../Attacked_Model/"+dataset+"/theta") as f: theta = float(f.read()) normal_op_ranges = pickle.load(open('dict_'+dataset+'.p', 'rb')) for att_number in attack_ids: variables = {} f = open("./constraints/"+dataset+"/constraint_variables_attack_"+str(att_number)+".txt", 'r').read() variables = eval(f) for max_concealable_variables in [2, 3, 4, 5, 6, 7, 8, 9, 10, 15, 20]: debug = False changed_columns = {} changed_variables = {} print('ATT NUMBER: '+str(att_number)) att_data = pd.read_csv( data_folder+'/attack_'+str(att_number)+'_from_test
print(error, previous_best_error)
conditional_block
constrained_attack.py
X : pandas DataFrame Dataframe Containing one row of sensor readings Returns -------- bool detection outcome float average recontrution error pandas dataframe reconstruction error vector for the considered sensor readings """ X_transformed = pd.DataFrame( index=[index], columns=xset, data=scaler.transform(X)) Yhat, error, temp, _ = autoencoder.detect( X_transformed, theta=theta, window=1, average=True) return Yhat, error, temp def scale_input_and_detect(index, X): """ given a row of the dataset we transform it with the scaler and we see if it is detected by the model. Parameters ---------- index : int row index X : pandas DataFrame Dataframe Containing one row of sensor readings Returns -------- bool detection outcome float average recontrution error pandas dataframe reconstruction error vector for the considered sensor readings """ X_transformed = pd.DataFrame( columns=xset, data=scaler.transform(X), index=X.index) _, error, _, _ = autoencoder.detect( X_transformed, theta=theta, window=1, average=True) error_df = pd.DataFrame({'error': error}) X = pd.concat([X, error_df], axis=1) X = X.iloc[X['error'].idxmin()] # print(X) error = X['error'] X = X.drop('error') X = pd.DataFrame([X]) # print(X) return X, error def compute_mutation_factor(att_data, newBest): """ compute how many columns values have been changed at the end of the transformation. Parameters ---------- att_data : pandas DataFrame original sensor readings newBest : pandas DataFrame concealed sensor readings """ X2 = pd.DataFrame(index=att_data.index, columns=xset, data=att_data[xset]) frames = [X2, newBest] merge = pd.concat(frames) merge.loc['Diff'] = merge.iloc[0] - merge.iloc[1] changed_columns[row_index] = merge.loc['Diff'].astype(bool).sum() print('changed tuples: ' + str(len(changed_columns))) def choose_column(row_index, prev_col_name, changed_variables, max_concealable_variables): """ select the sensor value to be manipulated depending on the constrints Parameters ---------- row_index : int prev_col_name : string changed_variables : list variables that can be manipulated max_concealable_variables : int number of variables that can manipulated Returns ---------- string the column that will be manipulated in the current iteration """ if(prev_col_name == None): return changed_variables[row_index][0] return changed_variables[row_index][(changed_variables[row_index].index(prev_col_name)+1) % max_concealable_variables] # this is the main algorithm, it actually transforms the input row trying to change its label. # second attempt, updates after 5 changes on the same variable the ranking and optimizes this variable def change_vector_label(row_index, att_data, solutions_found, changed_variables, variables): """ this is the main algorithm, it actually transforms the input row trying to change its predicted label. updates after 5 changes on the same variable the ranking and optimizes the new ranked 1 variable Parameters ---------- row_index : int att_data : pandas DataFrame original data to be concealed solutions_found : int counter of found solution Returns -------- pandas dataframe solution found int updated counter of solutions """ original_vector = att_data.copy() changes = 0 found_solution = 0 _, error, temp = scale_input_and_detect_single(row_index, att_data) previous_best_error = error[row_index] temp = sort_temp_and_drop(row_index, temp) prev_col_name = None num_changes_without_optimizations = 0 last_optimization = 0 newBest = att_data.copy() optimized = False changed_variables[row_index] = variables[max_concealable_variables] while changes < budget and (changes - last_optimization) < patience and not(found_solution): col_name = choose_column(row_index, temp, prev_col_name, num_changes_without_optimizations, changed_variables, max_concealable_variables) prev_col_name = col_name if debug: print('______________________________') print(col_name) print('______________________________') values = np.arange( normal_op_ranges[col_name]['min'], normal_op_ranges[col_name]['max']+0.1, normal_op_ranges[col_name]['step']) # print(values) att_data = att_data.append( [att_data] * (len(values)), ignore_index=True) att_data = att_data[:-1] # delete eccessive lenght # substitute column values usign normal operations att_data[col_name] = values att_data, error = scale_input_and_detect(row_index, att_data) if error < previous_best_error: if debug: print(error, previous_best_error) previous_best_error = error newBest = att_data.copy() last_optimization = changes num_changes_without_optimizations = 0 optimized = True try: if not(col_name) in changed_variables[row_index]: changed_variables[row_index].append(col_name) except: changed_variables[row_index] = [col_name] else: optimized = False if error < theta: solutions_found = solutions_found + 1 found_solution = 1 print('Found solution number: ' + str(solutions_found)) if optimized == False: num_changes_without_optimizations = num_changes_without_optimizations + 1 att_data = newBest.copy() _, error, temp = scale_input_and_detect_single( row_index, att_data) temp = sort_temp_and_drop(row_index, temp) changes = changes + 1 if debug: print(temp) print('--__--__--') print(changes) print('--__--__--') compute_mutation_factor(original_vector, att_data.copy()) return newBest.copy(), solutions_found """ Select wich dataset are you considering (we are not allowed to publish WADI data, please request them itrust Singapore website) """ dataset = 'BATADAL' #'WADI' data_folder = '../../Data/'+dataset if dataset == 'BATADAL': attack_ids = range(1,15) att_data = pd.read_csv(data_folder+'/attack_1_from_test_dataset.csv') xset = [col for col in att_data.columns if col not in [ 'Unnamed: 0', 'DATETIME', 'ATT_FLAG']] budget = 200 patience = 15 if dataset == 'WADI': attack_ids = [1, 2, 3, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15] att_data = pd.read_csv(data_folder+'/attack_1_from_test_dataset.csv') xset = [col for col in att_data.columns if col not in [ 'Row', 'DATETIME','ATT_FLAG', '2_MV_001_STATUS', '2_LT_001_PV', '2_MV_002_STATUS']] budget = 300 patience = 40 yset = ['ATT_FLAG'] autoencoder = load_AEED("../../Attacked_Model/"+dataset+"/autoencoder.json", "../../Attacked_Model/"+dataset+"/autoencoder.h5") scaler = pickle.load(open("../../Attacked_Model/"+dataset+"/scaler.p", "rb")) with open("../../Attacked_Model/"+dataset+"/theta") as f: theta = float(f.read()) normal_op_ranges = pickle.load(open('dict_'+dataset+'.p', 'rb')) for att_number in attack_ids: variables = {} f = open("./constraints/"+dataset+"/constraint_variables_attack_"+str(att_number)+".txt", 'r').read() variables = eval(f) for max_concealable_variables in [2, 3, 4, 5, 6, 7, 8, 9, 10, 15, 20]: debug = False changed_columns = {} changed_variables = {} print('ATT NUMBER: '+str(att_number)) att_data = pd.read_csv( data_folder+'/attack_'+str(att_number)+'_from_test_dataset.csv') y_att = att_data[yset] X = pd.DataFrame(index=att_data.index, columns=xset, data=att_data[xset]) new_tuples = pd.DataFrame(columns=xset) # main loop that iterates over every row in the dataset changed_rows = 0 solutions_found = 0 max_spent_time = 0 sum_spent = 0 times = [] import time for row_index, row in X.iterrows():
prov = pd.DataFrame(index=[row_index], columns=xset, data=att_data[xset]) Yhat, original_error, temp = scale_input_and_detect_single( row_index, prov)
random_line_split
constrained_attack.py
as DataFrame Autoencoder output vector Returns -------- recontrution vector sorted descending and dropped """ # print(temp) temp = temp.sort_values(by=row_index, axis=1, ascending=False) i = 0 for col in temp.columns: if temp.loc[row_index, col] < theta: break i = i + 1 temp = temp.drop(columns=temp.columns[i:43]) return(temp.copy()) def scale_input_and_detect_single(index, X): """ given a row of the dataset we transform it with the scaler and we see if it is detected by the model. Parameters ---------- index : int row index X : pandas DataFrame Dataframe Containing one row of sensor readings Returns -------- bool detection outcome float average recontrution error pandas dataframe reconstruction error vector for the considered sensor readings """ X_transformed = pd.DataFrame( index=[index], columns=xset, data=scaler.transform(X)) Yhat, error, temp, _ = autoencoder.detect( X_transformed, theta=theta, window=1, average=True) return Yhat, error, temp def scale_input_and_detect(index, X): """ given a row of the dataset we transform it with the scaler and we see if it is detected by the model. Parameters ---------- index : int row index X : pandas DataFrame Dataframe Containing one row of sensor readings Returns -------- bool detection outcome float average recontrution error pandas dataframe reconstruction error vector for the considered sensor readings """ X_transformed = pd.DataFrame( columns=xset, data=scaler.transform(X), index=X.index) _, error, _, _ = autoencoder.detect( X_transformed, theta=theta, window=1, average=True) error_df = pd.DataFrame({'error': error}) X = pd.concat([X, error_df], axis=1) X = X.iloc[X['error'].idxmin()] # print(X) error = X['error'] X = X.drop('error') X = pd.DataFrame([X]) # print(X) return X, error def compute_mutation_factor(att_data, newBest): """ compute how many columns values have been changed at the end of the transformation. Parameters ---------- att_data : pandas DataFrame original sensor readings newBest : pandas DataFrame concealed sensor readings """ X2 = pd.DataFrame(index=att_data.index, columns=xset, data=att_data[xset]) frames = [X2, newBest] merge = pd.concat(frames) merge.loc['Diff'] = merge.iloc[0] - merge.iloc[1] changed_columns[row_index] = merge.loc['Diff'].astype(bool).sum() print('changed tuples: ' + str(len(changed_columns))) def choose_column(row_index, prev_col_name, changed_variables, max_concealable_variables): """ select the sensor value to be manipulated depending on the constrints Parameters ---------- row_index : int prev_col_name : string changed_variables : list variables that can be manipulated max_concealable_variables : int number of variables that can manipulated Returns ---------- string the column that will be manipulated in the current iteration """ if(prev_col_name == None): return changed_variables[row_index][0] return changed_variables[row_index][(changed_variables[row_index].index(prev_col_name)+1) % max_concealable_variables] # this is the main algorithm, it actually transforms the input row trying to change its label. # second attempt, updates after 5 changes on the same variable the ranking and optimizes this variable def change_vector_label(row_index, att_data, solutions_found, changed_variables, variables):
original_vector = att_data.copy() changes = 0 found_solution = 0 _, error, temp = scale_input_and_detect_single(row_index, att_data) previous_best_error = error[row_index] temp = sort_temp_and_drop(row_index, temp) prev_col_name = None num_changes_without_optimizations = 0 last_optimization = 0 newBest = att_data.copy() optimized = False changed_variables[row_index] = variables[max_concealable_variables] while changes < budget and (changes - last_optimization) < patience and not(found_solution): col_name = choose_column(row_index, temp, prev_col_name, num_changes_without_optimizations, changed_variables, max_concealable_variables) prev_col_name = col_name if debug: print('______________________________') print(col_name) print('______________________________') values = np.arange( normal_op_ranges[col_name]['min'], normal_op_ranges[col_name]['max']+0.1, normal_op_ranges[col_name]['step']) # print(values) att_data = att_data.append( [att_data] * (len(values)), ignore_index=True) att_data = att_data[:-1] # delete eccessive lenght # substitute column values usign normal operations att_data[col_name] = values att_data, error = scale_input_and_detect(row_index, att_data) if error < previous_best_error: if debug: print(error, previous_best_error) previous_best_error = error newBest = att_data.copy() last_optimization = changes num_changes_without_optimizations = 0 optimized = True try: if not(col_name) in changed_variables[row_index]: changed_variables[row_index].append(col_name) except: changed_variables[row_index] = [col_name] else: optimized = False if error < theta: solutions_found = solutions_found + 1 found_solution = 1 print('Found solution number: ' + str(solutions_found)) if optimized == False: num_changes_without_optimizations = num_changes_without_optimizations + 1 att_data = newBest.copy() _, error, temp = scale_input_and_detect_single( row_index, att_data) temp = sort_temp_and_drop(row_index, temp) changes = changes + 1 if debug: print(temp) print('--__--__--') print(changes) print('--__--__--') compute_mutation_factor(original_vector, att_data.copy()) return newBest.copy(), solutions_found """ Select wich dataset are you considering (we are not allowed to publish WADI data, please request them itrust Singapore website) """ dataset = 'BATADAL' #'WADI' data_folder = '../../Data/'+dataset if dataset == 'BATADAL': attack_ids = range(1,15) att_data = pd.read_csv(data_folder+'/attack_1_from_test_dataset.csv') xset = [col for col in att_data.columns if col not in [ 'Unnamed: 0', 'DATETIME', 'ATT_FLAG']] budget = 200 patience = 15 if dataset == 'WADI': attack_ids = [1, 2, 3, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15] att_data = pd.read_csv(data_folder+'/attack_1_from_test_dataset.csv') xset = [col for col in att_data.columns if col not in [ 'Row', 'DATETIME','ATT_FLAG', '2_MV_001_STATUS', '2_LT_001_PV', '2_MV_002_STATUS']] budget = 300 patience = 40 yset = ['ATT_FLAG'] autoencoder = load_AEED("../../Attacked_Model/"+dataset+"/autoencoder.json", "../../Attacked_Model/"+dataset+"/autoencoder.h5") scaler = pickle.load(open("../../Attacked_Model/"+dataset+"/scaler.p", "rb")) with open("../../Attacked_Model/"+dataset+"/theta") as f: theta = float(f.read()) normal_op_ranges = pickle.load(open('dict_'+dataset+'.p', 'rb')) for att_number in attack_ids: variables = {} f = open("./constraints/"+dataset+"/constraint_variables_attack_"+str(att_number)+".txt", 'r').read() variables = eval(f) for max_concealable_variables in [2, 3, 4, 5, 6, 7, 8, 9, 10, 15, 20]: debug = False changed_columns = {} changed_variables = {} print('ATT NUMBER: '+str(att_number)) att_data = pd.read_csv( data_folder+'/attack_'+str(att_number)+'_from_test
""" this is the main algorithm, it actually transforms the input row trying to change its predicted label. updates after 5 changes on the same variable the ranking and optimizes the new ranked 1 variable Parameters ---------- row_index : int att_data : pandas DataFrame original data to be concealed solutions_found : int counter of found solution Returns -------- pandas dataframe solution found int updated counter of solutions """
identifier_body
watchman-diag
(buf) == 0: break blob.append(buf) os.close(fd) if len(blob) == 0: return None return "".join(blob) def parsefdinfo(self, blob): watches = 0 for line in blob.split("\n"): if line.find("inotify wd") != -1: watches = watches + 1 return watches def get_watches(self): watches = [("PID", "EXE", "FD", "WATCHES")] for fddir in glob.glob("/proc/*/fd"): for fdnode in glob.glob(fddir + "/*"): try: l = os.readlink(fdnode) if l != "anon_inode:inotify": continue _, _, pid, _, fdnum = fdnode.split("/") info = self.read_fdinfo("/proc/%s/fdinfo/%s" % (pid, fdnum)) if info is None: watches.append( (pid, self.procname(pid), fdnum, "<unknown> (see t8692428)") ) continue watches.append( (pid, self.procname(pid), fdnum, str(self.parsefdinfo(info))) ) except Exception as e: pass return watches def walk_root(root, case_sensitive, ignores): """Generate a map of file nodes for the given dir by looking at the filesystem""" # we can't use os.walk because it insists on stating and derefing # dewey and gvfs symlinks (== slow) results = {} # the queue of dirs to analyze dirs = [root] while len(dirs) > 0: dir = dirs.pop() for ent in os.listdir(dir): full = os.path.join(dir, ent) rel = os.path.relpath(full, root) if rel in ignores: continue st = os.lstat(full) if stat.S_ISDIR(st.st_mode): # add this child to our dir queue dirs.append(full) item = (rel, st) if not case_sensitive: rel = rel.lower() results[rel] = item return results def collect_watch_info(watchman, watch): root_config = watchman.query("get-config", watch)["config"] watchmanconfig_file = os.path.join(watch, ".watchmanconfig") file_config = {} if os.path.exists(watchmanconfig_file): with open(watchmanconfig_file) as f: file_config = json.load(f) if file_config != root_config: print("Watchman root %s is using this configuration: %s" % (watch, root_config)) print("%s has this configuration: %s" % (watchmanconfig_file, file_config)) print_status( ( "** You should run: `watchman watch-del %s ; " + "watchman watch %s` to reload .watchmanconfig **\n" ) % (watch, watch) ) if not is_eden(watch): # Eden mounts don't use the sparse extension, so skip this bit print("\nSparse configuration for %s:" % watch) passthru("cd %s && hg sparse" % watch, shell=True) # Eden watcher is stateless and doesn't have this print("\nContent hash cache stats for %s:" % watch) passthru(["watchman", "debug-contenthash", watch]) print("\nSymlink target cache stats for %s:" % watch) passthru(["watchman", "debug-symlink-target-cache", watch]) print("\nSubscriptions for %s:" % watch) passthru(["watchman", "debug-get-subscriptions", watch]) print("\nAsserted states for %s:" % watch) passthru(["watchman", "debug-get-asserted-states", watch]) def is_eden(dirpath): if sys.platform == "win32": return os.path.isfile(os.path.join(dirpath, ".eden", "config")) return os.path.islink(os.path.join(dirpath, ".eden", "root")) def cross_check_watch(watchman, watch, case_sensitive): if is_eden(watch): # We don't keep any state in watchman for eden mounts # that is worth testing against an O(repo) crawl print_status( "\nSkipping filesystem sanity check for %s as it is an eden mount\n" % watch ) return print_status( "\nSanity checking the filesystem at %s against watchman; this may take a couple of minutes." % watch ) root_config = watchman.query("get-config", watch)["config"] ignores = [] if "ignore_dirs" in root_config: ignores = root_config["ignore_dirs"] ignores.append(".hg") ignores.append(".git") ignores.append(".svn") print_status("Crawling %s..." % watch) start = time.time() fs = walk_root(watch, case_sensitive, ignores) print_status("(took %ds)" % (time.time() - start)) start = time.time() print_status("Interrogating watchman about %s..." % watch) fields = ["name", "mode", "size", "mtime_f", "oclock"] if os.name == "posix": fields.append("ino") files = watchman.query( "query", watch, { "expression": [ "allof", [ "not", [ "anyof", ["dirname", ".git"], ["dirname", ".hg"], ["dirname", ".svn"], ["name", ".git", "wholename"], ["name", ".svn", "wholename"], ["name", ".hg", "wholename"], ], ], "exists", ], "fields": fields, }, ) print_status("(took %ds)" % (time.time() - start)) print_status("Comparing results...") phantoms = [] bad_deletes = [] mismatched = [] missing = [] all_names_in_watchman = set() def diff_item(w_item, fs_item): diffs = [] if w_item["name"] != fs_item[0]: diffs.append( "watchman name is `%s` vs fs `%s" % (w_item["name"], fs_item[0]) ) st = fs_item[1] if w_item["mode"] != st.st_mode: diffs.append( "watchman mode is 0%o vs fs 0%o" % (w_item["mode"], st.st_mode) ) if w_item["size"] != st.st_size and not stat.S_ISDIR(st.st_mode): diffs.append("watchman size is %d vs fs %d" % (w_item["size"], st.st_size)) if w_item["mtime_f"] != st.st_mtime: diffs.append( "watchman mtime is %s vs fs %s" % (w_item["mtime_f"], st.st_mtime) ) if os.name == "posix" and (w_item["ino"] != st.st_ino): diffs.append("watchman ino is %d vs fs %d" % (w_item["ino"], st.st_ino)) if len(diffs) > 0: diffs.append(" oclock is %s" % w_item["oclock"]) return diffs return None for f in files["files"]: key = f["name"] if not case_sensitive: key = key.lower() all_names_in_watchman.add(key) if key not in fs: phantoms.append(f) else: diff = diff_item(f, fs[key]) if diff: print("Conflicting information for %s:" % f["name"]) for d in diff: print(d) for key in fs: if not key in all_names_in_watchman: missing.append(fs[key]) print_status( "There are %d items reported by watchman that do not exist on the fs:" % len(phantoms) ) if len(phantoms) > 0: for item in phantoms: print(item) print_status( "There are %d items on the filesystem not reported by watchman:" % len(missing) ) if len(missing) > 0: # Let's see if watchman had previously seen any of these names = ["anyof"] for item in missing: name = item[0] print(name, item[1]) names.append(["name", name, "wholename"]) files = watchman.query("query", watch, {"expression": names, "fields": fields}) print("This is what watchman knows about this set of files:") for f in files["files"]: print(f) def passthru(*args, **kwargs): sys.stdout.flush() try: subprocess.call(*args, **kwargs) except Exception as e: print("Error while running %s %s: %s" % (args, kwargs, e)) if not os.isatty(sys.stdout.fileno()): sys.stderr.write( "(most output is going to your pipe, some short summary will show here on stderr)\n"
) #
random_line_split
SurfstoreClientUtils.go
) if err == nil { *localFileMeta = remoteFileMeta } } } else { err := downloadFile(client, nil, &remoteFileMeta) if err == nil { localFileMeta := remoteFileMeta fileMetaMap[remoteFilename] = &localFileMeta } } } // working on files only on local -> upload for localFilename, localFileMeta := range fileMetaMap { if _, ok := remoteFileMetaMap[localFilename]; !ok { isUploadFailed = isUploadFailed || !uploadFile(client, localFileMeta) } } if !isUploadFailed { break } } // ==================================Finally, Write into a index file============================= writeIndexFile(client, fileMetaMap) } func uploadFile(client RPCClient, fileMeta *FileMetaData) bool { // divide into blocks filename := fileMeta.Filename if fileMeta.IsTombstone() { var latestVersion int err := client.UpdateFile(fileMeta, &latestVersion) if err != nil { return false } return fileMeta.Version == latestVersion } file, err := os.Open(filepath.Join(client.BaseDir, filename)) if err != nil { log.Println("uploadFile: Failed to open file", filename, err) return false } blockSize := client.BlockSize fileInfo, _ := file.Stat() fileSize := fileInfo.Size() numBlocks := fileSize / int64(blockSize) if fileSize%int64(blockSize) != 0 { numBlocks++ } if numBlocks == 0 { // for empty file blockBuffer := make([]byte, 0) block := Block{BlockData: blockBuffer, BlockSize: 0} succ := false err := client.PutBlock(block, &succ) if !succ || err != nil { log.Println("uploadFile: Failed to put empty block to the server") return false } } else { for i := int64(0); i < numBlocks; i++ { currentBlockOffset := i * int64(blockSize) var currentBlockSize int if blockSize < int(fileSize-currentBlockOffset) { currentBlockSize = blockSize } else { currentBlockSize = int(fileSize - currentBlockOffset) } block := NewBlock(currentBlockSize) _, err := file.Read(block.BlockData) if err != nil { return false } // write block to server // if there is error -> get block fail -> put block // if the error is nil -> get block succ -> no need succ := false err = client.HasBlock(block.Hash(), &succ) if err != nil { return false } if !succ { succ := false err := client.PutBlock(block, &succ) if !succ || err != nil { log.Println("uploadFile: Failed to put block to server") return false } } } } latestVersion := -1 err = client.UpdateFile(fileMeta, &latestVersion) if err != nil { return false } return fileMeta.Version == latestVersion } func readIndexFile(client RPCClient) map[string]*FileMetaData { // For read access. indexFilename := filepath.Join(client.BaseDir, "index.txt") indexFile, err := os.Open(indexFilename) if err != nil { // index.txt does not exit indexFile, err = os.Create(indexFilename) if err != nil { panic(err) } } defer indexFile.Close() fileMetaMap := make(map[string]*FileMetaData) // read index file reader := bufio.NewReader(indexFile) isReaderEnded := false for !isReaderEnded { line, err := reader.ReadString('\n') isReaderEnded = err == io.EOF if err != nil && err != io.EOF { panic(err) } if line == "" { break } text := strings.TrimSuffix(line, "\n") lineParts := strings.Split(text, ",") if len(lineParts) == 3 { filename := lineParts[0] version, _ := strconv.Atoi(lineParts[1]) blockHasheListString := lineParts[2] blockHasheList := strings.Split(blockHasheListString, " ") fileMeta := FileMetaData{ Filename: filename, Version: version, BlockHashList: blockHasheList, } fileMetaMap[filename] = &fileMeta } else { panic("Invalid index.txt") } } return fileMetaMap } func updateFileMetaMapWithLocalFiles(client RPCClient, fileMetaMap map[string]*FileMetaData) map[string]*FileMetaData { localFileMap := getLocalFileHashBlockListMap(client) // iterate over the file meta map and see if old file exists for filename, fileMeta := range fileMetaMap { if localBlockHashList, ok := localFileMap[filename]; ok { // find the existing file if len(localBlockHashList) != len(fileMeta.BlockHashList) { fileMeta.BlockHashList = localBlockHashList fileMeta.Version++ } else { isFileUpdated := false for i, blockHash := range localBlockHashList { if blockHash != fileMeta.BlockHashList[i] { fileMeta.BlockHashList[i] = blockHash isFileUpdated = true } } if isFileUpdated { fileMeta.Version++ } } } else { // file does not exist in dir, shoud be deleted // if file is not mark as deleted in file meta, update it if !fileMeta.IsTombstone() { fileMeta.MarkTombstone() fileMeta.Version++ } } } // iterate over the local files and create new files for filename, localBlockHashList := range localFileMap { if _, ok := fileMetaMap[filename]; !ok { fileMeta := FileMetaData{ Filename: filename, Version: 1, BlockHashList: localBlockHashList, } fileMetaMap[filename] = &fileMeta } } return fileMetaMap } func
(client RPCClient) map[string][]string { // open directory localFileInfos, err := ioutil.ReadDir(client.BaseDir) if err != nil { panic(err) } localFileMap := make(map[string][]string) // iterate over all the local files for _, fileInfo := range localFileInfos { if fileInfo.Name() == "index.txt" { continue } // check if the file is modified file, err := os.Open(filepath.Join(client.BaseDir, fileInfo.Name())) if err != nil { panic(err) } // divide into blocks fileSize := fileInfo.Size() blockSize := client.BlockSize numBlocks := fileSize / int64(blockSize) if fileSize%int64(blockSize) != 0 { numBlocks++ } var blockHashList []string // for empty file if numBlocks == 0 { // write to hash block := NewBlock(0) blockHashList = append(blockHashList, block.Hash()) } for i := int64(0); i < numBlocks; i++ { currentBlockOffset := i * int64(blockSize) var currentBlockSize int if blockSize < int(fileSize-currentBlockOffset) { currentBlockSize = blockSize } else { currentBlockSize = int(fileSize - currentBlockOffset) } block := NewBlock(currentBlockSize) _, err := file.Read(block.BlockData) if err != nil { panic("Invalid file read") } blockHashList = append(blockHashList, block.Hash()) } localFileMap[fileInfo.Name()] = blockHashList } return localFileMap } func writeIndexFile(client RPCClient, fileMetaMap map[string]*FileMetaData) { // err := os.Truncate(filepath.Join(client.BaseDir, "index.txt"), 0) file, err := os.OpenFile(filepath.Join(client.BaseDir, "index.txt"), os.O_RDWR|os.O_TRUNC, 0755) if err != nil { panic(err) } for _, fileMeta := range fileMetaMap { line := fmt.Sprintf( "%s,%d,%s", fileMeta.Filename, fileMeta.Version, strings.Join(fileMeta.BlockHashList, " "), ) line = strings.TrimSpace(line) _, err := file.WriteString(line + "\n") if err != nil { panic(err) } } err = file.Sync() if err != nil { panic(err) } } func downloadFile(client RPCClient, localFileMeta *FileMetaData, remoteFileMeta *FileMetaData) error { if remoteFileMeta == nil { return nil } if localFileMeta != nil && len(localFileMeta.BlockHashList) == len(remoteFileMeta.BlockHashList)
getLocalFileHashBlockListMap
identifier_name
SurfstoreClientUtils.go
) if err == nil { *localFileMeta = remoteFileMeta } } } else { err := downloadFile(client, nil, &remoteFileMeta) if err == nil { localFileMeta := remoteFileMeta fileMetaMap[remoteFilename] = &localFileMeta } } } // working on files only on local -> upload for localFilename, localFileMeta := range fileMetaMap { if _, ok := remoteFileMetaMap[localFilename]; !ok { isUploadFailed = isUploadFailed || !uploadFile(client, localFileMeta) } } if !isUploadFailed { break } } // ==================================Finally, Write into a index file============================= writeIndexFile(client, fileMetaMap) } func uploadFile(client RPCClient, fileMeta *FileMetaData) bool { // divide into blocks filename := fileMeta.Filename if fileMeta.IsTombstone() { var latestVersion int err := client.UpdateFile(fileMeta, &latestVersion) if err != nil { return false } return fileMeta.Version == latestVersion } file, err := os.Open(filepath.Join(client.BaseDir, filename)) if err != nil { log.Println("uploadFile: Failed to open file", filename, err) return false } blockSize := client.BlockSize fileInfo, _ := file.Stat() fileSize := fileInfo.Size() numBlocks := fileSize / int64(blockSize) if fileSize%int64(blockSize) != 0 { numBlocks++ } if numBlocks == 0 { // for empty file blockBuffer := make([]byte, 0) block := Block{BlockData: blockBuffer, BlockSize: 0} succ := false err := client.PutBlock(block, &succ) if !succ || err != nil
} else { for i := int64(0); i < numBlocks; i++ { currentBlockOffset := i * int64(blockSize) var currentBlockSize int if blockSize < int(fileSize-currentBlockOffset) { currentBlockSize = blockSize } else { currentBlockSize = int(fileSize - currentBlockOffset) } block := NewBlock(currentBlockSize) _, err := file.Read(block.BlockData) if err != nil { return false } // write block to server // if there is error -> get block fail -> put block // if the error is nil -> get block succ -> no need succ := false err = client.HasBlock(block.Hash(), &succ) if err != nil { return false } if !succ { succ := false err := client.PutBlock(block, &succ) if !succ || err != nil { log.Println("uploadFile: Failed to put block to server") return false } } } } latestVersion := -1 err = client.UpdateFile(fileMeta, &latestVersion) if err != nil { return false } return fileMeta.Version == latestVersion } func readIndexFile(client RPCClient) map[string]*FileMetaData { // For read access. indexFilename := filepath.Join(client.BaseDir, "index.txt") indexFile, err := os.Open(indexFilename) if err != nil { // index.txt does not exit indexFile, err = os.Create(indexFilename) if err != nil { panic(err) } } defer indexFile.Close() fileMetaMap := make(map[string]*FileMetaData) // read index file reader := bufio.NewReader(indexFile) isReaderEnded := false for !isReaderEnded { line, err := reader.ReadString('\n') isReaderEnded = err == io.EOF if err != nil && err != io.EOF { panic(err) } if line == "" { break } text := strings.TrimSuffix(line, "\n") lineParts := strings.Split(text, ",") if len(lineParts) == 3 { filename := lineParts[0] version, _ := strconv.Atoi(lineParts[1]) blockHasheListString := lineParts[2] blockHasheList := strings.Split(blockHasheListString, " ") fileMeta := FileMetaData{ Filename: filename, Version: version, BlockHashList: blockHasheList, } fileMetaMap[filename] = &fileMeta } else { panic("Invalid index.txt") } } return fileMetaMap } func updateFileMetaMapWithLocalFiles(client RPCClient, fileMetaMap map[string]*FileMetaData) map[string]*FileMetaData { localFileMap := getLocalFileHashBlockListMap(client) // iterate over the file meta map and see if old file exists for filename, fileMeta := range fileMetaMap { if localBlockHashList, ok := localFileMap[filename]; ok { // find the existing file if len(localBlockHashList) != len(fileMeta.BlockHashList) { fileMeta.BlockHashList = localBlockHashList fileMeta.Version++ } else { isFileUpdated := false for i, blockHash := range localBlockHashList { if blockHash != fileMeta.BlockHashList[i] { fileMeta.BlockHashList[i] = blockHash isFileUpdated = true } } if isFileUpdated { fileMeta.Version++ } } } else { // file does not exist in dir, shoud be deleted // if file is not mark as deleted in file meta, update it if !fileMeta.IsTombstone() { fileMeta.MarkTombstone() fileMeta.Version++ } } } // iterate over the local files and create new files for filename, localBlockHashList := range localFileMap { if _, ok := fileMetaMap[filename]; !ok { fileMeta := FileMetaData{ Filename: filename, Version: 1, BlockHashList: localBlockHashList, } fileMetaMap[filename] = &fileMeta } } return fileMetaMap } func getLocalFileHashBlockListMap(client RPCClient) map[string][]string { // open directory localFileInfos, err := ioutil.ReadDir(client.BaseDir) if err != nil { panic(err) } localFileMap := make(map[string][]string) // iterate over all the local files for _, fileInfo := range localFileInfos { if fileInfo.Name() == "index.txt" { continue } // check if the file is modified file, err := os.Open(filepath.Join(client.BaseDir, fileInfo.Name())) if err != nil { panic(err) } // divide into blocks fileSize := fileInfo.Size() blockSize := client.BlockSize numBlocks := fileSize / int64(blockSize) if fileSize%int64(blockSize) != 0 { numBlocks++ } var blockHashList []string // for empty file if numBlocks == 0 { // write to hash block := NewBlock(0) blockHashList = append(blockHashList, block.Hash()) } for i := int64(0); i < numBlocks; i++ { currentBlockOffset := i * int64(blockSize) var currentBlockSize int if blockSize < int(fileSize-currentBlockOffset) { currentBlockSize = blockSize } else { currentBlockSize = int(fileSize - currentBlockOffset) } block := NewBlock(currentBlockSize) _, err := file.Read(block.BlockData) if err != nil { panic("Invalid file read") } blockHashList = append(blockHashList, block.Hash()) } localFileMap[fileInfo.Name()] = blockHashList } return localFileMap } func writeIndexFile(client RPCClient, fileMetaMap map[string]*FileMetaData) { // err := os.Truncate(filepath.Join(client.BaseDir, "index.txt"), 0) file, err := os.OpenFile(filepath.Join(client.BaseDir, "index.txt"), os.O_RDWR|os.O_TRUNC, 0755) if err != nil { panic(err) } for _, fileMeta := range fileMetaMap { line := fmt.Sprintf( "%s,%d,%s", fileMeta.Filename, fileMeta.Version, strings.Join(fileMeta.BlockHashList, " "), ) line = strings.TrimSpace(line) _, err := file.WriteString(line + "\n") if err != nil { panic(err) } } err = file.Sync() if err != nil { panic(err) } } func downloadFile(client RPCClient, localFileMeta *FileMetaData, remoteFileMeta *FileMetaData) error { if remoteFileMeta == nil { return nil } if localFileMeta != nil && len(localFileMeta.BlockHashList) == len(remoteFileMeta.BlockHashList
{ log.Println("uploadFile: Failed to put empty block to the server") return false }
conditional_block
SurfstoreClientUtils.go
4(blockSize) if fileSize%int64(blockSize) != 0 { numBlocks++ } if numBlocks == 0 { // for empty file blockBuffer := make([]byte, 0) block := Block{BlockData: blockBuffer, BlockSize: 0} succ := false err := client.PutBlock(block, &succ) if !succ || err != nil { log.Println("uploadFile: Failed to put empty block to the server") return false } } else { for i := int64(0); i < numBlocks; i++ { currentBlockOffset := i * int64(blockSize) var currentBlockSize int if blockSize < int(fileSize-currentBlockOffset) { currentBlockSize = blockSize } else { currentBlockSize = int(fileSize - currentBlockOffset) } block := NewBlock(currentBlockSize) _, err := file.Read(block.BlockData) if err != nil { return false } // write block to server // if there is error -> get block fail -> put block // if the error is nil -> get block succ -> no need succ := false err = client.HasBlock(block.Hash(), &succ) if err != nil { return false } if !succ { succ := false err := client.PutBlock(block, &succ) if !succ || err != nil { log.Println("uploadFile: Failed to put block to server") return false } } } } latestVersion := -1 err = client.UpdateFile(fileMeta, &latestVersion) if err != nil { return false } return fileMeta.Version == latestVersion } func readIndexFile(client RPCClient) map[string]*FileMetaData { // For read access. indexFilename := filepath.Join(client.BaseDir, "index.txt") indexFile, err := os.Open(indexFilename) if err != nil { // index.txt does not exit indexFile, err = os.Create(indexFilename) if err != nil { panic(err) } } defer indexFile.Close() fileMetaMap := make(map[string]*FileMetaData) // read index file reader := bufio.NewReader(indexFile) isReaderEnded := false for !isReaderEnded { line, err := reader.ReadString('\n') isReaderEnded = err == io.EOF if err != nil && err != io.EOF { panic(err) } if line == "" { break } text := strings.TrimSuffix(line, "\n") lineParts := strings.Split(text, ",") if len(lineParts) == 3 { filename := lineParts[0] version, _ := strconv.Atoi(lineParts[1]) blockHasheListString := lineParts[2] blockHasheList := strings.Split(blockHasheListString, " ") fileMeta := FileMetaData{ Filename: filename, Version: version, BlockHashList: blockHasheList, } fileMetaMap[filename] = &fileMeta } else { panic("Invalid index.txt") } } return fileMetaMap } func updateFileMetaMapWithLocalFiles(client RPCClient, fileMetaMap map[string]*FileMetaData) map[string]*FileMetaData { localFileMap := getLocalFileHashBlockListMap(client) // iterate over the file meta map and see if old file exists for filename, fileMeta := range fileMetaMap { if localBlockHashList, ok := localFileMap[filename]; ok { // find the existing file if len(localBlockHashList) != len(fileMeta.BlockHashList) { fileMeta.BlockHashList = localBlockHashList fileMeta.Version++ } else { isFileUpdated := false for i, blockHash := range localBlockHashList { if blockHash != fileMeta.BlockHashList[i] { fileMeta.BlockHashList[i] = blockHash isFileUpdated = true } } if isFileUpdated { fileMeta.Version++ } } } else { // file does not exist in dir, shoud be deleted // if file is not mark as deleted in file meta, update it if !fileMeta.IsTombstone() { fileMeta.MarkTombstone() fileMeta.Version++ } } } // iterate over the local files and create new files for filename, localBlockHashList := range localFileMap { if _, ok := fileMetaMap[filename]; !ok { fileMeta := FileMetaData{ Filename: filename, Version: 1, BlockHashList: localBlockHashList, } fileMetaMap[filename] = &fileMeta } } return fileMetaMap } func getLocalFileHashBlockListMap(client RPCClient) map[string][]string { // open directory localFileInfos, err := ioutil.ReadDir(client.BaseDir) if err != nil { panic(err) } localFileMap := make(map[string][]string) // iterate over all the local files for _, fileInfo := range localFileInfos { if fileInfo.Name() == "index.txt" { continue } // check if the file is modified file, err := os.Open(filepath.Join(client.BaseDir, fileInfo.Name())) if err != nil { panic(err) } // divide into blocks fileSize := fileInfo.Size() blockSize := client.BlockSize numBlocks := fileSize / int64(blockSize) if fileSize%int64(blockSize) != 0 { numBlocks++ } var blockHashList []string // for empty file if numBlocks == 0 { // write to hash block := NewBlock(0) blockHashList = append(blockHashList, block.Hash()) } for i := int64(0); i < numBlocks; i++ { currentBlockOffset := i * int64(blockSize) var currentBlockSize int if blockSize < int(fileSize-currentBlockOffset) { currentBlockSize = blockSize } else { currentBlockSize = int(fileSize - currentBlockOffset) } block := NewBlock(currentBlockSize) _, err := file.Read(block.BlockData) if err != nil { panic("Invalid file read") } blockHashList = append(blockHashList, block.Hash()) } localFileMap[fileInfo.Name()] = blockHashList } return localFileMap } func writeIndexFile(client RPCClient, fileMetaMap map[string]*FileMetaData) { // err := os.Truncate(filepath.Join(client.BaseDir, "index.txt"), 0) file, err := os.OpenFile(filepath.Join(client.BaseDir, "index.txt"), os.O_RDWR|os.O_TRUNC, 0755) if err != nil { panic(err) } for _, fileMeta := range fileMetaMap { line := fmt.Sprintf( "%s,%d,%s", fileMeta.Filename, fileMeta.Version, strings.Join(fileMeta.BlockHashList, " "), ) line = strings.TrimSpace(line) _, err := file.WriteString(line + "\n") if err != nil { panic(err) } } err = file.Sync() if err != nil { panic(err) } } func downloadFile(client RPCClient, localFileMeta *FileMetaData, remoteFileMeta *FileMetaData) error { if remoteFileMeta == nil { return nil } if localFileMeta != nil && len(localFileMeta.BlockHashList) == len(remoteFileMeta.BlockHashList) { isHashListEqual := true for i, hash := range localFileMeta.BlockHashList { if hash != remoteFileMeta.BlockHashList[i] { isHashListEqual = false break } } if isHashListEqual { return nil } } var fileBlocks []*Block if !remoteFileMeta.IsTombstone() { // get block map for remote blockMap := make(map[string]*Block) for _, hash := range remoteFileMeta.BlockHashList { blockMap[hash] = nil } // update map with local blocks with existing files if localFileMeta != nil && !localFileMeta.IsTombstone() { var fileInfo os.FileInfo file, err := os.Open(filepath.Join(client.BaseDir, localFileMeta.Filename)) if err == nil { fileInfo, err = file.Stat() } if err == nil { // successfully access local file // divide into blocks fileSize := fileInfo.Size() blockSize := client.BlockSize // for empty file if len(localFileMeta.BlockHashList) == 0 { // write to hash localBlock := NewBlock(0) blockHash := localBlock.Hash() block, found := blockMap[blockHash] if found && block == nil { blockMap[blockHash] = &localBlock } } for i, localBlockHash := range localFileMeta.BlockHashList {
random_line_split
SurfstoreClientUtils.go
) if err == nil { *localFileMeta = remoteFileMeta } } } else { err := downloadFile(client, nil, &remoteFileMeta) if err == nil { localFileMeta := remoteFileMeta fileMetaMap[remoteFilename] = &localFileMeta } } } // working on files only on local -> upload for localFilename, localFileMeta := range fileMetaMap { if _, ok := remoteFileMetaMap[localFilename]; !ok { isUploadFailed = isUploadFailed || !uploadFile(client, localFileMeta) } } if !isUploadFailed { break } } // ==================================Finally, Write into a index file============================= writeIndexFile(client, fileMetaMap) } func uploadFile(client RPCClient, fileMeta *FileMetaData) bool { // divide into blocks filename := fileMeta.Filename if fileMeta.IsTombstone() { var latestVersion int err := client.UpdateFile(fileMeta, &latestVersion) if err != nil { return false } return fileMeta.Version == latestVersion } file, err := os.Open(filepath.Join(client.BaseDir, filename)) if err != nil { log.Println("uploadFile: Failed to open file", filename, err) return false } blockSize := client.BlockSize fileInfo, _ := file.Stat() fileSize := fileInfo.Size() numBlocks := fileSize / int64(blockSize) if fileSize%int64(blockSize) != 0 { numBlocks++ } if numBlocks == 0 { // for empty file blockBuffer := make([]byte, 0) block := Block{BlockData: blockBuffer, BlockSize: 0} succ := false err := client.PutBlock(block, &succ) if !succ || err != nil { log.Println("uploadFile: Failed to put empty block to the server") return false } } else { for i := int64(0); i < numBlocks; i++ { currentBlockOffset := i * int64(blockSize) var currentBlockSize int if blockSize < int(fileSize-currentBlockOffset) { currentBlockSize = blockSize } else { currentBlockSize = int(fileSize - currentBlockOffset) } block := NewBlock(currentBlockSize) _, err := file.Read(block.BlockData) if err != nil { return false } // write block to server // if there is error -> get block fail -> put block // if the error is nil -> get block succ -> no need succ := false err = client.HasBlock(block.Hash(), &succ) if err != nil { return false } if !succ { succ := false err := client.PutBlock(block, &succ) if !succ || err != nil { log.Println("uploadFile: Failed to put block to server") return false } } } } latestVersion := -1 err = client.UpdateFile(fileMeta, &latestVersion) if err != nil { return false } return fileMeta.Version == latestVersion } func readIndexFile(client RPCClient) map[string]*FileMetaData
line, err := reader.ReadString('\n') isReaderEnded = err == io.EOF if err != nil && err != io.EOF { panic(err) } if line == "" { break } text := strings.TrimSuffix(line, "\n") lineParts := strings.Split(text, ",") if len(lineParts) == 3 { filename := lineParts[0] version, _ := strconv.Atoi(lineParts[1]) blockHasheListString := lineParts[2] blockHasheList := strings.Split(blockHasheListString, " ") fileMeta := FileMetaData{ Filename: filename, Version: version, BlockHashList: blockHasheList, } fileMetaMap[filename] = &fileMeta } else { panic("Invalid index.txt") } } return fileMetaMap } func updateFileMetaMapWithLocalFiles(client RPCClient, fileMetaMap map[string]*FileMetaData) map[string]*FileMetaData { localFileMap := getLocalFileHashBlockListMap(client) // iterate over the file meta map and see if old file exists for filename, fileMeta := range fileMetaMap { if localBlockHashList, ok := localFileMap[filename]; ok { // find the existing file if len(localBlockHashList) != len(fileMeta.BlockHashList) { fileMeta.BlockHashList = localBlockHashList fileMeta.Version++ } else { isFileUpdated := false for i, blockHash := range localBlockHashList { if blockHash != fileMeta.BlockHashList[i] { fileMeta.BlockHashList[i] = blockHash isFileUpdated = true } } if isFileUpdated { fileMeta.Version++ } } } else { // file does not exist in dir, shoud be deleted // if file is not mark as deleted in file meta, update it if !fileMeta.IsTombstone() { fileMeta.MarkTombstone() fileMeta.Version++ } } } // iterate over the local files and create new files for filename, localBlockHashList := range localFileMap { if _, ok := fileMetaMap[filename]; !ok { fileMeta := FileMetaData{ Filename: filename, Version: 1, BlockHashList: localBlockHashList, } fileMetaMap[filename] = &fileMeta } } return fileMetaMap } func getLocalFileHashBlockListMap(client RPCClient) map[string][]string { // open directory localFileInfos, err := ioutil.ReadDir(client.BaseDir) if err != nil { panic(err) } localFileMap := make(map[string][]string) // iterate over all the local files for _, fileInfo := range localFileInfos { if fileInfo.Name() == "index.txt" { continue } // check if the file is modified file, err := os.Open(filepath.Join(client.BaseDir, fileInfo.Name())) if err != nil { panic(err) } // divide into blocks fileSize := fileInfo.Size() blockSize := client.BlockSize numBlocks := fileSize / int64(blockSize) if fileSize%int64(blockSize) != 0 { numBlocks++ } var blockHashList []string // for empty file if numBlocks == 0 { // write to hash block := NewBlock(0) blockHashList = append(blockHashList, block.Hash()) } for i := int64(0); i < numBlocks; i++ { currentBlockOffset := i * int64(blockSize) var currentBlockSize int if blockSize < int(fileSize-currentBlockOffset) { currentBlockSize = blockSize } else { currentBlockSize = int(fileSize - currentBlockOffset) } block := NewBlock(currentBlockSize) _, err := file.Read(block.BlockData) if err != nil { panic("Invalid file read") } blockHashList = append(blockHashList, block.Hash()) } localFileMap[fileInfo.Name()] = blockHashList } return localFileMap } func writeIndexFile(client RPCClient, fileMetaMap map[string]*FileMetaData) { // err := os.Truncate(filepath.Join(client.BaseDir, "index.txt"), 0) file, err := os.OpenFile(filepath.Join(client.BaseDir, "index.txt"), os.O_RDWR|os.O_TRUNC, 0755) if err != nil { panic(err) } for _, fileMeta := range fileMetaMap { line := fmt.Sprintf( "%s,%d,%s", fileMeta.Filename, fileMeta.Version, strings.Join(fileMeta.BlockHashList, " "), ) line = strings.TrimSpace(line) _, err := file.WriteString(line + "\n") if err != nil { panic(err) } } err = file.Sync() if err != nil { panic(err) } } func downloadFile(client RPCClient, localFileMeta *FileMetaData, remoteFileMeta *FileMetaData) error { if remoteFileMeta == nil { return nil } if localFileMeta != nil && len(localFileMeta.BlockHashList) == len(remoteFileMeta.BlockHashList)
{ // For read access. indexFilename := filepath.Join(client.BaseDir, "index.txt") indexFile, err := os.Open(indexFilename) if err != nil { // index.txt does not exit indexFile, err = os.Create(indexFilename) if err != nil { panic(err) } } defer indexFile.Close() fileMetaMap := make(map[string]*FileMetaData) // read index file reader := bufio.NewReader(indexFile) isReaderEnded := false for !isReaderEnded {
identifier_body
player.rs
, visible_game: &VisibleGame, discard_pile: &mut Vec<Card>) -> bool { // Removes and returns the given card from the player's hand. fn remove_from_hand(hand: &mut Vec<Card>, card: &Card) -> Card { let index = hand.iter().position(|c| c == card).unwrap(); hand.swap_remove(index) } if self.can_play(action, visible_game) { match action { Action::Build(card) => { let card_from_hand = remove_from_hand(&mut self.hand, card); self.built_structures.push(card_from_hand); self.coins -= card_from_hand.cost().coins; // TODO: deal with borrowed resources } Action::Wonder(_) => todo!(), Action::Discard(card) => { discard_pile.push(remove_from_hand(&mut self.hand, card)); self.coins += 3; } } true } else { false } } /// Replaces this player's hand with the given cards, returning the hand the player had before the swap. pub fn swap_hand(&mut self, new_hand: Vec<Card>) -> Vec<Card> { mem::replace(&mut self.hand, new_hand) } fn evaluate_green(colour_cards: &[Card]) -> f32 { let mut science_items_count: HashMap<ScienceItem, i32> = HashMap::new(); science_items_count.insert(ScienceItem::Compass, 0); science_items_count.insert(ScienceItem::Cog, 0); science_items_count.insert(ScienceItem::Tablet, 0); for card in colour_cards.iter() { if let Power::Science(science_items) = card.power() { for science_item in science_items.iter() { let count = science_items_count.entry(*science_item).or_insert(0); *count += 1; } } } let score_for_sets_of_identical_symbols: f32 = science_items_count.iter() .filter(|(_, count)| **count > 0) .map(|(_, count)| { (*count as f32).powf(2f32) }) .sum(); let score_for_all_symbol_groups: f32 = 7f32 * *science_items_count.iter().min_by_key(|(_, count)| *count).unwrap().1 as f32; score_for_all_symbol_groups + score_for_sets_of_identical_symbols } fn evaluate_colour(cards_of_given_colour: &[Card]) -> f32 { let colour = cards_of_given_colour.get(0).unwrap().colour(); match colour { Colour::Green => Self::evaluate_green(cards_of_given_colour), _ => cards_of_given_colour.iter().map(|card| card.immediate_strength()).sum(), } } fn strength_internal(cards: &[Card]) -> f32 { let mut colour_to_structure = HashMap::new(); for structure in cards.iter() { let colour_structures = colour_to_structure.entry(structure.colour()).or_insert_with(Vec::new); colour_structures.push(*structure) } colour_to_structure.iter() .map(|colour_entry| Self::evaluate_colour(colour_entry.1)) .sum() } /// Returns this player's "strength" -- a number where a higher value means the player is doing better than a lower /// value. pub fn strength(&self) -> f32 { Self::strength_internal(&self.built_structures) } pub fn can_play(&self, action: &Action, visible_game: &VisibleGame) -> bool { match action { Action::Build(card) => self.can_play_card(card, visible_game), Action::Wonder(_) => todo!(), Action::Discard(_) => true, } } /// Returns `true` if the user can afford to play the given card, given the resources the player /// has access to. /// /// TODO: doesn't currently deal with borrowing resources from neighbours. fn can_play_card(&self, card: &Card, _visible_game: &VisibleGame) -> bool { if !self.hand.iter().any(|c| c == card) { return false; } // Initialise a Resources struct with the number of coins we have. let mut available_resources = Resources::coins(self.coins); // Add all the other resources we always have access to (ie. those that are not resource // "choice" cards. At the same time, make a vector of resources choices available to us. let mut choices = Vec::new(); for card in &self.built_structures { match card.power() { // TODO: can we write these four options more succinctly? Power::PurchasableProducer(ProducedResources::Single(resources)) => { available_resources += &resources; } Power::Producer(ProducedResources::Single(resources)) => { available_resources += &resources; } Power::PurchasableProducer(ProducedResources::Choice(choice)) => { choices.push(choice); } Power::Producer(ProducedResources::Choice(choice)) => { choices.push(choice); } _ => {} } } // Add Wonder starting resources. available_resources += &self.wonder.starting_resource(); if available_resources.can_afford(&card.cost()) { return true; } if !choices.is_empty() { // Iterate through all possible combinations of the choices we have. Use the iteration // index to work out which choice to make for each card. let combinations: u32 = choices.iter() .fold(1, |x, y| x * y.len() as u32); for combination in 0..combinations { let mut available_resources_option = available_resources.clone(); let mut combination = combination; for choice in &choices { let index = combination % choice.len() as u32; available_resources_option += &choice[index as usize]; combination /= choice.len() as u32; } if available_resources_option.can_afford(&card.cost()) { return true; } } } false } } /// Represents the aspects of [`Player`] that are public knowledge (ie. visible on the table). Things like a player's /// current hand are not included. pub struct PublicPlayer { pub wonder: WonderBoard, pub built_structures: Vec<Card>, pub coins: u32, } impl PublicPlayer { /// Creates a [`PublicPlayer`] from a [`Player`], copy/cloning the values so the originals can be mutated later /// without issue. pub fn new(player: &Player) -> PublicPlayer { PublicPlayer { wonder: player.wonder, built_structures: player.built_structures.clone(), coins: player.coins, } } } #[cfg(test)] mod tests { use Card::*; use super::*; use crate::algorithms::random::Random; #[test] fn can_play_returns_true_when_player_can_afford_card() { // TODO: @Before etc let player = new_player(vec![LumberYard]); assert_eq!(true, player.can_play(&Action::Build(LumberYard), &visible_game())); } #[test] fn can_play_returns_true_after_player_builds_required_resources() { let mut player = new_player(vec![StonePit, Quarry, Aqueduct]); player.do_action(&Action::Build(StonePit), &visible_game(), &mut vec![]); assert_eq!(false, player.can_play(&Action::Build(Aqueduct), &visible_game())); assert_eq!(true, player.do_action(&Action::Build(Quarry), &visible_game(), &mut vec![])); assert_eq!(true, player.can_play(&Action::Build(Aqueduct), &visible_game())); } #[test] fn strength_returns_sum_of_card_strengths() { assert_eq!(0.0, Player::strength_internal(&vec![StonePit])); assert_eq!(5.0, Player::strength_internal(&vec![StonePit, Quarry, Aqueduct])); assert_eq!(6.0, Player::strength_internal(&vec![StonePit, Quarry, Aqueduct, Loom1, Apothecary])); } #[test] fn strength_returns_correct_strength_of_green_structures() { assert_eq!(1.0, Player::strength_internal(&vec![Lodge])); assert_eq!(4.0, Player::strength_internal(&vec![Lodge, Apothecary])); assert_eq!(9.0, Player::strength_internal(&vec![Lodge, Apothecary, Dispensary])); assert_eq!(10.0, Player::strength_internal(&vec![Lodge, Workshop, Library])); assert_eq!(21.0, Player::strength_internal(&vec![Lodge, Apothecary, Dispensary, Laboratory, Workshop, Library])); // rulebook example } #[test] fn can_play_returns_false_when_player_cannot_pay() { let mut player = new_player(vec![]); player.coins = 0; //TODO introduce a Bank type to allow for double-entry bookkeeping instead of this assert_eq!(false, player.can_play(&Action::Build(TreeFarm), &visible_game())); } #[test] fn can_play_returns_false_when_both_choice_resources_needed()
{ // TODO implement }
identifier_body
player.rs
<Card>) -> bool { // Removes and returns the given card from the player's hand. fn remove_from_hand(hand: &mut Vec<Card>, card: &Card) -> Card { let index = hand.iter().position(|c| c == card).unwrap(); hand.swap_remove(index) } if self.can_play(action, visible_game) { match action { Action::Build(card) => { let card_from_hand = remove_from_hand(&mut self.hand, card); self.built_structures.push(card_from_hand); self.coins -= card_from_hand.cost().coins; // TODO: deal with borrowed resources } Action::Wonder(_) => todo!(), Action::Discard(card) => { discard_pile.push(remove_from_hand(&mut self.hand, card)); self.coins += 3; } } true } else { false } } /// Replaces this player's hand with the given cards, returning the hand the player had before the swap. pub fn swap_hand(&mut self, new_hand: Vec<Card>) -> Vec<Card> { mem::replace(&mut self.hand, new_hand) } fn evaluate_green(colour_cards: &[Card]) -> f32 { let mut science_items_count: HashMap<ScienceItem, i32> = HashMap::new(); science_items_count.insert(ScienceItem::Compass, 0); science_items_count.insert(ScienceItem::Cog, 0); science_items_count.insert(ScienceItem::Tablet, 0); for card in colour_cards.iter() { if let Power::Science(science_items) = card.power() { for science_item in science_items.iter() { let count = science_items_count.entry(*science_item).or_insert(0); *count += 1; } } } let score_for_sets_of_identical_symbols: f32 = science_items_count.iter() .filter(|(_, count)| **count > 0) .map(|(_, count)| { (*count as f32).powf(2f32) }) .sum(); let score_for_all_symbol_groups: f32 = 7f32 * *science_items_count.iter().min_by_key(|(_, count)| *count).unwrap().1 as f32; score_for_all_symbol_groups + score_for_sets_of_identical_symbols } fn evaluate_colour(cards_of_given_colour: &[Card]) -> f32 { let colour = cards_of_given_colour.get(0).unwrap().colour(); match colour { Colour::Green => Self::evaluate_green(cards_of_given_colour), _ => cards_of_given_colour.iter().map(|card| card.immediate_strength()).sum(), } } fn strength_internal(cards: &[Card]) -> f32 { let mut colour_to_structure = HashMap::new(); for structure in cards.iter() { let colour_structures = colour_to_structure.entry(structure.colour()).or_insert_with(Vec::new); colour_structures.push(*structure) } colour_to_structure.iter() .map(|colour_entry| Self::evaluate_colour(colour_entry.1)) .sum() } /// Returns this player's "strength" -- a number where a higher value means the player is doing better than a lower /// value. pub fn strength(&self) -> f32 { Self::strength_internal(&self.built_structures) } pub fn can_play(&self, action: &Action, visible_game: &VisibleGame) -> bool { match action { Action::Build(card) => self.can_play_card(card, visible_game), Action::Wonder(_) => todo!(), Action::Discard(_) => true, } } /// Returns `true` if the user can afford to play the given card, given the resources the player /// has access to. /// /// TODO: doesn't currently deal with borrowing resources from neighbours. fn can_play_card(&self, card: &Card, _visible_game: &VisibleGame) -> bool { if !self.hand.iter().any(|c| c == card) { return false; } // Initialise a Resources struct with the number of coins we have. let mut available_resources = Resources::coins(self.coins); // Add all the other resources we always have access to (ie. those that are not resource // "choice" cards. At the same time, make a vector of resources choices available to us. let mut choices = Vec::new(); for card in &self.built_structures { match card.power() { // TODO: can we write these four options more succinctly? Power::PurchasableProducer(ProducedResources::Single(resources)) => { available_resources += &resources; } Power::Producer(ProducedResources::Single(resources)) => { available_resources += &resources; } Power::PurchasableProducer(ProducedResources::Choice(choice)) => { choices.push(choice); } Power::Producer(ProducedResources::Choice(choice)) => { choices.push(choice); } _ => {} } } // Add Wonder starting resources. available_resources += &self.wonder.starting_resource(); if available_resources.can_afford(&card.cost()) { return true; } if !choices.is_empty() { // Iterate through all possible combinations of the choices we have. Use the iteration // index to work out which choice to make for each card. let combinations: u32 = choices.iter() .fold(1, |x, y| x * y.len() as u32); for combination in 0..combinations { let mut available_resources_option = available_resources.clone(); let mut combination = combination; for choice in &choices { let index = combination % choice.len() as u32; available_resources_option += &choice[index as usize]; combination /= choice.len() as u32; } if available_resources_option.can_afford(&card.cost()) { return true; } } } false } } /// Represents the aspects of [`Player`] that are public knowledge (ie. visible on the table). Things like a player's /// current hand are not included. pub struct PublicPlayer { pub wonder: WonderBoard, pub built_structures: Vec<Card>, pub coins: u32, } impl PublicPlayer { /// Creates a [`PublicPlayer`] from a [`Player`], copy/cloning the values so the originals can be mutated later /// without issue. pub fn new(player: &Player) -> PublicPlayer { PublicPlayer { wonder: player.wonder, built_structures: player.built_structures.clone(), coins: player.coins, } } } #[cfg(test)] mod tests { use Card::*; use super::*; use crate::algorithms::random::Random; #[test] fn can_play_returns_true_when_player_can_afford_card() { // TODO: @Before etc let player = new_player(vec![LumberYard]); assert_eq!(true, player.can_play(&Action::Build(LumberYard), &visible_game())); } #[test] fn can_play_returns_true_after_player_builds_required_resources() { let mut player = new_player(vec![StonePit, Quarry, Aqueduct]); player.do_action(&Action::Build(StonePit), &visible_game(), &mut vec![]); assert_eq!(false, player.can_play(&Action::Build(Aqueduct), &visible_game())); assert_eq!(true, player.do_action(&Action::Build(Quarry), &visible_game(), &mut vec![])); assert_eq!(true, player.can_play(&Action::Build(Aqueduct), &visible_game())); } #[test] fn strength_returns_sum_of_card_strengths() { assert_eq!(0.0, Player::strength_internal(&vec![StonePit])); assert_eq!(5.0, Player::strength_internal(&vec![StonePit, Quarry, Aqueduct])); assert_eq!(6.0, Player::strength_internal(&vec![StonePit, Quarry, Aqueduct, Loom1, Apothecary])); } #[test] fn strength_returns_correct_strength_of_green_structures() { assert_eq!(1.0, Player::strength_internal(&vec![Lodge])); assert_eq!(4.0, Player::strength_internal(&vec![Lodge, Apothecary])); assert_eq!(9.0, Player::strength_internal(&vec![Lodge, Apothecary, Dispensary])); assert_eq!(10.0, Player::strength_internal(&vec![Lodge, Workshop, Library])); assert_eq!(21.0, Player::strength_internal(&vec![Lodge, Apothecary, Dispensary, Laboratory, Workshop, Library])); // rulebook example } #[test] fn can_play_returns_false_when_player_cannot_pay() { let mut player = new_player(vec![]); player.coins = 0; //TODO introduce a Bank type to allow for double-entry bookkeeping instead of this assert_eq!(false, player.can_play(&Action::Build(TreeFarm), &visible_game())); } #[test] fn can_play_returns_false_when_both_choice_resources_needed() { // TODO implement } #[test] fn
do_action_returns_false_if_action_not_playable
identifier_name
player.rs
} pub fn coins(&self) -> u32 { self.coins } pub fn hand(&self) -> &Vec<Card> { &self.hand } /// Performs the given [`Action`] on the current player, for example moving a card from the player's hand into the /// player's built structures. Returns `true` if the action is legal, `false` otherwise (in which case this function /// otherwise does nothing). pub fn do_action(&mut self, action: &Action, visible_game: &VisibleGame, discard_pile: &mut Vec<Card>) -> bool { // Removes and returns the given card from the player's hand. fn remove_from_hand(hand: &mut Vec<Card>, card: &Card) -> Card { let index = hand.iter().position(|c| c == card).unwrap(); hand.swap_remove(index) } if self.can_play(action, visible_game) { match action { Action::Build(card) => { let card_from_hand = remove_from_hand(&mut self.hand, card); self.built_structures.push(card_from_hand); self.coins -= card_from_hand.cost().coins; // TODO: deal with borrowed resources } Action::Wonder(_) => todo!(), Action::Discard(card) => { discard_pile.push(remove_from_hand(&mut self.hand, card)); self.coins += 3; } } true } else { false } } /// Replaces this player's hand with the given cards, returning the hand the player had before the swap. pub fn swap_hand(&mut self, new_hand: Vec<Card>) -> Vec<Card> { mem::replace(&mut self.hand, new_hand) } fn evaluate_green(colour_cards: &[Card]) -> f32 { let mut science_items_count: HashMap<ScienceItem, i32> = HashMap::new(); science_items_count.insert(ScienceItem::Compass, 0); science_items_count.insert(ScienceItem::Cog, 0); science_items_count.insert(ScienceItem::Tablet, 0); for card in colour_cards.iter() { if let Power::Science(science_items) = card.power() { for science_item in science_items.iter() { let count = science_items_count.entry(*science_item).or_insert(0); *count += 1; } } } let score_for_sets_of_identical_symbols: f32 = science_items_count.iter() .filter(|(_, count)| **count > 0) .map(|(_, count)| { (*count as f32).powf(2f32) }) .sum(); let score_for_all_symbol_groups: f32 = 7f32 * *science_items_count.iter().min_by_key(|(_, count)| *count).unwrap().1 as f32; score_for_all_symbol_groups + score_for_sets_of_identical_symbols } fn evaluate_colour(cards_of_given_colour: &[Card]) -> f32 { let colour = cards_of_given_colour.get(0).unwrap().colour(); match colour { Colour::Green => Self::evaluate_green(cards_of_given_colour), _ => cards_of_given_colour.iter().map(|card| card.immediate_strength()).sum(), } } fn strength_internal(cards: &[Card]) -> f32 { let mut colour_to_structure = HashMap::new(); for structure in cards.iter() { let colour_structures = colour_to_structure.entry(structure.colour()).or_insert_with(Vec::new); colour_structures.push(*structure) } colour_to_structure.iter() .map(|colour_entry| Self::evaluate_colour(colour_entry.1)) .sum() } /// Returns this player's "strength" -- a number where a higher value means the player is doing better than a lower /// value. pub fn strength(&self) -> f32 { Self::strength_internal(&self.built_structures) } pub fn can_play(&self, action: &Action, visible_game: &VisibleGame) -> bool { match action { Action::Build(card) => self.can_play_card(card, visible_game), Action::Wonder(_) => todo!(), Action::Discard(_) => true, } } /// Returns `true` if the user can afford to play the given card, given the resources the player /// has access to. /// /// TODO: doesn't currently deal with borrowing resources from neighbours. fn can_play_card(&self, card: &Card, _visible_game: &VisibleGame) -> bool { if !self.hand.iter().any(|c| c == card) { return false; }
// "choice" cards. At the same time, make a vector of resources choices available to us. let mut choices = Vec::new(); for card in &self.built_structures { match card.power() { // TODO: can we write these four options more succinctly? Power::PurchasableProducer(ProducedResources::Single(resources)) => { available_resources += &resources; } Power::Producer(ProducedResources::Single(resources)) => { available_resources += &resources; } Power::PurchasableProducer(ProducedResources::Choice(choice)) => { choices.push(choice); } Power::Producer(ProducedResources::Choice(choice)) => { choices.push(choice); } _ => {} } } // Add Wonder starting resources. available_resources += &self.wonder.starting_resource(); if available_resources.can_afford(&card.cost()) { return true; } if !choices.is_empty() { // Iterate through all possible combinations of the choices we have. Use the iteration // index to work out which choice to make for each card. let combinations: u32 = choices.iter() .fold(1, |x, y| x * y.len() as u32); for combination in 0..combinations { let mut available_resources_option = available_resources.clone(); let mut combination = combination; for choice in &choices { let index = combination % choice.len() as u32; available_resources_option += &choice[index as usize]; combination /= choice.len() as u32; } if available_resources_option.can_afford(&card.cost()) { return true; } } } false } } /// Represents the aspects of [`Player`] that are public knowledge (ie. visible on the table). Things like a player's /// current hand are not included. pub struct PublicPlayer { pub wonder: WonderBoard, pub built_structures: Vec<Card>, pub coins: u32, } impl PublicPlayer { /// Creates a [`PublicPlayer`] from a [`Player`], copy/cloning the values so the originals can be mutated later /// without issue. pub fn new(player: &Player) -> PublicPlayer { PublicPlayer { wonder: player.wonder, built_structures: player.built_structures.clone(), coins: player.coins, } } } #[cfg(test)] mod tests { use Card::*; use super::*; use crate::algorithms::random::Random; #[test] fn can_play_returns_true_when_player_can_afford_card() { // TODO: @Before etc let player = new_player(vec![LumberYard]); assert_eq!(true, player.can_play(&Action::Build(LumberYard), &visible_game())); } #[test] fn can_play_returns_true_after_player_builds_required_resources() { let mut player = new_player(vec![StonePit, Quarry, Aqueduct]); player.do_action(&Action::Build(StonePit), &visible_game(), &mut vec![]); assert_eq!(false, player.can_play(&Action::Build(Aqueduct), &visible_game())); assert_eq!(true, player.do_action(&Action::Build(Quarry), &visible_game(), &mut vec![])); assert_eq!(true, player.can_play(&Action::Build(Aqueduct), &visible_game())); } #[test] fn strength_returns_sum_of_card_strengths() { assert_eq!(0.0, Player::strength_internal(&vec![StonePit])); assert_eq!(5.0, Player::strength_internal(&vec![StonePit, Quarry, Aqueduct])); assert_eq!(6.0, Player::strength_internal(&vec![StonePit, Quarry, Aqueduct, Loom1, Apothecary])); } #[test] fn strength_returns_correct_strength_of_green_structures() { assert_eq!(1.0, Player::strength_internal(&vec![Lodge])); assert_eq!(4.0, Player::strength_internal(&vec![Lodge, Apothecary])); assert_eq!(9.0, Player::strength_internal(&vec![Lodge, Apothecary, Dispensary])); assert_eq!(10.0, Player::strength_internal(&vec![Lodge, Workshop, Library])); assert_eq!(21.0, Player::strength_internal(&vec![Lodge, Apothecary, Dispensary, Laboratory
// Initialise a Resources struct with the number of coins we have. let mut available_resources = Resources::coins(self.coins); // Add all the other resources we always have access to (ie. those that are not resource
random_line_split
runtime.rs
or threads, you /// will be able to access the runtime from those tasks or threads. /// /// The difference between <code>[Arc]\<Runtime></code> and [`Handle`] is that /// an <code>[Arc]\<Runtime></code> will prevent the runtime from shutting down, /// whereas a [`Handle`] does not prevent that. This is because shutdown of the /// runtime happens when the destructor of the `Runtime` object runs. /// /// Calls to [`shutdown_background`] and [`shutdown_timeout`] require exclusive /// ownership of the `Runtime` type. When using an <code>[Arc]\<Runtime></code>, /// this can be achieved via [`Arc::try_unwrap`] when only one strong count /// reference is left over. /// /// The runtime context is entered using the [`Runtime::enter`] or /// [`Handle::enter`] methods, which use a thread-local variable to store the /// current runtime. Whenever you are inside the runtime context, methods such /// as [`tokio::spawn`] will use the runtime whose context you are inside. /// /// [timer]: crate::time /// [mod]: index.html /// [`new`]: method@Self::new /// [`Builder`]: struct@Builder /// [`Handle`]: struct@Handle /// [`tokio::spawn`]: crate::spawn /// [`Arc::try_unwrap`]: std::sync::Arc::try_unwrap /// [Arc]: std::sync::Arc /// [`shutdown_background`]: method@Runtime::shutdown_background /// [`shutdown_timeout`]: method@Runtime::shutdown_timeout #[derive(Debug)] pub struct Runtime { /// Task scheduler scheduler: Scheduler, /// Handle to runtime, also contains driver handles handle: Handle, /// Blocking pool handle, used to signal shutdown blocking_pool: BlockingPool, } /// The flavor of a `Runtime`. /// /// This is the return type for [`Handle::runtime_flavor`](crate::runtime::Handle::runtime_flavor()). #[derive(Debug, PartialEq, Eq)] #[non_exhaustive] pub enum RuntimeFlavor { /// The flavor that executes all tasks on the current thread. CurrentThread, /// The flavor that executes tasks across multiple threads. MultiThread, /// The flavor that executes tasks across multiple threads. #[cfg(tokio_unstable)] MultiThreadAlt, } /// The runtime scheduler is either a multi-thread or a current-thread executor. #[derive(Debug)] pub(super) enum Scheduler { /// Execute all tasks on the current-thread. CurrentThread(CurrentThread), /// Execute tasks across multiple threads. #[cfg(all(feature = "rt-multi-thread", not(target_os = "wasi")))] MultiThread(MultiThread), /// Execute tasks across multiple threads. #[cfg(all(tokio_unstable, feature = "rt-multi-thread", not(target_os = "wasi")))] MultiThreadAlt(MultiThreadAlt), } impl Runtime { pub(super) fn from_parts( scheduler: Scheduler, handle: Handle, blocking_pool: BlockingPool, ) -> Runtime { Runtime { scheduler, handle, blocking_pool, } } cfg_not_wasi! { /// Creates a new runtime instance with default configuration values. /// /// This results in the multi threaded scheduler, I/O driver, and time driver being /// initialized. /// /// Most applications will not need to call this function directly. Instead, /// they will use the [`#[tokio::main]` attribute][main]. When a more complex /// configuration is necessary, the [runtime builder] may be used. /// /// See [module level][mod] documentation for more details. /// /// # Examples /// /// Creating a new `Runtime` with default configuration values. /// /// ``` /// use tokio::runtime::Runtime; /// /// let rt = Runtime::new() /// .unwrap(); /// /// // Use the runtime... /// ```
/// [runtime builder]: crate::runtime::Builder #[cfg(feature = "rt-multi-thread")] #[cfg_attr(docsrs, doc(cfg(feature = "rt-multi-thread")))] pub fn new() -> std::io::Result<Runtime> { Builder::new_multi_thread().enable_all().build() } } /// Returns a handle to the runtime's spawner. /// /// The returned handle can be used to spawn tasks that run on this runtime, and can /// be cloned to allow moving the `Handle` to other threads. /// /// Calling [`Handle::block_on`] on a handle to a `current_thread` runtime is error-prone. /// Refer to the documentation of [`Handle::block_on`] for more. /// /// # Examples /// /// ``` /// use tokio::runtime::Runtime; /// /// let rt = Runtime::new() /// .unwrap(); /// /// let handle = rt.handle(); /// /// // Use the handle... /// ``` pub fn handle(&self) -> &Handle { &self.handle } /// Spawns a future onto the Tokio runtime. /// /// This spawns the given future onto the runtime's executor, usually a /// thread pool. The thread pool is then responsible for polling the future /// until it completes. /// /// The provided future will start running in the background immediately /// when `spawn` is called, even if you don't await the returned /// `JoinHandle`. /// /// See [module level][mod] documentation for more details. /// /// [mod]: index.html /// /// # Examples /// /// ``` /// use tokio::runtime::Runtime; /// /// # fn dox() { /// // Create the runtime /// let rt = Runtime::new().unwrap(); /// /// // Spawn a future onto the runtime /// rt.spawn(async { /// println!("now running on a worker thread"); /// }); /// # } /// ``` #[track_caller] pub fn spawn<F>(&self, future: F) -> JoinHandle<F::Output> where F: Future + Send + 'static, F::Output: Send + 'static, { self.handle.spawn(future) } /// Runs the provided function on an executor dedicated to blocking operations. /// /// # Examples /// /// ``` /// use tokio::runtime::Runtime; /// /// # fn dox() { /// // Create the runtime /// let rt = Runtime::new().unwrap(); /// /// // Spawn a blocking function onto the runtime /// rt.spawn_blocking(|| { /// println!("now running on a worker thread"); /// }); /// # } /// ``` #[track_caller] pub fn spawn_blocking<F, R>(&self, func: F) -> JoinHandle<R> where F: FnOnce() -> R + Send + 'static, R: Send + 'static, { self.handle.spawn_blocking(func) } /// Runs a future to completion on the Tokio runtime. This is the /// runtime's entry point. /// /// This runs the given future on the current thread, blocking until it is /// complete, and yielding its resolved result. Any tasks or timers /// which the future spawns internally will be executed on the runtime. /// /// # Non-worker future /// /// Note that the future required by this function does not run as a /// worker. The expectation is that other tasks are spawned by the future here. /// Awaiting on other futures from the future provided here will not /// perform as fast as those spawned as workers. /// /// # Multi thread scheduler /// /// When the multi thread scheduler is used this will allow futures /// to run within the io driver and timer context of the overall runtime. /// /// Any spawned tasks will continue running after `block_on` returns. /// /// # Current thread scheduler /// /// When the current thread scheduler is enabled `block_on` /// can be called concurrently from multiple threads. The first call /// will take ownership of the io and timer drivers. This means /// other threads which do not own the drivers will hook into that one. /// When the first `block_on` completes, other threads will be able to /// "steal" the driver to allow continued execution of their futures. /// /// Any spawned tasks will be suspended after `block_on` returns. Calling /// `block_on` again will resume previously spawned tasks. /// /// # Panics /// /// This function panics if the provided future panics, or if called within an /// asynchronous execution context. /// /// # Examples /// /// ```no_run /// use tokio::runtime::Runtime; /// /// // Create the runtime /// let rt = Runtime::new().unwrap(); /// /// // Execute the future, blocking the current thread until completion /// rt.block_on(async { /// println!("hello"); /// }); /// ``` /// /// [handle]: fn@Handle::block_on #[track_caller] pub fn block_on<F: Future>(&self, future: F) -> F::
/// /// [mod]: index.html /// [main]: ../attr.main.html /// [threaded scheduler]: index.html#threaded-scheduler
random_line_split
runtime.rs
threads, you /// will be able to access the runtime from those tasks or threads. /// /// The difference between <code>[Arc]\<Runtime></code> and [`Handle`] is that /// an <code>[Arc]\<Runtime></code> will prevent the runtime from shutting down, /// whereas a [`Handle`] does not prevent that. This is because shutdown of the /// runtime happens when the destructor of the `Runtime` object runs. /// /// Calls to [`shutdown_background`] and [`shutdown_timeout`] require exclusive /// ownership of the `Runtime` type. When using an <code>[Arc]\<Runtime></code>, /// this can be achieved via [`Arc::try_unwrap`] when only one strong count /// reference is left over. /// /// The runtime context is entered using the [`Runtime::enter`] or /// [`Handle::enter`] methods, which use a thread-local variable to store the /// current runtime. Whenever you are inside the runtime context, methods such /// as [`tokio::spawn`] will use the runtime whose context you are inside. /// /// [timer]: crate::time /// [mod]: index.html /// [`new`]: method@Self::new /// [`Builder`]: struct@Builder /// [`Handle`]: struct@Handle /// [`tokio::spawn`]: crate::spawn /// [`Arc::try_unwrap`]: std::sync::Arc::try_unwrap /// [Arc]: std::sync::Arc /// [`shutdown_background`]: method@Runtime::shutdown_background /// [`shutdown_timeout`]: method@Runtime::shutdown_timeout #[derive(Debug)] pub struct Runtime { /// Task scheduler scheduler: Scheduler, /// Handle to runtime, also contains driver handles handle: Handle, /// Blocking pool handle, used to signal shutdown blocking_pool: BlockingPool, } /// The flavor of a `Runtime`. /// /// This is the return type for [`Handle::runtime_flavor`](crate::runtime::Handle::runtime_flavor()). #[derive(Debug, PartialEq, Eq)] #[non_exhaustive] pub enum RuntimeFlavor { /// The flavor that executes all tasks on the current thread. CurrentThread, /// The flavor that executes tasks across multiple threads. MultiThread, /// The flavor that executes tasks across multiple threads. #[cfg(tokio_unstable)] MultiThreadAlt, } /// The runtime scheduler is either a multi-thread or a current-thread executor. #[derive(Debug)] pub(super) enum Scheduler { /// Execute all tasks on the current-thread. CurrentThread(CurrentThread), /// Execute tasks across multiple threads. #[cfg(all(feature = "rt-multi-thread", not(target_os = "wasi")))] MultiThread(MultiThread), /// Execute tasks across multiple threads. #[cfg(all(tokio_unstable, feature = "rt-multi-thread", not(target_os = "wasi")))] MultiThreadAlt(MultiThreadAlt), } impl Runtime { pub(super) fn
( scheduler: Scheduler, handle: Handle, blocking_pool: BlockingPool, ) -> Runtime { Runtime { scheduler, handle, blocking_pool, } } cfg_not_wasi! { /// Creates a new runtime instance with default configuration values. /// /// This results in the multi threaded scheduler, I/O driver, and time driver being /// initialized. /// /// Most applications will not need to call this function directly. Instead, /// they will use the [`#[tokio::main]` attribute][main]. When a more complex /// configuration is necessary, the [runtime builder] may be used. /// /// See [module level][mod] documentation for more details. /// /// # Examples /// /// Creating a new `Runtime` with default configuration values. /// /// ``` /// use tokio::runtime::Runtime; /// /// let rt = Runtime::new() /// .unwrap(); /// /// // Use the runtime... /// ``` /// /// [mod]: index.html /// [main]: ../attr.main.html /// [threaded scheduler]: index.html#threaded-scheduler /// [runtime builder]: crate::runtime::Builder #[cfg(feature = "rt-multi-thread")] #[cfg_attr(docsrs, doc(cfg(feature = "rt-multi-thread")))] pub fn new() -> std::io::Result<Runtime> { Builder::new_multi_thread().enable_all().build() } } /// Returns a handle to the runtime's spawner. /// /// The returned handle can be used to spawn tasks that run on this runtime, and can /// be cloned to allow moving the `Handle` to other threads. /// /// Calling [`Handle::block_on`] on a handle to a `current_thread` runtime is error-prone. /// Refer to the documentation of [`Handle::block_on`] for more. /// /// # Examples /// /// ``` /// use tokio::runtime::Runtime; /// /// let rt = Runtime::new() /// .unwrap(); /// /// let handle = rt.handle(); /// /// // Use the handle... /// ``` pub fn handle(&self) -> &Handle { &self.handle } /// Spawns a future onto the Tokio runtime. /// /// This spawns the given future onto the runtime's executor, usually a /// thread pool. The thread pool is then responsible for polling the future /// until it completes. /// /// The provided future will start running in the background immediately /// when `spawn` is called, even if you don't await the returned /// `JoinHandle`. /// /// See [module level][mod] documentation for more details. /// /// [mod]: index.html /// /// # Examples /// /// ``` /// use tokio::runtime::Runtime; /// /// # fn dox() { /// // Create the runtime /// let rt = Runtime::new().unwrap(); /// /// // Spawn a future onto the runtime /// rt.spawn(async { /// println!("now running on a worker thread"); /// }); /// # } /// ``` #[track_caller] pub fn spawn<F>(&self, future: F) -> JoinHandle<F::Output> where F: Future + Send + 'static, F::Output: Send + 'static, { self.handle.spawn(future) } /// Runs the provided function on an executor dedicated to blocking operations. /// /// # Examples /// /// ``` /// use tokio::runtime::Runtime; /// /// # fn dox() { /// // Create the runtime /// let rt = Runtime::new().unwrap(); /// /// // Spawn a blocking function onto the runtime /// rt.spawn_blocking(|| { /// println!("now running on a worker thread"); /// }); /// # } /// ``` #[track_caller] pub fn spawn_blocking<F, R>(&self, func: F) -> JoinHandle<R> where F: FnOnce() -> R + Send + 'static, R: Send + 'static, { self.handle.spawn_blocking(func) } /// Runs a future to completion on the Tokio runtime. This is the /// runtime's entry point. /// /// This runs the given future on the current thread, blocking until it is /// complete, and yielding its resolved result. Any tasks or timers /// which the future spawns internally will be executed on the runtime. /// /// # Non-worker future /// /// Note that the future required by this function does not run as a /// worker. The expectation is that other tasks are spawned by the future here. /// Awaiting on other futures from the future provided here will not /// perform as fast as those spawned as workers. /// /// # Multi thread scheduler /// /// When the multi thread scheduler is used this will allow futures /// to run within the io driver and timer context of the overall runtime. /// /// Any spawned tasks will continue running after `block_on` returns. /// /// # Current thread scheduler /// /// When the current thread scheduler is enabled `block_on` /// can be called concurrently from multiple threads. The first call /// will take ownership of the io and timer drivers. This means /// other threads which do not own the drivers will hook into that one. /// When the first `block_on` completes, other threads will be able to /// "steal" the driver to allow continued execution of their futures. /// /// Any spawned tasks will be suspended after `block_on` returns. Calling /// `block_on` again will resume previously spawned tasks. /// /// # Panics /// /// This function panics if the provided future panics, or if called within an /// asynchronous execution context. /// /// # Examples /// /// ```no_run /// use tokio::runtime::Runtime; /// /// // Create the runtime /// let rt = Runtime::new().unwrap(); /// /// // Execute the future, blocking the current thread until completion /// rt.block_on(async { /// println!("hello"); /// }); /// ``` /// /// [handle]: fn@Handle::block_on #[track_caller] pub fn block_on<F: Future>(&self, future: F) -> F
from_parts
identifier_name
runtime.rs
/// // Spawn a future onto the runtime /// rt.spawn(async { /// println!("now running on a worker thread"); /// }); /// # } /// ``` #[track_caller] pub fn spawn<F>(&self, future: F) -> JoinHandle<F::Output> where F: Future + Send + 'static, F::Output: Send + 'static, { self.handle.spawn(future) } /// Runs the provided function on an executor dedicated to blocking operations. /// /// # Examples /// /// ``` /// use tokio::runtime::Runtime; /// /// # fn dox() { /// // Create the runtime /// let rt = Runtime::new().unwrap(); /// /// // Spawn a blocking function onto the runtime /// rt.spawn_blocking(|| { /// println!("now running on a worker thread"); /// }); /// # } /// ``` #[track_caller] pub fn spawn_blocking<F, R>(&self, func: F) -> JoinHandle<R> where F: FnOnce() -> R + Send + 'static, R: Send + 'static, { self.handle.spawn_blocking(func) } /// Runs a future to completion on the Tokio runtime. This is the /// runtime's entry point. /// /// This runs the given future on the current thread, blocking until it is /// complete, and yielding its resolved result. Any tasks or timers /// which the future spawns internally will be executed on the runtime. /// /// # Non-worker future /// /// Note that the future required by this function does not run as a /// worker. The expectation is that other tasks are spawned by the future here. /// Awaiting on other futures from the future provided here will not /// perform as fast as those spawned as workers. /// /// # Multi thread scheduler /// /// When the multi thread scheduler is used this will allow futures /// to run within the io driver and timer context of the overall runtime. /// /// Any spawned tasks will continue running after `block_on` returns. /// /// # Current thread scheduler /// /// When the current thread scheduler is enabled `block_on` /// can be called concurrently from multiple threads. The first call /// will take ownership of the io and timer drivers. This means /// other threads which do not own the drivers will hook into that one. /// When the first `block_on` completes, other threads will be able to /// "steal" the driver to allow continued execution of their futures. /// /// Any spawned tasks will be suspended after `block_on` returns. Calling /// `block_on` again will resume previously spawned tasks. /// /// # Panics /// /// This function panics if the provided future panics, or if called within an /// asynchronous execution context. /// /// # Examples /// /// ```no_run /// use tokio::runtime::Runtime; /// /// // Create the runtime /// let rt = Runtime::new().unwrap(); /// /// // Execute the future, blocking the current thread until completion /// rt.block_on(async { /// println!("hello"); /// }); /// ``` /// /// [handle]: fn@Handle::block_on #[track_caller] pub fn block_on<F: Future>(&self, future: F) -> F::Output { #[cfg(all( tokio_unstable, tokio_taskdump, feature = "rt", target_os = "linux", any(target_arch = "aarch64", target_arch = "x86", target_arch = "x86_64") ))] let future = super::task::trace::Trace::root(future); #[cfg(all(tokio_unstable, feature = "tracing"))] let future = crate::util::trace::task( future, "block_on", None, crate::runtime::task::Id::next().as_u64(), ); let _enter = self.enter(); match &self.scheduler { Scheduler::CurrentThread(exec) => exec.block_on(&self.handle.inner, future), #[cfg(all(feature = "rt-multi-thread", not(target_os = "wasi")))] Scheduler::MultiThread(exec) => exec.block_on(&self.handle.inner, future), #[cfg(all(tokio_unstable, feature = "rt-multi-thread", not(target_os = "wasi")))] Scheduler::MultiThreadAlt(exec) => exec.block_on(&self.handle.inner, future), } } /// Enters the runtime context. /// /// This allows you to construct types that must have an executor /// available on creation such as [`Sleep`] or [`TcpStream`]. It will /// also allow you to call methods such as [`tokio::spawn`]. /// /// [`Sleep`]: struct@crate::time::Sleep /// [`TcpStream`]: struct@crate::net::TcpStream /// [`tokio::spawn`]: fn@crate::spawn /// /// # Example /// /// ``` /// use tokio::runtime::Runtime; /// /// fn function_that_spawns(msg: String) { /// // Had we not used `rt.enter` below, this would panic. /// tokio::spawn(async move { /// println!("{}", msg); /// }); /// } /// /// fn main() { /// let rt = Runtime::new().unwrap(); /// /// let s = "Hello World!".to_string(); /// /// // By entering the context, we tie `tokio::spawn` to this executor. /// let _guard = rt.enter(); /// function_that_spawns(s); /// } /// ``` pub fn enter(&self) -> EnterGuard<'_> { self.handle.enter() } /// Shuts down the runtime, waiting for at most `duration` for all spawned /// work to stop. /// /// See the [struct level documentation](Runtime#shutdown) for more details. /// /// # Examples /// /// ``` /// use tokio::runtime::Runtime; /// use tokio::task; /// /// use std::thread; /// use std::time::Duration; /// /// fn main() { /// let runtime = Runtime::new().unwrap(); /// /// runtime.block_on(async move { /// task::spawn_blocking(move || { /// thread::sleep(Duration::from_secs(10_000)); /// }); /// }); /// /// runtime.shutdown_timeout(Duration::from_millis(100)); /// } /// ``` pub fn shutdown_timeout(mut self, duration: Duration) { // Wakeup and shutdown all the worker threads self.handle.inner.shutdown(); self.blocking_pool.shutdown(Some(duration)); } /// Shuts down the runtime, without waiting for any spawned work to stop. /// /// This can be useful if you want to drop a runtime from within another runtime. /// Normally, dropping a runtime will block indefinitely for spawned blocking tasks /// to complete, which would normally not be permitted within an asynchronous context. /// By calling `shutdown_background()`, you can drop the runtime from such a context. /// /// Note however, that because we do not wait for any blocking tasks to complete, this /// may result in a resource leak (in that any blocking tasks are still running until they /// return. /// /// See the [struct level documentation](Runtime#shutdown) for more details. /// /// This function is equivalent to calling `shutdown_timeout(Duration::from_nanos(0))`. /// /// ``` /// use tokio::runtime::Runtime; /// /// fn main() { /// let runtime = Runtime::new().unwrap(); /// /// runtime.block_on(async move { /// let inner_runtime = Runtime::new().unwrap(); /// // ... /// inner_runtime.shutdown_background(); /// }); /// } /// ``` pub fn shutdown_background(self) { self.shutdown_timeout(Duration::from_nanos(0)) } } #[allow(clippy::single_match)] // there are comments in the error branch, so we don't want if-let impl Drop for Runtime { fn drop(&mut self)
{ match &mut self.scheduler { Scheduler::CurrentThread(current_thread) => { // This ensures that tasks spawned on the current-thread // runtime are dropped inside the runtime's context. let _guard = context::try_set_current(&self.handle.inner); current_thread.shutdown(&self.handle.inner); } #[cfg(all(feature = "rt-multi-thread", not(target_os = "wasi")))] Scheduler::MultiThread(multi_thread) => { // The threaded scheduler drops its tasks on its worker threads, which is // already in the runtime's context. multi_thread.shutdown(&self.handle.inner); } #[cfg(all(tokio_unstable, feature = "rt-multi-thread", not(target_os = "wasi")))] Scheduler::MultiThreadAlt(multi_thread) => { // The threaded scheduler drops its tasks on its worker threads, which is // already in the runtime's context. multi_thread.shutdown(&self.handle.inner); }
identifier_body
multiuser.go
MultiDB) systemDBFile() string { return filepath.Join(m.BaseDir(), "system.sqlite") } func validUserName(name string) bool { var validUser = regexp.MustCompile(`^\p{L}+[_0-9\p{L}]*$`) return validUser.MatchString(name) } func validDir(dir string) bool { if _, err := os.Stat(dir); os.IsNotExist(err) { return false } return true } // CreateDirIfNotExists creates a directory including all subdirectories needed, // or returns an error. func CreateDirIfNotExist(dir string) error { if _, err := os.Stat(dir); os.IsNotExist(err) { err = os.MkdirAll(dir, 0755) if err != nil { return err } } return nil } func validateUser(name string, basedir string) error { if !validUserName(name) { return Fail(`invalid user name "%s"`, name) } if !validDir(basedir) { return Fail(`the base directory for user "%s" does not exist: %s`, name, basedir) } src, err := os.Stat(basedir) if err != nil { return err } if !src.IsDir() { return Fail(`not a directory: %s`, basedir) } return nil } // ErrCode types represent errors instead of error structures. type ErrCode int // Error codes returned by the functions. const ( ErrAuthenticationFailed ErrCode = iota + 1 // User authentication has failed (wrong password). OK // No error has occured. ErrUsernameInUse // The user name is already being used. ErrEmailInUse // The email is already being used. ErrCryptoRandFailure // The random number generator has failed. ErrInvalidParams // One or more parameters were invalid. ErrUnknownUser // The user is not known. ErrNotEnoughSalt // Insufficiently long salt has been supplied. ErrInvalidUser // The user name or email is invalid. ErrDBClosed // The internal housekeeping DB is locked, corrupted, or closed. ErrDBFail // A database operation has failed. ErrFileSystem // A directory or file could not be created. ErrNoHome // The user's DB home directory does not exist. ErrCloseFailed // Could not close the user database. ErrOpenFailed // Could not open the user database. ErrPackFail // Compressing user data failed. ErrInvalidKey // A given salted key is invalid (either nil, or other problems). ErrTransactionFail // Could not perform op because of a failed transaction. ) func (m *MultiDB) isExisting(field, query string) bool { q, _ := ParseQuery(fmt.Sprintf("User %s=%s", field, query)) results, err := m.system.Find(q, 1) if err != nil || len(results) < 1 { return false } return true } func (m *MultiDB)
(username string) Item { q, err := ParseQuery(fmt.Sprintf("User Username=%s", username)) if err != nil { return 0 } results, err := m.system.Find(q, 1) if err != nil || len(results) != 1 { return 0 } return results[0] } // ExistingUser returns true if a user with the given user name exists, false otherwise. func (m *MultiDB) ExistingUser(username string) bool { result := m.isExisting("Username", username) return result } // ExistingEmail returns true if a user with this email address exists, false otherwise. func (m *MultiDB) ExistingEmail(email string) bool { result := m.isExisting("Email", email) return result } // NewUser creates a new user with given username, email, and password. Based on a strong // salt that is only used internally and the Argon2 algorithm with the given parameters // an internal key is created and stored in an internal database. The user and OK are returned // unless an error has occurred. The integer returned is a numeric error code to make it easier to distinguish // certain cases: EmailInUse - the email has already been registered, UsernameInUse - a user with the same // user name has already been registered. Both emails and usernames must be unique and cannot be // registered twice. func (m *MultiDB) NewUser(username, email string, key *saltedKey) (*User, ErrCode, error) { // validate inputs if err := validateUser(username, m.BaseDir()); err != nil { return nil, ErrInvalidUser, err } reply, err := key.validate() if err != nil || reply != OK { return nil, reply, err } user := User{name: username} if m.system == nil { return nil, ErrDBClosed, Fail(`internal DB is nil`) } // check if user and email exist if m.ExistingUser(username) { return nil, ErrUsernameInUse, Fail(`user "%s" already exists!`, username) } if m.ExistingEmail(email) { return nil, ErrEmailInUse, Fail(`email "%s" is already in use!`, email) } // now start adding the user user.id, err = m.system.NewItem("User") if err != nil { return nil, ErrDBFail, err } tx, err := m.Begin() if err != nil { return nil, ErrTransactionFail, err } if err := tx.Set("User", user.id, "Username", []Value{NewString(username)}); err != nil { return nil, ErrDBFail, err } if err := tx.Set("User", user.id, "Email", []Value{NewString(email)}); err != nil { return nil, ErrDBFail, err } salt := make([]byte, key.p.InternalSaltLength) n, err := rand.Read(salt) if uint32(n) != key.p.InternalSaltLength || err != nil { return nil, ErrCryptoRandFailure, Fail(`random number generator failed to generate salt`) } if err := tx.Set("User", user.id, "InternalSalt", []Value{NewBytes(salt)}); err != nil { return nil, ErrDBFail, Fail(`could not store salt in multiuser database: %s`, err) } realkey := argon2.IDKey(key.pwd, salt, key.p.Argon2Iterations, key.p.Argon2Memory, key.p.Argon2Parallelism, key.p.KeyLength) if err := tx.Set("User", user.id, "Key", []Value{NewBytes(realkey)}); err != nil { return nil, ErrDBFail, Fail(`could not store key in multiuser database: %s`, err) } if err := tx.Set("User", user.id, "ExternalSalt", []Value{NewBytes(key.sel)}); err != nil { return nil, ErrDBFail, Fail(`could not store the external salt in multiuser database: %s`, err) } now := NewDate(time.Now()) if err := tx.Set("User", user.id, "Created", []Value{now}); err != nil { return nil, ErrDBFail, err } if err := tx.Set("User", user.id, "Modified", []Value{now}); err != nil { return nil, ErrDBFail, err } if err := tx.Commit(); err != nil { return nil, ErrDBFail, Fail(`multiuser database error: %s`, err) } dirpath := m.UserDir(&user) err = CreateDirIfNotExist(dirpath) if err != nil { return nil, ErrFileSystem, err } return &user, OK, nil } // ExternalSalt is the salt associated with a user. It is stored in the database and may // be used for hashing the password prior to authentication. The external salt is not used // for internal key derivation. func (m *MultiDB) ExternalSalt(username string) ([]byte, ErrCode, error) { if !m.ExistingUser(username) { return nil, ErrUnknownUser, Fail(`unknown user "%s"`, username) } tx, err := m.Begin() if err != nil { return nil, ErrTransactionFail, Fail(`transaction failed: %s`, err) } defer tx.Rollback() id := m.userID(username) if id == 0 { return nil, ErrUnknownUser, Fail(`unknown user "%s"`, username) } result, err := m.system.Get("User", id, "ExternalSalt") if err != nil || len(result) != 1 { return nil, ErrNotEnoughSalt, Fail(`user "%s" salt not found, the user database might be corrupted`, username) } err = tx.Commit() if err != nil { return nil, ErrTransactionFail, Fail(`transaction commit failed: %s`, err) } return result[0].Bytes(), OK, nil } // GenerateExternalSalt returns some new external salt of the length specified in params. // This salt should be passed to NewUser and can be used for passphrase hashing prior to // calling NewUser. It is stored in the user database and can be retrieved as ExternalSalt. func GenerateExternalSalt(params *Params) []byte { salt := make([]byte, params.ExternalSaltLength) n, err := rand.Read(salt) if err
userID
identifier_name
multiuser.go
ErrDBFail, err } if err := tx.Set("User", user.id, "Modified", []Value{now}); err != nil { return nil, ErrDBFail, err } if err := tx.Commit(); err != nil { return nil, ErrDBFail, Fail(`multiuser database error: %s`, err) } dirpath := m.UserDir(&user) err = CreateDirIfNotExist(dirpath) if err != nil { return nil, ErrFileSystem, err } return &user, OK, nil } // ExternalSalt is the salt associated with a user. It is stored in the database and may // be used for hashing the password prior to authentication. The external salt is not used // for internal key derivation. func (m *MultiDB) ExternalSalt(username string) ([]byte, ErrCode, error) { if !m.ExistingUser(username) { return nil, ErrUnknownUser, Fail(`unknown user "%s"`, username) } tx, err := m.Begin() if err != nil { return nil, ErrTransactionFail, Fail(`transaction failed: %s`, err) } defer tx.Rollback() id := m.userID(username) if id == 0 { return nil, ErrUnknownUser, Fail(`unknown user "%s"`, username) } result, err := m.system.Get("User", id, "ExternalSalt") if err != nil || len(result) != 1 { return nil, ErrNotEnoughSalt, Fail(`user "%s" salt not found, the user database might be corrupted`, username) } err = tx.Commit() if err != nil { return nil, ErrTransactionFail, Fail(`transaction commit failed: %s`, err) } return result[0].Bytes(), OK, nil } // GenerateExternalSalt returns some new external salt of the length specified in params. // This salt should be passed to NewUser and can be used for passphrase hashing prior to // calling NewUser. It is stored in the user database and can be retrieved as ExternalSalt. func GenerateExternalSalt(params *Params) []byte { salt := make([]byte, params.ExternalSaltLength) n, err := rand.Read(salt) if err != nil || uint32(n) < params.ExternalSaltLength { return nil } return salt } type saltedKey struct { pwd []byte sel []byte p *Params } func (key *saltedKey) validate() (ErrCode, error) { if key == nil { return ErrInvalidKey, Fail(`key is nil`) } if key.p == nil { return ErrInvalidParams, Fail(`key parameters are nil`) } if key.pwd == nil { return ErrInvalidKey, Fail(`key password is empty`) } if key.sel == nil { return ErrNotEnoughSalt, Fail(`key salt is nil`) } if !key.p.validate() { return ErrInvalidParams, Fail(`invalid parameters`) } if uint32(len(key.sel)) < key.p.ExternalSaltLength { return ErrNotEnoughSalt, Fail(`external key salt length is less than required by key params`) } return OK, nil } // GenerateKey takes a password and some salt, and generates a salted key of length 64 bytes. // Use the ExternalSalt as salt and the original, unaltered password. The function // use Blake2b-512 for key derivation. func GenerateKey(password string, salt []byte, params *Params) *saltedKey { unsalted := blake2b.Sum512([]byte(password)) salted := saltedKey{pwd: append(salt, unsalted[:]...), sel: salt, p: params} return &salted } // Authenticate a user by given name and salted password. // Returns the user and OK if successful, otherwise nil, a numeric error code and the error. // Notice that the external salt is not passed to this function. Instead, the password string // should have been prepared (securely hashed, whitened, etc.) before calling this function // on the basis of the user's ExternalSalt. func (m *MultiDB) Authenticate(username string, key *saltedKey) (*User, ErrCode, error) { if err := validateUser(username, m.BaseDir()); err != nil { return nil, ErrInvalidUser, err } if !m.ExistingUser(username) { return nil, ErrUnknownUser, Fail(`user "%s" does not exist`, username) } reply, err := key.validate() if err != nil || reply != OK { return nil, reply, err } user := User{name: username} dirpath := m.UserDir(&user) if _, err := os.Stat(dirpath); os.IsNotExist(err) { return nil, ErrNoHome, Fail(`user "%s" home directory does not exist: %s`, username, dirpath) } user.id = m.userID(username) if user.id == 0 { return nil, ErrUnknownUser, Fail(`user "%s" does not exist`, username) } // get the strong salt and hash with it using argon2, compare to stored key result, err := m.system.Get("User", user.id, "InternalSalt") if err != nil || len(result) != 1 { return nil, ErrNotEnoughSalt, Fail(`user "%s"'s internal salt was not found, the user database might be corrupted`, username) } salt := result[0].Bytes() if len(salt) != int(key.p.InternalSaltLength) { return nil, ErrInvalidParams, Fail(`invalid params, user "%s"'s internal salt length does not match internal salt length in params, given %d, expected %d`, username, len(salt), key.p.InternalSaltLength) } keyA := argon2.IDKey(key.pwd, salt, key.p.Argon2Iterations, key.p.Argon2Memory, key.p.Argon2Parallelism, key.p.KeyLength) keyresult, err := m.system.Get("User", user.id, "Key") if err != nil || len(keyresult) != 1 { return nil, ErrAuthenticationFailed, Fail(`user "%s" key was not found in user database, the database might be corrupted`, username) } keyB := keyresult[0].Bytes() if !bytes.Equal(keyA, keyB) { return nil, ErrAuthenticationFailed, Fail(`authentication failure`) } return &user, OK, nil } // Close the MultiDB, closing the internal housekeeping and all open user databases. func (m *MultiDB) Close() (ErrCode, error) { errcount := 0 s := "" for _, v := range m.userdbs { if v != nil { err := v.Close() if err != nil { s = fmt.Sprintf("%s, %s", s, err.Error()) errcount++ } } } for k := range m.userdbs { delete(m.userdbs, k) } if err := m.system.Close(); err != nil { s = fmt.Sprintf("%s, %s", s, err.Error()) errcount++ } if errcount > 0 { return ErrCloseFailed, Fail(`errors closing multi user DB: %s`, s) } return OK, nil } // UserDB returns the database of the given user. func (m *MultiDB) UserDB(user *User) (*MDB, ErrCode, error) { var err error if user.id == 0 { return nil, ErrUnknownUser, Fail(`user "%s" does not exist`, user.name) } db := m.userdbs[user.id] if db == nil { db, err = Open(m.driver, m.userDBFile(user)) if err != nil { return nil, ErrOpenFailed, err } } return db, OK, nil } // DeleteUserContent deletes a user's content in the multiuser database, i.e., all the user data. // This action cannot be undone. func (m *MultiDB) DeleteUserContent(user *User) (ErrCode, error) { db, _, _ := m.UserDB(user) db.Close() if err := removeContents(m.UserDir(user)); err != nil { return ErrFileSystem, err } return OK, nil } // DeleteUser deletes a user and all associated user content from a multiuser database. func (m *MultiDB) DeleteUser(user *User) (ErrCode, error) { tx, err := m.Begin() if err != nil { return ErrTransactionFail, err } defer tx.Rollback() if err := tx.RemoveItem("User", user.ID()); err != nil { return ErrDBFail, err } errcode, err := m.DeleteUserContent(user) delete(m.userdbs, user.ID()) if err != nil { return errcode, err } err = tx.Commit() if err != nil { return ErrTransactionFail, err } return OK, nil } func removeContents(dir string) error { d, err := os.Open(dir) if err != nil { return err } defer d.Close() names, err := d.Readdirnames(-1) if err != nil { return err } for _, name := range names
{ err = os.RemoveAll(filepath.Join(dir, name)) if err != nil { return err } }
conditional_block
multiuser.go
// BaseDir returns the base directory of the multiuser database. This directory contains databases // for all users. func (m *MultiDB) BaseDir() string { return m.basepath } func (m *MultiDB) userFile(user *User, file string) string { return filepath.Join(m.UserDir(user), file) } func (m *MultiDB) userDBFile(user *User) string { return m.userFile(user, "data.sqlite") } func (m *MultiDB) systemDBFile() string { return filepath.Join(m.BaseDir(), "system.sqlite") } func validUserName(name string) bool { var validUser = regexp.MustCompile(`^\p{L}+[_0-9\p{L}]*$`) return validUser.MatchString(name) } func validDir(dir string) bool { if _, err := os.Stat(dir); os.IsNotExist(err) { return false } return true } // CreateDirIfNotExists creates a directory including all subdirectories needed, // or returns an error. func CreateDirIfNotExist(dir string) error { if _, err := os.Stat(dir); os.IsNotExist(err) { err = os.MkdirAll(dir, 0755) if err != nil { return err } } return nil } func validateUser(name string, basedir string) error { if !validUserName(name) { return Fail(`invalid user name "%s"`, name) } if !validDir(basedir) { return Fail(`the base directory for user "%s" does not exist: %s`, name, basedir) } src, err := os.Stat(basedir) if err != nil { return err } if !src.IsDir() { return Fail(`not a directory: %s`, basedir) } return nil } // ErrCode types represent errors instead of error structures. type ErrCode int // Error codes returned by the functions. const ( ErrAuthenticationFailed ErrCode = iota + 1 // User authentication has failed (wrong password). OK // No error has occured. ErrUsernameInUse // The user name is already being used. ErrEmailInUse // The email is already being used. ErrCryptoRandFailure // The random number generator has failed. ErrInvalidParams // One or more parameters were invalid. ErrUnknownUser // The user is not known. ErrNotEnoughSalt // Insufficiently long salt has been supplied. ErrInvalidUser // The user name or email is invalid. ErrDBClosed // The internal housekeeping DB is locked, corrupted, or closed. ErrDBFail // A database operation has failed. ErrFileSystem // A directory or file could not be created. ErrNoHome // The user's DB home directory does not exist. ErrCloseFailed // Could not close the user database. ErrOpenFailed // Could not open the user database. ErrPackFail // Compressing user data failed. ErrInvalidKey // A given salted key is invalid (either nil, or other problems). ErrTransactionFail // Could not perform op because of a failed transaction. ) func (m *MultiDB) isExisting(field, query string) bool { q, _ := ParseQuery(fmt.Sprintf("User %s=%s", field, query)) results, err := m.system.Find(q, 1) if err != nil || len(results) < 1 { return false } return true } func (m *MultiDB) userID(username string) Item { q, err := ParseQuery(fmt.Sprintf("User Username=%s", username)) if err != nil { return 0 } results, err := m.system.Find(q, 1) if err != nil || len(results) != 1 { return 0 } return results[0] } // ExistingUser returns true if a user with the given user name exists, false otherwise. func (m *MultiDB) ExistingUser(username string) bool { result := m.isExisting("Username", username) return result } // ExistingEmail returns true if a user with this email address exists, false otherwise. func (m *MultiDB) ExistingEmail(email string) bool { result := m.isExisting("Email", email) return result } // NewUser creates a new user with given username, email, and password. Based on a strong // salt that is only used internally and the Argon2 algorithm with the given parameters // an internal key is created and stored in an internal database. The user and OK are returned // unless an error has occurred. The integer returned is a numeric error code to make it easier to distinguish // certain cases: EmailInUse - the email has already been registered, UsernameInUse - a user with the same // user name has already been registered. Both emails and usernames must be unique and cannot be // registered twice. func (m *MultiDB) NewUser(username, email string, key *saltedKey) (*User, ErrCode, error) { // validate inputs if err := validateUser(username, m.BaseDir()); err != nil { return nil, ErrInvalidUser, err } reply, err := key.validate() if err != nil || reply != OK { return nil, reply, err } user := User{name: username} if m.system == nil { return nil, ErrDBClosed, Fail(`internal DB is nil`) } // check if user and email exist if m.ExistingUser(username) { return nil, ErrUsernameInUse, Fail(`user "%s" already exists!`, username) } if m.ExistingEmail(email) { return nil, ErrEmailInUse, Fail(`email "%s" is already in use!`, email) } // now start adding the user user.id, err = m.system.NewItem("User") if err != nil { return nil, ErrDBFail, err } tx, err := m.Begin() if err != nil { return nil, ErrTransactionFail, err } if err := tx.Set("User", user.id, "Username", []Value{NewString(username)}); err != nil { return nil, ErrDBFail, err } if err := tx.Set("User", user.id, "Email", []Value{NewString(email)}); err != nil { return nil, ErrDBFail, err } salt := make([]byte, key.p.InternalSaltLength) n, err := rand.Read(salt) if uint32(n) != key.p.InternalSaltLength || err != nil { return nil, ErrCryptoRandFailure, Fail(`random number generator failed to generate salt`) } if err := tx.Set("User", user.id, "InternalSalt", []Value{NewBytes(salt)}); err != nil { return nil, ErrDBFail, Fail(`could not store salt in multiuser database: %s`, err) } realkey := argon2.IDKey(key.pwd, salt, key.p.Argon2Iterations, key.p.Argon2Memory, key.p.Argon2Parallelism, key.p.KeyLength) if err := tx.Set("User", user.id, "Key", []Value{NewBytes(realkey)}); err != nil { return nil, ErrDBFail, Fail(`could not store key in multiuser database: %s`, err) } if err := tx.Set("User", user.id, "ExternalSalt", []Value{NewBytes(key.sel)}); err != nil { return nil, ErrDBFail, Fail(`could not store the external salt in multiuser database: %s`, err) } now := NewDate(time.Now()) if err := tx.Set("User", user.id, "Created", []Value{now}); err != nil { return nil, ErrDBFail, err } if err := tx.Set("User", user.id, "Modified", []Value{now}); err != nil { return nil, ErrDBFail, err } if err := tx.Commit(); err != nil { return nil, ErrDBFail, Fail(`multiuser database error: %s`, err) } dirpath := m.UserDir(&user) err = CreateDirIfNotExist(dirpath) if err != nil { return nil, ErrFileSystem, err } return &user, OK, nil } // ExternalSalt is the salt associated with a user. It is stored in the database and may // be used for hashing the password prior to authentication. The external salt is not used // for internal key derivation. func (m *MultiDB) ExternalSalt(username string) ([]byte, ErrCode, error) { if !m.ExistingUser(username) { return nil, ErrUnknownUser, Fail(`unknown user "%s"`, username) } tx, err := m.Begin() if err != nil { return nil, ErrTransactionFail, Fail(`transaction failed: %s`, err) } defer tx.Rollback() id := m.userID(username) if id == 0 { return nil, ErrUnknownUser, Fail(`unknown user "%s"`, username) } result, err := m.system.Get("User", id, "ExternalSalt") if err != nil || len(result) != 1 { return nil, ErrNotEnoughSalt, Fail(`user "%s" salt not found, the user database might be corrupted`, username) } err = tx.Commit() if err != nil { return nil, ErrTransactionFail, Fail(`
{ return filepath.Join(m.basepath, user.name) }
identifier_body
multiuser.go
) } now := NewDate(time.Now()) if err := tx.Set("User", user.id, "Created", []Value{now}); err != nil { return nil, ErrDBFail, err } if err := tx.Set("User", user.id, "Modified", []Value{now}); err != nil { return nil, ErrDBFail, err } if err := tx.Commit(); err != nil { return nil, ErrDBFail, Fail(`multiuser database error: %s`, err) } dirpath := m.UserDir(&user) err = CreateDirIfNotExist(dirpath) if err != nil { return nil, ErrFileSystem, err } return &user, OK, nil } // ExternalSalt is the salt associated with a user. It is stored in the database and may // be used for hashing the password prior to authentication. The external salt is not used // for internal key derivation. func (m *MultiDB) ExternalSalt(username string) ([]byte, ErrCode, error) { if !m.ExistingUser(username) { return nil, ErrUnknownUser, Fail(`unknown user "%s"`, username) } tx, err := m.Begin() if err != nil { return nil, ErrTransactionFail, Fail(`transaction failed: %s`, err) } defer tx.Rollback() id := m.userID(username) if id == 0 { return nil, ErrUnknownUser, Fail(`unknown user "%s"`, username) } result, err := m.system.Get("User", id, "ExternalSalt") if err != nil || len(result) != 1 { return nil, ErrNotEnoughSalt, Fail(`user "%s" salt not found, the user database might be corrupted`, username) } err = tx.Commit() if err != nil { return nil, ErrTransactionFail, Fail(`transaction commit failed: %s`, err) } return result[0].Bytes(), OK, nil } // GenerateExternalSalt returns some new external salt of the length specified in params. // This salt should be passed to NewUser and can be used for passphrase hashing prior to // calling NewUser. It is stored in the user database and can be retrieved as ExternalSalt. func GenerateExternalSalt(params *Params) []byte { salt := make([]byte, params.ExternalSaltLength) n, err := rand.Read(salt) if err != nil || uint32(n) < params.ExternalSaltLength { return nil } return salt } type saltedKey struct { pwd []byte sel []byte p *Params } func (key *saltedKey) validate() (ErrCode, error) { if key == nil { return ErrInvalidKey, Fail(`key is nil`) } if key.p == nil { return ErrInvalidParams, Fail(`key parameters are nil`) } if key.pwd == nil { return ErrInvalidKey, Fail(`key password is empty`) } if key.sel == nil { return ErrNotEnoughSalt, Fail(`key salt is nil`) } if !key.p.validate() { return ErrInvalidParams, Fail(`invalid parameters`) } if uint32(len(key.sel)) < key.p.ExternalSaltLength { return ErrNotEnoughSalt, Fail(`external key salt length is less than required by key params`) } return OK, nil } // GenerateKey takes a password and some salt, and generates a salted key of length 64 bytes. // Use the ExternalSalt as salt and the original, unaltered password. The function // use Blake2b-512 for key derivation. func GenerateKey(password string, salt []byte, params *Params) *saltedKey { unsalted := blake2b.Sum512([]byte(password)) salted := saltedKey{pwd: append(salt, unsalted[:]...), sel: salt, p: params} return &salted } // Authenticate a user by given name and salted password. // Returns the user and OK if successful, otherwise nil, a numeric error code and the error. // Notice that the external salt is not passed to this function. Instead, the password string // should have been prepared (securely hashed, whitened, etc.) before calling this function // on the basis of the user's ExternalSalt. func (m *MultiDB) Authenticate(username string, key *saltedKey) (*User, ErrCode, error) { if err := validateUser(username, m.BaseDir()); err != nil { return nil, ErrInvalidUser, err } if !m.ExistingUser(username) { return nil, ErrUnknownUser, Fail(`user "%s" does not exist`, username) } reply, err := key.validate() if err != nil || reply != OK { return nil, reply, err } user := User{name: username} dirpath := m.UserDir(&user) if _, err := os.Stat(dirpath); os.IsNotExist(err) { return nil, ErrNoHome, Fail(`user "%s" home directory does not exist: %s`, username, dirpath) } user.id = m.userID(username) if user.id == 0 { return nil, ErrUnknownUser, Fail(`user "%s" does not exist`, username) } // get the strong salt and hash with it using argon2, compare to stored key result, err := m.system.Get("User", user.id, "InternalSalt") if err != nil || len(result) != 1 { return nil, ErrNotEnoughSalt, Fail(`user "%s"'s internal salt was not found, the user database might be corrupted`, username) } salt := result[0].Bytes() if len(salt) != int(key.p.InternalSaltLength) { return nil, ErrInvalidParams, Fail(`invalid params, user "%s"'s internal salt length does not match internal salt length in params, given %d, expected %d`, username, len(salt), key.p.InternalSaltLength) } keyA := argon2.IDKey(key.pwd, salt, key.p.Argon2Iterations, key.p.Argon2Memory, key.p.Argon2Parallelism, key.p.KeyLength) keyresult, err := m.system.Get("User", user.id, "Key") if err != nil || len(keyresult) != 1 { return nil, ErrAuthenticationFailed, Fail(`user "%s" key was not found in user database, the database might be corrupted`, username) } keyB := keyresult[0].Bytes() if !bytes.Equal(keyA, keyB) { return nil, ErrAuthenticationFailed, Fail(`authentication failure`) } return &user, OK, nil } // Close the MultiDB, closing the internal housekeeping and all open user databases. func (m *MultiDB) Close() (ErrCode, error) { errcount := 0 s := "" for _, v := range m.userdbs { if v != nil { err := v.Close() if err != nil { s = fmt.Sprintf("%s, %s", s, err.Error()) errcount++ } } } for k := range m.userdbs { delete(m.userdbs, k) } if err := m.system.Close(); err != nil { s = fmt.Sprintf("%s, %s", s, err.Error()) errcount++ } if errcount > 0 { return ErrCloseFailed, Fail(`errors closing multi user DB: %s`, s) } return OK, nil } // UserDB returns the database of the given user. func (m *MultiDB) UserDB(user *User) (*MDB, ErrCode, error) { var err error if user.id == 0 { return nil, ErrUnknownUser, Fail(`user "%s" does not exist`, user.name) } db := m.userdbs[user.id] if db == nil { db, err = Open(m.driver, m.userDBFile(user)) if err != nil { return nil, ErrOpenFailed, err } } return db, OK, nil } // DeleteUserContent deletes a user's content in the multiuser database, i.e., all the user data. // This action cannot be undone. func (m *MultiDB) DeleteUserContent(user *User) (ErrCode, error) { db, _, _ := m.UserDB(user) db.Close() if err := removeContents(m.UserDir(user)); err != nil { return ErrFileSystem, err } return OK, nil } // DeleteUser deletes a user and all associated user content from a multiuser database. func (m *MultiDB) DeleteUser(user *User) (ErrCode, error) { tx, err := m.Begin() if err != nil { return ErrTransactionFail, err } defer tx.Rollback() if err := tx.RemoveItem("User", user.ID()); err != nil { return ErrDBFail, err } errcode, err := m.DeleteUserContent(user) delete(m.userdbs, user.ID()) if err != nil { return errcode, err } err = tx.Commit() if err != nil { return ErrTransactionFail, err } return OK, nil } func removeContents(dir string) error { d, err := os.Open(dir) if err != nil { return err } defer d.Close() names, err := d.Readdirnames(-1)
if err != nil { return err
random_line_split
B-Server.go
.Println("| Broadcast Program |") fmt.Println("|_____________________________|\n") } func main() { // 0. Inicializando servicos adicionais; raven.SetDSN("https://[email protected]/1831452") // 0.1. blabla2 raven.SetDefaultLoggerName("saulopinedo") raven.SetDebug(true) raven.SetEnvironment("staging") raven.SetRelease("Xiaomi") raven.SetSampleRate(1.0) // 1. Pre-definindo as variaveis fundamentais; clientCount := 0 var serverConn net.Conn var ip string var mutex sync.Mutex // 1.2. Associacoes entre canais de mensagens e atributos; nameChannels := make(map[uint32]string) passChannels := make(map[uint32]string) // 1.2. Banco de enderecos de servidores; chanAddresses := make(map[uint32][]string)
PanicOnError(err) defer jsonFile.Close() byteValueJSON, _ := ioutil.ReadAll(jsonFile) config := Properties{} json.Unmarshal(byteValueJSON, &config) channelCount := config.IDchanBegin nameChannels[0] = "Global" // 3. Preparando o cabecalho; cmd := exec.Command("cmd", "/c", "cls") cmd.Stdout = os.Stdout cmd.Run() head() // 4. Preparando a conexao dos clientes; http.HandleFunc("/echo", handler) http.HandleFunc("/", func(w http.ResponseWriter, r *http.Request) { http.ServeFile(w, r, "websockets.html") }) go func() { http.ListenAndServe("127.0.0.1:"+config.PortClients, nil) // Tratar erro! }() // 4.1. Estabelendo conexao com o servico RabbitMQ; rabbit, err := amqp.Dial("amqp://guest:guest@localhost:5672/") PanicOnError(err) defer rabbit.Close() // 4.2. Dedicando canal de servico RabbitMQ; ch, err := rabbit.Channel() PanicOnError(err) defer ch.Close() // 4.3. Dedicando exchange; err = ch.ExchangeDeclare("syslog", "fanout", true, false, false, false, nil) PanicOnError(err) err = ch.ExchangeDeclare("msglog", "fanout", true, false, false, false, nil) PanicOnError(err) err = ch.ExchangeDeclare("setlog", "fanout", true, false, false, false, nil) PanicOnError(err) // 4.4. Dedicando filas de mensagens; sysqueue, err := ch.QueueDeclare("", false, false, true, false, nil) PanicOnError(err) msgqueue, err := ch.QueueDeclare("", false, false, true, false, nil) PanicOnError(err) setqueue, err := ch.QueueDeclare("", false, false, true, false, nil) PanicOnError(err) // 4.5. Criando vinculacoes; err = ch.QueueBind(sysqueue.Name, "", "syslog", false, nil) PanicOnError(err) err = ch.QueueBind(msgqueue.Name, "", "msglog", false, nil) PanicOnError(err) err = ch.QueueBind(setqueue.Name, "", "setlog", false, nil) PanicOnError(err) // 5. Conectando e recebendo lista estruturada de servidores do Master Server; go func() { mutex.Lock() serverConn, ip = reconnect(config) mutex.Unlock() for { var list List dec := gob.NewDecoder(serverConn) err := dec.Decode(&list) if err != nil && err.Error() != "gob: unknown type id or corrupted data" { fmt.Println("The Master Server is offline now.") mutex.Lock() serverConn, ip = reconnect(config) mutex.Unlock() continue } else if list.Name == nil { continue } for name := range serverAddresses { delete(serverAddresses, name) } for i := 0; i < len(list.Name); i++ { serverAddresses[list.Name[i]] = list.AddrS[i] } delete(serverAddresses, config.ThisServerName) chanAddresses[0] = list.AddrS } }() // 7. Recebendo mensagens estruturadas de servidores; msgClient, err := ch.Consume(msgqueue.Name, "", true, false, false, false, nil) PanicOnError(err) msgSystem, err := ch.Consume(sysqueue.Name, "", true, false, false, false, nil) PanicOnError(err) msgConfig, err := ch.Consume(setqueue.Name, "", true, false, false, false, nil) PanicOnError(err) // Se for uma mensagem de cliente, proceder: go func() { for d := range msgClient { msg, _ := deserialize(d.Body) Message <- msg } }() // Se for uma mensagem global de sistema: go func() { for d := range msgSystem { msg, _ := deserialize(d.Body) for c := range nameClients { if msg.Message2 != nameClients[c] { err := c.WriteMessage(1, []byte(fmt.Sprint(msg.Message1+"\n"))) if err != nil { deadConnection <- c } } } } }() // Se forem atributos de canal, proceder: go func() { for d := range msgConfig { msg, _ := deserialize(d.Body) if str, ok := chanAddresses[msg.RoomID]; ok { str = append(str, msg.Message2) chanAddresses[msg.RoomID] = str str2 := strings.Split(msg.Message1, ":") nameChannels[msg.RoomID] = str2[0] passChannels[msg.RoomID] = str2[1] fmt.Println(msg) // } else { str = append(str, msg.Message2) chanAddresses[msg.RoomID] = str str2 := strings.Split(msg.Message1, ":") nameChannels[msg.RoomID] = str2[0] passChannels[msg.RoomID] = str2[1] fmt.Println(msg) // } } }() // 8. Preparando para receber conexoes de novos clientes; //clientListener, err := net.Listen("tcp", ":"+config.PortClients) //PanicOnError(err) // 9. Recebendo novos clientes; // go func(cl net.Listener) { // for { // conn, err := cl.Accept() // if err != nil { // raven.CaptureError(err, nil) // log.Println(err) // continue // } // clientName, _ := get(conn) // nameClients[conn] = clientName // newConnection <- conn // } // }(clientListener) // 10. Qual alternativa tomar? go func() { for { select { // 10.1. Se um novo cliente se conectar, proceder: case conn := <-newConnection: fmt.Println("A new client has arrived: " + nameClients[conn]) sysMessageGlobal <- Chat{2, nil, 0, fmt.Sprint("{System} " + nameClients[conn] + " is online now."), nameClients[conn]} chanClients[conn] = 0 clientCount++ time.Sleep(100 * time.Millisecond) // 9.1.2. Recebendo mensagens desse cliente indefinidamente; // go func(conn net.Conn) { // for { // message, err := get(conn) // if err != nil { // break // } // if message[0:1] != "/" { // // ID 0 significa um tipo de mensagem. // messageQueue <- Chat{0, nil, chanClients[conn], message, nameClients[conn]} // } else { // message = strings.TrimLeft(message, "/") // sub := strings.Split(message, " ") // // ID nos comandos nao possuem um significado. // commandQueue <- Command{0, conn, sub[0], sub[1:]} // } // } // // 9.1.2.1. Se o loop quebrar, significa que o cliente se desconectou; // deadConnection <- conn // }(conn) // 10.2. Se existe uma mensagem para os clientes locais, proceder: case msg := <-Message: fmt.Println("TESTE2") for conn := range chanClients { if msg.RoomID == chanClients[conn] && msg
serverAddresses := make(map[string]string) // 2. Lendo configuracoes; jsonFile, err := os.Open(`b-properties.json`)
random_line_split
B-Server.go
else { message = strings.TrimLeft(message, "/") sub := strings.Split(message, " ") // ID nos comandos nao possuem um significado. commandQueue <- Command{0, socket, sub[0], sub[1:]} } } // Se o loop quebrar, significa que o cliente se desconectou; deadConnection <- socket } func PanicOnError(err error) { if err != nil { raven.CaptureErrorAndWait(err, nil) log.Panic(err) } } func PrintOnError(err error) { if err != nil { raven.CaptureError(err, nil) log.Println(err) } } func serialize(message Chat) ([]byte, error) { var b bytes.Buffer encoder := gob.NewEncoder(&b) err := encoder.Encode(message) return b.Bytes(), err } func deserialize(b []byte) (Chat, error) { var msg Chat buf := bytes.NewBuffer(b) decoder := gob.NewDecoder(buf) err := decoder.Decode(&msg) return msg, err } func get(reader io.Reader) (message string, err error) { message, err = bufio.NewReader(reader).ReadString('\n') message = strings.TrimRight(message, "\r\n") return } func reconnect(config Properties) (serverConn net.Conn, ip string) { fmt.Println("Trying to connect to Master Server...") fmt.Println("It will not possible to know other new servers until this moment.") fmt.Println() time.Sleep(1000 * time.Millisecond) for { sc, err := net.Dial("tcp", config.MasterIP+":"+config.MasterPort) serverConn = sc sW := bufio.NewWriter(serverConn) if err != nil { time.Sleep(5000 * time.Millisecond) continue } else { fmt.Println("Connected to Master Server!") fmt.Println("Now, new global servers can be known.") fmt.Println() if config.PublicServer == false { ip = "127.0.0.1" sW.WriteString(config.ThisServerName + ":" + ip + ":" + config.PortServers + ":" + config.PortClients + "\n") sW.Flush() } else if config.PublicServer == true { var buf bytes.Buffer resp, _ := http.Get("http://myexternalip.com/raw") io.Copy(&buf, resp.Body) resp.Body.Close() ip = buf.String() sW.WriteString(config.ThisServerName + ":" + ip + ":" + config.PortServers + ":" + config.PortClients + "\n") sW.Flush() } break } } return } func head() { fmt.Println(" _____________________________") fmt.Println("| |") fmt.Println("| XIAOMI Chat System |") fmt.Println("| Developed by Saulo Pinedo |") fmt.Println("| --- |") fmt.Println("| Broadcast Program |") fmt.Println("|_____________________________|\n") } func main() { // 0. Inicializando servicos adicionais; raven.SetDSN("https://[email protected]/1831452") // 0.1. blabla2 raven.SetDefaultLoggerName("saulopinedo") raven.SetDebug(true) raven.SetEnvironment("staging") raven.SetRelease("Xiaomi") raven.SetSampleRate(1.0) // 1. Pre-definindo as variaveis fundamentais; clientCount := 0 var serverConn net.Conn var ip string var mutex sync.Mutex // 1.2. Associacoes entre canais de mensagens e atributos; nameChannels := make(map[uint32]string) passChannels := make(map[uint32]string) // 1.2. Banco de enderecos de servidores; chanAddresses := make(map[uint32][]string) serverAddresses := make(map[string]string) // 2. Lendo configuracoes; jsonFile, err := os.Open(`b-properties.json`) PanicOnError(err) defer jsonFile.Close() byteValueJSON, _ := ioutil.ReadAll(jsonFile) config := Properties{} json.Unmarshal(byteValueJSON, &config) channelCount := config.IDchanBegin nameChannels[0] = "Global" // 3. Preparando o cabecalho; cmd := exec.Command("cmd", "/c", "cls") cmd.Stdout = os.Stdout cmd.Run() head() // 4. Preparando a conexao dos clientes; http.HandleFunc("/echo", handler) http.HandleFunc("/", func(w http.ResponseWriter, r *http.Request) { http.ServeFile(w, r, "websockets.html") }) go func() { http.ListenAndServe("127.0.0.1:"+config.PortClients, nil) // Tratar erro! }() // 4.1. Estabelendo conexao com o servico RabbitMQ; rabbit, err := amqp.Dial("amqp://guest:guest@localhost:5672/") PanicOnError(err) defer rabbit.Close() // 4.2. Dedicando canal de servico RabbitMQ; ch, err := rabbit.Channel() PanicOnError(err) defer ch.Close() // 4.3. Dedicando exchange; err = ch.ExchangeDeclare("syslog", "fanout", true, false, false, false, nil) PanicOnError(err) err = ch.ExchangeDeclare("msglog", "fanout", true, false, false, false, nil) PanicOnError(err) err = ch.ExchangeDeclare("setlog", "fanout", true, false, false, false, nil) PanicOnError(err) // 4.4. Dedicando filas de mensagens; sysqueue, err := ch.QueueDeclare("", false, false, true, false, nil) PanicOnError(err) msgqueue, err := ch.QueueDeclare("", false, false, true, false, nil) PanicOnError(err) setqueue, err := ch.QueueDeclare("", false, false, true, false, nil) PanicOnError(err) // 4.5. Criando vinculacoes; err = ch.QueueBind(sysqueue.Name, "", "syslog", false, nil) PanicOnError(err) err = ch.QueueBind(msgqueue.Name, "", "msglog", false, nil) PanicOnError(err) err = ch.QueueBind(setqueue.Name, "", "setlog", false, nil) PanicOnError(err) // 5. Conectando e recebendo lista estruturada de servidores do Master Server; go func() { mutex.Lock() serverConn, ip = reconnect(config) mutex.Unlock() for { var list List dec := gob.NewDecoder(serverConn) err := dec.Decode(&list) if err != nil && err.Error() != "gob: unknown type id or corrupted data" { fmt.Println("The Master Server is offline now.") mutex.Lock() serverConn, ip = reconnect(config) mutex.Unlock() continue } else if list.Name == nil { continue } for name := range serverAddresses { delete(serverAddresses, name) } for i := 0; i < len(list.Name); i++ { serverAddresses[list.Name[i]] = list.AddrS[i] } delete(serverAddresses, config.ThisServerName) chanAddresses[0] = list.AddrS } }() // 7. Recebendo mensagens estruturadas de servidores; msgClient, err := ch.Consume(msgqueue.Name, "", true, false, false, false, nil) PanicOnError(err) msgSystem, err := ch.Consume(sysqueue.Name, "", true, false, false, false, nil) PanicOnError(err) msgConfig, err := ch.Consume(setqueue.Name, "", true, false, false, false, nil) PanicOnError(err) // Se for uma mensagem de cliente, proceder: go func() { for d := range msgClient { msg, _ := deserialize(d.Body) Message <- msg } }() // Se for uma mensagem global de sistema: go func() { for d := range msgSystem { msg, _ := deserialize(d.Body) for c := range nameClients { if msg.Message2 != nameClients[c] { err := c.WriteMessage(1, []byte(fmt.Sprint(msg.Message1+"\n"))) if err != nil { deadConnection <- c } } } } }() // Se forem atributos de canal, proceder: go func() { for d := range msgConfig { msg, _ := deserialize(d.Body) if str, ok := chanAddresses[msg.RoomID]; ok { str = append(str, msg.Message2) chanAddresses[msg.RoomID] = str str2 := strings.Split(msg.Message1, ":") nameChannels[msg.RoomID] = str2[0] passChannels[msg.RoomID] = str2[1] fmt.Println(msg) // } else { str = append(str, msg.Message2) chan
{ // ID 0 significa um tipo de mensagem. messageQueue <- Chat{0, nil, chanClients[socket], message, nameClients[socket]} }
conditional_block
B-Server.go
(writer http.ResponseWriter, request *http.Request) { socket, err := upgrader.Upgrade(writer, request, nil) if err != nil { fmt.Println(err) } _, msg, err := socket.ReadMessage() nameClients[socket] = string(msg) // msg contem o nome do usuario. newConnection <- socket for { // Recebendo mensagens desse cliente indefinidamente: _, msg, err := socket.ReadMessage() if err != nil { break } message := string(msg) if message[0:1] != "/" { // ID 0 significa um tipo de mensagem. messageQueue <- Chat{0, nil, chanClients[socket], message, nameClients[socket]} } else { message = strings.TrimLeft(message, "/") sub := strings.Split(message, " ") // ID nos comandos nao possuem um significado. commandQueue <- Command{0, socket, sub[0], sub[1:]} } } // Se o loop quebrar, significa que o cliente se desconectou; deadConnection <- socket } func PanicOnError(err error) { if err != nil { raven.CaptureErrorAndWait(err, nil) log.Panic(err) } } func PrintOnError(err error) { if err != nil { raven.CaptureError(err, nil) log.Println(err) } } func serialize(message Chat) ([]byte, error) { var b bytes.Buffer encoder := gob.NewEncoder(&b) err := encoder.Encode(message) return b.Bytes(), err } func deserialize(b []byte) (Chat, error) { var msg Chat buf := bytes.NewBuffer(b) decoder := gob.NewDecoder(buf) err := decoder.Decode(&msg) return msg, err } func get(reader io.Reader) (message string, err error) { message, err = bufio.NewReader(reader).ReadString('\n') message = strings.TrimRight(message, "\r\n") return } func reconnect(config Properties) (serverConn net.Conn, ip string) { fmt.Println("Trying to connect to Master Server...") fmt.Println("It will not possible to know other new servers until this moment.") fmt.Println() time.Sleep(1000 * time.Millisecond) for { sc, err := net.Dial("tcp", config.MasterIP+":"+config.MasterPort) serverConn = sc sW := bufio.NewWriter(serverConn) if err != nil { time.Sleep(5000 * time.Millisecond) continue } else { fmt.Println("Connected to Master Server!") fmt.Println("Now, new global servers can be known.") fmt.Println() if config.PublicServer == false { ip = "127.0.0.1" sW.WriteString(config.ThisServerName + ":" + ip + ":" + config.PortServers + ":" + config.PortClients + "\n") sW.Flush() } else if config.PublicServer == true { var buf bytes.Buffer resp, _ := http.Get("http://myexternalip.com/raw") io.Copy(&buf, resp.Body) resp.Body.Close() ip = buf.String() sW.WriteString(config.ThisServerName + ":" + ip + ":" + config.PortServers + ":" + config.PortClients + "\n") sW.Flush() } break } } return } func head() { fmt.Println(" _____________________________") fmt.Println("| |") fmt.Println("| XIAOMI Chat System |") fmt.Println("| Developed by Saulo Pinedo |") fmt.Println("| --- |") fmt.Println("| Broadcast Program |") fmt.Println("|_____________________________|\n") } func main() { // 0. Inicializando servicos adicionais; raven.SetDSN("https://[email protected]/1831452") // 0.1. blabla2 raven.SetDefaultLoggerName("saulopinedo") raven.SetDebug(true) raven.SetEnvironment("staging") raven.SetRelease("Xiaomi") raven.SetSampleRate(1.0) // 1. Pre-definindo as variaveis fundamentais; clientCount := 0 var serverConn net.Conn var ip string var mutex sync.Mutex // 1.2. Associacoes entre canais de mensagens e atributos; nameChannels := make(map[uint32]string) passChannels := make(map[uint32]string) // 1.2. Banco de enderecos de servidores; chanAddresses := make(map[uint32][]string) serverAddresses := make(map[string]string) // 2. Lendo configuracoes; jsonFile, err := os.Open(`b-properties.json`) PanicOnError(err) defer jsonFile.Close() byteValueJSON, _ := ioutil.ReadAll(jsonFile) config := Properties{} json.Unmarshal(byteValueJSON, &config) channelCount := config.IDchanBegin nameChannels[0] = "Global" // 3. Preparando o cabecalho; cmd := exec.Command("cmd", "/c", "cls") cmd.Stdout = os.Stdout cmd.Run() head() // 4. Preparando a conexao dos clientes; http.HandleFunc("/echo", handler) http.HandleFunc("/", func(w http.ResponseWriter, r *http.Request) { http.ServeFile(w, r, "websockets.html") }) go func() { http.ListenAndServe("127.0.0.1:"+config.PortClients, nil) // Tratar erro! }() // 4.1. Estabelendo conexao com o servico RabbitMQ; rabbit, err := amqp.Dial("amqp://guest:guest@localhost:5672/") PanicOnError(err) defer rabbit.Close() // 4.2. Dedicando canal de servico RabbitMQ; ch, err := rabbit.Channel() PanicOnError(err) defer ch.Close() // 4.3. Dedicando exchange; err = ch.ExchangeDeclare("syslog", "fanout", true, false, false, false, nil) PanicOnError(err) err = ch.ExchangeDeclare("msglog", "fanout", true, false, false, false, nil) PanicOnError(err) err = ch.ExchangeDeclare("setlog", "fanout", true, false, false, false, nil) PanicOnError(err) // 4.4. Dedicando filas de mensagens; sysqueue, err := ch.QueueDeclare("", false, false, true, false, nil) PanicOnError(err) msgqueue, err := ch.QueueDeclare("", false, false, true, false, nil) PanicOnError(err) setqueue, err := ch.QueueDeclare("", false, false, true, false, nil) PanicOnError(err) // 4.5. Criando vinculacoes; err = ch.QueueBind(sysqueue.Name, "", "syslog", false, nil) PanicOnError(err) err = ch.QueueBind(msgqueue.Name, "", "msglog", false, nil) PanicOnError(err) err = ch.QueueBind(setqueue.Name, "", "setlog", false, nil) PanicOnError(err) // 5. Conectando e recebendo lista estruturada de servidores do Master Server; go func() { mutex.Lock() serverConn, ip = reconnect(config) mutex.Unlock() for { var list List dec := gob.NewDecoder(serverConn) err := dec.Decode(&list) if err != nil && err.Error() != "gob: unknown type id or corrupted data" { fmt.Println("The Master Server is offline now.") mutex.Lock() serverConn, ip = reconnect(config) mutex.Unlock() continue } else if list.Name == nil { continue } for name := range serverAddresses { delete(serverAddresses, name) } for i := 0; i < len(list.Name); i++ { serverAddresses[list.Name[i]] = list.AddrS[i] } delete(serverAddresses, config.ThisServerName) chanAddresses[0] = list.AddrS } }() // 7. Recebendo mensagens estruturadas de servidores; msgClient, err := ch.Consume(msgqueue.Name, "", true, false, false, false, nil) PanicOnError(err) msgSystem, err := ch.Consume(sysqueue.Name, "", true, false, false, false, nil) PanicOnError(err) msgConfig, err := ch.Consume(setqueue.Name, "", true, false, false, false, nil) PanicOnError(err) // Se for uma mensagem de cliente, proceder: go func() { for d := range msgClient { msg, _ := deserialize(d.Body) Message <- msg } }() // Se for uma mensagem global de sistema: go func() { for d := range msgSystem { msg, _ := deserialize(d.Body) for c := range nameClients { if msg.Message2 != nameClients[c] { err := c.WriteMessage(1, []byte(fmt.Sprint(msg.Message1+"\n"))) if err != nil { deadConnection <- c } } } } }() // Se
handler
identifier_name
B-Server.go
func PrintOnError(err error) { if err != nil { raven.CaptureError(err, nil) log.Println(err) } } func serialize(message Chat) ([]byte, error) { var b bytes.Buffer encoder := gob.NewEncoder(&b) err := encoder.Encode(message) return b.Bytes(), err } func deserialize(b []byte) (Chat, error) { var msg Chat buf := bytes.NewBuffer(b) decoder := gob.NewDecoder(buf) err := decoder.Decode(&msg) return msg, err } func get(reader io.Reader) (message string, err error) { message, err = bufio.NewReader(reader).ReadString('\n') message = strings.TrimRight(message, "\r\n") return } func reconnect(config Properties) (serverConn net.Conn, ip string) { fmt.Println("Trying to connect to Master Server...") fmt.Println("It will not possible to know other new servers until this moment.") fmt.Println() time.Sleep(1000 * time.Millisecond) for { sc, err := net.Dial("tcp", config.MasterIP+":"+config.MasterPort) serverConn = sc sW := bufio.NewWriter(serverConn) if err != nil { time.Sleep(5000 * time.Millisecond) continue } else { fmt.Println("Connected to Master Server!") fmt.Println("Now, new global servers can be known.") fmt.Println() if config.PublicServer == false { ip = "127.0.0.1" sW.WriteString(config.ThisServerName + ":" + ip + ":" + config.PortServers + ":" + config.PortClients + "\n") sW.Flush() } else if config.PublicServer == true { var buf bytes.Buffer resp, _ := http.Get("http://myexternalip.com/raw") io.Copy(&buf, resp.Body) resp.Body.Close() ip = buf.String() sW.WriteString(config.ThisServerName + ":" + ip + ":" + config.PortServers + ":" + config.PortClients + "\n") sW.Flush() } break } } return } func head() { fmt.Println(" _____________________________") fmt.Println("| |") fmt.Println("| XIAOMI Chat System |") fmt.Println("| Developed by Saulo Pinedo |") fmt.Println("| --- |") fmt.Println("| Broadcast Program |") fmt.Println("|_____________________________|\n") } func main() { // 0. Inicializando servicos adicionais; raven.SetDSN("https://[email protected]/1831452") // 0.1. blabla2 raven.SetDefaultLoggerName("saulopinedo") raven.SetDebug(true) raven.SetEnvironment("staging") raven.SetRelease("Xiaomi") raven.SetSampleRate(1.0) // 1. Pre-definindo as variaveis fundamentais; clientCount := 0 var serverConn net.Conn var ip string var mutex sync.Mutex // 1.2. Associacoes entre canais de mensagens e atributos; nameChannels := make(map[uint32]string) passChannels := make(map[uint32]string) // 1.2. Banco de enderecos de servidores; chanAddresses := make(map[uint32][]string) serverAddresses := make(map[string]string) // 2. Lendo configuracoes; jsonFile, err := os.Open(`b-properties.json`) PanicOnError(err) defer jsonFile.Close() byteValueJSON, _ := ioutil.ReadAll(jsonFile) config := Properties{} json.Unmarshal(byteValueJSON, &config) channelCount := config.IDchanBegin nameChannels[0] = "Global" // 3. Preparando o cabecalho; cmd := exec.Command("cmd", "/c", "cls") cmd.Stdout = os.Stdout cmd.Run() head() // 4. Preparando a conexao dos clientes; http.HandleFunc("/echo", handler) http.HandleFunc("/", func(w http.ResponseWriter, r *http.Request) { http.ServeFile(w, r, "websockets.html") }) go func() { http.ListenAndServe("127.0.0.1:"+config.PortClients, nil) // Tratar erro! }() // 4.1. Estabelendo conexao com o servico RabbitMQ; rabbit, err := amqp.Dial("amqp://guest:guest@localhost:5672/") PanicOnError(err) defer rabbit.Close() // 4.2. Dedicando canal de servico RabbitMQ; ch, err := rabbit.Channel() PanicOnError(err) defer ch.Close() // 4.3. Dedicando exchange; err = ch.ExchangeDeclare("syslog", "fanout", true, false, false, false, nil) PanicOnError(err) err = ch.ExchangeDeclare("msglog", "fanout", true, false, false, false, nil) PanicOnError(err) err = ch.ExchangeDeclare("setlog", "fanout", true, false, false, false, nil) PanicOnError(err) // 4.4. Dedicando filas de mensagens; sysqueue, err := ch.QueueDeclare("", false, false, true, false, nil) PanicOnError(err) msgqueue, err := ch.QueueDeclare("", false, false, true, false, nil) PanicOnError(err) setqueue, err := ch.QueueDeclare("", false, false, true, false, nil) PanicOnError(err) // 4.5. Criando vinculacoes; err = ch.QueueBind(sysqueue.Name, "", "syslog", false, nil) PanicOnError(err) err = ch.QueueBind(msgqueue.Name, "", "msglog", false, nil) PanicOnError(err) err = ch.QueueBind(setqueue.Name, "", "setlog", false, nil) PanicOnError(err) // 5. Conectando e recebendo lista estruturada de servidores do Master Server; go func() { mutex.Lock() serverConn, ip = reconnect(config) mutex.Unlock() for { var list List dec := gob.NewDecoder(serverConn) err := dec.Decode(&list) if err != nil && err.Error() != "gob: unknown type id or corrupted data" { fmt.Println("The Master Server is offline now.") mutex.Lock() serverConn, ip = reconnect(config) mutex.Unlock() continue } else if list.Name == nil { continue } for name := range serverAddresses { delete(serverAddresses, name) } for i := 0; i < len(list.Name); i++ { serverAddresses[list.Name[i]] = list.AddrS[i] } delete(serverAddresses, config.ThisServerName) chanAddresses[0] = list.AddrS } }() // 7. Recebendo mensagens estruturadas de servidores; msgClient, err := ch.Consume(msgqueue.Name, "", true, false, false, false, nil) PanicOnError(err) msgSystem, err := ch.Consume(sysqueue.Name, "", true, false, false, false, nil) PanicOnError(err) msgConfig, err := ch.Consume(setqueue.Name, "", true, false, false, false, nil) PanicOnError(err) // Se for uma mensagem de cliente, proceder: go func() { for d := range msgClient { msg, _ := deserialize(d.Body) Message <- msg } }() // Se for uma mensagem global de sistema: go func() { for d := range msgSystem { msg, _ := deserialize(d.Body) for c := range nameClients { if msg.Message2 != nameClients[c] { err := c.WriteMessage(1, []byte(fmt.Sprint(msg.Message1+"\n"))) if err != nil { deadConnection <- c } } } } }() // Se forem atributos de canal, proceder: go func() { for d := range msgConfig { msg, _ := deserialize(d.Body) if str, ok := chanAddresses[msg.RoomID]; ok { str = append(str, msg.Message2) chanAddresses[msg.RoomID] = str str2 := strings.Split(msg.Message1, ":") nameChannels[msg.RoomID] = str2[0] passChannels[msg.RoomID] = str2[1] fmt.Println(msg) // } else { str = append(str, msg.Message2) chanAddresses[msg.RoomID] = str str2 := strings.Split(msg.Message1, ":") nameChannels[msg.RoomID] = str2[0] passChannels[msg.RoomID] = str2[1] fmt.Println(msg) // } } }() // 8. Preparando para receber conexoes de novos clientes; //clientListener, err := net.Listen("tcp", ":"+config.PortClients) //PanicOnError(err) // 9. Recebendo novos clientes; // go func(cl net.Listener) { // for {
{ if err != nil { raven.CaptureErrorAndWait(err, nil) log.Panic(err) } }
identifier_body
train_fineall.py
1)') parser.add_argument('--epochs', type=int, default=600, metavar='N', help='number of epochs to train (default: 300)') parser.add_argument('--batchsize', type=int, default=128, metavar='N', help='input batch size for training (default: 64)') parser.add_argument('--lr', type=float, default=0.1, metavar='LR', help='learning rate (default: 0.001)') parser.add_argument('--momentum', type=float, default=0.9, metavar='M', help='SGD momentum (default: 0.9)') parser.add_argument('--weight_decay', type=float, default=5e-4, help='SGD weight_decay (default: 5e-4)') parser.add_argument('--automonous_stopping', type=int,default=0, help='automonous_stopping') parser.add_argument('--data',default='data',metavar='NT', help='the data directory') parser.add_argument('--modelpos',default='models/resnet18_pretrained.t7',metavar='NT', help='the data directory') args=parser.parse_args() args.cuda = not args.no_cuda and torch.cuda.is_available() torch.manual_seed(args.seed) if args.cuda: torch.cuda.manual_seed(args.seed) PATH_TO_MODEL = os.path.join('models', 'color_models') category_num=24 history=[0.01]*1000 historyMax=0.01 Hloss=[0.01]*1000 lossMin=1000000.0 hash2=[0, 1, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 2, 20, 21, 22, 23, 3, 4, 5, 6, 7, 8, 9] colorMax=[0.0001]*24 Misspre=np.array([[0.01 for col in range(0,24)] for row in range(0,24)]) # loading pre-trained model print(colored('initializing the model ...',"blue")) print(colored('load model at '+args.modelpos,"blue")) #model = models.resnet18(pretrained=True)#at least 224*224 model=torch.load(args.modelpos) model.fc=nn.Linear(512,24) for param in model.parameters(): param.requires_grads=False model.fc.requires_grads=True model.layer4.requires_grads=True model.layer3.requires_grads=True model.layer2.requires_grads=True model.layer1.requires_grads=True if args.cuda: model.cuda() print(colored('model ==> ',"green")) print(model) print(colored('initializing done.',"blue")) # Data loading code traindir = os.path.join(args.data, 'train') valdir = os.path.join(args.data, 'val') normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]) traindir='data/train' valdir='data/val' #print(os.listdir(traindir)) train_loader1= torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.RandomSizedCrop(224), transforms.RandomHorizontalFlip(), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) train_loader2= torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.Scale(224), #transforms.CenterCrop(224), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) train_loader3= torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.RandomSizedCrop(224), transforms.RandomHorizontalFlip(), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) val_loader = torch.utils.data.DataLoader( datasets.ImageFolder(valdir, transforms.Compose([ transforms.Scale(256), transforms.CenterCrop(224), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=False, num_workers=4, pin_memory=True) def imshow(img):
#dataiter = iter(train_loader) #print(len(train_loader)) #print(len(val_loader)) #images,labels= dataiter.next() #print(images) #print(labels) # print images #imshow(torchvision.utils.make_grid(images)) # define loss function (criterion) and pptimizer criterion = nn.CrossEntropyLoss().cuda() optimizer = optim.SGD(model.parameters(), lr=args.lr, momentum=args.momentum,weight_decay=args.weight_decay) def save_model(model, name): print('saving the model ...') if not os.path.exists(PATH_TO_MODEL): os.mkdir(PATH_TO_MODEL) torch.save(model,PATH_TO_MODEL+'/'+str(historyMax)+'.t7') print('done.') def visualize(data): for i in range(0,24): print('color '+str(i)+ '!') for j in range(0,24): if(data[i][j]>5): print '[%d][%d] = %.2f ' %(i,j,data[i][j]) , print() for i in range(0,24): print(colored('color %d precision: %.2f !' %(i,data[i][i]),'green')) def save_color_model(model, color): print('saving the color model for %d ...' %color) visualize(Misspre) PATH_TO_COLOR=PATH_TO_MODEL+str(color) if not os.path.exists(PATH_TO_COLOR): os.mkdir(PATH_TO_COLOR) torch.save(model,PATH_TO_COLOR+'/'+str(colorMax[color])+'.t7') print('done.') def train(epoch): model.train() print(colored('training epoch '+ str(epoch) + ' !','blue')) print(colored('loading data!','green')) if epoch%3==0: train_loader=train_loader1 elif epoch%3==1: train_loader=torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.Scale(224), #transforms.CenterCrop(224), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) else: train_loader=train_loader3 print(colored('done!','green')) tot_loss=0.0 num=0 right=0 for i, (inputs, targets) in tqdm(enumerate(train_loader)): if args.cuda: inputs=inputs.cuda(async=True) targets=targets.cuda(async=True) inputs_var, targets_var = Variable(inputs), Variable(targets) optimizer.zero_grad() outputs = model(inputs_var) loss=criterion(outputs, targets_var)/(args.batchsize*1.0) loss.backward() optimizer.step() tot_loss=tot_loss+loss.data[0] num=num+inputs.size(0) _,indices=torch.max(outputs.data,1) indices=indices.view(inputs.size(0)) right=right+sum(indices==targets) #print(output.data) #averageloss=2.3 #averageloss=(tot_loss*1.0)/(num*1.00) print(colored("totloss: %.8f ! " %tot_loss,'red')) precision=2.3 precision=(right*100.0)/(num*1.00) print(colored("precision: %.2f%c ! " %(precision,'%'),'red')) global Hloss,lossMin Hloss[epoch]=tot_loss if epoch==1: lossMin=tot_loss else: lossMin=min(lossMin,tot_loss) #print(colored("right: %d ! " %right,'red')) def test(epoch): Miss=[[0 for col in range(0,24)] for row in range(0,24)]# suppose i to be j model.eval() print(colored('Testing!','blue')) tot_loss=0.0 num=0 right=0 for i, (inputs, targets) in tqdm(enumerate(val_loader)): if args.cuda: inputs=inputs.cuda(async=True) targets=targets.cuda(async=True) inputs_var, targets_var = Variable(inputs), Variable(targets) #optimizer.zero_grad() outputs = model(inputs_var) loss=criterion(outputs, targets_var)/(args.batchsize*1.0) #loss.backward() #optimizer.step() tot_loss=tot_loss+loss.data[0] num=num+inputs.size(0)
img = img / 2 + 0.5 # unnormalize npimg = img.numpy() plt.imshow(np.transpose(npimg, (1,2,0))) plt.show()
identifier_body
train_fineall.py
1)') parser.add_argument('--epochs', type=int, default=600, metavar='N', help='number of epochs to train (default: 300)') parser.add_argument('--batchsize', type=int, default=128, metavar='N', help='input batch size for training (default: 64)') parser.add_argument('--lr', type=float, default=0.1, metavar='LR', help='learning rate (default: 0.001)') parser.add_argument('--momentum', type=float, default=0.9, metavar='M', help='SGD momentum (default: 0.9)') parser.add_argument('--weight_decay', type=float, default=5e-4, help='SGD weight_decay (default: 5e-4)') parser.add_argument('--automonous_stopping', type=int,default=0, help='automonous_stopping') parser.add_argument('--data',default='data',metavar='NT', help='the data directory') parser.add_argument('--modelpos',default='models/resnet18_pretrained.t7',metavar='NT', help='the data directory') args=parser.parse_args() args.cuda = not args.no_cuda and torch.cuda.is_available() torch.manual_seed(args.seed) if args.cuda: torch.cuda.manual_seed(args.seed) PATH_TO_MODEL = os.path.join('models', 'color_models') category_num=24 history=[0.01]*1000 historyMax=0.01 Hloss=[0.01]*1000 lossMin=1000000.0 hash2=[0, 1, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 2, 20, 21, 22, 23, 3, 4, 5, 6, 7, 8, 9] colorMax=[0.0001]*24 Misspre=np.array([[0.01 for col in range(0,24)] for row in range(0,24)]) # loading pre-trained model print(colored('initializing the model ...',"blue")) print(colored('load model at '+args.modelpos,"blue")) #model = models.resnet18(pretrained=True)#at least 224*224 model=torch.load(args.modelpos) model.fc=nn.Linear(512,24) for param in model.parameters(): param.requires_grads=False model.fc.requires_grads=True model.layer4.requires_grads=True model.layer3.requires_grads=True model.layer2.requires_grads=True model.layer1.requires_grads=True if args.cuda: model.cuda() print(colored('model ==> ',"green")) print(model) print(colored('initializing done.',"blue")) # Data loading code traindir = os.path.join(args.data, 'train') valdir = os.path.join(args.data, 'val') normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]) traindir='data/train' valdir='data/val' #print(os.listdir(traindir)) train_loader1= torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.RandomSizedCrop(224), transforms.RandomHorizontalFlip(), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) train_loader2= torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.Scale(224), #transforms.CenterCrop(224), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) train_loader3= torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.RandomSizedCrop(224), transforms.RandomHorizontalFlip(), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) val_loader = torch.utils.data.DataLoader( datasets.ImageFolder(valdir, transforms.Compose([ transforms.Scale(256), transforms.CenterCrop(224), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=False, num_workers=4, pin_memory=True) def imshow(img): img = img / 2 + 0.5 # unnormalize npimg = img.numpy() plt.imshow(np.transpose(npimg, (1,2,0))) plt.show() #dataiter = iter(train_loader) #print(len(train_loader)) #print(len(val_loader)) #images,labels= dataiter.next() #print(images) #print(labels) # print images #imshow(torchvision.utils.make_grid(images)) # define loss function (criterion) and pptimizer criterion = nn.CrossEntropyLoss().cuda() optimizer = optim.SGD(model.parameters(), lr=args.lr, momentum=args.momentum,weight_decay=args.weight_decay) def
(model, name): print('saving the model ...') if not os.path.exists(PATH_TO_MODEL): os.mkdir(PATH_TO_MODEL) torch.save(model,PATH_TO_MODEL+'/'+str(historyMax)+'.t7') print('done.') def visualize(data): for i in range(0,24): print('color '+str(i)+ '!') for j in range(0,24): if(data[i][j]>5): print '[%d][%d] = %.2f ' %(i,j,data[i][j]) , print() for i in range(0,24): print(colored('color %d precision: %.2f !' %(i,data[i][i]),'green')) def save_color_model(model, color): print('saving the color model for %d ...' %color) visualize(Misspre) PATH_TO_COLOR=PATH_TO_MODEL+str(color) if not os.path.exists(PATH_TO_COLOR): os.mkdir(PATH_TO_COLOR) torch.save(model,PATH_TO_COLOR+'/'+str(colorMax[color])+'.t7') print('done.') def train(epoch): model.train() print(colored('training epoch '+ str(epoch) + ' !','blue')) print(colored('loading data!','green')) if epoch%3==0: train_loader=train_loader1 elif epoch%3==1: train_loader=torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.Scale(224), #transforms.CenterCrop(224), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) else: train_loader=train_loader3 print(colored('done!','green')) tot_loss=0.0 num=0 right=0 for i, (inputs, targets) in tqdm(enumerate(train_loader)): if args.cuda: inputs=inputs.cuda(async=True) targets=targets.cuda(async=True) inputs_var, targets_var = Variable(inputs), Variable(targets) optimizer.zero_grad() outputs = model(inputs_var) loss=criterion(outputs, targets_var)/(args.batchsize*1.0) loss.backward() optimizer.step() tot_loss=tot_loss+loss.data[0] num=num+inputs.size(0) _,indices=torch.max(outputs.data,1) indices=indices.view(inputs.size(0)) right=right+sum(indices==targets) #print(output.data) #averageloss=2.3 #averageloss=(tot_loss*1.0)/(num*1.00) print(colored("totloss: %.8f ! " %tot_loss,'red')) precision=2.3 precision=(right*100.0)/(num*1.00) print(colored("precision: %.2f%c ! " %(precision,'%'),'red')) global Hloss,lossMin Hloss[epoch]=tot_loss if epoch==1: lossMin=tot_loss else: lossMin=min(lossMin,tot_loss) #print(colored("right: %d ! " %right,'red')) def test(epoch): Miss=[[0 for col in range(0,24)] for row in range(0,24)]# suppose i to be j model.eval() print(colored('Testing!','blue')) tot_loss=0.0 num=0 right=0 for i, (inputs, targets) in tqdm(enumerate(val_loader)): if args.cuda: inputs=inputs.cuda(async=True) targets=targets.cuda(async=True) inputs_var, targets_var = Variable(inputs), Variable(targets) #optimizer.zero_grad() outputs = model(inputs_var) loss=criterion(outputs, targets_var)/(args.batchsize*1.0) #loss.backward() #optimizer.step() tot_loss=tot_loss+loss.data[0] num=num+inputs.size(0)
save_model
identifier_name
train_fineall.py
net18(pretrained=True)#at least 224*224 model=torch.load(args.modelpos) model.fc=nn.Linear(512,24) for param in model.parameters(): param.requires_grads=False model.fc.requires_grads=True model.layer4.requires_grads=True model.layer3.requires_grads=True model.layer2.requires_grads=True model.layer1.requires_grads=True if args.cuda: model.cuda() print(colored('model ==> ',"green")) print(model) print(colored('initializing done.',"blue")) # Data loading code traindir = os.path.join(args.data, 'train') valdir = os.path.join(args.data, 'val') normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]) traindir='data/train' valdir='data/val' #print(os.listdir(traindir)) train_loader1= torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.RandomSizedCrop(224), transforms.RandomHorizontalFlip(), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) train_loader2= torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.Scale(224), #transforms.CenterCrop(224), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) train_loader3= torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.RandomSizedCrop(224), transforms.RandomHorizontalFlip(), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) val_loader = torch.utils.data.DataLoader( datasets.ImageFolder(valdir, transforms.Compose([ transforms.Scale(256), transforms.CenterCrop(224), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=False, num_workers=4, pin_memory=True) def imshow(img): img = img / 2 + 0.5 # unnormalize npimg = img.numpy() plt.imshow(np.transpose(npimg, (1,2,0))) plt.show() #dataiter = iter(train_loader) #print(len(train_loader)) #print(len(val_loader)) #images,labels= dataiter.next() #print(images) #print(labels) # print images #imshow(torchvision.utils.make_grid(images)) # define loss function (criterion) and pptimizer criterion = nn.CrossEntropyLoss().cuda() optimizer = optim.SGD(model.parameters(), lr=args.lr, momentum=args.momentum,weight_decay=args.weight_decay) def save_model(model, name): print('saving the model ...') if not os.path.exists(PATH_TO_MODEL): os.mkdir(PATH_TO_MODEL) torch.save(model,PATH_TO_MODEL+'/'+str(historyMax)+'.t7') print('done.') def visualize(data): for i in range(0,24): print('color '+str(i)+ '!') for j in range(0,24): if(data[i][j]>5): print '[%d][%d] = %.2f ' %(i,j,data[i][j]) , print() for i in range(0,24): print(colored('color %d precision: %.2f !' %(i,data[i][i]),'green')) def save_color_model(model, color): print('saving the color model for %d ...' %color) visualize(Misspre) PATH_TO_COLOR=PATH_TO_MODEL+str(color) if not os.path.exists(PATH_TO_COLOR): os.mkdir(PATH_TO_COLOR) torch.save(model,PATH_TO_COLOR+'/'+str(colorMax[color])+'.t7') print('done.') def train(epoch): model.train() print(colored('training epoch '+ str(epoch) + ' !','blue')) print(colored('loading data!','green')) if epoch%3==0: train_loader=train_loader1 elif epoch%3==1: train_loader=torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.Scale(224), #transforms.CenterCrop(224), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) else: train_loader=train_loader3 print(colored('done!','green')) tot_loss=0.0 num=0 right=0 for i, (inputs, targets) in tqdm(enumerate(train_loader)): if args.cuda: inputs=inputs.cuda(async=True) targets=targets.cuda(async=True) inputs_var, targets_var = Variable(inputs), Variable(targets) optimizer.zero_grad() outputs = model(inputs_var) loss=criterion(outputs, targets_var)/(args.batchsize*1.0) loss.backward() optimizer.step() tot_loss=tot_loss+loss.data[0] num=num+inputs.size(0) _,indices=torch.max(outputs.data,1) indices=indices.view(inputs.size(0)) right=right+sum(indices==targets) #print(output.data) #averageloss=2.3 #averageloss=(tot_loss*1.0)/(num*1.00) print(colored("totloss: %.8f ! " %tot_loss,'red')) precision=2.3 precision=(right*100.0)/(num*1.00) print(colored("precision: %.2f%c ! " %(precision,'%'),'red')) global Hloss,lossMin Hloss[epoch]=tot_loss if epoch==1: lossMin=tot_loss else: lossMin=min(lossMin,tot_loss) #print(colored("right: %d ! " %right,'red')) def test(epoch): Miss=[[0 for col in range(0,24)] for row in range(0,24)]# suppose i to be j model.eval() print(colored('Testing!','blue')) tot_loss=0.0 num=0 right=0 for i, (inputs, targets) in tqdm(enumerate(val_loader)): if args.cuda: inputs=inputs.cuda(async=True) targets=targets.cuda(async=True) inputs_var, targets_var = Variable(inputs), Variable(targets) #optimizer.zero_grad() outputs = model(inputs_var) loss=criterion(outputs, targets_var)/(args.batchsize*1.0) #loss.backward() #optimizer.step() tot_loss=tot_loss+loss.data[0] num=num+inputs.size(0) _,indices=torch.max(outputs.data,1) indices=indices.view(inputs.size(0)) right=right+sum(indices==targets) for j in range(0,inputs.size(0)): #if targets[j]>24 or indices[j]>24: #print(targets[j],indices[j]) Miss[hash2[int(targets[j])]][hash2[int(indices[j])]]+=1 #print(output.data) #averageloss=2.3 #averageloss=(tot_loss*1.0)/(num*1.00) print(colored("totloss: %.8f ! " %tot_loss,'red')) precision=2.3 precision=(right*100.0)/(num*1.00) print(colored("precision: %.2f%c ! " %(precision,'%'),'red')) global historyMax,history history[epoch]=precision historyMax=max(historyMax,precision) global Misspre for i in range(0,24): for j in range(0,24): Misspre[hash2[i]][hash2[j]]=(100*Miss[hash2[i]][hash2[j]]*1.000)/(sum(Miss[hash2[i]])*1.000) #torch.save(Misspre,'visualize/visualize.t7') #visualize(Misspre) for i in range(0,24): if(Misspre[hash2[i]][hash2[i]]>colorMax[hash2[i]]): colorMax[hash2[i]]=Misspre[hash2[i]][hash2[i]] if((Misspre[hash2[i]][hash2[i]]>85.00) and (precision>65)): save_color_model(model,hash2[i]) if __name__ == '__main__': #save_model(model, PATH_TO_MODEL) #print(np.shape(X)) for epoch in range(1, args.epochs + 1): train(epoch) test(epoch) if historyMax==history[epoch]: save_model(model, PATH_TO_MODEL) if args.automonous_stopping==1:
haha=0 if(epoch>10): for i in range(epoch-10+1,epoch+1): if(Hloss[i]==lossMin): haha=1 if haha==0: break
conditional_block
train_fineall.py
: 1)') parser.add_argument('--epochs', type=int, default=600, metavar='N', help='number of epochs to train (default: 300)') parser.add_argument('--batchsize', type=int, default=128, metavar='N', help='input batch size for training (default: 64)') parser.add_argument('--lr', type=float, default=0.1, metavar='LR', help='learning rate (default: 0.001)') parser.add_argument('--momentum', type=float, default=0.9, metavar='M', help='SGD momentum (default: 0.9)') parser.add_argument('--weight_decay', type=float, default=5e-4, help='SGD weight_decay (default: 5e-4)') parser.add_argument('--automonous_stopping', type=int,default=0, help='automonous_stopping') parser.add_argument('--data',default='data',metavar='NT', help='the data directory') parser.add_argument('--modelpos',default='models/resnet18_pretrained.t7',metavar='NT', help='the data directory') args=parser.parse_args() args.cuda = not args.no_cuda and torch.cuda.is_available() torch.manual_seed(args.seed) if args.cuda: torch.cuda.manual_seed(args.seed) PATH_TO_MODEL = os.path.join('models', 'color_models') category_num=24 history=[0.01]*1000 historyMax=0.01 Hloss=[0.01]*1000 lossMin=1000000.0 hash2=[0, 1, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 2, 20, 21, 22, 23, 3, 4, 5, 6, 7, 8, 9] colorMax=[0.0001]*24 Misspre=np.array([[0.01 for col in range(0,24)] for row in range(0,24)]) # loading pre-trained model print(colored('initializing the model ...',"blue")) print(colored('load model at '+args.modelpos,"blue")) #model = models.resnet18(pretrained=True)#at least 224*224 model=torch.load(args.modelpos) model.fc=nn.Linear(512,24) for param in model.parameters(): param.requires_grads=False model.fc.requires_grads=True model.layer4.requires_grads=True model.layer3.requires_grads=True model.layer2.requires_grads=True model.layer1.requires_grads=True if args.cuda: model.cuda() print(colored('model ==> ',"green")) print(model)
traindir = os.path.join(args.data, 'train') valdir = os.path.join(args.data, 'val') normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]) traindir='data/train' valdir='data/val' #print(os.listdir(traindir)) train_loader1= torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.RandomSizedCrop(224), transforms.RandomHorizontalFlip(), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) train_loader2= torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.Scale(224), #transforms.CenterCrop(224), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) train_loader3= torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.RandomSizedCrop(224), transforms.RandomHorizontalFlip(), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) val_loader = torch.utils.data.DataLoader( datasets.ImageFolder(valdir, transforms.Compose([ transforms.Scale(256), transforms.CenterCrop(224), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=False, num_workers=4, pin_memory=True) def imshow(img): img = img / 2 + 0.5 # unnormalize npimg = img.numpy() plt.imshow(np.transpose(npimg, (1,2,0))) plt.show() #dataiter = iter(train_loader) #print(len(train_loader)) #print(len(val_loader)) #images,labels= dataiter.next() #print(images) #print(labels) # print images #imshow(torchvision.utils.make_grid(images)) # define loss function (criterion) and pptimizer criterion = nn.CrossEntropyLoss().cuda() optimizer = optim.SGD(model.parameters(), lr=args.lr, momentum=args.momentum,weight_decay=args.weight_decay) def save_model(model, name): print('saving the model ...') if not os.path.exists(PATH_TO_MODEL): os.mkdir(PATH_TO_MODEL) torch.save(model,PATH_TO_MODEL+'/'+str(historyMax)+'.t7') print('done.') def visualize(data): for i in range(0,24): print('color '+str(i)+ '!') for j in range(0,24): if(data[i][j]>5): print '[%d][%d] = %.2f ' %(i,j,data[i][j]) , print() for i in range(0,24): print(colored('color %d precision: %.2f !' %(i,data[i][i]),'green')) def save_color_model(model, color): print('saving the color model for %d ...' %color) visualize(Misspre) PATH_TO_COLOR=PATH_TO_MODEL+str(color) if not os.path.exists(PATH_TO_COLOR): os.mkdir(PATH_TO_COLOR) torch.save(model,PATH_TO_COLOR+'/'+str(colorMax[color])+'.t7') print('done.') def train(epoch): model.train() print(colored('training epoch '+ str(epoch) + ' !','blue')) print(colored('loading data!','green')) if epoch%3==0: train_loader=train_loader1 elif epoch%3==1: train_loader=torch.utils.data.DataLoader( datasets.ImageFolder(traindir, transforms.Compose([ transforms.Scale(224), #transforms.CenterCrop(224), transforms.ToTensor(), normalize, ])), batch_size=args.batchsize, shuffle=True, num_workers=4, pin_memory=True) else: train_loader=train_loader3 print(colored('done!','green')) tot_loss=0.0 num=0 right=0 for i, (inputs, targets) in tqdm(enumerate(train_loader)): if args.cuda: inputs=inputs.cuda(async=True) targets=targets.cuda(async=True) inputs_var, targets_var = Variable(inputs), Variable(targets) optimizer.zero_grad() outputs = model(inputs_var) loss=criterion(outputs, targets_var)/(args.batchsize*1.0) loss.backward() optimizer.step() tot_loss=tot_loss+loss.data[0] num=num+inputs.size(0) _,indices=torch.max(outputs.data,1) indices=indices.view(inputs.size(0)) right=right+sum(indices==targets) #print(output.data) #averageloss=2.3 #averageloss=(tot_loss*1.0)/(num*1.00) print(colored("totloss: %.8f ! " %tot_loss,'red')) precision=2.3 precision=(right*100.0)/(num*1.00) print(colored("precision: %.2f%c ! " %(precision,'%'),'red')) global Hloss,lossMin Hloss[epoch]=tot_loss if epoch==1: lossMin=tot_loss else: lossMin=min(lossMin,tot_loss) #print(colored("right: %d ! " %right,'red')) def test(epoch): Miss=[[0 for col in range(0,24)] for row in range(0,24)]# suppose i to be j model.eval() print(colored('Testing!','blue')) tot_loss=0.0 num=0 right=0 for i, (inputs, targets) in tqdm(enumerate(val_loader)): if args.cuda: inputs=inputs.cuda(async=True) targets=targets.cuda(async=True) inputs_var, targets_var = Variable(inputs), Variable(targets) #optimizer.zero_grad() outputs = model(inputs_var) loss=criterion(outputs, targets_var)/(args.batchsize*1.0) #loss.backward() #optimizer.step() tot_loss=tot_loss+loss.data[0] num=num+inputs.size(0)
print(colored('initializing done.',"blue")) # Data loading code
random_line_split
lines.py
visible(lines[0:mid]) struct2 = visible(lines[mid:]) struct = combine(struct1, struct2) # print("visibleD ", struct) return struct def combine(struct1, struct2): # print("combining''''''''''''''''''''''''''''''''''''''") # print(struct1) # print(struct2) # IDEA : We assume that struct1 is from the lower slope half, struct2 from the higher slop half. # Now, we merge the intersection points in struct1 and struct2. We get something like: # p11, p12, p21, p13, p22 etc. some random order. The insight is the point of intersection we're looking for # Must lie between consec. p's. (Or straight up outside the range). # So we sort of pick each interleaving, find the corresponding lines # in that region, and check if their intersection is also in that region. # Another approach is to notice that as we approach form -infinity, struct2 must be lower, # and as we approach +infinity, struct2 must be higer. # The point of intersection is where this flip happens. This is also a reasonable approach, # but the corner casses etc. need to be considered. # Unsaid here is the assumption that there is one and only one point of intersection. # I can't come up with a definite proof, but it seems reasonable nonetheless. # The flippy approach. # Struct1 is required by intergalactic law to be low-slope struct. # if the infinity lines intersect at x < x10 and x20, we are done. Similarly for x > x1n and x2n. infx = intersection(struct1[0][2], struct2[0][2])[0] # print("infx", infx) inf2x = intersection(struct1[-1][2], struct2[-1][2])[0] # print("inf2x", inf2x) if infx <= min(struct1[0][1], struct2[0][1]): final = [[float("-inf"), infx, struct1[0][2]], [infx, struct2[0][1], struct2[0][2]]] + struct2[1:] elif inf2x >= max(struct1[-1][0], struct2[-1][0]): final = struct1[0:-1] + [[struct1[-1][0], inf2x, struct1[-1][2]], [inf2x, float("inf"), struct2[-1][2]]] # Otherwise we truncate the structs to finite lengths. Find the intersection using flipping. else: minx = min(struct1[0][1], struct2[0][1]) maxx = max(struct1[-1][0], struct2[-1][0]) struct1a = confine(struct1, minx, maxx) struct2a = confine(struct2, minx, maxx) intersectionx = struct_intersection(struct1a, struct2a) pos1 = getindex(intersectionx, struct1) pos2 = getindex(intersectionx, struct2) final1 = struct1[0:pos1] + [[struct1[pos1][0], intersectionx, struct1[pos1][2]]] final2 = [[intersectionx, struct2[pos2][1], struct2[pos2][2]]] + struct2[pos2 + 1:] final = final1 + final2 flag = False if flag: print("=1=1=1=11=1=1=1=1=1=1=1=1=1=1=1=1=1=1=1") print(struct1, struct2) print("seem to have combined into") print(final) print("@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@") return final def confine(struct, x1, x2): # print("confinig", struct, x1, x2) newstruct = struct[0:] if newstruct[0][1] > x1: newstruct[0] = [x1, newstruct[0][1], newstruct[0][2]] elif newstruct[0][1] == x1: newstruct = newstruct[1:] if newstruct[-1][0] < x2: newstruct[-1] = [newstruct[-1][0], x2, newstruct[-1][2]] elif newstruct[-1][0] == x2: newstruct = newstruct[:-1] # print("CONNFFFIIIINNNNNEEEEEEDDDDDDD", newstruct) return newstruct def struct_intersection(struct1, struct2): pos1 = binary_flip_search(struct1, struct2) pos2 = binary_flip_search(struct2, struct1) intersectionx = intersection(struct1[pos1][2], struct2[pos2][2])[0] return intersectionx def binary_flip_search(struct, cand): # print("-----------------------------") # print("binary flip search", struct, cand) if len(struct) == 1: if higher(struct[0], cand) is 0: return 0 else: print("ERROR. Flipping didn't happen in: ", struct, cand) mid = len(struct) / 2 higher1 = higher(struct[0], cand) highern = higher(struct[-1], cand) higher_mid = higher(struct[mid], cand) if higher1 is 0: return 0 if highern is 0: return len(struct) - 1 if higher_mid is 0: return mid if higher1 == higher_mid: # print("in call case0|||||||||||||||||||||||") return mid + 1 + binary_flip_search(struct[mid + 1:-1], cand) else: # print("in call case1||||||||||||||||||||||||||")
def higher(region, cand): point1 = [region[0], gety(region[0], region[2])] point2 = [region[1], gety(region[1], region[2])] high1 = high(point1, cand) high2 = high(point2, cand) if high1 and high2: return 1 elif not (high1 or high2): return -1 else: return 0 def high(point, struct): # print("HIGHHIGHHIG", point, struct) line = struct[getindex(point[0], struct)][2] y = gety(point[0], line) # print("Results for :", point, struct, line, y) if point[1] >= y: return True else: return False def getindex(x, struct): if len(struct) == 1: if struct[0][0] <= x <= struct[0][1]: return 0 else: return "Out of range of struct." else: mid = len(struct) / 2 if struct[mid][0] <= x <= struct[mid][1]: return mid elif x < struct[mid][0]: return getindex(x, struct[0:mid]) elif x > struct[mid][1]: return mid + 1 + getindex(x, struct[mid + 1:]) def gety(x, line): return line[0] * x + line[1] def reader(infile): linelist = [] infile = open(infile) lines = infile.readlines() for i in range(1, int(lines[0]) + 1): line = lines[i].split(":") linelist += [[float(line[0]), float(line[1]), i]] return linelist def writer(outfile, struct): outfile = open(outfile, "w") visibles = [] for i in range(0, len(struct)): visibles += [struct[i][2][2]] visibles = sorted(list(set(visibles))) s = str(visibles) s = s[1:-1] s = s.replace("'", "").replace(' ', '') # print(s) outfile.write(s) outfile.close() return s def clean(lines): if len(lines) < 2: return lines i = 1 while i < len(lines): now = lines[i][0] prv = lines[i - 1][0] if now == prv: # print(len(lines)) # print("hahaha. lele fuckru") lines = lines[0:i - 1] + lines[i:] # i += 1 # print(len(lines)) else: i += 1 return lines def runner(inf, outf): lines = reader(inf) lines.sort() lines = clean(lines) # sure = superbrute(lines) struct = visible(lines) s = writer(outf, struct) # surelines = [] # for line in sure: # surelines += [line[2]] # s = str((sorted(surelines)))
return 1 + binary_flip_search(struct[1:mid], cand)
conditional_block
lines.py
10 and x20, we are done. Similarly for x > x1n and x2n. infx = intersection(struct1[0][2], struct2[0][2])[0] # print("infx", infx) inf2x = intersection(struct1[-1][2], struct2[-1][2])[0] # print("inf2x", inf2x) if infx <= min(struct1[0][1], struct2[0][1]): final = [[float("-inf"), infx, struct1[0][2]], [infx, struct2[0][1], struct2[0][2]]] + struct2[1:] elif inf2x >= max(struct1[-1][0], struct2[-1][0]): final = struct1[0:-1] + [[struct1[-1][0], inf2x, struct1[-1][2]], [inf2x, float("inf"), struct2[-1][2]]] # Otherwise we truncate the structs to finite lengths. Find the intersection using flipping. else: minx = min(struct1[0][1], struct2[0][1]) maxx = max(struct1[-1][0], struct2[-1][0]) struct1a = confine(struct1, minx, maxx) struct2a = confine(struct2, minx, maxx) intersectionx = struct_intersection(struct1a, struct2a) pos1 = getindex(intersectionx, struct1) pos2 = getindex(intersectionx, struct2) final1 = struct1[0:pos1] + [[struct1[pos1][0], intersectionx, struct1[pos1][2]]] final2 = [[intersectionx, struct2[pos2][1], struct2[pos2][2]]] + struct2[pos2 + 1:] final = final1 + final2 flag = False if flag: print("=1=1=1=11=1=1=1=1=1=1=1=1=1=1=1=1=1=1=1") print(struct1, struct2) print("seem to have combined into") print(final) print("@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@") return final def confine(struct, x1, x2): # print("confinig", struct, x1, x2) newstruct = struct[0:] if newstruct[0][1] > x1: newstruct[0] = [x1, newstruct[0][1], newstruct[0][2]] elif newstruct[0][1] == x1: newstruct = newstruct[1:] if newstruct[-1][0] < x2: newstruct[-1] = [newstruct[-1][0], x2, newstruct[-1][2]] elif newstruct[-1][0] == x2: newstruct = newstruct[:-1] # print("CONNFFFIIIINNNNNEEEEEEDDDDDDD", newstruct) return newstruct def struct_intersection(struct1, struct2): pos1 = binary_flip_search(struct1, struct2) pos2 = binary_flip_search(struct2, struct1) intersectionx = intersection(struct1[pos1][2], struct2[pos2][2])[0] return intersectionx def binary_flip_search(struct, cand): # print("-----------------------------") # print("binary flip search", struct, cand) if len(struct) == 1: if higher(struct[0], cand) is 0: return 0 else: print("ERROR. Flipping didn't happen in: ", struct, cand) mid = len(struct) / 2 higher1 = higher(struct[0], cand) highern = higher(struct[-1], cand) higher_mid = higher(struct[mid], cand) if higher1 is 0: return 0 if highern is 0: return len(struct) - 1 if higher_mid is 0: return mid if higher1 == higher_mid: # print("in call case0|||||||||||||||||||||||") return mid + 1 + binary_flip_search(struct[mid + 1:-1], cand) else: # print("in call case1||||||||||||||||||||||||||") return 1 + binary_flip_search(struct[1:mid], cand) def higher(region, cand): point1 = [region[0], gety(region[0], region[2])] point2 = [region[1], gety(region[1], region[2])] high1 = high(point1, cand) high2 = high(point2, cand) if high1 and high2: return 1 elif not (high1 or high2): return -1 else: return 0 def high(point, struct): # print("HIGHHIGHHIG", point, struct) line = struct[getindex(point[0], struct)][2] y = gety(point[0], line) # print("Results for :", point, struct, line, y) if point[1] >= y: return True else: return False def getindex(x, struct): if len(struct) == 1: if struct[0][0] <= x <= struct[0][1]: return 0 else: return "Out of range of struct." else: mid = len(struct) / 2 if struct[mid][0] <= x <= struct[mid][1]: return mid elif x < struct[mid][0]: return getindex(x, struct[0:mid]) elif x > struct[mid][1]: return mid + 1 + getindex(x, struct[mid + 1:]) def gety(x, line): return line[0] * x + line[1] def reader(infile): linelist = [] infile = open(infile) lines = infile.readlines() for i in range(1, int(lines[0]) + 1): line = lines[i].split(":") linelist += [[float(line[0]), float(line[1]), i]] return linelist def writer(outfile, struct): outfile = open(outfile, "w") visibles = [] for i in range(0, len(struct)): visibles += [struct[i][2][2]] visibles = sorted(list(set(visibles))) s = str(visibles) s = s[1:-1] s = s.replace("'", "").replace(' ', '') # print(s) outfile.write(s) outfile.close() return s def clean(lines): if len(lines) < 2: return lines i = 1 while i < len(lines): now = lines[i][0] prv = lines[i - 1][0] if now == prv: # print(len(lines)) # print("hahaha. lele fuckru") lines = lines[0:i - 1] + lines[i:] # i += 1 # print(len(lines)) else: i += 1 return lines def runner(inf, outf): lines = reader(inf) lines.sort() lines = clean(lines) # sure = superbrute(lines) struct = visible(lines) s = writer(outf, struct) # surelines = [] # for line in sure: # surelines += [line[2]] # s = str((sorted(surelines))) # s = s[1:-1].replace(' ', '') print(s) return s infile = "input.txt" outfile = "output.txt" def superbrute(lines): visibles = [] for line in lines: if brute(lines, line): visibles += [line] print(visibles) return visibles def brute(lines, mine): # print(len(lines)) intersections = [] for line in lines: if not mine == line: intersections += [intersection(line, mine)[0]] # intersections.sort() ivisible = False print(intersections) for x in intersections: my = gety(x, mine) print('my',x,my) high = True for line in lines: if not mine == line: print('ot',x,gety(x, line)) if gety(x, line) > my: print('other was higher') high = False if high: ivisible = True # print(mine) return ivisible return ivisible import random def generate(n):
mylines = [] for i in range(1, n + 1): m = float(random.uniform(-100000, 100000)) c = float(random.uniform(-100000, 100000)) mylines += [[m, c, i]] f = open('input.txt', 'w') f.write(str(n) + '\n') for line in mylines: f.write(str(line[0]) + ':' + str(line[1]) + '\n') return mylines
identifier_body
lines.py
visible(lines[0:mid]) struct2 = visible(lines[mid:]) struct = combine(struct1, struct2) # print("visibleD ", struct) return struct def combine(struct1, struct2): # print("combining''''''''''''''''''''''''''''''''''''''") # print(struct1) # print(struct2) # IDEA : We assume that struct1 is from the lower slope half, struct2 from the higher slop half. # Now, we merge the intersection points in struct1 and struct2. We get something like: # p11, p12, p21, p13, p22 etc. some random order. The insight is the point of intersection we're looking for # Must lie between consec. p's. (Or straight up outside the range). # So we sort of pick each interleaving, find the corresponding lines # in that region, and check if their intersection is also in that region. # Another approach is to notice that as we approach form -infinity, struct2 must be lower, # and as we approach +infinity, struct2 must be higer. # The point of intersection is where this flip happens. This is also a reasonable approach, # but the corner casses etc. need to be considered. # Unsaid here is the assumption that there is one and only one point of intersection. # I can't come up with a definite proof, but it seems reasonable nonetheless. # The flippy approach. # Struct1 is required by intergalactic law to be low-slope struct. # if the infinity lines intersect at x < x10 and x20, we are done. Similarly for x > x1n and x2n. infx = intersection(struct1[0][2], struct2[0][2])[0] # print("infx", infx) inf2x = intersection(struct1[-1][2], struct2[-1][2])[0] # print("inf2x", inf2x) if infx <= min(struct1[0][1], struct2[0][1]): final = [[float("-inf"), infx, struct1[0][2]], [infx, struct2[0][1], struct2[0][2]]] + struct2[1:] elif inf2x >= max(struct1[-1][0], struct2[-1][0]): final = struct1[0:-1] + [[struct1[-1][0], inf2x, struct1[-1][2]], [inf2x, float("inf"), struct2[-1][2]]] # Otherwise we truncate the structs to finite lengths. Find the intersection using flipping. else: minx = min(struct1[0][1], struct2[0][1]) maxx = max(struct1[-1][0], struct2[-1][0]) struct1a = confine(struct1, minx, maxx) struct2a = confine(struct2, minx, maxx) intersectionx = struct_intersection(struct1a, struct2a) pos1 = getindex(intersectionx, struct1) pos2 = getindex(intersectionx, struct2) final1 = struct1[0:pos1] + [[struct1[pos1][0], intersectionx, struct1[pos1][2]]] final2 = [[intersectionx, struct2[pos2][1], struct2[pos2][2]]] + struct2[pos2 + 1:] final = final1 + final2 flag = False if flag: print("=1=1=1=11=1=1=1=1=1=1=1=1=1=1=1=1=1=1=1") print(struct1, struct2) print("seem to have combined into") print(final) print("@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@") return final def confine(struct, x1, x2): # print("confinig", struct, x1, x2) newstruct = struct[0:] if newstruct[0][1] > x1: newstruct[0] = [x1, newstruct[0][1], newstruct[0][2]] elif newstruct[0][1] == x1: newstruct = newstruct[1:] if newstruct[-1][0] < x2: newstruct[-1] = [newstruct[-1][0], x2, newstruct[-1][2]] elif newstruct[-1][0] == x2: newstruct = newstruct[:-1] # print("CONNFFFIIIINNNNNEEEEEEDDDDDDD", newstruct) return newstruct def struct_intersection(struct1, struct2): pos1 = binary_flip_search(struct1, struct2) pos2 = binary_flip_search(struct2, struct1) intersectionx = intersection(struct1[pos1][2], struct2[pos2][2])[0] return intersectionx def binary_flip_search(struct, cand): # print("-----------------------------") # print("binary flip search", struct, cand) if len(struct) == 1: if higher(struct[0], cand) is 0: return 0 else: print("ERROR. Flipping didn't happen in: ", struct, cand) mid = len(struct) / 2 higher1 = higher(struct[0], cand) highern = higher(struct[-1], cand) higher_mid = higher(struct[mid], cand) if higher1 is 0: return 0 if highern is 0: return len(struct) - 1
return mid if higher1 == higher_mid: # print("in call case0|||||||||||||||||||||||") return mid + 1 + binary_flip_search(struct[mid + 1:-1], cand) else: # print("in call case1||||||||||||||||||||||||||") return 1 + binary_flip_search(struct[1:mid], cand) def higher(region, cand): point1 = [region[0], gety(region[0], region[2])] point2 = [region[1], gety(region[1], region[2])] high1 = high(point1, cand) high2 = high(point2, cand) if high1 and high2: return 1 elif not (high1 or high2): return -1 else: return 0 def high(point, struct): # print("HIGHHIGHHIG", point, struct) line = struct[getindex(point[0], struct)][2] y = gety(point[0], line) # print("Results for :", point, struct, line, y) if point[1] >= y: return True else: return False def getindex(x, struct): if len(struct) == 1: if struct[0][0] <= x <= struct[0][1]: return 0 else: return "Out of range of struct." else: mid = len(struct) / 2 if struct[mid][0] <= x <= struct[mid][1]: return mid elif x < struct[mid][0]: return getindex(x, struct[0:mid]) elif x > struct[mid][1]: return mid + 1 + getindex(x, struct[mid + 1:]) def gety(x, line): return line[0] * x + line[1] def reader(infile): linelist = [] infile = open(infile) lines = infile.readlines() for i in range(1, int(lines[0]) + 1): line = lines[i].split(":") linelist += [[float(line[0]), float(line[1]), i]] return linelist def writer(outfile, struct): outfile = open(outfile, "w") visibles = [] for i in range(0, len(struct)): visibles += [struct[i][2][2]] visibles = sorted(list(set(visibles))) s = str(visibles) s = s[1:-1] s = s.replace("'", "").replace(' ', '') # print(s) outfile.write(s) outfile.close() return s def clean(lines): if len(lines) < 2: return lines i = 1 while i < len(lines): now = lines[i][0] prv = lines[i - 1][0] if now == prv: # print(len(lines)) # print("hahaha. lele fuckru") lines = lines[0:i - 1] + lines[i:] # i += 1 # print(len(lines)) else: i += 1 return lines def runner(inf, outf): lines = reader(inf) lines.sort() lines = clean(lines) # sure = superbrute(lines) struct = visible(lines) s = writer(outf, struct) # surelines = [] # for line in sure: # surelines += [line[2]] # s = str((sorted(surelines))) # s
if higher_mid is 0:
random_line_split
lines.py
must be lower, # and as we approach +infinity, struct2 must be higer. # The point of intersection is where this flip happens. This is also a reasonable approach, # but the corner casses etc. need to be considered. # Unsaid here is the assumption that there is one and only one point of intersection. # I can't come up with a definite proof, but it seems reasonable nonetheless. # The flippy approach. # Struct1 is required by intergalactic law to be low-slope struct. # if the infinity lines intersect at x < x10 and x20, we are done. Similarly for x > x1n and x2n. infx = intersection(struct1[0][2], struct2[0][2])[0] # print("infx", infx) inf2x = intersection(struct1[-1][2], struct2[-1][2])[0] # print("inf2x", inf2x) if infx <= min(struct1[0][1], struct2[0][1]): final = [[float("-inf"), infx, struct1[0][2]], [infx, struct2[0][1], struct2[0][2]]] + struct2[1:] elif inf2x >= max(struct1[-1][0], struct2[-1][0]): final = struct1[0:-1] + [[struct1[-1][0], inf2x, struct1[-1][2]], [inf2x, float("inf"), struct2[-1][2]]] # Otherwise we truncate the structs to finite lengths. Find the intersection using flipping. else: minx = min(struct1[0][1], struct2[0][1]) maxx = max(struct1[-1][0], struct2[-1][0]) struct1a = confine(struct1, minx, maxx) struct2a = confine(struct2, minx, maxx) intersectionx = struct_intersection(struct1a, struct2a) pos1 = getindex(intersectionx, struct1) pos2 = getindex(intersectionx, struct2) final1 = struct1[0:pos1] + [[struct1[pos1][0], intersectionx, struct1[pos1][2]]] final2 = [[intersectionx, struct2[pos2][1], struct2[pos2][2]]] + struct2[pos2 + 1:] final = final1 + final2 flag = False if flag: print("=1=1=1=11=1=1=1=1=1=1=1=1=1=1=1=1=1=1=1") print(struct1, struct2) print("seem to have combined into") print(final) print("@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@") return final def confine(struct, x1, x2): # print("confinig", struct, x1, x2) newstruct = struct[0:] if newstruct[0][1] > x1: newstruct[0] = [x1, newstruct[0][1], newstruct[0][2]] elif newstruct[0][1] == x1: newstruct = newstruct[1:] if newstruct[-1][0] < x2: newstruct[-1] = [newstruct[-1][0], x2, newstruct[-1][2]] elif newstruct[-1][0] == x2: newstruct = newstruct[:-1] # print("CONNFFFIIIINNNNNEEEEEEDDDDDDD", newstruct) return newstruct def struct_intersection(struct1, struct2): pos1 = binary_flip_search(struct1, struct2) pos2 = binary_flip_search(struct2, struct1) intersectionx = intersection(struct1[pos1][2], struct2[pos2][2])[0] return intersectionx def binary_flip_search(struct, cand): # print("-----------------------------") # print("binary flip search", struct, cand) if len(struct) == 1: if higher(struct[0], cand) is 0: return 0 else: print("ERROR. Flipping didn't happen in: ", struct, cand) mid = len(struct) / 2 higher1 = higher(struct[0], cand) highern = higher(struct[-1], cand) higher_mid = higher(struct[mid], cand) if higher1 is 0: return 0 if highern is 0: return len(struct) - 1 if higher_mid is 0: return mid if higher1 == higher_mid: # print("in call case0|||||||||||||||||||||||") return mid + 1 + binary_flip_search(struct[mid + 1:-1], cand) else: # print("in call case1||||||||||||||||||||||||||") return 1 + binary_flip_search(struct[1:mid], cand) def higher(region, cand): point1 = [region[0], gety(region[0], region[2])] point2 = [region[1], gety(region[1], region[2])] high1 = high(point1, cand) high2 = high(point2, cand) if high1 and high2: return 1 elif not (high1 or high2): return -1 else: return 0 def high(point, struct): # print("HIGHHIGHHIG", point, struct) line = struct[getindex(point[0], struct)][2] y = gety(point[0], line) # print("Results for :", point, struct, line, y) if point[1] >= y: return True else: return False def getindex(x, struct): if len(struct) == 1: if struct[0][0] <= x <= struct[0][1]: return 0 else: return "Out of range of struct." else: mid = len(struct) / 2 if struct[mid][0] <= x <= struct[mid][1]: return mid elif x < struct[mid][0]: return getindex(x, struct[0:mid]) elif x > struct[mid][1]: return mid + 1 + getindex(x, struct[mid + 1:]) def gety(x, line): return line[0] * x + line[1] def reader(infile): linelist = [] infile = open(infile) lines = infile.readlines() for i in range(1, int(lines[0]) + 1): line = lines[i].split(":") linelist += [[float(line[0]), float(line[1]), i]] return linelist def writer(outfile, struct): outfile = open(outfile, "w") visibles = [] for i in range(0, len(struct)): visibles += [struct[i][2][2]] visibles = sorted(list(set(visibles))) s = str(visibles) s = s[1:-1] s = s.replace("'", "").replace(' ', '') # print(s) outfile.write(s) outfile.close() return s def clean(lines): if len(lines) < 2: return lines i = 1 while i < len(lines): now = lines[i][0] prv = lines[i - 1][0] if now == prv: # print(len(lines)) # print("hahaha. lele fuckru") lines = lines[0:i - 1] + lines[i:] # i += 1 # print(len(lines)) else: i += 1 return lines def runner(inf, outf): lines = reader(inf) lines.sort() lines = clean(lines) # sure = superbrute(lines) struct = visible(lines) s = writer(outf, struct) # surelines = [] # for line in sure: # surelines += [line[2]] # s = str((sorted(surelines))) # s = s[1:-1].replace(' ', '') print(s) return s infile = "input.txt" outfile = "output.txt" def superbrute(lines): visibles = [] for line in lines: if brute(lines, line): visibles += [line] print(visibles) return visibles def brute(lines, mine): # print(len(lines)) intersections = [] for line in lines: if not mine == line: intersections += [intersection(line, mine)[0]] # intersections.sort() ivisible = False print(intersections) for x in intersections: my = gety(x, mine) print('my',x,my) high = True for line in lines: if not mine == line: print('ot',x,gety(x, line)) if gety(x, line) > my: print('other was higher') high = False if high: ivisible = True # print(mine) return ivisible return ivisible import random def
generate
identifier_name
tower-of-power.py
?:[\w-]+(?:\s*,\s*[\w-]+)*)?\s*)\)' clss_re = r'(?:\.([\w-]+))?' text_re = r':\s*(.*)?' line_re = re.compile( wsop_re + name_re + wsop_re + deps_re + wsop_re + clss_re + wsop_re + text_re ) match = line_re.match(line) if match: name = match.group(1) deps = match.group(2).replace(',', ' ').split() clss = match.group(3) or 'box-generic' text = match.group(4) if len(deps) == 0: deps = [dag.ROOT] self.insert(name, deps) self.clss[name] = clss self.text[name] = text elif len(line.split()) == 0 or line[0] == '#': pass else: raise DAGException("Syntax error on line: `" + line + "`") def load_file(self, text): lines = text.split('\n') for line in lines: self.load_line(line) # A lot of the invariants on this DAG are maintained by complaining loudly as # soon as they are violated. So, you have to insert nodes in the correct order. # Other than that, this insertion routine is unexciting. def insert(self, name, deps): if name in self.get_nodes(): raise DAGException("Already added this name to DAG.") for d in deps: if d not in self.deps: raise DAGException("Unknown dependency: %s (for %s)"%(d, name)) self.deps[name] = list(deps) # This is also unexciting. def get_nodes(self): return self.deps.keys() # This, however, is kind of interesting. Notice that Parker's tower doesn't # distinguish between "direct dependencies" and "transitive dependencies": if # A depends on B and C, and B itself depends on C, then the box for A doesn't # need to sit on the boxes for both B and C (in fact, it can't!). The correct # drawing is "A on B, B on C", with the understanding that A "of course" # depends on C as well. # # The following predicates sort out this mess, by giving me a way to check if a # dependency is "direct" or "transitive" in that sense. Transitive dependencies # can be ignored. def get_dependencies(self, node): return self.deps[node] def is_dependency(self, node, dep): if (node, dep) in self.dep_cache: return self.dep_cache[(node, dep)] shallow_deps = self.get_dependencies(node) if dep in shallow_deps: self.dep_cache[(node, dep)] = True return True for sd in shallow_deps: if self.is_dependency(sd, dep): self.dep_cache[(node, dep)] = True return True self.dep_cache[(node, dep)] = False return False def is_transitive_dependency(self, node, dep): for sd in self.get_dependencies(node): if self.is_dependency(sd, dep): return True return False def is_direct_dependency(self, node, dep): return self.is_dependency(node, dep) and\ not self.is_transitive_dependency(node, dep) # Okay, okay, fine, I'll start talking about the solver now. def solve(self): # The way it works is, each box is represented by four symbolic integers, # representing the X/Y coordinates of its top-left and bottom-right vertices. # (Note, however, that because computers are silly, the Y coordinates DECREASE # as you go UP the tower. Just something to keep in mind. Otherwise we get # upside-down stalactite-towers.) svs = {} solver = z3.Solver() for node in self.get_nodes(): svs[node] = ( (z3.Int(node+'_x0'), z3.Int(node+'_y0')), (z3.Int(node+'_x1'), z3.Int(node+'_y1')) ) # Almost immediately, we need to make some sanity assertions. We want the # top-left corner to actually be "to the left" and "on top of" the bottom-right # corner, so we have to tell the solver that. solver.add(svs[node][0][0] < svs[node][1][0]) solver.add(svs[node][0][1] < svs[node][1][1]) # There's also a bit of logic here to automatically make boxes taller if they # have a lot of text, so that text doesn't overflow awkwardly. This is janky, # but it works! solver.add( svs[node][1][1] - svs[node][0][1] >=\ (len(self.text[node].split('\\')) - len(self.text[node].split('\\')) / 2) ) # And finally, we enforce that everything happens in the first quadrant. solver.add(svs[node][0][0] >= 0) # Now we can put root (recall, the "ground") literally on the ground! solver.add(svs[dag.ROOT][0][0] == 0) solver.add(svs[dag.ROOT][0][1] == 0) # Up next, we enforce that no boxes intersect. This is done by checking if the # X and Y ranges are disjoint (at least one needs to be -- but not necessarily # both!). def ranges_disjoint(x0min, x0max, x1min, x1max): return z3.Or(x0min >= x1max, x0max <= x1min) for node1 in self.get_nodes(): for node2 in self.get_nodes(): if node1 != node2: solver.add( z3.Or( ranges_disjoint( svs[node1][0][0], svs[node1][1][0], svs[node2][0][0], svs[node2][1][0] ), ranges_disjoint( svs[node1][0][1], svs[node1][1][1], svs[node2][0][1], svs[node2][1][1] ) ) ) # This is the hard one: for each pair of nodes, it creates an "A is on top of # B" assertion, and then asserts either it or its negation, depending on # whether or not B is a direct dependency of A. for node in self.get_nodes(): for dep in self.get_nodes(): on_top = z3.And( # When is "A" on top of "B"? There are two conditions: # First, A's box's floor is directly on B's box's ceiling. svs[node][1][1] == svs[dep][0][1], # Second, the boxes have intersecting X ranges. z3.Not( ranges_disjoint( svs[node][0][0], svs[node][1][0], svs[dep] [0][0], svs[dep] [1][0] ) ) ) if self.is_direct_dependency(node, dep): solver.add(on_top) else: solver.add(z3.Not(on_top)) # Finally, for the sake of ~aesthetics~, there's a bit of logic to # automatically minimize the total perimeter of all the blocks. (Why not area, # you ask? Because area is nonlinear and the solver takes *much* longer to work # with such constrants!) def perimeter(node): return (svs[node][1][0] - svs[node][0][0]) + (svs[node][1][1] - svs[node][0][1]) total_perim = sum([perimeter(node) for node in self.get_nodes()]) # (That's what the loop is for: it keeps asking the solver to "do better" until # the solver can't do any better and gives up. It may or may not be a metaphor # for life.) rects = None perim_tgt = len(self.get_nodes()) * 4 * 3 while True: perim_tgt -= 1 solver.add(total_perim < perim_tgt) check = solver.check() if check == z3.sat: model = solver.model() rects = [] # I translate the solver output into SVG coordinates using some hardcoded # scaling factors and randomized fudge factors. for node in self.get_nodes():
x0 = model.eval(svs[node][0][0]) y0 = model.eval(svs[node][0][1]) x1 = model.eval(svs[node][1][0]) y1 = model.eval(svs[node][1][1]) import random x0 = int(str(x0)) * 160 + 10 + random.choice(range(10)) y0 = int(str(y0)) * 50 x1 = int(str(x1)) * 160 - 10 + random.choice(range(10)) y1 = int(str(y1)) * 50 rects.append((node, x0, y0, x1, y1))
conditional_block
tower-of-power.py
some algorithms below self.insert(dag.ROOT, []) # The following incomprehensible regular expressions define the BOX file format # which I invented to make it easier to specify DAGs for this project. def load_line(self, line): import re wsop_re = r'\s*' name_re = r'([\w-]+)' deps_re = r'\((\s*(?:[\w-]+(?:\s*,\s*[\w-]+)*)?\s*)\)' clss_re = r'(?:\.([\w-]+))?' text_re = r':\s*(.*)?' line_re = re.compile( wsop_re + name_re + wsop_re + deps_re + wsop_re + clss_re + wsop_re + text_re ) match = line_re.match(line) if match: name = match.group(1) deps = match.group(2).replace(',', ' ').split() clss = match.group(3) or 'box-generic' text = match.group(4) if len(deps) == 0: deps = [dag.ROOT] self.insert(name, deps) self.clss[name] = clss self.text[name] = text elif len(line.split()) == 0 or line[0] == '#': pass else: raise DAGException("Syntax error on line: `" + line + "`") def load_file(self, text): lines = text.split('\n') for line in lines: self.load_line(line) # A lot of the invariants on this DAG are maintained by complaining loudly as # soon as they are violated. So, you have to insert nodes in the correct order. # Other than that, this insertion routine is unexciting. def insert(self, name, deps): if name in self.get_nodes(): raise DAGException("Already added this name to DAG.") for d in deps: if d not in self.deps: raise DAGException("Unknown dependency: %s (for %s)"%(d, name)) self.deps[name] = list(deps) # This is also unexciting. def get_nodes(self): return self.deps.keys() # This, however, is kind of interesting. Notice that Parker's tower doesn't # distinguish between "direct dependencies" and "transitive dependencies": if # A depends on B and C, and B itself depends on C, then the box for A doesn't # need to sit on the boxes for both B and C (in fact, it can't!). The correct # drawing is "A on B, B on C", with the understanding that A "of course" # depends on C as well. # # The following predicates sort out this mess, by giving me a way to check if a # dependency is "direct" or "transitive" in that sense. Transitive dependencies # can be ignored. def get_dependencies(self, node): return self.deps[node] def is_dependency(self, node, dep): if (node, dep) in self.dep_cache: return self.dep_cache[(node, dep)] shallow_deps = self.get_dependencies(node) if dep in shallow_deps: self.dep_cache[(node, dep)] = True return True for sd in shallow_deps: if self.is_dependency(sd, dep): self.dep_cache[(node, dep)] = True return True self.dep_cache[(node, dep)] = False return False def is_transitive_dependency(self, node, dep):
def is_direct_dependency(self, node, dep): return self.is_dependency(node, dep) and\ not self.is_transitive_dependency(node, dep) # Okay, okay, fine, I'll start talking about the solver now. def solve(self): # The way it works is, each box is represented by four symbolic integers, # representing the X/Y coordinates of its top-left and bottom-right vertices. # (Note, however, that because computers are silly, the Y coordinates DECREASE # as you go UP the tower. Just something to keep in mind. Otherwise we get # upside-down stalactite-towers.) svs = {} solver = z3.Solver() for node in self.get_nodes(): svs[node] = ( (z3.Int(node+'_x0'), z3.Int(node+'_y0')), (z3.Int(node+'_x1'), z3.Int(node+'_y1')) ) # Almost immediately, we need to make some sanity assertions. We want the # top-left corner to actually be "to the left" and "on top of" the bottom-right # corner, so we have to tell the solver that. solver.add(svs[node][0][0] < svs[node][1][0]) solver.add(svs[node][0][1] < svs[node][1][1]) # There's also a bit of logic here to automatically make boxes taller if they # have a lot of text, so that text doesn't overflow awkwardly. This is janky, # but it works! solver.add( svs[node][1][1] - svs[node][0][1] >=\ (len(self.text[node].split('\\')) - len(self.text[node].split('\\')) / 2) ) # And finally, we enforce that everything happens in the first quadrant. solver.add(svs[node][0][0] >= 0) # Now we can put root (recall, the "ground") literally on the ground! solver.add(svs[dag.ROOT][0][0] == 0) solver.add(svs[dag.ROOT][0][1] == 0) # Up next, we enforce that no boxes intersect. This is done by checking if the # X and Y ranges are disjoint (at least one needs to be -- but not necessarily # both!). def ranges_disjoint(x0min, x0max, x1min, x1max): return z3.Or(x0min >= x1max, x0max <= x1min) for node1 in self.get_nodes(): for node2 in self.get_nodes(): if node1 != node2: solver.add( z3.Or( ranges_disjoint( svs[node1][0][0], svs[node1][1][0], svs[node2][0][0], svs[node2][1][0] ), ranges_disjoint( svs[node1][0][1], svs[node1][1][1], svs[node2][0][1], svs[node2][1][1] ) ) ) # This is the hard one: for each pair of nodes, it creates an "A is on top of # B" assertion, and then asserts either it or its negation, depending on # whether or not B is a direct dependency of A. for node in self.get_nodes(): for dep in self.get_nodes(): on_top = z3.And( # When is "A" on top of "B"? There are two conditions: # First, A's box's floor is directly on B's box's ceiling. svs[node][1][1] == svs[dep][0][1], # Second, the boxes have intersecting X ranges. z3.Not( ranges_disjoint( svs[node][0][0], svs[node][1][0], svs[dep] [0][0], svs[dep] [1][0] ) ) ) if self.is_direct_dependency(node, dep): solver.add(on_top) else: solver.add(z3.Not(on_top)) # Finally, for the sake of ~aesthetics~, there's a bit of logic to # automatically minimize the total perimeter of all the blocks. (Why not area, # you ask? Because area is nonlinear and the solver takes *much* longer to work # with such constrants!) def perimeter(node): return (svs[node][1][0] - svs[node][0][0]) + (svs[node][1][1] - svs[node][0][1]) total_perim = sum([perimeter(node) for node in self.get_nodes()]) # (That's what the loop is for: it keeps asking the solver to "do better" until # the solver can't do any better and gives up. It may or may not be a metaphor # for life.) rects = None perim_tgt = len(self.get_nodes()) * 4 * 3 while True: perim_tgt -= 1 solver.add(total_perim < perim_tgt) check = solver.check() if check == z3.sat: model = solver.model() rects = [] # I translate the solver output into SVG coordinates using some hardcoded # scaling factors and randomized fudge factors. for node in self.get_nodes(): x0 = model.eval(svs[node][0][0]) y0 = model.eval(svs[node][0][1]) x1 = model.eval(svs[node][1][0]) y1 = model.eval(svs[node][1][1]) import random x0 = int(str(x0)) * 1
for sd in self.get_dependencies(node): if self.is_dependency(sd, dep): return True return False
identifier_body
tower-of-power.py
some algorithms below self.insert(dag.ROOT, []) # The following incomprehensible regular expressions define the BOX file format # which I invented to make it easier to specify DAGs for this project. def load_line(self, line): import re wsop_re = r'\s*' name_re = r'([\w-]+)' deps_re = r'\((\s*(?:[\w-]+(?:\s*,\s*[\w-]+)*)?\s*)\)' clss_re = r'(?:\.([\w-]+))?' text_re = r':\s*(.*)?' line_re = re.compile( wsop_re + name_re + wsop_re + deps_re + wsop_re + clss_re + wsop_re + text_re ) match = line_re.match(line) if match: name = match.group(1) deps = match.group(2).replace(',', ' ').split() clss = match.group(3) or 'box-generic' text = match.group(4) if len(deps) == 0: deps = [dag.ROOT] self.insert(name, deps) self.clss[name] = clss self.text[name] = text elif len(line.split()) == 0 or line[0] == '#': pass else: raise DAGException("Syntax error on line: `" + line + "`") def load_file(self, text): lines = text.split('\n') for line in lines: self.load_line(line) # A lot of the invariants on this DAG are maintained by complaining loudly as # soon as they are violated. So, you have to insert nodes in the correct order. # Other than that, this insertion routine is unexciting. def insert(self, name, deps): if name in self.get_nodes(): raise DAGException("Already added this name to DAG.") for d in deps: if d not in self.deps: raise DAGException("Unknown dependency: %s (for %s)"%(d, name)) self.deps[name] = list(deps) # This is also unexciting. def get_nodes(self): return self.deps.keys() # This, however, is kind of interesting. Notice that Parker's tower doesn't # distinguish between "direct dependencies" and "transitive dependencies": if # A depends on B and C, and B itself depends on C, then the box for A doesn't # need to sit on the boxes for both B and C (in fact, it can't!). The correct # drawing is "A on B, B on C", with the understanding that A "of course" # depends on C as well. # # The following predicates sort out this mess, by giving me a way to check if a # dependency is "direct" or "transitive" in that sense. Transitive dependencies # can be ignored. def get_dependencies(self, node): return self.deps[node] def is_dependency(self, node, dep): if (node, dep) in self.dep_cache: return self.dep_cache[(node, dep)] shallow_deps = self.get_dependencies(node) if dep in shallow_deps: self.dep_cache[(node, dep)] = True return True for sd in shallow_deps: if self.is_dependency(sd, dep): self.dep_cache[(node, dep)] = True return True self.dep_cache[(node, dep)] = False return False def is_transitive_dependency(self, node, dep): for sd in self.get_dependencies(node): if self.is_dependency(sd, dep): return True return False def is_direct_dependency(self, node, dep): return self.is_dependency(node, dep) and\ not self.is_transitive_dependency(node, dep) # Okay, okay, fine, I'll start talking about the solver now. def solve(self): # The way it works is, each box is represented by four symbolic integers, # representing the X/Y coordinates of its top-left and bottom-right vertices. # (Note, however, that because computers are silly, the Y coordinates DECREASE # as you go UP the tower. Just something to keep in mind. Otherwise we get # upside-down stalactite-towers.) svs = {} solver = z3.Solver() for node in self.get_nodes(): svs[node] = ( (z3.Int(node+'_x0'), z3.Int(node+'_y0')), (z3.Int(node+'_x1'), z3.Int(node+'_y1')) ) # Almost immediately, we need to make some sanity assertions. We want the # top-left corner to actually be "to the left" and "on top of" the bottom-right # corner, so we have to tell the solver that. solver.add(svs[node][0][0] < svs[node][1][0]) solver.add(svs[node][0][1] < svs[node][1][1]) # There's also a bit of logic here to automatically make boxes taller if they # have a lot of text, so that text doesn't overflow awkwardly. This is janky, # but it works! solver.add( svs[node][1][1] - svs[node][0][1] >=\ (len(self.text[node].split('\\')) - len(self.text[node].split('\\')) / 2) ) # And finally, we enforce that everything happens in the first quadrant. solver.add(svs[node][0][0] >= 0) # Now we can put root (recall, the "ground") literally on the ground! solver.add(svs[dag.ROOT][0][0] == 0) solver.add(svs[dag.ROOT][0][1] == 0) # Up next, we enforce that no boxes intersect. This is done by checking if the # X and Y ranges are disjoint (at least one needs to be -- but not necessarily # both!). def
(x0min, x0max, x1min, x1max): return z3.Or(x0min >= x1max, x0max <= x1min) for node1 in self.get_nodes(): for node2 in self.get_nodes(): if node1 != node2: solver.add( z3.Or( ranges_disjoint( svs[node1][0][0], svs[node1][1][0], svs[node2][0][0], svs[node2][1][0] ), ranges_disjoint( svs[node1][0][1], svs[node1][1][1], svs[node2][0][1], svs[node2][1][1] ) ) ) # This is the hard one: for each pair of nodes, it creates an "A is on top of # B" assertion, and then asserts either it or its negation, depending on # whether or not B is a direct dependency of A. for node in self.get_nodes(): for dep in self.get_nodes(): on_top = z3.And( # When is "A" on top of "B"? There are two conditions: # First, A's box's floor is directly on B's box's ceiling. svs[node][1][1] == svs[dep][0][1], # Second, the boxes have intersecting X ranges. z3.Not( ranges_disjoint( svs[node][0][0], svs[node][1][0], svs[dep] [0][0], svs[dep] [1][0] ) ) ) if self.is_direct_dependency(node, dep): solver.add(on_top) else: solver.add(z3.Not(on_top)) # Finally, for the sake of ~aesthetics~, there's a bit of logic to # automatically minimize the total perimeter of all the blocks. (Why not area, # you ask? Because area is nonlinear and the solver takes *much* longer to work # with such constrants!) def perimeter(node): return (svs[node][1][0] - svs[node][0][0]) + (svs[node][1][1] - svs[node][0][1]) total_perim = sum([perimeter(node) for node in self.get_nodes()]) # (That's what the loop is for: it keeps asking the solver to "do better" until # the solver can't do any better and gives up. It may or may not be a metaphor # for life.) rects = None perim_tgt = len(self.get_nodes()) * 4 * 3 while True: perim_tgt -= 1 solver.add(total_perim < perim_tgt) check = solver.check() if check == z3.sat: model = solver.model() rects = [] # I translate the solver output into SVG coordinates using some hardcoded # scaling factors and randomized fudge factors. for node in self.get_nodes(): x0 = model.eval(svs[node][0][0]) y0 = model.eval(svs[node][0][1]) x1 = model.eval(svs[node][1][0]) y1 = model.eval(svs[node][1][1]) import random x0 = int(str(x0)) *
ranges_disjoint
identifier_name
tower-of-power.py
some algorithms below self.insert(dag.ROOT, []) # The following incomprehensible regular expressions define the BOX file format # which I invented to make it easier to specify DAGs for this project. def load_line(self, line): import re wsop_re = r'\s*' name_re = r'([\w-]+)' deps_re = r'\((\s*(?:[\w-]+(?:\s*,\s*[\w-]+)*)?\s*)\)' clss_re = r'(?:\.([\w-]+))?' text_re = r':\s*(.*)?' line_re = re.compile( wsop_re + name_re + wsop_re + deps_re + wsop_re + clss_re + wsop_re + text_re ) match = line_re.match(line) if match: name = match.group(1) deps = match.group(2).replace(',', ' ').split() clss = match.group(3) or 'box-generic' text = match.group(4) if len(deps) == 0: deps = [dag.ROOT] self.insert(name, deps) self.clss[name] = clss self.text[name] = text elif len(line.split()) == 0 or line[0] == '#': pass else: raise DAGException("Syntax error on line: `" + line + "`") def load_file(self, text): lines = text.split('\n') for line in lines: self.load_line(line) # A lot of the invariants on this DAG are maintained by complaining loudly as # soon as they are violated. So, you have to insert nodes in the correct order. # Other than that, this insertion routine is unexciting. def insert(self, name, deps): if name in self.get_nodes(): raise DAGException("Already added this name to DAG.") for d in deps: if d not in self.deps: raise DAGException("Unknown dependency: %s (for %s)"%(d, name)) self.deps[name] = list(deps) # This is also unexciting. def get_nodes(self): return self.deps.keys() # This, however, is kind of interesting. Notice that Parker's tower doesn't # distinguish between "direct dependencies" and "transitive dependencies": if # A depends on B and C, and B itself depends on C, then the box for A doesn't # need to sit on the boxes for both B and C (in fact, it can't!). The correct # drawing is "A on B, B on C", with the understanding that A "of course" # depends on C as well. # # The following predicates sort out this mess, by giving me a way to check if a # dependency is "direct" or "transitive" in that sense. Transitive dependencies # can be ignored. def get_dependencies(self, node): return self.deps[node] def is_dependency(self, node, dep): if (node, dep) in self.dep_cache: return self.dep_cache[(node, dep)] shallow_deps = self.get_dependencies(node) if dep in shallow_deps: self.dep_cache[(node, dep)] = True return True for sd in shallow_deps: if self.is_dependency(sd, dep): self.dep_cache[(node, dep)] = True return True self.dep_cache[(node, dep)] = False return False def is_transitive_dependency(self, node, dep): for sd in self.get_dependencies(node): if self.is_dependency(sd, dep): return True return False def is_direct_dependency(self, node, dep): return self.is_dependency(node, dep) and\ not self.is_transitive_dependency(node, dep) # Okay, okay, fine, I'll start talking about the solver now. def solve(self): # The way it works is, each box is represented by four symbolic integers, # representing the X/Y coordinates of its top-left and bottom-right vertices. # (Note, however, that because computers are silly, the Y coordinates DECREASE # as you go UP the tower. Just something to keep in mind. Otherwise we get
svs = {} solver = z3.Solver() for node in self.get_nodes(): svs[node] = ( (z3.Int(node+'_x0'), z3.Int(node+'_y0')), (z3.Int(node+'_x1'), z3.Int(node+'_y1')) ) # Almost immediately, we need to make some sanity assertions. We want the # top-left corner to actually be "to the left" and "on top of" the bottom-right # corner, so we have to tell the solver that. solver.add(svs[node][0][0] < svs[node][1][0]) solver.add(svs[node][0][1] < svs[node][1][1]) # There's also a bit of logic here to automatically make boxes taller if they # have a lot of text, so that text doesn't overflow awkwardly. This is janky, # but it works! solver.add( svs[node][1][1] - svs[node][0][1] >=\ (len(self.text[node].split('\\')) - len(self.text[node].split('\\')) / 2) ) # And finally, we enforce that everything happens in the first quadrant. solver.add(svs[node][0][0] >= 0) # Now we can put root (recall, the "ground") literally on the ground! solver.add(svs[dag.ROOT][0][0] == 0) solver.add(svs[dag.ROOT][0][1] == 0) # Up next, we enforce that no boxes intersect. This is done by checking if the # X and Y ranges are disjoint (at least one needs to be -- but not necessarily # both!). def ranges_disjoint(x0min, x0max, x1min, x1max): return z3.Or(x0min >= x1max, x0max <= x1min) for node1 in self.get_nodes(): for node2 in self.get_nodes(): if node1 != node2: solver.add( z3.Or( ranges_disjoint( svs[node1][0][0], svs[node1][1][0], svs[node2][0][0], svs[node2][1][0] ), ranges_disjoint( svs[node1][0][1], svs[node1][1][1], svs[node2][0][1], svs[node2][1][1] ) ) ) # This is the hard one: for each pair of nodes, it creates an "A is on top of # B" assertion, and then asserts either it or its negation, depending on # whether or not B is a direct dependency of A. for node in self.get_nodes(): for dep in self.get_nodes(): on_top = z3.And( # When is "A" on top of "B"? There are two conditions: # First, A's box's floor is directly on B's box's ceiling. svs[node][1][1] == svs[dep][0][1], # Second, the boxes have intersecting X ranges. z3.Not( ranges_disjoint( svs[node][0][0], svs[node][1][0], svs[dep] [0][0], svs[dep] [1][0] ) ) ) if self.is_direct_dependency(node, dep): solver.add(on_top) else: solver.add(z3.Not(on_top)) # Finally, for the sake of ~aesthetics~, there's a bit of logic to # automatically minimize the total perimeter of all the blocks. (Why not area, # you ask? Because area is nonlinear and the solver takes *much* longer to work # with such constrants!) def perimeter(node): return (svs[node][1][0] - svs[node][0][0]) + (svs[node][1][1] - svs[node][0][1]) total_perim = sum([perimeter(node) for node in self.get_nodes()]) # (That's what the loop is for: it keeps asking the solver to "do better" until # the solver can't do any better and gives up. It may or may not be a metaphor # for life.) rects = None perim_tgt = len(self.get_nodes()) * 4 * 3 while True: perim_tgt -= 1 solver.add(total_perim < perim_tgt) check = solver.check() if check == z3.sat: model = solver.model() rects = [] # I translate the solver output into SVG coordinates using some hardcoded # scaling factors and randomized fudge factors. for node in self.get_nodes(): x0 = model.eval(svs[node][0][0]) y0 = model.eval(svs[node][0][1]) x1 = model.eval(svs[node][1][0]) y1 = model.eval(svs[node][1][1]) import random x0 = int(str(x0)) * 1
# upside-down stalactite-towers.)
random_line_split
pool.rs
a usize /// which will the be stored for display. pub fn custom_subgraph<OP, R, START, END, S>(tag: &'static str, start: START, end: END, op: OP) -> R where OP: FnOnce() -> R, START: FnOnce() -> S, END: FnOnce(S) -> usize, { let s = start(); start_subgraph(tag); let r = op(); let measured_value = end(s); end_subgraph(tag, measured_value); r } /// Stop current task (virtually) and start a subgraph. /// You most likely don't need to call this function directly but `subgraph` instead. pub fn start_subgraph(tag: &'static str) { let subgraph_start_task_id = next_task_id(); logs!( // log child's work and dependencies. RayonEvent::Child(subgraph_start_task_id), // end current task RayonEvent::TaskEnd(now()), // execute full sequential task RayonEvent::TaskStart(subgraph_start_task_id, now()), RayonEvent::SubgraphStart(tag) ); } /// Stop current task (virtually) and end a subgraph. /// You most likely don't need to call this function directly but `subgraph` instead. pub fn end_subgraph(tag: &'static str, measured_value: usize) { let continuation_task_id = next_task_id(); logs!( RayonEvent::SubgraphEnd(tag, measured_value), RayonEvent::Child(continuation_task_id), RayonEvent::TaskEnd(now()), // start continuation task RayonEvent::TaskStart(continuation_task_id, now()) ); } /// Identical to `join`, except that the closures have a parameter /// that provides context for the way the closure has been called, /// especially indicating whether they're executing on a different /// thread than where `join_context` was called. This will occur if /// the second job is stolen by a different thread, or if /// `join_context` was called from outside the thread pool to begin /// with. pub fn join_context<A, B, RA, RB>(oper_a: A, oper_b: B) -> (RA, RB) where A: FnOnce(FnContext) -> RA + Send, B: FnOnce(FnContext) -> RB + Send, RA: Send, RB: Send, { let id_c = next_task_id(); let id_a = next_task_id(); let ca = |c| { log(RayonEvent::TaskStart(id_a, now())); let result = oper_a(c); logs!(RayonEvent::Child(id_c), RayonEvent::TaskEnd(now())); result }; let id_b = next_task_id(); let cb = |c| { log(RayonEvent::TaskStart(id_b, now())); let result = oper_b(c); logs!(RayonEvent::Child(id_c), RayonEvent::TaskEnd(now())); result }; logs!( RayonEvent::Child(id_a), RayonEvent::Child(id_b), RayonEvent::TaskEnd(now()) ); let r = rayon::join_context(ca, cb); log(RayonEvent::TaskStart(id_c, now())); r } /// Takes two closures and *potentially* runs them in parallel. It /// returns a pair of the results from those closures. /// /// Conceptually, calling `join()` is similar to spawning two threads, /// one executing each of the two closures. However, the /// implementation is quite different and incurs very low /// overhead. The underlying technique is called "work stealing": the /// Rayon runtime uses a fixed pool of worker threads and attempts to /// only execute code in parallel when there are idle CPUs to handle /// it. /// /// When `join` is called from outside the thread pool, the calling /// thread will block while the closures execute in the pool. When /// `join` is called within the pool, the calling thread still actively /// participates in the thread pool. It will begin by executing closure /// A (on the current thread). While it is doing that, it will advertise /// closure B as being available for other threads to execute. Once closure A /// has completed, the current thread will try to execute closure B; /// if however closure B has been stolen, then it will look for other work /// while waiting for the thief to fully execute closure B. (This is the /// typical work-stealing strategy). /// /// # Examples /// /// This example uses join to perform a quick-sort (note this is not a /// particularly optimized implementation: if you **actually** want to /// sort for real, you should prefer [the `par_sort` method] offered /// by Rayon). /// /// [the `par_sort` method]: ../rayon/slice/trait.ParallelSliceMut.html#method.par_sort /// /// ```rust /// let mut v = vec![5, 1, 8, 22, 0, 44]; /// quick_sort(&mut v); /// assert_eq!(v, vec![0, 1, 5, 8, 22, 44]); /// /// fn quick_sort<T:PartialOrd+Send>(v: &mut [T]) { /// if v.len() > 1 { /// let mid = partition(v); /// let (lo, hi) = v.split_at_mut(mid); /// rayon::join(|| quick_sort(lo), /// || quick_sort(hi)); /// } /// } /// /// // Partition rearranges all items `<=` to the pivot /// // item (arbitrary selected to be the last item in the slice) /// // to the first half of the slice. It then returns the /// // "dividing point" where the pivot is placed. /// fn partition<T:PartialOrd+Send>(v: &mut [T]) -> usize { /// let pivot = v.len() - 1; /// let mut i = 0; /// for j in 0..pivot { /// if v[j] <= v[pivot] { /// v.swap(i, j); /// i += 1; /// } /// } /// v.swap(i, pivot); /// i /// } /// ``` /// /// # Warning about blocking I/O /// /// The assumption is that the closures given to `join()` are /// CPU-bound tasks that do not perform I/O or other blocking /// operations. If you do perform I/O, and that I/O should block /// (e.g., waiting for a network request), the overall performance may /// be poor. Moreover, if you cause one closure to be blocked waiting /// on another (for example, using a channel), that could lead to a /// deadlock. /// /// # Panics /// /// No matter what happens, both closures will always be executed. If /// a single closure panics, whether it be the first or second /// closure, that panic will be propagated and hence `join()` will /// panic with the same panic value. If both closures panic, `join()` /// will panic with the panic value from the first closure. pub fn join<A, B, RA, RB>(oper_a: A, oper_b: B) -> (RA, RB) where A: FnOnce() -> RA + Send, B: FnOnce() -> RB + Send, RA: Send, RB: Send, { let id_c = next_task_id(); let id_a = next_task_id(); let ca = || { log(RayonEvent::TaskStart(id_a, now())); let result = oper_a(); logs!(RayonEvent::Child(id_c), RayonEvent::TaskEnd(now())); result }; let id_b = next_task_id(); let cb = || { log(RayonEvent::TaskStart(id_b, now())); let result = oper_b(); logs!(RayonEvent::Child(id_c), RayonEvent::TaskEnd(now())); result }; logs!( RayonEvent::Child(id_a), RayonEvent::Child(id_b), RayonEvent::TaskEnd(now()) ); let r = rayon::join(ca, cb); log(RayonEvent::TaskStart(id_c, now())); r } // small global counter to increment file names static INSTALL_COUNT: AtomicUsize = AtomicUsize::new(0); /// We wrap rayon's pool into our own struct to overload the install method. pub struct ThreadPool { pub(crate) logs: Arc<Mutex<Vec<Arc<Storage<RayonEvent>>>>>, pub(crate) pool: rayon::ThreadPool, } impl ThreadPool { /// Reset all logs and counters to initial condition. fn reset(&self) { NEXT_TASK_ID.store(0, Ordering::SeqCst); NEXT_ITERATOR_ID.store(0, Ordering::SeqCst); let logs = &*self.logs.lock().unwrap(); // oh yeah baby for log in logs { log.clear(); } } /// Execute given closure in the thread pool, logging it's task as the initial one.
/// After running, we post-process the logs and return a `RunLog` together with the closure's /// result. pub fn logging_install<OP, R>(&self, op: OP) -> (R, RunLog)
random_line_split
pool.rs
/// /// <div> /// <img /// src="http://www-id.imag.fr/Laboratoire/Membres/Wagner_Frederic/images/downgraded_manual_max.svg"/> /// </div> /// /// Using it we obtain the graph below. /// On the real file you can hover but javascript and toggle the display of the different tags but /// it is disabled with rustdoc so I downgraded the file /// for this display. pub fn subgraph<OP, R>(work_type: &'static str, work_amount: usize, op: OP) -> R where OP: FnOnce() -> R, { custom_subgraph(work_type, || (), |_| work_amount, op) } /// Same as the subgraph function, but we can log a hardware event /// /// (from: https://github.com/gz/rust-perfcnt) /// /// Events: /// /// * ```HardwareEventType::CPUCycles``` /// /// * ```HardwareEventType::Instructions``` /// /// * ```HardwareEventType::CacheReferences``` /// /// * ```HardwareEventType::CacheMisses``` /// /// * ```HardwareEventType::BranchInstructions``` /// /// * ```HardwareEventType::BranchMisses``` /// /// * ```HardwareEventType::BusCycles``` /// /// * ```HardwareEventType::StalledCyclesFrontend``` /// /// * ```HardwareEventType::StalledCyclesBackend``` /// /// * ```HardwareEventType::RefCPUCycles``` /// /// You will have to import the events from rayon_logs /// and to use the nightly version of the compiler. /// note that It is **freaking slow**: 1 full second to set up the counter. #[cfg(feature = "perf")] pub fn subgraph_hardware_event<OP, R>(tag: &'static str, event: HardwareEventType, op: OP) -> R where OP: FnOnce() -> R, { custom_subgraph( tag, || { let pc: PerfCounter = PerfCounterBuilderLinux::from_hardware_event(event) .exclude_idle() .exclude_kernel() .finish() .expect("Could not create counter"); pc.start().expect("Can not start the counter"); pc }, |mut pc| { pc.stop().expect("Can not stop the counter"); let counted_value = pc.read().unwrap() as usize; pc.reset().expect("Can not reset the counter"); counted_value }, op, ) } /// Same as the subgraph function, but we can log a software event /// /// (from: https://github.com/gz/rust-perfcnt) /// /// Events: /// /// * ```SoftwareEventType::CpuClock``` /// /// * ```SoftwareEventType::TaskClock``` /// /// * ```SoftwareEventType::PageFaults``` /// /// * ```SoftwareEventType::CacheMisses``` /// /// * ```SoftwareEventType::ContextSwitches``` /// /// * ```SoftwareEventType::CpuMigrations``` /// /// * ```SoftwareEventType::PageFaultsMin``` /// /// * ```SoftwareEventType::PageFaultsMin``` /// /// * ```SoftwareEventType::PageFaultsMaj``` /// /// * ```SoftwareEventType::AlignmentFaults``` /// /// * ```SoftwareEventType::EmulationFaults``` /// /// You will have to import the events from rayon_logs /// and to use the nightly version of the compiler #[cfg(feature = "perf")] pub fn subgraph_software_event<OP, R>(tag: &'static str, event: SoftwareEventType, op: OP) -> R where OP: FnOnce() -> R, { //TODO: avoid code duplication by abstracting over events custom_subgraph( tag, || { let pc: PerfCounter = PerfCounterBuilderLinux::from_software_event(event) .exclude_idle() .exclude_kernel() .finish() .expect("Could not create counter"); pc.start().expect("Can not start the counter"); pc }, |mut pc| { pc.stop().expect("Can not stop the counter"); let counted_value = pc.read().unwrap() as usize; pc.reset().expect("Can not reset the counter"); counted_value }, op, ) } /// Same as the subgraph function, but we can log a cache event /// /// (from: https://github.com/gz/rust-perfcnt) /// /// CacheId: /// /// * ```CacheId::L1D``` /// /// * ```CacheId::L1I``` /// /// * ```CacheId::LL``` /// /// * ```CacheId::DTLB``` /// /// * ```CacheId::ITLB``` /// /// * ```CacheId::BPU``` /// /// * ```CacheId::Node``` /// /// CacheOpId: /// /// * ```CacheOpId::Read``` /// /// * ```CacheOpId::Write``` /// /// * ```CacheOpId::Prefetch``` /// /// CacheOpResultId: /// /// * ```CacheOpResultId::Access``` /// /// * ```CacheOpResultId::Miss``` /// /// /// You will have to import the events from rayon_logs /// and to use the nightly version of the compiler /// #[cfg(feature = "perf")] pub fn subgraph_cache_event<OP, R>( tag: &'static str, cache_id: CacheId, cache_op_id: CacheOpId, cache_op_result_id: CacheOpResultId, op: OP, ) -> R where OP: FnOnce() -> R, { //TODO: avoid code duplication by abstracting over events custom_subgraph( tag, || { let pc: PerfCounter = PerfCounterBuilderLinux::from_cache_event( cache_id, cache_op_id, cache_op_result_id, ) .exclude_idle() .exclude_kernel() .finish() .expect("Could not create counter"); pc.start().expect("Can not start the counter"); pc }, |mut pc| { pc.stop().expect("Can not stop the counter"); let counted_value = pc.read().unwrap() as usize; pc.reset().expect("Can not reset the counter"); counted_value }, op, ) } /// Tag a subgraph with a custom value. /// The start function will be called just before running the graph and produce an S. /// The end function will be called just after running the graph on this S and produce a usize /// which will the be stored for display. pub fn custom_subgraph<OP, R, START, END, S>(tag: &'static str, start: START, end: END, op: OP) -> R where OP: FnOnce() -> R, START: FnOnce() -> S, END: FnOnce(S) -> usize, { let s = start(); start_subgraph(tag); let r = op(); let measured_value = end(s); end_subgraph(tag, measured_value); r } /// Stop current task (virtually) and start a subgraph. /// You most likely don't need to call this function directly but `subgraph` instead. pub fn start_subgraph(tag: &'static str) { let subgraph_start_task_id = next_task_id(); logs!( // log child's work and dependencies. RayonEvent::Child(subgraph_start_task_id), // end current task RayonEvent::TaskEnd(now()), // execute full sequential task RayonEvent::TaskStart(subgraph_start_task_id, now()), RayonEvent::SubgraphStart(tag) ); } /// Stop current task (virtually) and end a subgraph. /// You most likely don't need to call this function directly but `subgraph` instead. pub fn end_subgraph(tag: &'static str, measured_value: usize)
/// Identical to `join`, except that the closures have a parameter /// that provides context for the way the closure has been called, /// especially indicating whether they're executing on a different /// thread than where `join_context` was called. This will occur if /// the second job is stolen by a different thread, or if /// `join_context` was called from outside the thread pool to begin /// with. pub fn join_context<A, B, RA, RB>(oper_a: A, oper_b: B) -> (RA, RB) where A: FnOnce(FnContext) -> RA + Send, B: FnOnce(FnContext) -> RB + Send, RA: Send, RB: Send, { let id_c = next_task_id(); let id_a = next_task_id(); let ca = |c| { log(RayonEvent::TaskStart(id_a, now())); let result = oper_a(c); logs!(RayonEvent::Child(id_c), RayonEvent::TaskEnd(now())); result }; let id_b = next_task_id(); let cb = |c| { log(RayonEvent::TaskStart(id_b, now())); let result = oper_b(c); logs!(RayonEvent::Child(id_c), RayonEvent::Task
{ let continuation_task_id = next_task_id(); logs!( RayonEvent::SubgraphEnd(tag, measured_value), RayonEvent::Child(continuation_task_id), RayonEvent::TaskEnd(now()), // start continuation task RayonEvent::TaskStart(continuation_task_id, now()) ); }
identifier_body
pool.rs
/// The end function will be called just after running the graph on this S and produce a usize /// which will the be stored for display. pub fn custom_subgraph<OP, R, START, END, S>(tag: &'static str, start: START, end: END, op: OP) -> R where OP: FnOnce() -> R, START: FnOnce() -> S, END: FnOnce(S) -> usize, { let s = start(); start_subgraph(tag); let r = op(); let measured_value = end(s); end_subgraph(tag, measured_value); r } /// Stop current task (virtually) and start a subgraph. /// You most likely don't need to call this function directly but `subgraph` instead. pub fn start_subgraph(tag: &'static str) { let subgraph_start_task_id = next_task_id(); logs!( // log child's work and dependencies. RayonEvent::Child(subgraph_start_task_id), // end current task RayonEvent::TaskEnd(now()), // execute full sequential task RayonEvent::TaskStart(subgraph_start_task_id, now()), RayonEvent::SubgraphStart(tag) ); } /// Stop current task (virtually) and end a subgraph. /// You most likely don't need to call this function directly but `subgraph` instead. pub fn end_subgraph(tag: &'static str, measured_value: usize) { let continuation_task_id = next_task_id(); logs!( RayonEvent::SubgraphEnd(tag, measured_value), RayonEvent::Child(continuation_task_id), RayonEvent::TaskEnd(now()), // start continuation task RayonEvent::TaskStart(continuation_task_id, now()) ); } /// Identical to `join`, except that the closures have a parameter /// that provides context for the way the closure has been called, /// especially indicating whether they're executing on a different /// thread than where `join_context` was called. This will occur if /// the second job is stolen by a different thread, or if /// `join_context` was called from outside the thread pool to begin /// with. pub fn join_context<A, B, RA, RB>(oper_a: A, oper_b: B) -> (RA, RB) where A: FnOnce(FnContext) -> RA + Send, B: FnOnce(FnContext) -> RB + Send, RA: Send, RB: Send, { let id_c = next_task_id(); let id_a = next_task_id(); let ca = |c| { log(RayonEvent::TaskStart(id_a, now())); let result = oper_a(c); logs!(RayonEvent::Child(id_c), RayonEvent::TaskEnd(now())); result }; let id_b = next_task_id(); let cb = |c| { log(RayonEvent::TaskStart(id_b, now())); let result = oper_b(c); logs!(RayonEvent::Child(id_c), RayonEvent::TaskEnd(now())); result }; logs!( RayonEvent::Child(id_a), RayonEvent::Child(id_b), RayonEvent::TaskEnd(now()) ); let r = rayon::join_context(ca, cb); log(RayonEvent::TaskStart(id_c, now())); r } /// Takes two closures and *potentially* runs them in parallel. It /// returns a pair of the results from those closures. /// /// Conceptually, calling `join()` is similar to spawning two threads, /// one executing each of the two closures. However, the /// implementation is quite different and incurs very low /// overhead. The underlying technique is called "work stealing": the /// Rayon runtime uses a fixed pool of worker threads and attempts to /// only execute code in parallel when there are idle CPUs to handle /// it. /// /// When `join` is called from outside the thread pool, the calling /// thread will block while the closures execute in the pool. When /// `join` is called within the pool, the calling thread still actively /// participates in the thread pool. It will begin by executing closure /// A (on the current thread). While it is doing that, it will advertise /// closure B as being available for other threads to execute. Once closure A /// has completed, the current thread will try to execute closure B; /// if however closure B has been stolen, then it will look for other work /// while waiting for the thief to fully execute closure B. (This is the /// typical work-stealing strategy). /// /// # Examples /// /// This example uses join to perform a quick-sort (note this is not a /// particularly optimized implementation: if you **actually** want to /// sort for real, you should prefer [the `par_sort` method] offered /// by Rayon). /// /// [the `par_sort` method]: ../rayon/slice/trait.ParallelSliceMut.html#method.par_sort /// /// ```rust /// let mut v = vec![5, 1, 8, 22, 0, 44]; /// quick_sort(&mut v); /// assert_eq!(v, vec![0, 1, 5, 8, 22, 44]); /// /// fn quick_sort<T:PartialOrd+Send>(v: &mut [T]) { /// if v.len() > 1 { /// let mid = partition(v); /// let (lo, hi) = v.split_at_mut(mid); /// rayon::join(|| quick_sort(lo), /// || quick_sort(hi)); /// } /// } /// /// // Partition rearranges all items `<=` to the pivot /// // item (arbitrary selected to be the last item in the slice) /// // to the first half of the slice. It then returns the /// // "dividing point" where the pivot is placed. /// fn partition<T:PartialOrd+Send>(v: &mut [T]) -> usize { /// let pivot = v.len() - 1; /// let mut i = 0; /// for j in 0..pivot { /// if v[j] <= v[pivot] { /// v.swap(i, j); /// i += 1; /// } /// } /// v.swap(i, pivot); /// i /// } /// ``` /// /// # Warning about blocking I/O /// /// The assumption is that the closures given to `join()` are /// CPU-bound tasks that do not perform I/O or other blocking /// operations. If you do perform I/O, and that I/O should block /// (e.g., waiting for a network request), the overall performance may /// be poor. Moreover, if you cause one closure to be blocked waiting /// on another (for example, using a channel), that could lead to a /// deadlock. /// /// # Panics /// /// No matter what happens, both closures will always be executed. If /// a single closure panics, whether it be the first or second /// closure, that panic will be propagated and hence `join()` will /// panic with the same panic value. If both closures panic, `join()` /// will panic with the panic value from the first closure. pub fn join<A, B, RA, RB>(oper_a: A, oper_b: B) -> (RA, RB) where A: FnOnce() -> RA + Send, B: FnOnce() -> RB + Send, RA: Send, RB: Send, { let id_c = next_task_id(); let id_a = next_task_id(); let ca = || { log(RayonEvent::TaskStart(id_a, now())); let result = oper_a(); logs!(RayonEvent::Child(id_c), RayonEvent::TaskEnd(now())); result }; let id_b = next_task_id(); let cb = || { log(RayonEvent::TaskStart(id_b, now())); let result = oper_b(); logs!(RayonEvent::Child(id_c), RayonEvent::TaskEnd(now())); result }; logs!( RayonEvent::Child(id_a), RayonEvent::Child(id_b), RayonEvent::TaskEnd(now()) ); let r = rayon::join(ca, cb); log(RayonEvent::TaskStart(id_c, now())); r } // small global counter to increment file names static INSTALL_COUNT: AtomicUsize = AtomicUsize::new(0); /// We wrap rayon's pool into our own struct to overload the install method. pub struct ThreadPool { pub(crate) logs: Arc<Mutex<Vec<Arc<Storage<RayonEvent>>>>>, pub(crate) pool: rayon::ThreadPool, } impl ThreadPool { /// Reset all logs and counters to initial condition. fn reset(&self) { NEXT_TASK_ID.store(0, Ordering::SeqCst); NEXT_ITERATOR_ID.store(0, Ordering::SeqCst); let logs = &*self.logs.lock().unwrap(); // oh yeah baby for log in logs { log.clear(); } } /// Execute given closure in the thread pool, logging it's task as the initial one. /// After running, we post-process the logs and return a `RunLog` together with the closure's /// result. pub fn
logging_install
identifier_name
slide-load.js
farCoords: {}, closeCoords: {}, transiton: {}, defaults: { transition: "bottom", ajax: { url: "/", dataType: 'html', type: 'POST', data: {}, }, }, /** * Initializes the creation of the loader and sets its options. * @param {Object} config A configuration object for the loader. */ _init: function( config, el ) { config = config || {}; $("#pl-loader").remove(); this.$el = this._proxy_check(el); this.settings = $.extend(true, {}, this.defaults, config); this.$loader = $('<div id="pl-loader" class="loading-screen" />'); this._load(); }, /** * Clones the invoking element and positions the clone directly on top of it. */ _create_loader: function() { var $content = this.$el.clone(); this.$loader.html($content); this._set_loader_style(); $(document).trigger("pl.clone", [this.$el]); this.$loader.appendTo('.faux-body'); }, /** * Gets two sets of coordinates for use in animating the loader. */ _set_coords: function() { this.offsets = this.$el.offset(); this.farCoords = this._get_far_coords(this.$el); this.closeCoords = this._get_close_coords(); }, /** * Gets the close coordinates of the element. * * The close coordinates are the distances from each side of the document * to that same side of the element. So distance from the left side * of the document to the left side of the element. * * @param {Object} coords An object containing top and left coordinates. * @return {Object} An object containing all four cooridantes. */ _get_close_coords: function() { var coords = {}; coords.bottom = $(".faux-body").outerHeight() - this.farCoords.bottom; coords.right = $(".faux-body").outerWidth() - this.farCoords.right; coords.top = this.offsets.top;
/** * Gets the far coordinates of the element. * * The far coordinates are the distances from each side of the document * to the opposite side of the element. So distance from the left side * of the document to the right side of the element. * * @param {jQuery} coords The jQuery object to get teh coordinates of. * @return {Object} An object containing all four cooridantes. */ _get_far_coords: function($cntr) { coords = {}; coords.top = $(".faux-body").outerHeight() - this.offsets.top; coords.left = $(".faux-body").outerWidth() - this.offsets.left; coords.bottom = $cntr.outerHeight() + this.offsets.top; coords.right = $cntr.outerWidth() + this.offsets.left; return coords; }, /** * This sets additional css properties of the loader clone. */ _set_loader_style: function() { this.$loader.css({ height: this.$el.height(), width: this.$el.width(), }); }, /** * This runs the actual AJAX call to get new content, then calls the callback * function. */ _load: function() { var self = this; $.ajax(this.settings.ajax).done(function(data) { self._show_content.call(self, data); }); }, /** * This runs all of the necessary functions to add the new content to the * container element, and then reveal it by animating the loader clone. * * @param {String} html The new HTML to insert into the container element. */ _show_content: function(html) { var self = this; this._set_coords(); this._create_loader(); this._prepare_cntr(html); this.transitions[ this.settings.transition ].call(this); this.$el.imagesLoaded( function() { self.$el.trigger("pl.animate", [self.$el]); var fade = self.$el.fadeTo({ duration: 750, queue: false }, 1).animateAuto("height", { duration: 500, queue: false }) var slide = self.$loader.animate( self.transiton, { queue: false, duration: 500, } ); $.when(fade, slide).done(function () { self.$el.removeAttr('style'); self._exit(); }); }); }, /** * Inserts the new content into the container element. * * This sets a few style properties to keep things from showing up before * they're ready. It sets the height to match the height of the previous content * so that the page doesn't expand in length before we're ready for it. It also * sets the opacity to 0, which is there primarily to aid the eventual transition, * but has the added bonus of providing a fallback to make sure no new content is * shown behind the loading screen. In some instances, the new content is meant * to replace the container, instead of being placed inside of it. (Think team * member to team member navigation.) This function will determine if that * replacement needs to happen. * * @param {String} html The HTML content to load into the element. */ _prepare_cntr: function(html) { var $new = $(html) , replace = this.$el.data("load-replace") $new = replace ? this._find_replacemnt($new) : $new; this.$el.css({ height: this.$el.height(), opacity: 0, }); this.$el.html($new); }, /** * Checks for a load proxy on the container element. * * A load proxy allows the content to be loaded into an element that is not * the container element. If the container element has a data-load="" property * set, then this function attempts to find the new container element. The * data-load property can be any valid jQuery selector. If there is a data-load * property set, but the jQuery selector returns an empty result, the initial * element is returned. * * @param {jQuery} $cntr A jQuery object containing the container element. * * @return {jQuery} A jQuery object containing the new element if found, * otherwise the initial object is returned. */ _proxy_check: function($cntr) { var proxy = $cntr.data("load-proxy"); return $(proxy).length ? $(proxy) : $cntr; }, /** * This function looks for a replacement container in the new content. * * In order to make sure that the new content is placed correctly, the replacement * is dependent on finding an element at the root of the new HTML that has the * data-load-replacement property. If an element with that property is found, * then the replacement is called, which copies attributes/properties of the replacement * onto the existing container. Then those properties.attributes are strripped from * the replacement container. This is done because if we were to wholesale replace * the existing container, all the events we have bound to it, and triggers in this * plugin would cease to exist, and therefore would fail. * * @param {jQuery} $new A jQuery object containing the new HTML object. * * @return {jQuery} A jQuery object containing the new HTML with the replacement made. */ _find_replacemnt: function($new) { var self = this , $copy = $new $copy.each(function(i, val) { if ($(this).data("load-replacement")) { self._replace_container($(this)); contents = $(this).contents().toArray(); $new.splice.apply($new, [i, 1].concat(contents)) } }); return $new; }, /** * This function copies attributes from one jQuery element to the container element. * * This function iterates over the properties of the jQuery element provided, and then * copies them onto the Container jQuery element. * * @param {jQuery} $new A jQuery element to copy the properties of. */ _replace_container: function($new) { var attr = $new.prop("attributes") self = this $.each(attr, function() { self.$el.attr(this.name, this.value); }); },
coords.left = this.offsets.left; // $.extend(coords, this.offsets); return coords; },
random_line_split
slide-load.js
four cooridantes. */ _get_close_coords: function() { var coords = {}; coords.bottom = $(".faux-body").outerHeight() - this.farCoords.bottom; coords.right = $(".faux-body").outerWidth() - this.farCoords.right; coords.top = this.offsets.top; coords.left = this.offsets.left; // $.extend(coords, this.offsets); return coords; }, /** * Gets the far coordinates of the element. * * The far coordinates are the distances from each side of the document * to the opposite side of the element. So distance from the left side * of the document to the right side of the element. * * @param {jQuery} coords The jQuery object to get teh coordinates of. * @return {Object} An object containing all four cooridantes. */ _get_far_coords: function($cntr) { coords = {}; coords.top = $(".faux-body").outerHeight() - this.offsets.top; coords.left = $(".faux-body").outerWidth() - this.offsets.left; coords.bottom = $cntr.outerHeight() + this.offsets.top; coords.right = $cntr.outerWidth() + this.offsets.left; return coords; }, /** * This sets additional css properties of the loader clone. */ _set_loader_style: function() { this.$loader.css({ height: this.$el.height(), width: this.$el.width(), }); }, /** * This runs the actual AJAX call to get new content, then calls the callback * function. */ _load: function() { var self = this; $.ajax(this.settings.ajax).done(function(data) { self._show_content.call(self, data); }); }, /** * This runs all of the necessary functions to add the new content to the * container element, and then reveal it by animating the loader clone. * * @param {String} html The new HTML to insert into the container element. */ _show_content: function(html) { var self = this; this._set_coords(); this._create_loader(); this._prepare_cntr(html); this.transitions[ this.settings.transition ].call(this); this.$el.imagesLoaded( function() { self.$el.trigger("pl.animate", [self.$el]); var fade = self.$el.fadeTo({ duration: 750, queue: false }, 1).animateAuto("height", { duration: 500, queue: false }) var slide = self.$loader.animate( self.transiton, { queue: false, duration: 500, } ); $.when(fade, slide).done(function () { self.$el.removeAttr('style'); self._exit(); }); }); }, /** * Inserts the new content into the container element. * * This sets a few style properties to keep things from showing up before * they're ready. It sets the height to match the height of the previous content * so that the page doesn't expand in length before we're ready for it. It also * sets the opacity to 0, which is there primarily to aid the eventual transition, * but has the added bonus of providing a fallback to make sure no new content is * shown behind the loading screen. In some instances, the new content is meant * to replace the container, instead of being placed inside of it. (Think team * member to team member navigation.) This function will determine if that * replacement needs to happen. * * @param {String} html The HTML content to load into the element. */ _prepare_cntr: function(html) { var $new = $(html) , replace = this.$el.data("load-replace") $new = replace ? this._find_replacemnt($new) : $new; this.$el.css({ height: this.$el.height(), opacity: 0, }); this.$el.html($new); }, /** * Checks for a load proxy on the container element. * * A load proxy allows the content to be loaded into an element that is not * the container element. If the container element has a data-load="" property * set, then this function attempts to find the new container element. The * data-load property can be any valid jQuery selector. If there is a data-load * property set, but the jQuery selector returns an empty result, the initial * element is returned. * * @param {jQuery} $cntr A jQuery object containing the container element. * * @return {jQuery} A jQuery object containing the new element if found, * otherwise the initial object is returned. */ _proxy_check: function($cntr) { var proxy = $cntr.data("load-proxy"); return $(proxy).length ? $(proxy) : $cntr; }, /** * This function looks for a replacement container in the new content. * * In order to make sure that the new content is placed correctly, the replacement * is dependent on finding an element at the root of the new HTML that has the * data-load-replacement property. If an element with that property is found, * then the replacement is called, which copies attributes/properties of the replacement * onto the existing container. Then those properties.attributes are strripped from * the replacement container. This is done because if we were to wholesale replace * the existing container, all the events we have bound to it, and triggers in this * plugin would cease to exist, and therefore would fail. * * @param {jQuery} $new A jQuery object containing the new HTML object. * * @return {jQuery} A jQuery object containing the new HTML with the replacement made. */ _find_replacemnt: function($new) { var self = this , $copy = $new $copy.each(function(i, val) { if ($(this).data("load-replacement")) { self._replace_container($(this)); contents = $(this).contents().toArray(); $new.splice.apply($new, [i, 1].concat(contents)) } }); return $new; }, /** * This function copies attributes from one jQuery element to the container element. * * This function iterates over the properties of the jQuery element provided, and then * copies them onto the Container jQuery element. * * @param {jQuery} $new A jQuery element to copy the properties of. */ _replace_container: function($new) { var attr = $new.prop("attributes") self = this $.each(attr, function() { self.$el.attr(this.name, this.value); }); }, /** * Finishes execution of the loader. * * Removes the loading screen element from the page and then triggers the * completion event on the document. */ _exit: function() { this.$loader.remove(); $(document).trigger("pl.complete", [this.$el]); }, transitions: { top: function() { this.$loader.css({ top: this.closeCoords.top, }); this.transiton = { height: 0, } }, bottom: function() { this.$loader.css({ top: this.closeCoords.top, bottom: this.closeCoords.bottom, }); this.transiton = { top: this.farCoords.bottom, height: 0, } }, left: function() { this.$loader.css({ top: this.closeCoords.top, left: this.closeCoords.left, }); this.transiton = { left: "-100%", } }, right: function() { this.$loader.css({ top: this.closeCoords.top, right: this.closeCoords.right, }); this.transiton = { right: "-100%", } }, }, }; $.fn.pushLoad = function() { // The second string is a pushLoad transition method if ( pushLoad.transitions[ arguments[1] ] ) { config = { transition: arguments[1], ajax: { url: arguments[0], }, } pushLoad._init.call( pushLoad, config, this ); } else if ( typeof arguments[0] === 'object' ) { pushLoad._init.call( pushLoad, arguments[0], this ); } else
{ $.error( 'pushLoad has no method '+arguments[0] ); }
conditional_block
main.rs
older // variable is still usable after assignment. Rust won’t let us annotate a type with the Copy trait if the // type, or any of its parts, has implemented the Drop trait. If the type needs something special to happen // when the value goes out of scope and we add the Copy annotation to that type, we’ll get a compile time error. // To learn about how to add the Copy annotation to your type, see Appendix C on Derivable Traits. // So what types are Copy? You can check the documentation for the given type to be sure, but as a general rule, // any group of simple scalar values can be Copy, and nothing that requires allocation or is some form of resource // is Copy. Here are some of the types that are Copy: // All the integer types, like u32. // The boolean type, bool, with values true and false. // All the floating point types, like f64. // Tuples, but only if they contain types that are also Copy. (i32, i32) is Copy, but (i32, String) is not. let s = String::from("hello"); // s comes into scope. //So in rust if we pass a variable into a function it loses it's ownership to the //function. Then once the function is over that variable no longer exists //because it is now out of scope. takes_ownership(s); // s's value moves into the function... // ... and so is no longer valid here. let x = 5; // x comes into scope. //If a variable has the copy trait then only a copy is made to the function and //we can still use the variable afterwards even though all the variables in the //function are now out of scope. makes_copy(x); // x would move into the function, // but i32 is Copy, so it’s okay to still // use x afterward. //we can give ownership of a variable from a function by having an expression at the end. //We could pass in a variable and then take back its ownership by doing this. However, I think this //is kinda of a pain. The people at Rust feel the same. let s1 = gives_ownership(); //Rust also let's return variables as tuples so which we can then can deconstruct this when //we get the returned values. //Now it's time to go over references and borrowing! let s1 = String::from("hello"); //The & creates a reference to a variable. They can be thought of a pointer to the original data. //By doing this we do not pass ownership of the variable to the function //Therefore when we go out of scope of the function we still have ownership of the variable //where the function call was made. //References as function parameters is called borrowing. let len = calculate_length(&s1); println!("The length of '{}' is {}.", s1, len); //We can not modify a borrowed variable. //change(&s1); let mut s1 = String::from("hello"); //We can fix this by making a mutable reference //We also need to make sure that our variable we're passing in is also mutable. change(&mut s1); println!("{}", s1); //You are only allowed one mutable reference to a particular piece of data in a particular scope. //This insures that we don't have any aliasing with our references refering to the same data. //The benefit of having this restriction is that Rust can prevent data races at compile time. //From the rust book //Whew! We also cannot have a mutable reference while we have an immutable one. //Users of an immutable reference don’t expect the values to suddenly change out from under them! //However, multiple immutable references are okay because no one who is just reading the data has //the ability to affect anyone else’s reading of the data. //let mut s = String::from("Hello"); //let r1 = &s; //Immutable reference //let r2 = &s; //Immutable reference //let r3 = &s; //Mutable reference -- big no no //The compiler does not dangling pointers/references. It therefore will error out on us. // let refernece_to_nothing = dangle(); //We are now going to go over slices. //From the rust book: Another data type that does not have ownership is the slice. //Slices let you reference a contiguous sequence of elements in a collection rather than the whole collection. // let mut s = String::from("hello world"); // let word = first_word(&s); // word will get the value 5. // s.clear(); // This empties the String, making it equal to "". // word still has the value 5 here, but there's no more string that // we could meaningfully use the value 5 with. word is now totally invalid! //The index we got is now completely out of sync with our original string. //If we end up having more indices we could get even more out of sync with our data. //For strings we can take advantage of a built in feature called string slices. //They create a reference to portions of a string. let s = String::from("hello world"); //Slicing is similar to slicing in python where you have a starting index and then //the ending value is +1 of the data you actually care about. let hello = &s[0..5]; // let hello = &s[..5]; //Equivalent to the above let world = &s[6..11]; // let world = &s[6..]; //Equivalent to the above let len = s.len(); let slice = &s[0..len]; // let slice = &s[..]; //Equivalent to the above // We now have a straightforward API that’s much harder to mess up, since the compiler will //ensure the references into the String remain valid. Remember the bug in the program in Listing 4-11, //when we got the index to the end of the first word but then cleared the string so our index was invalid? //That code was logically incorrect but didn’t show any immediate errors. The problems would show up later //if we kept trying to use the first word index with an emptied string. Slices make this bug impossible //and let us know we have a problem with our code much sooner. Using the slice version of first_word //will throw a compile time error: // let mut s = String::from("hello world"); // let word = first_word(&s); // s.clear(); // Error! // Recall from the borrowing rules that if we have an immutable reference to something, we cannot also // take a mutable reference. Because clear needs to truncate the String, it tries to take a mutable reference, // which fails. Not only has Rust made our API easier to use, but it has also eliminated an entire class of errors // at compile time! let s = "Hello, world!"; // The type of s here is &str: it’s a slice pointing to that specific point of the binary. This is also why string // literals are immutable; &str is an immutable reference. let my_string = String::from("hello world"); // first_word works on slices of `String`s let word = first_word(&my_string[..]); let my_string_literal = "hello world"; // first_word works on slices of string literals let word = first_word(&my_string_literal[..]); // since string literals *are* string slices already, // this works too, without the slice syntax! let word = first_word(my_string_literal); let a = [1, 2, 3, 4, 5]; let slice = &a[1..3]; // This slice has the type &[i32]. It works the same way as string slices do, by storing a reference to the // first element and a length. You’ll use this kind of slice for all sorts of other collections. We’ll discuss // these collections in detail when we talk about vectors in Chapter 8 } // Here, x goes out of scope, then s. But since s's value was moved, nothing // special happens. fn takes_ownership(some_string: String) { // some_string comes into scope. println!("{}", some_string); } // Here, some_string goes out of scope and `drop` is called. The backing // memory is freed. fn makes_copy(some_integer: i32) { // some_integer comes into scope. println!("{}", some_integer); } // Here, some_integer goes out of scope. Nothing special happens. //Tell what type the function will return fn gives_ownership() -> String { // gives_ownership will move its // return value into the function // that calls it. let some_string = String::from("hello"); // some_string comes into scope. some_string // some_string is returned and // moves out to the calling // function. } fn calculate_length(s: &String) -> usize { s.len() } //This fun
ction will erro
identifier_body
main.rs
this //is kinda of a pain. The people at Rust feel the same. let s1 = gives_ownership(); //Rust also let's return variables as tuples so which we can then can deconstruct this when //we get the returned values. //Now it's time to go over references and borrowing! let s1 = String::from("hello"); //The & creates a reference to a variable. They can be thought of a pointer to the original data. //By doing this we do not pass ownership of the variable to the function //Therefore when we go out of scope of the function we still have ownership of the variable //where the function call was made. //References as function parameters is called borrowing. let len = calculate_length(&s1); println!("The length of '{}' is {}.", s1, len); //We can not modify a borrowed variable. //change(&s1); let mut s1 = String::from("hello"); //We can fix this by making a mutable reference //We also need to make sure that our variable we're passing in is also mutable. change(&mut s1); println!("{}", s1); //You are only allowed one mutable reference to a particular piece of data in a particular scope. //This insures that we don't have any aliasing with our references refering to the same data. //The benefit of having this restriction is that Rust can prevent data races at compile time. //From the rust book //Whew! We also cannot have a mutable reference while we have an immutable one. //Users of an immutable reference don’t expect the values to suddenly change out from under them! //However, multiple immutable references are okay because no one who is just reading the data has //the ability to affect anyone else’s reading of the data. //let mut s = String::from("Hello"); //let r1 = &s; //Immutable reference //let r2 = &s; //Immutable reference //let r3 = &s; //Mutable reference -- big no no //The compiler does not dangling pointers/references. It therefore will error out on us. // let refernece_to_nothing = dangle(); //We are now going to go over slices. //From the rust book: Another data type that does not have ownership is the slice. //Slices let you reference a contiguous sequence of elements in a collection rather than the whole collection. // let mut s = String::from("hello world"); // let word = first_word(&s); // word will get the value 5. // s.clear(); // This empties the String, making it equal to "". // word still has the value 5 here, but there's no more string that // we could meaningfully use the value 5 with. word is now totally invalid! //The index we got is now completely out of sync with our original string. //If we end up having more indices we could get even more out of sync with our data. //For strings we can take advantage of a built in feature called string slices. //They create a reference to portions of a string. let s = String::from("hello world"); //Slicing is similar to slicing in python where you have a starting index and then //the ending value is +1 of the data you actually care about. let hello = &s[0..5]; // let hello = &s[..5]; //Equivalent to the above let world = &s[6..11]; // let world = &s[6..]; //Equivalent to the above let len = s.len(); let slice = &s[0..len]; // let slice = &s[..]; //Equivalent to the above // We now have a straightforward API that’s much harder to mess up, since the compiler will //ensure the references into the String remain valid. Remember the bug in the program in Listing 4-11, //when we got the index to the end of the first word but then cleared the string so our index was invalid? //That code was logically incorrect but didn’t show any immediate errors. The problems would show up later //if we kept trying to use the first word index with an emptied string. Slices make this bug impossible //and let us know we have a problem with our code much sooner. Using the slice version of first_word //will throw a compile time error: // let mut s = String::from("hello world"); // let word = first_word(&s); // s.clear(); // Error! // Recall from the borrowing rules that if we have an immutable reference to something, we cannot also // take a mutable reference. Because clear needs to truncate the String, it tries to take a mutable reference, // which fails. Not only has Rust made our API easier to use, but it has also eliminated an entire class of errors // at compile time! let s = "Hello, world!"; // The type of s here is &str: it’s a slice pointing to that specific point of the binary. This is also why string // literals are immutable; &str is an immutable reference. let my_string = String::from("hello world"); // first_word works on slices of `String`s let word = first_word(&my_string[..]); let my_string_literal = "hello world"; // first_word works on slices of string literals let word = first_word(&my_string_literal[..]); // since string literals *are* string slices already, // this works too, without the slice syntax! let word = first_word(my_string_literal); let a = [1, 2, 3, 4, 5]; let slice = &a[1..3]; // This slice has the type &[i32]. It works the same way as string slices do, by storing a reference to the // first element and a length. You’ll use this kind of slice for all sorts of other collections. We’ll discuss // these collections in detail when we talk about vectors in Chapter 8 } // Here, x goes out of scope, then s. But since s's value was moved, nothing // special happens. fn takes_ownership(some_string: String) { // some_string comes into scope. println!("{}", some_string); } // Here, some_string goes out of scope and `drop` is called. The backing // memory is freed. fn makes_copy(some_integer: i32) { // some_integer comes into scope. println!("{}", some_integer); } // Here, some_integer goes out of scope. Nothing special happens. //Tell what type the function will return fn gives_ownership() -> String { // gives_ownership will move its // return value into the function // that calls it. let some_string = String::from("hello"); // some_string comes into scope. some_string // some_string is returned and // moves out to the calling // function. } fn calculate_length(s: &String) -> usize { s.len() } //This function will error on us since we are trying to //modify a borrowed variable. We will always get an //error for this function even if we never call it. // fn change(some_string: &String) { // some_string.push_str(", world"); // } //This fixes the above code by making a mutable reference that we can now modify. fn change(some_string: &mut String) { some_string.push_str(", world"); } //The below code creates a dangling pointer/reference. //So when the data goes out of scope at the end of the function //our reference now points to memory that has been freed. //The compiler catches this and errors out on us. // fn dangle() -> &String { // let s = String::from("hello"); // &s // } //This version doesn't create slices of the data so things become out of index with each other //We are going to rewrite it with a new version // fn first_word(s: &String) -> usize { // //We are converting our string into a byte // let bytes = s.as_bytes(); // //We now iterate through the string using iter. // //the enumerate function packages up each part of the // //iterator as a tuple with an index and a reference to the value // for (i, &item) in bytes.iter().enumerate() { // //We check to see if the byte literal of the space is // //equal to our item. // //If it is then we return that index. // if item == b' ' { // return i; // } // } // //If we don't run across a space at all then we return the length of the string. // s.len() // } //We can change the following to the current function signature // fn first_word(s: &String) -> &str { //The new signature now allows us to operate on both Strings and str types fn first_word(s: &str) -> &str { let bytes = s.as_bytes(); for (i, &item) in bytes.iter().enumerate() { if item == b' ' { return &s[0.
.i]; } } &s[..] }
conditional_block
main.rs
To learn about how to add the Copy annotation to your type, see Appendix C on Derivable Traits. // So what types are Copy? You can check the documentation for the given type to be sure, but as a general rule, // any group of simple scalar values can be Copy, and nothing that requires allocation or is some form of resource // is Copy. Here are some of the types that are Copy: // All the integer types, like u32. // The boolean type, bool, with values true and false. // All the floating point types, like f64. // Tuples, but only if they contain types that are also Copy. (i32, i32) is Copy, but (i32, String) is not. let s = String::from("hello"); // s comes into scope. //So in rust if we pass a variable into a function it loses it's ownership to the //function. Then once the function is over that variable no longer exists //because it is now out of scope. takes_ownership(s); // s's value moves into the function... // ... and so is no longer valid here. let x = 5; // x comes into scope. //If a variable has the copy trait then only a copy is made to the function and //we can still use the variable afterwards even though all the variables in the //function are now out of scope. makes_copy(x); // x would move into the function, // but i32 is Copy, so it’s okay to still // use x afterward. //we can give ownership of a variable from a function by having an expression at the end. //We could pass in a variable and then take back its ownership by doing this. However, I think this //is kinda of a pain. The people at Rust feel the same. let s1 = gives_ownership(); //Rust also let's return variables as tuples so which we can then can deconstruct this when //we get the returned values. //Now it's time to go over references and borrowing! let s1 = String::from("hello"); //The & creates a reference to a variable. They can be thought of a pointer to the original data. //By doing this we do not pass ownership of the variable to the function //Therefore when we go out of scope of the function we still have ownership of the variable //where the function call was made. //References as function parameters is called borrowing. let len = calculate_length(&s1); println!("The length of '{}' is {}.", s1, len); //We can not modify a borrowed variable. //change(&s1); let mut s1 = String::from("hello"); //We can fix this by making a mutable reference //We also need to make sure that our variable we're passing in is also mutable. change(&mut s1); println!("{}", s1); //You are only allowed one mutable reference to a particular piece of data in a particular scope. //This insures that we don't have any aliasing with our references refering to the same data. //The benefit of having this restriction is that Rust can prevent data races at compile time. //From the rust book //Whew! We also cannot have a mutable reference while we have an immutable one. //Users of an immutable reference don’t expect the values to suddenly change out from under them! //However, multiple immutable references are okay because no one who is just reading the data has //the ability to affect anyone else’s reading of the data. //let mut s = String::from("Hello"); //let r1 = &s; //Immutable reference //let r2 = &s; //Immutable reference //let r3 = &s; //Mutable reference -- big no no //The compiler does not dangling pointers/references. It therefore will error out on us. // let refernece_to_nothing = dangle(); //We are now going to go over slices. //From the rust book: Another data type that does not have ownership is the slice. //Slices let you reference a contiguous sequence of elements in a collection rather than the whole collection. // let mut s = String::from("hello world"); // let word = first_word(&s); // word will get the value 5. // s.clear(); // This empties the String, making it equal to "". // word still has the value 5 here, but there's no more string that // we could meaningfully use the value 5 with. word is now totally invalid! //The index we got is now completely out of sync with our original string. //If we end up having more indices we could get even more out of sync with our data. //For strings we can take advantage of a built in feature called string slices. //They create a reference to portions of a string. let s = String::from("hello world"); //Slicing is similar to slicing in python where you have a starting index and then //the ending value is +1 of the data you actually care about. let hello = &s[0..5]; // let hello = &s[..5]; //Equivalent to the above let world = &s[6..11]; // let world = &s[6..]; //Equivalent to the above let len = s.len(); let slice = &s[0..len]; // let slice = &s[..]; //Equivalent to the above // We now have a straightforward API that’s much harder to mess up, since the compiler will //ensure the references into the String remain valid. Remember the bug in the program in Listing 4-11, //when we got the index to the end of the first word but then cleared the string so our index was invalid? //That code was logically incorrect but didn’t show any immediate errors. The problems would show up later //if we kept trying to use the first word index with an emptied string. Slices make this bug impossible //and let us know we have a problem with our code much sooner. Using the slice version of first_word //will throw a compile time error: // let mut s = String::from("hello world"); // let word = first_word(&s); // s.clear(); // Error! // Recall from the borrowing rules that if we have an immutable reference to something, we cannot also // take a mutable reference. Because clear needs to truncate the String, it tries to take a mutable reference, // which fails. Not only has Rust made our API easier to use, but it has also eliminated an entire class of errors // at compile time! let s = "Hello, world!"; // The type of s here is &str: it’s a slice pointing to that specific point of the binary. This is also why string // literals are immutable; &str is an immutable reference. let my_string = String::from("hello world"); // first_word works on slices of `String`s let word = first_word(&my_string[..]); let my_string_literal = "hello world"; // first_word works on slices of string literals let word = first_word(&my_string_literal[..]); // since string literals *are* string slices already, // this works too, without the slice syntax! let word = first_word(my_string_literal); let a = [1, 2, 3, 4, 5]; let slice = &a[1..3]; // This slice has the type &[i32]. It works the same way as string slices do, by storing a reference to the // first element and a length. You’ll use this kind of slice for all sorts of other collections. We’ll discuss // these collections in detail when we talk about vectors in Chapter 8 } // Here, x goes out of scope, then s. But since s's value was moved, nothing // special happens. fn takes_ownership(some_string: String) { // some_string comes into scope. println!("{}", some_string); } // Here, some_string goes out of scope and `drop` is called. The backing // memory is freed. fn makes_copy(some_integer: i32) { // some_integer comes into scope. println!("{}", some_integer); } // Here, some_integer goes out of scope. Nothing special happens. //Tell what type the function will return fn gives_ownership() -> String { // gives_ownership will move its // return value into the function // that calls it. let some_string = String::from("hello"); // some_string comes into scope. some_string // some_string is returned and // moves out to the calling // function. } fn calculate_length(s: &String) -> usize { s.len() } //This function will error on us since we are trying to //modify a borrowed variable. We will always get an //error for this function even if we never call it. // fn change(some_string: &String) { // some_string.push_str(", world"); // } //This fixes the above code by making a mutable reference that we can now modify. fn change(some_string: &mut S
tring)
identifier_name
main.rs
//Note: In C++, this pattern of deallocating resources at the end of an item’s lifetime is sometimes //called Resource Acquisition Is Initialization (RAII). The drop function in Rust will be familiar //to you if you’ve used RAII patterns. let x = 5; let y = x;// y is just a copy of x since they are simple types and have a fixed size let s1 = String::from("hello"); let s2 = s1; // s2 is a copy of the pointer to the data that s1 points to // this errors out because s1 does not have a copy trait which meanse //we made a shallow copy instead of a deep copy. Rust does not like this // if we tried to use s1. If we use s2 we are fine since s1 is invalidated //after we assign s2 to s1 values. This operation is called a move. // println!("{}", s2); let s1 = String::from("hello"); let s2 = s1.clone(); // This creates a deep copy of of s1. We can now use s1 in other places with out // it being invalid // println!("{}",s1); //Info about what things that make a deep copy when you do let x = something; let y = x; // Rust has a special annotation called the Copy trait that we can place on types like integers that are // stored on the stack (we’ll talk more about traits in Chapter 10). If a type has the Copy trait, an older // variable is still usable after assignment. Rust won’t let us annotate a type with the Copy trait if the // type, or any of its parts, has implemented the Drop trait. If the type needs something special to happen // when the value goes out of scope and we add the Copy annotation to that type, we’ll get a compile time error. // To learn about how to add the Copy annotation to your type, see Appendix C on Derivable Traits. // So what types are Copy? You can check the documentation for the given type to be sure, but as a general rule, // any group of simple scalar values can be Copy, and nothing that requires allocation or is some form of resource // is Copy. Here are some of the types that are Copy: // All the integer types, like u32. // The boolean type, bool, with values true and false. // All the floating point types, like f64. // Tuples, but only if they contain types that are also Copy. (i32, i32) is Copy, but (i32, String) is not. let s = String::from("hello"); // s comes into scope. //So in rust if we pass a variable into a function it loses it's ownership to the //function. Then once the function is over that variable no longer exists //because it is now out of scope. takes_ownership(s); // s's value moves into the function... // ... and so is no longer valid here. let x = 5; // x comes into scope. //If a variable has the copy trait then only a copy is made to the function and //we can still use the variable afterwards even though all the variables in the //function are now out of scope. makes_copy(x); // x would move into the function, // but i32 is Copy, so it’s okay to still // use x afterward. //we can give ownership of a variable from a function by having an expression at the end. //We could pass in a variable and then take back its ownership by doing this. However, I think this //is kinda of a pain. The people at Rust feel the same. let s1 = gives_ownership(); //Rust also let's return variables as tuples so which we can then can deconstruct this when //we get the returned values. //Now it's time to go over references and borrowing! let s1 = String::from("hello"); //The & creates a reference to a variable. They can be thought of a pointer to the original data. //By doing this we do not pass ownership of the variable to the function //Therefore when we go out of scope of the function we still have ownership of the variable //where the function call was made. //References as function parameters is called borrowing. let len = calculate_length(&s1); println!("The length of '{}' is {}.", s1, len); //We can not modify a borrowed variable. //change(&s1); let mut s1 = String::from("hello"); //We can fix this by making a mutable reference //We also need to make sure that our variable we're passing in is also mutable. change(&mut s1); println!("{}", s1); //You are only allowed one mutable reference to a particular piece of data in a particular scope. //This insures that we don't have any aliasing with our references refering to the same data. //The benefit of having this restriction is that Rust can prevent data races at compile time. //From the rust book //Whew! We also cannot have a mutable reference while we have an immutable one. //Users of an immutable reference don’t expect the values to suddenly change out from under them! //However, multiple immutable references are okay because no one who is just reading the data has //the ability to affect anyone else’s reading of the data. //let mut s = String::from("Hello"); //let r1 = &s; //Immutable reference //let r2 = &s; //Immutable reference //let r3 = &s; //Mutable reference -- big no no //The compiler does not dangling pointers/references. It therefore will error out on us. // let refernece_to_nothing = dangle(); //We are now going to go over slices. //From the rust book: Another data type that does not have ownership is the slice. //Slices let you reference a contiguous sequence of elements in a collection rather than the whole collection. // let mut s = String::from("hello world"); // let word = first_word(&s); // word will get the value 5. // s.clear(); // This empties the String, making it equal to "". // word still has the value 5 here, but there's no more string that // we could meaningfully use the value 5 with. word is now totally invalid! //The index we got is now completely out of sync with our original string. //If we end up having more indices we could get even more out of sync with our data. //For strings we can take advantage of a built in feature called string slices. //They create a reference to portions of a string. let s = String::from("hello world"); //Slicing is similar to slicing in python where you have a starting index and then //the ending value is +1 of the data you actually care about. let hello = &s[0..5]; // let hello = &s[..5]; //Equivalent to the above let world = &s[6..11]; // let world = &s[6..]; //Equivalent to the above let len = s.len(); let slice = &s[0..len]; // let slice = &s[..]; //Equivalent to the above // We now have a straightforward API that’s much harder to mess up, since the compiler will //ensure the references into the String remain valid. Remember the bug in the program in Listing 4-11, //when we got the index to the end of the first word but then cleared the string so our index was invalid? //That code was logically incorrect but didn’t show any immediate errors. The problems would show up later //if we kept trying to use the first word index with an emptied string. Slices make this bug impossible //and let us know we have a problem with our code much sooner. Using the slice version of first_word //will throw a compile time error: // let mut s = String::from("hello world"); // let word = first_word(&s); // s.clear(); // Error! // Recall from the borrowing rules that if we have an immutable reference to something, we cannot also // take a mutable reference. Because clear needs to truncate the String, it tries to take a mutable reference, // which fails. Not only has Rust made our API easier to use, but it has also eliminated an entire class of errors // at compile time! let s = "Hello, world!"; // The type of s here is &str: it’s a slice pointing to that specific point of the binary. This is also why string // literals are immutable; &str is an immutable reference. let my_string = String::from("hello world"); // first_word works on slices of `String`s let word = first_word(&my_string[..]); let my_string_literal = "hello world"; // first_word works on slices of string literals let word = first_word(&my_string_literal[..]); // since string literals *are* string
println!("{}", s);
random_line_split
mod.rs
// Parity Ethereum is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with Parity Ethereum. If not, see <http://www.gnu.org/licenses/>. //! Utilities and helpers for transaction dispatch. pub(crate) mod light; mod full; mod prospective_signer; #[cfg(any(test, feature = "accounts"))] mod signing; #[cfg(not(any(test, feature = "accounts")))] mod signing { use super::*; use v1::helpers::errors; /// Dummy signer implementation #[derive(Debug, Clone)] pub struct Signer; impl Signer { /// Create new instance of dummy signer (accept any AccountProvider) pub fn new<T>(_ap: T) -> Self { Signer } } impl super::Accounts for Signer { fn sign_transaction(&self, _filled: FilledTransactionRequest, _chain_id: Option<u64>, _nonce: U256, _password: SignWith) -> Result<WithToken<SignedTransaction>> { Err(errors::account("Signing unsupported", "See #9997")) } fn sign_message(&self, _address: Address, _password: SignWith, _hash: SignMessage) -> Result<WithToken<Signature>> { Err(errors::account("Signing unsupported", "See #9997")) } fn decrypt(&self, _address: Address, _password: SignWith, _data: Bytes) -> Result<WithToken<Bytes>> { Err(errors::account("Signing unsupported", "See #9997")) } fn supports_prospective_signing(&self, _address: &Address, _password: &SignWith) -> bool { false } fn default_account(&self) -> Address { Default::default() } fn is_unlocked(&self, _address: &Address) -> bool { false } } } pub use self::light::LightDispatcher; pub use self::full::FullDispatcher; pub use self::signing::Signer; pub use v1::helpers::nonce::Reservations; use std::fmt::Debug; use std::ops::Deref; use std::sync::Arc; use bytes::Bytes; use client_traits::BlockChainClient; use ethcore::miner::MinerService; use ethereum_types::{H520, H256, U256, Address}; use ethkey::{Password, Signature}; use hash::keccak; use types::transaction::{SignedTransaction, PendingTransaction}; use jsonrpc_core::{BoxFuture, Result, Error}; use jsonrpc_core::futures::{future, Future, IntoFuture}; use v1::helpers::{TransactionRequest, FilledTransactionRequest, ConfirmationPayload}; use v1::types::{ Bytes as RpcBytes, RichRawTransaction as RpcRichRawTransaction, ConfirmationPayload as RpcConfirmationPayload, ConfirmationResponse, EthSignRequest as RpcEthSignRequest, EIP191SignRequest as RpcSignRequest, DecryptRequest as RpcDecryptRequest, }; /// Has the capability to dispatch, sign, and decrypt. /// /// Requires a clone implementation, with the implication that it be cheap; /// usually just bumping a reference count or two. pub trait Dispatcher: Send + Sync + Clone { // TODO: when ATC exist, use zero-cost // type Out<T>: IntoFuture<T, Error> /// Fill optional fields of a transaction request, fetching gas price but not nonce. fn fill_optional_fields(&self, request: TransactionRequest, default_sender: Address, force_nonce: bool) -> BoxFuture<FilledTransactionRequest>; /// Sign the given transaction request without dispatching, fetching appropriate nonce. fn sign<P>( &self, filled: FilledTransactionRequest, signer: &Arc<dyn Accounts>, password: SignWith, post_sign: P, ) -> BoxFuture<P::Item> where P: PostSign + 'static, <P::Out as futures::future::IntoFuture>::Future: Send; /// Converts a `SignedTransaction` into `RichRawTransaction` fn enrich(&self, SignedTransaction) -> RpcRichRawTransaction; /// "Dispatch" a local transaction. fn dispatch_transaction(&self, signed_transaction: PendingTransaction) -> Result<H256>; } /// Payload to sign pub enum SignMessage { /// Eth-sign kind data (requires prefixing) Data(Bytes), /// Prefixed data hash Hash(H256), } /// Abstract transaction signer. /// /// NOTE This signer is semi-correct, it's a temporary measure to avoid moving too much code. /// If accounts are ultimately removed all password-dealing endpoints will be wiped out. pub trait Accounts: Send + Sync { /// Sign given filled transaction request for the specified chain_id. fn sign_transaction(&self, filled: FilledTransactionRequest, chain_id: Option<u64>, nonce: U256, password: SignWith) -> Result<WithToken<SignedTransaction>>; /// Sign given message. fn sign_message(&self, address: Address, password: SignWith, hash: SignMessage) -> Result<WithToken<Signature>>; /// Decrypt given message. fn decrypt(&self, address: Address, password: SignWith, data: Bytes) -> Result<WithToken<Bytes>>; /// Returns `true` if the accounts can sign multiple times. fn supports_prospective_signing(&self, address: &Address, password: &SignWith) -> bool; /// Returns default account. fn default_account(&self) -> Address; /// Returns true if account is unlocked (i.e. can sign without a password) fn is_unlocked(&self, address: &Address) -> bool; } /// action to execute after signing /// e.g importing a transaction into the chain pub trait PostSign: Send { /// item that this PostSign returns type Item: Send; /// incase you need to perform async PostSign actions type Out: IntoFuture<Item = Self::Item, Error = Error> + Send; /// perform an action with the signed transaction fn execute(self, signer: WithToken<SignedTransaction>) -> Self::Out; } impl PostSign for () { type Item = WithToken<SignedTransaction>; type Out = Result<Self::Item>; fn execute(self, signed: WithToken<SignedTransaction>) -> Self::Out { Ok(signed) } } impl<F: Send, T: Send> PostSign for F where F: FnOnce(WithToken<SignedTransaction>) -> Result<T> { type Item = T; type Out = Result<Self::Item>; fn execute(self, signed: WithToken<SignedTransaction>) -> Self::Out { (self)(signed) } } /// Single-use account token. pub type AccountToken = Password; /// Values used to unlock accounts for signing. #[derive(Clone, PartialEq)] pub enum SignWith { /// Nothing -- implies the account is already unlocked. Nothing, /// Unlock with password. Password(Password), /// Unlock with single-use token. Token(AccountToken), } impl SignWith { #[cfg(any(test, feature = "accounts"))] fn is_password(&self) -> bool { if let SignWith::Password(_) = *self { true } else { false } } } /// A value, potentially accompanied by a signing token. pub enum WithToken<T> { /// No token. No(T), /// With token. Yes(T, AccountToken), } impl<T: Debug> Deref for WithToken<T> { type Target = T; fn deref(&self) -> &Self::Target { match *self { WithToken::No(ref v) => v, WithToken::Yes(ref v, _) => v, } } } impl<T: Debug> WithToken<T> { /// Map the value with the given closure, preserving the token. pub fn map<S, F>(self, f: F) -> WithToken<S> where S: Debug, F: FnOnce(T) -> S, { match self { WithToken::No(v) => WithToken::No(f(v)), WithToken::Yes(v, token) => WithToken::Yes(f(v), token), } } /// Convert into inner value, ignoring possible token. pub fn into_value(self) -> T { match self { WithToken::No(v) => v, WithToken::Yes(v, _) => v, } } /// Convert the `WithToken` into a tuple. pub fn into_tuple(self) -> (T, Option<AccountToken>) { match self { WithToken::No(v) => (v, None), WithToken::Yes(v, token) => (v, Some(token)) } } } impl<T: Debug> From<(T, AccountToken)> for WithToken<T> { fn from(tuple: (T, AccountToken)) -> Self { WithToken::Yes(tuple.0, tuple.1) } } impl<T: Debug> From<(T, Option<AccountToken>)> for WithToken<T> { fn from(tuple: (T, Option<AccountToken>)) -> Self { match
// (at your option) any later version.
random_line_split
mod.rs
64>, _nonce: U256, _password: SignWith) -> Result<WithToken<SignedTransaction>> { Err(errors::account("Signing unsupported", "See #9997")) } fn sign_message(&self, _address: Address, _password: SignWith, _hash: SignMessage) -> Result<WithToken<Signature>> { Err(errors::account("Signing unsupported", "See #9997")) } fn decrypt(&self, _address: Address, _password: SignWith, _data: Bytes) -> Result<WithToken<Bytes>> { Err(errors::account("Signing unsupported", "See #9997")) } fn supports_prospective_signing(&self, _address: &Address, _password: &SignWith) -> bool { false } fn default_account(&self) -> Address { Default::default() } fn is_unlocked(&self, _address: &Address) -> bool { false } } } pub use self::light::LightDispatcher; pub use self::full::FullDispatcher; pub use self::signing::Signer; pub use v1::helpers::nonce::Reservations; use std::fmt::Debug; use std::ops::Deref; use std::sync::Arc; use bytes::Bytes; use client_traits::BlockChainClient; use ethcore::miner::MinerService; use ethereum_types::{H520, H256, U256, Address}; use ethkey::{Password, Signature}; use hash::keccak; use types::transaction::{SignedTransaction, PendingTransaction}; use jsonrpc_core::{BoxFuture, Result, Error}; use jsonrpc_core::futures::{future, Future, IntoFuture}; use v1::helpers::{TransactionRequest, FilledTransactionRequest, ConfirmationPayload}; use v1::types::{ Bytes as RpcBytes, RichRawTransaction as RpcRichRawTransaction, ConfirmationPayload as RpcConfirmationPayload, ConfirmationResponse, EthSignRequest as RpcEthSignRequest, EIP191SignRequest as RpcSignRequest, DecryptRequest as RpcDecryptRequest, }; /// Has the capability to dispatch, sign, and decrypt. /// /// Requires a clone implementation, with the implication that it be cheap; /// usually just bumping a reference count or two. pub trait Dispatcher: Send + Sync + Clone { // TODO: when ATC exist, use zero-cost // type Out<T>: IntoFuture<T, Error> /// Fill optional fields of a transaction request, fetching gas price but not nonce. fn fill_optional_fields(&self, request: TransactionRequest, default_sender: Address, force_nonce: bool) -> BoxFuture<FilledTransactionRequest>; /// Sign the given transaction request without dispatching, fetching appropriate nonce. fn sign<P>( &self, filled: FilledTransactionRequest, signer: &Arc<dyn Accounts>, password: SignWith, post_sign: P, ) -> BoxFuture<P::Item> where P: PostSign + 'static, <P::Out as futures::future::IntoFuture>::Future: Send; /// Converts a `SignedTransaction` into `RichRawTransaction` fn enrich(&self, SignedTransaction) -> RpcRichRawTransaction; /// "Dispatch" a local transaction. fn dispatch_transaction(&self, signed_transaction: PendingTransaction) -> Result<H256>; } /// Payload to sign pub enum SignMessage { /// Eth-sign kind data (requires prefixing) Data(Bytes), /// Prefixed data hash Hash(H256), } /// Abstract transaction signer. /// /// NOTE This signer is semi-correct, it's a temporary measure to avoid moving too much code. /// If accounts are ultimately removed all password-dealing endpoints will be wiped out. pub trait Accounts: Send + Sync { /// Sign given filled transaction request for the specified chain_id. fn sign_transaction(&self, filled: FilledTransactionRequest, chain_id: Option<u64>, nonce: U256, password: SignWith) -> Result<WithToken<SignedTransaction>>; /// Sign given message. fn sign_message(&self, address: Address, password: SignWith, hash: SignMessage) -> Result<WithToken<Signature>>; /// Decrypt given message. fn decrypt(&self, address: Address, password: SignWith, data: Bytes) -> Result<WithToken<Bytes>>; /// Returns `true` if the accounts can sign multiple times. fn supports_prospective_signing(&self, address: &Address, password: &SignWith) -> bool; /// Returns default account. fn default_account(&self) -> Address; /// Returns true if account is unlocked (i.e. can sign without a password) fn is_unlocked(&self, address: &Address) -> bool; } /// action to execute after signing /// e.g importing a transaction into the chain pub trait PostSign: Send { /// item that this PostSign returns type Item: Send; /// incase you need to perform async PostSign actions type Out: IntoFuture<Item = Self::Item, Error = Error> + Send; /// perform an action with the signed transaction fn execute(self, signer: WithToken<SignedTransaction>) -> Self::Out; } impl PostSign for () { type Item = WithToken<SignedTransaction>; type Out = Result<Self::Item>; fn execute(self, signed: WithToken<SignedTransaction>) -> Self::Out { Ok(signed) } } impl<F: Send, T: Send> PostSign for F where F: FnOnce(WithToken<SignedTransaction>) -> Result<T> { type Item = T; type Out = Result<Self::Item>; fn execute(self, signed: WithToken<SignedTransaction>) -> Self::Out { (self)(signed) } } /// Single-use account token. pub type AccountToken = Password; /// Values used to unlock accounts for signing. #[derive(Clone, PartialEq)] pub enum SignWith { /// Nothing -- implies the account is already unlocked. Nothing, /// Unlock with password. Password(Password), /// Unlock with single-use token. Token(AccountToken), } impl SignWith { #[cfg(any(test, feature = "accounts"))] fn is_password(&self) -> bool { if let SignWith::Password(_) = *self { true } else { false } } } /// A value, potentially accompanied by a signing token. pub enum WithToken<T> { /// No token. No(T), /// With token. Yes(T, AccountToken), } impl<T: Debug> Deref for WithToken<T> { type Target = T; fn
(&self) -> &Self::Target { match *self { WithToken::No(ref v) => v, WithToken::Yes(ref v, _) => v, } } } impl<T: Debug> WithToken<T> { /// Map the value with the given closure, preserving the token. pub fn map<S, F>(self, f: F) -> WithToken<S> where S: Debug, F: FnOnce(T) -> S, { match self { WithToken::No(v) => WithToken::No(f(v)), WithToken::Yes(v, token) => WithToken::Yes(f(v), token), } } /// Convert into inner value, ignoring possible token. pub fn into_value(self) -> T { match self { WithToken::No(v) => v, WithToken::Yes(v, _) => v, } } /// Convert the `WithToken` into a tuple. pub fn into_tuple(self) -> (T, Option<AccountToken>) { match self { WithToken::No(v) => (v, None), WithToken::Yes(v, token) => (v, Some(token)) } } } impl<T: Debug> From<(T, AccountToken)> for WithToken<T> { fn from(tuple: (T, AccountToken)) -> Self { WithToken::Yes(tuple.0, tuple.1) } } impl<T: Debug> From<(T, Option<AccountToken>)> for WithToken<T> { fn from(tuple: (T, Option<AccountToken>)) -> Self { match tuple.1 { Some(token) => WithToken::Yes(tuple.0, token), None => WithToken::No(tuple.0), } } } /// Execute a confirmation payload. pub fn execute<D: Dispatcher + 'static>( dispatcher: D, signer: &Arc<dyn Accounts>, payload: ConfirmationPayload, pass: SignWith ) -> BoxFuture<WithToken<ConfirmationResponse>> { match payload { ConfirmationPayload::SendTransaction(request) => { let condition = request.condition.clone().map(Into::into); let cloned_dispatcher = dispatcher.clone(); let post_sign = move |with_token_signed: WithToken<SignedTransaction>| { let (signed, token) = with_token_signed.into_tuple(); let signed_transaction = PendingTransaction::new(signed, condition); cloned_dispatcher.dispatch_transaction(signed_transaction) .map(|hash| (hash, token)) }; Box::new( dispatcher.sign(request, &signer, pass, post_sign).map(|(hash, token)| { WithToken::from((ConfirmationResponse::SendTransaction(hash), token)) }) )
deref
identifier_name
environment.py
logger.info("Note: {e} environment variable is specified, but tests are " "still run locally\n" "Check other values required to run tests against existing " "deployent".format(e=env_var_name)) def _missing_api_token_warning(env_var_name): if os.environ.get(env_var_name): logger.info("OK: {name} environment is set and will be used as " "authorization token".format(name=env_var_name)) else: logger.info("Warning: the {name} environment variable is not" " set.\n" "Most tests that require authorization will probably fail".format( name=env_var_name)) def _check_api_tokens_presence(): # we need RECOMMENDER_API_TOKEN or RECOMMENDER_REFRESH_TOKEN to be set if not os.environ.get("RECOMMENDER_REFRESH_TOKEN"): _missing_api_token_warning("RECOMMENDER_API_TOKEN") else: _missing_api_token_warning("RECOMMENDER_REFRESH_TOKEN") _missing_api_token_warning("JOB_API_TOKEN") def _check_env_var_presence_s3_db(env_var_name): """Check if given environment variable exist. Check the existence of environment variable needed to connect to the AWS S3 database. """ if os.environ.get(env_var_name) is None: logger.info("Warning: the {name} environment variable is not set.\n" "All tests that access AWS S3 database will fail\n".format( name=env_var_name)) def _parse_int_env_var(env_var_name): val = os.environ.get(env_var_name) try: return int(val) except (TypeError, ValueError): return None def _read_url_from_env_var(env_var_name): return _add_slash(os.environ.get(env_var_name, None)) def check_test_environment(context, coreapi_url): """Check the test environent - whether tests are run locally or in Docker.""" if context.running_locally: logger.info("Note: integration tests are running localy via docker-compose") if coreapi_url: _check_env_for_remote_tests("F8A_API_URL") _check_env_for_remote_tests("F8A_JOB_API_URL") _check_env_for_remote_tests("F8A_GEMINI_API_URL") else: logger.info("Note: integration tests are running against existing deployment") _check_api_tokens_presence() def _running_locally(coreapi_url, jobs_api_url): """Check if tests are running locally.""" return not (coreapi_url and jobs_api_url) def _get_url(context, actual, attribute_name, port): """Get the URL + port for the selected service.""" return actual or _get_api_url(context, attribute_name, port) def check_token_structure(data): """Check the basic structure of response with access token.""" assert "token" in data token_structure = data["token"] assert "access_token" in token_structure assert "token_type" in token_structure assert "expires_in" in token_structure def before_all(context): """Perform the setup before the first event.""" context.config.setup_logging() context.start_system = _start_system context.teardown_system = _teardown_system context.restart_system = _restart_system context.run_command_in_service = _run_command_in_service context.exec_command_in_container = _exec_command_in_container context.is_running = _is_running context.is_jobs_debug_api_running = _is_jobs_debug_api_running context.is_component_search_service_running = _is_component_search_service_running context.is_master_tag_list_service_running = _is_master_tag_list_service_running context.wait_for_master_tag_list_service = _wait_for_master_tag_list_service context.is_get_untagged_component_service_running = _is_get_untagged_component_service_running context.wait_for_get_untagged_component_service = _wait_for_get_untagged_component_service context.send_json_file = _send_json_file context.wait_for_jobs_debug_api_service = _wait_for_jobs_debug_api_service context.wait_for_component_search_service = _wait_for_component_search_service context.is_3scale_staging_running = _is_3scale_staging_running context.is_3scale_preview_running = _is_3scale_preview_running context.is_backbone_api_running = _is_backbone_api_running context.is_gemini_api_running = _is_gemini_api_running # Configure container logging context.dump_logs = _read_boolean_setting(context, 'dump_logs') tail_logs = int(context.config.userdata.get('tail_logs', 0)) dump_errors = _read_boolean_setting(context, 'dump_errors') if tail_logs: dump_errors = True else: tail_logs = 50 context.dump_errors = dump_errors context.tail_logs = tail_logs # Configure system under test context.kubernetes_dir_path = context.config.userdata.get('kubernetes_dir', None) if context.kubernetes_dir_path is not None: context.docker_compose_path = None else: # If we're not running Kubernetes, use the local Docker Compose setup _set_default_compose_path(context) # for now, we just assume we know what compose file looks like (what services need what images) context.images = {} context.images['bayesian/bayesian-api'] = context.config.userdata.get( 'coreapi_server_image', 'registry.devshift.net/bayesian/bayesian-api') context.images['bayesian/cucos-worker'] = context.config.userdata.get( 'coreapi_worker_image', 'registry.devshift.net/bayesian/cucos-worker') core_v2_api_url = _read_url_from_env_var('F8A_API_V2_URL') coreapi_url = _read_url_from_env_var('F8A_API_URL') core_v2_api_url = _read_url_from_env_var('F8A_API_V2_URL') valid_synk_token = os.environ.get("SNYK_TOKEN") uuid = os.environ.get("REGISTERED_USER_UUID") jobs_api_url = _read_url_from_env_var('F8A_JOB_API_URL') gremlin_url = _read_url_from_env_var('F8A_GREMLIN_URL') threescale_url = _read_url_from_env_var('F8A_3SCALE_URL') threescale_preview_url = _read_url_from_env_var('F8A_THREE_SCALE_PREVIEW_URL') backbone_api_url = _read_url_from_env_var('F8A_BACKBONE_API_URL') service_id = _read_url_from_env_var('F8A_SERVICE_ID') gemini_api_url = _read_url_from_env_var('F8A_GEMINI_API_URL') license_service_url = _read_url_from_env_var('F8A_LICENSE_SERVICE_URL') context.running_locally = _running_locally(coreapi_url, jobs_api_url) check_test_environment(context, coreapi_url) context.coreapi_url = _get_url(context, coreapi_url, 'coreapi_url', _FABRIC8_ANALYTICS_SERVER) context.core_v2_api_url = core_v2_api_url context.jobs_api_url = _get_url(context, jobs_api_url, 'jobs_api_url', _FABRIC8_ANALYTICS_JOBS) context.gremlin_url = _get_url(context, gremlin_url, "gremlin_url", _FABRIC8_GREMLIN_SERVICE) context.license_service_url = _get_url(context, license_service_url, 'license_service_url', _FABRIC8_LICENSE_SERVICE) context.core_v2_api_url = core_v2_api_url context.threescale_url = threescale_url context.valid_synk_token = valid_synk_token context.uuid = uuid context.threescale_preview_url = threescale_preview_url context.backbone_api_url = backbone_api_url context.service_id = service_id context.gemini_api_url = gemini_api_url # we can retrieve access token by using refresh/offline token # informations needed to access S3 database from tests _check_env_var_presence_s3_db('AWS_ACCESS_KEY_ID') _check_env_var_presence_s3_db('AWS_SECRET_ACCESS_KEY') _check_env_var_presence_s3_db('S3_REGION_NAME') aws_access_key_id = os.environ.get('AWS_ACCESS_KEY_ID') aws_secret_access_key = os.environ.get('AWS_SECRET_ACCESS_KEY') s3_region_name = os.environ.get('S3_REGION_NAME') deployment_prefix = os.environ.get('DEPLOYMENT_PREFIX', 'STAGE') context.reports_bucket = os.environ.get('DEVELOPER_ANALYTICS_REPORTS_BUCKET') context.s3interface = S3Interface(aws_access_key_id, aws_secret_access_key, s3_region_name, deployment_prefix) context.client = None # timeout values can be overwritten by environment variables stack_analysis_timeout = _parse_int_env_var('F8A_STACK_ANALYSIS_TIMEOUT') component_analysis_timeout = _parse_int_env_var('F8A_COMPONENT_ANALYSIS_TIMEOUT') context.stack_analysis_timeout = stack_analysis_timeout or _DEFAULT_STACK_ANALYSIS_TIMEOUT context.component_analysis_timeout = component_analysis_timeout \ or _DEFAULT_COMPONENT_ANALYSIS_TIMEOUT if context.running_locally: context.client = docker.AutoVersionClient() for desired, actual in context.images.items(): desired = 'registry.devshift.net/' + desired if desired != actual:
context.client.tag(actual, desired, force=True)
conditional_block
environment.py
output def _get_k8s_volumes_to_delete(): # universal_newlines decodes output on Python 3.x out = subprocess.check_output(['kubectl', 'get', 'pods', '-o', 'json'], universal_newlines=True) j = json.loads(out) volumes = [] for pod in j['items']: pod_vols = pod['spec'].get('volumes', []) for pod_vol in pod_vols: if 'hostPath' in pod_vol: volumes.append(pod_vol['hostPath']['path']) return volumes def _dump_server_logs(context, tail=None): if context.docker_compose_path: cmd = _make_compose_command(context, 'logs') if tail is not None: cmd.append('--tail={:d}'.format(tail)) subprocess.check_call(cmd, stderr=subprocess.STDOUT) else: pass # No current support for dumping logs under k8s def _teardown_system(context): cmds = [] if context.docker_compose_path: cmds.append(_make_compose_command(context, 'kill')) cmds.append(_make_compose_command(context, 'rm', '-fv')) if hasattr(context, "container"): cmds.append(['docker', "kill", context.container]) cmds.append(['docker', "rm", "-fv", "--rm-all", context.container]) _set_default_compose_path(context) else: cmds.append(['kubectl', 'delete', '--ignore-not-found', '-f', context.kubernetes_dir_path]) volumes = _get_k8s_volumes_to_delete() for volume in volumes: # TODO: the sudo thing is not very nice, but... cmds.append(['sudo', 'rm', '-rf', volume]) cmds.append(['sudo', 'mkdir', volume]) for cmd in cmds: subprocess.check_output(cmd, stderr=subprocess.STDOUT) def _post_startup(context, started_all, wait_for_server): """Post startup actions.""" if started_all: # let's give the whole system a while to breathe time.sleep(float(context.config.userdata.get('breath_time', 5))) else: raise Exception('Server failed to start in under {s} seconds'. format(s=wait_for_server)) def _wait_for_system(context, wait_for_server=60): start = datetime.datetime.utcnow() wait_till = start + datetime.timedelta(seconds=wait_for_server) # try to wait for server to start for some time while datetime.datetime.utcnow() < wait_till: time.sleep(1) started_all = False if context.kubernetes_dir_path: res = json.loads(subprocess.check_output(['kubectl', 'get', 'pods', '-o', 'json'])) for pod in res['items']: status = pod.get('status', {}) conditions = status.get('conditions', []) phase = status.get('phase', '') if status == {}: continue if phase != 'Running': continue for condition in conditions: if condition['type'] == 'Ready' and condition['status'] != 'True': continue # if we got here, then everything is running started_all = True break else: if _is_running(context): started_all = True break _post_startup(context, started_all, wait_for_server) def _wait_for_api(context, wait_for_service, check_function): for _ in range(wait_for_service): if check_function(context): break time.sleep(1) else: raise Exception('Timeout waiting for the API service') def _wait_for_jobs_debug_api_service(context, wait_for_service=60): _wait_for_api(context, wait_for_service, _is_jobs_debug_api_running) def _wait_for_component_search_service(context, wait_for_service=60): _wait_for_api(context, wait_for_service, _is_component_search_service_running) def _wait_for_master_tag_list_service(context, wait_for_service=60): _wait_for_api(context, wait_for_service, _is_master_tag_list_service_running) def _wait_for_get_untagged_component_service(context, wait_for_service=60): _wait_for_api(context, wait_for_service, _is_get_untagged_component_service_running) def _restart_system(context, wait_for_server=60): # NOTE: it does make sense to restart the local system only if context.running_locally: try: _teardown_system(context) _start_system(context) _wait_for_system(context, wait_for_server) except subprocess.CalledProcessError as e: raise Exception('Failed to restart system. Command "{c}" failed:\n{o}'. format(c=' '.join(e.cmd), o=e.output)) def _is_api_running(url, accepted_codes=None): accepted_codes = accepted_codes or {200, 401} try: res = requests.get(url) if res.status_code in accepted_codes: return True except requests.exceptions.ConnectionError: pass return False def _is_3scale_staging_running(threescale_url, accepted_codes={200, 401}):
def _is_3scale_preview_running(context, accepted_codes={200, 403, 401}): try: res = requests.post(context.threescale_preview_url) if res.status_code in accepted_codes: return True except requests.exceptions.ConnectionError: pass return False def _is_backbone_api_running(backbone_api_url, accepted_codes={200}): try: url = '%s/api/v1/readiness' % backbone_api_url res = requests.get(url) if res.status_code in accepted_codes: return True except requests.exceptions.ConnectionError: pass return False def _is_gemini_api_running(gemini_api_url, accepted_codes={200}): try: url = '%s/api/v1/readiness' % gemini_api_url res = requests.get(url) if res.status_code in accepted_codes: return True except requests.exceptions.ConnectionError: pass return False def _is_api_running_post(url): try: res = requests.post(url) if res.status_code in {200, 401}: return True except requests.exceptions.ConnectionError: pass return False def _is_running(context): return _is_api_running(context.coreapi_url + _API_ENDPOINT) and \ _is_api_running(context.jobs_api_url + _API_ENDPOINT) and \ _is_api_running(context.gremlin_url, {400}) def _is_jobs_debug_api_running(context): return _is_api_running(context.jobs_api_url + _JOBS_DEBUG_API + "/analyses-report?ecosystem=maven") def _is_component_search_service_running(context): return _is_api_running(context.coreapi_url + _API_ENDPOINT + "/component-search/any-component") def _is_master_tag_list_service_running(context): return _is_api_running(context.coreapi_url + _API_ENDPOINT + "/master-tags/maven") def _is_get_untagged_component_service_running(context): return _is_api_running_post(context.coreapi_url + _API_ENDPOINT + "/get-next-component/maven") def _read_boolean_setting(context, setting_name): setting = context.config.userdata.get(setting_name, '').lower() if setting in ('1', 'yes', 'true', 'on'): return True if setting in ('', '0', 'no', 'false', 'off'): return False msg = '{!r} is not a valid option for boolean setting {!r}' raise ValueError(msg.format(setting, setting_name)) def _add_slash(url): if url and not url.endswith('/'): url += '/' return url def _get_api_url(context, attribute, port): return _add_slash(context.config.userdata.get(attribute, 'http://localhost:{port}/'.format(port=port))) def _send_json_file(endpoint, filename, custom_headers=None): """Send the JSON file to the selected API endpoint. The optional custom header is used (given it is provided). """ headers = {'Content-Type': 'application/json', 'Accept': 'application/json'} if custom_headers is not None: headers.update(custom_headers) with open(filename) as json_data: response = requests.post(endpoint, data=json_data, headers=headers) return response def _check_env_for_remote_tests(env_var_name): if os.environ.get(env_var_name): logger.info("Note: {e} environment variable is specified, but tests are " "still run locally\n" "Check other values required to run tests against existing " "deployent".format(e=env_var_name)) def _missing_api_token_warning(env_var_name): if os.environ.get(env_var_name): logger.info("OK: {name} environment is set and will be used as " "authorization token".format(name=env_var_name)) else: logger.info("Warning: the {name} environment variable is not" " set.\n" "Most tests that require authorization will probably fail".format( name=env_var_name)) def _check_api_tokens_presence(): # we need
try: res = requests.post(threescale_url) if res.status_code in accepted_codes: return True except requests.exceptions.ConnectionError: pass return False
identifier_body