content
stringlengths
228
999k
pred_label
stringclasses
1 value
pred_score
float64
0.5
1
Модуль:DateI18n Матеріал з Релігія в огні Документацію для цього модуля можна створити у Модуль:DateI18n/документація --[[ __ __ _ _ ____ _ ___ _ ___ | \/ | ___ __| |_ _| | ___ _| _ \ __ _| |_ ___|_ _/ |( _ ) _ __ | |\/| |/ _ \ / _` | | | | |/ _ (_) | | |/ _` | __/ _ \| || |/ _ \| '_ \ | | | | (_) | (_| | |_| | | __/_| |_| | (_| | || __/| || | (_) | | | | |_| |_|\___/ \__,_|\__,_|_|\___(_)____/ \__,_|\__\___|___|_|\___/|_| |_| This module is intended for processing of date strings. Please do not modify this code without applying the changes first at Module:Date/sandbox and testing at Module:Date/sandbox/testcases and Module talk:Date/sandbox/testcases. Authors and maintainers: * User:Parent5446 - original version of the function mimicking template:ISOdate * User:Jarekt - original version of the functions mimicking template:Date ]] -- ======================================= -- === Dependencies ====================== -- ======================================= require('strict') -- ======================================= -- === Local Functions =================== -- ======================================= ------------------------------------------------------------------------------ --[[ (copied from Module:Core) Function allowing for consistent treatment of boolean-like wikitext input. Inputs: 1) val - value to be evaluated, outputs as a function of values: true : true (boolean), 1 (number), or strings: "yes", "y", "true", "1" false : false (boolean), 0 (number), or strings: "no", "n", "false", "0" 2) default - value to return otherwise See Also: It works similarly to Module:Yesno ]] local function yesno(val, default) if type(val) == 'boolean' then return val elseif type(val) == 'number' then val = tostring(val) end if type(val) == 'string' then local LUT = { yes=true , y=true , ['true'] =true , t=true , ['1']=true , on =true, no =false, n=false, ['false']=false, f=false, ['0']=false, off=false } val = LUT[mw.ustring.lower(val)] -- put in lower case if (val~=nil) then return val end end return default end --------------------------------------------------------------------------------------- -- trim leading zeros in years prior to year 1000 -- INPUT: -- * datestr - translated date string -- * lang - language of translation -- OUTPUT: -- * datestr - updated date string local function trimYear(datestr, year, lang) local yearStr0, yearStr1, yearStr2, zeroStr yearStr0 = string.format('%04i', year ) -- 4 digit year in standard form "0123" yearStr1 = mw.language.new(lang):formatDate( 'Y', yearStr0) -- same as calling {{#time}} parser function --yearStr1 = mw.getCurrentFrame():callParserFunction( "#time", { 'Y', yearStr0, lang } ) -- translate to a language if yearStr0==yearStr1 then -- most of languages use standard form of year yearStr2 = tostring(year) else -- some languages use different characters for numbers yearStr2 = yearStr1 zeroStr = mw.ustring.sub(yearStr1,1,1) -- get "0" in whatever language for i=1,3 do -- trim leading zeros if mw.ustring.sub(yearStr2,1,1)==zeroStr then yearStr2 = mw.ustring.sub(yearStr2, 2, 5-i) else break end end end return string.gsub(datestr, yearStr1, yearStr2 ) -- in datestr replace long year with trimmed one end --------------------------------------------------------------------------------------- -- Look up proper format string to be passed to {{#time}} parser function -- INPUTS: -- * datecode: YMDhms, YMDhm, YMD, YM, Y, MDhms, MDhm, MD, or M -- * day : Number between 1 and 31 (not needed for most languages) -- * lang : language -- OUTPUT: -- * dFormat : input to {{#time}} function local function getDateFormat(datecode, day, lang) local function parseFormat(dFormat, day) if dFormat:find('default') and #dFormat>10 then -- Special (and messy) case of dFormat code depending on a day number, where data is a -- JSON-encoded table {”default”:”*”,”dDD”:”*”} including fields for specific 2-digit days. -- Change curly double quotes (possibly used for easier editing in tabular data) in dFormat -- to straight ASCII double quotes (required for parsing of this JSON-encoded table). local D = mw.text.jsonDecode(mw.ustring.gsub(dFormat, '[„“‟”]', '"')) --com = mw.dumpObject(D) -- If the desired day is not in that JSON table, then use its "default" case. dFormat = D[string.format('d%02i', day)] or D.default -- Change ASCII single quotes to ASCII double quotes used for {{#time}} marking. -- Apostrophes needed in plain-text must not use ASCII single quotes but curly apostrophe -- e.g. { ‟default”: ‟j”, ‟d01”: ‟j’'o'” }, not { ‟default”: ‟j”, ‟d01”: ‟j''o'” }. end dFormat = dFormat:gsub("'", '"') return dFormat end local T = {} local tab = mw.ext.data.get('DateI18n.tab', lang) for _, row in pairs(tab.data) do -- convert the output into a dictionary table local id, _, msg = unpack(row) T[id] = msg end -- Compatibility of legacy data using 'HMS' or 'HM', where 'M' is ambiguous T.YMDhms = T.YMDhms or T.YMDHMS T.YMDhm = T.YMDhm or T.YMDHM datecode = datecode == 'YMDHMS' and 'YMDhms' or datecode == 'YMDHM' and 'YMDhm' or datecode local dFormat = T[datecode] if dFormat == 'default' and (datecode == 'YMDhms' or datecode == 'YMDhm') then -- For most languages adding hour:minute:second is done by adding ", HH:ii:ss to the -- day precission date, those languages are skipped in DateI18n.tab and default to -- English which stores word "default" dFormat = parseFormat(T['YMD'], day).. ', H:i' if datecode == 'YMDhms' then dFormat = dFormat .. ':s' end else dFormat = parseFormat(dFormat, day) end return dFormat end --------------------------------------------------------------------------------------- -- Look up proper format string to be passed to {{#time}} parser function -- INPUTS: -- * month : month number -- * case : gramatic case abbriviation, like "ins", "loc" -- * lang : language -- OUTPUT: -- * dFormat : input to {{#time}} function local function MonthCase(month, case, lang) if month == nil or case == nil then return nil end local T = {{},{},{},{},{},{},{},{},{},{},{},{}} local tab = mw.ext.data.get('I18n/MonthCases.tab', lang) for _, row in pairs(tab.data) do local mth, cs, msg = unpack(row) T[mth][cs] = msg end return T[month][case] end -- ================================================== -- === External functions =========================== -- ================================================== local p = {} -- =========================================================================== -- === Functions accesible from the outside to allow unit-testing -- === Please do not use directly as they could change in the future -- =========================================================================== --------------------------------------------------------------------------------------- -- Single string replacement that ignores part of the string in "..." function p.strReplace(String, old, new) if String:find('"') then local T={} for i, str in ipairs(mw.text.split( String, '"', true )) do if i%2==1 then str = str:gsub(old, new, 1) end table.insert(T, str) end return table.concat(T,'"') else return String:gsub(old, new, 1) end end --------------------------------------------------------------------------------------- -- process datevec -- INPUT: -- * datevec - Array of {year,month,day,hour,minute,second, tzhour, tzmin} containing broken -- down date-time component strings or numbers -- OUTPUT: -- * datenum - same array but holding only numbers or nuls function p.clean_datevec(datevec) -- create datecode based on which variables are provided and check for out-of-bound values -- check special case of month provided as a name local month = datevec[2] if type(month) == 'string' and month ~= '' and not tonumber(month) then -- When the month is not a number, check if it's a month name in the project's language. datevec[2] = mw.getContentLanguage():formatDate('n', month) end -- check bounds local maxval = { 1/0, 12, 31, 23, 59, 59, 23, 59 } -- max values (or 1/0=+inf) for year, month, day, hour, minute, second, tzhour, tzmin local minval = { -1/0, 01, 01, 00, 00, 00, -23, 00 } -- min values (or -1/0=-inf) for year, month, ... local datenum = {} -- date-time encoded as a vector = [year, month, ... , second, tzhour, tzmin] for i = 1, 8 do local val = tonumber(datevec[i]) if val and val >= minval[i] and val <= maxval[i] then -- These tests work with infinite min/max values. datenum[i] = val end end -- leap second if tonumber(datevec[6]) == 60 then -- leap second '60' is valid only at end of 23:59 UTC, on 30 June or 31 December of specific years -- datenum[6] = 60 local MDhm = table.concat({unpack(datenum,2,5)}, ',') if (MDhm == table.concat({6, 30, 23, 59}, ',')) or (MDhm == table.concat({12, 31, 23, 59}, ',')) then datenum[6] = 60 end end return datenum end --------------------------------------------------------------------------------------- -- process datevec -- INPUT: -- * datenum - Array of {year,month,day,hour,minute,second, tzhour, tzmin} as numbers or nuls -- OUTPUT: -- * timeStamp - date string in the format taken by mw.language:formatDate lua function and {{#time}} parser function -- https://www.mediawiki.org/wiki/Extension:Scribunto/Lua_reference_manual#mw.language:formatDate -- https://www.mediawiki.org/wiki/Help:Extension:ParserFunctions#.23time -- * datecode - a code specifying content of the array where Y' is year, 'M' is month, -- 'D' is day, 'h' is hour, 'm' minute, 's' is second. -- Output has to be one of YMDhms, YMDhm, YMD, YM, Y, MDhms, MDhm, MD, M. function p.getTimestamp(datenum) -- create datecode based on datenum local codes = { 'Y', 'M', 'D', 'h', 'm', 's'} local datecode = '' -- a string signifying which combination of variables was provided for i, c in ipairs(codes) do datecode = datecode .. (datenum[i] and c or '') -- if datenum[i] than append codes[i] to datecode end -- create timestamp string (for example 2000-02-20 02:20:20) based on which variables were provided local timeStamp -- date starting by a year if datecode == 'YMDhms' then timeStamp = string.format('%04i-%02i-%02i %02i:%02i:%02i', datenum[1], datenum[2], datenum[3], datenum[4], datenum[5], datenum[6] ) elseif datecode == 'YMDhm' then timeStamp = string.format('%04i-%02i-%02i %02i:%02i', datenum[1], datenum[2], datenum[3], datenum[4], datenum[5] ) elseif datecode:sub(1,3)=='YMD' then timeStamp = string.format('%04i-%02i-%02i', datenum[1], datenum[2], datenum[3] ) datecode = 'YMD' -- 'YMDhms', 'YMDhm' and 'YMD' are the only supported format starting with 'YMD'; all others will be converted to 'YMD'. elseif datecode:sub(1,2) == 'YM' then timeStamp = string.format('%04i-%02i', datenum[1], datenum[2] ) datecode = 'YM' elseif datecode:sub(1,1)=='Y' then timeStamp = string.format('%04i', datenum[1] ) datecode = 'Y' -- date starting by a month (the implied year is 2000) elseif datecode== 'MDhms' then timeStamp = string.format('%04i-%02i-%02i %02i:%02i:%02i', 2000, datenum[2], datenum[3], datenum[4], datenum[5], datenum[6] ) elseif datecode == 'MDhm' then timeStamp = string.format('%04i-%02i-%02i %02i:%02i', 2000, datenum[2], datenum[3], datenum[4], datenum[5] ) elseif datecode:sub(1,2) == 'MD' then timeStamp = string.format('%04i-%02i-%02i', 2000, datenum[2], datenum[3] ) datecode = 'MD' -- 'MDhms', 'MDhm' and 'MD' are the only supported format starting with 'MD'; all others will be converted to 'MD' elseif datecode:sub(1,1) == 'M' then -- Ambiguous: could mean minutes, but here means month (when parsed as a name/abbrev, not as a number). timeStamp = string.format('%04i-%02i-%02i', 2000, datenum[2], 1 ) datecode = 'M' -- other possible but unrecognized formats (e.g. 'DHis', 'DHi', 'D', 'His', 'Hi'); -- note that 'Dh', 'D', 'h', 's' may eventually work, but not 'm' for minute only, which is ambiguous with 'M' for month only. else timeStamp = nil -- format not supported end return timeStamp, datecode end -- =========================================================================== -- === Version of the function to be called from other LUA codes -- =========================================================================== --[[ ======================================================================================== Date This function is the core part of the ISOdate template. Usage: local Date = require('Module:DateI18n')._Date local dateStr = Date({2020, 12, 30, 12, 20, 11}, lang) Parameters: * {year,month,day,hour,minute,second, tzhour, tzmin}: broken down date-time component strings or numbers tzhour, tzmin are timezone offsets from UTC, hours and minutes * lang: The language to display it in * case: Language format (genitive, etc.) for some languages * class: CSS class for the <time> node, use "" for no metadata at all ]] function p._Date(datevec, lang, case, class, trim_year) -- make sure inputs are in the right format if not lang or not mw.language.isValidCode( lang ) then lang = mw.getCurrentFrame():callParserFunction( "int", "lang" ) -- get user's chosen language end if lang == 'be-tarask' then lang = 'be-x-old' end -- process datevec and extract timeStamp and datecode strings as well as numeric datenum array local datenum = p.clean_datevec(datevec) local year, month, day = datenum[1], datenum[2], datenum[3] local timeStamp, datecode = p.getTimestamp(datenum) if not timeStamp then -- something went wrong in parserDatevec return '' end -- Commons [[Data:DateI18n.tab]] page stores prefered formats for diferent -- languages and datecodes (specifying year-month-day or just year of month-day, etc) -- Look up country specific format input to {{#time}} function local dFormat = getDateFormat(datecode, day, lang) -- By default the gramatical case is not specified (case=='') allowing the format to be specified -- in [[Data:DateI18n.tab]]. You can overwrite the default grammatical case of the month by -- specifying "case" variable. This is needed mostly by Slavic languages to create more complex -- phrases as it is done in [[c:Module:Complex date]] case = case or '' if (lang=='qu' or lang=='qug') and case=='nom' then -- Special case related to Quechua and Kichwa languages. The form in the I18n is -- Genitive case with suffix "pi" added to month names provided by {#time}} -- in Nominative case that "pi" should be removed -- see https://commons.wikimedia.org/wiki/Template_talk:Date#Quechua from 2014 dFormat = dFormat:gsub('F"pi"', 'F') elseif case == 'gen' then dFormat = p.strReplace(dFormat, "F", "xg") elseif case == 'nom' then dFormat = p.strReplace(dFormat, "xg", "F") elseif case ~= '' and month ~= nil then -- see is page [[Data:I18n/MonthCases.tab]] on Commons have name of the month -- in specific gramatic case in desired language. If we have it than replace -- "F" and xg" in dFormat local monthMsg = MonthCase(month, case, lang) if monthMsg and monthMsg ~= '' then -- make sure it exists dFormat = p.strReplace(dFormat, 'F', '"'..monthMsg..'"') -- replace default month with month name we already looked up dFormat = p.strReplace(dFormat, 'xg', '"'..monthMsg..'"') end end -- Translate the date using specified format. -- See https://www.mediawiki.org/wiki/Extension:Scribunto/Lua_reference_manual#mw.language:formatDate and -- https://www.mediawiki.org/wiki/Help:Extension:ParserFunctions##time for explanation of the format local langObj = mw.language.new(lang) local datestr = langObj:formatDate(dFormat, timeStamp) -- same as using {{#time}} parser function -- Special case related to Thai solar calendar: prior to 1940 new-year was at different time of year, -- so just year (datecode == 'Y') is ambiguous and is replaced by "YYYY or YYYY" phrase if lang=='th' and datecode=='Y' and year<=1940 then datestr = string.format('%04i หรือ %04i', year+542, year+543 ) end -- If year < 1000 than either keep the date padded to the length of 4 digits or trim it. -- Decide if the year will stay padded with zeros (for years in 0-999 range). if year and year < 1000 then trim_year = yesno(trim_year, trim_year or '100-999') if type(trim_year) == 'string' then -- If `trim_year` not a simple boolean, then it's a range of dates. -- For example '100-999' means to pad 1-or-2-digit years to be 4-digit long, while keeping 3-digit years as is. local YMin, YMax = trim_year:match( '(%d+)-(%d+)' ) trim_year = YMin and year >= tonumber(YMin) and year <= tonumber(YMax) end if trim_year then datestr = trimYear(datestr, year, lang) -- in datestr replace long year with trimmed one end end -- Append a timezone if present (after the hour and minute of the day). if datenum[7] and (datecode:sub(1, 5) == 'YMDhm' or datecode:sub(1, 4) == 'MDhm') then -- Use {{#time}} parser function to create timezone string, so that we use the correct character set. local sign = (datenum[7]<0) and '−' or '+' timeStamp = string.format("2000-01-01 %02i:%02i:00", math.abs(datenum[7]), datenum[8] or 0) local timezone = langObj:formatDate('H:i', timeStamp) -- same as using {{#time}} parser function datestr = string.format("%s %s%s", datestr, sign, timezone ) end -- HTML formating of date string and tagging for microformats (only for absolute dates with a year). if class and class ~= '' and class ~= '-' and datecode:sub(1,1) == 'Y' then local pat = '<time class="%s" datetime="%s" lang="%s" dir="%s" style="white-space:nowrap">%s</time>' datestr = pat:format(class, timeStamp, lang, langObj:getDir(), datestr) end return datestr end -- =========================================================================== -- === Versions of the function to be called from template namespace -- =========================================================================== --[[ ======================================================================================== Date This function is the core part of the ISOdate template. Usage: {{#invoke:DateI18n|Date|year=|month=|day=|hour=|minute=|second=|tzhour=|tzmin=|lang=en}} Parameters: * year, month, day, hour, minute, second: broken down date-time component strings * tzhour, tzmin: timezone offset from UTC, hours and minutes * lang: The language to display it in * case: Language format (genitive, etc.) for some languages * class: CSS class for the <time> node, use "" for no metadata at all ]] function p.Date(frame) local args = {} for name, value in pairs( frame.args ) do name = string.gsub( string.lower(name), ' ', '_') args[name] = value end return p._Date( { args.year, args.month, args.day, args.hour, args.minute, args.second, args.tzhour, args.tzmin }, args.lang, -- language args.case, -- allows to specify grammatical case for the month for languages that use them args.class or 'dtstart', -- allows to set the html class of the time node where the date is included. This is useful for microformats. args.trim_year or '100-999' -- by default pad one and 2 digit years to be 4 digit long, while keeping 3 digit years as is ) end return p
__label__pos
0.635854
0 I asked some questions about algorithms, and some replied they are abstract concepts, however I don't know what abstraction means for algorithms or how it applies. For objects, I know if we remove the details of an object we have made it abstract, for example a Person is more abstract than Student I thought maybe they mean removing details of an algorithm and just specifying it with its input and output for example A = Sort(A) is more abstract than bubble sort implementation. Am I correct? • 1 Based on your edit it looks like you have a new question. Ask another question (make a new post), don't just tack it onto this one. Link back to this one if appropriate. – Becuzz Feb 18 '15 at 17:54 1 Don't get afraid by words : Abstraction is the process of removing every aspect of the issue that is not useful to solve it. -So you can concentrate on only what matters- Abstraction is so widely used because there exist a number of 'patterns' in programming that keeps repeating in every application. Find the pattern corresponding to your issue, find the abstract solution to it, implement it and you're done. Even better : most (?all?) coding langages provides some built-in abstract patterns, which are most easy to use. Some API also provides more advanced patterns. Abstracting is removing, so you can reach different level of abstraction by removing more or less aspect of your issue. But if you remove too much elements, you'll say nothing about nothing (expl : A=sort(A) won't solve much issues..). On the other hand, if you keep too much details you might not see what's relevant to solve the issue. Balance. Let's take the sorting issue : It's not relevant what kind of object you want to use, neither the kind of storage you are using for those objects. ( And not either the platform or the programming language you're using). Sorting all breaks down to : - having a collection of object in which you can iterate. - having a way to compare two elements. Then the purpose of the sort is to have every two successive items correctly ordered. Now you can think bubble sort / quick sort / ... for any kind of iterable collection of comparable items. Solve the issue once, and use it any time you encounter this issue, it doesn't matter if you are using a list, an array, a stack.... to store dates, or graphic objects, or ... : the abstract algorithm will be the same. Javascript for instance as a built-in Collection called an Array (it is in fact a dynamic array that can be sparse), which has a sort method. The only thing you have to do to sort the array is to provide the comparison function to the sort method -Great !-. small example : // let's define a collection of information about travels. var travels = [ travel1, travel2, travel3, ... ] ; // sort the travels on their start date // : just provide the right comparison function to the sort method. travels.sort( function ( t1, t2 ) { return t1.startDate - t2.startDate } ); ==> done ! Rq : Abstraction is used for algorithm, but also for general software design : that's what design patterns are about. Just one small example : imagine you have one object very costly to create that you might not even use, what can you do ? (see that this question is very 'abstract' there's no useless details about the object type, the resource being used, ..., there's only the problem's skeleton). Answer is -->>> use the lazy initialization pattern : do not create the object until it's actually asked by your application. class SomeClass { public get someHeavyObject { // create object if it was not created already if (_someHeavyObject == null) _someHeavyObject = new SomeHeavyObject(); // return the object return _someHeavyObject; } private _someHeavyObject = null; } (more on design patterns here http://en.wikipedia.org/wiki/Software_design_pattern ) Another great advantage of abstraction is that when you're seeking help, be it on forums or with a mate, not everyone will want to hear/read the 30 minutes description of what you're building. If you go straight to the point by abstracting the issue, you greatly raise your chances to be understood and to get your reply. You will win a great deal of time by reusing the right abstract data types / algorithm / patterns, so do not hesitate to dig into it. | improve this answer | | 3 No, you're not correct about what that person meant; your reference to objects is a rather technical detail of OO languages (which concerns abstractions the code is modelling), and talking about an algorithm only in terms of input and output is a different level of abstraction, one step too high (but at the same time too low because you seem to think about it again in terms of a programming language). Algorithms are abstract concepts in the sense that the definition of a specific algorithm (let's say Bubblesort) is independent of how it's implemented. You can implement Bubblesort in any programming language you like, Haskell, Java, C, assembler, or even directly in hardware. You can even do it manually. But it will always be Bubblesort, it will always do things in the same way, and always have the characteristics (O(n^2) worst case running time, O(n) on presorted data) of Bubblesort. | improve this answer | | • Thank you, it was my another guess about abstraction in algorithms. but there are other aspects too, for example should we specify the type of items which are sorted in an abstract algorithm!? or just call them comparable objects? – Ahmad Feb 18 '15 at 11:27 • Moreover, you said specifying an algorithm in term of input and output could be a different level of abstraction, when or where (in which field) that type of abstraction is discussed?? OO? – Ahmad Feb 18 '15 at 11:28 • @Ahmad "comparable objects" is exactly the kind of thing that should be part of an abstract description of an algorithm. There are in fact sorting algoritms (e.g. radix sort) where that is not sufficient. Abstracting things in terms of input and output is part of software engineering or (more specifically) requirements engineering. – Michael Borgwardt Feb 18 '15 at 12:00 0 In general an abstraction is a simplification or a simplified representation of something, so to produce an abstraction of an algorithm would be state its form in the simplest way - for a sort this would be state it without the details of how the comparisons or the substitutions worked, for instance. | improve this answer | | Your Answer By clicking “Post Your Answer”, you agree to our terms of service, privacy policy and cookie policy Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.888007
Take the 2-minute tour × Stack Overflow is a question and answer site for professional and enthusiast programmers. It's 100% free, no registration required. This may at first seem like an odd question, but when a cin request is made, if it receives the wrong type it still continues but the status of cin changes. How do I loop until cin is OK, e.g. when I ask for a number it "accepts" a string if no extra code is given e.g. a loop? Finally when I use cin multiple times in a row it does the first cin as expected but then skips the rest; how do I fix this? If you need more information just ask in a comment. // Example cout << "Enter a number: "; cin >> num; // A string is given cout << "Enter another number: "; cin >> num2; In the example above the string would be kinda accepted and the second cin would most likely skip for some reason. I had a while ago find the answer to this question but I lost the snippet of the loop I used so :/ share|improve this question 3   A small snippet of code that almost does what you want would be helpful. –  larsmans Apr 14 '11 at 21:52 1 Answer 1 up vote 2 down vote accepted example: int value; while(!(cin >> value)) { cin.clear(); cin.ignore(); // eat one character } while(!(cin >> value)) { cin.clear(); cin.ignore(10000,'\n'); // eat the rest of the line } share|improve this answer      +1, Also cplusplus.com/reference/iostream/ios/operatornot –  Dacav Oct 13 '11 at 11:52 Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.905063
Mathematics Stack Exchange is a question and answer site for people studying math at any level and professionals in related fields. Join them; it only takes a minute: Sign up Here's how it works: 1. Anybody can ask a question 2. Anybody can answer 3. The best answers are voted up and rise to the top (The integers modulo 3) permit unrestricted subtraction (so that, for example, $1-2=2$), and they permit division restricted only by the exclusion of the denominator 0 (so that, for example, $\frac{1}{2} = 2$). Could someone please help me understand these operations on this finite field? I had a couple of thoughts about subtraction: if the numbers are arranged from left to right in increasing order and I "moved $2$ places to the left for subraction" I would get $2= 1-2$ (as confirmed in the book). This would imply that $0-2 =1$, which seems correct to me since $1+2 =3$... But I wasn't sure if this is the proper way to think about this... Is there a better way? With the division, I don't know: why is $\frac{1}{2} =2$? Thank you for your help. share|cite|improve this question 1   The division is "best" understood via its definition. The number $x=1/2$ is defined as the solution to $2x=1$. (in modular arithmetic this equation doesn't necessarily need to have a unique solution, but for mod $p$ with $p$ prime it does). – Fabian Apr 17 '11 at 7:28      By the way, it is usually best to think about the numbers as being arranged on a circle. – Fabian Apr 17 '11 at 7:29      @Fabian: thanks for the helpful tips – ghshtalt Apr 17 '11 at 10:18 up vote 5 down vote accepted The integers $\mathbb Z$ are a ring: That means it has addition, subtraction, multiplication and some axioms about them. By $3 \mathbb Z$ I denote $\{ 3x \in \mathbb Z \mid x \in \mathbb Z \} = \{\cdots,-6,-3,0,3,6,\cdots\}$. The idea of modular arithmetic (mod 3) is that -6 = -3 = 0 = 3 = 6 = ... and ... = -5 = -2 = 1 = 4 = 6 = ... and so on. The first step now is to make an equivalence relation $\sim$ that expresses this (i.e. $0\sim 3$, $2 \sim 8$, $1 \not \sim 5$) and this is quite easy! Define $x \sim y :\!\!\iff x + 3\mathbb Z = y + 3\mathbb Z$. Since all we have done is applied the function $\varphi(x) = x + 3\mathbb Z$ to both sides this is automatically an equivalence relation. We can see that it is the one we want as well: • $0\sim 3 \iff 0 + 3\mathbb Z = 3 + 3\mathbb Z \iff \{\cdots,-6,-3,0,3,6,\cdots\} = \{\cdots,-3,0,3,6,9,\cdots\} \iff \text{true}$. • $2\sim 8 \iff 2 + 3\mathbb Z = 8 + 3\mathbb Z \iff \{\cdots,-4,-1,2,5,8,\cdots\} = \{\cdots,2,5,8,11,14,\cdots\} \iff \text{true}$. • $1\not\sim 5 \iff 1 + 3\mathbb Z \not = 5 + 3\mathbb Z \iff \{\cdots,-5,-2,1,4,7,\cdots\} \not = \{\cdots,-1,2,5,8,14,\cdots\} \iff \text{true}$. We can now define arithmetic operations on the image $\varphi(\mathbb Z) = \mathbb Z / 3 \mathbb Z$. • $\varphi(a)+\varphi(b):=\varphi(a+b)$ • $-\varphi(a):=\varphi(-a)$ • $\varphi(a)\cdot \varphi(b):=\varphi(a\cdot b)$ To see that e.g. + is actually a function it is necessary to prove that it "respects the equivalence relation" in the sense that if $\varphi(x) = \varphi(x')$ and $\varphi(y) = \varphi(y')$ then $\varphi(x) + \varphi(y) = \varphi(x') + \varphi(y')$. Here is a proof: • $(x + 3 \mathbb Z) + (y + 3 \mathbb Z) = \{\cdots,x-6,x-3,x,x+3,x+6,\cdots\}+ \{\cdots,y-6,y-3,y,y+3,y+6,\cdots\} = \{x+y+i+j\in \mathbb Z | i \in 3 \mathbb Z, j \in 3 \mathbb Z\} = (x + y) + 3 \mathbb Z$. The same type of calculation proves that negation and multiplication are respectful functions. Since the function is respectful it respects each of the ring axioms, this proves that $\mathbb Z/3 \mathbb Z$ is a ring and $\varphi$ is a ring homomorphism. It should be clear that nothing depends on special properties of the number 3 so far and the arguments above are fully general. The standard notation for working in this ring is not $\varphi(x) = \varphi(y)$ but $x \equiv y \pmod 3$ where $x$ is implicitly mapped from $\mathbb Z$ to $\mathbb Z / 3 \mathbb Z$ if needed. The fact that it is furthermore a field is quite miraculous and depends the fact that 3 is a prime number. For $p$ prime every nonzero element of $\mathbb Z/p \mathbb Z$ is invertible. The proof of this depends on details from number theory rather than algebra. First the condition for a number $x$ to be invertible (in any ring) is that there exists some number $x^{-1}$ such that $x \cdot x^{-1} = 1$. In the ring of rationals $\mathbb Q$ this number is $\frac{1}{x}$ (the rationals are also a field because $1 \not = 0$ and every nonzero element is invertible). Given $(a,b)=1$, that is, $a$,$b$ coprime there exists $x$,$y$ such that $ax + by = 1$. You can compute this by the Euclidean algorithm. In terms of modular arithmetic this tells us that given $(a,b) = 1$ then there exists $x$ such that $ax \equiv 1 \pmod b$! Of course when "b" is prime every element except 0 is coprime and thus has an inverse. Since $1 \not \equiv 0 \pmod p$ this proves that $\mathbb Z/p \mathbb Z$ is a field too. Now we can compute $2^{-1} \pmod 3$, it must be the unique number $\varphi(x)$ such that $2x + 3y = 1$, $x = 2, y = -1$ will do so we have $2^{-1} \equiv 2 \pmod 3$. share|cite|improve this answer      @quanta: thank you for the answer. I'm still digesting it, but I wanted to check if I am understanding part of the idea. If I wanted to divide on say integers modulo 5 (where there are a couple more examples) then for say $\frac{3}{4}=x$ I would first always calculate $4^{-1}$ and then multiply? And because of the gcd stuff you showed, I know that $1\equiv 4x \mod 5 \Rightarrow x = 4 \Rightarrow \frac{3}{4} = 2$? So in these cases it is about finding the inverses first? – ghshtalt Apr 17 '11 at 9:44      @quanta: ok, yeah I meant by that 3 divided by 4 in integers modulo 5 (and I realize the way I used $x$ in my above comment was nonsense). But isn't that how you would carry out division of 3 by 4 in integers modulo 5? Since $4$ is the inverse of $4$, $3*4^{-1}=2$ in integers modulo 5, no? – ghshtalt Apr 17 '11 at 9:53      In general modular arithmetic the use of fractions is a mistake but even though prime modulus gives a field there is not homomorphism from $\mathbb Q$ to $\mathbb Z/p \mathbb Z$ so every fraction is not meaningful (consider $1/2 \pmod 2$). Overall I don't think the fraction is a useful concept in modular arithmetic. – quanta Apr 17 '11 at 9:55      "3 divided by 4" or "3/4" could be used to mean $3 \cdot 4^{-1}$ or anything you like. I find it easier to avoid completely but of course just do it whatever way you like. – quanta Apr 17 '11 at 9:56      @quanta: I'm sorry if I'm just completely missing something here, but are you saying there is a problem with the way I am trying to write down/express an idea, or with the idea itself that I am trying to carry out division on the integers modulo 5? – ghshtalt Apr 17 '11 at 10:00 Do you understand multiplication in the field? If so, then division's easy; $a/b=c$ means $a=bc$. In particular, $1/2=2$ because $2\times2=1$. If you don't understand multiplication in the field, just let me know, I'll try something else. Oh, and if you understand addition, then you understand subtraction, because $a-b=c$ means $a=b+c$. share|cite|improve this answer      Thank you for this answer – ghshtalt Apr 17 '11 at 10:19 Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.989425
FFmpeg mjpegdec.c Go to the documentation of this file. 1 /* 2  * MJPEG decoder 3  * Copyright (c) 2000, 2001 Fabrice Bellard 4  * Copyright (c) 2003 Alex Beregszaszi 5  * Copyright (c) 2003-2004 Michael Niedermayer 6  * 7  * Support for external huffman table, various fixes (AVID workaround), 8  * aspecting, new decode_frame mechanism and apple mjpeg-b support 9  * by Alex Beregszaszi 10  * 11  * This file is part of FFmpeg. 12  * 13  * FFmpeg is free software; you can redistribute it and/or 14  * modify it under the terms of the GNU Lesser General Public 15  * License as published by the Free Software Foundation; either 16  * version 2.1 of the License, or (at your option) any later version. 17  * 18  * FFmpeg is distributed in the hope that it will be useful, 19  * but WITHOUT ANY WARRANTY; without even the implied warranty of 20  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 21  * Lesser General Public License for more details. 22  * 23  * You should have received a copy of the GNU Lesser General Public 24  * License along with FFmpeg; if not, write to the Free Software 25  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA 26  */ 27  28 /** 29  * @file 30  * MJPEG decoder. 31  */ 32  33 #include "libavutil/display.h" 34 #include "libavutil/imgutils.h" 35 #include "libavutil/avassert.h" 36 #include "libavutil/opt.h" 37 #include "avcodec.h" 38 #include "blockdsp.h" 39 #include "copy_block.h" 40 #include "decode.h" 41 #include "hwconfig.h" 42 #include "idctdsp.h" 43 #include "internal.h" 44 #include "jpegtables.h" 45 #include "mjpeg.h" 46 #include "mjpegdec.h" 47 #include "jpeglsdec.h" 48 #include "profiles.h" 49 #include "put_bits.h" 50 #include "tiff.h" 51 #include "exif.h" 52 #include "bytestream.h" 53  54  56 { 57  static const struct { 58  int class; 59  int index; 60  const uint8_t *bits; 61  const uint8_t *values; 62  int length; 63  } ht[] = { 65  ff_mjpeg_val_dc, 12 }, 67  ff_mjpeg_val_dc, 12 }, 76  }; 77  int i, ret; 78  79  for (i = 0; i < FF_ARRAY_ELEMS(ht); i++) { 80  ff_free_vlc(&s->vlcs[ht[i].class][ht[i].index]); 81  ret = ff_mjpeg_build_vlc(&s->vlcs[ht[i].class][ht[i].index], 82  ht[i].bits, ht[i].values, 83  ht[i].class == 1, s->avctx); 84  if (ret < 0) 85  return ret; 86  87  if (ht[i].class < 2) { 88  memcpy(s->raw_huffman_lengths[ht[i].class][ht[i].index], 89  ht[i].bits + 1, 16); 90  memcpy(s->raw_huffman_values[ht[i].class][ht[i].index], 91  ht[i].values, ht[i].length); 92  } 93  } 94  95  return 0; 96 } 97  98 static void parse_avid(MJpegDecodeContext *s, uint8_t *buf, int len) 99 { 100  s->buggy_avid = 1; 101  if (len > 14 && buf[12] == 1) /* 1 - NTSC */ 102  s->interlace_polarity = 1; 103  if (len > 14 && buf[12] == 2) /* 2 - PAL */ 104  s->interlace_polarity = 0; 105  if (s->avctx->debug & FF_DEBUG_PICT_INFO) 106  av_log(s->avctx, AV_LOG_INFO, "AVID: len:%d %d\n", len, len > 14 ? buf[12] : -1); 107 } 108  109 static void init_idct(AVCodecContext *avctx) 110 { 111  MJpegDecodeContext *s = avctx->priv_data; 112  113  ff_idctdsp_init(&s->idsp, avctx); 114  ff_init_scantable(s->idsp.idct_permutation, &s->scantable, 116 } 117  119 { 120  MJpegDecodeContext *s = avctx->priv_data; 121  int ret; 122  123  if (!s->picture_ptr) { 124  s->picture = av_frame_alloc(); 125  if (!s->picture) 126  return AVERROR(ENOMEM); 127  s->picture_ptr = s->picture; 128  } 129  130  s->pkt = avctx->internal->in_pkt; 131  132  s->avctx = avctx; 133  ff_blockdsp_init(&s->bdsp, avctx); 134  ff_hpeldsp_init(&s->hdsp, avctx->flags); 135  init_idct(avctx); 136  s->buffer_size = 0; 137  s->buffer = NULL; 138  s->start_code = -1; 139  s->first_picture = 1; 140  s->got_picture = 0; 141  s->orig_height = avctx->coded_height; 143  avctx->colorspace = AVCOL_SPC_BT470BG; 144  s->hwaccel_pix_fmt = s->hwaccel_sw_pix_fmt = AV_PIX_FMT_NONE; 145  146  if ((ret = init_default_huffman_tables(s)) < 0) 147  return ret; 148  149  if (s->extern_huff) { 150  av_log(avctx, AV_LOG_INFO, "using external huffman table\n"); 151  if ((ret = init_get_bits(&s->gb, avctx->extradata, avctx->extradata_size * 8)) < 0) 152  return ret; 153  if (ff_mjpeg_decode_dht(s)) { 154  av_log(avctx, AV_LOG_ERROR, 155  "error using external huffman table, switching back to internal\n"); 156  if ((ret = init_default_huffman_tables(s)) < 0) 157  return ret; 158  } 159  } 160  if (avctx->field_order == AV_FIELD_BB) { /* quicktime icefloe 019 */ 161  s->interlace_polarity = 1; /* bottom field first */ 162  av_log(avctx, AV_LOG_DEBUG, "bottom field first\n"); 163  } else if (avctx->field_order == AV_FIELD_UNKNOWN) { 164  if (avctx->codec_tag == AV_RL32("MJPG")) 165  s->interlace_polarity = 1; 166  } 167  168  if (avctx->codec_id == AV_CODEC_ID_SMVJPEG) { 169  if (avctx->extradata_size >= 4) 170  s->smv_frames_per_jpeg = AV_RL32(avctx->extradata); 171  172  if (s->smv_frames_per_jpeg <= 0) { 173  av_log(avctx, AV_LOG_ERROR, "Invalid number of frames per jpeg.\n"); 174  return AVERROR_INVALIDDATA; 175  } 176  177  s->smv_frame = av_frame_alloc(); 178  if (!s->smv_frame) 179  return AVERROR(ENOMEM); 180  } else if (avctx->extradata_size > 8 181  && AV_RL32(avctx->extradata) == 0x2C 182  && AV_RL32(avctx->extradata+4) == 0x18) { 183  parse_avid(s, avctx->extradata, avctx->extradata_size); 184  } 185  186  if (avctx->codec->id == AV_CODEC_ID_AMV) 187  s->flipped = 1; 188  189  return 0; 190 } 191  192  193 /* quantize tables */ 195 { 196  int len, index, i; 197  198  len = get_bits(&s->gb, 16) - 2; 199  200  if (8*len > get_bits_left(&s->gb)) { 201  av_log(s->avctx, AV_LOG_ERROR, "dqt: len %d is too large\n", len); 202  return AVERROR_INVALIDDATA; 203  } 204  205  while (len >= 65) { 206  int pr = get_bits(&s->gb, 4); 207  if (pr > 1) { 208  av_log(s->avctx, AV_LOG_ERROR, "dqt: invalid precision\n"); 209  return AVERROR_INVALIDDATA; 210  } 211  index = get_bits(&s->gb, 4); 212  if (index >= 4) 213  return -1; 214  av_log(s->avctx, AV_LOG_DEBUG, "index=%d\n", index); 215  /* read quant table */ 216  for (i = 0; i < 64; i++) { 217  s->quant_matrixes[index][i] = get_bits(&s->gb, pr ? 16 : 8); 218  if (s->quant_matrixes[index][i] == 0) { 219  int log_level = s->avctx->err_recognition & AV_EF_EXPLODE ? AV_LOG_ERROR : AV_LOG_WARNING; 220  av_log(s->avctx, log_level, "dqt: 0 quant value\n"); 221  if (s->avctx->err_recognition & AV_EF_EXPLODE) 222  return AVERROR_INVALIDDATA; 223  } 224  } 225  226  // XXX FIXME fine-tune, and perhaps add dc too 227  s->qscale[index] = FFMAX(s->quant_matrixes[index][1], 228  s->quant_matrixes[index][8]) >> 1; 229  av_log(s->avctx, AV_LOG_DEBUG, "qscale[%d]: %d\n", 230  index, s->qscale[index]); 231  len -= 1 + 64 * (1+pr); 232  } 233  return 0; 234 } 235  236 /* decode huffman tables and build VLC decoders */ 238 { 239  int len, index, i, class, n, v; 240  uint8_t bits_table[17]; 241  uint8_t val_table[256]; 242  int ret = 0; 243  244  len = get_bits(&s->gb, 16) - 2; 245  246  if (8*len > get_bits_left(&s->gb)) { 247  av_log(s->avctx, AV_LOG_ERROR, "dht: len %d is too large\n", len); 248  return AVERROR_INVALIDDATA; 249  } 250  251  while (len > 0) { 252  if (len < 17) 253  return AVERROR_INVALIDDATA; 254  class = get_bits(&s->gb, 4); 255  if (class >= 2) 256  return AVERROR_INVALIDDATA; 257  index = get_bits(&s->gb, 4); 258  if (index >= 4) 259  return AVERROR_INVALIDDATA; 260  n = 0; 261  for (i = 1; i <= 16; i++) { 262  bits_table[i] = get_bits(&s->gb, 8); 263  n += bits_table[i]; 264  } 265  len -= 17; 266  if (len < n || n > 256) 267  return AVERROR_INVALIDDATA; 268  269  for (i = 0; i < n; i++) { 270  v = get_bits(&s->gb, 8); 271  val_table[i] = v; 272  } 273  len -= n; 274  275  /* build VLC and flush previous vlc if present */ 276  ff_free_vlc(&s->vlcs[class][index]); 277  av_log(s->avctx, AV_LOG_DEBUG, "class=%d index=%d nb_codes=%d\n", 278  class, index, n); 279  if ((ret = ff_mjpeg_build_vlc(&s->vlcs[class][index], bits_table, 280  val_table, class > 0, s->avctx)) < 0) 281  return ret; 282  283  if (class > 0) { 284  ff_free_vlc(&s->vlcs[2][index]); 285  if ((ret = ff_mjpeg_build_vlc(&s->vlcs[2][index], bits_table, 286  val_table, 0, s->avctx)) < 0) 287  return ret; 288  } 289  290  for (i = 0; i < 16; i++) 291  s->raw_huffman_lengths[class][index][i] = bits_table[i + 1]; 292  for (i = 0; i < 256; i++) 293  s->raw_huffman_values[class][index][i] = val_table[i]; 294  } 295  return 0; 296 } 297  299 { 300  int len, nb_components, i, width, height, bits, ret, size_change; 301  unsigned pix_fmt_id; 302  int h_count[MAX_COMPONENTS] = { 0 }; 303  int v_count[MAX_COMPONENTS] = { 0 }; 304  305  s->cur_scan = 0; 306  memset(s->upscale_h, 0, sizeof(s->upscale_h)); 307  memset(s->upscale_v, 0, sizeof(s->upscale_v)); 308  309  len = get_bits(&s->gb, 16); 310  bits = get_bits(&s->gb, 8); 311  312  if (bits > 16 || bits < 1) { 313  av_log(s->avctx, AV_LOG_ERROR, "bits %d is invalid\n", bits); 314  return AVERROR_INVALIDDATA; 315  } 316  317  if (s->avctx->bits_per_raw_sample != bits) { 318  av_log(s->avctx, s->avctx->bits_per_raw_sample > 0 ? AV_LOG_INFO : AV_LOG_DEBUG, "Changing bps from %d to %d\n", s->avctx->bits_per_raw_sample, bits); 319  s->avctx->bits_per_raw_sample = bits; 320  init_idct(s->avctx); 321  } 322  if (s->pegasus_rct) 323  bits = 9; 324  if (bits == 9 && !s->pegasus_rct) 325  s->rct = 1; // FIXME ugly 326  327  if(s->lossless && s->avctx->lowres){ 328  av_log(s->avctx, AV_LOG_ERROR, "lowres is not possible with lossless jpeg\n"); 329  return -1; 330  } 331  332  height = get_bits(&s->gb, 16); 333  width = get_bits(&s->gb, 16); 334  335  // HACK for odd_height.mov 336  if (s->interlaced && s->width == width && s->height == height + 1) 337  height= s->height; 338  339  av_log(s->avctx, AV_LOG_DEBUG, "sof0: picture: %dx%d\n", width, height); 340  if (av_image_check_size(width, height, 0, s->avctx) < 0) 341  return AVERROR_INVALIDDATA; 342  if (s->buf_size && (width + 7) / 8 * ((height + 7) / 8) > s->buf_size * 4LL) 343  return AVERROR_INVALIDDATA; 344  345  nb_components = get_bits(&s->gb, 8); 346  if (nb_components <= 0 || 347  nb_components > MAX_COMPONENTS) 348  return -1; 349  if (s->interlaced && (s->bottom_field == !s->interlace_polarity)) { 350  if (nb_components != s->nb_components) { 351  av_log(s->avctx, AV_LOG_ERROR, 352  "nb_components changing in interlaced picture\n"); 353  return AVERROR_INVALIDDATA; 354  } 355  } 356  if (s->ls && !(bits <= 8 || nb_components == 1)) { 358  "JPEG-LS that is not <= 8 " 359  "bits/component or 16-bit gray"); 360  return AVERROR_PATCHWELCOME; 361  } 362  if (len != 8 + 3 * nb_components) { 363  av_log(s->avctx, AV_LOG_ERROR, "decode_sof0: error, len(%d) mismatch %d components\n", len, nb_components); 364  return AVERROR_INVALIDDATA; 365  } 366  367  s->nb_components = nb_components; 368  s->h_max = 1; 369  s->v_max = 1; 370  for (i = 0; i < nb_components; i++) { 371  /* component id */ 372  s->component_id[i] = get_bits(&s->gb, 8) - 1; 373  h_count[i] = get_bits(&s->gb, 4); 374  v_count[i] = get_bits(&s->gb, 4); 375  /* compute hmax and vmax (only used in interleaved case) */ 376  if (h_count[i] > s->h_max) 377  s->h_max = h_count[i]; 378  if (v_count[i] > s->v_max) 379  s->v_max = v_count[i]; 380  s->quant_index[i] = get_bits(&s->gb, 8); 381  if (s->quant_index[i] >= 4) { 382  av_log(s->avctx, AV_LOG_ERROR, "quant_index is invalid\n"); 383  return AVERROR_INVALIDDATA; 384  } 385  if (!h_count[i] || !v_count[i]) { 386  av_log(s->avctx, AV_LOG_ERROR, 387  "Invalid sampling factor in component %d %d:%d\n", 388  i, h_count[i], v_count[i]); 389  return AVERROR_INVALIDDATA; 390  } 391  392  av_log(s->avctx, AV_LOG_DEBUG, "component %d %d:%d id: %d quant:%d\n", 393  i, h_count[i], v_count[i], 394  s->component_id[i], s->quant_index[i]); 395  } 396  if ( nb_components == 4 397  && s->component_id[0] == 'C' - 1 398  && s->component_id[1] == 'M' - 1 399  && s->component_id[2] == 'Y' - 1 400  && s->component_id[3] == 'K' - 1) 401  s->adobe_transform = 0; 402  403  if (s->ls && (s->h_max > 1 || s->v_max > 1)) { 404  avpriv_report_missing_feature(s->avctx, "Subsampling in JPEG-LS"); 405  return AVERROR_PATCHWELCOME; 406  } 407  408  if (s->bayer) { 409  if (nb_components == 2) { 410  /* Bayer images embedded in DNGs can contain 2 interleaved components and the 411  width stored in their SOF3 markers is the width of each one. We only output 412  a single component, therefore we need to adjust the output image width. We 413  handle the deinterleaving (but not the debayering) in this file. */ 414  width *= 2; 415  } 416  /* They can also contain 1 component, which is double the width and half the height 417  of the final image (rows are interleaved). We don't handle the decoding in this 418  file, but leave that to the TIFF/DNG decoder. */ 419  } 420  421  /* if different size, realloc/alloc picture */ 422  if (width != s->width || height != s->height || bits != s->bits || 423  memcmp(s->h_count, h_count, sizeof(h_count)) || 424  memcmp(s->v_count, v_count, sizeof(v_count))) { 425  size_change = 1; 426  427  s->width = width; 428  s->height = height; 429  s->bits = bits; 430  memcpy(s->h_count, h_count, sizeof(h_count)); 431  memcpy(s->v_count, v_count, sizeof(v_count)); 432  s->interlaced = 0; 433  s->got_picture = 0; 434  435  /* test interlaced mode */ 436  if (s->first_picture && 437  (s->multiscope != 2 || s->avctx->time_base.den >= 25 * s->avctx->time_base.num) && 438  s->orig_height != 0 && 439  s->height < ((s->orig_height * 3) / 4)) { 440  s->interlaced = 1; 441  s->bottom_field = s->interlace_polarity; 442  s->picture_ptr->interlaced_frame = 1; 443  s->picture_ptr->top_field_first = !s->interlace_polarity; 444  height *= 2; 445  } 446  447  ret = ff_set_dimensions(s->avctx, width, height); 448  if (ret < 0) 449  return ret; 450  451  if ((s->avctx->codec_tag == MKTAG('A', 'V', 'R', 'n') || 452  s->avctx->codec_tag == MKTAG('A', 'V', 'D', 'J')) && 453  s->orig_height < height) 454  s->avctx->height = AV_CEIL_RSHIFT(s->orig_height, s->avctx->lowres); 455  456  s->first_picture = 0; 457  } else { 458  size_change = 0; 459  } 460  461  if (s->avctx->codec_id == AV_CODEC_ID_SMVJPEG) { 462  s->avctx->height = s->avctx->coded_height / s->smv_frames_per_jpeg; 463  if (s->avctx->height <= 0) 464  return AVERROR_INVALIDDATA; 465  } 466  467  if (s->got_picture && s->interlaced && (s->bottom_field == !s->interlace_polarity)) { 468  if (s->progressive) { 469  avpriv_request_sample(s->avctx, "progressively coded interlaced picture"); 470  return AVERROR_INVALIDDATA; 471  } 472  } else { 473  if (s->v_max == 1 && s->h_max == 1 && s->lossless==1 && (nb_components==3 || nb_components==4)) 474  s->rgb = 1; 475  else if (!s->lossless) 476  s->rgb = 0; 477  /* XXX: not complete test ! */ 478  pix_fmt_id = ((unsigned)s->h_count[0] << 28) | (s->v_count[0] << 24) | 479  (s->h_count[1] << 20) | (s->v_count[1] << 16) | 480  (s->h_count[2] << 12) | (s->v_count[2] << 8) | 481  (s->h_count[3] << 4) | s->v_count[3]; 482  av_log(s->avctx, AV_LOG_DEBUG, "pix fmt id %x\n", pix_fmt_id); 483  /* NOTE we do not allocate pictures large enough for the possible 484  * padding of h/v_count being 4 */ 485  if (!(pix_fmt_id & 0xD0D0D0D0)) 486  pix_fmt_id -= (pix_fmt_id & 0xF0F0F0F0) >> 1; 487  if (!(pix_fmt_id & 0x0D0D0D0D)) 488  pix_fmt_id -= (pix_fmt_id & 0x0F0F0F0F) >> 1; 489  490  for (i = 0; i < 8; i++) { 491  int j = 6 + (i&1) - (i&6); 492  int is = (pix_fmt_id >> (4*i)) & 0xF; 493  int js = (pix_fmt_id >> (4*j)) & 0xF; 494  495  if (is == 1 && js != 2 && (i < 2 || i > 5)) 496  js = (pix_fmt_id >> ( 8 + 4*(i&1))) & 0xF; 497  if (is == 1 && js != 2 && (i < 2 || i > 5)) 498  js = (pix_fmt_id >> (16 + 4*(i&1))) & 0xF; 499  500  if (is == 1 && js == 2) { 501  if (i & 1) s->upscale_h[j/2] = 1; 502  else s->upscale_v[j/2] = 1; 503  } 504  } 505  506  if (s->bayer) { 507  if (pix_fmt_id != 0x11110000 && pix_fmt_id != 0x11000000) 508  goto unk_pixfmt; 509  } 510  511  switch (pix_fmt_id) { 512  case 0x11110000: /* for bayer-encoded huffman lossless JPEGs embedded in DNGs */ 513  if (!s->bayer) 514  goto unk_pixfmt; 515  s->avctx->pix_fmt = AV_PIX_FMT_GRAY16LE; 516  break; 517  case 0x11111100: 518  if (s->rgb) 519  s->avctx->pix_fmt = s->bits <= 9 ? AV_PIX_FMT_BGR24 : AV_PIX_FMT_BGR48; 520  else { 521  if ( s->adobe_transform == 0 522  || s->component_id[0] == 'R' - 1 && s->component_id[1] == 'G' - 1 && s->component_id[2] == 'B' - 1) { 523  s->avctx->pix_fmt = s->bits <= 8 ? AV_PIX_FMT_GBRP : AV_PIX_FMT_GBRP16; 524  } else { 525  if (s->bits <= 8) s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV444P : AV_PIX_FMT_YUVJ444P; 526  else s->avctx->pix_fmt = AV_PIX_FMT_YUV444P16; 527  s->avctx->color_range = s->cs_itu601 ? AVCOL_RANGE_MPEG : AVCOL_RANGE_JPEG; 528  } 529  } 530  av_assert0(s->nb_components == 3); 531  break; 532  case 0x11111111: 533  if (s->rgb) 534  s->avctx->pix_fmt = s->bits <= 9 ? AV_PIX_FMT_ABGR : AV_PIX_FMT_RGBA64; 535  else { 536  if (s->adobe_transform == 0 && s->bits <= 8) { 537  s->avctx->pix_fmt = AV_PIX_FMT_GBRAP; 538  } else { 539  s->avctx->pix_fmt = s->bits <= 8 ? AV_PIX_FMT_YUVA444P : AV_PIX_FMT_YUVA444P16; 540  s->avctx->color_range = s->cs_itu601 ? AVCOL_RANGE_MPEG : AVCOL_RANGE_JPEG; 541  } 542  } 543  av_assert0(s->nb_components == 4); 544  break; 545  case 0x22111122: 546  case 0x22111111: 547  if (s->adobe_transform == 0 && s->bits <= 8) { 548  s->avctx->pix_fmt = AV_PIX_FMT_GBRAP; 549  s->upscale_v[1] = s->upscale_v[2] = 1; 550  s->upscale_h[1] = s->upscale_h[2] = 1; 551  } else if (s->adobe_transform == 2 && s->bits <= 8) { 552  s->avctx->pix_fmt = AV_PIX_FMT_YUVA444P; 553  s->upscale_v[1] = s->upscale_v[2] = 1; 554  s->upscale_h[1] = s->upscale_h[2] = 1; 555  s->avctx->color_range = s->cs_itu601 ? AVCOL_RANGE_MPEG : AVCOL_RANGE_JPEG; 556  } else { 557  if (s->bits <= 8) s->avctx->pix_fmt = AV_PIX_FMT_YUVA420P; 558  else s->avctx->pix_fmt = AV_PIX_FMT_YUVA420P16; 559  s->avctx->color_range = s->cs_itu601 ? AVCOL_RANGE_MPEG : AVCOL_RANGE_JPEG; 560  } 561  av_assert0(s->nb_components == 4); 562  break; 563  case 0x12121100: 564  case 0x22122100: 565  case 0x21211100: 566  case 0x21112100: 567  case 0x22211200: 568  case 0x22221100: 569  case 0x22112200: 570  case 0x11222200: 571  if (s->bits <= 8) s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV444P : AV_PIX_FMT_YUVJ444P; 572  else 573  goto unk_pixfmt; 574  s->avctx->color_range = s->cs_itu601 ? AVCOL_RANGE_MPEG : AVCOL_RANGE_JPEG; 575  break; 576  case 0x11000000: 577  case 0x13000000: 578  case 0x14000000: 579  case 0x31000000: 580  case 0x33000000: 581  case 0x34000000: 582  case 0x41000000: 583  case 0x43000000: 584  case 0x44000000: 585  if(s->bits <= 8) 586  s->avctx->pix_fmt = s->force_pal8 ? AV_PIX_FMT_PAL8 : AV_PIX_FMT_GRAY8; 587  else 588  s->avctx->pix_fmt = AV_PIX_FMT_GRAY16; 589  break; 590  case 0x12111100: 591  case 0x14121200: 592  case 0x14111100: 593  case 0x22211100: 594  case 0x22112100: 595  if (s->component_id[0] == 'Q' && s->component_id[1] == 'F' && s->component_id[2] == 'A') { 596  if (s->bits <= 8) s->avctx->pix_fmt = AV_PIX_FMT_GBRP; 597  else 598  goto unk_pixfmt; 599  s->upscale_v[0] = s->upscale_v[1] = 1; 600  } else { 601  if (pix_fmt_id == 0x14111100) 602  s->upscale_v[1] = s->upscale_v[2] = 1; 603  if (s->bits <= 8) s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV440P : AV_PIX_FMT_YUVJ440P; 604  else 605  goto unk_pixfmt; 606  s->avctx->color_range = s->cs_itu601 ? AVCOL_RANGE_MPEG : AVCOL_RANGE_JPEG; 607  } 608  break; 609  case 0x21111100: 610  if (s->component_id[0] == 'Q' && s->component_id[1] == 'F' && s->component_id[2] == 'A') { 611  if (s->bits <= 8) s->avctx->pix_fmt = AV_PIX_FMT_GBRP; 612  else 613  goto unk_pixfmt; 614  s->upscale_h[0] = s->upscale_h[1] = 1; 615  } else { 616  if (s->bits <= 8) s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV422P : AV_PIX_FMT_YUVJ422P; 617  else s->avctx->pix_fmt = AV_PIX_FMT_YUV422P16; 618  s->avctx->color_range = s->cs_itu601 ? AVCOL_RANGE_MPEG : AVCOL_RANGE_JPEG; 619  } 620  break; 621  case 0x31111100: 622  if (s->bits > 8) 623  goto unk_pixfmt; 624  s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV444P : AV_PIX_FMT_YUVJ444P; 625  s->avctx->color_range = s->cs_itu601 ? AVCOL_RANGE_MPEG : AVCOL_RANGE_JPEG; 626  s->upscale_h[1] = s->upscale_h[2] = 2; 627  break; 628  case 0x22121100: 629  case 0x22111200: 630  if (s->bits <= 8) s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV422P : AV_PIX_FMT_YUVJ422P; 631  else 632  goto unk_pixfmt; 633  s->avctx->color_range = s->cs_itu601 ? AVCOL_RANGE_MPEG : AVCOL_RANGE_JPEG; 634  break; 635  case 0x22111100: 636  case 0x23111100: 637  case 0x42111100: 638  case 0x24111100: 639  if (s->bits <= 8) s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV420P : AV_PIX_FMT_YUVJ420P; 640  else s->avctx->pix_fmt = AV_PIX_FMT_YUV420P16; 641  s->avctx->color_range = s->cs_itu601 ? AVCOL_RANGE_MPEG : AVCOL_RANGE_JPEG; 642  if (pix_fmt_id == 0x42111100) { 643  if (s->bits > 8) 644  goto unk_pixfmt; 645  s->upscale_h[1] = s->upscale_h[2] = 1; 646  } else if (pix_fmt_id == 0x24111100) { 647  if (s->bits > 8) 648  goto unk_pixfmt; 649  s->upscale_v[1] = s->upscale_v[2] = 1; 650  } else if (pix_fmt_id == 0x23111100) { 651  if (s->bits > 8) 652  goto unk_pixfmt; 653  s->upscale_v[1] = s->upscale_v[2] = 2; 654  } 655  break; 656  case 0x41111100: 657  if (s->bits <= 8) s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV411P : AV_PIX_FMT_YUVJ411P; 658  else 659  goto unk_pixfmt; 660  s->avctx->color_range = s->cs_itu601 ? AVCOL_RANGE_MPEG : AVCOL_RANGE_JPEG; 661  break; 662  default: 663  unk_pixfmt: 664  avpriv_report_missing_feature(s->avctx, "Pixel format 0x%x bits:%d", pix_fmt_id, s->bits); 665  memset(s->upscale_h, 0, sizeof(s->upscale_h)); 666  memset(s->upscale_v, 0, sizeof(s->upscale_v)); 667  return AVERROR_PATCHWELCOME; 668  } 669  if ((AV_RB32(s->upscale_h) || AV_RB32(s->upscale_v)) && s->avctx->lowres) { 670  avpriv_report_missing_feature(s->avctx, "Lowres for weird subsampling"); 671  return AVERROR_PATCHWELCOME; 672  } 673  if ((AV_RB32(s->upscale_h) || AV_RB32(s->upscale_v)) && s->progressive && s->avctx->pix_fmt == AV_PIX_FMT_GBRP) { 674  avpriv_report_missing_feature(s->avctx, "progressive for weird subsampling"); 675  return AVERROR_PATCHWELCOME; 676  } 677  if (s->ls) { 678  memset(s->upscale_h, 0, sizeof(s->upscale_h)); 679  memset(s->upscale_v, 0, sizeof(s->upscale_v)); 680  if (s->nb_components == 3) { 681  s->avctx->pix_fmt = AV_PIX_FMT_RGB24; 682  } else if (s->nb_components != 1) { 683  av_log(s->avctx, AV_LOG_ERROR, "Unsupported number of components %d\n", s->nb_components); 684  return AVERROR_PATCHWELCOME; 685  } else if ((s->palette_index || s->force_pal8) && s->bits <= 8) 686  s->avctx->pix_fmt = AV_PIX_FMT_PAL8; 687  else if (s->bits <= 8) 688  s->avctx->pix_fmt = AV_PIX_FMT_GRAY8; 689  else 690  s->avctx->pix_fmt = AV_PIX_FMT_GRAY16; 691  } 692  693  s->pix_desc = av_pix_fmt_desc_get(s->avctx->pix_fmt); 694  if (!s->pix_desc) { 695  av_log(s->avctx, AV_LOG_ERROR, "Could not get a pixel format descriptor.\n"); 696  return AVERROR_BUG; 697  } 698  699  if (s->avctx->pix_fmt == s->hwaccel_sw_pix_fmt && !size_change) { 700  s->avctx->pix_fmt = s->hwaccel_pix_fmt; 701  } else { 702  enum AVPixelFormat pix_fmts[] = { 703 #if CONFIG_MJPEG_NVDEC_HWACCEL 705 #endif 706 #if CONFIG_MJPEG_VAAPI_HWACCEL 708 #endif 709  s->avctx->pix_fmt, 711  }; 712  s->hwaccel_pix_fmt = ff_get_format(s->avctx, pix_fmts); 713  if (s->hwaccel_pix_fmt < 0) 714  return AVERROR(EINVAL); 715  716  s->hwaccel_sw_pix_fmt = s->avctx->pix_fmt; 717  s->avctx->pix_fmt = s->hwaccel_pix_fmt; 718  } 719  720  if (s->avctx->skip_frame == AVDISCARD_ALL) { 721  s->picture_ptr->pict_type = AV_PICTURE_TYPE_I; 722  s->picture_ptr->key_frame = 1; 723  s->got_picture = 1; 724  return 0; 725  } 726  727  av_frame_unref(s->picture_ptr); 728  if (ff_get_buffer(s->avctx, s->picture_ptr, AV_GET_BUFFER_FLAG_REF) < 0) 729  return -1; 730  s->picture_ptr->pict_type = AV_PICTURE_TYPE_I; 731  s->picture_ptr->key_frame = 1; 732  s->got_picture = 1; 733  734  // Lets clear the palette to avoid leaving uninitialized values in it 735  if (s->avctx->pix_fmt == AV_PIX_FMT_PAL8) 736  memset(s->picture_ptr->data[1], 0, 1024); 737  738  for (i = 0; i < 4; i++) 739  s->linesize[i] = s->picture_ptr->linesize[i] << s->interlaced; 740  741  ff_dlog(s->avctx, "%d %d %d %d %d %d\n", 742  s->width, s->height, s->linesize[0], s->linesize[1], 743  s->interlaced, s->avctx->height); 744  745  } 746  747  if ((s->rgb && !s->lossless && !s->ls) || 748  (!s->rgb && s->ls && s->nb_components > 1) || 749  (s->avctx->pix_fmt == AV_PIX_FMT_PAL8 && !s->ls) 750  ) { 751  av_log(s->avctx, AV_LOG_ERROR, "Unsupported coding and pixel format combination\n"); 752  return AVERROR_PATCHWELCOME; 753  } 754  755  /* totally blank picture as progressive JPEG will only add details to it */ 756  if (s->progressive) { 757  int bw = (width + s->h_max * 8 - 1) / (s->h_max * 8); 758  int bh = (height + s->v_max * 8 - 1) / (s->v_max * 8); 759  for (i = 0; i < s->nb_components; i++) { 760  int size = bw * bh * s->h_count[i] * s->v_count[i]; 761  av_freep(&s->blocks[i]); 762  av_freep(&s->last_nnz[i]); 763  s->blocks[i] = av_calloc(size, sizeof(**s->blocks)); 764  s->last_nnz[i] = av_calloc(size, sizeof(**s->last_nnz)); 765  if (!s->blocks[i] || !s->last_nnz[i]) 766  return AVERROR(ENOMEM); 767  s->block_stride[i] = bw * s->h_count[i]; 768  } 769  memset(s->coefs_finished, 0, sizeof(s->coefs_finished)); 770  } 771  772  if (s->avctx->hwaccel) { 773  s->hwaccel_picture_private = 774  av_mallocz(s->avctx->hwaccel->frame_priv_data_size); 775  if (!s->hwaccel_picture_private) 776  return AVERROR(ENOMEM); 777  778  ret = s->avctx->hwaccel->start_frame(s->avctx, s->raw_image_buffer, 779  s->raw_image_buffer_size); 780  if (ret < 0) 781  return ret; 782  } 783  784  return 0; 785 } 786  787 static inline int mjpeg_decode_dc(MJpegDecodeContext *s, int dc_index) 788 { 789  int code; 790  code = get_vlc2(&s->gb, s->vlcs[0][dc_index].table, 9, 2); 791  if (code < 0 || code > 16) { 792  av_log(s->avctx, AV_LOG_WARNING, 793  "mjpeg_decode_dc: bad vlc: %d:%d (%p)\n", 794  0, dc_index, &s->vlcs[0][dc_index]); 795  return 0xfffff; 796  } 797  798  if (code) 799  return get_xbits(&s->gb, code); 800  else 801  return 0; 802 } 803  804 /* decode block and dequantize */ 805 static int decode_block(MJpegDecodeContext *s, int16_t *block, int component, 806  int dc_index, int ac_index, uint16_t *quant_matrix) 807 { 808  int code, i, j, level, val; 809  810  /* DC coef */ 811  val = mjpeg_decode_dc(s, dc_index); 812  if (val == 0xfffff) { 813  av_log(s->avctx, AV_LOG_ERROR, "error dc\n"); 814  return AVERROR_INVALIDDATA; 815  } 816  val = val * (unsigned)quant_matrix[0] + s->last_dc[component]; 817  val = av_clip_int16(val); 818  s->last_dc[component] = val; 819  block[0] = val; 820  /* AC coefs */ 821  i = 0; 822  {OPEN_READER(re, &s->gb); 823  do { 824  UPDATE_CACHE(re, &s->gb); 825  GET_VLC(code, re, &s->gb, s->vlcs[1][ac_index].table, 9, 2); 826  827  i += ((unsigned)code) >> 4; 828  code &= 0xf; 829  if (code) { 830  if (code > MIN_CACHE_BITS - 16) 831  UPDATE_CACHE(re, &s->gb); 832  833  { 834  int cache = GET_CACHE(re, &s->gb); 835  int sign = (~cache) >> 31; 836  level = (NEG_USR32(sign ^ cache,code) ^ sign) - sign; 837  } 838  839  LAST_SKIP_BITS(re, &s->gb, code); 840  841  if (i > 63) { 842  av_log(s->avctx, AV_LOG_ERROR, "error count: %d\n", i); 843  return AVERROR_INVALIDDATA; 844  } 845  j = s->scantable.permutated[i]; 846  block[j] = level * quant_matrix[i]; 847  } 848  } while (i < 63); 849  CLOSE_READER(re, &s->gb);} 850  851  return 0; 852 } 853  855  int component, int dc_index, 856  uint16_t *quant_matrix, int Al) 857 { 858  unsigned val; 859  s->bdsp.clear_block(block); 860  val = mjpeg_decode_dc(s, dc_index); 861  if (val == 0xfffff) { 862  av_log(s->avctx, AV_LOG_ERROR, "error dc\n"); 863  return AVERROR_INVALIDDATA; 864  } 865  val = (val * (quant_matrix[0] << Al)) + s->last_dc[component]; 866  s->last_dc[component] = val; 867  block[0] = val; 868  return 0; 869 } 870  871 /* decode block and dequantize - progressive JPEG version */ 873  uint8_t *last_nnz, int ac_index, 874  uint16_t *quant_matrix, 875  int ss, int se, int Al, int *EOBRUN) 876 { 877  int code, i, j, val, run; 878  unsigned level; 879  880  if (*EOBRUN) { 881  (*EOBRUN)--; 882  return 0; 883  } 884  885  { 886  OPEN_READER(re, &s->gb); 887  for (i = ss; ; i++) { 888  UPDATE_CACHE(re, &s->gb); 889  GET_VLC(code, re, &s->gb, s->vlcs[2][ac_index].table, 9, 2); 890  891  run = ((unsigned) code) >> 4; 892  code &= 0xF; 893  if (code) { 894  i += run; 895  if (code > MIN_CACHE_BITS - 16) 896  UPDATE_CACHE(re, &s->gb); 897  898  { 899  int cache = GET_CACHE(re, &s->gb); 900  int sign = (~cache) >> 31; 901  level = (NEG_USR32(sign ^ cache,code) ^ sign) - sign; 902  } 903  904  LAST_SKIP_BITS(re, &s->gb, code); 905  906  if (i >= se) { 907  if (i == se) { 908  j = s->scantable.permutated[se]; 909  block[j] = level * (quant_matrix[se] << Al); 910  break; 911  } 912  av_log(s->avctx, AV_LOG_ERROR, "error count: %d\n", i); 913  return AVERROR_INVALIDDATA; 914  } 915  j = s->scantable.permutated[i]; 916  block[j] = level * (quant_matrix[i] << Al); 917  } else { 918  if (run == 0xF) {// ZRL - skip 15 coefficients 919  i += 15; 920  if (i >= se) { 921  av_log(s->avctx, AV_LOG_ERROR, "ZRL overflow: %d\n", i); 922  return AVERROR_INVALIDDATA; 923  } 924  } else { 925  val = (1 << run); 926  if (run) { 927  UPDATE_CACHE(re, &s->gb); 928  val += NEG_USR32(GET_CACHE(re, &s->gb), run); 929  LAST_SKIP_BITS(re, &s->gb, run); 930  } 931  *EOBRUN = val - 1; 932  break; 933  } 934  } 935  } 936  CLOSE_READER(re, &s->gb); 937  } 938  939  if (i > *last_nnz) 940  *last_nnz = i; 941  942  return 0; 943 } 944  945 #define REFINE_BIT(j) { \ 946  UPDATE_CACHE(re, &s->gb); \ 947  sign = block[j] >> 15; \ 948  block[j] += SHOW_UBITS(re, &s->gb, 1) * \ 949  ((quant_matrix[i] ^ sign) - sign) << Al; \ 950  LAST_SKIP_BITS(re, &s->gb, 1); \ 951 } 952  953 #define ZERO_RUN \ 954 for (; ; i++) { \ 955  if (i > last) { \ 956  i += run; \ 957  if (i > se) { \ 958  av_log(s->avctx, AV_LOG_ERROR, "error count: %d\n", i); \ 959  return -1; \ 960  } \ 961  break; \ 962  } \ 963  j = s->scantable.permutated[i]; \ 964  if (block[j]) \ 965  REFINE_BIT(j) \ 966  else if (run-- == 0) \ 967  break; \ 968 } 969  970 /* decode block and dequantize - progressive JPEG refinement pass */ 972  uint8_t *last_nnz, 973  int ac_index, uint16_t *quant_matrix, 974  int ss, int se, int Al, int *EOBRUN) 975 { 976  int code, i = ss, j, sign, val, run; 977  int last = FFMIN(se, *last_nnz); 978  979  OPEN_READER(re, &s->gb); 980  if (*EOBRUN) { 981  (*EOBRUN)--; 982  } else { 983  for (; ; i++) { 984  UPDATE_CACHE(re, &s->gb); 985  GET_VLC(code, re, &s->gb, s->vlcs[2][ac_index].table, 9, 2); 986  987  if (code & 0xF) { 988  run = ((unsigned) code) >> 4; 989  UPDATE_CACHE(re, &s->gb); 990  val = SHOW_UBITS(re, &s->gb, 1); 991  LAST_SKIP_BITS(re, &s->gb, 1); 992  ZERO_RUN; 993  j = s->scantable.permutated[i]; 994  val--; 995  block[j] = ((quant_matrix[i] << Al) ^ val) - val; 996  if (i == se) { 997  if (i > *last_nnz) 998  *last_nnz = i; 999  CLOSE_READER(re, &s->gb); 1000  return 0; 1001  } 1002  } else { 1003  run = ((unsigned) code) >> 4; 1004  if (run == 0xF) { 1005  ZERO_RUN; 1006  } else { 1007  val = run; 1008  run = (1 << run); 1009  if (val) { 1010  UPDATE_CACHE(re, &s->gb); 1011  run += SHOW_UBITS(re, &s->gb, val); 1012  LAST_SKIP_BITS(re, &s->gb, val); 1013  } 1014  *EOBRUN = run - 1; 1015  break; 1016  } 1017  } 1018  } 1019  1020  if (i > *last_nnz) 1021  *last_nnz = i; 1022  } 1023  1024  for (; i <= last; i++) { 1025  j = s->scantable.permutated[i]; 1026  if (block[j]) 1027  REFINE_BIT(j) 1028  } 1029  CLOSE_READER(re, &s->gb); 1030  1031  return 0; 1032 } 1033 #undef REFINE_BIT 1034 #undef ZERO_RUN 1035  1036 static int handle_rstn(MJpegDecodeContext *s, int nb_components) 1037 { 1038  int i; 1039  int reset = 0; 1040  1041  if (s->restart_interval) { 1042  s->restart_count--; 1043  if(s->restart_count == 0 && s->avctx->codec_id == AV_CODEC_ID_THP){ 1044  align_get_bits(&s->gb); 1045  for (i = 0; i < nb_components; i++) /* reset dc */ 1046  s->last_dc[i] = (4 << s->bits); 1047  } 1048  1049  i = 8 + ((-get_bits_count(&s->gb)) & 7); 1050  /* skip RSTn */ 1051  if (s->restart_count == 0) { 1052  if( show_bits(&s->gb, i) == (1 << i) - 1 1053  || show_bits(&s->gb, i) == 0xFF) { 1054  int pos = get_bits_count(&s->gb); 1055  align_get_bits(&s->gb); 1056  while (get_bits_left(&s->gb) >= 8 && show_bits(&s->gb, 8) == 0xFF) 1057  skip_bits(&s->gb, 8); 1058  if (get_bits_left(&s->gb) >= 8 && (get_bits(&s->gb, 8) & 0xF8) == 0xD0) { 1059  for (i = 0; i < nb_components; i++) /* reset dc */ 1060  s->last_dc[i] = (4 << s->bits); 1061  reset = 1; 1062  } else 1063  skip_bits_long(&s->gb, pos - get_bits_count(&s->gb)); 1064  } 1065  } 1066  } 1067  return reset; 1068 } 1069  1070 /* Handles 1 to 4 components */ 1071 static int ljpeg_decode_rgb_scan(MJpegDecodeContext *s, int nb_components, int predictor, int point_transform) 1072 { 1073  int i, mb_x, mb_y; 1074  unsigned width; 1075  uint16_t (*buffer)[4]; 1076  int left[4], top[4], topleft[4]; 1077  const int linesize = s->linesize[0]; 1078  const int mask = ((1 << s->bits) - 1) << point_transform; 1079  int resync_mb_y = 0; 1080  int resync_mb_x = 0; 1081  int vpred[6]; 1082  1083  if (!s->bayer && s->nb_components < 3) 1084  return AVERROR_INVALIDDATA; 1085  if (s->bayer && s->nb_components > 2) 1086  return AVERROR_INVALIDDATA; 1087  if (s->nb_components <= 0 || s->nb_components > 4) 1088  return AVERROR_INVALIDDATA; 1089  if (s->v_max != 1 || s->h_max != 1 || !s->lossless) 1090  return AVERROR_INVALIDDATA; 1091  1092  1093  s->restart_count = s->restart_interval; 1094  1095  if (s->restart_interval == 0) 1096  s->restart_interval = INT_MAX; 1097  1098  if (s->bayer) 1099  width = s->mb_width / nb_components; /* Interleaved, width stored is the total so need to divide */ 1100  else 1101  width = s->mb_width; 1102  1103  av_fast_malloc(&s->ljpeg_buffer, &s->ljpeg_buffer_size, width * 4 * sizeof(s->ljpeg_buffer[0][0])); 1104  if (!s->ljpeg_buffer) 1105  return AVERROR(ENOMEM); 1106  1107  buffer = s->ljpeg_buffer; 1108  1109  for (i = 0; i < 4; i++) 1110  buffer[0][i] = 1 << (s->bits - 1); 1111  1112  for (mb_y = 0; mb_y < s->mb_height; mb_y++) { 1113  uint8_t *ptr = s->picture_ptr->data[0] + (linesize * mb_y); 1114  1115  if (s->interlaced && s->bottom_field) 1116  ptr += linesize >> 1; 1117  1118  for (i = 0; i < 4; i++) 1119  top[i] = left[i] = topleft[i] = buffer[0][i]; 1120  1121  if ((mb_y * s->width) % s->restart_interval == 0) { 1122  for (i = 0; i < 6; i++) 1123  vpred[i] = 1 << (s->bits-1); 1124  } 1125  1126  for (mb_x = 0; mb_x < width; mb_x++) { 1127  int modified_predictor = predictor; 1128  1129  if (get_bits_left(&s->gb) < 1) { 1130  av_log(s->avctx, AV_LOG_ERROR, "bitstream end in rgb_scan\n"); 1131  return AVERROR_INVALIDDATA; 1132  } 1133  1134  if (s->restart_interval && !s->restart_count){ 1135  s->restart_count = s->restart_interval; 1136  resync_mb_x = mb_x; 1137  resync_mb_y = mb_y; 1138  for(i=0; i<4; i++) 1139  top[i] = left[i]= topleft[i]= 1 << (s->bits - 1); 1140  } 1141  if (mb_y == resync_mb_y || mb_y == resync_mb_y+1 && mb_x < resync_mb_x || !mb_x) 1142  modified_predictor = 1; 1143  1144  for (i=0;i<nb_components;i++) { 1145  int pred, dc; 1146  1147  topleft[i] = top[i]; 1148  top[i] = buffer[mb_x][i]; 1149  1150  dc = mjpeg_decode_dc(s, s->dc_index[i]); 1151  if(dc == 0xFFFFF) 1152  return -1; 1153  1154  if (!s->bayer || mb_x) { 1155  pred = left[i]; 1156  } else { /* This path runs only for the first line in bayer images */ 1157  vpred[i] += dc; 1158  pred = vpred[i] - dc; 1159  } 1160  1161  PREDICT(pred, topleft[i], top[i], pred, modified_predictor); 1162  1163  left[i] = buffer[mb_x][i] = 1164  mask & (pred + (unsigned)(dc * (1 << point_transform))); 1165  } 1166  1167  if (s->restart_interval && !--s->restart_count) { 1168  align_get_bits(&s->gb); 1169  skip_bits(&s->gb, 16); /* skip RSTn */ 1170  } 1171  } 1172  if (s->rct && s->nb_components == 4) { 1173  for (mb_x = 0; mb_x < s->mb_width; mb_x++) { 1174  ptr[4*mb_x + 2] = buffer[mb_x][0] - ((buffer[mb_x][1] + buffer[mb_x][2] - 0x200) >> 2); 1175  ptr[4*mb_x + 1] = buffer[mb_x][1] + ptr[4*mb_x + 2]; 1176  ptr[4*mb_x + 3] = buffer[mb_x][2] + ptr[4*mb_x + 2]; 1177  ptr[4*mb_x + 0] = buffer[mb_x][3]; 1178  } 1179  } else if (s->nb_components == 4) { 1180  for(i=0; i<nb_components; i++) { 1181  int c= s->comp_index[i]; 1182  if (s->bits <= 8) { 1183  for(mb_x = 0; mb_x < s->mb_width; mb_x++) { 1184  ptr[4*mb_x+3-c] = buffer[mb_x][i]; 1185  } 1186  } else if(s->bits == 9) { 1187  return AVERROR_PATCHWELCOME; 1188  } else { 1189  for(mb_x = 0; mb_x < s->mb_width; mb_x++) { 1190  ((uint16_t*)ptr)[4*mb_x+c] = buffer[mb_x][i]; 1191  } 1192  } 1193  } 1194  } else if (s->rct) { 1195  for (mb_x = 0; mb_x < s->mb_width; mb_x++) { 1196  ptr[3*mb_x + 1] = buffer[mb_x][0] - ((buffer[mb_x][1] + buffer[mb_x][2] - 0x200) >> 2); 1197  ptr[3*mb_x + 0] = buffer[mb_x][1] + ptr[3*mb_x + 1]; 1198  ptr[3*mb_x + 2] = buffer[mb_x][2] + ptr[3*mb_x + 1]; 1199  } 1200  } else if (s->pegasus_rct) { 1201  for (mb_x = 0; mb_x < s->mb_width; mb_x++) { 1202  ptr[3*mb_x + 1] = buffer[mb_x][0] - ((buffer[mb_x][1] + buffer[mb_x][2]) >> 2); 1203  ptr[3*mb_x + 0] = buffer[mb_x][1] + ptr[3*mb_x + 1]; 1204  ptr[3*mb_x + 2] = buffer[mb_x][2] + ptr[3*mb_x + 1]; 1205  } 1206  } else if (s->bayer) { 1207  if (nb_components == 1) { 1208  /* Leave decoding to the TIFF/DNG decoder (see comment in ff_mjpeg_decode_sof) */ 1209  for (mb_x = 0; mb_x < width; mb_x++) 1210  ((uint16_t*)ptr)[mb_x] = buffer[mb_x][0]; 1211  } else if (nb_components == 2) { 1212  for (mb_x = 0; mb_x < width; mb_x++) { 1213  ((uint16_t*)ptr)[2*mb_x + 0] = buffer[mb_x][0]; 1214  ((uint16_t*)ptr)[2*mb_x + 1] = buffer[mb_x][1]; 1215  } 1216  } 1217  } else { 1218  for(i=0; i<nb_components; i++) { 1219  int c= s->comp_index[i]; 1220  if (s->bits <= 8) { 1221  for(mb_x = 0; mb_x < s->mb_width; mb_x++) { 1222  ptr[3*mb_x+2-c] = buffer[mb_x][i]; 1223  } 1224  } else if(s->bits == 9) { 1225  return AVERROR_PATCHWELCOME; 1226  } else { 1227  for(mb_x = 0; mb_x < s->mb_width; mb_x++) { 1228  ((uint16_t*)ptr)[3*mb_x+2-c] = buffer[mb_x][i]; 1229  } 1230  } 1231  } 1232  } 1233  } 1234  return 0; 1235 } 1236  1238  int point_transform, int nb_components) 1239 { 1240  int i, mb_x, mb_y, mask; 1241  int bits= (s->bits+7)&~7; 1242  int resync_mb_y = 0; 1243  int resync_mb_x = 0; 1244  1245  point_transform += bits - s->bits; 1246  mask = ((1 << s->bits) - 1) << point_transform; 1247  1248  av_assert0(nb_components>=1 && nb_components<=4); 1249  1250  for (mb_y = 0; mb_y < s->mb_height; mb_y++) { 1251  for (mb_x = 0; mb_x < s->mb_width; mb_x++) { 1252  if (get_bits_left(&s->gb) < 1) { 1253  av_log(s->avctx, AV_LOG_ERROR, "bitstream end in yuv_scan\n"); 1254  return AVERROR_INVALIDDATA; 1255  } 1256  if (s->restart_interval && !s->restart_count){ 1257  s->restart_count = s->restart_interval; 1258  resync_mb_x = mb_x; 1259  resync_mb_y = mb_y; 1260  } 1261  1262  if(!mb_x || mb_y == resync_mb_y || mb_y == resync_mb_y+1 && mb_x < resync_mb_x || s->interlaced){ 1263  int toprow = mb_y == resync_mb_y || mb_y == resync_mb_y+1 && mb_x < resync_mb_x; 1264  int leftcol = !mb_x || mb_y == resync_mb_y && mb_x == resync_mb_x; 1265  for (i = 0; i < nb_components; i++) { 1266  uint8_t *ptr; 1267  uint16_t *ptr16; 1268  int n, h, v, x, y, c, j, linesize; 1269  n = s->nb_blocks[i]; 1270  c = s->comp_index[i]; 1271  h = s->h_scount[i]; 1272  v = s->v_scount[i]; 1273  x = 0; 1274  y = 0; 1275  linesize= s->linesize[c]; 1276  1277  if(bits>8) linesize /= 2; 1278  1279  for(j=0; j<n; j++) { 1280  int pred, dc; 1281  1282  dc = mjpeg_decode_dc(s, s->dc_index[i]); 1283  if(dc == 0xFFFFF) 1284  return -1; 1285  if ( h * mb_x + x >= s->width 1286  || v * mb_y + y >= s->height) { 1287  // Nothing to do 1288  } else if (bits<=8) { 1289  ptr = s->picture_ptr->data[c] + (linesize * (v * mb_y + y)) + (h * mb_x + x); //FIXME optimize this crap 1290  if(y==0 && toprow){ 1291  if(x==0 && leftcol){ 1292  pred= 1 << (bits - 1); 1293  }else{ 1294  pred= ptr[-1]; 1295  } 1296  }else{ 1297  if(x==0 && leftcol){ 1298  pred= ptr[-linesize]; 1299  }else{ 1300  PREDICT(pred, ptr[-linesize-1], ptr[-linesize], ptr[-1], predictor); 1301  } 1302  } 1303  1304  if (s->interlaced && s->bottom_field) 1305  ptr += linesize >> 1; 1306  pred &= mask; 1307  *ptr= pred + ((unsigned)dc << point_transform); 1308  }else{ 1309  ptr16 = (uint16_t*)(s->picture_ptr->data[c] + 2*(linesize * (v * mb_y + y)) + 2*(h * mb_x + x)); //FIXME optimize this crap 1310  if(y==0 && toprow){ 1311  if(x==0 && leftcol){ 1312  pred= 1 << (bits - 1); 1313  }else{ 1314  pred= ptr16[-1]; 1315  } 1316  }else{ 1317  if(x==0 && leftcol){ 1318  pred= ptr16[-linesize]; 1319  }else{ 1320  PREDICT(pred, ptr16[-linesize-1], ptr16[-linesize], ptr16[-1], predictor); 1321  } 1322  } 1323  1324  if (s->interlaced && s->bottom_field) 1325  ptr16 += linesize >> 1; 1326  pred &= mask; 1327  *ptr16= pred + ((unsigned)dc << point_transform); 1328  } 1329  if (++x == h) { 1330  x = 0; 1331  y++; 1332  } 1333  } 1334  } 1335  } else { 1336  for (i = 0; i < nb_components; i++) { 1337  uint8_t *ptr; 1338  uint16_t *ptr16; 1339  int n, h, v, x, y, c, j, linesize, dc; 1340  n = s->nb_blocks[i]; 1341  c = s->comp_index[i]; 1342  h = s->h_scount[i]; 1343  v = s->v_scount[i]; 1344  x = 0; 1345  y = 0; 1346  linesize = s->linesize[c]; 1347  1348  if(bits>8) linesize /= 2; 1349  1350  for (j = 0; j < n; j++) { 1351  int pred; 1352  1353  dc = mjpeg_decode_dc(s, s->dc_index[i]); 1354  if(dc == 0xFFFFF) 1355  return -1; 1356  if ( h * mb_x + x >= s->width 1357  || v * mb_y + y >= s->height) { 1358  // Nothing to do 1359  } else if (bits<=8) { 1360  ptr = s->picture_ptr->data[c] + 1361  (linesize * (v * mb_y + y)) + 1362  (h * mb_x + x); //FIXME optimize this crap 1363  PREDICT(pred, ptr[-linesize-1], ptr[-linesize], ptr[-1], predictor); 1364  1365  pred &= mask; 1366  *ptr = pred + ((unsigned)dc << point_transform); 1367  }else{ 1368  ptr16 = (uint16_t*)(s->picture_ptr->data[c] + 2*(linesize * (v * mb_y + y)) + 2*(h * mb_x + x)); //FIXME optimize this crap 1369  PREDICT(pred, ptr16[-linesize-1], ptr16[-linesize], ptr16[-1], predictor); 1370  1371  pred &= mask; 1372  *ptr16= pred + ((unsigned)dc << point_transform); 1373  } 1374  1375  if (++x == h) { 1376  x = 0; 1377  y++; 1378  } 1379  } 1380  } 1381  } 1382  if (s->restart_interval && !--s->restart_count) { 1383  align_get_bits(&s->gb); 1384  skip_bits(&s->gb, 16); /* skip RSTn */ 1385  } 1386  } 1387  } 1388  return 0; 1389 } 1390  1392  uint8_t *dst, const uint8_t *src, 1393  int linesize, int lowres) 1394 { 1395  switch (lowres) { 1396  case 0: s->hdsp.put_pixels_tab[1][0](dst, src, linesize, 8); 1397  break; 1398  case 1: copy_block4(dst, src, linesize, linesize, 4); 1399  break; 1400  case 2: copy_block2(dst, src, linesize, linesize, 2); 1401  break; 1402  case 3: *dst = *src; 1403  break; 1404  } 1405 } 1406  1407 static void shift_output(MJpegDecodeContext *s, uint8_t *ptr, int linesize) 1408 { 1409  int block_x, block_y; 1410  int size = 8 >> s->avctx->lowres; 1411  if (s->bits > 8) { 1412  for (block_y=0; block_y<size; block_y++) 1413  for (block_x=0; block_x<size; block_x++) 1414  *(uint16_t*)(ptr + 2*block_x + block_y*linesize) <<= 16 - s->bits; 1415  } else { 1416  for (block_y=0; block_y<size; block_y++) 1417  for (block_x=0; block_x<size; block_x++) 1418  *(ptr + block_x + block_y*linesize) <<= 8 - s->bits; 1419  } 1420 } 1421  1422 static int mjpeg_decode_scan(MJpegDecodeContext *s, int nb_components, int Ah, 1423  int Al, const uint8_t *mb_bitmask, 1424  int mb_bitmask_size, 1425  const AVFrame *reference) 1426 { 1427  int i, mb_x, mb_y, chroma_h_shift, chroma_v_shift, chroma_width, chroma_height; 1428  uint8_t *data[MAX_COMPONENTS]; 1429  const uint8_t *reference_data[MAX_COMPONENTS]; 1430  int linesize[MAX_COMPONENTS]; 1431  GetBitContext mb_bitmask_gb = {0}; // initialize to silence gcc warning 1432  int bytes_per_pixel = 1 + (s->bits > 8); 1433  1434  if (mb_bitmask) { 1435  if (mb_bitmask_size != (s->mb_width * s->mb_height + 7)>>3) { 1436  av_log(s->avctx, AV_LOG_ERROR, "mb_bitmask_size mismatches\n"); 1437  return AVERROR_INVALIDDATA; 1438  } 1439  init_get_bits(&mb_bitmask_gb, mb_bitmask, s->mb_width * s->mb_height); 1440  } 1441  1442  s->restart_count = 0; 1443  1444  av_pix_fmt_get_chroma_sub_sample(s->avctx->pix_fmt, &chroma_h_shift, 1445  &chroma_v_shift); 1446  chroma_width = AV_CEIL_RSHIFT(s->width, chroma_h_shift); 1447  chroma_height = AV_CEIL_RSHIFT(s->height, chroma_v_shift); 1448  1449  for (i = 0; i < nb_components; i++) { 1450  int c = s->comp_index[i]; 1451  data[c] = s->picture_ptr->data[c]; 1452  reference_data[c] = reference ? reference->data[c] : NULL; 1453  linesize[c] = s->linesize[c]; 1454  s->coefs_finished[c] |= 1; 1455  } 1456  1457  for (mb_y = 0; mb_y < s->mb_height; mb_y++) { 1458  for (mb_x = 0; mb_x < s->mb_width; mb_x++) { 1459  const int copy_mb = mb_bitmask && !get_bits1(&mb_bitmask_gb); 1460  1461  if (s->restart_interval && !s->restart_count) 1462  s->restart_count = s->restart_interval; 1463  1464  if (get_bits_left(&s->gb) < 0) { 1465  av_log(s->avctx, AV_LOG_ERROR, "overread %d\n", 1466  -get_bits_left(&s->gb)); 1467  return AVERROR_INVALIDDATA; 1468  } 1469  for (i = 0; i < nb_components; i++) { 1470  uint8_t *ptr; 1471  int n, h, v, x, y, c, j; 1472  int block_offset; 1473  n = s->nb_blocks[i]; 1474  c = s->comp_index[i]; 1475  h = s->h_scount[i]; 1476  v = s->v_scount[i]; 1477  x = 0; 1478  y = 0; 1479  for (j = 0; j < n; j++) { 1480  block_offset = (((linesize[c] * (v * mb_y + y) * 8) + 1481  (h * mb_x + x) * 8 * bytes_per_pixel) >> s->avctx->lowres); 1482  1483  if (s->interlaced && s->bottom_field) 1484  block_offset += linesize[c] >> 1; 1485  if ( 8*(h * mb_x + x) < ((c == 1) || (c == 2) ? chroma_width : s->width) 1486  && 8*(v * mb_y + y) < ((c == 1) || (c == 2) ? chroma_height : s->height)) { 1487  ptr = data[c] + block_offset; 1488  } else 1489  ptr = NULL; 1490  if (!s->progressive) { 1491  if (copy_mb) { 1492  if (ptr) 1493  mjpeg_copy_block(s, ptr, reference_data[c] + block_offset, 1494  linesize[c], s->avctx->lowres); 1495  1496  } else { 1497  s->bdsp.clear_block(s->block); 1498  if (decode_block(s, s->block, i, 1499  s->dc_index[i], s->ac_index[i], 1500  s->quant_matrixes[s->quant_sindex[i]]) < 0) { 1501  av_log(s->avctx, AV_LOG_ERROR, 1502  "error y=%d x=%d\n", mb_y, mb_x); 1503  return AVERROR_INVALIDDATA; 1504  } 1505  if (ptr) { 1506  s->idsp.idct_put(ptr, linesize[c], s->block); 1507  if (s->bits & 7) 1508  shift_output(s, ptr, linesize[c]); 1509  } 1510  } 1511  } else { 1512  int block_idx = s->block_stride[c] * (v * mb_y + y) + 1513  (h * mb_x + x); 1514  int16_t *block = s->blocks[c][block_idx]; 1515  if (Ah) 1516  block[0] += get_bits1(&s->gb) * 1517  s->quant_matrixes[s->quant_sindex[i]][0] << Al; 1518  else if (decode_dc_progressive(s, block, i, s->dc_index[i], 1519  s->quant_matrixes[s->quant_sindex[i]], 1520  Al) < 0) { 1521  av_log(s->avctx, AV_LOG_ERROR, 1522  "error y=%d x=%d\n", mb_y, mb_x); 1523  return AVERROR_INVALIDDATA; 1524  } 1525  } 1526  ff_dlog(s->avctx, "mb: %d %d processed\n", mb_y, mb_x); 1527  ff_dlog(s->avctx, "%d %d %d %d %d %d %d %d \n", 1528  mb_x, mb_y, x, y, c, s->bottom_field, 1529  (v * mb_y + y) * 8, (h * mb_x + x) * 8); 1530  if (++x == h) { 1531  x = 0; 1532  y++; 1533  } 1534  } 1535  } 1536  1537  handle_rstn(s, nb_components); 1538  } 1539  } 1540  return 0; 1541 } 1542  1544  int se, int Ah, int Al) 1545 { 1546  int mb_x, mb_y; 1547  int EOBRUN = 0; 1548  int c = s->comp_index[0]; 1549  uint16_t *quant_matrix = s->quant_matrixes[s->quant_sindex[0]]; 1550  1551  av_assert0(ss>=0 && Ah>=0 && Al>=0); 1552  if (se < ss || se > 63) { 1553  av_log(s->avctx, AV_LOG_ERROR, "SS/SE %d/%d is invalid\n", ss, se); 1554  return AVERROR_INVALIDDATA; 1555  } 1556  1557  // s->coefs_finished is a bitmask for coefficients coded 1558  // ss and se are parameters telling start and end coefficients 1559  s->coefs_finished[c] |= (2ULL << se) - (1ULL << ss); 1560  1561  s->restart_count = 0; 1562  1563  for (mb_y = 0; mb_y < s->mb_height; mb_y++) { 1564  int block_idx = mb_y * s->block_stride[c]; 1565  int16_t (*block)[64] = &s->blocks[c][block_idx]; 1566  uint8_t *last_nnz = &s->last_nnz[c][block_idx]; 1567  if (get_bits_left(&s->gb) <= 0) { 1568  av_log(s->avctx, AV_LOG_ERROR, "bitstream truncated in mjpeg_decode_scan_progressive_ac\n"); 1569  return AVERROR_INVALIDDATA; 1570  } 1571  for (mb_x = 0; mb_x < s->mb_width; mb_x++, block++, last_nnz++) { 1572  int ret; 1573  if (s->restart_interval && !s->restart_count) 1574  s->restart_count = s->restart_interval; 1575  1576  if (Ah) 1577  ret = decode_block_refinement(s, *block, last_nnz, s->ac_index[0], 1578  quant_matrix, ss, se, Al, &EOBRUN); 1579  else 1580  ret = decode_block_progressive(s, *block, last_nnz, s->ac_index[0], 1581  quant_matrix, ss, se, Al, &EOBRUN); 1582  1583  if (ret >= 0 && get_bits_left(&s->gb) < 0) 1585  if (ret < 0) { 1586  av_log(s->avctx, AV_LOG_ERROR, 1587  "error y=%d x=%d\n", mb_y, mb_x); 1588  return AVERROR_INVALIDDATA; 1589  } 1590  1591  if (handle_rstn(s, 0)) 1592  EOBRUN = 0; 1593  } 1594  } 1595  return 0; 1596 } 1597  1599 { 1600  int mb_x, mb_y; 1601  int c; 1602  const int bytes_per_pixel = 1 + (s->bits > 8); 1603  const int block_size = s->lossless ? 1 : 8; 1604  1605  for (c = 0; c < s->nb_components; c++) { 1606  uint8_t *data = s->picture_ptr->data[c]; 1607  int linesize = s->linesize[c]; 1608  int h = s->h_max / s->h_count[c]; 1609  int v = s->v_max / s->v_count[c]; 1610  int mb_width = (s->width + h * block_size - 1) / (h * block_size); 1611  int mb_height = (s->height + v * block_size - 1) / (v * block_size); 1612  1613  if (~s->coefs_finished[c]) 1614  av_log(s->avctx, AV_LOG_WARNING, "component %d is incomplete\n", c); 1615  1616  if (s->interlaced && s->bottom_field) 1617  data += linesize >> 1; 1618  1619  for (mb_y = 0; mb_y < mb_height; mb_y++) { 1620  uint8_t *ptr = data + (mb_y * linesize * 8 >> s->avctx->lowres); 1621  int block_idx = mb_y * s->block_stride[c]; 1622  int16_t (*block)[64] = &s->blocks[c][block_idx]; 1623  for (mb_x = 0; mb_x < mb_width; mb_x++, block++) { 1624  s->idsp.idct_put(ptr, linesize, *block); 1625  if (s->bits & 7) 1626  shift_output(s, ptr, linesize); 1627  ptr += bytes_per_pixel*8 >> s->avctx->lowres; 1628  } 1629  } 1630  } 1631 } 1632  1633 int ff_mjpeg_decode_sos(MJpegDecodeContext *s, const uint8_t *mb_bitmask, 1634  int mb_bitmask_size, const AVFrame *reference) 1635 { 1636  int len, nb_components, i, h, v, predictor, point_transform; 1637  int index, id, ret; 1638  const int block_size = s->lossless ? 1 : 8; 1639  int ilv, prev_shift; 1640  1641  if (!s->got_picture) { 1642  av_log(s->avctx, AV_LOG_WARNING, 1643  "Can not process SOS before SOF, skipping\n"); 1644  return -1; 1645  } 1646  1647  if (reference) { 1648  if (reference->width != s->picture_ptr->width || 1649  reference->height != s->picture_ptr->height || 1650  reference->format != s->picture_ptr->format) { 1651  av_log(s->avctx, AV_LOG_ERROR, "Reference mismatching\n"); 1652  return AVERROR_INVALIDDATA; 1653  } 1654  } 1655  1656  /* XXX: verify len field validity */ 1657  len = get_bits(&s->gb, 16); 1658  nb_components = get_bits(&s->gb, 8); 1659  if (nb_components == 0 || nb_components > MAX_COMPONENTS) { 1661  "decode_sos: nb_components (%d)", 1662  nb_components); 1663  return AVERROR_PATCHWELCOME; 1664  } 1665  if (len != 6 + 2 * nb_components) { 1666  av_log(s->avctx, AV_LOG_ERROR, "decode_sos: invalid len (%d)\n", len); 1667  return AVERROR_INVALIDDATA; 1668  } 1669  for (i = 0; i < nb_components; i++) { 1670  id = get_bits(&s->gb, 8) - 1; 1671  av_log(s->avctx, AV_LOG_DEBUG, "component: %d\n", id); 1672  /* find component index */ 1673  for (index = 0; index < s->nb_components; index++) 1674  if (id == s->component_id[index]) 1675  break; 1676  if (index == s->nb_components) { 1677  av_log(s->avctx, AV_LOG_ERROR, 1678  "decode_sos: index(%d) out of components\n", index); 1679  return AVERROR_INVALIDDATA; 1680  } 1681  /* Metasoft MJPEG codec has Cb and Cr swapped */ 1682  if (s->avctx->codec_tag == MKTAG('M', 'T', 'S', 'J') 1683  && nb_components == 3 && s->nb_components == 3 && i) 1684  index = 3 - i; 1685  1686  s->quant_sindex[i] = s->quant_index[index]; 1687  s->nb_blocks[i] = s->h_count[index] * s->v_count[index]; 1688  s->h_scount[i] = s->h_count[index]; 1689  s->v_scount[i] = s->v_count[index]; 1690  1691  if((nb_components == 1 || nb_components == 3) && s->nb_components == 3 && s->avctx->pix_fmt == AV_PIX_FMT_GBR24P) 1692  index = (index+2)%3; 1693  1694  s->comp_index[i] = index; 1695  1696  s->dc_index[i] = get_bits(&s->gb, 4); 1697  s->ac_index[i] = get_bits(&s->gb, 4); 1698  1699  if (s->dc_index[i] < 0 || s->ac_index[i] < 0 || 1700  s->dc_index[i] >= 4 || s->ac_index[i] >= 4) 1701  goto out_of_range; 1702  if (!s->vlcs[0][s->dc_index[i]].table || !(s->progressive ? s->vlcs[2][s->ac_index[0]].table : s->vlcs[1][s->ac_index[i]].table)) 1703  goto out_of_range; 1704  } 1705  1706  predictor = get_bits(&s->gb, 8); /* JPEG Ss / lossless JPEG predictor /JPEG-LS NEAR */ 1707  ilv = get_bits(&s->gb, 8); /* JPEG Se / JPEG-LS ILV */ 1708  if(s->avctx->codec_tag != AV_RL32("CJPG")){ 1709  prev_shift = get_bits(&s->gb, 4); /* Ah */ 1710  point_transform = get_bits(&s->gb, 4); /* Al */ 1711  }else 1712  prev_shift = point_transform = 0; 1713  1714  if (nb_components > 1) { 1715  /* interleaved stream */ 1716  s->mb_width = (s->width + s->h_max * block_size - 1) / (s->h_max * block_size); 1717  s->mb_height = (s->height + s->v_max * block_size - 1) / (s->v_max * block_size); 1718  } else if (!s->ls) { /* skip this for JPEG-LS */ 1719  h = s->h_max / s->h_scount[0]; 1720  v = s->v_max / s->v_scount[0]; 1721  s->mb_width = (s->width + h * block_size - 1) / (h * block_size); 1722  s->mb_height = (s->height + v * block_size - 1) / (v * block_size); 1723  s->nb_blocks[0] = 1; 1724  s->h_scount[0] = 1; 1725  s->v_scount[0] = 1; 1726  } 1727  1728  if (s->avctx->debug & FF_DEBUG_PICT_INFO) 1729  av_log(s->avctx, AV_LOG_DEBUG, "%s %s p:%d >>:%d ilv:%d bits:%d skip:%d %s comp:%d\n", 1730  s->lossless ? "lossless" : "sequential DCT", s->rgb ? "RGB" : "", 1731  predictor, point_transform, ilv, s->bits, s->mjpb_skiptosod, 1732  s->pegasus_rct ? "PRCT" : (s->rct ? "RCT" : ""), nb_components); 1733  1734  1735  /* mjpeg-b can have padding bytes between sos and image data, skip them */ 1736  for (i = s->mjpb_skiptosod; i > 0; i--) 1737  skip_bits(&s->gb, 8); 1738  1739 next_field: 1740  for (i = 0; i < nb_components; i++) 1741  s->last_dc[i] = (4 << s->bits); 1742  1743  if (s->avctx->hwaccel) { 1744  int bytes_to_start = get_bits_count(&s->gb) / 8; 1745  av_assert0(bytes_to_start >= 0 && 1746  s->raw_scan_buffer_size >= bytes_to_start); 1747  1748  ret = s->avctx->hwaccel->decode_slice(s->avctx, 1749  s->raw_scan_buffer + bytes_to_start, 1750  s->raw_scan_buffer_size - bytes_to_start); 1751  if (ret < 0) 1752  return ret; 1753  1754  } else if (s->lossless) { 1755  av_assert0(s->picture_ptr == s->picture); 1756  if (CONFIG_JPEGLS_DECODER && s->ls) { 1757 // for () { 1758 // reset_ls_coding_parameters(s, 0); 1759  1761  point_transform, ilv)) < 0) 1762  return ret; 1763  } else { 1764  if (s->rgb || s->bayer) { 1765  if ((ret = ljpeg_decode_rgb_scan(s, nb_components, predictor, point_transform)) < 0) 1766  return ret; 1767  } else { 1769  point_transform, 1770  nb_components)) < 0) 1771  return ret; 1772  } 1773  } 1774  } else { 1775  if (s->progressive && predictor) { 1776  av_assert0(s->picture_ptr == s->picture); 1778  ilv, prev_shift, 1779  point_transform)) < 0) 1780  return ret; 1781  } else { 1782  if ((ret = mjpeg_decode_scan(s, nb_components, 1783  prev_shift, point_transform, 1784  mb_bitmask, mb_bitmask_size, reference)) < 0) 1785  return ret; 1786  } 1787  } 1788  1789  if (s->interlaced && 1790  get_bits_left(&s->gb) > 32 && 1791  show_bits(&s->gb, 8) == 0xFF) { 1792  GetBitContext bak = s->gb; 1793  align_get_bits(&bak); 1794  if (show_bits(&bak, 16) == 0xFFD1) { 1795  av_log(s->avctx, AV_LOG_DEBUG, "AVRn interlaced picture marker found\n"); 1796  s->gb = bak; 1797  skip_bits(&s->gb, 16); 1798  s->bottom_field ^= 1; 1799  1800  goto next_field; 1801  } 1802  } 1803  1804  emms_c(); 1805  return 0; 1806  out_of_range: 1807  av_log(s->avctx, AV_LOG_ERROR, "decode_sos: ac/dc index out of range\n"); 1808  return AVERROR_INVALIDDATA; 1809 } 1810  1812 { 1813  if (get_bits(&s->gb, 16) != 4) 1814  return AVERROR_INVALIDDATA; 1815  s->restart_interval = get_bits(&s->gb, 16); 1816  s->restart_count = 0; 1817  av_log(s->avctx, AV_LOG_DEBUG, "restart interval: %d\n", 1818  s->restart_interval); 1819  1820  return 0; 1821 } 1822  1824 { 1825  int len, id, i; 1826  1827  len = get_bits(&s->gb, 16); 1828  if (len < 6) { 1829  if (s->bayer) { 1830  // Pentax K-1 (digital camera) JPEG images embedded in DNG images contain unknown APP0 markers 1831  av_log(s->avctx, AV_LOG_WARNING, "skipping APPx (len=%"PRId32") for bayer-encoded image\n", len); 1832  skip_bits(&s->gb, len); 1833  return 0; 1834  } else 1835  return AVERROR_INVALIDDATA; 1836  } 1837  if (8 * len > get_bits_left(&s->gb)) 1838  return AVERROR_INVALIDDATA; 1839  1840  id = get_bits_long(&s->gb, 32); 1841  len -= 6; 1842  1843  if (s->avctx->debug & FF_DEBUG_STARTCODE) 1844  av_log(s->avctx, AV_LOG_DEBUG, "APPx (%s / %8X) len=%d\n", 1845  av_fourcc2str(av_bswap32(id)), id, len); 1846  1847  /* Buggy AVID, it puts EOI only at every 10th frame. */ 1848  /* Also, this fourcc is used by non-avid files too, it holds some 1849  information, but it's always present in AVID-created files. */ 1850  if (id == AV_RB32("AVI1")) { 1851  /* structure: 1852  4bytes AVI1 1853  1bytes polarity 1854  1bytes always zero 1855  4bytes field_size 1856  4bytes field_size_less_padding 1857  */ 1858  s->buggy_avid = 1; 1859  i = get_bits(&s->gb, 8); len--; 1860  av_log(s->avctx, AV_LOG_DEBUG, "polarity %d\n", i); 1861  goto out; 1862  } 1863  1864  if (id == AV_RB32("JFIF")) { 1865  int t_w, t_h, v1, v2; 1866  if (len < 8) 1867  goto out; 1868  skip_bits(&s->gb, 8); /* the trailing zero-byte */ 1869  v1 = get_bits(&s->gb, 8); 1870  v2 = get_bits(&s->gb, 8); 1871  skip_bits(&s->gb, 8); 1872  1873  s->avctx->sample_aspect_ratio.num = get_bits(&s->gb, 16); 1874  s->avctx->sample_aspect_ratio.den = get_bits(&s->gb, 16); 1875  if ( s->avctx->sample_aspect_ratio.num <= 0 1876  || s->avctx->sample_aspect_ratio.den <= 0) { 1877  s->avctx->sample_aspect_ratio.num = 0; 1878  s->avctx->sample_aspect_ratio.den = 1; 1879  } 1880  1881  if (s->avctx->debug & FF_DEBUG_PICT_INFO) 1882  av_log(s->avctx, AV_LOG_INFO, 1883  "mjpeg: JFIF header found (version: %x.%x) SAR=%d/%d\n", 1884  v1, v2, 1885  s->avctx->sample_aspect_ratio.num, 1886  s->avctx->sample_aspect_ratio.den); 1887  1888  len -= 8; 1889  if (len >= 2) { 1890  t_w = get_bits(&s->gb, 8); 1891  t_h = get_bits(&s->gb, 8); 1892  if (t_w && t_h) { 1893  /* skip thumbnail */ 1894  if (len -10 - (t_w * t_h * 3) > 0) 1895  len -= t_w * t_h * 3; 1896  } 1897  len -= 2; 1898  } 1899  goto out; 1900  } 1901  1902  if ( id == AV_RB32("Adob") 1903  && len >= 7 1904  && show_bits(&s->gb, 8) == 'e' 1905  && show_bits_long(&s->gb, 32) != AV_RB32("e_CM")) { 1906  skip_bits(&s->gb, 8); /* 'e' */ 1907  skip_bits(&s->gb, 16); /* version */ 1908  skip_bits(&s->gb, 16); /* flags0 */ 1909  skip_bits(&s->gb, 16); /* flags1 */ 1910  s->adobe_transform = get_bits(&s->gb, 8); 1911  if (s->avctx->debug & FF_DEBUG_PICT_INFO) 1912  av_log(s->avctx, AV_LOG_INFO, "mjpeg: Adobe header found, transform=%d\n", s->adobe_transform); 1913  len -= 7; 1914  goto out; 1915  } 1916  1917  if (id == AV_RB32("LJIF")) { 1918  int rgb = s->rgb; 1919  int pegasus_rct = s->pegasus_rct; 1920  if (s->avctx->debug & FF_DEBUG_PICT_INFO) 1921  av_log(s->avctx, AV_LOG_INFO, 1922  "Pegasus lossless jpeg header found\n"); 1923  skip_bits(&s->gb, 16); /* version ? */ 1924  skip_bits(&s->gb, 16); /* unknown always 0? */ 1925  skip_bits(&s->gb, 16); /* unknown always 0? */ 1926  skip_bits(&s->gb, 16); /* unknown always 0? */ 1927  switch (i=get_bits(&s->gb, 8)) { 1928  case 1: 1929  rgb = 1; 1930  pegasus_rct = 0; 1931  break; 1932  case 2: 1933  rgb = 1; 1934  pegasus_rct = 1; 1935  break; 1936  default: 1937  av_log(s->avctx, AV_LOG_ERROR, "unknown colorspace %d\n", i); 1938  } 1939  1940  len -= 9; 1941  if (s->got_picture) 1942  if (rgb != s->rgb || pegasus_rct != s->pegasus_rct) { 1943  av_log(s->avctx, AV_LOG_WARNING, "Mismatching LJIF tag\n"); 1944  goto out; 1945  } 1946  1947  s->rgb = rgb; 1948  s->pegasus_rct = pegasus_rct; 1949  1950  goto out; 1951  } 1952  if (id == AV_RL32("colr") && len > 0) { 1953  s->colr = get_bits(&s->gb, 8); 1954  if (s->avctx->debug & FF_DEBUG_PICT_INFO) 1955  av_log(s->avctx, AV_LOG_INFO, "COLR %d\n", s->colr); 1956  len --; 1957  goto out; 1958  } 1959  if (id == AV_RL32("xfrm") && len > 0) { 1960  s->xfrm = get_bits(&s->gb, 8); 1961  if (s->avctx->debug & FF_DEBUG_PICT_INFO) 1962  av_log(s->avctx, AV_LOG_INFO, "XFRM %d\n", s->xfrm); 1963  len --; 1964  goto out; 1965  } 1966  1967  /* JPS extension by VRex */ 1968  if (s->start_code == APP3 && id == AV_RB32("_JPS") && len >= 10) { 1969  int flags, layout, type; 1970  if (s->avctx->debug & FF_DEBUG_PICT_INFO) 1971  av_log(s->avctx, AV_LOG_INFO, "_JPSJPS_\n"); 1972  1973  skip_bits(&s->gb, 32); len -= 4; /* JPS_ */ 1974  skip_bits(&s->gb, 16); len -= 2; /* block length */ 1975  skip_bits(&s->gb, 8); /* reserved */ 1976  flags = get_bits(&s->gb, 8); 1977  layout = get_bits(&s->gb, 8); 1978  type = get_bits(&s->gb, 8); 1979  len -= 4; 1980  1981  av_freep(&s->stereo3d); 1982  s->stereo3d = av_stereo3d_alloc(); 1983  if (!s->stereo3d) { 1984  goto out; 1985  } 1986  if (type == 0) { 1987  s->stereo3d->type = AV_STEREO3D_2D; 1988  } else if (type == 1) { 1989  switch (layout) { 1990  case 0x01: 1991  s->stereo3d->type = AV_STEREO3D_LINES; 1992  break; 1993  case 0x02: 1994  s->stereo3d->type = AV_STEREO3D_SIDEBYSIDE; 1995  break; 1996  case 0x03: 1997  s->stereo3d->type = AV_STEREO3D_TOPBOTTOM; 1998  break; 1999  } 2000  if (!(flags & 0x04)) { 2001  s->stereo3d->flags = AV_STEREO3D_FLAG_INVERT; 2002  } 2003  } 2004  goto out; 2005  } 2006  2007  /* EXIF metadata */ 2008  if (s->start_code == APP1 && id == AV_RB32("Exif") && len >= 2) { 2009  GetByteContext gbytes; 2010  int ret, le, ifd_offset, bytes_read; 2011  const uint8_t *aligned; 2012  2013  skip_bits(&s->gb, 16); // skip padding 2014  len -= 2; 2015  2016  // init byte wise reading 2017  aligned = align_get_bits(&s->gb); 2018  bytestream2_init(&gbytes, aligned, len); 2019  2020  // read TIFF header 2021  ret = ff_tdecode_header(&gbytes, &le, &ifd_offset); 2022  if (ret) { 2023  av_log(s->avctx, AV_LOG_ERROR, "mjpeg: invalid TIFF header in EXIF data\n"); 2024  } else { 2025  bytestream2_seek(&gbytes, ifd_offset, SEEK_SET); 2026  2027  // read 0th IFD and store the metadata 2028  // (return values > 0 indicate the presence of subimage metadata) 2029  ret = ff_exif_decode_ifd(s->avctx, &gbytes, le, 0, &s->exif_metadata); 2030  if (ret < 0) { 2031  av_log(s->avctx, AV_LOG_ERROR, "mjpeg: error decoding EXIF data\n"); 2032  } 2033  } 2034  2035  bytes_read = bytestream2_tell(&gbytes); 2036  skip_bits(&s->gb, bytes_read << 3); 2037  len -= bytes_read; 2038  2039  goto out; 2040  } 2041  2042  /* Apple MJPEG-A */ 2043  if ((s->start_code == APP1) && (len > (0x28 - 8))) { 2044  id = get_bits_long(&s->gb, 32); 2045  len -= 4; 2046  /* Apple MJPEG-A */ 2047  if (id == AV_RB32("mjpg")) { 2048  /* structure: 2049  4bytes field size 2050  4bytes pad field size 2051  4bytes next off 2052  4bytes quant off 2053  4bytes huff off 2054  4bytes image off 2055  4bytes scan off 2056  4bytes data off 2057  */ 2058  if (s->avctx->debug & FF_DEBUG_PICT_INFO) 2059  av_log(s->avctx, AV_LOG_INFO, "mjpeg: Apple MJPEG-A header found\n"); 2060  } 2061  } 2062  2063  if (s->start_code == APP2 && id == AV_RB32("ICC_") && len >= 10) { 2064  int id2; 2065  unsigned seqno; 2066  unsigned nummarkers; 2067  2068  id = get_bits_long(&s->gb, 32); 2069  id2 = get_bits(&s->gb, 24); 2070  len -= 7; 2071  if (id != AV_RB32("PROF") || id2 != AV_RB24("ILE")) { 2072  av_log(s->avctx, AV_LOG_WARNING, "Invalid ICC_PROFILE header in APP2\n"); 2073  goto out; 2074  } 2075  2076  skip_bits(&s->gb, 8); 2077  seqno = get_bits(&s->gb, 8); 2078  len -= 2; 2079  if (seqno == 0) { 2080  av_log(s->avctx, AV_LOG_WARNING, "Invalid sequence number in APP2\n"); 2081  goto out; 2082  } 2083  2084  nummarkers = get_bits(&s->gb, 8); 2085  len -= 1; 2086  if (nummarkers == 0) { 2087  av_log(s->avctx, AV_LOG_WARNING, "Invalid number of markers coded in APP2\n"); 2088  goto out; 2089  } else if (s->iccnum != 0 && nummarkers != s->iccnum) { 2090  av_log(s->avctx, AV_LOG_WARNING, "Mistmatch in coded number of ICC markers between markers\n"); 2091  goto out; 2092  } else if (seqno > nummarkers) { 2093  av_log(s->avctx, AV_LOG_WARNING, "Mismatching sequence number and coded number of ICC markers\n"); 2094  goto out; 2095  } 2096  2097  /* Allocate if this is the first APP2 we've seen. */ 2098  if (s->iccnum == 0) { 2099  if (!FF_ALLOCZ_TYPED_ARRAY(s->iccentries, nummarkers)) { 2100  av_log(s->avctx, AV_LOG_ERROR, "Could not allocate ICC data arrays\n"); 2101  return AVERROR(ENOMEM); 2102  } 2103  s->iccnum = nummarkers; 2104  } 2105  2106  if (s->iccentries[seqno - 1].data) { 2107  av_log(s->avctx, AV_LOG_WARNING, "Duplicate ICC sequence number\n"); 2108  goto out; 2109  } 2110  2111  s->iccentries[seqno - 1].length = len; 2112  s->iccentries[seqno - 1].data = av_malloc(len); 2113  if (!s->iccentries[seqno - 1].data) { 2114  av_log(s->avctx, AV_LOG_ERROR, "Could not allocate ICC data buffer\n"); 2115  return AVERROR(ENOMEM); 2116  } 2117  2118  memcpy(s->iccentries[seqno - 1].data, align_get_bits(&s->gb), len); 2119  skip_bits(&s->gb, len << 3); 2120  len = 0; 2121  s->iccread++; 2122  2123  if (s->iccread > s->iccnum) 2124  av_log(s->avctx, AV_LOG_WARNING, "Read more ICC markers than are supposed to be coded\n"); 2125  } 2126  2127 out: 2128  /* slow but needed for extreme adobe jpegs */ 2129  if (len < 0) 2130  av_log(s->avctx, AV_LOG_ERROR, 2131  "mjpeg: error, decode_app parser read over the end\n"); 2132  while (--len > 0) 2133  skip_bits(&s->gb, 8); 2134  2135  return 0; 2136 } 2137  2139 { 2140  int len = get_bits(&s->gb, 16); 2141  if (len >= 2 && 8 * len - 16 <= get_bits_left(&s->gb)) { 2142  int i; 2143  char *cbuf = av_malloc(len - 1); 2144  if (!cbuf) 2145  return AVERROR(ENOMEM); 2146  2147  for (i = 0; i < len - 2; i++) 2148  cbuf[i] = get_bits(&s->gb, 8); 2149  if (i > 0 && cbuf[i - 1] == '\n') 2150  cbuf[i - 1] = 0; 2151  else 2152  cbuf[i] = 0; 2153  2154  if (s->avctx->debug & FF_DEBUG_PICT_INFO) 2155  av_log(s->avctx, AV_LOG_INFO, "comment: '%s'\n", cbuf); 2156  2157  /* buggy avid, it puts EOI only at every 10th frame */ 2158  if (!strncmp(cbuf, "AVID", 4)) { 2159  parse_avid(s, cbuf, len); 2160  } else if (!strcmp(cbuf, "CS=ITU601")) 2161  s->cs_itu601 = 1; 2162  else if ((!strncmp(cbuf, "Intel(R) JPEG Library, version 1", 32) && s->avctx->codec_tag) || 2163  (!strncmp(cbuf, "Metasoft MJPEG Codec", 20))) 2164  s->flipped = 1; 2165  else if (!strcmp(cbuf, "MULTISCOPE II")) { 2166  s->avctx->sample_aspect_ratio = (AVRational) { 1, 2 }; 2167  s->multiscope = 2; 2168  } 2169  2170  av_free(cbuf); 2171  } 2172  2173  return 0; 2174 } 2175  2176 /* return the 8 bit start code value and update the search 2177  state. Return -1 if no start code found */ 2178 static int find_marker(const uint8_t **pbuf_ptr, const uint8_t *buf_end) 2179 { 2180  const uint8_t *buf_ptr; 2181  unsigned int v, v2; 2182  int val; 2183  int skipped = 0; 2184  2185  buf_ptr = *pbuf_ptr; 2186  while (buf_end - buf_ptr > 1) { 2187  v = *buf_ptr++; 2188  v2 = *buf_ptr; 2189  if ((v == 0xff) && (v2 >= SOF0) && (v2 <= COM) && buf_ptr < buf_end) { 2190  val = *buf_ptr++; 2191  goto found; 2192  } 2193  skipped++; 2194  } 2195  buf_ptr = buf_end; 2196  val = -1; 2197 found: 2198  ff_dlog(NULL, "find_marker skipped %d bytes\n", skipped); 2199  *pbuf_ptr = buf_ptr; 2200  return val; 2201 } 2202  2204  const uint8_t **buf_ptr, const uint8_t *buf_end, 2205  const uint8_t **unescaped_buf_ptr, 2206  int *unescaped_buf_size) 2207 { 2208  int start_code; 2209  start_code = find_marker(buf_ptr, buf_end); 2210  2211  av_fast_padded_malloc(&s->buffer, &s->buffer_size, buf_end - *buf_ptr); 2212  if (!s->buffer) 2213  return AVERROR(ENOMEM); 2214  2215  /* unescape buffer of SOS, use special treatment for JPEG-LS */ 2216  if (start_code == SOS && !s->ls) { 2217  const uint8_t *src = *buf_ptr; 2218  const uint8_t *ptr = src; 2219  uint8_t *dst = s->buffer; 2220  2221  #define copy_data_segment(skip) do { \ 2222  ptrdiff_t length = (ptr - src) - (skip); \ 2223  if (length > 0) { \ 2224  memcpy(dst, src, length); \ 2225  dst += length; \ 2226  src = ptr; \ 2227  } \ 2228  } while (0) 2229  2230  if (s->avctx->codec_id == AV_CODEC_ID_THP) { 2231  ptr = buf_end; 2232  copy_data_segment(0); 2233  } else { 2234  while (ptr < buf_end) { 2235  uint8_t x = *(ptr++); 2236  2237  if (x == 0xff) { 2238  ptrdiff_t skip = 0; 2239  while (ptr < buf_end && x == 0xff) { 2240  x = *(ptr++); 2241  skip++; 2242  } 2243  2244  /* 0xFF, 0xFF, ... */ 2245  if (skip > 1) { 2246  copy_data_segment(skip); 2247  2248  /* decrement src as it is equal to ptr after the 2249  * copy_data_segment macro and we might want to 2250  * copy the current value of x later on */ 2251  src--; 2252  } 2253  2254  if (x < RST0 || x > RST7) { 2255  copy_data_segment(1); 2256  if (x) 2257  break; 2258  } 2259  } 2260  } 2261  if (src < ptr) 2262  copy_data_segment(0); 2263  } 2264  #undef copy_data_segment 2265  2266  *unescaped_buf_ptr = s->buffer; 2267  *unescaped_buf_size = dst - s->buffer; 2268  memset(s->buffer + *unescaped_buf_size, 0, 2270  2271  av_log(s->avctx, AV_LOG_DEBUG, "escaping removed %"PTRDIFF_SPECIFIER" bytes\n", 2272  (buf_end - *buf_ptr) - (dst - s->buffer)); 2273  } else if (start_code == SOS && s->ls) { 2274  const uint8_t *src = *buf_ptr; 2275  uint8_t *dst = s->buffer; 2276  int bit_count = 0; 2277  int t = 0, b = 0; 2278  PutBitContext pb; 2279  2280  /* find marker */ 2281  while (src + t < buf_end) { 2282  uint8_t x = src[t++]; 2283  if (x == 0xff) { 2284  while ((src + t < buf_end) && x == 0xff) 2285  x = src[t++]; 2286  if (x & 0x80) { 2287  t -= FFMIN(2, t); 2288  break; 2289  } 2290  } 2291  } 2292  bit_count = t * 8; 2293  init_put_bits(&pb, dst, t); 2294  2295  /* unescape bitstream */ 2296  while (b < t) { 2297  uint8_t x = src[b++]; 2298  put_bits(&pb, 8, x); 2299  if (x == 0xFF && b < t) { 2300  x = src[b++]; 2301  if (x & 0x80) { 2302  av_log(s->avctx, AV_LOG_WARNING, "Invalid escape sequence\n"); 2303  x &= 0x7f; 2304  } 2305  put_bits(&pb, 7, x); 2306  bit_count--; 2307  } 2308  } 2309  flush_put_bits(&pb); 2310  2311  *unescaped_buf_ptr = dst; 2312  *unescaped_buf_size = (bit_count + 7) >> 3; 2313  memset(s->buffer + *unescaped_buf_size, 0, 2315  } else { 2316  *unescaped_buf_ptr = *buf_ptr; 2317  *unescaped_buf_size = buf_end - *buf_ptr; 2318  } 2319  2320  return start_code; 2321 } 2322  2324 { 2325  int i; 2326  2327  if (s->iccentries) { 2328  for (i = 0; i < s->iccnum; i++) 2329  av_freep(&s->iccentries[i].data); 2330  av_freep(&s->iccentries); 2331  } 2332  2333  s->iccread = 0; 2334  s->iccnum = 0; 2335 } 2336  2337 // SMV JPEG just stacks several output frames into one JPEG picture 2338 // we handle that by setting up the cropping parameters appropriately 2340 { 2341  MJpegDecodeContext *s = avctx->priv_data; 2342  int ret; 2343  2344  if (s->smv_next_frame > 0) { 2345  av_assert0(s->smv_frame->buf[0]); 2347  ret = av_frame_ref(frame, s->smv_frame); 2348  if (ret < 0) 2349  return ret; 2350  } else { 2351  av_assert0(frame->buf[0]); 2352  av_frame_unref(s->smv_frame); 2353  ret = av_frame_ref(s->smv_frame, frame); 2354  if (ret < 0) 2355  return ret; 2356  } 2357  2358  av_assert0((s->smv_next_frame + 1) * avctx->height <= avctx->coded_height); 2359  2360  frame->width = avctx->coded_width; 2361  frame->height = avctx->coded_height; 2362  frame->crop_top = FFMIN(s->smv_next_frame * avctx->height, frame->height); 2363  frame->crop_bottom = frame->height - (s->smv_next_frame + 1) * avctx->height; 2364  2365  s->smv_next_frame = (s->smv_next_frame + 1) % s->smv_frames_per_jpeg; 2366  2367  if (s->smv_next_frame == 0) 2368  av_frame_unref(s->smv_frame); 2369  2370  return 0; 2371 } 2372  2374 { 2375  MJpegDecodeContext *s = avctx->priv_data; 2376  int ret; 2377  2378  av_packet_unref(s->pkt); 2379  ret = ff_decode_get_packet(avctx, s->pkt); 2380  if (ret < 0) 2381  return ret; 2382  2383 #if CONFIG_SP5X_DECODER || CONFIG_AMV_DECODER 2384  if (avctx->codec_id == AV_CODEC_ID_SP5X || 2385  avctx->codec_id == AV_CODEC_ID_AMV) { 2386  ret = ff_sp5x_process_packet(avctx, s->pkt); 2387  if (ret < 0) 2388  return ret; 2389  } 2390 #endif 2391  2392  s->buf_size = s->pkt->size; 2393  2394  return 0; 2395 } 2396  2398 { 2399  MJpegDecodeContext *s = avctx->priv_data; 2400  const uint8_t *buf_end, *buf_ptr; 2401  const uint8_t *unescaped_buf_ptr; 2402  int hshift, vshift; 2403  int unescaped_buf_size; 2404  int start_code; 2405  int i, index; 2406  int ret = 0; 2407  int is16bit; 2408  AVDictionaryEntry *e = NULL; 2409  2410  s->force_pal8 = 0; 2411  2412  if (avctx->codec_id == AV_CODEC_ID_SMVJPEG && s->smv_next_frame > 0) 2413  return smv_process_frame(avctx, frame); 2414  2415  av_dict_free(&s->exif_metadata); 2416  av_freep(&s->stereo3d); 2417  s->adobe_transform = -1; 2418  2419  if (s->iccnum != 0) 2421  2422  ret = mjpeg_get_packet(avctx); 2423  if (ret < 0) 2424  return ret; 2425 redo_for_pal8: 2426  buf_ptr = s->pkt->data; 2427  buf_end = s->pkt->data + s->pkt->size; 2428  while (buf_ptr < buf_end) { 2429  /* find start next marker */ 2430  start_code = ff_mjpeg_find_marker(s, &buf_ptr, buf_end, 2431  &unescaped_buf_ptr, 2432  &unescaped_buf_size); 2433  /* EOF */ 2434  if (start_code < 0) { 2435  break; 2436  } else if (unescaped_buf_size > INT_MAX / 8) { 2437  av_log(avctx, AV_LOG_ERROR, 2438  "MJPEG packet 0x%x too big (%d/%d), corrupt data?\n", 2439  start_code, unescaped_buf_size, s->pkt->size); 2440  return AVERROR_INVALIDDATA; 2441  } 2442  av_log(avctx, AV_LOG_DEBUG, "marker=%x avail_size_in_buf=%"PTRDIFF_SPECIFIER"\n", 2443  start_code, buf_end - buf_ptr); 2444  2445  ret = init_get_bits8(&s->gb, unescaped_buf_ptr, unescaped_buf_size); 2446  2447  if (ret < 0) { 2448  av_log(avctx, AV_LOG_ERROR, "invalid buffer\n"); 2449  goto fail; 2450  } 2451  2452  s->start_code = start_code; 2453  if (s->avctx->debug & FF_DEBUG_STARTCODE) 2454  av_log(avctx, AV_LOG_DEBUG, "startcode: %X\n", start_code); 2455  2456  /* process markers */ 2457  if (start_code >= RST0 && start_code <= RST7) { 2458  av_log(avctx, AV_LOG_DEBUG, 2459  "restart marker: %d\n", start_code & 0x0f); 2460  /* APP fields */ 2461  } else if (start_code >= APP0 && start_code <= APP15) { 2462  if ((ret = mjpeg_decode_app(s)) < 0) 2463  av_log(avctx, AV_LOG_ERROR, "unable to decode APP fields: %s\n", 2464  av_err2str(ret)); 2465  /* Comment */ 2466  } else if (start_code == COM) { 2467  ret = mjpeg_decode_com(s); 2468  if (ret < 0) 2469  return ret; 2470  } else if (start_code == DQT) { 2472  if (ret < 0) 2473  return ret; 2474  } 2475  2476  ret = -1; 2477  2478  if (!CONFIG_JPEGLS_DECODER && 2479  (start_code == SOF48 || start_code == LSE)) { 2480  av_log(avctx, AV_LOG_ERROR, "JPEG-LS support not enabled.\n"); 2481  return AVERROR(ENOSYS); 2482  } 2483  2484  if (avctx->skip_frame == AVDISCARD_ALL) { 2485  switch(start_code) { 2486  case SOF0: 2487  case SOF1: 2488  case SOF2: 2489  case SOF3: 2490  case SOF48: 2491  case SOI: 2492  case SOS: 2493  case EOI: 2494  break; 2495  default: 2496  goto skip; 2497  } 2498  } 2499  2500  switch (start_code) { 2501  case SOI: 2502  s->restart_interval = 0; 2503  s->restart_count = 0; 2504  s->raw_image_buffer = buf_ptr; 2505  s->raw_image_buffer_size = buf_end - buf_ptr; 2506  /* nothing to do on SOI */ 2507  break; 2508  case DHT: 2509  if ((ret = ff_mjpeg_decode_dht(s)) < 0) { 2510  av_log(avctx, AV_LOG_ERROR, "huffman table decode error\n"); 2511  goto fail; 2512  } 2513  break; 2514  case SOF0: 2515  case SOF1: 2516  if (start_code == SOF0) 2517  s->avctx->profile = FF_PROFILE_MJPEG_HUFFMAN_BASELINE_DCT; 2518  else 2520  s->lossless = 0; 2521  s->ls = 0; 2522  s->progressive = 0; 2523  if ((ret = ff_mjpeg_decode_sof(s)) < 0) 2524  goto fail; 2525  break; 2526  case SOF2: 2527  s->avctx->profile = FF_PROFILE_MJPEG_HUFFMAN_PROGRESSIVE_DCT; 2528  s->lossless = 0; 2529  s->ls = 0; 2530  s->progressive = 1; 2531  if ((ret = ff_mjpeg_decode_sof(s)) < 0) 2532  goto fail; 2533  break; 2534  case SOF3: 2535  s->avctx->profile = FF_PROFILE_MJPEG_HUFFMAN_LOSSLESS; 2536  s->avctx->properties |= FF_CODEC_PROPERTY_LOSSLESS; 2537  s->lossless = 1; 2538  s->ls = 0; 2539  s->progressive = 0; 2540  if ((ret = ff_mjpeg_decode_sof(s)) < 0) 2541  goto fail; 2542  break; 2543  case SOF48: 2544  s->avctx->profile = FF_PROFILE_MJPEG_JPEG_LS; 2545  s->avctx->properties |= FF_CODEC_PROPERTY_LOSSLESS; 2546  s->lossless = 1; 2547  s->ls = 1; 2548  s->progressive = 0; 2549  if ((ret = ff_mjpeg_decode_sof(s)) < 0) 2550  goto fail; 2551  break; 2552  case LSE: 2553  if (!CONFIG_JPEGLS_DECODER || 2554  (ret = ff_jpegls_decode_lse(s)) < 0) 2555  goto fail; 2556  if (ret == 1) 2557  goto redo_for_pal8; 2558  break; 2559  case EOI: 2560 eoi_parser: 2561  if (!avctx->hwaccel && avctx->skip_frame != AVDISCARD_ALL && 2562  s->progressive && s->cur_scan && s->got_picture) 2564  s->cur_scan = 0; 2565  if (!s->got_picture) { 2566  av_log(avctx, AV_LOG_WARNING, 2567  "Found EOI before any SOF, ignoring\n"); 2568  break; 2569  } 2570  if (s->interlaced) { 2571  s->bottom_field ^= 1; 2572  /* if not bottom field, do not output image yet */ 2573  if (s->bottom_field == !s->interlace_polarity) 2574  break; 2575  } 2576  if (avctx->skip_frame == AVDISCARD_ALL) { 2577  s->got_picture = 0; 2578  ret = AVERROR(EAGAIN); 2579  goto the_end_no_picture; 2580  } 2581  if (s->avctx->hwaccel) { 2582  ret = s->avctx->hwaccel->end_frame(s->avctx); 2583  if (ret < 0) 2584  return ret; 2585  2586  av_freep(&s->hwaccel_picture_private); 2587  } 2588  if ((ret = av_frame_ref(frame, s->picture_ptr)) < 0) 2589  return ret; 2590  s->got_picture = 0; 2591  2592  frame->pkt_dts = s->pkt->dts; 2593  2594  if (!s->lossless && avctx->debug & FF_DEBUG_QP) { 2595  int qp = FFMAX3(s->qscale[0], 2596  s->qscale[1], 2597  s->qscale[2]); 2598  2599  av_log(avctx, AV_LOG_DEBUG, "QP: %d\n", qp); 2600  } 2601  2602  goto the_end; 2603  case SOS: 2604  s->raw_scan_buffer = buf_ptr; 2605  s->raw_scan_buffer_size = buf_end - buf_ptr; 2606  2607  s->cur_scan++; 2608  if (avctx->skip_frame == AVDISCARD_ALL) { 2609  skip_bits(&s->gb, get_bits_left(&s->gb)); 2610  break; 2611  } 2612  2613  if ((ret = ff_mjpeg_decode_sos(s, NULL, 0, NULL)) < 0 && 2614  (avctx->err_recognition & AV_EF_EXPLODE)) 2615  goto fail; 2616  break; 2617  case DRI: 2618  if ((ret = mjpeg_decode_dri(s)) < 0) 2619  return ret; 2620  break; 2621  case SOF5: 2622  case SOF6: 2623  case SOF7: 2624  case SOF9: 2625  case SOF10: 2626  case SOF11: 2627  case SOF13: 2628  case SOF14: 2629  case SOF15: 2630  case JPG: 2631  av_log(avctx, AV_LOG_ERROR, 2632  "mjpeg: unsupported coding type (%x)\n", start_code); 2633  break; 2634  } 2635  2636 skip: 2637  /* eof process start code */ 2638  buf_ptr += (get_bits_count(&s->gb) + 7) / 8; 2639  av_log(avctx, AV_LOG_DEBUG, 2640  "marker parser used %d bytes (%d bits)\n", 2641  (get_bits_count(&s->gb) + 7) / 8, get_bits_count(&s->gb)); 2642  } 2643  if (s->got_picture && s->cur_scan) { 2644  av_log(avctx, AV_LOG_WARNING, "EOI missing, emulating\n"); 2645  goto eoi_parser; 2646  } 2647  av_log(avctx, AV_LOG_FATAL, "No JPEG data found in image\n"); 2648  return AVERROR_INVALIDDATA; 2649 fail: 2650  s->got_picture = 0; 2651  return ret; 2652 the_end: 2653  2654  is16bit = av_pix_fmt_desc_get(s->avctx->pix_fmt)->comp[0].step > 1; 2655  2656  if (AV_RB32(s->upscale_h)) { 2657  int p; 2659  avctx->pix_fmt == AV_PIX_FMT_YUV444P || 2660  avctx->pix_fmt == AV_PIX_FMT_YUVJ440P || 2661  avctx->pix_fmt == AV_PIX_FMT_YUV440P || 2662  avctx->pix_fmt == AV_PIX_FMT_YUVA444P || 2663  avctx->pix_fmt == AV_PIX_FMT_YUVJ420P || 2664  avctx->pix_fmt == AV_PIX_FMT_YUV420P || 2665  avctx->pix_fmt == AV_PIX_FMT_YUV420P16|| 2666  avctx->pix_fmt == AV_PIX_FMT_YUVA420P || 2667  avctx->pix_fmt == AV_PIX_FMT_YUVA420P16|| 2668  avctx->pix_fmt == AV_PIX_FMT_GBRP || 2669  avctx->pix_fmt == AV_PIX_FMT_GBRAP 2670  ); 2671  ret = av_pix_fmt_get_chroma_sub_sample(s->avctx->pix_fmt, &hshift, &vshift); 2672  if (ret) 2673  return ret; 2674  2675  av_assert0(s->nb_components == av_pix_fmt_count_planes(s->picture_ptr->format)); 2676  for (p = 0; p<s->nb_components; p++) { 2677  uint8_t *line = s->picture_ptr->data[p]; 2678  int w = s->width; 2679  int h = s->height; 2680  if (!s->upscale_h[p]) 2681  continue; 2682  if (p==1 || p==2) { 2683  w = AV_CEIL_RSHIFT(w, hshift); 2684  h = AV_CEIL_RSHIFT(h, vshift); 2685  } 2686  if (s->upscale_v[p] == 1) 2687  h = (h+1)>>1; 2688  av_assert0(w > 0); 2689  for (i = 0; i < h; i++) { 2690  if (s->upscale_h[p] == 1) { 2691  if (is16bit) ((uint16_t*)line)[w - 1] = ((uint16_t*)line)[(w - 1) / 2]; 2692  else line[w - 1] = line[(w - 1) / 2]; 2693  for (index = w - 2; index > 0; index--) { 2694  if (is16bit) 2695  ((uint16_t*)line)[index] = (((uint16_t*)line)[index / 2] + ((uint16_t*)line)[(index + 1) / 2]) >> 1; 2696  else 2697  line[index] = (line[index / 2] + line[(index + 1) / 2]) >> 1; 2698  } 2699  } else if (s->upscale_h[p] == 2) { 2700  if (is16bit) { 2701  ((uint16_t*)line)[w - 1] = ((uint16_t*)line)[(w - 1) / 3]; 2702  if (w > 1) 2703  ((uint16_t*)line)[w - 2] = ((uint16_t*)line)[w - 1]; 2704  } else { 2705  line[w - 1] = line[(w - 1) / 3]; 2706  if (w > 1) 2707  line[w - 2] = line[w - 1]; 2708  } 2709  for (index = w - 3; index > 0; index--) { 2710  line[index] = (line[index / 3] + line[(index + 1) / 3] + line[(index + 2) / 3] + 1) / 3; 2711  } 2712  } 2713  line += s->linesize[p]; 2714  } 2715  } 2716  } 2717  if (AV_RB32(s->upscale_v)) { 2718  int p; 2720  avctx->pix_fmt == AV_PIX_FMT_YUV444P || 2721  avctx->pix_fmt == AV_PIX_FMT_YUVJ422P || 2722  avctx->pix_fmt == AV_PIX_FMT_YUV422P || 2723  avctx->pix_fmt == AV_PIX_FMT_YUVJ420P || 2724  avctx->pix_fmt == AV_PIX_FMT_YUV420P || 2725  avctx->pix_fmt == AV_PIX_FMT_YUV440P || 2726  avctx->pix_fmt == AV_PIX_FMT_YUVJ440P || 2727  avctx->pix_fmt == AV_PIX_FMT_YUVA444P || 2728  avctx->pix_fmt == AV_PIX_FMT_YUVA420P || 2729  avctx->pix_fmt == AV_PIX_FMT_YUVA420P16|| 2730  avctx->pix_fmt == AV_PIX_FMT_GBRP || 2731  avctx->pix_fmt == AV_PIX_FMT_GBRAP 2732  ); 2733  ret = av_pix_fmt_get_chroma_sub_sample(s->avctx->pix_fmt, &hshift, &vshift); 2734  if (ret) 2735  return ret; 2736  2737  av_assert0(s->nb_components == av_pix_fmt_count_planes(s->picture_ptr->format)); 2738  for (p = 0; p < s->nb_components; p++) { 2739  uint8_t *dst; 2740  int w = s->width; 2741  int h = s->height; 2742  if (!s->upscale_v[p]) 2743  continue; 2744  if (p==1 || p==2) { 2745  w = AV_CEIL_RSHIFT(w, hshift); 2746  h = AV_CEIL_RSHIFT(h, vshift); 2747  } 2748  dst = &((uint8_t *)s->picture_ptr->data[p])[(h - 1) * s->linesize[p]]; 2749  for (i = h - 1; i; i--) { 2750  uint8_t *src1 = &((uint8_t *)s->picture_ptr->data[p])[i * s->upscale_v[p] / (s->upscale_v[p] + 1) * s->linesize[p]]; 2751  uint8_t *src2 = &((uint8_t *)s->picture_ptr->data[p])[(i + 1) * s->upscale_v[p] / (s->upscale_v[p] + 1) * s->linesize[p]]; 2752  if (s->upscale_v[p] != 2 && (src1 == src2 || i == h - 1)) { 2753  memcpy(dst, src1, w); 2754  } else { 2755  for (index = 0; index < w; index++) 2756  dst[index] = (src1[index] + src2[index]) >> 1; 2757  } 2758  dst -= s->linesize[p]; 2759  } 2760  } 2761  } 2762  if (s->flipped && !s->rgb) { 2763  int j; 2764  ret = av_pix_fmt_get_chroma_sub_sample(s->avctx->pix_fmt, &hshift, &vshift); 2765  if (ret) 2766  return ret; 2767  2768  av_assert0(s->nb_components == av_pix_fmt_count_planes(s->picture_ptr->format)); 2769  for (index=0; index<s->nb_components; index++) { 2770  uint8_t *dst = s->picture_ptr->data[index]; 2771  int w = s->picture_ptr->width; 2772  int h = s->picture_ptr->height; 2773  if(index && index<3){ 2774  w = AV_CEIL_RSHIFT(w, hshift); 2775  h = AV_CEIL_RSHIFT(h, vshift); 2776  } 2777  if(dst){ 2778  uint8_t *dst2 = dst + s->picture_ptr->linesize[index]*(h-1); 2779  for (i=0; i<h/2; i++) { 2780  for (j=0; j<w; j++) 2781  FFSWAP(int, dst[j], dst2[j]); 2782  dst += s->picture_ptr->linesize[index]; 2783  dst2 -= s->picture_ptr->linesize[index]; 2784  } 2785  } 2786  } 2787  } 2788  if (s->adobe_transform == 0 && s->avctx->pix_fmt == AV_PIX_FMT_GBRAP) { 2789  int w = s->picture_ptr->width; 2790  int h = s->picture_ptr->height; 2791  av_assert0(s->nb_components == 4); 2792  for (i=0; i<h; i++) { 2793  int j; 2794  uint8_t *dst[4]; 2795  for (index=0; index<4; index++) { 2796  dst[index] = s->picture_ptr->data[index] 2797  + s->picture_ptr->linesize[index]*i; 2798  } 2799  for (j=0; j<w; j++) { 2800  int k = dst[3][j]; 2801  int r = dst[0][j] * k; 2802  int g = dst[1][j] * k; 2803  int b = dst[2][j] * k; 2804  dst[0][j] = g*257 >> 16; 2805  dst[1][j] = b*257 >> 16; 2806  dst[2][j] = r*257 >> 16; 2807  dst[3][j] = 255; 2808  } 2809  } 2810  } 2811  if (s->adobe_transform == 2 && s->avctx->pix_fmt == AV_PIX_FMT_YUVA444P) { 2812  int w = s->picture_ptr->width; 2813  int h = s->picture_ptr->height; 2814  av_assert0(s->nb_components == 4); 2815  for (i=0; i<h; i++) { 2816  int j; 2817  uint8_t *dst[4]; 2818  for (index=0; index<4; index++) { 2819  dst[index] = s->picture_ptr->data[index] 2820  + s->picture_ptr->linesize[index]*i; 2821  } 2822  for (j=0; j<w; j++) { 2823  int k = dst[3][j]; 2824  int r = (255 - dst[0][j]) * k; 2825  int g = (128 - dst[1][j]) * k; 2826  int b = (128 - dst[2][j]) * k; 2827  dst[0][j] = r*257 >> 16; 2828  dst[1][j] = (g*257 >> 16) + 128; 2829  dst[2][j] = (b*257 >> 16) + 128; 2830  dst[3][j] = 255; 2831  } 2832  } 2833  } 2834  2835  if (s->stereo3d) { 2837  if (stereo) { 2838  stereo->type = s->stereo3d->type; 2839  stereo->flags = s->stereo3d->flags; 2840  } 2841  av_freep(&s->stereo3d); 2842  } 2843  2844  if (s->iccnum != 0 && s->iccnum == s->iccread) { 2845  AVFrameSideData *sd; 2846  size_t offset = 0; 2847  int total_size = 0; 2848  int i; 2849  2850  /* Sum size of all parts. */ 2851  for (i = 0; i < s->iccnum; i++) 2852  total_size += s->iccentries[i].length; 2853  2855  if (!sd) { 2856  av_log(s->avctx, AV_LOG_ERROR, "Could not allocate frame side data\n"); 2857  return AVERROR(ENOMEM); 2858  } 2859  2860  /* Reassemble the parts, which are now in-order. */ 2861  for (i = 0; i < s->iccnum; i++) { 2862  memcpy(sd->data + offset, s->iccentries[i].data, s->iccentries[i].length); 2863  offset += s->iccentries[i].length; 2864  } 2865  } 2866  2867  if (e = av_dict_get(s->exif_metadata, "Orientation", e, AV_DICT_IGNORE_SUFFIX)) { 2868  char *value = e->value + strspn(e->value, " \n\t\r"), *endptr; 2869  int orientation = strtol(value, &endptr, 0); 2870  2871  if (!*endptr) { 2872  AVFrameSideData *sd = NULL; 2873  2874  if (orientation >= 2 && orientation <= 8) { 2875  int32_t *matrix; 2876  2878  if (!sd) { 2879  av_log(s->avctx, AV_LOG_ERROR, "Could not allocate frame side data\n"); 2880  return AVERROR(ENOMEM); 2881  } 2882  2883  matrix = (int32_t *)sd->data; 2884  2885  switch (orientation) { 2886  case 2: 2887  av_display_rotation_set(matrix, 0.0); 2888  av_display_matrix_flip(matrix, 1, 0); 2889  break; 2890  case 3: 2891  av_display_rotation_set(matrix, 180.0); 2892  break; 2893  case 4: 2894  av_display_rotation_set(matrix, 180.0); 2895  av_display_matrix_flip(matrix, 1, 0); 2896  break; 2897  case 5: 2898  av_display_rotation_set(matrix, 90.0); 2899  av_display_matrix_flip(matrix, 1, 0); 2900  break; 2901  case 6: 2902  av_display_rotation_set(matrix, 90.0); 2903  break; 2904  case 7: 2905  av_display_rotation_set(matrix, -90.0); 2906  av_display_matrix_flip(matrix, 1, 0); 2907  break; 2908  case 8: 2909  av_display_rotation_set(matrix, -90.0); 2910  break; 2911  default: 2912  av_assert0(0); 2913  } 2914  } 2915  } 2916  } 2917  2918  av_dict_copy(&frame->metadata, s->exif_metadata, 0); 2919  av_dict_free(&s->exif_metadata); 2920  2921  if (avctx->codec_id == AV_CODEC_ID_SMVJPEG) { 2922  ret = smv_process_frame(avctx, frame); 2923  if (ret < 0) { 2925  return ret; 2926  } 2927  } 2928  if ((avctx->codec_tag == MKTAG('A', 'V', 'R', 'n') || 2929  avctx->codec_tag == MKTAG('A', 'V', 'D', 'J')) && 2930  avctx->coded_height > s->orig_height) { 2931  frame->height = AV_CEIL_RSHIFT(avctx->coded_height, avctx->lowres); 2932  frame->crop_top = frame->height - avctx->height; 2933  } 2934  2935  ret = 0; 2936  2937 the_end_no_picture: 2938  av_log(avctx, AV_LOG_DEBUG, "decode frame unused %"PTRDIFF_SPECIFIER" bytes\n", 2939  buf_end - buf_ptr); 2940  2941  return ret; 2942 } 2943  2944 /* mxpeg may call the following function (with a blank MJpegDecodeContext) 2945  * even without having called ff_mjpeg_decode_init(). */ 2947 { 2948  MJpegDecodeContext *s = avctx->priv_data; 2949  int i, j; 2950  2951  if (s->interlaced && s->bottom_field == !s->interlace_polarity && s->got_picture && !avctx->frame_number) { 2952  av_log(avctx, AV_LOG_INFO, "Single field\n"); 2953  } 2954  2955  if (s->picture) { 2956  av_frame_free(&s->picture); 2957  s->picture_ptr = NULL; 2958  } else if (s->picture_ptr) 2959  av_frame_unref(s->picture_ptr); 2960  2961  av_frame_free(&s->smv_frame); 2962  2963  av_freep(&s->buffer); 2964  av_freep(&s->stereo3d); 2965  av_freep(&s->ljpeg_buffer); 2966  s->ljpeg_buffer_size = 0; 2967  2968  for (i = 0; i < 3; i++) { 2969  for (j = 0; j < 4; j++) 2970  ff_free_vlc(&s->vlcs[i][j]); 2971  } 2972  for (i = 0; i < MAX_COMPONENTS; i++) { 2973  av_freep(&s->blocks[i]); 2974  av_freep(&s->last_nnz[i]); 2975  } 2976  av_dict_free(&s->exif_metadata); 2977  2979  2980  av_freep(&s->hwaccel_picture_private); 2981  av_freep(&s->jls_state); 2982  2983  return 0; 2984 } 2985  2986 static void decode_flush(AVCodecContext *avctx) 2987 { 2988  MJpegDecodeContext *s = avctx->priv_data; 2989  s->got_picture = 0; 2990  2991  s->smv_next_frame = 0; 2992  av_frame_unref(s->smv_frame); 2993 } 2994  2995 #if CONFIG_MJPEG_DECODER 2996 #define OFFSET(x) offsetof(MJpegDecodeContext, x) 2997 #define VD AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_DECODING_PARAM 2998 static const AVOption options[] = { 2999  { "extern_huff", "Use external huffman table.", 3000  OFFSET(extern_huff), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, VD }, 3001  { NULL }, 3002 }; 3003  3004 static const AVClass mjpegdec_class = { 3005  .class_name = "MJPEG decoder", 3006  .item_name = av_default_item_name, 3007  .option = options, 3008  .version = LIBAVUTIL_VERSION_INT, 3009 }; 3010  3011 const AVCodec ff_mjpeg_decoder = { 3012  .name = "mjpeg", 3013  .long_name = NULL_IF_CONFIG_SMALL("MJPEG (Motion JPEG)"), 3014  .type = AVMEDIA_TYPE_VIDEO, 3015  .id = AV_CODEC_ID_MJPEG, 3016  .priv_data_size = sizeof(MJpegDecodeContext), 3018  .close = ff_mjpeg_decode_end, 3020  .flush = decode_flush, 3021  .capabilities = AV_CODEC_CAP_DR1, 3022  .max_lowres = 3, 3023  .priv_class = &mjpegdec_class, 3027  .hw_configs = (const AVCodecHWConfigInternal *const []) { 3028 #if CONFIG_MJPEG_NVDEC_HWACCEL 3029  HWACCEL_NVDEC(mjpeg), 3030 #endif 3031 #if CONFIG_MJPEG_VAAPI_HWACCEL 3032  HWACCEL_VAAPI(mjpeg), 3033 #endif 3034  NULL 3035  }, 3036 }; 3037 #endif 3038 #if CONFIG_THP_DECODER 3039 const AVCodec ff_thp_decoder = { 3040  .name = "thp", 3041  .long_name = NULL_IF_CONFIG_SMALL("Nintendo Gamecube THP video"), 3042  .type = AVMEDIA_TYPE_VIDEO, 3043  .id = AV_CODEC_ID_THP, 3044  .priv_data_size = sizeof(MJpegDecodeContext), 3046  .close = ff_mjpeg_decode_end, 3048  .flush = decode_flush, 3049  .capabilities = AV_CODEC_CAP_DR1, 3050  .max_lowres = 3, 3053 }; 3054 #endif 3055  3056 #if CONFIG_SMVJPEG_DECODER 3057 const AVCodec ff_smvjpeg_decoder = { 3058  .name = "smvjpeg", 3059  .long_name = NULL_IF_CONFIG_SMALL("SMV JPEG"), 3060  .type = AVMEDIA_TYPE_VIDEO, 3061  .id = AV_CODEC_ID_SMVJPEG, 3062  .priv_data_size = sizeof(MJpegDecodeContext), 3064  .close = ff_mjpeg_decode_end, 3066  .flush = decode_flush, 3067  .capabilities = AV_CODEC_CAP_DR1, 3070 }; 3071 #endif FF_ALLOCZ_TYPED_ARRAY #define FF_ALLOCZ_TYPED_ARRAY(p, nelem) Definition: internal.h:98 hwconfig.h av_packet_unref void av_packet_unref(AVPacket *pkt) Wipe the packet. Definition: avpacket.c:424 AVCodecContext::hwaccel const struct AVHWAccel * hwaccel Hardware accelerator in use. Definition: avcodec.h:1359 AVCodec AVCodec. Definition: codec.h:202 skip_bits_long static void skip_bits_long(GetBitContext *s, int n) Skips the specified number of bits. Definition: get_bits.h:292 ff_decode_get_packet int ff_decode_get_packet(AVCodecContext *avctx, AVPacket *pkt) Called by decoders to get the next packet for decoding. Definition: decode.c:224 AV_LOG_WARNING #define AV_LOG_WARNING Something somehow does not look correct. Definition: log.h:186 FF_CODEC_CAP_INIT_THREADSAFE #define FF_CODEC_CAP_INIT_THREADSAFE The codec does not modify any global variables in the init function, allowing to call the init functi... Definition: internal.h:42 AV_PIX_FMT_CUDA @ AV_PIX_FMT_CUDA HW acceleration through CUDA. Definition: pixfmt.h:225 ff_init_scantable av_cold void ff_init_scantable(uint8_t *permutation, ScanTable *st, const uint8_t *src_scantable) Definition: idctdsp.c:29 AVPixelFormat AVPixelFormat Pixel format. Definition: pixfmt.h:64 jpegtables.h FF_CODEC_CAP_SETS_PKT_DTS #define FF_CODEC_CAP_SETS_PKT_DTS Decoders marked with FF_CODEC_CAP_SETS_PKT_DTS want to set AVFrame.pkt_dts manually. Definition: internal.h:57 mjpeg.h level uint8_t level Definition: svq3.c:204 show_bits_long static unsigned int show_bits_long(GetBitContext *s, int n) Show 0-32 bits. Definition: get_bits.h:603 blockdsp.h get_bits_left static int get_bits_left(GetBitContext *gb) Definition: get_bits.h:850 r const char * r Definition: vf_curves.c:116 AVERROR Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions opt.h AVCodecContext::colorspace enum AVColorSpace colorspace YUV colorspace type. Definition: avcodec.h:960 ff_get_format int ff_get_format(AVCodecContext *avctx, const enum AVPixelFormat *fmt) Select the (possibly hardware accelerated) pixel format. Definition: decode.c:1089 out FILE * out Definition: movenc.c:54 SOS @ SOS Definition: mjpeg.h:72 mjpeg_copy_block static av_always_inline void mjpeg_copy_block(MJpegDecodeContext *s, uint8_t *dst, const uint8_t *src, int linesize, int lowres) Definition: mjpegdec.c:1391 is The official guide to swscale for confused that is Definition: swscale.txt:28 SOF48 @ SOF48 JPEG-LS. Definition: mjpeg.h:103 av_frame_new_side_data AVFrameSideData * av_frame_new_side_data(AVFrame *frame, enum AVFrameSideDataType type, size_t size) Add a new side data to a frame. Definition: frame.c:605 GetByteContext Definition: bytestream.h:33 APP1 @ APP1 Definition: mjpeg.h:80 decode_flush static void decode_flush(AVCodecContext *avctx) Definition: mjpegdec.c:2986 av_pix_fmt_desc_get const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt) Definition: pixdesc.c:2660 ZERO_RUN #define ZERO_RUN Definition: mjpegdec.c:953 SOF0 @ SOF0 Definition: mjpeg.h:39 AVCodecContext::err_recognition int err_recognition Error recognition; may misdetect some more or less valid parts as errors. Definition: avcodec.h:1324 GET_VLC #define GET_VLC(code, name, gb, table, bits, max_depth) If the vlc code is invalid and max_depth=1, then no bits will be removed. Definition: get_bits.h:707 get_bits_long static unsigned int get_bits_long(GetBitContext *s, int n) Read 0-32 bits. Definition: get_bits.h:547 init_put_bits static void init_put_bits(PutBitContext *s, uint8_t *buffer, int buffer_size) Initialize the PutBitContext s. Definition: put_bits.h:61 se #define se(name, range_min, range_max) Definition: cbs_h2645.c:275 get_bits_count static int get_bits_count(const GetBitContext *s) Definition: get_bits.h:220 init_idct static void init_idct(AVCodecContext *avctx) Definition: mjpegdec.c:109 RST7 @ RST7 Definition: mjpeg.h:68 av_frame_free void av_frame_free(AVFrame **frame) Free the frame and any dynamically allocated objects in it, e.g. Definition: frame.c:109 mjpegdec.h start_code static const uint8_t start_code[] Definition: videotoolboxenc.c:192 bytestream2_seek static av_always_inline int bytestream2_seek(GetByteContext *g, int offset, int whence) Definition: bytestream.h:212 AVFrame This structure describes decoded (raw) audio or video data. Definition: frame.h:317 put_bits static void put_bits(Jpeg2000EncoderContext *s, int val, int n) put n times val bit Definition: j2kenc.c:220 index fg index Definition: ffmpeg_filter.c:167 AVFrame::width int width Definition: frame.h:389 AV_PIX_FMT_YUVA420P16 #define AV_PIX_FMT_YUVA420P16 Definition: pixfmt.h:446 w uint8_t w Definition: llviddspenc.c:38 AVCOL_RANGE_JPEG @ AVCOL_RANGE_JPEG Full range content. Definition: pixfmt.h:597 FF_PROFILE_MJPEG_JPEG_LS #define FF_PROFILE_MJPEG_JPEG_LS Definition: avcodec.h:1629 smv_process_frame static int smv_process_frame(AVCodecContext *avctx, AVFrame *frame) Definition: mjpegdec.c:2339 av_display_matrix_flip void av_display_matrix_flip(int32_t matrix[9], int hflip, int vflip) Flip the input matrix horizontally and/or vertically. Definition: display.c:65 internal.h SOF11 @ SOF11 Definition: mjpeg.h:50 AVCodecContext::field_order enum AVFieldOrder field_order Field order. Definition: avcodec.h:989 AVOption AVOption. Definition: opt.h:247 b #define b Definition: input.c:40 mjpeg_decode_dc static int mjpeg_decode_dc(MJpegDecodeContext *s, int dc_index) Definition: mjpegdec.c:787 jpeglsdec.h data const char data[16] Definition: mxf.c:143 AVComponentDescriptor::step int step Number of elements between 2 horizontally consecutive pixels. Definition: pixdesc.h:40 AV_DICT_IGNORE_SUFFIX #define AV_DICT_IGNORE_SUFFIX Return first entry in a dictionary whose first part corresponds to the search key,... Definition: dict.h:68 ff_mjpeg_val_dc const uint8_t ff_mjpeg_val_dc[] Definition: jpegtabs.h:34 get_vlc2 static av_always_inline int get_vlc2(GetBitContext *s, VLC_TYPE(*table)[2], int bits, int max_depth) Parse a vlc code. Definition: get_bits.h:798 mjpeg_get_packet static int mjpeg_get_packet(AVCodecContext *avctx) Definition: mjpegdec.c:2373 AV_PIX_FMT_BGR24 @ AV_PIX_FMT_BGR24 packed RGB 8:8:8, 24bpp, BGRBGR... Definition: pixfmt.h:69 av_display_rotation_set void av_display_rotation_set(int32_t matrix[9], double angle) Initialize a transformation matrix describing a pure clockwise rotation by the specified angle (in de... Definition: display.c:50 AV_FRAME_DATA_DISPLAYMATRIX @ AV_FRAME_DATA_DISPLAYMATRIX This side data contains a 3x3 transformation matrix describing an affine transformation that needs to... Definition: frame.h:84 AV_PIX_FMT_YUV440P @ AV_PIX_FMT_YUV440P planar YUV 4:4:0 (1 Cr & Cb sample per 1x2 Y samples) Definition: pixfmt.h:99 UPDATE_CACHE #define UPDATE_CACHE(name, gb) Definition: get_bits.h:179 FFMAX #define FFMAX(a, b) Definition: macros.h:47 ff_mjpeg_bits_ac_chrominance const uint8_t ff_mjpeg_bits_ac_chrominance[] Definition: jpegtabs.h:66 AV_CODEC_ID_THP @ AV_CODEC_ID_THP Definition: codec_id.h:150 init_get_bits static int init_get_bits(GetBitContext *s, const uint8_t *buffer, int bit_size) Initialize GetBitContext. Definition: get_bits.h:660 FF_DEBUG_PICT_INFO #define FF_DEBUG_PICT_INFO Definition: avcodec.h:1303 AVFrame::data uint8_t * data[AV_NUM_DATA_POINTERS] pointer to the picture/channel planes. Definition: frame.h:338 av_malloc #define av_malloc(s) Definition: tableprint_vlc.h:31 APP15 @ APP15 Definition: mjpeg.h:94 init static int init Definition: av_tx.c:47 GET_CACHE #define GET_CACHE(name, gb) Definition: get_bits.h:216 skip_bits static void skip_bits(GetBitContext *s, int n) Definition: get_bits.h:468 AV_STEREO3D_SIDEBYSIDE @ AV_STEREO3D_SIDEBYSIDE Views are next to each other. Definition: stereo3d.h:67 av_pix_fmt_count_planes int av_pix_fmt_count_planes(enum AVPixelFormat pix_fmt) Definition: pixdesc.c:2700 AVCOL_SPC_BT470BG @ AVCOL_SPC_BT470BG also ITU-R BT601-6 625 / ITU-R BT1358 625 / ITU-R BT1700 625 PAL & SECAM / IEC 61966-2-4 xvYCC601 Definition: pixfmt.h:529 get_bits static unsigned int get_bits(GetBitContext *s, int n) Read 1-25 bits. Definition: get_bits.h:380 rgb Definition: rpzaenc.c:59 ff_mjpeg_decode_dht int ff_mjpeg_decode_dht(MJpegDecodeContext *s) Definition: mjpegdec.c:237 ljpeg_decode_yuv_scan static int ljpeg_decode_yuv_scan(MJpegDecodeContext *s, int predictor, int point_transform, int nb_components) Definition: mjpegdec.c:1237 shift_output static void shift_output(MJpegDecodeContext *s, uint8_t *ptr, int linesize) Definition: mjpegdec.c:1407 AV_PIX_FMT_GBRAP @ AV_PIX_FMT_GBRAP planar GBRA 4:4:4:4 32bpp Definition: pixfmt.h:205 AVCodecContext::codec const struct AVCodec * codec Definition: avcodec.h:392 ff_mjpeg_decode_init av_cold int ff_mjpeg_decode_init(AVCodecContext *avctx) Definition: mjpegdec.c:118 AVCodecContext::skip_frame enum AVDiscard skip_frame Skip decoding for selected frames. Definition: avcodec.h:1673 fail #define fail() Definition: checkasm.h:127 AV_STEREO3D_2D @ AV_STEREO3D_2D Video is not stereoscopic (and metadata has to be there). Definition: stereo3d.h:55 AV_PIX_FMT_YUVA444P16 #define AV_PIX_FMT_YUVA444P16 Definition: pixfmt.h:448 SOF3 @ SOF3 Definition: mjpeg.h:42 FF_PROFILE_MJPEG_HUFFMAN_BASELINE_DCT #define FF_PROFILE_MJPEG_HUFFMAN_BASELINE_DCT Definition: avcodec.h:1625 GetBitContext Definition: get_bits.h:62 mjpeg_decode_com static int mjpeg_decode_com(MJpegDecodeContext *s) Definition: mjpegdec.c:2138 init_default_huffman_tables static int init_default_huffman_tables(MJpegDecodeContext *s) Definition: mjpegdec.c:55 AVCodecContext::flags int flags AV_CODEC_FLAG_*. Definition: avcodec.h:463 val static double val(void *priv, double ch) Definition: aeval.c:76 av_pix_fmt_get_chroma_sub_sample int av_pix_fmt_get_chroma_sub_sample(enum AVPixelFormat pix_fmt, int *h_shift, int *v_shift) Utility function to access log2_chroma_w log2_chroma_h from the pixel format AVPixFmtDescriptor. Definition: pixdesc.c:2688 type it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf type Definition: writing_filters.txt:86 AVCodecContext::coded_height int coded_height Definition: avcodec.h:571 AV_PIX_FMT_GRAY16 #define AV_PIX_FMT_GRAY16 Definition: pixfmt.h:388 ff_sp5x_process_packet int ff_sp5x_process_packet(AVCodecContext *avctx, AVPacket *avpkt) Definition: sp5xdec.c:33 ss #define ss(width, name, subs,...) Definition: cbs_vp9.c:261 FF_CODEC_CAP_SKIP_FRAME_FILL_PARAM #define FF_CODEC_CAP_SKIP_FRAME_FILL_PARAM The decoder extracts and fills its parameters even if the frame is skipped due to the skip_frame sett... Definition: internal.h:62 av_frame_alloc AVFrame * av_frame_alloc(void) Allocate an AVFrame and set its fields to default values. Definition: frame.c:97 av_bswap32 #define av_bswap32 Definition: bswap.h:33 AV_PIX_FMT_YUVJ411P @ AV_PIX_FMT_YUVJ411P planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples) full scale (JPEG), deprecated in favor ... Definition: pixfmt.h:248 ff_mjpeg_profiles const AVProfile ff_mjpeg_profiles[] Definition: profiles.c:169 ff_exif_decode_ifd int ff_exif_decode_ifd(void *logctx, GetByteContext *gbytes, int le, int depth, AVDictionary **metadata) Definition: exif.c:115 aligned static int aligned(int val) Definition: dashdec.c:169 avassert.h AV_LOG_ERROR #define AV_LOG_ERROR Something went wrong and cannot losslessly be recovered. Definition: log.h:180 FF_ARRAY_ELEMS #define FF_ARRAY_ELEMS(a) Definition: sinewin_tablegen.c:29 av_cold #define av_cold Definition: attributes.h:90 decode_dc_progressive static int decode_dc_progressive(MJpegDecodeContext *s, int16_t *block, int component, int dc_index, uint16_t *quant_matrix, int Al) Definition: mjpegdec.c:854 AV_PIX_FMT_YUV422P16 #define AV_PIX_FMT_YUV422P16 Definition: pixfmt.h:416 init_get_bits8 static int init_get_bits8(GetBitContext *s, const uint8_t *buffer, int byte_size) Initialize GetBitContext. Definition: get_bits.h:678 FF_CODEC_PROPERTY_LOSSLESS #define FF_CODEC_PROPERTY_LOSSLESS Definition: avcodec.h:1823 COM @ COM Definition: mjpeg.h:111 FF_PROFILE_MJPEG_HUFFMAN_PROGRESSIVE_DCT #define FF_PROFILE_MJPEG_HUFFMAN_PROGRESSIVE_DCT Definition: avcodec.h:1627 mask static const uint16_t mask[17] Definition: lzw.c:38 av_dict_get AVDictionaryEntry * av_dict_get(const AVDictionary *m, const char *key, const AVDictionaryEntry *prev, int flags) Get a dictionary entry with matching key. Definition: dict.c:40 handle_rstn static int handle_rstn(MJpegDecodeContext *s, int nb_components) Definition: mjpegdec.c:1036 AV_PIX_FMT_YUVJ422P @ AV_PIX_FMT_YUVJ422P planar YUV 4:2:2, 16bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV422P and setting col... Definition: pixfmt.h:79 CLOSE_READER #define CLOSE_READER(name, gb) Definition: get_bits.h:150 SOF5 @ SOF5 Definition: mjpeg.h:44 AVCodecContext::extradata_size int extradata_size Definition: avcodec.h:485 width #define width AV_STEREO3D_LINES @ AV_STEREO3D_LINES Views are packed per line, as if interlaced. Definition: stereo3d.h:129 s #define s(width, name) Definition: cbs_vp9.c:257 AV_PIX_FMT_YUVA420P @ AV_PIX_FMT_YUVA420P planar YUV 4:2:0, 20bpp, (1 Cr & Cb sample per 2x2 Y & A samples) Definition: pixfmt.h:101 parse_avid static void parse_avid(MJpegDecodeContext *s, uint8_t *buf, int len) Definition: mjpegdec.c:98 AV_PIX_FMT_YUV444P16 #define AV_PIX_FMT_YUV444P16 Definition: pixfmt.h:417 AV_CEIL_RSHIFT #define AV_CEIL_RSHIFT(a, b) Definition: common.h:51 FF_PROFILE_MJPEG_HUFFMAN_LOSSLESS #define FF_PROFILE_MJPEG_HUFFMAN_LOSSLESS Definition: avcodec.h:1628 g const char * g Definition: vf_curves.c:117 APP3 @ APP3 Definition: mjpeg.h:82 AV_GET_BUFFER_FLAG_REF #define AV_GET_BUFFER_FLAG_REF The decoder will keep a reference to the frame and may reuse it later. Definition: avcodec.h:361 ff_jpegls_decode_picture int ff_jpegls_decode_picture(MJpegDecodeContext *s, int near, int point_transform, int ilv) Definition: jpeglsdec.c:355 bits uint8_t bits Definition: vp3data.h:141 av_assert0 #define av_assert0(cond) assert() equivalent, that is always enabled. Definition: avassert.h:37 pix_fmts static enum AVPixelFormat pix_fmts[] Definition: libkvazaar.c:296 AV_FIELD_UNKNOWN @ AV_FIELD_UNKNOWN Definition: codec_par.h:37 AV_LOG_DEBUG #define AV_LOG_DEBUG Stuff which is only useful for libav* developers. Definition: log.h:201 AV_PIX_FMT_GBR24P @ AV_PIX_FMT_GBR24P Definition: pixfmt.h:159 AV_PIX_FMT_YUV420P16 #define AV_PIX_FMT_YUV420P16 Definition: pixfmt.h:415 ff_thp_decoder const AVCodec ff_thp_decoder RST0 @ RST0 Definition: mjpeg.h:61 decode.h reset_icc_profile static void reset_icc_profile(MJpegDecodeContext *s) Definition: mjpegdec.c:2323 ff_mjpeg_decode_end av_cold int ff_mjpeg_decode_end(AVCodecContext *avctx) Definition: mjpegdec.c:2946 ff_free_vlc void ff_free_vlc(VLC *vlc) Definition: bitstream.c:431 AV_PIX_FMT_YUV420P @ AV_PIX_FMT_YUV420P planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples) Definition: pixfmt.h:66 PutBitContext Definition: put_bits.h:49 AV_PIX_FMT_YUVJ444P @ AV_PIX_FMT_YUVJ444P planar YUV 4:4:4, 24bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV444P and setting col... Definition: pixfmt.h:80 ff_hpeldsp_init av_cold void ff_hpeldsp_init(HpelDSPContext *c, int flags) Definition: hpeldsp.c:338 AVCodecContext::codec_id enum AVCodecID codec_id Definition: avcodec.h:393 AVStereo3D::flags int flags Additional information about the frame packing. Definition: stereo3d.h:185 if if(ret) Definition: filter_design.txt:179 ff_mjpeg_receive_frame int ff_mjpeg_receive_frame(AVCodecContext *avctx, AVFrame *frame) Definition: mjpegdec.c:2397 AVDISCARD_ALL @ AVDISCARD_ALL discard all Definition: defs.h:54 AV_PIX_FMT_GBRP16 #define AV_PIX_FMT_GBRP16 Definition: pixfmt.h:423 AV_PIX_FMT_RGBA64 #define AV_PIX_FMT_RGBA64 Definition: pixfmt.h:394 LIBAVUTIL_VERSION_INT #define LIBAVUTIL_VERSION_INT Definition: version.h:85 AV_CODEC_ID_SP5X @ AV_CODEC_ID_SP5X Definition: codec_id.h:60 AVClass Describe the class of an AVClass context structure. Definition: log.h:66 av_clip_int16 #define av_clip_int16 Definition: common.h:111 PTRDIFF_SPECIFIER #define PTRDIFF_SPECIFIER Definition: internal.h:192 ff_smvjpeg_decoder const AVCodec ff_smvjpeg_decoder AV_PIX_FMT_BGR48 #define AV_PIX_FMT_BGR48 Definition: pixfmt.h:395 flush static void flush(AVCodecContext *avctx) Definition: aacdec_template.c:593 NULL #define NULL Definition: coverity.c:32 mjpeg_idct_scan_progressive_ac static void mjpeg_idct_scan_progressive_ac(MJpegDecodeContext *s) Definition: mjpegdec.c:1598 copy_block2 static void copy_block2(uint8_t *dst, const uint8_t *src, ptrdiff_t dstStride, ptrdiff_t srcStride, int h) Definition: copy_block.h:27 AVERROR_PATCHWELCOME #define AVERROR_PATCHWELCOME Not yet implemented in FFmpeg, patches welcome. Definition: error.h:64 run uint8_t run Definition: svq3.c:203 AVRational Rational number (pair of numerator and denominator). Definition: rational.h:58 ff_mjpeg_decode_dqt int ff_mjpeg_decode_dqt(MJpegDecodeContext *s) Definition: mjpegdec.c:194 SOF13 @ SOF13 Definition: mjpeg.h:52 AVCodecContext::internal struct AVCodecInternal * internal Private context used for internal data. Definition: avcodec.h:418 receive_frame static CopyRet receive_frame(AVCodecContext *avctx, AVFrame *frame, int *got_frame) Definition: crystalhd.c:559 AV_PIX_FMT_YUVJ420P @ AV_PIX_FMT_YUVJ420P planar YUV 4:2:0, 12bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV420P and setting col... Definition: pixfmt.h:78 tiff.h av_default_item_name const char * av_default_item_name(void *ptr) Return the context name. Definition: log.c:235 get_bits1 static unsigned int get_bits1(GetBitContext *s) Definition: get_bits.h:499 AV_PICTURE_TYPE_I @ AV_PICTURE_TYPE_I Intra. Definition: avutil.h:274 profiles.h src #define src Definition: vp8dsp.c:255 AV_FRAME_DATA_ICC_PROFILE @ AV_FRAME_DATA_ICC_PROFILE The data contains an ICC profile as an opaque octet buffer following the format described by ISO 1507... Definition: frame.h:143 LAST_SKIP_BITS #define LAST_SKIP_BITS(name, gb, num) Definition: get_bits.h:200 MJpegDecodeContext Definition: mjpegdec.h:54 mjpeg_decode_scan static int mjpeg_decode_scan(MJpegDecodeContext *s, int nb_components, int Ah, int Al, const uint8_t *mb_bitmask, int mb_bitmask_size, const AVFrame *reference) Definition: mjpegdec.c:1422 decode_block_refinement static int decode_block_refinement(MJpegDecodeContext *s, int16_t *block, uint8_t *last_nnz, int ac_index, uint16_t *quant_matrix, int ss, int se, int Al, int *EOBRUN) Definition: mjpegdec.c:971 lowres static int lowres Definition: ffplay.c:334 mjpeg_decode_scan_progressive_ac static int mjpeg_decode_scan_progressive_ac(MJpegDecodeContext *s, int ss, int se, int Ah, int Al) Definition: mjpegdec.c:1543 ff_mjpeg_val_ac_chrominance const uint8_t ff_mjpeg_val_ac_chrominance[] Definition: jpegtabs.h:69 AV_PIX_FMT_GRAY8 @ AV_PIX_FMT_GRAY8 Y , 8bpp. Definition: pixfmt.h:74 AV_EF_EXPLODE #define AV_EF_EXPLODE abort decoding on minor error detection Definition: avcodec.h:1335 FF_CODEC_CAP_EXPORTS_CROPPING #define FF_CODEC_CAP_EXPORTS_CROPPING The decoder sets the cropping fields in the output frames manually. Definition: internal.h:68 AV_PIX_FMT_ABGR @ AV_PIX_FMT_ABGR packed ABGR 8:8:8:8, 32bpp, ABGRABGR... Definition: pixfmt.h:94 DRI @ DRI Definition: mjpeg.h:75 c Undefined Behavior In the C some operations are like signed integer dereferencing freed accessing outside allocated Undefined Behavior must not occur in a C it is not safe even if the output of undefined operations is unused The unsafety may seem nit picking but Optimizing compilers have in fact optimized code on the assumption that no undefined Behavior occurs Optimizing code based on wrong assumptions can and has in some cases lead to effects beyond the output of computations The signed integer overflow problem in speed critical code Code which is highly optimized and works with signed integers sometimes has the problem that often the output of the computation does not c Definition: undefined.txt:32 bytestream2_tell static av_always_inline int bytestream2_tell(GetByteContext *g) Definition: bytestream.h:192 for for(j=16;j >0;--j) Definition: h264pred_template.c:469 ff_dlog #define ff_dlog(a,...) Definition: tableprint_vlc.h:29 copy_data_segment #define copy_data_segment(skip) AVCodecContext::lowres int lowres low resolution decoding, 1-> 1/2 size, 2->1/4 size Definition: avcodec.h:1432 options const OptionDef options[] copy_mb static void copy_mb(CinepakEncContext *s, uint8_t *a_data[4], int a_linesize[4], uint8_t *b_data[4], int b_linesize[4]) Definition: cinepakenc.c:508 ff_get_buffer int ff_get_buffer(AVCodecContext *avctx, AVFrame *frame, int flags) Get a buffer for a frame. Definition: decode.c:1652 AV_PIX_FMT_RGB24 @ AV_PIX_FMT_RGB24 packed RGB 8:8:8, 24bpp, RGBRGB... Definition: pixfmt.h:68 AV_CODEC_CAP_DR1 #define AV_CODEC_CAP_DR1 Codec uses get_buffer() or get_encode_buffer() for allocating buffers and supports custom allocators. Definition: codec.h:52 ljpeg_decode_rgb_scan static int ljpeg_decode_rgb_scan(MJpegDecodeContext *s, int nb_components, int predictor, int point_transform) Definition: mjpegdec.c:1071 ff_mjpeg_val_ac_luminance const uint8_t ff_mjpeg_val_ac_luminance[] Definition: jpegtabs.h:42 dc Tag MUST be and< 10hcoeff half pel interpolation filter coefficients, hcoeff[0] are the 2 middle coefficients[1] are the next outer ones and so on, resulting in a filter like:...eff[2], hcoeff[1], hcoeff[0], hcoeff[0], hcoeff[1], hcoeff[2] ... the sign of the coefficients is not explicitly stored but alternates after each coeff and coeff[0] is positive, so ...,+,-,+,-,+,+,-,+,-,+,... hcoeff[0] is not explicitly stored but found by subtracting the sum of all stored coefficients with signs from 32 hcoeff[0]=32 - hcoeff[1] - hcoeff[2] - ... a good choice for hcoeff and htaps is htaps=6 hcoeff={40,-10, 2} an alternative which requires more computations at both encoder and decoder side and may or may not be better is htaps=8 hcoeff={42,-14, 6,-2}ref_frames minimum of the number of available reference frames and max_ref_frames for example the first frame after a key frame always has ref_frames=1spatial_decomposition_type wavelet type 0 is a 9/7 symmetric compact integer wavelet 1 is a 5/3 symmetric compact integer wavelet others are reserved stored as delta from last, last is reset to 0 if always_reset||keyframeqlog quality(logarithmic quantizer scale) stored as delta from last, last is reset to 0 if always_reset||keyframemv_scale stored as delta from last, last is reset to 0 if always_reset||keyframe FIXME check that everything works fine if this changes between framesqbias dequantization bias stored as delta from last, last is reset to 0 if always_reset||keyframeblock_max_depth maximum depth of the block tree stored as delta from last, last is reset to 0 if always_reset||keyframequant_table quantization tableHighlevel bitstream structure:==============================--------------------------------------------|Header|--------------------------------------------|------------------------------------|||Block0||||split?||||yes no||||......... intra?||||:Block01 :yes no||||:Block02 :....... ..........||||:Block03 ::y DC ::ref index:||||:Block04 ::cb DC ::motion x :||||......... :cr DC ::motion y :||||....... ..........|||------------------------------------||------------------------------------|||Block1|||...|--------------------------------------------|------------ ------------ ------------|||Y subbands||Cb subbands||Cr subbands||||--- ---||--- ---||--- ---|||||LL0||HL0||||LL0||HL0||||LL0||HL0|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||LH0||HH0||||LH0||HH0||||LH0||HH0|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||HL1||LH1||||HL1||LH1||||HL1||LH1|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||HH1||HL2||||HH1||HL2||||HH1||HL2|||||...||...||...|||------------ ------------ ------------|--------------------------------------------Decoding process:=================------------|||Subbands|------------||||------------|Intra DC||||LL0 subband prediction ------------|\ Dequantization ------------------- \||Reference frames|\ IDWT|------- -------|Motion \|||Frame 0||Frame 1||Compensation . OBMC v -------|------- -------|--------------. \------> Frame n output Frame Frame<----------------------------------/|...|------------------- Range Coder:============Binary Range Coder:------------------- The implemented range coder is an adapted version based upon "Range encoding: an algorithm for removing redundancy from a digitised message." by G. N. N. Martin. The symbols encoded by the Snow range coder are bits(0|1). The associated probabilities are not fix but change depending on the symbol mix seen so far. bit seen|new state ---------+----------------------------------------------- 0|256 - state_transition_table[256 - old_state];1|state_transition_table[old_state];state_transition_table={ 0, 0, 0, 0, 0, 0, 0, 0, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 190, 191, 192, 194, 194, 195, 196, 197, 198, 199, 200, 201, 202, 202, 204, 205, 206, 207, 208, 209, 209, 210, 211, 212, 213, 215, 215, 216, 217, 218, 219, 220, 220, 222, 223, 224, 225, 226, 227, 227, 229, 229, 230, 231, 232, 234, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 248, 0, 0, 0, 0, 0, 0, 0};FIXME Range Coding of integers:------------------------- FIXME Neighboring Blocks:===================left and top are set to the respective blocks unless they are outside of the image in which case they are set to the Null block top-left is set to the top left block unless it is outside of the image in which case it is set to the left block if this block has no larger parent block or it is at the left side of its parent block and the top right block is not outside of the image then the top right block is used for top-right else the top-left block is used Null block y, cb, cr are 128 level, ref, mx and my are 0 Motion Vector Prediction:=========================1. the motion vectors of all the neighboring blocks are scaled to compensate for the difference of reference frames scaled_mv=(mv *(256 *(current_reference+1)/(mv.reference+1))+128)> the median of the scaled top and top right vectors is used as motion vector prediction the used motion vector is the sum of the predictor and(mvx_diff, mvy_diff) *mv_scale Intra DC Prediction block[y][x] dc[1] Definition: snow.txt:400 NULL_IF_CONFIG_SMALL #define NULL_IF_CONFIG_SMALL(x) Return NULL if CONFIG_SMALL is true, otherwise the argument without modification. Definition: internal.h:117 id enum AVCodecID id Definition: extract_extradata_bsf.c:325 av_frame_ref int av_frame_ref(AVFrame *dst, const AVFrame *src) Set up a new reference to the data described by the source frame. Definition: frame.c:325 SOF14 @ SOF14 Definition: mjpeg.h:53 ff_jpegls_decode_lse int ff_jpegls_decode_lse(MJpegDecodeContext *s) Decode LSE block with initialization parameters. Definition: jpeglsdec.c:51 decode_block_progressive static int decode_block_progressive(MJpegDecodeContext *s, int16_t *block, uint8_t *last_nnz, int ac_index, uint16_t *quant_matrix, int ss, int se, int Al, int *EOBRUN) Definition: mjpegdec.c:872 av_err2str #define av_err2str(errnum) Convenience macro, the return value should be used only directly in function arguments but never stan... Definition: error.h:121 ff_mjpeg_decode_sos int ff_mjpeg_decode_sos(MJpegDecodeContext *s, const uint8_t *mb_bitmask, int mb_bitmask_size, const AVFrame *reference) Definition: mjpegdec.c:1633 ff_mjpeg_bits_ac_luminance const uint8_t ff_mjpeg_bits_ac_luminance[] Definition: jpegtabs.h:40 size int size Definition: twinvq_data.h:10344 AV_CODEC_ID_SMVJPEG @ AV_CODEC_ID_SMVJPEG Definition: codec_id.h:263 AV_RB32 uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_WL32 unsigned int_TMPL AV_WL24 unsigned int_TMPL AV_WL16 uint64_t_TMPL AV_WB64 unsigned int_TMPL AV_RB32 Definition: bytestream.h:96 avpriv_report_missing_feature void avpriv_report_missing_feature(void *avc, const char *msg,...) av_printf_format(2 Log a generic warning message about a missing feature. AVFrameSideData::data uint8_t * data Definition: frame.h:225 SOF15 @ SOF15 Definition: mjpeg.h:54 AVFrame::format int format format of the frame, -1 if unknown or unset Values correspond to enum AVPixelFormat for video frames,... Definition: frame.h:404 AVCodecHWConfigInternal Definition: hwconfig.h:29 ff_mjpeg_bits_dc_luminance const uint8_t ff_mjpeg_bits_dc_luminance[] Definition: jpegtabs.h:32 OPEN_READER #define OPEN_READER(name, gb) Definition: get_bits.h:139 height #define height AV_PIX_FMT_YUVA444P @ AV_PIX_FMT_YUVA444P planar YUV 4:4:4 32bpp, (1 Cr & Cb sample per 1x1 Y & A samples) Definition: pixfmt.h:167 VD #define VD Definition: av1dec.c:1213 offset it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf offset Definition: writing_filters.txt:86 line Definition: graph2dot.c:48 get_xbits static int get_xbits(GetBitContext *s, int n) Read MPEG-1 dc-style VLC (sign bit + mantissa with no MSB). Definition: get_bits.h:322 av_dict_free void av_dict_free(AVDictionary **pm) Free all the memory allocated for an AVDictionary struct and all keys and values. Definition: dict.c:203 HWACCEL_NVDEC #define HWACCEL_NVDEC(codec) Definition: hwconfig.h:71 predictor static void predictor(uint8_t *src, ptrdiff_t size) Definition: exrenc.c:164 find_marker static int find_marker(const uint8_t **pbuf_ptr, const uint8_t *buf_end) Definition: mjpegdec.c:2178 AV_STEREO3D_FLAG_INVERT #define AV_STEREO3D_FLAG_INVERT Inverted views, Right/Bottom represents the left view. Definition: stereo3d.h:167 AV_PIX_FMT_VAAPI @ AV_PIX_FMT_VAAPI Hardware acceleration through VA-API, data[3] contains a VASurfaceID. Definition: pixfmt.h:119 DQT @ DQT Definition: mjpeg.h:73 AV_LOG_INFO #define AV_LOG_INFO Standard information. Definition: log.h:191 AVCodec::id enum AVCodecID id Definition: codec.h:216 layout Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel layout Definition: filter_design.txt:18 SOF10 @ SOF10 Definition: mjpeg.h:49 AV_FIELD_BB @ AV_FIELD_BB Definition: codec_par.h:40 AV_CODEC_ID_MJPEG @ AV_CODEC_ID_MJPEG Definition: codec_id.h:57 src1 #define src1 Definition: h264pred.c:140 NEG_USR32 #define NEG_USR32(a, s) Definition: mathops.h:166 copy_block4 static void copy_block4(uint8_t *dst, const uint8_t *src, ptrdiff_t dstStride, ptrdiff_t srcStride, int h) Definition: copy_block.h:37 interlaced uint8_t interlaced Definition: mxfenc.c:2040 decode_block static int decode_block(MJpegDecodeContext *s, int16_t *block, int component, int dc_index, int ac_index, uint16_t *quant_matrix) Definition: mjpegdec.c:805 i #define i(width, name, range_min, range_max) Definition: cbs_h2645.c:271 code and forward the test the status of outputs and forward it to the corresponding return FFERROR_NOT_READY If the filters stores internally one or a few frame for some it can consider them to be part of the FIFO and delay acknowledging a status change accordingly Example code Definition: filter_design.txt:178 EOI @ EOI Definition: mjpeg.h:71 copy_block.h AVCodecContext::extradata uint8_t * extradata some codecs need / can use extradata like Huffman tables. Definition: avcodec.h:484 show_bits static unsigned int show_bits(GetBitContext *s, int n) Show 1-25 bits. Definition: get_bits.h:447 FF_CODEC_CAP_INIT_CLEANUP #define FF_CODEC_CAP_INIT_CLEANUP The codec allows calling the close function for deallocation even if the init function returned a fai... Definition: internal.h:50 FF_PROFILE_MJPEG_HUFFMAN_EXTENDED_SEQUENTIAL_DCT #define FF_PROFILE_MJPEG_HUFFMAN_EXTENDED_SEQUENTIAL_DCT Definition: avcodec.h:1626 display.h AV_STEREO3D_TOPBOTTOM @ AV_STEREO3D_TOPBOTTOM Views are on top of each other. Definition: stereo3d.h:79 mjpeg_decode_dri static int mjpeg_decode_dri(MJpegDecodeContext *s) Definition: mjpegdec.c:1811 AVCodecInternal::in_pkt AVPacket * in_pkt This packet is used to hold the packet given to decoders implementing the .decode API; it is unused b... Definition: internal.h:144 MIN_CACHE_BITS #define MIN_CACHE_BITS Definition: get_bits.h:129 av_fast_padded_malloc void av_fast_padded_malloc(void *ptr, unsigned int *size, size_t min_size) Same behaviour av_fast_malloc but the buffer has additional AV_INPUT_BUFFER_PADDING_SIZE at the end w... Definition: utils.c:50 SOF9 @ SOF9 Definition: mjpeg.h:48 av_always_inline #define av_always_inline Definition: attributes.h:49 value it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf default value Definition: writing_filters.txt:86 FF_DEBUG_STARTCODE #define FF_DEBUG_STARTCODE Definition: avcodec.h:1310 FFMIN #define FFMIN(a, b) Definition: macros.h:49 AV_PIX_FMT_YUVJ440P @ AV_PIX_FMT_YUVJ440P planar YUV 4:4:0 full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV440P and setting color_range Definition: pixfmt.h:100 av_frame_unref void av_frame_unref(AVFrame *frame) Unreference all the buffers referenced by frame and reset the frame fields. Definition: frame.c:435 ff_idctdsp_init av_cold void ff_idctdsp_init(IDCTDSPContext *c, AVCodecContext *avctx) Definition: idctdsp.c:238 av_mallocz void * av_mallocz(size_t size) Allocate a memory block with alignment suitable for all memory accesses (including vectors if availab... Definition: mem.c:263 AVCodec::name const char * name Name of the codec implementation. Definition: codec.h:209 AVCodecContext::chroma_sample_location enum AVChromaLocation chroma_sample_location This defines the location of chroma samples. Definition: avcodec.h:974 len int len Definition: vorbis_enc_data.h:426 exif.h DHT @ DHT Definition: mjpeg.h:56 AVCodecContext::height int height Definition: avcodec.h:556 AVCodecContext::pix_fmt enum AVPixelFormat pix_fmt Pixel format, see AV_PIX_FMT_xxx. Definition: avcodec.h:593 AVCOL_RANGE_MPEG @ AVCOL_RANGE_MPEG Narrow or limited range content. Definition: pixfmt.h:580 av_calloc void * av_calloc(size_t nmemb, size_t size) Definition: mem.c:271 idctdsp.h avcodec.h ff_zigzag_direct const uint8_t ff_zigzag_direct[64] Definition: mathtables.c:98 AV_PIX_FMT_PAL8 @ AV_PIX_FMT_PAL8 8 bits with AV_PIX_FMT_RGB32 palette Definition: pixfmt.h:77 REFINE_BIT #define REFINE_BIT(j) Definition: mjpegdec.c:945 ret ret Definition: filter_design.txt:187 AV_LOG_FATAL #define AV_LOG_FATAL Something went wrong and recovery is not possible. Definition: log.h:174 pred static const float pred[4] Definition: siprdata.h:259 av_stereo3d_alloc AVStereo3D * av_stereo3d_alloc(void) Allocate an AVStereo3D structure and set its fields to default values. Definition: stereo3d.c:28 FFSWAP #define FFSWAP(type, a, b) Definition: macros.h:52 AVClass::class_name const char * class_name The name of the class; usually it is the same name as the context structure type to which the AVClass... Definition: log.h:71 frame these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame Definition: filter_design.txt:264 AVStereo3D::type enum AVStereo3DType type How views are packed within the video. Definition: stereo3d.h:180 SOF2 @ SOF2 Definition: mjpeg.h:41 align_get_bits static const uint8_t * align_get_bits(GetBitContext *s) Definition: get_bits.h:694 pos unsigned int pos Definition: spdifenc.c:412 LSE @ LSE JPEG-LS extension parameters. Definition: mjpeg.h:104 FF_DEBUG_QP #define FF_DEBUG_QP Definition: avcodec.h:1307 AV_INPUT_BUFFER_PADDING_SIZE #define AV_INPUT_BUFFER_PADDING_SIZE Definition: defs.h:40 left Tag MUST be and< 10hcoeff half pel interpolation filter coefficients, hcoeff[0] are the 2 middle coefficients[1] are the next outer ones and so on, resulting in a filter like:...eff[2], hcoeff[1], hcoeff[0], hcoeff[0], hcoeff[1], hcoeff[2] ... the sign of the coefficients is not explicitly stored but alternates after each coeff and coeff[0] is positive, so ...,+,-,+,-,+,+,-,+,-,+,... hcoeff[0] is not explicitly stored but found by subtracting the sum of all stored coefficients with signs from 32 hcoeff[0]=32 - hcoeff[1] - hcoeff[2] - ... a good choice for hcoeff and htaps is htaps=6 hcoeff={40,-10, 2} an alternative which requires more computations at both encoder and decoder side and may or may not be better is htaps=8 hcoeff={42,-14, 6,-2}ref_frames minimum of the number of available reference frames and max_ref_frames for example the first frame after a key frame always has ref_frames=1spatial_decomposition_type wavelet type 0 is a 9/7 symmetric compact integer wavelet 1 is a 5/3 symmetric compact integer wavelet others are reserved stored as delta from last, last is reset to 0 if always_reset||keyframeqlog quality(logarithmic quantizer scale) stored as delta from last, last is reset to 0 if always_reset||keyframemv_scale stored as delta from last, last is reset to 0 if always_reset||keyframe FIXME check that everything works fine if this changes between framesqbias dequantization bias stored as delta from last, last is reset to 0 if always_reset||keyframeblock_max_depth maximum depth of the block tree stored as delta from last, last is reset to 0 if always_reset||keyframequant_table quantization tableHighlevel bitstream structure:==============================--------------------------------------------|Header|--------------------------------------------|------------------------------------|||Block0||||split?||||yes no||||......... intra?||||:Block01 :yes no||||:Block02 :....... ..........||||:Block03 ::y DC ::ref index:||||:Block04 ::cb DC ::motion x :||||......... :cr DC ::motion y :||||....... ..........|||------------------------------------||------------------------------------|||Block1|||...|--------------------------------------------|------------ ------------ ------------|||Y subbands||Cb subbands||Cr subbands||||--- ---||--- ---||--- ---|||||LL0||HL0||||LL0||HL0||||LL0||HL0|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||LH0||HH0||||LH0||HH0||||LH0||HH0|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||HL1||LH1||||HL1||LH1||||HL1||LH1|||||--- ---||--- ---||--- ---||||--- ---||--- ---||--- ---|||||HH1||HL2||||HH1||HL2||||HH1||HL2|||||...||...||...|||------------ ------------ ------------|--------------------------------------------Decoding process:=================------------|||Subbands|------------||||------------|Intra DC||||LL0 subband prediction ------------|\ Dequantization ------------------- \||Reference frames|\ IDWT|------- -------|Motion \|||Frame 0||Frame 1||Compensation . OBMC v -------|------- -------|--------------. \------> Frame n output Frame Frame<----------------------------------/|...|------------------- Range Coder:============Binary Range Coder:------------------- The implemented range coder is an adapted version based upon "Range encoding: an algorithm for removing redundancy from a digitised message." by G. N. N. Martin. The symbols encoded by the Snow range coder are bits(0|1). The associated probabilities are not fix but change depending on the symbol mix seen so far. bit seen|new state ---------+----------------------------------------------- 0|256 - state_transition_table[256 - old_state];1|state_transition_table[old_state];state_transition_table={ 0, 0, 0, 0, 0, 0, 0, 0, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 190, 191, 192, 194, 194, 195, 196, 197, 198, 199, 200, 201, 202, 202, 204, 205, 206, 207, 208, 209, 209, 210, 211, 212, 213, 215, 215, 216, 217, 218, 219, 220, 220, 222, 223, 224, 225, 226, 227, 227, 229, 229, 230, 231, 232, 234, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 248, 0, 0, 0, 0, 0, 0, 0};FIXME Range Coding of integers:------------------------- FIXME Neighboring Blocks:===================left and top are set to the respective blocks unless they are outside of the image in which case they are set to the Null block top-left is set to the top left block unless it is outside of the image in which case it is set to the left block if this block has no larger parent block or it is at the left side of its parent block and the top right block is not outside of the image then the top right block is used for top-right else the top-left block is used Null block y, cb, cr are 128 level, ref, mx and my are 0 Motion Vector Prediction:=========================1. the motion vectors of all the neighboring blocks are scaled to compensate for the difference of reference frames scaled_mv=(mv *(256 *(current_reference+1)/(mv.reference+1))+128)> the median of the scaled left Definition: snow.txt:386 AV_RL32 uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_RL32 Definition: bytestream.h:92 ff_mjpeg_find_marker int ff_mjpeg_find_marker(MJpegDecodeContext *s, const uint8_t **buf_ptr, const uint8_t *buf_end, const uint8_t **unescaped_buf_ptr, int *unescaped_buf_size) Definition: mjpegdec.c:2203 AV_CODEC_ID_AMV @ AV_CODEC_ID_AMV Definition: codec_id.h:157 OFFSET it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf default minimum maximum flags name is the option keep it simple and lowercase description are in without and describe what they for example set the foo of the bar offset is the offset of the field in your see the OFFSET() macro AVCodecContext main external API structure. Definition: avcodec.h:383 AVFrame::height int height Definition: frame.h:389 SHOW_UBITS #define SHOW_UBITS(name, gb, num) Definition: get_bits.h:212 buffer the frame and frame reference mechanism is intended to as much as expensive copies of that data while still allowing the filters to produce correct results The data is stored in buffers represented by AVFrame structures Several references can point to the same frame buffer Definition: filter_design.txt:49 AVCHROMA_LOC_CENTER @ AVCHROMA_LOC_CENTER MPEG-1 4:2:0, JPEG 4:2:0, H.263 4:2:0. Definition: pixfmt.h:619 AV_PIX_FMT_NONE @ AV_PIX_FMT_NONE Definition: pixfmt.h:65 APP2 @ APP2 Definition: mjpeg.h:81 values these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return values Definition: filter_design.txt:263 AVPixFmtDescriptor::comp AVComponentDescriptor comp[4] Parameters that describe how pixels are packed. Definition: pixdesc.h:105 ff_tdecode_header int ff_tdecode_header(GetByteContext *gb, int *le, int *ifd_offset) Decodes a TIFF header from the input bytestream and sets the endianness in *le and the offset to the ... Definition: tiff_common.c:228 profiles static const AVProfile profiles[] Definition: libfdk-aacenc.c:429 AV_PIX_FMT_YUV444P @ AV_PIX_FMT_YUV444P planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples) Definition: pixfmt.h:71 ff_mjpeg_bits_dc_chrominance const uint8_t ff_mjpeg_bits_dc_chrominance[] Definition: jpegtabs.h:37 AVCodecContext::debug int debug debug Definition: avcodec.h:1302 ff_mjpeg_decode_sof int ff_mjpeg_decode_sof(MJpegDecodeContext *s) Definition: mjpegdec.c:298 APP0 @ APP0 Definition: mjpeg.h:79 AV_PIX_FMT_GBRP @ AV_PIX_FMT_GBRP planar GBR 4:4:4 24bpp Definition: pixfmt.h:158 AVCodecContext::coded_width int coded_width Bitstream width / height, may be different from width/height e.g. Definition: avcodec.h:571 AVMEDIA_TYPE_VIDEO @ AVMEDIA_TYPE_VIDEO Definition: avutil.h:201 AV_PIX_FMT_GRAY16LE @ AV_PIX_FMT_GRAY16LE Y , 16bpp, little-endian. Definition: pixfmt.h:98 AV_PIX_FMT_YUV422P @ AV_PIX_FMT_YUV422P planar YUV 4:2:2, 16bpp, (1 Cr & Cb sample per 2x1 Y samples) Definition: pixfmt.h:70 ff_set_dimensions int ff_set_dimensions(AVCodecContext *s, int width, int height) Check that the provided frame dimensions are valid and set them on the codec context. Definition: utils.c:86 SOI @ SOI Definition: mjpeg.h:70 ff_mjpeg_decoder const AVCodec ff_mjpeg_decoder mjpeg_decode_app static int mjpeg_decode_app(MJpegDecodeContext *s) Definition: mjpegdec.c:1823 AVCodecContext::frame_number int frame_number Frame counter, set by libavcodec. Definition: avcodec.h:1023 av_stereo3d_create_side_data AVStereo3D * av_stereo3d_create_side_data(AVFrame *frame) Allocate a complete AVFrameSideData and add it to the frame. Definition: stereo3d.c:33 avpriv_request_sample #define avpriv_request_sample(...) Definition: tableprint_vlc.h:37 AVFrameSideData Structure to hold side data for an AVFrame. Definition: frame.h:223 flush_put_bits static void flush_put_bits(PutBitContext *s) Pad the end of the output stream with zeros. Definition: put_bits.h:142 SOF1 @ SOF1 Definition: mjpeg.h:40 av_free #define av_free(p) Definition: tableprint_vlc.h:34 AVDictionaryEntry Definition: dict.h:79 AVCodecContext::codec_tag unsigned int codec_tag fourcc (LSB first, so "ABCD" -> ('D'<<24) + ('C'<<16) + ('B'<<8) + 'A'). Definition: avcodec.h:408 ff_mjpeg_build_vlc int ff_mjpeg_build_vlc(VLC *vlc, const uint8_t *bits_table, const uint8_t *val_table, int is_ac, void *logctx) Definition: mjpegdec_common.c:41 AVCodecContext::priv_data void * priv_data Definition: avcodec.h:410 AV_OPT_TYPE_BOOL @ AV_OPT_TYPE_BOOL Definition: opt.h:241 av_freep #define av_freep(p) Definition: tableprint_vlc.h:35 av_fast_malloc void av_fast_malloc(void *ptr, unsigned int *size, size_t min_size) Allocate a buffer, reusing the given one if large enough. Definition: mem.c:560 av_dict_copy int av_dict_copy(AVDictionary **dst, const AVDictionary *src, int flags) Copy entries from one AVDictionary struct into another. Definition: dict.c:217 AV_PIX_FMT_YUV411P @ AV_PIX_FMT_YUV411P planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples) Definition: pixfmt.h:73 HWACCEL_VAAPI #define HWACCEL_VAAPI(codec) Definition: hwconfig.h:73 FFMAX3 #define FFMAX3(a, b, c) Definition: macros.h:48 int32_t int32_t Definition: audioconvert.c:56 bytestream.h imgutils.h bytestream2_init static av_always_inline void bytestream2_init(GetByteContext *g, const uint8_t *buf, int buf_size) Definition: bytestream.h:137 flags #define flags(name, subs,...) Definition: cbs_av1.c:561 AVERROR_BUG #define AVERROR_BUG Internal bug, also see AVERROR_BUG2. Definition: error.h:52 MAX_COMPONENTS #define MAX_COMPONENTS Definition: mjpegdec.h:45 rgb static const SheerTable rgb[2] Definition: sheervideodata.h:32 block The exact code depends on how similar the blocks are and how related they are to the block Definition: filter_design.txt:207 av_log #define av_log(a,...) Definition: tableprint_vlc.h:28 AVERROR_INVALIDDATA #define AVERROR_INVALIDDATA Invalid data found when processing input. Definition: error.h:61 MKTAG #define MKTAG(a, b, c, d) Definition: macros.h:55 h h Definition: vp9dsp_template.c:2038 SOF7 @ SOF7 Definition: mjpeg.h:46 AVStereo3D Stereo 3D type: this structure describes how two videos are packed within a single video surface,... Definition: stereo3d.h:176 AVDictionaryEntry::value char * value Definition: dict.h:81 av_image_check_size int av_image_check_size(unsigned int w, unsigned int h, int log_offset, void *log_ctx) Check if the given dimension of an image is valid, meaning that all bytes of the image can be address... Definition: imgutils.c:318 ff_blockdsp_init av_cold void ff_blockdsp_init(BlockDSPContext *c, AVCodecContext *avctx) Definition: blockdsp.c:59 AV_RB24 uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_WL32 unsigned int_TMPL AV_WL24 unsigned int_TMPL AV_WL16 uint64_t_TMPL AV_WB64 unsigned int_TMPL AV_WB32 unsigned int_TMPL AV_RB24 Definition: bytestream.h:97 PREDICT #define PREDICT(ret, topleft, top, left, predictor) Definition: mjpeg.h:118 put_bits.h SOF6 @ SOF6 Definition: mjpeg.h:45 JPG @ JPG Definition: mjpeg.h:47 av_fourcc2str #define av_fourcc2str(fourcc) Definition: avutil.h:348 re float re Definition: fft.c:78
__label__pos
0.950082
Using the Internet as a Survey Tool The internet is a two-way interactive medium.  As such, it has the potential of being used as a survey instrument.  This is particularly significant at a time when other survey modes (such as face-to-face personal interviews, mail and telephone) are confronting declining cooperation rates in these changing times (e.g. security concerns, inundation of unsolicited mail, aggressive telemarketing, unlisted telephone calls, answering machines, caller ID, opt-out listings, etc). The internet is obviously the logical survey tool for the universe of all internet users.  But what is the validity of extrapolating the results of internet surveys to the non-internet universe?  That depends on the characteristics and behavior of the internet users.  Now, a great deal is known about the characteristics of internet users vis-à-vis the general population.  There are many data sources that document the internet population, and we will cite some survey results from the 2002 MARS study.  This is a mail survey of adults 18+ in the 50 states of the USA conducted during the first quarter of 2002.  Within that study, 59.5% of the respondents indicated that they had used the Internet in the last 30 days.   In the next two charts, we show the incidence of internet users separately by age/sex groups, educational level, occupation and personal income (among those who are employed). (source: 2002 MARS study) (source: 2002 MARS study) These charts support the common belief that the internet population is younger, better educated, hold better jobs and are better paid.  Therefore, any simple sample of the internet population will not be directly representative of the general population.  That much is not in dispute.  But we are not writing this article to argue this point.  Rather, we wish to discuss the issue of whether it is possible to conduct an internet survey, weight that sample to the general population's universe estimates and then claim to have accurate and unbiased estimates for that general population.  Absent any other proof, we find this to be an unsubstantiated claim.  Here is the simplest test that we can construct --- if we commission someone to conduct an internet survey of the incidence of internet usage, they can weight that sample any which way they want and the result would still be ... 100%, and quite wrong at that! We now offer an empirical test of the representativeness of an internet survey.  Within the MARS 2002 survey, there is a simple random sample of 7,293 adults in the USA.  To compensate for the incompleteness of the coverage of the sampling frame and the differential response rates by population subgroups, that sample was weighted to US census figures on demographic characteristics such as age, sex, education, occupation, personal income, household income, geography, household size, race, ethnicity, etc.  This is classical post-stratification weighting used by most survey organizations.  Then, audiences estimates were obtained for 96 consumer magazine titles.   Next, we took the subset of 4,632 adults in the MARS sample who had used the internet in the last 30 days and we weighted them to the same US census figures in exactly the same way.  Thus, this sample will match the full sample on the same demographic characteristics such as age, sex, education, occupation, personal income, household income, geographical, household size, race, ethnicity, etc.  Then, audience estimates were obtained for the same 96 consumer magazine titles.  Here, we are assuming that conducting the same survey using a different survey mode on the same set of persons will yield identical responses.  This is known NOT to be true, but we will gloss over issues of survey mode differences in this discussion. In the next chart, we present the two sets of magazine audiences in the form of scatterplot.  It would appear that the two series are highly correlated.  Indeed, the correlation coefficient is a very high 0.993 (where 1.000 represents perfection).  In this scatterplot, we have also plotted the best fitting straight line.  If we look at that line carefully, we would realize that it lies above the 45 degree diagonal line (which represents perfection).  The fact that the line lies above the 45 degree line means that the Internet-based audience estimates tend to be higher. (source: 2002 MARS study) We can look at the numbers in another way.  For each magazine title, we define a magazine audience index as 100 times the internet-based estimates divided by the general population estimate.  An index of 100 would represent perfection.  An index greater than 100 implies a larger internet estimates and vice versa.  In the next chart, we show a histogram of the distrubtion of the 96 magazine audience indices.  The average index is 110, which means that the internet-based magazine audience estimates are 10% high on the average.  This overestimate is not consistent across the magazine titles, as some are significantly higher and some are significantly lower.  The differences are sufficiently large that they would be unacceptable for the purpose of media planning (that is, some magazine would be advantaged by 40% while another would be disadvantaged by 25%). (source: 2002 MARS study) This is not say that we have proclaimed the internet to be totally useless for magazine audience estimation.  It is certainly useful, and in fact quite natural, for measuring magazine readership among internet users.  In the case for professional technology magazines, for example.  And as more people get on the internet, it will be a better approximant of the general population.  It is also possible that certain types of data adjustments to the internet-based survey estimates may yield accurate estimates for the general population.  But these data adjustments will require some knowledge about the relationship between non-internet usage and the survey outcome variables and such relationships cannot be revealed solely on the basis of an internet-based survey. We should point out that, in this article, we have not even remotely hinted at the difficulty of obtaining a sample that is representative of the internet.  There is no universal frame for internet users.  As a result, an internet survey sample is often assembled from pop-up intercepts at selected cooperating websites or permission-based emails, which incur an unknown degree of bias. (posted by Roland Soong on 04/07/2003) (Return to Zona Latina's Home Page)
__label__pos
0.506178
Probability Mass function | Its Complete Overview with 5 Examples Discrete Random Variable and Mathematical Expectation-II As already we now familiar with the discrete random variable, it is the random variable which takes countable number of possible values in a sequence. The two important concept related to the discrete random variables are the probability of discrete random variable and distribution function we restrict the name to such probability and distribution function as, Probability Mass function (p.m.f)                 The Probability Mass function is the probability of the discrete random variable, so for any discrete random variables  x1, x2, x3, x4,……, xk  the corresponding probabilities P(x1), P(x2), P(x3), P(x4)……, P(xk) are the corresponding probability mass functions. Specifically, for X=a, P(a)=P(X=a) is its p.m.f We here onwards use probability mass function for discrete random variables probability.  All the probability characteristics for the probability will obviously applicable to probability mass function like positivity and summation of all p.m.f will be one e.t.c. Cumulative Distribution Function (c.d.f)/Distribution Function   The distribution function defined as F(x)=P(X<=x) for discrete random variable with probability mass function is the cumulative distribution function (c.d.f.) of the random variable. and mathematical expectation for such random variable we defined was E(g(x))= \sum\limits_{i}x_{i}p_{i} we now see some of the results of mathematical expectations 1. If x1, x2, x3, x4,….. are the discrete random variables with respective probabilities P(x1), P(x2), P(x3), P(x4) … the expectation for the real valued function g will be E(g(x))= \sum\limits_{i}g(x_{i})p(x_{i}) Example: for the following probability mass functions find the E(X3) probability mass function Here the g(X)=X3 So, E(g(x))= \sum\limits_{i}g(x_{i})p(x_{i}) E(X^{3}) = \sum\limits_{i}x_{i}^{3}p(x_{i}) E(X^{3}) = (-1)^{3}<em>0.2+(0)^{3}</em>0.5+(1)^{^{3}}*0.3 E(X^{3}) = 0.1 In the similar way for any nth order we can write E[X^{n}] = \sum\limits_{x:p(x)>0}x^{n}p(x) Which is known as nth moment. 2. If a and b are constants then E[aX + b]=aE[X] + b This we can understand easily as E[aX + b] = \sum\limits_{x:p(x)>0}(ax + b)p(x) = a \sum\limits_{x:p(x)>0} xp(x) +b \sum\limits_{x:p(x)>0}p(x) =aE[X] + b Variance in terms of Expectation.                 For the mean denoted by μ the variance of the discrete random variable X denoted by var(X) or σ in terms of expectation will be Var(X) =E[(X- μ)2] and this we can further simplify as Var(X) =E[(X- μ)2] = \sum\limits_{x} (x-\mu)^{2} p(x) = \sum\limits_{x} (x^{2}-2x\mu+\mu^{2}) p(x) = \sum\limits_{x} (x^{2}p(x) -2\mu \sum\limits_{x}xp(x) +\mu^{2} \sum\limits_{x} p(x) = E[X^{2}] -2 \mu^{2} + \mu^{2} = E[X^{2}] - \mu^{2} this means we can write the variance as the difference of the expectation of random variable square and square of expectation of random variable. i.e. Var (X)= E[X2] – (E[X])2 Example:  when a die is thrown calculate the variance. Solution:  here we know when die thrown the probabilities for each face will be p(1)=p(2)=p(3)=p(4)=p(5)=p(6)=1/6 hence for calculating variance we will find expectation of random variable and its square as E[X]=1.(1/6)+2.(1/6)+3.(1/6)+4.(1/6)+5.(1/6)+6.(1/6)=(7/2) E[X2] =12.(1/6)+22.(1/6)+32.(1/6)+42.(1/6)+52.(1/6)+62.(1/6) =(1/6)(91) and we just obtained the variance as Var (X) =E[X2] – (E[X])2 so Var (X)=(91/6) -(7/2)2 =35/12 One of the important identity for variance is 1. For the arbitrary constants a and b we have Var(aX + b) =a2 Var(X) This we can show easily as Var(aX + b) =E[(aX+ b -aμ-b)2 ] =E[a2(X – μ)2] =a2 E[(X-μ)2] =a2 Var(X) Bernoulli Random variable       A Swiss mathematician James Bernoulli define the Bernoulli random variable as a random variable having either success or failure as only two outcomes for the random experiment. i.e When the outcome is success X=1 When the outcome is failure X=0 So the probability mass function for the Bernoulli random variable is p(0) = P{X=0}=1-p p(1) =P{X=1}=p where p is the probability of success and 1-p will be the probability of failure. Here we can take 1-p=q also where q is the probability of failure. As this type of random variable is obviously discrete so this is one of discrete random variable. Example: Tossing a coin. Binomial Random Variable If for a random experiment which is having only outcome as success or failure we take n trials so each time we will get either success or failure then the random variable X representing outcome for such n trial random experiment is known as Binomial random variable.                 In other words if p is the probability mass function for the success in  the single Bernoulli trial and q=1-p is the probability for the failure then the probability for happening of event ‘x or i’ times in n trials will be f(x)=P(X=x)=\binom{n}{x}p^{x}q^{n-x}=\frac{n!}{x!(n-x!)}p^{x}q^{n-x} or p(i)=\binom{n}{i}p^{i}(1-p)^{n-i} where i =0,1,2,….n Example: If we toss two coins six times and getting head is success and remaining occurrences are failures then its probability will be f(x)=P(X=x)=\binom{n}{x}p^{x}q^{n-x}=\frac{n!}{x!(n-x!)}p^{x}q^{n-x} P(X=2)=\binom{6}{2}(\frac{1}{2})^{2}(\frac{1}{2})^{6-2}=\frac{6!}{2!4!}(\frac{1}{2})^{2}(\frac{1}{2})^{4}=\frac{15}{64} in the similar way we can calculate for any such experiment. The Binomial random variable is having the name Binomial because it represent the expansion of latex^{n}=q^{n}+\binom{n}{1}q^{n-1}p+\binom{n}{2}q^{n-2}p^{2}+…….+p^{n}=\sum\limits_{i = 1}^n\binom{n}{x}p^{x}q^{n-x}[/latex] If we put in place of n=1 then this would turn into the Bernoulli’s random variable. Example: If five coins were tossed and the outcome is taken independently then what would be the probability for number of heads occurred. Here if we take random variable X as the number of heads then it would turns to the binomial random variable with n=5 and probability of  success as ½ So by following the probability mass function for the binomial random variable we will get P{X=0}=\binom{5}{0}(\frac{1}{2})^{0}(\frac{1}{2})^{5}=\frac{1}{32} P{X=1}=\binom{5}{1}(\frac{1}{2})^{1}(\frac{1}{2})^{4}=\frac{5}{32} P{X=2}=\binom{5}{2}(\frac{1}{2})^{2}(\frac{1}{2})^{3}=\frac{10}{32} P{X=3}=\binom{5}{3}(\frac{1}{2})^{3}(\frac{1}{2})^{2}=\frac{10}{32} P{X=4}=\binom{5}{4}(\frac{1}{2})^{4}(\frac{1}{2})^{1}=\frac{5}{32} Example: In a certain company the probability of defective is 0.01 from the production. The company manufacture and sells the product in a pack of 10 and to its customers offer money back guarantee that at most 1 of the 10 product is defective, so what proportion of sold products pack the company must replace. Here If X is the random variable representing the defective products then it is of the binomial type with n=10 and p=0.01 then the probability that the pack will return is P({X\geq 1})=1-P(X=0)-P(X=1)=1-\binom{10}{0}(0.01)^{0}(0.99)^{10}-\binom{10}{1}(0.01)^{1}(0.99)^{9} \approx 0.004 Example: (chuck-a-luck/ wheel of fortune) In a specific game of fortune in hotel a player bet on any of the numbers from 1 to 6, three dice then rolled and if the number appears bet by the player once, twice or thrice the player that much units means if appear once then 1 unit if on two dice then 2 units and if on three dice then 3 units, check with the help of probability the game is fair for the player or not. If we assume there will be no unfair means with the dice and con techniques then by assuming the outcome of the dice independently the probability of success for each dice is 1/6 and failure will be  1-1/6 so this turns to be the example of binomial random variable with n=3 so first we will calculate the winning probabilities by assigning x as players win P(X=0)=\binom{3}{0}(\frac{1}{6})^{0}(\frac{5}{6})^{3}=\frac{125}{216} P(X=1)=\binom{3}{1}(\frac{1}{6})^{1}(\frac{5}{6})^{2}=\frac{75}{216} P(X=2)=\binom{3}{2}(\frac{1}{6})^{2}(\frac{5}{6})^{1}=\frac{15}{216} P(X=3)=\binom{3}{3}(\frac{1}{6})^{3}(\frac{5}{6})^{0}=\frac{1}{216} Now to calculate the game is fair for the player or not we will calculate the expectation of the random variable E[X]=\frac{-125+75+30+3}{216} =-\frac{17}{216} This means the probability of losing the game for the player when he plays 216 times is 17. Conclusion:    In this article we discussed some of the basic properties of a discrete random variable, probability mass function and variance.  In addition we has seen some types of a discrete random variable, Before we start the continuous random variable we try to cover all the types and properties of discrete random variable, if you want further reading then go through: Schaum’s Outlines of Probability and Statistics https://en.wikipedia.org/wiki/Probability For more Topics on Mathematics, please follow this link About DR. MOHAMMED MAZHAR UL HAQUE I am DR. Mohammed Mazhar Ul Haque , Assistant professor in Mathematics. Having 12 years of experience in teaching. Having vast knowledge in Pure Mathematics , precisely on Algebra. Having the immense ability of problem designing and solving. Capable of Motivating candidates to enhance their performance. I love to contribute to Lambdageeks to make Mathematics Simple , Interesting & Self Explanatory for beginners as well as experts. Let's connect through LinkedIn - https://www.linkedin.com/in/dr-mohammed-mazhar-ul-haque-58747899/
__label__pos
0.999033
Free Shipping | Easy Returns search Search account_circle Login Search What is Bluetooth: a Beginner’s Guide • 5 min read What is Bluetooth: a Beginner’s Guide - Auris, Inc Chances are, you’re reading this article on a device that is Bluetooth compatible. These are usually your smartphones, tablets, and laptops that have the amazing functionality to stream data wirelessly through the air. Bluetooth has even more applications than you might have realized, and it’s becoming more and more difficult to imagine a world without it. As we quickly transition from wired tech to a wireless society, it’s important to have some knowledge of what Bluetooth can offer for your own personal life. So if you’re fairly new to Bluetooth and are ready to make the giant leap forward, we’re here to guide you through everything you need to know about this fantastic wireless protocol.   History of Bluetooth The concept of Bluetooth was originally conceived back in 1994 by Dr. Jaap Haartsen and Ericcson. It was named after a prominent Viking who united Denmark and Norway. The Viking, named Harald, got the nickname “Bluetooth” because of a rotten tooth in his mouth that took a bluish hue. Bluetooth’s iconic logo is a combination of the Nordic letters “H” and “B” for Harald Bluetooth The protocol was designed to replace telecommunication cables by using short-range radio waves. The first-ever consumer Bluetooth device was released in 1999 as a hands-free mobile headset. Alongside the headset came the first version of Bluetooth 1.0, allowing developers to create wireless chipsets, dongles, etc. Bluetooth has continually been updated since then, with its latest 5.0 version introducing new features that would have been farfetched just over 20 years ago. With technology that is always evolving, it is exciting to see how far it has come and how far it will go. How Does Bluetooth Work? Bluetooth works by broadcasting signals via radio waves at a 2.4GHz frequency. It uses a network organization known as “star topology.” When devices are set-up in this manner, it creates something called a piconet. A piconet is composed of a master device (something like your phone or a laptop) and one or multiple “slave” devices (your headphones, speakers, etc.). They all share the same radio frequency and are able to communicate with one another with specific wavelengths that are dictated by the master device. At this frequency, the devices share information with each other called “packets.” These might contain audio data, textual data, or even image data. These are what’s sent between your devices when you change the song from your phone or print something from your computer. While this admittedly a simplified version of what’s actually happening between your Bluetooth peripherals, it can give you a sense of just exactly how complex the mechanisms are that allow you to enjoy wireless applications. What Are the Benefits of Bluetooth? The reason that Bluetooth has become so ingrained in our culture is because of one main facet: convenience. Wired technology is very linear, meaning that typically only two devices are able to communicate with each other through one wired system. With Bluetooth, you can control multiple devices at one time. On top of that, wires limit our range of motion, tethering us down to a small range of mobility. The newest Bluetooth supports a range of up to 800 feet, so you could theoretically keep listening to your music from anywhere inside the house. Not to mention, the world is quickly transitioning into a fully wireless landscape. In a market report by Bluetooth themselves, they revealed that annual Bluetooth device shipments have increased from 2.7 billion in 2014 to 4.4 billion in 2020. With that in mind, it’s a way of the future that it is important to embrace. What Are the Limitations of Bluetooth? As great as it is, it’s not without some flaws. One of the largest limitations of Bluetooth is the sound quality that it is able to produce. Bluetooth uses codecs that must compress audio files into small, manageable bits that can be transmitted through the air. Then, it must be decompressed into a larger file size that can be listened to from your favorite speakers.  This process is a type of lossy audio, meaning that the end result is never as pristine as the original recording. Compared to analog systems like stereos or record players, which maintain the authentic sound, it can sound a bit worse than we might hope. Bluetooth also isn’t yet capable of transmitting visual elements, so video streaming isn’t yet available. What Can Bluetooth Be Used For? Bluetooth has a wide variety of uses that you can probably utilize in many different facets of your daily routine. Here are a few examples: Speakers and Headphones Arguably the most popular usage of Bluetooth is in wireless headphones and speaker systems. These allow you to instantly stream music from your phone’s library directly to the device without ever needing to use a wire. They’re perfect for the gym or getting work done around the house, just because you won’t be held down by those annoying cables. They’re also great for making hands-free calls and texts, as most headphones come with a microphone that can interact with elements on your smartphone. Smart Appliances A new way of using Bluetooth that is on the rise is using wireless tech to communicate with smart appliances. These are products that can utilize wi-fi or Bluetooth networks to be controlled through apps on your phone. Smart lights are a popular usage for this. These connect to your network, allowing you to use an app on your phone to control their color, dimness, and even set automated timers to switch on and off. Turning Wired Devices Into Wireless Since we know that wired systems have better quality sound than wireless Bluetooth devices but just aren’t nearly as convenient, wouldn’t it be great if we could combine those two benefits into one? The great news is that you can with a Bluetooth receiver. This device can attach to any existing analog or digital stereo system, allowing you to connect a master device and have full control. It means that you’ll be able to play your favorite Spotify playlists without ever having to sacrifice the authentic sound that you know and love. Similarly, you could turn your television into a wirelessly compatible device by getting a Bluetooth transmitter. The speakers on a TV aren’t the greatest for watching movies from the safety of your home. But with a transmitter, you can use Bluetooth technology to broadcast the TVs sound to a quality wireless speaker or even headphones. You could pair up to two devices at one time, so you could even create your own surround sound set-up right at home. If Bluetooth weren’t so convenient, these types of gadgets would never have been invented in the first place. They’re really great to have on hand, especially if good audio quality is something you want to be able to maintain. File Transfers Transferring files between a phone and computer used to be a nauseating endeavor because of all of the cables needed to make it work. But Bluetooth has made it easier than ever before. You could even use Bluetooth to share information between two phones. Apple has recently introduced a feature called “Airdropping,” which uses Bluetooth frequencies to easily send files to another iPhone within range at just the click of a button. While these methods might not be as fast as a traditional wire method, there is really no doubting that it’s a whole lot simpler. In Conclusion If you don’t have much experience working with Bluetooth, it can be overwhelming to figure out where to start. However, once you learn a little bit about how it works, you’ll never go back to traditional wired devices. Since its inception in the late 90s, Bluetooth has allowed for seamless music streaming, smart device control, file transfer, and wired compatibility like we’ve never seen before. As it continues to grow in popularity, so too does its vast catalog of uses and features. Bluetooth shows no signs of slowing down anytime soon, and we are excited to see just exactly how it will be engineered to assist us in our busy lives.   Sources: Bluetooth: Why Modern Tech is Named After Powerful King of Denmark and Norway | Ancient Origins Radio Frequency Radiation and Cell Phones | FDA 2019 Bluetooth Market Update | Bluetooth® Technology Website
__label__pos
0.763595
Skip to main content Cloud function runner A Function-as-a-Service (FaaS) runner is a platform that allows developers to deploy and run small, single-purpose functions in the cloud. FaaS runners typically provide a serverless architecture, which means that developers do not have to worry about infrastructure management or the scaling, as the platform automatically handles these tasks. Case study For example, imagine you have an e-commerce application that uses FaaS to process orders. When a customer places an order, multiple functions may need to be executed, such as validating the order, processing the payment, and updating the inventory. Each function may be executed independently by the FaaS platform and may take varying amounts of time to complete. Those functions may also be executed for historical reason on different platforms like AWS Lambda, Google Cloud Functions, or Azure Functions. To collect the results of all the functions in a timely manner, you need to ensure that each function is executed in the correct order and that you are not waiting for a slow function to complete before moving on to the next function. Metatype's solution To solve the use case of executing multiple functions and collecting their results, Metatype provides two key features. 1. Function composition/chaining: functions can be chained together to form a pipeline. The output of one function can be used as the input of the next function in the pipeline. This allows us to execute multiple functions in a specific order. 2. Embedded runner: you can easily write a function that glues together multiple functions and executes them in a specific order. This allows you to execute multiple functions in a specific order. Currently, both Python and Typescript are supported. Loading...
__label__pos
0.978511
I have 3? 5? files with lumps of java tracing code in. THIS IS THE INDEX / DIRECTORY to my trace files. (My "testjava.html" also has useful tracing functions in it, but only for DOM stuff.) There are three writing methods: 1) alert(); / pops up a window, done by Java 2) Edit DOM tree: reserve a place on the web page by pre-putting an HTML area for this, and java inserts a string: see file 'trace1', function trace1() [and mystudents.php] 3) Open new window and do writeln: see file 'trace3' [and DOMtrace] ==== TRACE0 General advice on tracing tactics; and: Catch java errors [flagerror() / onerror()], so you don't miss their existence. One-liner kind of stuff: java snippets to copy into body of java, to get tracing output. E.g. how to print out stuff. Lumps to be assimilated elsewhere. ==== TRACE1 My basic routine for inserting errmsgs into a prepared bit of a browser page. ==== TRACE2 Printing info about a given object, either java object or DOM object. Typically insert a suitable "onclick", and pass "this". It uses the trace1 method for output, though could easily do alert(). function prnodeJava(node) { alert(myprnode(node)); }; function prnodeJava2(node) { trace(myprnode(node)); }; // Prints short info on the node, or any Java object, given it. // and calls prnodeShort if DOM object function prnodeShort(node) { trace(myprnodeshort(node) + '\n'); } Brief info on one DOM node: a bit more than 'prnodeJava' e.g. name-value, id, name, form. function prnodeMed(node) { trace(myprnodemedium(node) + '\n'); } Ditto, but gives the brief info on each of: self, parent, children. function prnodeSubtree(node) { printDOMtree(node,null, ==== TRACE3 DOM tree printing; to a new window it creates. function prnodeSubtree(node) { printDOMtree(node,null,0); }; // Prints subtree in a new window. function prnodeFull(node) { printDOMtree(node, null, 1); } Prints ALL fields of one node-object in a new window. Good for when you want the name of a field in these objects. function supertrace(obj) { // Opens a window and does a 4 part inspection of the object: a) java type b) short summary of DOM props c) Subtree from object's parent downwards d) Full all-field print of objects' fields.
__label__pos
0.988752
Tell me more × Stack Overflow is a question and answer site for professional and enthusiast programmers. It's 100% free, no registration required. #include <iostream> using namespace std; int main() { int score; char grade; cout << "Enter your score:" << endl; cin >> score; if (score >= 90) grade = 'a'; if (score >= 80) grade = 'b'; if (score >= 70) grade = 'c'; if (score >= 60) grade = 'd'; else grade = 'f'; cout << grade << endl; switch (grade) { case 'a': cout << "Good job" << endl; break; case 'c': cout << "Fair job" << endl; break; case 'f': cout << "Failure" << endl; break; default: cout << "invalid" << endl; } cin.get(); return 0; } why is it giving me my default switch case when i enter 95 when i should be getting case a share|improve this question 1   Why are you converting the number to a letter and then the letter to a longer worded description? Skip the switch and print both in your first if (after you fix it of course). Over-complication is our worst curse, fight against it! – Blindy Sep 28 '11 at 19:20 because case 'd': is missing did you notice which grade was being output? – AJG85 Sep 28 '11 at 19:21 1   did you checked the output of the grade before the switch start is it returns a – punit Sep 28 '11 at 19:23 @Blindy probably because this is a homework assignment illustrating differences between nested if-else-if and switch cases. – AJG85 Sep 28 '11 at 19:23 9 Answers You're missing a bunch of elses, or doing the comparisons in the wrong order. 95 is greater than 90, but it's also greater than 80, 70 and 60. So you'll get a 'd'. (And you're not handling 'd' in your switch.) share|improve this answer I believe you want if (score >= 90) grade = 'a'; else if (score >= 80) grade = 'b'; else if (score >= 70) grade = 'c'; else if (score >= 60) grade = 'd'; else grade = 'f'; What you have does not mutually exclude any but the last two cases, 60 and above and lower. Your code doesn't short circuit, it checks all of 1 through 5. if (score >= 90) // 1. grade = 'a'; if (score >= 80) // 2. grade = 'b'; if (score >= 70) // 4. grade = 'c'; if (score >= 60) // 5. grade = 'd'; else grade = 'f'; share|improve this answer I think you want to use 'else if', it is falling down to the last if "score >= 60" which is true, and grade then equals "d", which produces the default case in your switch statement. share|improve this answer You have specified it such that your 95 satisfies all the cases: 95 is bigger than 90, but also bigger than 80 and than 70 etc... In this case, the last one wins. You can solve it by either using elses, or by wrapping it in a function and returning as soon as you know the grade you need: char grade( int score ){ if( score >= 90 ) return 'a'; if( score >= 80 ) return 'b'; ... } share|improve this answer It's because of your if statements up top. you should be using else ifs instead of individual ifs. Whats happening is your if for 90 is following through, and then so are all the others. Your letter a is essentially being overwritten because 95 is >= to all of the other coniditons. Using an else if will break the rest of the checks when a true one is found. if (score >= 90) grade = 'a'; else if (score >= 80) grade = 'b'; else if (score >= 70) grade = 'c'; else if (score >= 60) grade = 'd'; else grade = 'f'; share|improve this answer Because all score comparisons are not combined with if/else if conditions. They are independent if statements. Thus grade gets overwritten for 95. share|improve this answer The if branches are ordered wrong (or you need to provide else branches like so:) See it live here: http://ideone.com/2uSZT #include <iostream> using namespace std; int main() { int score; char grade; cout << "Enter your score:" << endl; cin >> score; if (score >= 90) grade = 'a'; else if (score >= 80) grade = 'b'; else if (score >= 70) grade = 'c'; else if (score >= 60) grade = 'd'; else grade = 'f'; cout << grade << endl; switch (grade) { case 'a': cout << "Good job" << endl; break; case 'c': cout << "Fair job" << endl; break; case 'f': cout << "Failure" << endl; break; default: cout << "invalid" << endl; } cin.get(); return 0; } share|improve this answer @mat: you must have flashed on my initial too-quick response? I thought I ninja'd it quick enough :) – sehe Sep 28 '11 at 19:39 yep, saw your answer pop up while I was writing mine. Got confused a bit by it too :) – Mat Sep 28 '11 at 19:43 @Mat: see, I know that it is good for something that I make it a point to disregard these popups. The only time it bites me is when I'm finally ready to press 'Submit' and it turns out that the question got deleted – sehe Sep 28 '11 at 19:57 you need to improve your if conditions, you are checking score >= no. the input 95 execute all the if statements and the last executed statement was the d now in your switch statement case d is not present so it's executes the default one. share|improve this answer You've gotten some answers already, but I think I'll suggest a slightly different possibility gets rid of most of the control flow and substitutes a bit of math: char grades[] = "00000012344"; char *messages[] = { "Excellent Job", "Good job", "Average job", "Mediocre Job", "Failure" }; if (score < 0 || score > 100) std::cout << "Invalid score"; else { int grade = grades[score/10]; std::cout << messages[grade]; } So, we use score/10 to turn scores of 0-100 to 0-10. We then look up the appropriate grade for a score, with f=0, d=1, c=2, b=3 and a=4. We use that to select and print out the appropriate message. I've added messages (that may or may not be quite what you want) for the letters you skipped. share|improve this answer Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.828413
Tuesday, July 19th, 2022 code html javascript utils • 158w Every single time you have a input text field and submit-like button next to it, at some point, you would like to bind the same action of the submit button to the press of the enter key. It's trivial, but also it's simply better to just copy-paste it. So here its is: function onenter(ele, f) { ele.addEventListener('keydown', function (e) { if (e.keyCode === 13) { f(e); e.preventDefault(); } }) } And also while we are here, instead of using the-not-be-named lib, just use these: function q(x) { return document.querySelector(x); } function qa(x) { return [...document.querySelectorAll(x)]; } function range(x) { return Array.from({length: x}, (v, i) => i); } For more just go here. done_
__label__pos
0.988482
Was this page helpful? Your feedback about this content is important. Let us know what you think. Additional feedback? 1500 characters remaining Export (0) Print Expand All NetworkStream Class Provides the underlying stream of data for network access. System.Object   System.MarshalByRefObject     System.IO.Stream       System.Net.Sockets.NetworkStream Namespace:  System.Net.Sockets Assembly:  System (in System.dll) public class NetworkStream : Stream The NetworkStream type exposes the following members.   NameDescription Public methodNetworkStream(Socket)Creates a new instance of the NetworkStream class for the specified Socket. Public methodNetworkStream(Socket, Boolean)Initializes a new instance of the NetworkStream class for the specified Socket with the specified Socket ownership. Public methodNetworkStream(Socket, FileAccess)Creates a new instance of the NetworkStream class for the specified Socket with the specified access rights. Public methodNetworkStream(Socket, FileAccess, Boolean)Creates a new instance of the NetworkStream class for the specified Socket with the specified access rights and the specified Socket ownership. Top   NameDescription Public propertyCanReadGets a value that indicates whether the NetworkStream supports reading. (Overrides Stream.CanRead.) Public propertyCanSeekGets a value that indicates whether the stream supports seeking. This property is not currently supported.This property always returns false. (Overrides Stream.CanSeek.) Public propertyCanTimeoutIndicates whether timeout properties are usable for NetworkStream. (Overrides Stream.CanTimeout.) Public propertyCanWriteGets a value that indicates whether the NetworkStream supports writing. (Overrides Stream.CanWrite.) Public propertyDataAvailableGets a value that indicates whether data is available on the NetworkStream to be read. Public propertyLengthGets the length of the data available on the stream. This property is not currently supported and always throws a NotSupportedException. (Overrides Stream.Length.) Public propertyPositionGets or sets the current position in the stream. This property is not currently supported and always throws a NotSupportedException. (Overrides Stream.Position.) Protected propertyReadableGets or sets a value that indicates whether the NetworkStream can be read. Public propertyReadTimeoutGets or sets the amount of time that a read operation blocks waiting for data. (Overrides Stream.ReadTimeout.) Protected propertySocketGets the underlying Socket. Protected propertyWriteableGets a value that indicates whether the NetworkStream is writable. Public propertyWriteTimeoutGets or sets the amount of time that a write operation blocks waiting for data. (Overrides Stream.WriteTimeout.) Top   NameDescription Public methodBeginReadBegins an asynchronous read from the NetworkStream. (Overrides Stream.BeginRead(Byte[], Int32, Int32, AsyncCallback, Object).) Public methodBeginWriteBegins an asynchronous write to a stream. (Overrides Stream.BeginWrite(Byte[], Int32, Int32, AsyncCallback, Object).) Public methodClose()Closes the current stream and releases any resources (such as sockets and file handles) associated with the current stream. Instead of calling this method, ensure that the stream is properly disposed. (Inherited from Stream.) Public methodClose(Int32)Closes the NetworkStream after waiting the specified time to allow data to be sent. Public methodCopyTo(Stream)Reads the bytes from the current stream and writes them to another stream. (Inherited from Stream.) Public methodCopyTo(Stream, Int32)Reads the bytes from the current stream and writes them to another stream, using a specified buffer size. (Inherited from Stream.) Public methodCopyToAsync(Stream)Asynchronously reads the bytes from the current stream and writes them to another stream. (Inherited from Stream.) Public methodCopyToAsync(Stream, Int32)Asynchronously reads the bytes from the current stream and writes them to another stream, using a specified buffer size. (Inherited from Stream.) Public methodCopyToAsync(Stream, Int32, CancellationToken)Asynchronously reads the bytes from the current stream and writes them to another stream, using a specified buffer size and cancellation token. (Inherited from Stream.) Public methodCreateObjRefCreates an object that contains all the relevant information required to generate a proxy used to communicate with a remote object. (Inherited from MarshalByRefObject.) Protected methodCreateWaitHandle Obsolete. Allocates a WaitHandle object. (Inherited from Stream.) Public methodDispose()Releases all resources used by the Stream. (Inherited from Stream.) Protected methodDispose(Boolean)Releases the unmanaged resources used by the NetworkStream and optionally releases the managed resources. (Overrides Stream.Dispose(Boolean).) Public methodEndReadHandles the end of an asynchronous read. (Overrides Stream.EndRead(IAsyncResult).) Public methodEndWriteHandles the end of an asynchronous write. (Overrides Stream.EndWrite(IAsyncResult).) Public methodEquals(Object)Determines whether the specified object is equal to the current object. (Inherited from Object.) Protected methodFinalizeReleases all resources used by the NetworkStream. (Overrides Object.Finalize().) Public methodFlushFlushes data from the stream. This method is reserved for future use. (Overrides Stream.Flush().) Public methodFlushAsync()Asynchronously clears all buffers for this stream and causes any buffered data to be written to the underlying device. (Inherited from Stream.) Public methodFlushAsync(CancellationToken)Flushes data from the stream as an asynchronous operation. (Overrides Stream.FlushAsync(CancellationToken).) Public methodGetHashCodeServes as the default hash function. (Inherited from Object.) Public methodGetLifetimeServiceRetrieves the current lifetime service object that controls the lifetime policy for this instance. (Inherited from MarshalByRefObject.) Public methodGetTypeGets the Type of the current instance. (Inherited from Object.) Public methodInitializeLifetimeServiceObtains a lifetime service object to control the lifetime policy for this instance. (Inherited from MarshalByRefObject.) Protected methodMemberwiseClone()Creates a shallow copy of the current Object. (Inherited from Object.) Protected methodMemberwiseClone(Boolean)Creates a shallow copy of the current MarshalByRefObject object. (Inherited from MarshalByRefObject.) Protected methodObjectInvariantInfrastructure. Obsolete. Provides support for a Contract. (Inherited from Stream.) Public methodReadReads data from the NetworkStream. (Overrides Stream.Read(Byte[], Int32, Int32).) Public methodReadAsync(Byte[], Int32, Int32)Asynchronously reads a sequence of bytes from the current stream and advances the position within the stream by the number of bytes read. (Inherited from Stream.) Public methodReadAsync(Byte[], Int32, Int32, CancellationToken)Asynchronously reads a sequence of bytes from the current stream, advances the position within the stream by the number of bytes read, and monitors cancellation requests. (Inherited from Stream.) Public methodReadByteReads a byte from the stream and advances the position within the stream by one byte, or returns -1 if at the end of the stream. (Inherited from Stream.) Public methodSeekSets the current position of the stream to the given value. This method is not currently supported and always throws a NotSupportedException. (Overrides Stream.Seek(Int64, SeekOrigin).) Public methodSetLengthSets the length of the stream. This method always throws a NotSupportedException. (Overrides Stream.SetLength(Int64).) Public methodToStringReturns a string that represents the current object. (Inherited from Object.) Public methodWriteWrites data to the NetworkStream. (Overrides Stream.Write(Byte[], Int32, Int32).) Public methodWriteAsync(Byte[], Int32, Int32)Asynchronously writes a sequence of bytes to the current stream and advances the current position within this stream by the number of bytes written. (Inherited from Stream.) Public methodWriteAsync(Byte[], Int32, Int32, CancellationToken)Asynchronously writes a sequence of bytes to the current stream, advances the current position within this stream by the number of bytes written, and monitors cancellation requests. (Inherited from Stream.) Public methodWriteByteWrites a byte to the current position in the stream and advances the position within the stream by one byte. (Inherited from Stream.) Top   NameDescription Public Extension MethodAsInputStreamConverts a managed stream in the .NET for Windows 8.x Store apps to an input stream in the Windows Runtime. (Defined by WindowsRuntimeStreamExtensions.) Public Extension MethodAsOutputStreamConverts a managed stream in the .NET for Windows 8.x Store apps to an output stream in the Windows Runtime. (Defined by WindowsRuntimeStreamExtensions.) Public Extension MethodAsRandomAccessStreamConverts the specified stream to a random access stream. (Defined by WindowsRuntimeStreamExtensions.) Top The NetworkStream class provides methods for sending and receiving data over Stream sockets in blocking mode. For more information about blocking versus nonblocking Sockets, see Using an Asynchronous Client Socket. You can use the NetworkStream class for both synchronous and asynchronous data transfer. For more information about synchronous and asynchronous communication, see Sockets. To create a NetworkStream, you must provide a connected Socket. You can also specify what FileAccess permission the NetworkStream has over the provided Socket. By default, closing the NetworkStream does not close the provided Socket. If you want the NetworkStream to have permission to close the provided Socket, you must specify true for the value of the ownsSocket parameter. Use the Write and Read methods for simple single thread synchronous blocking I/O. If you want to process your I/O using separate threads, consider using the BeginWrite and EndWrite methods, or the BeginRead and EndRead methods for communication. The NetworkStream does not support random access to the network data stream. The value of the CanSeek property, which indicates whether the stream supports seeking, is always false; reading the Position property, reading the Length property, or calling the Seek method will throw a NotSupportedException. Read and write operations can be performed simultaneously on an instance of the NetworkStream class without the need for synchronization. As long as there is one unique thread for the write operations and one unique thread for the read operations, there will be no cross-interference between read and write threads and no synchronization is required. The following code example demonstrates how to create a NetworkStream from a connected Stream Socket and perform basic synchronous blocking I/O. // Examples for constructors that do not specify file permission.  // Create the NetworkStream for communicating with the remote host. NetworkStream myNetworkStream; if (networkStreamOwnsSocket){ myNetworkStream = new NetworkStream(mySocket, true); } else{ myNetworkStream = new NetworkStream(mySocket); } .NET Framework Supported in: 4.6, 4.5, 4, 3.5, 3.0, 2.0, 1.1 .NET Framework Client Profile Supported in: 4, 3.5 SP1 Any public static (Shared in Visual Basic) members of this type are thread safe. Any instance members are not guaranteed to be thread safe. Show: © 2015 Microsoft
__label__pos
0.674248
Understanding The Loop Understanding how the Loop functions will help you understand how you can control it. Controlling the Loop to display exactly the content you want will be one of your most used tools in developing WordPress-powered web sites. Because the Loop is at the heart of every WordPress theme, being able to customize the display content opens up the doors to making WordPress look and act however you want. To understand the Loop, it helps to break down the steps WordPress takes to generate a page's content: > The URL is matched against existing files and directories in the WordPress installation. If the file is there, it is loaded by the web server. WordPress doesn't actually get involved in this decision; it's up to your web server and the .htaccess file created by WordPress to decide if the URL is something handled by the web server or to be turned into a WordPress content query. This was covered in Chapter 4. > If the URL is passed to WordPress, it has to determine what content to load. For example, when visiting a specific tag page like http: //example.com/tag/bacon, WordPress will determine that you are viewing a tag and load the appropriate template, select the posts saved with that tag, and generate the output for the tag page. > The translation of URL to content selection magic happens inside of the parse_query () method within the WP_Query object that WordPress created early on in its processing. WordPress parses the URL first into a set of query parameters that are described in the next section. All query strings from the URL are passed into WordPress to determine what content to display, even if they look like nicely formatted pathnames. If your site is using pretty permalinks, the values between slashes in those permalinks are merely parameters for query strings. For example, http://example.com/tag/bacon is the same as http: //example.com?tag=bacon, which conveys a query string of "tag with a value of bacon.'' > WordPress then converts the query specification parameters into a MySQL database query to retrieve the content. The workhorse here is the get_ posts () method within the WP_Query object that we describe later in this chapter. The get_ posts () method takes all of those query parameters and turns them into SQL statements, eventually invoking the SQL string on the MySQL database server and extracting the desired content. The content returned from the database is then saved in the WP_Query object to be used in the WordPress Loop and cached to speed up other references to the same posts made before another database query is executed. > Once the content is retrieved, WordPress sets all of the is_ conditional tags such as is_home and is_ page. These are set as part of executing the default query based on the URL parsing, and we'll discuss cases where you may need to reset these tags. > WordPress picks a template from your theme based on the type of query and the number of posts returned, for example, a single post or a category only query, and the output of the query is passed to this default invocation of the Loop. The Loop can be customized for different web site purposes. For example, a news site might use the Loop to display the latest news headlines. A business directory could use the Loop to display local businesses alphabetically by name, or always put posts about sponsoring businesses at the top of every displayed page. A photo blog might use the Loop to display the most recent photos loaded into the web site. The possibilities are endless when customizing the Loop in WordPress because it gives you complete control over what content is selected and the order in which it is rendered for display. Six Figure Blog Marketing Six Figure Blog Marketing Take Advantage Of This Technology In The World Of WordPress™ Blogging And Start Making Real Residual Money Again. Even If You Are New To The Playing Field, It's Easy To Make Money Again On The Web, With A Free WordPress™ Blog, Some Good Advice And A New Get My Free Ebook Post a comment
__label__pos
0.628297
Results 1 to 5 of 5 1.    #1   I copied Saved Preferences.prc from RAM to a folder on SD card using Explorer and FileZ. The copies were made within seconds of each other. The files have different names on the SD card (but that was explained in a separate thread earlier). However, the file sizes and # of records in them were very different as well. Why? Here are the details: Saved by FileZ: Name: Saved Preferences.prc Type: sprf Creator: psys Size: 33810 b, 236 records Saved by Explorer: Name: Saved%20Preferences.prc Type: sprf Creator: psys Size: 79092 b, 278 records The Unsaved Preferences.prc sizes are different too. Saved by FileZ: 52 K Saved by Explorer: 32 K What's the explanation for these differences? I use these files to restore the preferences after recent crashes that result in lost preferences. I would like to know if the 2 programs (FileZ and Explorer) are equally reliable in saving and restoring these files. -- Aloke Cingular GSM Software:Treo650-1.17-CNG Firmware:01.51 Hardware:A 2. #2   Try this before jumping to any conclusions: Copy the "Saved Preferences.prc" file with FileZ, then exit the program and open Resco. Do not save the same file with Resco. Instead use Resco to save a different .prc file, then exit. Open FileZ again, and save the "Saved Preferences.prc" file again, but under a different name or folder. Finally compare those two copies of the same file, both copied by FileZ, and see if they are the identical or not. I have a feeling that FileZ or Resco is saving info to the saved preferences file at some point (as certain apps do), which is causing the file to grow and/or shrink in between switching/using apps to make copies. Honestly, I wouldnt worry. Whatever you copy it with, that last copy is going to be the most current. Also, if you hotsync regularly you have a copy there too. . 3.    #3   The size of Saved Preferences in my Backup folder (on pc) is 34K, similar to the size saved by FileZ. I have asked Resco for an explanation as well. Let's see what they say about this... -- Aloke Cingular GSM Software:Treo650-1.17-CNG Firmware:01.51 Hardware:A 4.    #4   Can anyone else duplicate my results? Copy Saved Preferences and Unsaved Preferences from RAM to a directory on the SD card using Resco Explorer. Now check the file properties (size, # of records) of the files in RAM and card using Explorer File Info. Do you see any differences? Last edited by aprasad; 12/28/2005 at 02:09 PM. Reason: more info -- Aloke Cingular GSM Software:Treo650-1.17-CNG Firmware:01.51 Hardware:A 5. #5   I tried to copy save pref to my card by using filez and rescoexplorer. Nothing Different in size or record items. Posting Permissions
__label__pos
0.505447
[RFC] Is a more expressive way to represent reductions useful? I’m reviewing [AArch64] Generate DOT instructions from matching IR by huntergr-arm · Pull Request #69583 · llvm/llvm-project · GitHub and find the need to perform block/loop analysis in order to identify a reduction to be bothersome. For the transformation to be safe the new pass wants to identify add operations that can be cross-lane reassociated. However, during vectorisation we know this is our intent so I figure we can represent this better at the IR level to maximise expressiveness. This goes beyond the PR’s DOT use case as at least for SVE there’s other instructions that are tricky to use once the reassociativity information is lost. We could add dedicated intrinsics like: Ty vector.reassociative.<binop>(Ty X, Ty Y): whereby X and Y are concatenated and then a binop is performed for each result lane by selecting any two input lanes once and all once. Or: Ty vector.shuffle.obscure(Ty X): which performs no real work other than to represent an unknown translation of input to output lanes. {Ty, Ty} vector.shuffle.obscure(Ty X, Ty T): I personally prefer the shuffle option because one or two intrinsics likely cover all cases but then perhaps this can be represented using instruction flags or metadata? My gut feeling is that would be dangerous in the face of CSE? What do others think? Is this a problem worth solving? or are passes like that in the PR the best solution? Thinking about this a little more I guess for best results we might also need to incorporate some kind of partial reduction properties to represent a reduction of element count. This makes me wonder if a solution exists that doesn’t necessitate having to add code generation support for something new, which is something I’ve been trying to avoid.
__label__pos
0.993736
Friday February 27, 2015 Homework Help: math Posted by Al on Monday, April 2, 2012 at 6:47am. 5 4 10b b 6 ? 5a 2a 40 2 2 5c The above numbers are presented in 3 sets of 4 boxes each. What is the missing algebraic expression in the box in the last set? Answer this Question First Name: School Subject: Answer: Related Questions math - 5 4 5a 2a 10b b 40 2 6 ? 2 5c What is the missing algebraic expression? math - 5 4 5a 2a 10b b 40 2 6 ? 2 5c What is the missing algebraic expression? ... math - Observe the patterns in the sequence below. What is the missing algebraic... 3rd grade math - sara wants to know how many boxes of cookies her entire girl ... Math - observe the first 3 sets of numbers to determine the missing value in the... math - Observe the first three sets of numbers to determine the missing value in... 2nd grade math - what is mental math and how does it work. the question is use ... Math Help Please - Let U = {9, 10, 11, 12, 13, 14, 15, 16, 17, 18}, A = {10, 12... math ( sets ) - { 30,31,32,....,100 } How many numbers in the above set are ... math (sets) - { 1,2.3,..........,210 } How many numbers in the above set are ... Members
__label__pos
0.572618
fatal error php handler Kieler Wisconsin Address 5739 Seemore Rdg, Potosi, WI 53820 Phone (608) 330-2794 Website Link http://www.oyentechnologies.com Hours fatal error php handler Kieler, Wisconsin I find it useful to throw an Exception when this error is caught, then you can use try/catch. go Reply Leave a Reply Cancel reply Your email address will not be published. How? Copyright 1999-2016 by Refsnes Data. Logical fallacy: X is bad, Y is worse, thus X is not bad The mortgage company is trying to force us to make repairs after an insurance claim This riddle could When an exception is thrown, code following the statement will not be executed, and PHP will attempt to find the first matching catch block. But this class acts a nice container for those functions. up down 4 pinkgothic at gmail dot com ¶6 years ago If you're handling sensitive data and you don't Now that we have created an error handling function we need to decide when it should be triggered. Which day of the week is today? However, I couldn't find out how to determine if the shutdown has occured due to a fatal error or due to the script reaching its end. How would they learn astronomy, those who don't see the stars? "Rollbacked" or "rolled back" the edit? share|improve this answer edited Jul 1 '15 at 1:20 answered Nov 10 '08 at 9:44 keparo 19.5k105164 23 Pfff, I remember those 650.000+ e-mails i got the following morning. Try it and see what happens! This is the handler signature since PHP 7: void handler ( Throwable $ex ) NULL may be passed instead, to reset this handler to its default state. Catch − A "catch" block retrieves an exception and creates an object containing the exception information. Pinpointing the file/line where it occured is enough. For example, if your PHP has a compile error, it makes sense that the compiler won’t be able to read your code and see that you’re trying to set a custom thanks Reply James says: May 31, 2016 at 2:53 PM Good tip, thanks! Recovering from fatal errors (memory limits for example) is not something that I would try to do, but making these errors discoverable (without customer submitting a support ticket) makes all the A better way to evaluate a certain determinant Did Sputnik 1 have attitude control? I assume the comments are because the code is posted here, since none of them provide extra information your code doesn't provide. The variables have meaningful names and the indentation is good. php fatal-error share|improve this question edited Dec 22 '13 at 19:26 aksu 3,94451434 asked Nov 10 '08 at 6:42 too much php 41.8k26106123 add a comment| 17 Answers 17 active oldest Code: Replace lines matching a pattern with lines from another file in order Is accuracy binary? Fatal errors are called that, because they are fatal. share|improve this answer answered May 2 '12 at 23:33 J.Money 3,28511838 add a comment| up vote 6 down vote Nice solution found in Zend Framework 2: /** * ErrorHandler that can Credits: Picture by John Trainor Please enable JavaScript to view the comments powered by Disqus. I wouldn't use in live production code but it's great to add to a page when a quick answer to what is failing is needed. Send an Error Message by E-Mail In the example below we will send an e-mail with an error message and end the script, if a specific error occurs: For those of you looking for an object orientated way to do this without looking down at Glen and Sean's examples (Lesson 1: ALWAYS read the logs!), here you go: However, if you use a custom error handler to convert errors to ErrorExceptions suddenly there are a multitude of new I was searching for this solution too long time and didn't find! The PHP manual describes an E_RECOVERABLE_ERROR as: Catchable fatal error. more stack exchange communities company blog Stack Exchange Inbox Reputation and Badges sign up log in tour help Tour Start here for a quick overview of the site Help Center Detailed Join them; it only takes a minute: Sign up Handle fatal errors in PHP using register_shutdown_function() up vote 21 down vote favorite 11 According to the comment on this answer it asked 7 years ago viewed 244368 times active 2 months ago Visit Chat Linked 0 is it possible to catch fatal error php on include file? 0 Is there a way Replace lines matching a pattern with lines from another file in order How to get this substring on bash script? This is very inconvinient. In still agreeing with keparo, catching fatal errors does defeat the purpose of "FATAL error" so its not really intended for you to do further processing. This is because processing should not typically be recovered after a fatal error. Books for chess traps With the passing of Thai King Bhumibol, are there any customs/etiquette as a traveler I should be aware of? Lets explain there new keyword related to exceptions. The script found something that might be an error, but could also happen when running a script normally 8 5 E_CORE_ERROR Fatal errors that occur during PHP's initial start-up. 16 6 Each try must have at least one corresponding catch block. Isn't that more expensive than an elevated system? That is, if you define your own global exception handler by calling set_exception_handler() and you throw an exception from inside it then this fatal error occurs.
__label__pos
0.797529
 performance measurements Each table row shows performance measurements for this Go program with a particular command-line input value N.  N  CPU secs Elapsed secs Memory KB Code B ≈ CPU Load 250,0001.420.5135,2921531  90% 78% 58% 61% 2,500,00013.394.5550,0241531  92% 79% 62% 64% 25,000,000136.3247.62263,2161531  87% 56% 57% 89% Read the ↓ make, command line, and program output logs to see how this program was run. Read k-nucleotide benchmark to see what this program should do.  notes go version go1.3 linux/amd64  k-nucleotide Go #2 program source code /* The Computer Language Benchmarks Game * http://benchmarksgame.alioth.debian.org/ * * contributed by Tylor Arndt */ package main import ( "bufio" "bytes" "fmt" "io/ioutil" "log" "os" "runtime" "sort" "sync" ) func main() { runtime.GOMAXPROCS(4) dna := readEncDNA() var wgs [7]*sync.WaitGroup for i := 0; i < 7; i++ { wgs[i] = new(sync.WaitGroup) } report(dna, 1, nil, wgs[0]) report(dna, 2, wgs[0], wgs[1]) report(dna, 3, wgs[1], wgs[2]) report(dna, 4, wgs[2], wgs[3]) report(dna, 6, wgs[3], wgs[4]) report(dna, 12, wgs[4], wgs[5]) report(dna, 18, wgs[5], wgs[6]) wgs[6].Wait() } func readEncDNA() []byte { in, startTok := bufio.NewReader(os.Stdin), []byte(">THREE ") for line, err := in.ReadSlice('\n'); !bytes.HasPrefix(line, startTok); line, err = in.ReadSlice('\n') { if err != nil { log.Panicf("Error: Could not read input from stdin; Details: %s", err) } } ascii, err := ioutil.ReadAll(in) if err != nil { log.Panicf("Error: Could not read input from stdin; Details: %s", err) } j := 0 for i, c, asciic := 0, byte(0), len(ascii); i < asciic; i++ { c = ascii[i] switch c { case 'a', 'A': c = 0 case 't', 'T': c = 1 case 'g', 'G': c = 2 case 'c', 'C': c = 3 case '\n': continue default: log.Fatalf("Error: Invalid nucleotide value: '%c'", ascii[i]) } ascii[j] = c j++ } return ascii[:j+1] } var targSeqs = []string{3: "GGT", 4: "GGTA", 6: "GGTATT", 12: "GGTATTTTAATT", 18: "GGTATTTTAATTTATAGT"} func report(dna []byte, n int, prev, done *sync.WaitGroup) { done.Add(1) go func() { tbl, output := count(dna, n), "" switch n { case 1, 2: output = freqReport(tbl, n) default: targ := targSeqs[n] output = fmt.Sprintf("%d\t%s\n", tbl[compStr(targ)], targ) } if prev != nil { prev.Wait() } fmt.Print(output) done.Done() }() } func count(dna []byte, n int) map[uint64]uint64 { tbl := make(map[uint64]uint64, (2<<16)+1) for i, end := 0, len(dna)-n; i < end; i++ { tbl[compress(dna[i:i+n])]++ } return tbl } func compress(dna []byte) uint64 { var val uint64 for i, dnac := 0, len(dna); i < dnac; i++ { val = (val << 2) | uint64(dna[i]) } return val } func compStr(dna string) uint64 { raw := []byte(dna) for i, rawc, c := 0, len(raw), byte(0); i < rawc; i++ { c = raw[i] switch c { case 'A': c = 0 case 'T': c = 1 case 'G': c = 2 case 'C': c = 3 } raw[i] = c } return compress(raw) } func decompToBytes(compDNA uint64, n int) []byte { buf := bytes.NewBuffer(make([]byte, 0, n)) var c byte for i := 0; i < n; i++ { switch compDNA & 3 { case 0: c = 'A' case 1: c = 'T' case 2: c = 'G' case 3: c = 'C' } buf.WriteByte(c) compDNA = compDNA >> 2 } if n > 1 { return reverse(buf.Bytes()) } return buf.Bytes() } func reverse(s []byte) []byte { for i, j := 0, len(s)-1; i < j; i, j = i+1, j-1 { s[i], s[j] = s[j], s[i] } return s } func freqReport(tbl map[uint64]uint64, n int) string { seqs := make(seqSlice, 0, len(tbl)) var val, count, sum uint64 for val, count = range tbl { seqs = append(seqs, seq{nuc: decompToBytes(val, n), n: count}) sum += count } sort.Sort(seqs) var buf bytes.Buffer sumFloat, entry := float64(sum), seq{} for _, entry = range seqs { fmt.Fprintf(&buf, "%s %.3f\n", entry.nuc, (100*float64(entry.n))/sumFloat) } buf.WriteByte('\n') return buf.String() } type seq struct { nuc []byte n uint64 } type seqSlice []seq func (seq seqSlice) Len() int { return len(seq) } func (seq seqSlice) Swap(i, j int) { seq[i], seq[j] = seq[j], seq[i] } func (seq seqSlice) Less(i, j int) bool { if seq[i].n == seq[j].n { return bytes.Compare(seq[i].nuc, seq[j].nuc) < 0 } return seq[i].n > seq[j].n }  make, command-line, and program output logs Thu, 19 Jun 2014 06:17:48 GMT MAKE: /usr/local/src/go/bin/go build -o knucleotide.go-2.go_run 0.48s to complete and log all make actions COMMAND LINE: ./knucleotide.go-2.go_run 0 < knucleotide-input25000000.txt PROGRAM OUTPUT: A 30.295 T 30.151 C 19.800 G 19.754 AA 9.177 TA 9.132 AT 9.131 TT 9.091 CA 6.002 AC 6.001 AG 5.987 GA 5.984 CT 5.971 TC 5.971 GT 5.957 TG 5.956 CC 3.917 GC 3.911 CG 3.909 GG 3.902 1471758 GGT 446535 GGTA 47336 GGTATT 893 GGTATTTTAATT 893 GGTATTTTAATTTATAGT Revised BSD license   Home   Conclusions   License   Play  
__label__pos
0.988643
Using lifecycle ignore_changes for attributes managed outside of terraform Hi, I have some deployed instances and EBS resources in AWS, all of which have EC2 tags on. I want to ensure that future changes ignore some but not all of the tags. I’ve followed the documentation but I’m getting some unexpected behaviour for tags changed outside of terrform. I’m using terraform v0.13.6 and aws provider 3.58. The docs (The lifecycle Meta-Argument - Configuration Language | Terraform by HashiCorp) state “The ignore_changes feature is intended to be used when a resource is created with references to data that may change in the future, but should not affect said resource after its creation. In some rare cases, settings of a remote object are modified by processes outside of Terraform, which Terraform would then attempt to “fix” on the next run. In order to make Terraform share management responsibilities of a single object with a separate process, the ignore_changes meta-argument specifies resource attributes that Terraform should ignore when planning updates to the associated remote object.” Since I have processes outside of terraform that add new tags, this sounds exactly what I’m looking for. ignore_changes seems to work with tags which are managed by terraform, but tries to set tags not managed by terraform to null. Here’s some sample config: resource "aws_instance" "fileserver" { ... lifecycle { ignore_changes = [ user_data, tags["App_version"], tags["Config_version"], tags["hostname"], ] } tags = { "App_version" = local.app_version "Config_version" = local.config_version } ... } resource "aws_ebs_volume" "volume" { ... lifecycle { ignore_changes = [ tags["App_version"], tags["Config_version"], tags["Drive_Letter"], ] } tags = { "App_version" = local.app_version "Config_version" = local.config_version } } My output is “unexpected” because if I run a tf plan where I have changed the input values for “Config_version”, this tag is ignored as expected, but the tags which are changed outside of terraform are not ignored, but instead set to null: # module.storage.aws_instance.fileserver will be updated in-place ~ resource "aws_instance" "fileserver" { ... ~ tags = { "App_version" = "develop" "Config_version" = "feature/QWE-1234" - "hostname" = "EC2AMAZ-5JJJ44G" -> null } } # module.storage.aws_ebs_volume.volume will be updated in-place ~ resource "aws_ebs_volume" "volume" { ... ~ tags = { "App_version" = "develop" "Config_version" = "feature/QWE-1234" - "Drive_Letter" = "D:" -> null } } Are the docs wrong, is there a bug, or was this feature introduced into a later version of terraform? Help appreciated. Thanks. Looks like this was fixed in v0.14.0: Thanks.
__label__pos
0.843518
Title: Managing email servers by prioritizing emails Kind Code: A1 Abstract: Disclosed are email server management methods and systems that protect the ability of the infrastructure of the email server to process legitimate emails in the presence of large spam volumes. During a period of server overload, priority classes of emails are identified, and emails are processed according to priority. In a typical embodiment, the server sends emails sequentially in a queue, and the queue has a limited capacity. When the server nears or reaches that capacity, the emails in the queue are analyzed to identify priority emails, and the priority emails are moved to the head of the queue. Inventors: Sen, Subhabrata (New Providence, NJ, US) Haffner, Patrick (Atlantic Highlands, NJ, US) Spatscheck, Oliver (Randolph, NJ, US) Venkataraman, Shobha (Pittsburgh, PA, US) Application Number: 11/977243 Publication Date: 04/30/2009 Filing Date: 10/24/2007 Primary Class: International Classes: G06F15/16 View Patent Images: Primary Examiner: PFIZENMAYER, MARK C Attorney, Agent or Firm: AT&T CORP. (ROOM 2A207, ONE AT&T WAY, BEDMINSTER, NJ, 07921, US) Claims: 1. Method for server management of email wherein the server receives X emails sequentially in an input queue, and sends E emails to email subscribers sequentially in an output queue, and the server queue has a capacity of C emails, comprising the steps of: 1) analyzing the emails to identify a class P of priority emails, where P is a fraction of X, 2) moving the P emails to the head of the E email queue. 2. The method of claim 2 wherein E is less than X. 3. The method of claim 2 wherein steps 1) and 2) are performed when X is approximately equal to C. 4. The method of claim 2 wherein steps 1) and 2) are performed when X is greater than 75% of C. 5. The method of claim 1 wherein the E emails comprise spam emails S and legitimate emails L. 6. The method of claim 5 wherein the P emails comprises L emails and a portion of S emails. 7. The method of claim 6 wherein the P emails are identified by identifying a least a portion of S emails, and subtracting the portion of S emails from E. 8. The method of claim 1 wherein the P emails are identified based on the reputation of emails. 9. The method of claim 8 wherein the reputation is based on IP address. 10. The method of claim 8 wherein the reputation is based on IP cluster identification. 11. The method of claim 8 wherein the reputation is based on both IP address and IP cluster identification. 12. The method of claim 1 wherein the P emails are identified based on the persistence of emails. 13. The method of claim 12 wherein the persistence is based on IP address. 14. The method of claim 12 wherein the persistence is based on IP cluster identification. 15. The method of claim 12 wherein the persistence is based on both IP address and IP cluster identification. Description: FIELD OF THE INVENTION This invention relates to systems and methods for prioritizing emails during periods of overload in an email server. More specifically, it involves sorting emails to establish one or more priority email classes, and queuing emails by priority class during periods of email server overload. BACKGROUND OF THE INVENTION Email has emerged as an indispensable and ubiquitous means of communication and is arguably one of the “killer” applications on the Internet. In many businesses, emails are at least as important as telephone calls, and in private communication emails have replaced writing letters by a large extent. Unfortunately, the utility of email is increasingly diminished by an ever larger volume of spam requiring both mail server and human resources to handle. Considerable effort has focused on reducing the amount of spam an email user will receive. Most Internet Service Providers (ISPs) operate some type of spam filtering to identify and remove spam emails before they are received by the end-user. Email software on an end-user's PC might add an additional layer of filtering to remove this unwanted traffic based on the typical email patterns of the end-user. On the other hand, there has been less attention paid to how these large volume of spam messages impact the ISP mail infrastructure which has to receive, filter and deliver mail appropriately. Spam is typically sent from zombies, and to a smaller extent, from open mail relays. Since zombie networks are very large, the spam that an attacker can generate is extremely elastic. The attacker can easily generate far many more messages per second than even the largest mail server can receive or process. However, the spammer has no interest in crashing a mail server since that would prevent the spam emails from being delivered. At the same time, there is a clear incentive to send large volumes of spam—the more spam a spammer sends the more likely it is that some of the spam will penetrate the spam filters deployed by ISPs. Given these observations, it is unsurprising that spammers would try to maximize the amount of spam they send by increasing the load on the mail infrastructure to a point at which the most spam will be received. In fact, this has been observed on mail servers of large ISPs. Mail servers typically respond to overloads by dropping emails at random. If the spammer increases the spam volume, more spam is likely to get accepted by the mail server. Thus, the spammer's optimal operation point is not the maximum capacity of the mail server, but the maximum load before the mail server will crash. This indicates that the approach of throwing more resources at the problem does not work in this case: increasing the mail server capacity will not work, unless it can be increased to a point larger than the largest botnet available to the spammer. This is typically not economically feasible, and so a different approach is needed. While it is not the objective of spammers to overload the server, overload conditions in servers do occur as the result of large spam volume, and result in denial of service (DoS) for at least some users. DoS events may also occur as the result of deliberate overloads caused by one or more malicious users. These are referred to as DoS attacks. Small email servers, serving, for example, local area networks (LANS) are especially susceptible to DoS attacks. BRIEF STATEMENT OF THE INVENTION We have designed systems, and operation of systems, that prevent or reduce either of these forms of DoS. In the primary case, these protect the ability of the infrastructure of an email server to process legitimate emails in the presence of large spam volumes. They operate by identifying priority classes of emails, and processing emails according to priority during a period of server overload. In this description, this operation will be referred to as priority sorting. In one embodiment, priority sorting is invoked by the server when the server volume is at or near capacity. In this embodiment, the server sends emails sequentially in a queue, and the queue has a limited capacity. When the server nears or reaches that capacity, the emails in the queue are analyzed to identify priority emails, and the priority emails are moved to the head of the queue. In another embodiment, which recognizes that once the tools for implementing priority sorting are in place for use during overload conditions, the option exists for operating the server using priority sorting during normal (non-overload) conditions as well. To implement priority sorting, it is necessary to invoke one or more methods for identifying priority email. The priority email is classified here as legitimate email, and can be categorized by identifying the legitimate email directly, or by deriving the legitimate email by identifying and separating out spam, or combinations of both. BRIEF DESCRIPTION OF THE DRAWING The invention may be better understood when considered in conjunction with the drawing in which: FIG. 1 is a plot of daily email volume vs. attempted SMTP connection, attempts received, emails where SpamAssassin™ is applied, and non-spam messages; FIG. 2 shows cumulative distribution function (CDF) of the spam ratios of individual IPs: FIG. 3 is a plot of legitimate emails sent vs. IP spam-ratio; FIG. 4 is a plot of spam emails sent vs. IP spam-ratio; FIG. 5 is a plot of the persistence in days of IP addresses; FIG. 6 is a plot of the persistence in days of good IP addresses; FIG. 7 is a plot of the persistence in days of IP addresses; FIG. 8 is a plot of the persistence in days of spam IP addresses; FIG. 9 shows the CDF of the frequency-fraction excess for several good k-sets; FIG. 10 shows the fraction of spam sent by spam IP addresses and spam clusters; FIG. 11 is a plot similar to FIG. 10 for legitimate email; and FIGS. 12 and 13 show persistence of network-aware clusters. FIG. 12 shows spam and FIG. 13 shows legitimate emails. DETAILED DESCRIPTION OF THE INVENTION Most known spam control techniques use a form of blacklist. Various forms of whitelists have also been proposed, but whitelists are inherently restrictive and thus typically not widely used. However, we propose a new variation of a whitelist approach to address the problem of server overload. The two main categories of emails discussed herein, i.e. legitimate emails and spam emails, are well known and easily recognized categories. Legitimate emails have information content and are sent usually once to a limited number of recipients. Spam emails typically have advertising content and are sent once or more than once to a large number of recipients, e.g. more than 50 recipients. In between there is a significant volume of email that is legitimate but sent to a large number of recipients, e.g. inter-company alerts, subscriber lists, etc., as well as a significant volume of spam that may initially be sent to a relatively limited number of relay recipients (e.g. zombies, i.e, computers of innocent users that are co-opted by a spam sender to relay spam to an innocent user's address list). The objective of the invention is to identify a class of legitimate emails with a relatively high confidence level. These are defined as “priority” emails. The focus of the invention is a technique to protect the ability of mail server infrastructure to process legitimate emails in the presence of large spam volumes. The goal is to increase the amount of legitimate mail that the server processes when under overload, and gain a performance improvement over the current approaches of dropping mail at random. To address this problem specifically, incoming emails are dropped selectively during overload situation. The selection process may be viewed from two related but distinct perspectives. One, the selected emails may be emails that are identified, with a high level of confidence, as spam, or: two, emails that are identified, with a high level of confidence, as legitimate emails. The email queue in the server is modified in the first case by dropping the spam emails from the queue, or in the second case by moving the legitimate emails to the front of the queue. The result in terms of averting an overloaded server is qualitatively the same. But the selection process may be different. It should be understood that since the goal is to maximize legitimate mail during overload, the priorities resulting from the selection process are different from regular spam-filtering. Spam filtering methods attempt to identify all spam. The approach here only requires identification of a significant portion of spam. Thus the selection process used here is much less demanding, and therefore less costly, than most spam filtering programs. Likewise, if the selection process is aimed at identifying legitimate emails that selection may be inexact also. The precision of the two selection approaches can be expressed in general as: • 1. Identify SOME of the spam email, or • 2. Identify ALL of the legitimate email, but since this identification can include SOME of the spam email, the selection need not be exact. To implement the inexact selection processes just mentioned, past historical behavior of IP addresses that send email is used to predict the likelihood of an incoming email being legitimate or spam, and of using IP-address reputations to drive the selective drop policy. This is referred to as “reputation”. The advantages of an IP-address reputation based filtering scheme are the ease of which the information can be collected, and the difficulty a spammer faces to hide the IP address of the zombie or open mail relay s/he utilizes. Obviously, using the IP address for classification is substantially cheaper then any content-based scheme. In fact, IP address based prioritization can even be implemented on modern routers or switches and can therefore be used to offload the processing of rejected senders from the mail server entirely. Further, IP based classification can be quite accurate, as demonstrated below. Consider that “good” mail servers, which are mail servers that try to actively block outgoing spam, typically belong to large organizations or ISPs, and rarely switch their IP address. On the other hand spammers mainly rely on botnets as well as poorly managed mail servers to relay their spam. Therefore, their IP addresses change more frequently, but stay within the IP prefix ranges. In some cases, these IP prefixes can be used as markers for compromised or poorly managed hosts. This leads to the hypothesis that good mail servers are mostly good and stay mostly good for a long time, and that bad prefixes send mainly spam and stay bad for a long time. If this hypothesis holds, the properties of both legitimate mail and spam can be used prioritize legitimate mail as needed. To verify useful selection processes, an extensive measurement study was performed to understand IP-based properties of legitimate mail and spam. With that data a simulation study was performed to evaluate how these properties can be used to prioritize legitimate mail when the mail server is overloaded. It was demonstrated that a suitable reputation-based policy has a potential performance improvement factor of 3 over the state-of-the-art, in terms of amount of legitimate mail accepted. While a very significant quantity of spam comes from IP addresses that are ephemeral, a significant of the legitimate mail volume comes from IP addresses that last for a long time. This suggests that using the history of good IP addresses—IP addresses that send a lot of legitimate mail—can be used as a mechanism for prioritizing mail in spam mitigation. Such an approach would be complementary to the usual blacklisting approaches. The analysis performed also explored so-called network-aware clusters as candidates that may exploit structure in the IP addresses. Results suggest that IP addresses responsible for the bulk of the spam are well-clustered. Clusters responsible for the bulk of the spam are very long-lived. This suggests that network-aware clusters may be used in place of individual IP addresses as a reputation scheme to identify spammers, many of whom are ephemeral. The cluster reputation selection process, while theoretically less exact than the IP address reputation process, is potentially easier and less expensive to implement. Since spam is so pervasive, much effort has been expended in mitigating spam, and understanding the characteristics of spammers. Traditionally, the two primary approaches to spam mitigation have used content-based spam-filtering and DNS blacklists. Content-based spam-filtering software is typically applied at the end of the mail processing queue, and there has been a lot of research in content-based analysis, and understanding its limits. Content-based analysis has been proposed to rate-limit spam at the router. However, content-based analysis is expensive to implement, and in some cases raises privacy concerns. The invention described here does not consider content of mail, but rather focuses on the history and structure of the IP address. DNS blacklists are another popular way to reduce spam. Studies on DNS blacklisting have shown that over 90% of the spamming IP addresses were present in at least one blacklist at their time of appearance. The invention described here involves selection that is complementary to blacklisting. The focus is to develop a whitelist of legitimate mail, typically using a reputation mechanism. Yet another approach to spam identification is a greylist process that delays incoming emails if recent emails from a mail server have been identified as spam, or if no history for a given mail server exists. In contrast, the selection methods recommended for use with the invention provide a more detailed analysis of how predictable the spam behavior of a mail server identified by an IP address is, using more up-to-date data. In some embodiments, the identification of good and bad mail servers is extended to clusters of IP addresses, and a continuum rather than a binary decision is used to accept or reject incoming mail. Data developed for the analysis consists of traces from the mail server of a large company serving one of the corporate locations with approximately 700 mailboxes taken over a period of 166 days from January to June 2006. The location runs a PostFix mail server with extensive logging that records the following: (a) every attempted SMTP connection, with its IP address and time stamp (b) whether the connection was rejected, along with a reason for rejection, (c) whether the connection was accepted, results of the mail server's customized spam-filtering checks, and if accepted for delivery, the results of running SpamAssassin™. FIG. 1 shows a daily summary of the data for six months. It shows four quantities each day: (a) the number of SMTP connection requests made (including those that are denied via blacklists), (b) the amount of mail received by the mail server, (c) the number of e-mails that were sent to SpamAssassin, and (d) the number of e-mails deemed legitimate by SpamAssassin. The relative sizes of these four quantities every day illustrates the scope of the problem: the spam is 20 times larger than the legitimate mail received. (In our data set, there were 1.4 million legitimate messages and 27 million spam messages.) Such a sharp imbalance indicates the possibility of a significant impact for applications like rate-limiting under server overload: if there is a way to prioritize legitimate mail, the server can handle it much more quickly, because the volume of legitimate mail is tiny in comparison to spam. In the analysis that follows, every message that is considered legitimate by SpamAssassin is counted as a legitimate message; every message that is considered spam by SpamAssassin, the mail server's spam filtering checks, or through denial by a blacklist is counted as spam. The behavior of individual IP addresses that send legitimate mail and spam can be analyzed with the goal of uncovering any significant differences in behavior patterns. The analysis focuses on the IP spam-ratio of an IP address, which is defined as the fraction of mail sent by the IP address that is spam. This is a simple, intuitive metric that captures the spamming behavior of an IP address: a low spam-ratio indicates that the IP address sends mostly legitimate mail; a high spam-ratio indicate that the IP address sends mostly spam. The goal is to see whether the historical communication behavior of IP addresses with similar spam-ratios yields clues to sufficiently distinguish between IP addresses of legitimate senders and spammers. As indicated earlier, the distinction between the legitimate senders and spammers need not be perfect; even with partially correct classification, benefit can be gained. For example, when all the mail cannot be accepted, a partial distinction would still help in increasing the amount of legitimate mail that is received. In the IP-based analysis, the following is addressed: • Distribution by IP Spam Ratio: What is the distribution of the number of IP addresses by their spam-ratio, and what fraction of legitimate mail and spam is contributed by IP addresses with different spam-ratios? • Persistence: Are IP addresses with low/high spam-ratios present in many days? If they are, do such IP addresses contribute to a significant fraction of the legitimate mail/spam? • Temporal Spam-Ratio Stability: Do many of the IP addresses that appear to be good on average fluctuate between having very low and very high spam-ratios? The answers to these three questions, taken together, gives an indication of the benefit derived in using the history of IP address behavior for the selection process used in the invention. Most IP addresses have a spam-ratio of 0% or 100% , but a significant amount of legitimate mail will come from IP addresses with spam-ratio exceeding zero. It is demonstrated below that a very significant fraction of the legitimate mail comes from IP addresses that persist for a long time, but only a small fraction of the spam comes from IP addresses that persist for a long time. It is also demonstrated below that most IP addresses have a very high temporal ratio-stability—they do not fluctuate between exhibiting a very low or very high spam ratio every day. Together, these three observations suggest that identifying IP addresses with low spam ratios that regularly send legitimate mail is useful in spam mitigation and prioritizing legitimate mail. To understand how IP-based filtering using spam ratio is useful and what kind of impact it has, the distribution of IP addresses and their associated mail volumes are studied as a function of the IP spam-ratios. Intuitively, we expect that most IP addresses either send mostly legitimate mail, or mostly spam, and that most of the legitimate mail and spam comes from these IP addresses. If this hypothesis holds, then for spam mitigation it will be sufficient if the IP addresses are identified as senders of legitimate mail or spammers. To test this hypothesis, the following two empirical distributions are identified: (a) the distribution of IP addresses as a function of the spam ratios, and (b) the distribution of legitimate mail/spam as a function of the spam ratio of the respective IP addresses. The first experiment shows that most IP addresses are present at either ends of the spectrum of spam ratios, but the second experiment shows that the distribution of legitimate mail volume is not as focused at the ends of the spectrum. The spam-ratio computed over a short time period is studied to understand the behavior of IP addresses, without being affected by their possible fluctuations in time. The analysis is for intervals of a day to cover possible time-of-day variations. FIG. 2 depicts, for a large number of randomly selected days across the observation period, the daily empirical cumulative distribution function (CDF) of the spam ratios of individual IPs that sent some email to the server on that day. This shows that for nearly six months, on any particular day, (i) most IP addresses send either mostly spam or mostly legitimate mail. (ii) Fewer than 1-2% of the active IP addresses have a spam-ratio of between 1%-99% , ie., there are very few IP addresses that send a non-trivial fraction of both spam and legitimate mail. (iii) the vast majority (nearly 90% ) of the IP addresses on any given day generate almost exclusively spam, having spam-ratios between 99%-100%. The above indicates that identifying IP addresses with low or high spam-ratios can identify most of the legitimate senders and spammers. For some applications, it would also be valuable to identify the IP addresses that send the bulk of the spam or the bulk of the legitimate mail. An example is the server overload problem, where the goal is to accept as much of the legitimate mail volume as possible. The distribution of the daily legitimate mail or spam volumes as a function of the IP spam-ratios are identified. IP addresses that have a spam-ratio of at most k are categorized as set Ik. FIG. 3 shows how the volume of legitimate mail sent by the set Ik depends on the spam-ratio k. Specifically, let Li(k) and Si(k)be the fractions of the total daily legitimate mail and spam that comes from all IPs in the set Ik, on day i. FIG. 3 plots Li(k)averaged over all the days, along with confidence intervals. FIG. 4 shows the analogous plot for the spam volume Si(k). These data show that the bulk of the legitimate mail (nearly 70% on average) comes from IP addresses with a very low spam-ratio (k≦5%). However, a modest quantity (over 7% on average) also comes from IP addresses with a high spam-ratio. (k≧80% ). It also shows that almost all (over 99% on average) of the spam sent every day comes from IP addresses with an extremely high spam-ratio (when k≧95% ). indeed the contribution of the IP addresses with a spam-ratios (k≦80% ) is a tiny fraction of the total. We observe that there is a sharp difference in how the distribution of legitimate mail and spam contributions vary with the spam-ratio k: There are two possible explanations for this more diffused behavior of the legitimate senders. First, spam-filtering software tends to be conservative, allowing some spam to marked as legitimate mail. Second, a lot of legitimate mail tends to come from large mail servers that cannot do perfect outgoing spam-filtering. Together the above results suggest that the IP spam-ratio appears to be a useful discriminating feature for spam mitigation. Specifically, assume a classification function that accepts all IP addresses with a spam-ratio of at most k, and rejects all IP addresses with a higher spam-ratio. Then, if k is set=95% , nearly all of the legitimate mail is accepted, and no more than 1% of the spam. The effectiveness of such a history-based classification function for spam mitigation depends both on the extent to which IP addresses are long lasting, how much of the legitimate email or spam are contributed by the long lasting IP addresses, and to what extent the spam ratio of an IP address varies over time. These effects are examined next. To understand how IP addresses can be identified as spammers or non-spammers, data is analyzed to determine whether there are legitimate long-term properties that can be exploited to differentiate between them. For example, it can be assumed that many of the IP addresses that send legitimate mail do so consistently, and a significant fraction of the legitimate mail is sent by these IP addresses. For this analysis, the spam ratio of each individual IP address is computed over the entire data set to show behavior over the lifetime of the address. Two properties are shown in this analysis: (i) IP addresses sending a lot of good mail last for a long time (persistence), and (ii) IP addresses sending a lot of good mail tend to have a bounded spam ratio each time they appear (temporal stability). These 2 properties directly influence the effectiveness of using historical reputation information for determining the “spaminess” of emails being sent by an individual IP address. Due to the community structure inherent in non-spam communication patterns, it seems reasonable that much legitimate mail will originate from IP addresses that appear and re-appear. Studies have also indicated that most of the spam comes from IP addresses that are extremely short-lived. If these hypotheses hold, together they suggest the existence of a potentially significant difference in the behavior of senders of legitimate mail and spammers with respect to persistence. This premise, and the quantifiable extent to which it holds, may be established by examining the persistence of individual IP addresses. The methodology proposed for understanding the persistence behavior of IP addresses is as follows: consider the set of all IP addresses with a low lifetime spam-ratio, and examine both how much legitimate mail they send, as well as how much of this is sent by IP addresses that are present for a long time. Such an understanding can indicate the potential of using a whitelist-based approach for mitigation in specified situations, like the server overload problem. If, for instance, the bulk of the legitimate mail comes from IP addresses that last for a long time, this property can be used to prioritize legitimate mail from long lasting IP addresses with low spam ratios. For this priority category the following definition is used: k-good IP address: an IP address whose lifetime spam-ratio is at most k. • A k-good set is the set of all k-good IP addresses. Thus, a 20-good set is the set of all IP addresses whose lifetime spam-ratio is no more than 20% . The number of IP addresses present in the k-good set for at least x distinct days is determined, as well as the fraction of legitimate mail contributed by IP addresses in the k-good set that are present in at least x distinct days. FIG. 5 shows the number of IP addresses that appear in at least x distinct days, for several different k, and drops by a factor of 10 to 2000 when x=10. FIG. 6 shows the fraction of the total legitimate mail that originates from IP addresses that are in the k-good set, and appear in at least x days, for each threshold k. Most of the IP addresses in a k-good set are not present very long, and the number of IP addresses falls quickly, especially in the first few days. However the contribution of IP addresses in a k-good set to the legitimate mail drops much more slowly as x increases. The result is that the few longer-lived IPs contribute to most of the legitimate mail from the a k-good set. For example, only 5% of all IP addresses in the 20-good set appear at least 10 distinct days, but they contribute to almost 87% of all legitimate mail for the 20-good set. FIGS. 6 and 7 indicate that, overall, IP addresses with low lifetime spam ratios (small k) tend to contribute a major portion of the total legitimate email, while only a small fraction of the IP addresses with a low lifetime spam-ratio addresses that appear over many days, constitute a significant portion of the legitimate mail. For instance, IP addresses in the 20-good set contribute 63.5% of the total legitimate mail received. Only 2.1% of those IP addresses are present for at least 30 days, but they contribute to over 50% of the total legitimate mail received. The graphs also suggest another trend: the longer an IP address lasts, the more stable its contribution to the legitimate mail. For example, 0.09% of the IP addresses in the 20-good set are for at least 60 days, but they contribute to over 40% of the total legitimate mail received. From this it can be inferred that an additional 1.2% of IP addresses in the 20-good set were present for 30-59 days, but they contributed only 10% of the total legitimate mail received. FIGS. 7 and 8 present a similar analysis of persistence for IP addresses with a high lifetime spam-ratio. These are “bad” IP addresses and are defined as: • k-bad IP address: A k-bad IP address is an IP address that has a lifetime spam-ratio of at least k. A k-bad set is the set of all k-bad IP addresses. FIG. 7 presents the number of IP addresses in the k-bad set that are present in at least x days, and FIG. 8 presents the fraction of the total spam sent by IP addresses in the k-bad set that are present in least x days. FIGS. 7 and 8 show that, overall, IP addresses with high lifetime spam ratios (large k) tend to contribute almost all of the spam, and most of these high spam-rate IPs last only a short time and account for a large proportion of the overall spam. It also shows that the small fraction of these IPs that do last several days still contribute a significant fraction of the overall spam. Only 1.5% of the IP addresses in the 80-bad set appear in at least 10 distinct days, and these contribute 35.4% of the volume of spam from the 80-bad set, and 34% of the total spam. The difference is more pronounced for 100-bad IP addresses: 2% of the 100-bad IP addresses last for 10 distinct days, and contribute 25% of the total spam volume. As in the case of the k-good IP addresses, the volume of spam coming from the k-bad IP addresses tends to get more stable with time. The above results have an implication for the design of spam filters, especially for applications where the goal is to prioritize legitimate mail, rather than discard the spam. While the spamming IP addresses that are persistent can be blacklisted, the scope of a purely blacklisting approach is limited. On the other hand, a very significant fraction of the legitimate mail can be prioritized using the sender history of the legitimate mail. The IP addresses in the k-good set can also be analyzed for temporal stability, i.e. is an IP address that appears in a k-good set (for small values of k) likely to have a high spam-ratio? The focus in this analysis is on k-good IP addresses; the results for the k-bad IP addresses are similar. For each IP address in a k-good set, how often does the daily spam-ratio exceed k (normalized by the number of appearances). This quantity is defined as the frequency-fraction excess. The CDF of the frequency-fraction excess of all IP addresses in the k-good set is plotted. Intuitively, the distribution of the frequency-fraction excess is a measure of how many IP addresses in the k-good set exceed k, and how often. FIG. 9 shows the CDF of the frequency-fraction excess for several k-good sets. It shows that the majority of the IP addresses in each k-good set have a frequency-fraction excess of 0, and that 95% have a frequency fraction excess of at most 0.1. To understand the implication of this to the temporal stability of IP addresses, the k-good set for k=20 is analyzed. This is the set of IP addresses with a lifetime spam-ratio bounded by 20% . Note that the frequency-fraction excess of 0 for 95% of the IP addresses implies that 95% of IP addresses in this k-good set do not send more than 20% spam any day. Note that 4.75% of the IP addresses in this k-good set have a frequency-fraction excess between 0-20%, which implies that for 80% of their appearances, 99.75% IP addresses have a daily spam ratio bounded by k=20% . FIG. 9 shows that for many k-good sets with small k-values, only a few IP addresses have a significant frequency-fraction excess. This implies that most IP addresses in each set do not exceed the value k. Since they would need to exceed k often to significantly change their spamming behavior, it follows that most IP addresses in the k-good set do not change spamming behavior significantly. In addition, the frequency-fraction excess is a strict measure, since it increases that even when k is exceeded slightly. Similarly, the measure that increases only when k is exceeded by 5% is computed. No more than 0.01% of IP address in the k-good set exceed k by 5%. Since the metric here is the temporal stability of IP addresses that last a long time, the frequency fraction-excess distribution for IP addresses that last 10, 20, 40 and 60 days is analyzed. In each case, almost no IP address exceeds k by 5% . The conclusion from this is that of the IP addresses present in the 20-good set, fewer than 0.01% have a daily spam-ratio exceeding 25% on any day throughout their lifetime. Fewer than 1% of them have a daily spam-ratio exceeding 20% for more than one-tenth of their appearances. Thus most IP addresses in k-good sets do not fluctuate significantly in their spamming behavior; and most that appear to be good on an average are good every individual day as well. This result allows an analysis of the behavior of k-good sets of IP addresses, constructed over their entire lifetimes, and use of that analysis to understand implications to the behavior in the daily time intervals. The analysis of these three properties of IP addresses indicates that a significant fraction of the legitimate mail comes from IP addresses that persistently appear in the traffic. These IP addresses tend to exhibit very stable behavior: they do not fluctuate significantly between sending spam and legitimate mail. However, there is still a significant portion of the mail that cannot be accounted for through the use of IP addresses only. These results lend weight to the hypothesis that spam mitigation efforts can benefit non-trivially by preferentially allocating resources to the stable and persistent senders of legitimate mail. A limitation of reputation schemes based on historical behavior of individual IP addresses is that while they are able to discern IPs that appeared in the past, they may not be very useful in distinguishing between newcomer legitimate senders of spam or legitimate emails. To address this issue, the data can be analyzed to determine if there are coarser aggregations, other than individual IP addresses, that might exhibit more persistence, and afford more effective discrimination power for spam mitigation. The premise is that for IP addresses with little or no past history, their current reputation can be derived based on the historical reputation of the aggregation they belong to. To implement this, network-aware clusters of IP addresses are used. Network-aware clusters are a set of unique network IP prefixes collected from a wide set of Border Gateway Protocol (BGP) routing table snapshots. An IP address belongs to a network-aware cluster if a prefix matches the prefix associated with the cluster. The motivation behind using network-aware clustering is that clusters represent IP addresses that are close in terms of network topology and, with high probability, represent regions of the IP space that are under the same administrative control and share similar security and spam policies. Thus they provide a mechanism for reputation-based classification of IP addresses. Analysis similar to that described above indicates that cluster spam-ratio is useful as an approximation of the IP spam-ratio described above. FIG. 10 shows how the volume of spam sent by IP addresses with a cluster or IP spam-ratio of at most k varies with k. Specifically, let C Si(k) and ISi(k) be the fraction of spam sent by the IP addresses with a cluster spam ratio (respectively IP spam ratio) of at most k on day i. FIG. 10 plots (i) C Si(k)and ISi(k) averaged over all the days in the data set, as a function of k, along with confidence intervals. These data show that almost all (over 95%) of the spam every day comes from IPs in clusters with a very high cluster spam-ratio (over 90%). A similar fraction (over 99% on average). of the spam every day comes from IP addresses with a very high IP spam-ratio (over 90%). This suggests that spammers responsible for a high volume of the total spam may be closely correlated with the clusters that have a very high spam-ratio. The graph indicates that if we use a spam ratio threshold of k≦90% for spam mitigation, using the IP spam-ratio rather than their cluster spam-ratio as the discriminating feature, would identify less than 2% additional spam. This suggests that cluster spam-ratios are a good approximation to IP spam-ratios for identifying the bulk of spam sent. Analogous to the earlier spam study, the distribution of legitimate mail according to cluster spam-ratios is considered. This is compared with IP spam-ratios in FIG. 11. Let C Li(k) and ILi(k) be the fraction of legitimate mail sent by IPs with a cluster spam-ratio and IP spam ratio respectively, of at most k. FIG. 11 plots C Li(k) and ILi(k) averaged over all the days in the data set, as a function of k, along with confidence intervals. FIG. 11 shows that a significant amount of legitimate mail is contributed by clusters with both low and high spam-ratios. A significant fraction of the legitimate mail (around 45% on average) comes from IP addresses with a low cluster spam-ratio (k≦20% ). However, a much larger fraction of the legitimate mail (around 70% , on average) originates from IP addresses with a similarly low IP spam-ratio. These data reveal that with spam-ratios as high as 30-40%, the cluster spam-ratios only distinguish, on average, around 50% of the legitimate mail. By contrast, IP spam-ratios can distinguish as much as 70%. This suggests that IP addresses responsible for the bulk of legitimate mail are less correlated with clusters of low spam-ratio. However, FIG. 11 suggests that, if the threshold is set to 90% or higher, a relatively small penalty is incurred in both legitimate mail acceptance and spam. However, there are two additional considerations. First, the bulk of the legitimate email comes from persistent k-good IP addresses. This suggests that more legitimate email can be identified by considering the persistent k-good IP addresses, in combination with cluster-level information. Second, for some applications, the correlation between high cluster spam-ratios and the bulk of the spam may be sufficient to justify using cluster-level analysis. For example, under the existing distribution of spam and legitimate mail, using a high cluster spam-ratio threshold would be sufficient to reduce the total volume of the mail accepted by the mail server. This has general implications for the server overload problem. Similar to the study of IP addresses, persistence is also a useful means for evaluating network-aware clusters. A cluster is considered to be present on a given day if at least one IP address that belongs to that cluster appears that day. Earlier results showed that clusters were at least as (and usually more) temporally stable than IP addresses. As in the earlier IP address analysis, k-good and k-bad cluster categories are used, and are based on the lifetime cluster spam-ratio: the ratio of the total spam mail sent by the cluster to the total mail sent by it over its lifetime. These are defined specifically as: • A k-good cluster is a cluster of IP addresses whose lifetime cluster spam-ratio is at most k. The k-good cluster-set is the set of all k-good clusters. • A k-bad cluster is a cluster of IP addresses whose lifetime cluster spam-ratio is at least k. The k-bad cluster-set is the set of all k-bad clusters. FIG. 12 examines the legitimate mail sent by k-good clusters, for small values of k. The k-good clusters, even when k=30% , contribute less than 40% of the total legitimate mail. However, the contribution from long-lived clusters is far more than from long-lived individual IPs. The difference from FIG. 6 is striking; indeed, k-good clusters (for all k) present for at least ten days contribute to almost 100% of total legitimate mail coming from k-good cluster-set. Further, k-good clusters present for at least 60 days contribute to nearly 99% of the legitimate mail from the k-good cluster set. This implies that any cluster accounting for a non-trivial volume of legitimate mail is present for at least 60 days. The legitimate mail volume drops to 90% of the total k-good cluster-set only in the case of clusters present for more than 120 days. FIG. 13 presents the same analysis for k-bad clusters. Here, there are some striking differences from the k-good clusters. First, the 90-bad cluster-set contributes nearly 95% of the total spam volume. A much larger fraction of spam comes from long-lived clusters than from long-lived IPs (FIG. 8). For example, over 95% of the spam in the 90-bad cluster set is contributed by clusters present for at least 10 days. This is in sharp contrast to the k-bad IP addresses, where only 20% of the total spam volume comes from IP addresses that last 20 or more days. Thus it is demonstrated that long-lived clusters tend to contribute the bulk of both legitimate emails and spam, and that network-aware clustering can be used to address the problem of transience of IP addresses in developing history-based reputations of IP addresses. Measurements show that senders of legitimate mail demonstrate stability and persistence, while spammers do not. However, the bulk of high volume spammers appears to be clustered within some network-aware clusters that persist very long. Together, this suggests a useful reputation mechanism based on the history of an IP address, and the history of a cluster to which it belongs. However, because mail rejection mechanisms should be conservative, such a reputation-based mechanism is primarily useful for prioritizing legitimate mail, rather than discarding suspected spammers. An email server has a finite capacity of the number of mails that can be processed in any time interval, and may choose the connections it accepts or rejects. As indicated earlier, the goal of the invention is for the email server to selectively accept connections in order to maximize the legitimate mail accepted. Email server overload is a significant problem. For example, assume an email server can process 100 emails per second, will start dropping new incoming SMTP connections when its load reaches 100 emails per second, and crashes if the offered load reaches 200 emails per second. Assume also that 20 legitimate emails are received per second. In such a scenario the spammer could increase the load of the mail server to 100% by sending 80 emails per second, all of which would be received by the email server. Alternatively, the spammer could also increase the load to 199% by offering 179 spam email per second, in which case nearly half the requests would not be served. In summary, it is established above that there are history-based reputation functions that may be used for prioritizing email to address server overload issues. As is evident the target identifications are: • Identify legitimate email • Identify spam Either identification may be derived from the other by subtraction, but the distinction is important since neither identification mechanism is expected to be exact. In the usual case, the nearer to perfection of either identification, the more likely the error. That is, for the case of most reputation functions, the confidence level for the identification category declines as the percentage increases. In most cases of overload, it is sufficient to identify just enough spam to alleviate the overload condition. This may be done with a relatively high level of confidence. It is then not important if legitimate emails are identified at all. In making the identification, characteristics of the emails are assessed. These may include: • IP addresses • IP clusters • IP addresses and IP clusters In each case the characteristic may be evaluated according to: • email sending rate (emails per unit time) • persistence In the preferred embodiment the email queue for the server is processed according to priority of the emails when the server queue reaches X % of server capacity C, where X is a threshold of, for example, 75 or above. Various additional modifications of this invention will occur to those skilled in the art. All deviations from the specific teachings of this specification that basically rely on the principles and their equivalents through which the art has been advanced are properly considered within the scope of the invention as described and claimed.
__label__pos
0.923728
Developer-friendly On Scroll Animation JavaScript Library Category: Animation , Javascript | May 7, 2021 Author:johnsonfash Views Total:149 views Official Page:Go to website Last Update:May 7, 2021 License:MIT Preview: Developer-friendly On Scroll Animation JavaScript Library Description: A modern and developer-friendly, scroll-triggered animation JavaScript library for the web & mobile. It enables you to animate any elements using custom CSS animations when they come into view. How to use it: 1. Install and import the module. # NPM $ npm i onscroll-animation import OnScrollAnimation from "onscroll-animation"; 2. Or include the bundled JavaScript file on the page. <script src="/dist/animate.bundle.js"></script> 3. Create a new instance of the OnScrollAnimation and define your own CSS animations for each element as follows: var animate = new OnScrollAnimation({ ".element1": { parameters: { animationDuration: "1s", animationDelay: "2s", animationFillMode: "forwards", animationTimeFunction: "ease-in", pixelCorrection: "-200px", run: "once", // more CSS animation properties here }, to: ["transform: translateX(-150px)"] }, "element2": { from: ["left: -600px"], to: { left: "0px" } }, ".element3": { from: ["right: -600px"], to: ["right: 0px"] }, // more elements here }); 4. Start tracking the scroll event and apply the animations to the elements when they’re scrolled into view. animate.init(); 5. Pass animation parameters globally using the defaultParams method. animate.defaultParams([ "animation-duration: 1s", "animation-delay: 2s", "animation-fill-mode: forwards", "animation-time-function: ease-in", "pixel-correction: -200px", "run: once", // ... ]); 6. You can also apply custom animations to elements directly using CSS @keyframes. .custom { animation: customAnimation 1s forwards; } @keyframes customAnimation { from { ... } to { ... } } const animate = new OnScrollAnimation({ ".element1": { css: "custom" } }); You Might Be Interested In: Leave a Reply
__label__pos
0.837727
Hammer Hammer - 1 year ago 60 Javascript Question Javascript Module inheritence Parent Module, var Component = function () { var _componentName; var _test = 'ts'; return { getName: function(){ console.log('getName() is called'); console.log(this._componentName); console.log(_test); return _componentName; } }; }; module.exports = Component; Child module, var Component = require('./component'); var Skip_button = function () { var skipBtn = Component(); skipBtn._componentName = 'Skip_Btn'; return skipBtn; }; module.exports = Skip_button; In another place, when I call var skipBtn = Skip_button(); skipBtn.getName(); if it is console.log(this._componentName); in Component, the value can be printed out successfully. However, if it is console.log(_componentName); , undefine error will be there. Any idea? [Update] Check this out. It works as expected. http://jsfiddle.net/nedvedyy/Lvxqjo9v/1 so the question is still, why it does not work in my original code example above if console.log(this._componentName); is changed to console.log(_componentName); Answer Source In Component funtion _componentName and _test are not in the this scope. but however you are setting skipBtn._componentName in the Skip_button function means you are setting _componentName in the this scope. thats why console.log(this._componentName) printing 'Skip_Btn' var Component = function () { var _componentName; var _test = 'ts'; return { getName: function(){ console.log('getName() is called'); console.log(this._componentName); console.log(this._test); return _componentName; } }; }; var Skip_button = function () { var skipBtn = Component(); skipBtn._componentName = 'Skip_Btn'; return skipBtn ; }; var skipBtn = Skip_button(); skipBtn.getName(); If you run the above script this._test will print 'undefined' becuase _test is not in the this scope. var Component = function () { var _componentName = 'test'; var _test = 'ts'; return { getName: function(){ console.log('getName() is called'); console.log(_componentName); console.log(_test); return _componentName; } }; }; var Skip_button = function () { var skipBtn = Component(); skipBtn._componentName = 'Skip_Btn'; return skipBtn ; }; var skipBtn = Skip_button(); skipBtn.getName(); If you run the above code, console.log(_componentName) will print 'test' becasue it has some value
__label__pos
0.998619
How to use a JS file in a Smarty template? by beatrice.lesch , in category: PHP , a year ago How to use a JS file in a Smarty template? Facebook Twitter LinkedIn Telegram Whatsapp 2 answers by madisen.leannon , a year ago @beatrice.lesch  To use a JavaScript (JS) file in a Smarty template, you can follow these steps: 1. Save your JavaScript file in a location accessible to your web application. 2. In your Smarty template, add a <script> tag to include the JavaScript file. You can do this in the <head> section of your template, for example: 1 2 3 <head> <script src="path/to/your/javascript/file.js"></script> </head> 1. If your JavaScript code relies on Smarty variables or functions, you can pass them to your JavaScript code by assigning them to global JavaScript variables in your Smarty template. For example: 1 2 3 4 5 6 <head> <script src="path/to/your/javascript/file.js"></script> <script> var mySmartyVariable = "{$smartyVariable}"; </script> </head> In the above example, {$smartyVariable} is a Smarty variable that is assigned to a global JavaScript variable mySmartyVariable. 1. You can then use the global JavaScript variable in your JavaScript code as needed. For example: 1 alert(mySmartyVariable); This will display the value of the mySmartyVariable variable in an alert box. Note that it's generally considered good practice to include JavaScript files at the end of your HTML code (just before the closing </body> tag), to ensure that the page content is loaded before any scripts. Member by viviane , 4 months ago @beatrice.lesch  You can also use the {literal} tag in Smarty templates to prevent any conflicts with curly braces and special characters that may be present in your JavaScript code. Here's an example of how you can incorporate the {literal} tag when including a JavaScript file in your Smarty template: 1 2 3 4 5 <head> {literal} <script src="path/to/your/javascript/file.js"></script> {/literal} </head> Including the {literal} tag ensures that the content within it will be treated as raw text and not interpreted as Smarty tags. This can be particularly useful when you have a significant amount of JavaScript code that you want to include without Smarty interpretation. Additionally, you can use Smarty variables directly in your JavaScript code by using the $smarty object inside the {literal} block. Here's an example: 1 2 3 4 5 6 {literal} <script> var mySmartyVariable = {$smarty->getVariable('smartyVariable')->value}; console.log(mySmartyVariable); </script> {/literal} In this example, the getVariable method of the $smarty object is used to retrieve the value of the Smarty variable smartyVariable and assign it to a JavaScript variable mySmartyVariable. By combining these techniques, you can effectively include and work with JavaScript files in your Smarty templates while ensuring proper handling of Smarty variables and syntax.
__label__pos
0.99831
logo codesdope Virtual and abstract Before going to virtual function, let's first have a look at early binding and late binding. Binding means matching the function call with the correct function definition by the compiler. It takes place either at compile time or at runtime. Early Binding In early binding, the compiler matches the function call with the correct function definition at compile time. It is also known as Static Binding or Compile-time Binding. By default, the compiler goes to the function definition which has been called during compile time. So, all the function calls you have studied till now are due to early binding. You have learned about function overriding in which the base and derived classes have functions with the same name, parameters and return type. In that case also, early binding takes place. In function overriding, we called the function with the objects of the classes. Now let's try to write the same example but this time calling the functions with the pointer to the base class i.e., refernce to the base class' object. #include <iostream> using namespace std; class Animals { public: void sound() { cout << "This is parent class" << endl; } }; class Dogs : public Animals { public: void sound() { cout << "Dogs bark" << endl; } }; int main() { Animals *a; Dogs d; a= &d; a -> sound(); // early binding return 0; } Output This is parent class Now in this example, we created a pointer a to the parent class Animals. Then by writing a= &d , the pointer 'a' started referring to the object d of the class Dogs. a -> sound(); - On calling the function sound() which is present in both the classes by the pointer 'a', the function of the parent class got called, even if the pointer is referring to the object of the class Dogs. This is due to Early Binding. We know that a is a pointer of the parent class referring to the object of the child class. Since early binding takes place at compile-time, therefore when the compiler saw that a is a pointer of the parent class, it matched the call with the 'sound()' function of the parent class without considering which object the pointer is referring to. Late Binding In the case of late binding, the compiler matches the function call with the correct function definition at runtime. It is also known as Dynamic Binding or Runtime Binding. In late binding, the compiler identifies the type of object at runtime and then matches the function call with the correct function definition. By default, early binding takes place. So if by any means we tell the compiler to perform late binding, then the problem in the previous example can be solved. This can be achieved by declaring a virtual function. Virtual function Virtual Function is a member function of the base class which is overridden in the derived class. The compiler performs late binding on this function. To make a function virtual, we write the keyword virtual before the function definition. #include <iostream> using namespace std; class Animals { public: virtual void sound() { cout << "This is parent class" << endl; } }; class Dogs : public Animals { public: void sound() { cout << "Dogs bark" << endl; } }; int main() { Animals *a; Dogs d; a= &d; a -> sound(); return 0; } Output Dogs bark Since the function sound() of the base class is made virtual, the compiler now performs late binding for this function. Now, the function call will be matched to the function definition at runtime. Since the compiler now identifies pointer a as referring to the object 'd' of the derived class Dogs, it will call the sound() function of the class Dogs. If we declare a member function in a base class as a virtual function, then that function automatically becomes virtual in all of its derived classes. If we make any function inside a base class virtual, then that function becomes virtual in all its derived classes. This means that we don't need to declare that function as virtual separately in its derived classes. We can also call private function of derived class from a base class pointer by declaring that function in the base class as virtual. Compiler checks if the members of a class are private, public or protected only at compile time and not at runtime. Since our function is being called at runtime, so we can call any type of function, private or public as shown in the following example. #include <iostream> using namespace std; class Animals { public: virtual void sound() { cout << "This is parent class" << endl; } }; class Dogs : public Animals { private: virtual void sound() { cout << "Dogs bark" << endl; } }; int main() { Animals *a; Dogs b; a = &b; a->sound(); return 0; } Output Dogs bark Since the same function (virtual function) having different definitions in different classes is called depending on the type of object that calls the function, this is also a part of Polymorphism. Pure Virtual Function Pure virtual function is a virtual function which has no definition. Pure virtual functions are also called abstract functions. To create a pure virtual function, we assign a value 0 to the function as follows. virtual void sound() = 0; Here sound() is a pure virtual area. Abstract Class An abstract class is a class whose instances (objects) can't be made. We can only make objects of its subclass (if they are not abstract). Abstract class is also known as abstract base class. An abstract class has at least one abstract function (pure virtual function). Let's look at an example of abstract class. Suppose there are some employees working in a firm. The firm hires only two types of employees- either driver or developer. Now, you have to develop a software to store information about them. So, here is an idea - There is no need to make objects of employee class. We will make objects to only driver or developer. Also, both must have some salary. So, there must be a common function to know about salary. This need will be best accomplished with abstract class. So, we can make 'Employee' an abstract class and 'Developer' and 'Driver' its subclasses. #include <iostream> using namespace std; class Employee // abstract base class { virtual int getSalary() = 0; // pure virtual function }; class Developer : public Employee { int salary; public: Developer(int s) { salary = s; } int getSalary() { return salary; } }; class Driver : public Employee { int salary; public: Driver(int t) { salary = t; } int getSalary() { return salary; } }; int main() { Developer d1(5000); Driver d2(3000); int a, b; a = d1.getSalary(); b = d2.getSalary(); cout << "Salary of Developer : " << a << endl; cout << "Salary of Driver : " << b << endl; return 0; } Output Salary of Developer : 5000 Salary of Driver : 3000 The getSalary() function in the class Employee is a pure virtual function. Since the Employee class contains this pure virtual function, therefore it is an abstract base class. Since the abstract function is defined in the subclasses, therefore the function 'getSalary()' is defined in both the subclasses of the class Employee. You must have understood the rest of the code. Subclasses of an abstract base class must define the abstract method, otherwise, they will also become abstract classes. In an abstract class, we can also have other functions and variables apart from pure virtual function. Let's see one more example of abstract base class. #include <iostream> using namespace std; class Animals { public: virtual void sound() = 0; }; class Dogs { public: void sound() { cout << "Dogs bark" << endl; } }; class Cats { public: void sound() { cout << "Cats meow" << endl; } }; class Pigs { public: void sound() { cout << "Pigs snort" << endl; } }; int main() { Dogs d; Cats c; Pigs p; d.sound(); c.sound(); p.sound(); return 0; } Output Dogs bark Cats meow Pigs snort Interface Interface or Interface class is a class which is the same as abstract class with a difference that all its functions are pure virtual and it has no member variables. Its derived classes must implement each of its virtual functions i.e., provide definition to each of the pure virtual functions of the base class. Like an abstract class, we can't create objects of an interface. We can also say that interface is an abstract class with no member variables and all its member functions pure virtual. Name of an interface class often begins with the letter I. Let's see an example of it. class IShape {     public:         virtual getArea() = 0;         virtual getPerimeter() = 0; }; IShape is an interface because it contains only pure virtual functions. #include <iostream> using namespace std; class IShape { public: virtual int getArea() = 0; virtual int getPerimeter() = 0; }; class Rectangle : public IShape { int length; int breadth; public: Rectangle(int l, int b) { length = l; breadth = b; } int getArea() { return length * breadth; } int getPerimeter() { return 2*(length + breadth); } }; class Square : public IShape { int side; public: Square(int a) { side = a; } int getArea() { return side * side; } int getPerimeter() { return 4 * side; } }; int main() { Rectangle rt(7, 4); Square s(4); cout << "Rectangle :" << endl; cout << "Area : " << rt.getArea() << " Perimeter : " << rt.getPerimeter() << endl; cout << "Square :" << endl; cout << "Area : " << s.getArea() << " Perimeter : " << s.getPerimeter() << endl; return 0; } Output Rectangle : Area : 28 Perimeter : 22 Square : Area : 16 Perimeter : 16 So we just saw that IShape is an interface with two pure virtual functions. These virtual functions are implemented (defined) in its subclasses Rectangle and Square according to their requirements. So, an interface is just an abstract class with all pure virtual methods. Everything is practice Share what you know Doubt? Ask question Close Welcome.please sign up. Close Welcome.please login.
__label__pos
0.99443
blob: 1fa3ac0f18fc4f5dd35087452d4839dcedee4969 [file] [log] [blame] // Copyright 2009 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. package walk import ( "cmd/compile/internal/base" "cmd/compile/internal/ir" "cmd/compile/internal/typecheck" "cmd/compile/internal/types" "cmd/internal/src" ) // directClosureCall rewrites a direct call of a function literal into // a normal function call with closure variables passed as arguments. // This avoids allocation of a closure object. // // For illustration, the following call: // // func(a int) { // println(byval) // byref++ // }(42) // // becomes: // // func(byval int, &byref *int, a int) { // println(byval) // (*&byref)++ // }(byval, &byref, 42) func directClosureCall(n *ir.CallExpr) { clo := n.X.(*ir.ClosureExpr) clofn := clo.Func if ir.IsTrivialClosure(clo) { return // leave for walkClosure to handle } // We are going to insert captured variables before input args. var params []*types.Field var decls []*ir.Name for _, v := range clofn.ClosureVars { if !v.Byval() { // If v of type T is captured by reference, // we introduce function param &v *T // and v remains PAUTOHEAP with &v heapaddr // (accesses will implicitly deref &v). addr := ir.NewNameAt(clofn.Pos(), typecheck.Lookup("&"+v.Sym().Name)) addr.Curfn = clofn addr.SetType(types.NewPtr(v.Type())) v.Heapaddr = addr v = addr } v.Class = ir.PPARAM decls = append(decls, v) fld := types.NewField(src.NoXPos, v.Sym(), v.Type()) fld.Nname = v params = append(params, fld) } // f is ONAME of the actual function. f := clofn.Nname typ := f.Type() // Create new function type with parameters prepended, and // then update type and declarations. typ = types.NewSignature(nil, append(params, typ.Params().FieldSlice()...), typ.Results().FieldSlice()) f.SetType(typ) clofn.Dcl = append(decls, clofn.Dcl...) // Rewrite call. n.X = f n.Args.Prepend(closureArgs(clo)...) // Update the call expression's type. We need to do this // because typecheck gave it the result type of the OCLOSURE // node, but we only rewrote the ONAME node's type. Logically, // they're the same, but the stack offsets probably changed. if typ.NumResults() == 1 { n.SetType(typ.Results().Field(0).Type) } else { n.SetType(typ.Results()) } // Add to Closures for enqueueFunc. It's no longer a proper // closure, but we may have already skipped over it in the // functions list as a non-trivial closure, so this just // ensures it's compiled. ir.CurFunc.Closures = append(ir.CurFunc.Closures, clofn) } func walkClosure(clo *ir.ClosureExpr, init *ir.Nodes) ir.Node { clofn := clo.Func // If no closure vars, don't bother wrapping. if ir.IsTrivialClosure(clo) { if base.Debug.Closure > 0 { base.WarnfAt(clo.Pos(), "closure converted to global") } return clofn.Nname } // The closure is not trivial or directly called, so it's going to stay a closure. ir.ClosureDebugRuntimeCheck(clo) clofn.SetNeedctxt(true) // The closure expression may be walked more than once if it appeared in composite // literal initialization (e.g, see issue #49029). // // Don't add the closure function to compilation queue more than once, since when // compiling a function twice would lead to an ICE. if !clofn.Walked() { clofn.SetWalked(true) ir.CurFunc.Closures = append(ir.CurFunc.Closures, clofn) } typ := typecheck.ClosureType(clo) clos := ir.NewCompLitExpr(base.Pos, ir.OCOMPLIT, typ, nil) clos.SetEsc(clo.Esc()) clos.List = append([]ir.Node{ir.NewUnaryExpr(base.Pos, ir.OCFUNC, clofn.Nname)}, closureArgs(clo)...) for i, value := range clos.List { clos.List[i] = ir.NewStructKeyExpr(base.Pos, typ.Field(i), value) } addr := typecheck.NodAddr(clos) addr.SetEsc(clo.Esc()) // Force type conversion from *struct to the func type. cfn := typecheck.ConvNop(addr, clo.Type()) // non-escaping temp to use, if any. if x := clo.Prealloc; x != nil { if !types.Identical(typ, x.Type()) { panic("closure type does not match order's assigned type") } addr.Prealloc = x clo.Prealloc = nil } return walkExpr(cfn, init) } // closureArgs returns a slice of expressions that an be used to // initialize the given closure's free variables. These correspond // one-to-one with the variables in clo.Func.ClosureVars, and will be // either an ONAME node (if the variable is captured by value) or an // OADDR-of-ONAME node (if not). func closureArgs(clo *ir.ClosureExpr) []ir.Node { fn := clo.Func args := make([]ir.Node, len(fn.ClosureVars)) for i, v := range fn.ClosureVars { var outer ir.Node outer = v.Outer if !v.Byval() { outer = typecheck.NodAddrAt(fn.Pos(), outer) } args[i] = typecheck.Expr(outer) } return args } func walkMethodValue(n *ir.SelectorExpr, init *ir.Nodes) ir.Node { // Create closure in the form of a composite literal. // For x.M with receiver (x) type T, the generated code looks like: // // clos = &struct{F uintptr; R T}{T.M·f, x} // // Like walkClosure above. if n.X.Type().IsInterface() { // Trigger panic for method on nil interface now. // Otherwise it happens in the wrapper and is confusing. n.X = cheapExpr(n.X, init) n.X = walkExpr(n.X, nil) tab := ir.NewUnaryExpr(base.Pos, ir.OITAB, n.X) check := ir.NewUnaryExpr(base.Pos, ir.OCHECKNIL, tab) init.Append(typecheck.Stmt(check)) } typ := typecheck.MethodValueType(n) clos := ir.NewCompLitExpr(base.Pos, ir.OCOMPLIT, typ, nil) clos.SetEsc(n.Esc()) clos.List = []ir.Node{ir.NewUnaryExpr(base.Pos, ir.OCFUNC, methodValueWrapper(n)), n.X} addr := typecheck.NodAddr(clos) addr.SetEsc(n.Esc()) // Force type conversion from *struct to the func type. cfn := typecheck.ConvNop(addr, n.Type()) // non-escaping temp to use, if any. if x := n.Prealloc; x != nil { if !types.Identical(typ, x.Type()) { panic("partial call type does not match order's assigned type") } addr.Prealloc = x n.Prealloc = nil } return walkExpr(cfn, init) } // methodValueWrapper returns the ONAME node representing the // wrapper function (*-fm) needed for the given method value. If the // wrapper function hasn't already been created yet, it's created and // added to typecheck.Target.Decls. func methodValueWrapper(dot *ir.SelectorExpr) *ir.Name { if dot.Op() != ir.OMETHVALUE { base.Fatalf("methodValueWrapper: unexpected %v (%v)", dot, dot.Op()) } meth := dot.Sel rcvrtype := dot.X.Type() sym := ir.MethodSymSuffix(rcvrtype, meth, "-fm") if sym.Uniq() { return sym.Def.(*ir.Name) } sym.SetUniq(true) base.FatalfAt(dot.Pos(), "missing wrapper for %v", meth) panic("unreachable") }
__label__pos
0.923927
Marine Rust SDK The marine-rs-sdk empowers developers to write services suitable for peer hosting in peer-to-peer networks using the Marine Virtual Machine by enabling the wasm32-wasi compile target for Marine. API The procedural macros [marine] and [marine_test] are the two primary features provided by the SDK. The [marine] macro can be applied to a function, external block or structure. The [marine_test] macro, on the other hand, allows the use of the familiar cargo test to execute tests over the actual Wasm module generated from the service code. Function Export Applying the [marine] macro to a function results in its export, which means that it can be called from other modules or AIR scripts. For the function to be compatible with this macro, its arguments must be of the ftype, which is defined as follows: ftype = bool, u8, u16, u32, u64, i8, i16, i32, i64, f32, f64, String ftype = ftype | Vec<ftype> ftype = ftype | Record<ftype> In other words, the arguments must be one of the types listed below: one of the following Rust basic types: bool, u8, u16, u32, u64, i8, i16, i32, i64, f32, f64, String a vector of elements of the above types a vector composed of vectors of the above type, where recursion is acceptable, e.g. the type Vec<Vec<Vec<u8>>> is permissible a record, where all fields are of the basic Rust types a record, where all fields are of any above types or other records The return type of a function must follow the same rules, but currently only one return type is possible. See the example below of an exposed function with a complex type signature and return value: 1 // export TestRecord as a public data structure bound by 2 // the IT type constraints 3 #[marine] 4 pub struct TestRecord { 5 pub field_0: i32, 6 pub field_1: Vec<Vec<u8>>, 7 } 8 9 // export foo as a public function bound by the 10 // IT type contraints 11 #[marine] # 12 pub fn foo(arg_1: Vec<Vec<Vec<Vec<TestRecord>>>>, arg_2: String) -> Vec<Vec<Vec<Vec<TestRecord>>>> { 13 unimplemented!() 14 } Copied! Function Export Requirements wrap a target function with the [marine] macro function arguments must by of ftype the function return type also must be of ftype Function Import The [marine] macro can also wrap an extern block. In this case, all functions declared in it are considered imported functions. If there are imported functions in some module, say, module A, then: There should be another module, module B, that exports the same functions. The name of module B is indicated in the link macro (see examples below). Module B should be loaded to Marine by the moment the loading of module A starts. Module A cannot be loaded if at least one imported function is absent in Marine. See the examples below for wrapped extern block usage: Example 1 Example 2 1 #[marine] 2 pub struct TestRecord { 3 pub field_0: i32, 4 pub field_1: Vec<Vec<u8>>, 5 } 6 7 // wrap the extern block with the marine macro to expose the function 8 // as an import to the Marine VM 9 #[marine] 10 #[link(wasm_import_module = "some_module")] 11 extern "C" { 12 pub fn foo(arg: Vec<Vec<Vec<Vec<TestRecord>>>>, arg_2: String) -> Vec<Vec<Vec<Vec<TestRecord>>>>; 13 } Copied! 1 [marine] 2 #[link(wasm_import_module = "some_module")] 3 extern "C" { 4 pub fn foo(arg: Vec<Vec<Vec<Vec<u8>>>>) -> Vec<Vec<Vec<Vec<u8>>>>; 5 } Copied! Function import requirements wrap an extern block with the function(s) to be imported with the [marine] macro all function(s) arguments must be of the ftype type the return type of the function(s) must be ftype Structures Finally, the [marine] macro can wrap a struct making possible to use it as a function argument or return type. Note that only macro-wrapped structures can be used as function arguments and return types all fields of the wrapped structure must be public and of the ftype. it is possible to have inner records in the macro-wrapped structure and to import wrapped structs from other crates See the example below for wrapping struct: Example 1 Example 2 Example 3 1 #[marine] 2 pub struct TestRecord0 { 3 pub field_0: i32, 4 } 5 6 #[marine] 7 pub struct TestRecord1 { 8 pub field_0: i32, 9 pub field_1: String, 10 pub field_2: Vec<u8>, 11 pub test_record_0: TestRecord0, 12 } 13 14 #[marine] 15 pub struct TestRecord2 { 16 pub test_record_0: TestRecord0, 17 pub test_record_1: TestRecord1, 18 } 19 20 #[marine] 21 fn foo(mut test_record: TestRecord2) -> TestRecord2 { unimplemented!(); } Copied! 1 #[fce] 2 pub struct TestRecord0 { 3 pub field_0: i32, 4 } 5 6 #[fce] 7 pub struct TestRecord1 { 8 pub field_0: i32, 9 pub field_1: String, 10 pub field_2: Vec<u8>, 11 pub test_record_0: TestRecord0, 12 } 13 14 #[fce] 15 pub struct TestRecord2 { 16 pub test_record_0: TestRecord0, 17 pub test_record_1: TestRecord1, 18 } 19 20 #[fce] 21 #[link(wasm_import_module = "some_module")] 22 extern "C" { 23 fn foo(mut test_record: TestRecord2) -> TestRecord2; 24 } Copied! 1 mod data_crate { 2 use fluence::marine; 3 #[marine] 4 pub struct Data { 5 pub name: String, 6 pub data: f64, 7 } 8 } 9 10 use data_crate::Data; 11 use fluence::marine; 12 13 fn main() {} 14 15 #[marine] 16 fn some_function() -> Data { 17 Data { 18 name: "example".into(), 19 data: 1.0, 20 } 21 } 22 Copied! Structure passing requirements wrap a structure with the [marine] macro all structure fields must be of the ftype the structure must be pointed to without preceding package import in a function signature, i.eStructureName but not package_name::module_name::StructureName wrapped structs can be imported from crates Call Parameters There is a special API function fluence::get_call_parameters() that returns an instance of the CallParameters structure defined as follows: 1 pub struct CallParameters { 2 /// Peer id of the AIR script initiator. 3 pub init_peer_id: String, 4 5 /// Id of the current service. 6 pub service_id: String, 7 8 /// Id of the service creator. 9 pub service_creator_peer_id: String, 10 11 /// Id of the host which run this service. 12 pub host_id: String, 13 14 /// Id of the particle which execution resulted a call this service. 15 pub particle_id: String, 16 17 /// Security tetraplets which described origin of the arguments. 18 pub tetraplets: Vec<Vec<SecurityTetraplet>>, 19 } Copied! CallParameters are especially useful in constructing authentication services: 1 // auth.rs 2 use fluence::{marine, CallParameters}; 3 use::marine; 4 5 pub fn is_owner() -> bool { 6 let meta = marine::get_call_parameters(); 7 let caller = meta.init_peer_id; 8 let owner = meta.service_creator_peer_id; 9 10 caller == owner 11 } 12 13 #[marine] 14 pub fn am_i_owner() -> bool { 15 is_owner() 16 } Copied! MountedBinaryResult Due to the inherent limitations of Wasm modules, such as a lack of sockets, it may be necessary for a module to interact with its host to bridge such gaps, e.g. use a https transport provider like curl. In order for a Wasm module to use a host's curl capabilities, we need to provide access to the binary, which at the code level is achieved through the Rust extern block: 1 // Importing a linked binary, curl, to a Wasm module 2 #![allow(improper_ctypes)] 3 4 use fluence::marine; 5 use fluence::module_manifest; 6 use fluence::MountedBinaryResult; 7 8 module_manifest!(); 9 10 pub fn main() {} 11 12 #[marine] 13 pub fn curl_request(curl_cmd: Vec<String>) -> MountedBinaryResult { 14 let response = curl(curl_cmd); 15 response 16 } 17 18 #[marine] 19 #[link(wasm_import_module = "host")] 20 extern "C" { 21 fn curl(cmd: Vec<String>) -> MountedBinaryResult; 22 } Copied! The above code creates a "curl adapter", i.e., a Wasm module that allows other Wasm modules to use the the curl_request function, which calls the imported curl binary in this case, to make http calls. Please note that we are wrapping the extern block with the [marine]macro and introduce a Marine-native data structure MountedBinaryResult as the linked-function return value. Please not that if you want to use curl_request with testing, see below, the curl call needs to be marked unsafe, e.g.: 1 let response = unsafe { curl(curl_cmd) }; Copied! since cargo does not access to the marine macro to handle unsafe. MountedBinaryResult itself is a Marine-compatible struct containing a binary's return process code, error string and stdout and stderr as byte arrays: 1 #[marine] 2 #[derive(Clone, PartialEq, Default, Eq, Debug, Serialize, Deserialize)] 3 pub struct MountedBinaryResult { 4 /// Return process exit code or host execution error code, where SUCCESS_CODE means success. 5 pub ret_code: i32, 6 7 /// Contains the string representation of an error, if ret_code != SUCCESS_CODE. 8 pub error: String, 9 10 /// The data that the process wrote to stdout. 11 pub stdout: Vec<u8>, 12 13 /// The data that the process wrote to stderr. 14 pub stderr: Vec<u8>, 15 } 16 Copied! MountedBinaryResult then can be used on a variety of match or conditional tests. Testing Since we are compiling to a wasm32-wasi target with ftype constrains, the basic cargo test is not all that useful or even usable for our purposes. To alleviate that limitation, Fluence has introduced the [marine-test] macro that does a lot of the heavy lifting to allow developers to use cargo test as intended. That is, [marine-test] macro generates the necessary code to call Marine, one instance per test function, based on the Wasm module and associated configuration file so that the actual test function is run against the Wasm module not the native code. To use the [marine-test] macro please add marine-rs-sdk-test crate to the [dev-dependencies] section of Config.toml: 1 [dev-dependencies] 2 marine-rs-sdk-test = "0.2.0" Copied! Let's have a look at an implementation example: 1 use marine_rs_sdk::marine; 2 use marine_rs_sdk::module_manifest; 3 4 module_manifest!(); 5 6 pub fn main() {} 7 8 #[marine] 9 pub fn greeting(name: String) -> String { // 1 10 format!("Hi, {}", name) 11 } 12 13 #[cfg(test)] 14 mod tests { 15 use marine_rs_sdk_test::marine_test; // 2 16 17 #[marine_test(config_path = "../Config.toml", modules_dir = "../artifacts")] // 3 18 fn empty_string(greeting: marine_test_env::greeting::ModuleInterface) { 19 let actual = greeting.greeting(String::new()); // 4 20 assert_eq!(actual, "Hi, "); 21 } 22 23 #[marine_test(config_path = "../Config.toml", modules_dir = "../artifacts")] 24 fn non_empty_string(greeting: marine_test_env::greeting::ModuleInterface) { 25 let actual = greeting.greeting("name".to_string()); 26 assert_eq!(actual, "Hi, name"); 27 } 28 } Copied! 1. We wrap a basic greeting function with the [marine] macro which results in the greeting.wasm module 2. We wrap our tests as usual with [cfg(test)] and import the marine test crate. Do not import super or the local crate. 3. Instead, we apply the [marine_test] macro to each of the test functions by providing the path to the config file, e.g., Config.toml, and the directory containing the Wasm module we obtained after compiling our project with marine build. Moreover, we add the type of the test as an argument in the function signature. It is imperative that project build precedes the test runner otherwise the required Wasm file will be missing. 4. The target of our tests is the pub fn greeting function. Since we are calling the function from the Wasm module we must prefix the function name with the module namespace -- greeting in this example case as specified in the function argument. Now that we have our Wasm module and tests in place, we can proceed with cargo test --release. Note that using the releaseflag vastly improves the import speed of the necessary Wasm modules. The same macro also allows testing data flow between multiple services, so you do not need to deploy anything to the network and write an Aqua app just for basic testing. Let's look at an example: test.rs producer.rs consumer.rs test_on_mod.rs 1 fn main() {} 2 3 #[cfg(test)] 4 mod tests { 5 use marine_rs_sdk_test::marine_test; 6 #[marine_test( // 1 7 producer( 8 config_path = "../producer/Config.toml", 9 modules_dir = "../producer/artifacts" 10 ), 11 consumer( 12 config_path = "../consumer/Config.toml", 13 modules_dir = "../consumer/artifacts" 14 ) 15 )] 16 fn test() { 17 let mut producer = marine_test_env::producer::ServiceInterface::new(); // 2 18 let mut consumer = marine_test_env::consumer::ServiceInterface::new(); 19 let input = marine_test_env::producer::Input { // 3 20 first_name: String::from("John"), 21 last_name: String::from("Doe"), 22 }; 23 let data = producer.produce(input); // 4 24 let result = consumer.consume(data); 25 assert_eq!(result, "John Doe") 26 } 27 } 28 Copied! 1 use marine_rs_sdk::marine; 2 use marine_rs_sdk::module_manifest; 3 4 module_manifest!(); 5 6 pub fn main() {} 7 8 #[marine] 9 pub struct Data { 10 pub name: String, 11 } 12 13 #[marine] 14 pub struct Input { 15 pub first_name: String, 16 pub last_name: String, 17 } 18 19 #[marine] 20 pub fn produce(data: Input) -> Data { 21 Data { 22 name: format!("{} {}", data.first_name, data.last_name), 23 } 24 } 25 Copied! 1 use marine_rs_sdk::marine; 2 use marine_rs_sdk::module_manifest; 3 4 module_manifest!(); 5 6 pub fn main() {} 7 8 #[marine] 9 pub struct Data { 10 pub name: String, 11 } 12 13 #[marine] 14 pub fn consume(data: Data) -> String { 15 data.name 16 } 17 Copied! 1 fn main() {} 2 3 #[cfg(test)] 4 #[marine_rs_sdk_test::marine_test( 5 producer( 6 config_path = "../producer/Config.toml", 7 modules_dir = "../producer/artifacts" 8 ), 9 consumer( 10 config_path = "../consumer/Config.toml", 11 modules_dir = "../consumer/artifacts" 12 ) 13 )] 14 mod tests_on_mod { 15 #[test] 16 fn test() { 17 let mut producer = marine_test_env::producer::ServiceInterface::new(); 18 let mut consumer = marine_test_env::consumer::ServiceInterface::new(); 19 let input = marine_test_env::producer::Input { 20 first_name: String::from("John"), 21 last_name: String::from("Doe"), 22 }; 23 let data = producer.produce(input); 24 let result = consumer.consume(data); 25 assert_eq!(result, "John Doe") 26 } 27 } 28 Copied! 1. We wrap the test function with the marine_test macro by providing named service configurations with module locations. Based on its arguments the macro defines a marine_test_env module with an interface to the services. 2. We create new services. Each ServiceInterface::new() runs a new marine runtime with the service. 3. We prepare data to pass to a service using structure definition from marine_test_env. The macro finds all structures used in the service interface functions and defines them in the corresponding submodule of marine_test_env . 4. We call a service function through the ServiceInterface object. 5. It is possible to use the result of one service call as an argument for a different service call. The interface types with the same structure have the same rust type in marine_test_env. In the test_on_mod.rs tab we can see another option — applying marine_test to a mod. The macro just defines the marine_test_env at the beginning of the module and then it can be used as usual everywhere inside the module. The full example is here. The marine_test macro also gives access to the interface of internal modules which may be useful for setting up a test environment. This feature is designed to be used in situations when it is simpler to set up a service for a test through internal functions than through the service interface. To illustrate this feature we have rewritten the previous example: 1 fn main() {} 2 3 #[cfg(test)] 4 mod tests { 5 use marine_rs_sdk_test::marine_test; 6 #[marine_test( 7 producer( 8 config_path = "../producer/Config.toml", 9 modules_dir = "../producer/artifacts" 10 ), 11 consumer( 12 config_path = "../consumer/Config.toml", 13 modules_dir = "../consumer/artifacts" 14 ) 15 )] 16 fn test() { 17 let mut producer = marine_test_env::producer::ServiceInterface::new(); 18 let mut consumer = marine_test_env::consumer::ServiceInterface::new(); 19 let input = marine_test_env::producer::modules::producer::Input { // 1 20 first_name: String::from("John"), 21 last_name: String::from("Doe"), 22 }; 23 let data = producer.modules.producer.produce(input); // 2 24 let consumer_data = marine_test_env::consumer::modules::consumer::Data { name: data.name } // 3; 25 let result = consumer.modules.consumer.consume(consumer_data); 26 assert_eq!(result, "John Doe") 27 } 28 } 29 Copied! 1. We access the internal service interface to construct an interface structure. To do so, we use the following pattern: marine_test_env::$service_name::modules::$module_name::$structure_name. 2. We access the internal service interface and directly call a function from one of the modules of this service. To do so, we use the following pattern: $service_object.modules.$module_name.$function_name . 3. In the previous example, the same interface types had the same rust types. It is limited when using internal modules: the property is true only when structures are defined in internal modules of one service, or when structures are defined in service interfaces of different services. So, we need to construct the proper type to pass data to the internals of another module. Testing sdk also has the interface for Cargo build scripts. Some IDEs can analyze files generated in build scripts, providing code completion and error highlighting for code generated in build scripts. But using it may be a little bit tricky because build scripts are not designed for such things. Actions required to set up IDE: CLion: in the Help -> Actions -> Experimental Futures enable org.rust.cargo.evaluate.build.scripts refresh cargo project in order to update generated code: change Cargo.toml and build from IDE or press Refresh Cargo Project in Cargo tab. VS Code: install rust-analyzer plugin change Cargo.toml to let plugin update code from generated files The update will not work instantly: you should build service to wasm, and then trigger build.rs run again, but for the native target. And here is the example of using this: build.rs src/main.rs Cargo.toml 1 use marine_rs_sdk_test::generate_marine_test_env; 2 use marine_rs_sdk_test::ServiceDescription; 3 fn main() { 4 let services = vec![ // <- 1 5 ("greeting".to_string(), ServiceDescription { 6 config_path: "Config.toml".to_string(), 7 modules_dir: Some("artifacts".to_string()), 8 }) 9 ]; 10 11 let target = std::env::var("CARGO_CFG_TARGET_ARCH").unwrap(); 12 if target != "wasm32" { // <- 2 13 generate_marine_test_env(services, "marine_test_env.rs", file!()); // <- 3 14 } 15 16 println!("cargo:rerun-if-changed=src/main.rs"); // <- 4 17 } Copied! 1 use marine_rs_sdk::marine; 2 use marine_rs_sdk::module_manifest; 3 4 module_manifest!(); 5 6 pub fn main() {} 7 8 #[marine] 9 pub fn greeting(name: String) -> String { 10 format!("Hi, {}", name) 11 } 12 13 #[cfg(test)] 14 mod built_tests { 15 marine_rs_sdk_test::include_test_env!("/marine_test_env.rs"); // <- 4 16 #[test] 17 fn non_empty_string() { 18 let mut greeting = marine_test_env::greeting::ServiceInterface::new(); 19 let actual = greeting.greeting("name".to_string()); 20 assert_eq!(actual, "Hi, name"); 21 } 22 } Copied! 1 [package] 2 name = "wasm-build-rs" 3 version = "0.1.0" 4 authors = ["Fluence Labs"] 5 description = "The greeting module for the Fluence network" 6 repository = "https://github.com/fluencelabs/marine/tree/master/examples/build_rs" 7 edition = "2018" 8 publish = false 9 10 [[bin]] 11 name = "build_rs_test" 12 path = "src/main.rs" 13 14 [dependencies] 15 marine-rs-sdk = "0.6.11" 16 17 [dev-dependencies] 18 marine-rs-sdk-test = "0.4.0" 19 20 [build-dependencies] 21 marine-rs-sdk-test = "0.4.0" # <- 5 22 Copied! 1. We create a vector of pairs (service_name, service_description) to pass to the generator. The structure is the same with multi-service marine_test. 2. We check if we build for a non-wasm target. As we build this marine service only for wasm32-wasi and tests are built for native target, we can generate marine_test_env only for tests. This is needed because our generator depends on the artifacts from wasm32-wasi build. We suggest using a separate crate for using build scripts for testing purposes. It is here for simplicity. 3. We pass our services, a name of the file to generate, and a path to the build script file to the marine_test_env generator. Just always use file!() for the last argument. The generated file will be in the directory specified by the OUT_DIR variable, which is set by cargo. The build script must not change any files outside of this directory. 4. We set up condition to re-run the build script. It must be customized, a good choice is to re-run the build script when .wasm files or Config.toml are changed. 5. We import the generated file with the marine_test_env definition to the project. 6. Do not forget to add marine-rs-sdk-test to the build-dependencies section of Cargo.toml. Features The SDK has two useful features: logger and debug. Logger Using logging is a simple way to assist in debugging without deploying the module(s) to a peer-to-peer network node. The logger feature allows you to use a special logger that is based at the top of the log crate. To enable logging please specify the logger feature of the Fluence SDK in Config.toml and add the log crate: 1 [dependencies] 2 log = "0.4.14" 3 fluence = { version = "0.6.9", features = ["logger"] } Copied! The logger should be initialized before its usage. This can be done in the main function as shown in the example below. 1 use fluence::marine; 2 use fluence::WasmLogger; 3 4 pub fn main() { 5 WasmLogger::new() 6 // with_log_level can be skipped, 7 // logger will be initialized with Info level in this case. 8 .with_log_level(log::Level::Info) 9 .build() 10 .unwrap(); 11 } 12 13 #[marine] 14 pub fn put(name: String, file_content: Vec<u8>) -> String { 15 log::info!("put called with file name {}", file_name); 16 unimplemented!() 17 } Copied! In addition to the standard log creation features, the Fluence logger allows the so-called target map to be configured during the initialization step. This allows you to filter out logs by logging_mask, which can be set for each module in the service configuration. Let's consider an example: 1 const TARGET_MAP: [(&str, i64); 4] = [ 2 ("instruction", 1 << 1), 3 ("data_cache", 1 << 2), 4 ("next_peer_pks", 1 << 3), 5 ("subtree_complete", 1 << 4), 6 ]; 7 8 pub fn main() { 9 use std::collections::HashMap; 10 use std::iter::FromIterator; 11 12 let target_map = HashMap::from_iter(TARGET_MAP.iter().cloned()); 13 14 fluence::WasmLogger::new() 15 .with_target_map(target_map) 16 .build() 17 .unwrap(); 18 } 19 20 #[marine] 21 pub fn foo() { 22 log::info!(target: "instruction", "this will print if (logging_mask & 1) != 0"); 23 log::info!(target: "data_cache", "this will print if (logging_mask & 2) != 0"); 24 } Copied! Here, an array called TARGET_MAP is defined and provided to a logger in the main function of a module. Each entry of this array contains a string (a target) and a number that represents the bit position in the 64-bit mask logging_mask. When you write a log message request log::info!, its target must coincide with one of the strings (the targets) defined in the TARGET_MAP array. The log will be printed if logging_mask for the module has the corresponding target bit set. REPL also uses the log crate to print logs from Wasm modules. Log messages will be printed ifRUST_LOG environment variable is specified. Debug The application of the second feature is limited to obtaining some of the internal details of the IT execution. Normally, this feature should not be used by a backend developer. Here you can see example of such details for the greeting service compiled with the debug feature: 1 # running the greeting service compiled with debug feature 2 ~ $ RUST_LOG="info" fce-repl Config.toml 3 Welcome to the Fluence FaaS REPL 4 app service's created with service id = e5cfa463-ff50-4996-98d8-4eced5ac5bb9 5 elapsed time 40.694769ms 6 7 1> call greeting greeting "user" 8 [greeting] sdk.allocate: 4 9 [greeting] sdk.set_result_ptr: 1114240 10 [greeting] sdk.set_result_size: 8 11 [greeting] sdk.get_result_ptr, returns 1114240 12 [greeting] sdk.get_result_size, returns 8 13 [greeting] sdk.get_result_ptr, returns 1114240 14 [greeting] sdk.get_result_size, returns 8 15 [greeting] sdk.deallocate: 0x110080 8 16 17 result: String("Hi, user") 18 elapsed time: 222.675µs Copied! The most important information these logs relates to the allocate/deallocate function calls. The sdk.allocate: 4 line corresponds to passing the 4-byte user string to the Wasm module, with the memory allocated inside the module and the string is copied there. Whereas sdk.deallocate: 0x110080 8 refers to passing the 8-byte resulting string Hi, user to the host side. Since all arguments and results are passed by value, deallocate is called to delete unnecessary memory inside the Wasm module. Module Manifest The module_manifest! macro embeds the Interface Type (IT), SDK and Rust project version as well as additional project and build information into Wasm module. For the macro to be usable, it needs to be imported and initialized in the main.rs file: 1 // main.rs 2 use fluence::marine; 3 use fluence::module_manifest; // import manifest macro 4 5 module_manifest!(); // initialize macro 6 7 fn main() {} 8 9 #[marine] 10 fn some_function() {} 11 } Copied! Using the Marine CLI, we can inspect a module's manifest with marine info: 1 mbp16~/localdev/struct-exp(main|) % marine info -i artifacts/*.wasm 2 it version: 0.20.1 3 sdk version: 0.6.0 4 authors: The Fluence Team 5 version: 0.1.0 6 description: foo-wasm, a Marine wasi module 7 repository: 8 build time: 2021-06-11 21:08:59.855352 +00:00 UTC Copied! Last modified 8d ago Copy link Contents API Features
__label__pos
0.975576
Frobenius numbers From Rosetta Code (Redirected from Frobenius primes) Frobenius numbers is a draft programming task. It is not yet considered ready to be promoted as a complete task, for reasons that should be found in its talk page. Task Find and display here on this page the Frobenius numbers that are   <   10,000. The series is defined by: FrobeniusNumber(n) = prime(n) * prime(n+1) - prime(n) - prime(n+1) where:   prime(1)   =   2   prime(2)   =   3   prime(3)   =   5   prime(4)   =   7   •   •   • 11l[edit] Translation of: Python F isPrime(v) I v <= 1 R 0B I v < 4 R 1B I v % 2 == 0 R 0B I v < 9 R 1B I v % 3 == 0 R 0B E V r = round(pow(v, 0.5)) V f = 5 L f <= r I v % f == 0 | v % (f + 2) == 0 R 0B f += 6 R 1B V pn = 2 V n = 0 L(i) (3..).step(2) I isPrime(i) n++ V f = (pn * i) - pn - i I f > 10000 L.break print(n‘ => ’f) pn = i Output: 1 => 1 2 => 7 3 => 23 4 => 59 5 => 119 6 => 191 7 => 287 8 => 395 9 => 615 10 => 839 11 => 1079 12 => 1439 13 => 1679 14 => 1931 15 => 2391 16 => 3015 17 => 3479 18 => 3959 19 => 4619 20 => 5039 21 => 5615 22 => 6395 23 => 7215 24 => 8447 25 => 9599 Action![edit] INCLUDE "H6:SIEVE.ACT" INT FUNC NextPrime(INT p BYTE ARRAY primes) DO p==+1 UNTIL primes(p) OD RETURN (p) PROC Main() DEFINE MAXNUM="200" BYTE ARRAY primes(MAXNUM+1) INT p1,p2,f Put(125) PutE() ;clear the screen Sieve(primes,MAXNUM+1) p2=2 DO p1=p2 p2=NextPrime(p2,primes) f=p1*p2-p1-p2 IF f<10000 THEN PrintI(f) Put(32) ELSE EXIT FI OD RETURN Output: Screenshot from Atari 8-bit computer 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 ALGOL 68[edit] BEGIN # find some Frobenius Numbers: # # Frobenius(n) = ( prime(n) * prime(n+1) ) - prime(n) - prime(n+1) # # reurns a list of primes up to n # PROC prime list = ( INT n )[]INT: BEGIN # sieve the primes to n # INT no = 0, yes = 1; [ 1 : n ]INT p; p[ 1 ] := no; p[ 2 ] := yes; FOR i FROM 3 BY 2 TO n DO p[ i ] := yes OD; FOR i FROM 4 BY 2 TO n DO p[ i ] := no OD; FOR i FROM 3 BY 2 TO ENTIER sqrt( n ) DO IF p[ i ] = yes THEN FOR s FROM i * i BY i + i TO n DO p[ s ] := no OD FI OD; # replace the sieve with a list # INT p pos := 0; FOR i TO n DO IF p[ i ] = yes THEN p[ p pos +:= 1 ] := i FI OD; p[ 1 : p pos ] END # prime list # ; # show Frobenius numbers up to 10 000 # INT max number = 10 000; []INT prime = prime list( max number ); FOR i TO max number - 1 WHILE INT frobenius number = ( ( prime[ i ] * prime[ i + 1 ] ) - prime[ i ] ) - prime[ i + 1 ]; frobenius number < max number DO print( ( " ", whole( frobenius number, 0 ) ) ) OD END Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 APL[edit] Works with: Dyalog APL (¯1(⊢×1)-⊢+1)(((/⍨)(~⊢∊∘.×))1↓⍳)(1+*.5) 10000 Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 AppleScript[edit] on isPrime(n) if (n < 4) then return (n > 1) if ((n mod 2 is 0) or (n mod 3 is 0)) then return false repeat with i from 5 to (n ^ 0.5) div 1 by 6 if ((n mod i is 0) or (n mod (i + 2) is 0)) then return false end repeat return true end isPrime on Frobenii(max) script o property frobs : {} end script set p to 2 set n to 3 repeat if (isPrime(n)) then set frob to p * n - p - n if (frob > max) then exit repeat set end of o's frobs to frob set p to n end if set n to n + 2 end repeat return o's frobs end Frobenii Frobenii(9999) Output: {1, 7, 23, 59, 119, 191, 287, 395, 615, 839, 1079, 1439, 1679, 1931, 2391, 3015, 3479, 3959, 4619, 5039, 5615, 6395, 7215, 8447, 9599} Arturo[edit] primes: select 0..10000 => prime? frobenius: function [n] -> sub sub primes\[n] * primes\[n+1] primes\[n] primes\[n+1] frob: 0 lst: new [] j: new 0 while [frob < 10000] [ 'lst ++ frob: <= frobenius j inc 'j ] loop split.every:10 chop lst 'a -> print map a => [pad to :string & 5] Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 AutoHotkey[edit] n := 0, i := 1, pn := 2 loop { if isprime(i+=2) { if ((f := pn*i - pn - i) > 10000) break result .= SubStr(" " f, -3) . (Mod(++n, 5) ? "`t" : "`n") pn := i } } MsgBox % result return isPrime(n, p=1) { if (n < 2) return false if !Mod(n, 2) return (n = 2) if !Mod(n, 3) return (n = 3) while ((p+=4) <= Sqrt(n)) if !Mod(n, p) return false else if !Mod(n, p+=2) return false return true } Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 AWK[edit] # syntax: GAWK -f FROBENIUS_NUMBERS.AWK # converted from FreeBASIC BEGIN { start = 3 stop = 9999 pn = 2 for (i=start; i<=stop; i++) { if (is_prime(i)) { f = pn * i - pn - i if (f > stop) { break } printf("%4d%1s",f,++count%10?"":"\n") pn = i } } printf("\nFrobenius numbers %d-%d: %d\n",start,stop,count) exit(0) } function is_prime(x, i) { if (x <= 1) { return(0) } for (i=2; i<=int(sqrt(x)); i++) { if (x % i == 0) { return(0) } } return(1) } Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 Frobenius numbers 3-9999: 25 BASIC[edit] 10 DEFINT A-Z 20 LM = 10000 30 M = SQR(LM)+1 40 DIM P(M) 50 FOR I=2 TO SQR(M) 60 IF P(I)=0 THEN FOR J=I+I TO M STEP I: P(J)=1: NEXT J 70 NEXT I 80 FOR I=2 TO M 90 IF P(I)=0 THEN P(C)=I: C=C+1 100 NEXT I 110 FOR N=0 TO C-2 120 PRINT P(N)*P(N+1)-P(N)-P(N+1), 130 NEXT N Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 BASIC256[edit] n = 0 lim = 10000 k = sqr(lim) + 1 dim P(k) for i = 2 to sqr(k) if P[i] = 0 then for j = i + i to k step i P[j] = 1 next j end if next i for i = 2 to k-1 if P[i] = 0 then P[n] = i: n += 1 next i for i = 0 to n - 2 print i+1; " => "; P[i] * P[i + 1] - P[i] - P[i + 1] next i end BCPL[edit] get "libhdr" manifest $( limit = 10000 $) // Integer square root let sqrt(s) = s <= 1 -> 1, valof $( let x0 = s >> 1 let x1 = (x0 + s/x0) >> 1 while x1 < x0 $( x0 := x1 x1 := (x0 + s/x0) >> 1 $) resultis x0 $) // Find primes up to n, store at v. // Returns amount of primes found let sieve(v, n) = valof $( let count = 0 // Sieve the primes for i=2 to n do v!i := true for i=2 to sqrt(n) if v!i then $( let j = i+i while j <= n $( v!j := false j := j + i $) $) // Filter the primes for i=2 to n if v!i then $( v!count := i count := count + 1 $) resultis count $) // Frobenius number given prime array let frob(p, n) = p!n * p!(n+1) - p!n - p!(n+1) let start() be $( // frob(n) is always less than p(n+1)^2 // sieving up to the square root of the limit is enough, // though whe need one extra since p(n+1) is necessary let primes = getvec(sqrt(limit)+1) let nprimes = sieve(primes, sqrt(limit)+1) // similarly, having found that many primes, we generate // one fewer Frobenius number for n = 0 to nprimes-2 do writef("%N*N", frob(primes, n)) freevec(primes) $) Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 C[edit] #include <stdio.h> #include <stdlib.h> #include <math.h> #define LIMIT 10000 /* Generate primes up to N */ unsigned int sieve(unsigned int n, unsigned int **list) { unsigned char *sieve = calloc(n+1, 1); unsigned int i, j, max = 0; for (i = 2; i*i <= n; i++) if (!sieve[i]) for (j = i+i; j <= n; j += i) sieve[j] = 1; for (i = 2; i <= n; i++) max += !sieve[i]; *list = malloc(max * sizeof(unsigned int)); for (i = 0, j = 2; j <= n; j++) if (!sieve[j]) (*list)[i++] = j; free(sieve); return i; } /* Frobenius number */ unsigned int frob(unsigned const int *primes, unsigned int n) { return primes[n] * primes[n+1] - primes[n] - primes[n+1]; } int main() { /* Same trick as in BCPL example. frob(n) < primes(n+1)^2, so we need primes up to sqrt(limit)+1. */ unsigned int *primes; unsigned int amount = sieve(sqrt(LIMIT)+1, &primes); unsigned int i; for (i=0; i<amount-1; i++) printf("%d\n", frob(primes, i)); free(primes); return 0; } Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 C#[edit] Asterisks mark the non-primes among the numbers. using System.Collections.Generic; using System.Linq; using static System.Console; using static System.Math; class Program { static bool ispr(int x) { int lim = (int)Sqrt((double)x); if (x < 2) return false; if ((x % 3) == 0) return x == 0; bool odd = false; for (int d = 5; d <= lim; d += (odd = !odd) ? 2 : 4) { if (x % d == 0) return false; } return true; } static void Main() { int c = 0, d = 0, f, lim = 1000000, l2 = lim / 100; var Frob = PG.Primes((int)Sqrt(lim) + 1).ToArray(); for (int n = 0, m = 1; m < Frob.Length; n = m++) { if ((f = Frob[n] * Frob[m] - Frob[n] - Frob[m]) < l2) d++; Write("{0,7:n0}{2} {1}", f , ++c % 10 == 0 ? "\n" : "", ispr(f) ? " " : "*"); } Write("\n\nCalculated {0} Frobenius numbers of consecutive primes under {1:n0}, " + "of which {2} were under {3:n0}", c, lim, d, l2); } } class PG { public static IEnumerable<int> Primes(int lim) { var flags = new bool[lim + 1]; int j = 3; yield return 2; for (int d = 8, sq = 9; sq <= lim; j += 2, sq += d += 8) if (!flags[j]) { yield return j; for (int k = sq, i = j << 1; k <= lim; k += i) flags[k] = true; } for (; j <= lim; j += 2) if (!flags[j]) yield return j; } } Output: 1* 7 23 59 119* 191 287* 395* 615* 839 1,079* 1,439 1,679* 1,931 2,391* 3,015* 3,479* 3,959* 4,619* 5,039 5,615* 6,395* 7,215* 8,447 9,599* 10,199* 10,811* 11,447 12,095* 14,111* 16,379* 17,679* 18,767* 20,423* 22,199* 23,399 25,271* 26,891 28,551* 30,615* 32,039* 34,199* 36,479 37,631* 38,807* 41,579 46,619 50,171* 51,527* 52,895* 55,215* 57,119 59,999 63,999* 67,071* 70,215* 72,359* 74,519* 77,279 78,959* 82,343* 89,351* 94,859* 96,719* 98,591* 104,279* 110,879 116,255* 120,407* 122,495* 126,015* 131,027* 136,151* 140,615* 144,395* 148,215* 153,647* 158,399* 163,199 170,543* 175,559* 180,599* 185,759* 189,215* 193,595* 198,015* 204,287* 209,759* 212,519* 215,291* 222,747* 232,307 238,139* 244,019* 249,995* 255,015* 264,159* 271,439* 281,879* 294,839* 303,575* 312,471* 319,215* 323,759 328,319* 337,535* 346,911* 354,015* 358,799* 363,599* 370,871 376,991* 380,687* 389,339* 403,199* 410,879* 414,731 421,191* 429,015* 434,279* 443,519* 454,271* 461,031* 470,579 482,999* 495,599* 508,343* 521,267 531,431* 540,215* 547,595* 556,499* 566,999 574,559* 583,679* 592,895* 606,791 625,655* 643,167* 654,479* 664,199 674,039* 678,971 683,927* 693,863* 713,975* 729,311* 734,447* 739,595* 755,111* 770,879* 776,159 781,451* 802,715* 824,459 835,379 851,903* 868,607* 879,839 889,239* 900,591* 919,631 937,019* 946,719* 958,431* 972,179* 986,039* Calculated 167 Frobenius numbers of consecutive primes under 1,000,000, of which 25 were under 10,000 C++[edit] Library: Primesieve #include <cstdint> #include <iomanip> #include <iostream> #include <primesieve.hpp> bool is_prime(uint64_t n) { if (n < 2) return false; if (n % 2 == 0) return n == 2; if (n % 3 == 0) return n == 3; for (uint64_t p = 5; p * p <= n; p += 4) { if (n % p == 0) return false; p += 2; if (n % p == 0) return false; } return true; } int main() { const uint64_t limit = 1000000; std::cout << "Frobenius numbers less than " << limit << " (asterisk marks primes):\n"; primesieve::iterator it; uint64_t prime1 = it.next_prime(); for (int count = 1;; ++count) { uint64_t prime2 = it.next_prime(); uint64_t frobenius = prime1 * prime2 - prime1 - prime2; if (frobenius >= limit) break; std::cout << std::setw(6) << frobenius << (is_prime(frobenius) ? '*' : ' ') << (count % 10 == 0 ? '\n' : ' '); prime1 = prime2; } std::cout << '\n'; } Output: Frobenius numbers less than 1000000 (asterisk marks primes): 1 7* 23* 59* 119 191* 287 395 615 839* 1079 1439* 1679 1931* 2391 3015 3479 3959 4619 5039* 5615 6395 7215 8447* 9599 10199 10811 11447* 12095 14111 16379 17679 18767 20423 22199 23399* 25271 26891* 28551 30615 32039 34199 36479* 37631 38807 41579* 46619* 50171 51527 52895 55215 57119* 59999* 63999 67071 70215 72359 74519 77279* 78959 82343 89351 94859 96719 98591 104279 110879* 116255 120407 122495 126015 131027 136151 140615 144395 148215 153647 158399 163199* 170543 175559 180599 185759 189215 193595 198015 204287 209759 212519 215291 222747 232307* 238139 244019 249995 255015 264159 271439 281879 294839 303575 312471 319215 323759* 328319 337535 346911 354015 358799 363599 370871* 376991 380687 389339 403199 410879 414731* 421191 429015 434279 443519 454271 461031 470579* 482999 495599 508343 521267* 531431 540215 547595 556499 566999* 574559 583679 592895 606791* 625655 643167 654479 664199* 674039 678971* 683927 693863 713975 729311 734447 739595 755111 770879 776159* 781451 802715 824459* 835379* 851903 868607 879839* 889239 900591 919631* 937019 946719 958431 972179 986039 Cowgol[edit] include "cowgol.coh"; const LIMIT := 10000; sub sqrt(n: intptr): (x0: intptr) is var x1: intptr; if n <= 1 then x0 := 1; else x0 := n >> 1; x1 := (x0 + n/x0) >> 1; while x1 < x0 loop x0 := x1; x1 := (x0 + n/x0) >> 1; end loop; end if; end sub; sub sieve(max: intptr, buf: [uint16]): (count: uint16) is var sbuf := buf as [uint8] + max; MemZero(sbuf, max); var i: intptr := 2; while i*i <= max loop if [sbuf+i] == 0 then var j := i+i; while j <= max loop [sbuf+j] := 1; j := j+i; end loop; end if; i := i+1; end loop; count := 0; i := 2; while i <= max loop if [sbuf+i] == 0 then [buf] := i as uint16; buf := @next buf; count := count + 1; end if; i := i + 1; end loop; end sub; var primes: uint16[LIMIT + 1]; var nprimes := sieve(sqrt(LIMIT)+1, &primes[0]); var n: uint16 := 0; while n < nprimes-1 loop print_i16(primes[n] * primes[n+1] - primes[n] - primes[n+1]); print_nl(); n := n + 1; end loop; Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 Delphi[edit] Works with: Delphi version 6.0 function IsPrime(N: integer): boolean; {Optimised prime test - about 40% faster than the naive approach} var I,Stop: integer; begin if (N = 2) or (N=3) then Result:=true else if (n <= 1) or ((n mod 2) = 0) or ((n mod 3) = 0) then Result:= false else begin I:=5; Stop:=Trunc(sqrt(N)); Result:=False; while I<=Stop do begin if ((N mod I) = 0) or ((N mod (i + 2)) = 0) then exit; Inc(I,6); end; Result:=True; end; end; function GetNextPrime(Start: integer): integer; {Get the next prime number after Start} begin repeat Inc(Start) until IsPrime(Start); Result:=Start; end; procedure ShowFrobeniusNumbers(Memo: TMemo); var N,N1,FN,Cnt: integer; begin N:=2; Cnt:=0; while true do begin Inc(Cnt); N1:=GetNextPrime(N); FN:=N * N1 - N - N1; N:=N1; if FN>10000 then break; Memo.Lines.Add(Format('%2d = %5d',[Cnt,FN])); end; end; Output: 1 = 1 2 = 7 3 = 23 4 = 59 5 = 119 6 = 191 7 = 287 8 = 395 9 = 615 10 = 839 11 = 1079 12 = 1439 13 = 1679 14 = 1931 15 = 2391 16 = 3015 17 = 3479 18 = 3959 19 = 4619 20 = 5039 21 = 5615 22 = 6395 23 = 7215 24 = 8447 25 = 9599 Factor[edit] Works with: Factor version 0.99 2021-02-05 USING: io kernel math math.primes prettyprint ; "Frobenius numbers < 10,000:" print 2 3 [ [ nip dup next-prime ] [ * ] [ [ - ] dip - ] 2tri dup 10,000 < ] [ . ] while 3drop Output: Frobenius numbers < 10,000: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 Fermat[edit] Function Frobenius(n)=Prime(n)*Prime(n+1)-Prime(n)-Prime(n+1). for n = 1 to 25 do !!Frobenius(n) od Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 FreeBASIC[edit] #include "isprime.bas" dim as integer pn=2, n=0, f for i as integer = 3 to 9999 step 2 if isprime(i) then n += 1 f = pn*i - pn - i if f > 10000 then end print n, f pn = i end if next i Output: 1 1 2 7 3 23 4 59 5 119 6 191 7 287 8 395 9 615 10 839 11 1079 12 1439 13 1679 14 1931 15 2391 16 3015 17 3479 18 3959 19 4619 20 5039 21 5615 22 6395 23 7215 24 8447 25 9599 FutureBasic[edit] include "NSLog.incl" local fn IsPrime( n as long ) as BOOL long i BOOL result = YES if ( n < 2 ) then result = NO : exit fn for i = 2 to n + 1 if ( i * i <= n ) and ( n mod i == 0 ) result = NO : exit fn end if next end fn = result void local fn ListFrobenius( upperLimit as long ) long i, pn = 2, n = 0, f, r = 0 NSLog( @"Frobenius numbers through %ld:", upperLimit ) for i = 3 to upperLimit - 1 step 2 if ( fn IsPrime(i) ) n++ f = pn * i - pn - i if ( f > upperLimit ) then break NSLog( @"%7ld\b", f ) r++ if r mod 5 == 0 then NSLog( @"" ) pn = i end if next end fn fn ListFrobenius( 100000 ) HandleEvents Output: Frobenius numbers through 100000: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 10199 10811 11447 12095 14111 16379 17679 18767 20423 22199 23399 25271 26891 28551 30615 32039 34199 36479 37631 38807 41579 46619 50171 51527 52895 55215 57119 59999 63999 67071 70215 72359 74519 77279 78959 82343 89351 94859 96719 98591 Go[edit] Translation of: Wren Library: Go-rcu package main import ( "fmt" "rcu" ) func main() { primes := rcu.Primes(101) var frobenius []int for i := 0; i < len(primes)-1; i++ { frob := primes[i]*primes[i+1] - primes[i] - primes[i+1] if frob >= 10000 { break } frobenius = append(frobenius, frob) } fmt.Println("Frobenius numbers under 10,000:") for i, n := range frobenius { fmt.Printf("%5s ", rcu.Commatize(n)) if (i+1)%9 == 0 { fmt.Println() } } fmt.Printf("\n\n%d such numbers found.\n", len(frobenius)) } Output: Frobenius numbers under 10,000: 1 7 23 59 119 191 287 395 615 839 1,079 1,439 1,679 1,931 2,391 3,015 3,479 3,959 4,619 5,039 5,615 6,395 7,215 8,447 9,599 25 such numbers found. Haskell[edit] primes = 2 : sieve [3,5..] where sieve (x:xs) = x : sieve (filter (\y -> y `mod` x /= 0) xs) frobenius = zipWith (\a b -> a*b - a - b) primes (tail primes) λ> takeWhile (< 10000) frobenius [1,7,23,59,119,191,287,395,615,839,1079,1439,1679,1931,2391,3015,3479,3959,4619,5039,5615,6395,7215,8447,9599] J[edit] frob =: (*&p: - +&p:) >: echo frob i. 25 (Note that frob counts prime numbers starting from 0 (which gives 2), so for some contexts the function to calculate frobenius numbers would be frob@<:.) Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 Java[edit] Uses the PrimeGenerator class from Extensible prime generator#Java. public class Frobenius { public static void main(String[] args) { final int limit = 1000000; System.out.printf("Frobenius numbers less than %d (asterisk marks primes):\n", limit); PrimeGenerator primeGen = new PrimeGenerator(1000, 100000); int prime1 = primeGen.nextPrime(); for (int count = 1; ; ++count) { int prime2 = primeGen.nextPrime(); int frobenius = prime1 * prime2 - prime1 - prime2; if (frobenius >= limit) break; System.out.printf("%6d%c%c", frobenius, isPrime(frobenius) ? '*' : ' ', count % 10 == 0 ? '\n' : ' '); prime1 = prime2; } System.out.println(); } private static boolean isPrime(int n) { if (n < 2) return false; if (n % 2 == 0) return n == 2; if (n % 3 == 0) return n == 3; for (int p = 5; p * p <= n; p += 4) { if (n % p == 0) return false; p += 2; if (n % p == 0) return false; } return true; } } Output: Frobenius numbers less than 1000000 (asterisk marks primes): 1 7* 23* 59* 119 191* 287 395 615 839* 1079 1439* 1679 1931* 2391 3015 3479 3959 4619 5039* 5615 6395 7215 8447* 9599 10199 10811 11447* 12095 14111 16379 17679 18767 20423 22199 23399* 25271 26891* 28551 30615 32039 34199 36479* 37631 38807 41579* 46619* 50171 51527 52895 55215 57119* 59999* 63999 67071 70215 72359 74519 77279* 78959 82343 89351 94859 96719 98591 104279 110879* 116255 120407 122495 126015 131027 136151 140615 144395 148215 153647 158399 163199* 170543 175559 180599 185759 189215 193595 198015 204287 209759 212519 215291 222747 232307* 238139 244019 249995 255015 264159 271439 281879 294839 303575 312471 319215 323759* 328319 337535 346911 354015 358799 363599 370871* 376991 380687 389339 403199 410879 414731* 421191 429015 434279 443519 454271 461031 470579* 482999 495599 508343 521267* 531431 540215 547595 556499 566999* 574559 583679 592895 606791* 625655 643167 654479 664199* 674039 678971* 683927 693863 713975 729311 734447 739595 755111 770879 776159* 781451 802715 824459* 835379* 851903 868607 879839* 889239 900591 919631* 937019 946719 958431 972179 986039 jq[edit] Works with: jq Works with gojq, the Go implementation of jq The solution offered here is based on a function that can in principle generate an unbounded stream of Frobenius numbers without relying on the precomputation or storage of an array of primes except as may be used by `is_prime`. The following is also designed to take advantage of gojq's support for unbounded-precision integer arithmetic. See e.g. Erdős-primes#jq for a suitable implementation of `is_prime`. # Generate a stream of Frobenius numbers up to an including `.`; # specify `null` or `infinite` to generate an unbounded stream. def frobenius: . as $limit | label $out | foreach (range(3;infinite;2) | select(is_prime)) as $p ({prev: 2}; (.prev * $p - .prev - $p) as $frob | if ($limit != null and $frob > $limit then break $out else .frob = $frob end | .prev = $p; .frob); 9999 | frobenius Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 Julia[edit] using Primes const primeslt10k = primes(10000) frobenius(n) = begin (x, y) = primeslt10k[n:n+1]; x * y - x - y end function frobeniuslessthan(maxnum) frobpairs = Pair{Int, Bool}[] for n in 1:maxnum frob = frobenius(n) frob > maxnum && break push!(frobpairs, Pair(frob, isprime(frob))) end return frobpairs end function testfrobenius() println("Frobenius numbers less than 1,000,000 (an asterisk marks the prime ones).") frobpairs = frobeniuslessthan(1_000_000) for (i, p) in enumerate(frobpairs) print(rpad(string(p[1]) * (p[2] ? "*" : ""), 8), i % 10 == 0 ? "\n" : "") end end testfrobenius() Output: Frobenius numbers less than 1,000,000 (an asterisk marks the prime ones). 1 7* 23* 59* 119 191* 287 395 615 839* 1079 1439* 1679 1931* 2391 3015 3479 3959 4619 5039* 5615 6395 7215 8447* 9599 10199 10811 11447* 12095 14111 16379 17679 18767 20423 22199 23399* 25271 26891* 28551 30615 32039 34199 36479* 37631 38807 41579* 46619* 50171 51527 52895 55215 57119* 59999* 63999 67071 70215 72359 74519 77279* 78959 82343 89351 94859 96719 98591 104279 110879* 116255 120407 122495 126015 131027 136151 140615 144395 148215 153647 158399 163199* 170543 175559 180599 185759 189215 193595 198015 204287 209759 212519 215291 222747 232307* 238139 244019 249995 255015 264159 271439 281879 294839 303575 312471 319215 323759* 328319 337535 346911 354015 358799 363599 370871* 376991 380687 389339 403199 410879 414731* 421191 429015 434279 443519 454271 461031 470579* 482999 495599 508343 521267* 531431 540215 547595 556499 566999* 574559 583679 592895 606791* 625655 643167 654479 664199* 674039 678971* 683927 693863 713975 729311 734447 739595 755111 770879 776159* 781451 802715 824459* 835379* 851903 868607 879839* 889239 900591 919631* 937019 946719 958431 972179 986039 Mathematica/Wolfram Language[edit] ClearAll[fn] fn[n_] := Prime[n] Prime[n + 1] - Prime[n] - Prime[n + 1] a = -1; i = 1; res = {}; While[a < 10^4, a = fn[i]; i++; If[a < 10^4, AppendTo[res, a]] ] res Output: {1,7,23,59,119,191,287,395,615,839,1079,1439,1679,1931,2391,3015,3479,3959,4619,5039,5615,6395,7215,8447,9599} Nim[edit] As I like iterators, I used one for (odd) primes and one for Frobenius numbers. Of course, there are other ways to proceed. import sequtils, strutils func isOddPrime(n: Positive): bool = if n mod 3 == 0: return n == 3 var d = 5 while d * d <= n: if n mod d == 0: return false inc d, 2 if n mod d == 0: return false inc d, 4 result = true iterator oddPrimes(): int = yield 3 var n = 5 while true: if n.isOddPrime: yield n inc n, 2 if n.isOddPrime: yield n inc n, 4 iterator frobenius(lim: Positive): int = var p1 = 2 for p2 in oddPrimes(): let f = p1 * p2 - p1 - p2 if f < lim: yield f else: break p1 = p2 const N = 10_000 var result = toSeq(frobenius(10_000)) echo "Found $1 Frobenius numbers less than $2:".format(result.len, N) echo result.join(" ") Output: Found 25 Frobenius numbers less than 10000: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 Perl[edit] Library: ntheory use strict; use warnings; use feature 'say'; use ntheory <nth_prime primes>; use List::MoreUtils qw(slide); # build adding one term at a time my(@F,$n); do { ++$n and push @F, nth_prime($n) * nth_prime($n+1) - (nth_prime($n) + nth_prime($n+1)) } until $F[-1] >= 10000; say "$#F matching numbers:\n" . join(' ', @F[0 .. $#F-1]); # process a list with a 2-wide sliding window my $limit = 10_000; say "\n" . join ' ', grep { $_ < $limit } slide { $a * $b - $a - $b } @{primes($limit)}; Output: 25 matching numbers: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 Phix[edit] for n=4 to 6 by 2 do integer lim = power(10,n), i=1 sequence frob = {} while true do integer p = get_prime(i), q = get_prime(i+1), frobenius = p*q-p-q if frobenius > lim then exit end if frob &= frobenius i += 1 end while frob = apply(true,sprintf,{{"%d"},frob}) printf(1,"%3d Frobenius numbers under %,9d: %s\n", {length(frob),lim,join(shorten(frob,"",5),", ")}) end for Output: 25 Frobenius numbers under 10,000: 1, 7, 23, 59, 119, ..., 5615, 6395, 7215, 8447, 9599 167 Frobenius numbers under 1,000,000: 1, 7, 23, 59, 119, ..., 937019, 946719, 958431, 972179, 986039 Python[edit] #!/usr/bin/python def isPrime(v): if v <= 1: return False if v < 4: return True if v % 2 == 0: return False if v < 9: return True if v % 3 == 0: return False else: r = round(pow(v,0.5)) f = 5 while f <= r: if v % f == 0 or v % (f + 2) == 0: return False f += 6 return True pn = 2 n = 0 for i in range(3, 9999, 2): if isPrime(i): n += 1 f = (pn * i) - pn - i if f > 10000: break print (n, ' => ', f) pn = i PL/M[edit] 100H: BDOS: PROCEDURE (FN, ARG); DECLARE FN BYTE, ARG ADDRESS; GO TO 5; END BDOS; EXIT: PROCEDURE; CALL BDOS(0,0); END EXIT; PRINT: PROCEDURE (S); DECLARE S ADDRESS; CALL BDOS(9,S); END PRINT; DECLARE LIMIT LITERALLY '10$000'; PRINT$NUMBER: PROCEDURE (N); DECLARE S (8) BYTE INITIAL ('.....',13,10,'$'); DECLARE (N, P) ADDRESS, C BASED P BYTE; P = .S(5); DIGIT: P = P - 1; C = N MOD 10 + '0'; N = N / 10; IF N > 0 THEN GO TO DIGIT; CALL PRINT(P); END PRINT$NUMBER; SQRT: PROCEDURE (N) ADDRESS; DECLARE (N, X0, X1) ADDRESS; IF N <= 1 THEN RETURN 1; X0 = SHR(N,1); LOOP: X1 = SHR(X0 + N/X0, 1); IF X1 >= X0 THEN RETURN X0; X0 = X1; GO TO LOOP; END SQRT; SIEVE: PROCEDURE (LIM, BASE) ADDRESS; DECLARE (LIM, BASE) ADDRESS; DECLARE PRIMES BASED BASE ADDRESS; DECLARE COUNT ADDRESS; DECLARE SBASE ADDRESS, SIEVE BASED SBASE BYTE; DECLARE (I, J, SQLIM) ADDRESS; SBASE = BASE + LIM; SQLIM = SQRT(LIM); DO I=2 TO LIM; SIEVE(I) = 1; END; DO I=2 TO SQLIM; IF SIEVE(I) THEN DO; DO J=I+I TO LIM BY I; SIEVE(J) = 0; END; END; END; COUNT = 0; DO I=2 TO LIM; IF SIEVE(I) THEN DO; PRIMES(COUNT) = I; COUNT = COUNT+1; END; END; RETURN COUNT; END SIEVE; FROBENIUS: PROCEDURE (N, PBASE) ADDRESS; DECLARE (PBASE, N, P BASED PBASE) ADDRESS; RETURN P(N) * P(N+1) - P(N) - P(N+1); END FROBENIUS; DECLARE NPRIMES ADDRESS; DECLARE N ADDRESS; NPRIMES = SIEVE(SQRT(LIMIT)+1, .MEMORY); DO N=0 TO NPRIMES-2; CALL PRINT$NUMBER(FROBENIUS(N, .MEMORY)); END; CALL EXIT; EOF Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 PureBasic[edit] Procedure isPrime(v.i) If v < = 1 : ProcedureReturn #False ElseIf v < 4 : ProcedureReturn #True ElseIf v % 2 = 0 : ProcedureReturn #False ElseIf v < 9 : ProcedureReturn #True ElseIf v % 3 = 0 : ProcedureReturn #False Else Protected r = Round(Sqr(v), #PB_Round_Down) Protected f = 5 While f <= r If v % f = 0 Or v % (f + 2) = 0 ProcedureReturn #False EndIf f + 6 Wend EndIf ProcedureReturn #True EndProcedure OpenConsole() pn.i = 2 n.i = 0 For i.i = 3 To 9999 Step 2 If isPrime(i) n + 1 f.i = pn * i - pn - i If f > 10000 Break EndIf Print(Str(n) + " => " + Str(f) + #CRLF$) pn = i EndIf Next i Input() CloseConsole() End Output: 1 => 1 2 => 7 3 => 23 4 => 59 5 => 119 6 => 191 7 => 287 8 => 395 9 => 615 10 => 839 11 => 1079 12 => 1439 13 => 1679 14 => 1931 15 => 2391 16 => 3015 17 => 3479 18 => 3959 19 => 4619 20 => 5039 21 => 5615 22 => 6395 23 => 7215 24 => 8447 25 => 9599 Quackery[edit] eratosthenes and isprime are defined at Sieve of Eratosthenes#Quackery. In this solution the primes and Frobenius numbers are zero indexed rather than one indexed as per the task. It simplifies the code a smidgeon, as Quackery nests are zero indexed. 200 eratosthenes [ [ [] 200 times [ i^ isprime if [ i^ join ] ] ] constant swap peek ] is prime ( n --> n ) [ dup prime swap 1+ prime 2dup * rot - swap - ] is frobenius ( n --> n ) [] 0 [ tuck frobenius dup 10000 < while join swap 1+ again ] drop nip echo Output: [ 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 ] Raku[edit] say "{+$_} matching numbers\n{.batch(10)».fmt('%4d').join: "\n"}\n" given (^1000).grep( *.is-prime ).rotor(2 => -1) .map( { (.[0] * .[1] - .[0] - .[1]) } ).grep(* < 10000); Output: 25 matching numbers 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 REXX[edit] /*REXX program finds Frobenius numbers where the numbers are less than some number N. */ parse arg hi cols . /*obtain optional argument from the CL.*/ if hi=='' | hi=="," then hi= 10000 /* " " " " " " */ if cols=='' | cols=="," then cols= 10 /* " " " " " " */ w= 10 /*the width of any column in the output*/ call genP /*build array of semaphores for primes.*/ title= ' Frobenius numbers that are < ' commas(hi) if cols>0 then say ' index │'center(title, 1 + cols*(w+1) ) if cols>0 then say '───────┼'center("" , 1 + cols*(w+1), '─') $=; idx= 1 /*list of Frobenius #s (so far); index.*/ do j=1; jp= j+1; y= @.j*@.jp - @.j - @.jp /*calculate a Frobenius number. */ if y>= hi then leave /*Is Y too high? Yes, then leave. */ if cols<=0 then iterate /*Build the list (to be shown later)? */ c= commas(y) /*maybe add commas to the number. */ $= $ right(c, max(w, length(c) ) ) /*add a Frobenius #──►list, allow big #*/ if j//cols\==0 then iterate /*have we populated a line of output? */ say center(idx, 7)'│' substr($, 2); $= /*display what we have so far (cols). */ idx= idx + cols /*bump the index count for the output*/ end /*j*/ if $\=='' then say center(idx, 7)"│" substr($, 2) /*possible display residual output.*/ if cols>0 then say '───────┴'center("" , 1 + cols*(w+1), '─') say say 'Found ' commas(j-1) title exit 0 /*stick a fork in it, we're all done. */ /*──────────────────────────────────────────────────────────────────────────────────────*/ commas: parse arg ?; do jc=length(?)-3 to 1 by -3; ?=insert(',', ?, jc); end; return ? /*──────────────────────────────────────────────────────────────────────────────────────*/ genP: @.1=2; @.2=3; @.3=5; @.4=7; @.5=11; @.6= 13 /*define some low primes. */ #=6; sq.#= @.# **2 /*number of primes so far; prime²*/ /* [↓] generate more primes ≤ high.*/ do j=@.#+2 by 2 for max(0,hi%2-@.#%2+1); parse var j '' -1 _ /*find odd primes*/ if _==5 then iterate; if j// 3==0 then iterate /*J ÷ by 5? J ÷ by 3? */ if j//7==0 then iterate; if j//11==0 then iterate /*" " " 7? " " " 11? */ do k=6 while sq.k<=j /* [↓] divide by the known odd primes.*/ if j//@.k==0 then iterate j /*Is J ÷ X? Then not prime. ___ */ end /*k*/ /* [↑] only process numbers ≤ √ J */ #= #+1; @.#= j; sq.#= j*j /*bump # Ps; assign next P; P squared*/ end /*j*/; return output   when using the default inputs: index │ Frobenius numbers that are < 10,000 ───────┼─────────────────────────────────────────────────────────────────────────────────────────────────────────────── 1 │ 1 7 23 59 119 191 287 395 615 839 11 │ 1,079 1,439 1,679 1,931 2,391 3,015 3,479 3,959 4,619 5,039 21 │ 5,615 6,395 7,215 8,447 9,599 ───────┴─────────────────────────────────────────────────────────────────────────────────────────────────────────────── Found 25 Frobenius numbers that are < 10,000 Ring[edit] load "stdlib.ring" # for isprime() function ? "working..." + nl + "Frobenius numbers are:" # create table of prime numbers between 2 and 101 inclusive Frob = [2] for n = 3 to 101 if isprime(n) Add(Frob,n) ok next m = 1 for n = 2 to len(Frob) fr = Frob[n] * Frob[m] - Frob[n] - Frob[m] see sf(fr, 4) + " " if m % 5 = 0 see nl ok m = n next ? nl + nl + "Found " + (m-1) + " Frobenius numbers" + nl + "done..." # a very plain string formatter, intended to even up columnar outputs def sf x, y s = string(x) l = len(s) if l > y y = l ok return substr(" ", 11 - y + l) + s Output: working... Frobenius numbers are: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 Found 25 Frobenius numbers done... Ruby[edit] require 'prime' Prime.each_cons(2) do |p1, p2| f = p1*p2-p1-p2 break if f > 10_000 puts f end Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 Rust[edit] // [dependencies] // primal = "0.3" fn frobenius_numbers() -> impl std::iter::Iterator<Item = (usize, bool)> { let mut primes = primal::Primes::all(); let mut prime = primes.next().unwrap(); std::iter::from_fn(move || { if let Some(p) = primes.by_ref().next() { let fnum = prime * p - prime - p; prime = p; return Some((fnum, primal::is_prime(fnum as u64))); } None }) } fn main() { let limit = 1000000; let mut count = 0; println!( "Frobenius numbers less than {} (asterisk marks primes):", limit ); for (fnum, is_prime) in frobenius_numbers().take_while(|(x, _)| *x < limit) { count += 1; let c = if is_prime { '*' } else { ' ' }; let s = if count % 10 == 0 { '\n' } else { ' ' }; print!("{:6}{}{}", fnum, c, s); } println!(); } Output: Frobenius numbers less than 1000000 (asterisk marks primes): 1 7* 23* 59* 119 191* 287 395 615 839* 1079 1439* 1679 1931* 2391 3015 3479 3959 4619 5039* 5615 6395 7215 8447* 9599 10199 10811 11447* 12095 14111 16379 17679 18767 20423 22199 23399* 25271 26891* 28551 30615 32039 34199 36479* 37631 38807 41579* 46619* 50171 51527 52895 55215 57119* 59999* 63999 67071 70215 72359 74519 77279* 78959 82343 89351 94859 96719 98591 104279 110879* 116255 120407 122495 126015 131027 136151 140615 144395 148215 153647 158399 163199* 170543 175559 180599 185759 189215 193595 198015 204287 209759 212519 215291 222747 232307* 238139 244019 249995 255015 264159 271439 281879 294839 303575 312471 319215 323759* 328319 337535 346911 354015 358799 363599 370871* 376991 380687 389339 403199 410879 414731* 421191 429015 434279 443519 454271 461031 470579* 482999 495599 508343 521267* 531431 540215 547595 556499 566999* 574559 583679 592895 606791* 625655 643167 654479 664199* 674039 678971* 683927 693863 713975 729311 734447 739595 755111 770879 776159* 781451 802715 824459* 835379* 851903 868607 879839* 889239 900591 919631* 937019 946719 958431 972179 986039 Sidef[edit] func frobenius_number(n) { prime(n) * prime(n+1) - prime(n) - prime(n+1) } say gather { 1..Inf -> each {|k| var n = frobenius_number(k) break if (n >= 10_000) take(n) } } Output: [1, 7, 23, 59, 119, 191, 287, 395, 615, 839, 1079, 1439, 1679, 1931, 2391, 3015, 3479, 3959, 4619, 5039, 5615, 6395, 7215, 8447, 9599] Wren[edit] Library: Wren-math Library: Wren-seq Library: Wren-fmt import "/math" for Int import "/seq" for Lst import "/fmt" for Fmt var primes = Int.primeSieve(101) var frobenius = [] for (i in 0...primes.count-1) { var frob = primes[i]*primes[i+1] - primes[i] - primes[i+1] if (frob >= 10000) break frobenius.add(frob) } System.print("Frobenius numbers under 10,000:") for (chunk in Lst.chunks(frobenius, 9)) Fmt.print("$,5d", chunk) System.print("\n%(frobenius.count) such numbers found.") Output: Frobenius numbers under 10,000: 1 7 23 59 119 191 287 395 615 839 1,079 1,439 1,679 1,931 2,391 3,015 3,479 3,959 4,619 5,039 5,615 6,395 7,215 8,447 9,599 25 such numbers found. XPL0[edit] func IsPrime(N); \Return 'true' if N is a prime number int N, I; [if N <= 1 then return false; for I:= 2 to sqrt(N) do if rem(N/I) = 0 then return false; return true; ]; int Count, M, Pn, Pn1, F; [Count:= 0; M:= 2; \first prime Pn:= M; loop [repeat M:= M+1 until IsPrime(M); Pn1:= M; F:= Pn*Pn1 - Pn - Pn1; if F >= 10_000 then quit; IntOut(0, F); Count:= Count+1; if rem(Count/10) = 0 then CrLf(0) else ChOut(0, 9\tab\); Pn:= Pn1; ]; CrLf(0); IntOut(0, Count); Text(0, " Frobenius numbers found below 10,000. "); ] Output: 1 7 23 59 119 191 287 395 615 839 1079 1439 1679 1931 2391 3015 3479 3959 4619 5039 5615 6395 7215 8447 9599 25 Frobenius numbers found below 10,000. Yabasic[edit] Translation of: PureBasic sub isPrime(v) if v < 2 then return False : fi if mod(v, 2) = 0 then return v = 2 : fi if mod(v, 3) = 0 then return v = 3 : fi d = 5 while d * d <= v if mod(v, d) = 0 then return False else d = d + 2 : fi wend return True end sub pn = 2 n = 0 for i = 3 to 9999 step 2 if isPrime(i) then n = n + 1 f = pn * i - pn - i if f > 10000 then break : fi print n, " => ", f pn = i end if next i end Output: Igual que la entrada de PureBasic.
__label__pos
0.898945
Skip to content This repository has been archived by the owner. It is now read-only. DEPRECATED: Merged into https://github.com/tendermint/tendermint under `abci` Go Python Shell Makefile Branch: master Clone or download README.md Application BlockChain Interface (ABCI) CircleCI Blockchains are systems for multi-master state machine replication. ABCI is an interface that defines the boundary between the replication engine (the blockchain), and the state machine (the application). Using a socket protocol, a consensus engine running in one process can manage an application state running in another. Previously, the ABCI was referred to as TMSP. The community has provided a number of addtional implementations, see the Tendermint Ecosystem Specification A detailed description of the ABCI methods and message types is contained in: For more background information on ABCI, motivations, and tendermint, please visit the documentation. The two guides to focus on are the Application Development Guide and Using ABCI-CLI. Protocl Buffers To compile the protobuf file, run: make protoc See protoc --help and the Protocol Buffers site for details on compiling for other languages. Note we also include a GRPC service definition. Install ABCI-CLI The abci-cli is a simple tool for debugging ABCI servers and running some example apps. To install it: go get github.com/tendermint/abci cd $GOPATH/src/github.com/tendermint/abci make get_vendor_deps make install Implementation We provide three implementations of the ABCI in Go: • Golang in-process • ABCI-socket • GRPC Note the GRPC version is maintained primarily to simplify onboarding and prototyping and is not receiving the same attention to security and performance as the others In Process The simplest implementation just uses function calls within Go. This means ABCI applications written in Golang can be compiled with TendermintCore and run as a single binary. See the examples below for more information. Socket (TSP) ABCI is best implemented as a streaming protocol. The socket implementation provides for asynchronous, ordered message passing over unix or tcp. Messages are serialized using Protobuf3 and length-prefixed with a signed Varint For example, if the Protobuf3 encoded ABCI message is 0xDEADBEEF (4 bytes), the length-prefixed message is 0x08DEADBEEF, since 0x08 is the signed varint encoding of 4. If the Protobuf3 encoded ABCI message is 65535 bytes long, the length-prefixed message would be like 0xFEFF07.... Note the benefit of using this varint encoding over the old version (where integers were encoded as <len of len><big endian len> is that it is the standard way to encode integers in Protobuf. It is also generally shorter. GRPC GRPC is an rpc framework native to Protocol Buffers with support in many languages. Implementing the ABCI using GRPC can allow for faster prototyping, but is expected to be much slower than the ordered, asynchronous socket protocol. The implementation has also not received as much testing or review. Note the length-prefixing used in the socket implementation does not apply for GRPC. Usage The abci-cli tool wraps an ABCI client and can be used for probing/testing an ABCI server. For instance, abci-cli test will run a test sequence against a listening server running the Counter application (see below). It can also be used to run some example applications. See the documentation for more details. Examples Check out the variety of example applications in the example directory. It also contains the code refered to by the counter and kvstore apps; these apps come built into the abci-cli binary. Counter The abci-cli counter application illustrates nonce checking in transactions. It's code looks like: func cmdCounter(cmd *cobra.Command, args []string) error { app := counter.NewCounterApplication(flagSerial) logger := log.NewTMLogger(log.NewSyncWriter(os.Stdout)) // Start the listener srv, err := server.NewServer(flagAddrC, flagAbci, app) if err != nil { return err } srv.SetLogger(logger.With("module", "abci-server")) if err := srv.Start(); err != nil { return err } // Wait forever cmn.TrapSignal(func() { // Cleanup srv.Stop() }) return nil } and can be found in this file. kvstore The abci-cli kvstore application, which illustrates a simple key-value Merkle tree func cmdKVStore(cmd *cobra.Command, args []string) error { logger := log.NewTMLogger(log.NewSyncWriter(os.Stdout)) // Create the application - in memory or persisted to disk var app types.Application if flagPersist == "" { app = kvstore.NewKVStoreApplication() } else { app = kvstore.NewPersistentKVStoreApplication(flagPersist) app.(*kvstore.PersistentKVStoreApplication).SetLogger(logger.With("module", "kvstore")) } // Start the listener srv, err := server.NewServer(flagAddrD, flagAbci, app) if err != nil { return err } srv.SetLogger(logger.With("module", "abci-server")) if err := srv.Start(); err != nil { return err } // Wait forever cmn.TrapSignal(func() { // Cleanup srv.Stop() }) return nil } You can’t perform that action at this time.
__label__pos
0.589976
对Excel表中的集美大学各省录取分数分析 2019/12/9 16:08:43 人评论 次浏览 分类:学习教程 版权声明:本文为博主原创文章,遵循 CC 4.0 BY-SA 版权协议,转载请附上原文出处链接和本声明。 本文链接:https://blog.csdn.net/CJY1204/article/details/103459686 对Excel表中的集美大学各省录取分数分析 主要功能 分析文件文件‘集美大学各省录取分数.xlsx’,完成: 集美大学2015-2018年间不同省份在本一批的平均分数(柱状图) 福建省这3年各批次成绩情况(折线图) 其他省份数据 首先要安装类库: xlrd库 numpy matplotlib.pyplot pandas pyecharts ##### 安装方式 pip install 库名 -i https://pypi.douban.com/simple import xlrd import numpy as np import matplotlib.pyplot as plt import pandas as pd from pyecharts import Line from pyecharts import Map plt.rcParams['font.sans-serif']=['SimHei'] class Exc(): def __init__(self,filePath,sheetName): self.data=xlrd.open_workbook(filePath) self.table=self.data.sheet_by_name(sheetName) self.rowNum = self.table.nrows print(u"总行数:", self.rowNum) self.colNum = self.table.ncols print(u"总列数:", self.colNum) self.j=1 self.databs={} for i in range(self.rowNum-1): values=self.table.row_values(self.j) self.databs[self.j]=values #print(self.databs[self.j]) self.j=self.j+1 def draw1(self,years): sfdict={} sflist=[] sfname=[] sfnum=[] qs={} for i in range(1,self.rowNum): if self.databs[i][7]==years and self.databs[i][1]=='本一批': try: sfdict[self.databs[i][0]]=(sfdict[self.databs[i][0]]+self.databs[i][6])/2 except KeyError: sfdict[self.databs[i][0]]=self.databs[i][6] mapsfname=list(sfdict.keys()) mapsfnum=list(sfdict.values()) map=Map('{}年全国各省本一批平均成绩'.format(years),'中国地图',width=1200,height=600) map.add("", mapsfname, mapsfnum, maptype='china',visual_range=[300, 700], is_visualmap=True, visual_text_color='#000') map.render(path="{}年全国各省本一批平均成绩.html".format(years)) sflist=sorted(zip(sfdict.values(),sfdict.keys())) for i in sflist[-10:]: sfname.append(i[1]) sfnum.append(i[0]) qs['省份']=sfname qs['平均分']=sfnum print(qs) qs=pd.DataFrame(qs) qs.plot(x='省份',y='平均分',kind='bar') for i in range(len(sfnum)): plt.text(i-0.2,sfnum[i],sfnum[i]) plt.title("{}年本一批平均分数排名前十的省份".format(years),fontsize=20) plt.xticks(fontsize=8,rotation=0) plt.savefig('{}年本一批平均分数排名前十的省份.jpg'.format(years),bbox_inches='tight') plt.show() def setV(sfpc,years,i): xdic={} for j in years: xdic[j]=sfpc[j][i] return xdic def draw2(self,sf): sfpc={} pc={} pc2016={} pc2017={} pc2018={} xdic={} pclist=[] sfpclist=[] sfpcname=[] sfpcnum=[] years=[2016,2017,2018] #for i in range(1,self.rowNum): # if self.databs[i][0]==sf and self.databs[i][2]=='理工': # if self.databs[i][1] not in pclist: # pclist.append(self.databs[i][1]) for i in range(1,self.rowNum): if self.databs[i][0]==sf and self.databs[i][2]=='理工' and self.databs[i][7]==2017: pc2017[self.databs[i][1]]=self.databs[i][5] sfpc[self.databs[i][7]]=pc2017 if self.databs[i][1] not in pclist: pclist.append(self.databs[i][1]) elif self.databs[i][0]==sf and self.databs[i][2]=='理工' and self.databs[i][7]==2018: pc2018[self.databs[i][1]]=self.databs[i][5] sfpc[self.databs[i][7]]=pc2018 elif self.databs[i][0]==sf and self.databs[i][2]=='理工' and self.databs[i][7]==2016: pc2016[self.databs[i][1]]=self.databs[i][5] sfpc[self.databs[i][7]]=pc2016 #print(sfpc) line = Line("福建省这3年各批次成绩情况",title_top="90%") n=0 for i in pclist: for j in years: sfpcnum.append(sfpc[j][i]) #print(i) line.add(i,years,sfpcnum[n:],is_label_show=True)#i 批次的名字 yeas 表示Y轴 sfpcnum 表示录取成绩 pc[i]=Exc.setV(sfpc,years,i) n=n+3 #pc=zip(sfpc.keys(),sfpc.values()) #print(pc) line.render('福建省这3年各批次成绩情况.html') if __name__=='__main__': filePath="E:\Python_str\集美录取分\集美录取分\\集美大学各省录取分数.xlsx" sheetName='集美大学各省录取分数' data=Exc(filePath,sheetName) #print(data.dict_data()) data.draw1(2015) data.draw1(2016) data.draw1(2017) data.draw1(2018) data.draw2('福建') #data.draw3() 实验结果 在这里插入图片描述 在这里插入图片描述 在这里插入图片描述 在这里插入图片描述 相关资讯 暂无相关的资讯... 共有访客发表了评论 网友评论 验证码: 看不清楚? -->
__label__pos
0.714408
Triangle calculator Please enter what you know about the triangle: Symbols definition of ABC triangle You have entered side a, angle β and angle γ. Right scalene triangle. Sides: a = 60   b = 34.64110161514   c = 69.28220323028 Area: T = 1039.233048454 Perimeter: p = 163.9233048454 Semiperimeter: s = 81.96215242271 Angle ∠ A = α = 60° = 1.04771975512 rad Angle ∠ B = β = 30° = 0.52435987756 rad Angle ∠ C = γ = 90° = 1.57107963268 rad Height: ha = 34.64110161514 Height: hb = 60 Height: hc = 30 Median: ma = 45.82657569496 Median: mb = 62.4549979984 Median: mc = 34.64110161514 Inradius: r = 12.67994919243 Circumradius: R = 34.64110161514 Vertex coordinates: A[69.28220323028; 0] B[0; 0] C[51.96215242271; 30] Centroid: CG[40.41545188433; 10] Coordinates of the circumscribed circle: U[34.64110161514; -0] Coordinates of the inscribed circle: I[47.32105080757; 12.67994919243] Exterior(or external, outer) angles of the triangle: ∠ A' = α' = 120° = 1.04771975512 rad ∠ B' = β' = 150° = 0.52435987756 rad ∠ C' = γ' = 90° = 1.57107963268 rad Calculate another triangle How did we calculate this triangle? 1. Input data entered: side a, angle β and angle γ. a = 60 ; ; beta = 30° ; ; gamma = 90° ; ; 2. From angle β and angle γ we calculate angle α: beta + gamma + alpha = 180° ; ; alpha = 180° - beta - gamma = 180° - 30 ° - 90 ° = 60 ° ; ; 3. From angle β, angle α and side a we calculate side b - By using the law of sines, we calculate unknown side b: fraction{ b }{ a } = fraction{ sin beta }{ sin alpha } ; ; ; ; b = a * fraction{ sin beta }{ sin alpha } ; ; ; ; b = 60 * fraction{ sin 30° }{ sin 60° } = 34.64 ; ; 4. From angle γ, angle α and side a we calculate side c - By using the law of sines, we calculate unknown side c: fraction{ c }{ a } = fraction{ sin gamma }{ sin alpha } ; ; ; ; c = a * fraction{ sin gamma }{ sin alpha } ; ; ; ; c = 60 * fraction{ sin 90° }{ sin 60° } = 69.28 ; ; Now we know the lengths of all three sides of the triangle and the triangle is uniquely determined. Next we calculate another its characteristics - same procedure as calculation of the triangle from the known three sides SSS. a = 60 ; ; b = 34.64 ; ; c = 69.28 ; ; 5. The triangle circumference is the sum of the lengths of its three sides p = a+b+c = 60+34.64+69.28 = 163.92 ; ; 6. Semiperimeter of the triangle s = fraction{ o }{ 2 } = fraction{ 163.92 }{ 2 } = 81.96 ; ; 7. The triangle area using Heron's formula T = sqrt{ s(s-a)(s-b)(s-c) } ; ; T = sqrt{ 81.96 * (81.96-60)(81.96-34.64)(81.96-69.28) } ; ; T = sqrt{ 1080000 } = 1039.23 ; ; 8. Calculate the heights of the triangle from its area. T = fraction{ a h _a }{ 2 } ; ; h _a = fraction{ 2 T }{ a } = fraction{ 2 * 1039.23 }{ 60 } = 34.64 ; ; h _b = fraction{ 2 T }{ b } = fraction{ 2 * 1039.23 }{ 34.64 } = 60 ; ; h _c = fraction{ 2 T }{ c } = fraction{ 2 * 1039.23 }{ 69.28 } = 30 ; ; 9. Calculation of the inner angles of the triangle using a Law of Cosines a**2 = b**2+c**2 - 2bc cos alpha ; ; alpha = arccos( fraction{ b**2+c**2-a**2 }{ 2bc } ) = arccos( fraction{ 34.64**2+69.28**2-60**2 }{ 2 * 34.64 * 69.28 } ) = 60° ; ; b**2 = a**2+c**2 - 2ac cos beta ; ; beta = arccos( fraction{ a**2+c**2-b**2 }{ 2ac } ) = arccos( fraction{ 60**2+69.28**2-34.64**2 }{ 2 * 60 * 69.28 } ) = 30° ; ; gamma = 180° - alpha - beta = 180° - 60° - 30° = 90° ; ; 10. Inradius T = rs ; ; r = fraction{ T }{ s } = fraction{ 1039.23 }{ 81.96 } = 12.68 ; ; 11. Circumradius R = fraction{ a }{ 2 * sin alpha } = fraction{ 60 }{ 2 * sin 60° } = 34.64 ; ; 12. Calculation of medians m_a = fraction{ sqrt{ 2 b**2+2c**2 - a**2 } }{ 2 } = fraction{ sqrt{ 2 * 34.64**2+2 * 69.28**2 - 60**2 } }{ 2 } = 45.826 ; ; m_b = fraction{ sqrt{ 2 c**2+2a**2 - b**2 } }{ 2 } = fraction{ sqrt{ 2 * 69.28**2+2 * 60**2 - 34.64**2 } }{ 2 } = 62.45 ; ; m_c = fraction{ sqrt{ 2 b**2+2a**2 - c**2 } }{ 2 } = fraction{ sqrt{ 2 * 34.64**2+2 * 60**2 - 69.28**2 } }{ 2 } = 34.641 ; ; Calculate another triangle Look also our friend's collection of math examples and problems: See more information about triangles or more details on solving triangles.
__label__pos
1
Blog How to Develop a QR Code: A Step-by-Step Guide In today’s fast-paced electronic earth, little organizations need successful resources for connecting with consumers and streamline operations. QR signal machines give you a easy yet powerful alternative for achieving these goals. Using a QR code turbine, small corporations can increase their marketing efforts, increase customer diamond, and simplify transactions. One of many principal great things about utilizing a QR rule turbine is the ability to produce easy-to-access digital content. For example, little companies can use QR limitations to url with their sites, on line possibilities, or promotional offers. By reading the QR code with their smartphones, clients may quickly entry that material and never having to key in long URLs or look for information. This convenience can lead to raised proposal and improved traffic to the business’s online platforms. QR rules provide a valuable chance for little businesses to QR Code Generator client information and monitor the potency of their marketing campaigns. Many QR code machines offer analytics characteristics that allow firms to monitor the amount of tests, the location of the tests, and the units used. That information can help corporations understand client behavior and preferences, permitting them to tailor their advertising methods more effectively. Yet another benefit of QR requirements is their versatility in streamlining operations. Little corporations can use QR rules to facilitate contactless obligations, streamline catalog administration, and actually handle customer devotion programs. As an example, a QR rule linked to a payment gateway allows clients to complete transactions quickly and safely, lowering the requirement for bodily income handling or card swipes. Furthermore, QR codes are cost-effective and an easy task to implement, creating them an ideal tool for small organizations with limited resources. Generating a QR code is normally free, and organizations may print them on different resources, such as for instance flyers, posters, or product packaging. This low-cost strategy allows small firms to achieve a larger market with out a substantial expense in new technology. In conclusion, employing a QR rule generator presents numerous advantages for small firms, including increased client wedding, useful data ideas, streamlined operations, and cost-effectiveness. By incorporating QR rules within their advertising and detailed techniques, small firms may stay aggressive and offer a more seamless knowledge for his or her customers. In a global wherever QR limitations are increasingly employed for marketing, conversation, and transactions, customization has become critical to position out. Customizing your QR rules not merely makes them successfully fascinating but additionally supports your manufacturer identity. Here’s how you can customize your QR codes for maximum affect employing a QR signal generator. The first step in customizing a QR code is to select a QR signal turbine that offers sophisticated modification options. While many generators allow you to develop simple black-and-white QR requirements, the others provide functions for adjusting the look, shade, and form of one’s QR code. Pick a turbine that aligns together with your customization wants and presents resources that enable for innovative freedom. One of the most common modification possibilities is changing the color system of your QR code. Rather than the common black-and-white style, you are able to choose colors that fit your brand’s visible identity. As an example, you can use your brand’s major shades for the QR code’s adventures and a complementary color for the background. However, it’s important to keep adequate distinction between the QR rule and its background to make sure it remains scannable. Introducing a logo or image to your QR code is yet another powerful method to personalize it. Several QR signal turbines allow you to place your brand’s brand or even a appropriate symbol in the middle of the QR code. This not only makes the QR code creatively distinctive but also supports manufacturer recognition. When introducing a brand, ensure it doesn’t block the QR code’s readability by maintaining it small and centrally placed. Shape modification is another choice that may produce your QR code more appealing. As opposed to the conventional square adventures, some QR code generators permit you to create requirements with spherical sides or other unique shapes. This can give your QR code a gentler, newer look. Also, you can style the corners of the QR rule with custom styles or icons that align with your brand’s aesthetic. LEAVE A RESPONSE Your email address will not be published. Required fields are marked * Related Posts
__label__pos
0.750096
0 I have a network originally written using Estimator API and I would like to take advantage of tensorflow TensorRT integration. I can load the model from SavedModel and run create_inference_graph on it. However, when I try importing the graph definition obtained from "create_inference_graph", I get the following error: File "/python2.7/site-packages/tensorflow/python/framework/importer.py", line 422, in import_graph_def raise ValueError(str(e)) ValueError: Node 'my_trt_op_6': Unknown input node 'my_trt_op_0' I get the same error when I freeze the tftrt graph and try loading it into tensorboard. Printing out the nodes in the tftrt GraphDef outputs a terminal full of nonsensical to me values looking like \000\000\000\000\000\002\000\000\000\000\000\200?\000\000\200?\n\000\000\000OutputPH_0\000\000\010\000\014\000\004\000\010\000\010\000\000\000\010\000\000\000\024\000\000\000\003\000\000\000\001\000\000\000\000\000\000\000\000\000\000\000\003\000\000\000\002\000\000\000\034\000\000\000\034\000\000\000 (this is just a small sample of the terminal output). I can run simple inference using sess.run(y_tensor, feed_dict={x_tensor: x_test}) on my loaded model before using tftrt optimization. I also tested all the code I am using (saving estimator model, loading model into a session without Estimator API, converting graph into trt-optimized graph and running accelerated inference) on a sample MNIST Estimator model from official tensorflow model repo, and the code worked fine for that model. I do get quite a few errors and warnings from convert_graph and trt_logger when I run create_inference_graph, but based on my results from converting other models, those errors and warning do not usually indicate fatal failures. I can post them here if it might be useful for understanding why this happens. How can I fix this issue and run inference on trt-optimized version of my model? Your Answer By clicking “Post Your Answer”, you agree to our terms of service, privacy policy and cookie policy Browse other questions tagged or ask your own question.
__label__pos
0.974336
Multimap From Wikipedia, the free encyclopedia Jump to: navigation, search This article is about the data type. For the mathematical concept, see Multivalued function. For the mapping website, see Multimap.com. In computer science, a multimap (sometimes also multihash) is a generalization of a map or associative array abstract data type in which more than one value may be associated with and returned for a given key. Both map and multimap are particular cases of containers (see for example C++ Standard Template Library containers). Often the multimap is implemented as a map with lists or sets as the map values. Examples[edit] • In a student enrollment system, where students may be enrolled in multiple classes simultaneously, there might be an association for each enrollment of a student in a course, where the key is the student ID and the value is the course ID. If a student is enrolled in three courses, there will be three associations containing the same key. • The index of a book may report any number of references for a given index term, and thus may be coded as a multimap from index terms to any number of reference locations. • Querystrings may have multiple values associated with a single field. This is commonly generated when a web form allows multiple check boxes or selections to be chosen in response to a single form element. Language support[edit] C++'s Standard Template Library provides the multimap container for the sorted multimap using a self-balancing binary search tree,[1] and SGI's STL extension provides the hash_multimap container, which implements a multimap using a hash table.[2] Apache Commons Collections provides a MultiMap interface for Java.[3] It also provides a MultiValueMap implementing class that makes a MultiMap out of a Map object and a type of Collection.[4] Google Guava (previously named Google Collections) also provides an interface Multimap and implementations. [5] The Scala programming language's API also provides Multimap and implementations[6] OCaml's standard library module Hashtbl implements a hash table where it's possible to store multiple values for a key. Quiver provides a Multimap for Dart.[7] See also[edit] References[edit] 1. ^ "multimap<Key, Data, Compare, Alloc>". Standard Template Library Programmer's Guide. Silicon Graphics International.  2. ^ "hash_multimap<Key, HashFcn, EqualKey, Alloc>". Standard Template Library Programmer's Guide. Silicon Graphics International.  3. ^ "Interface MultiMap". Commons Collections 3.2.1 API, Apache Commons.  4. ^ "Class MultiValueMap". Commons Collections 3.2.1 API, Apache Commons.  5. ^ "Interface Multimap<K,V>". Guava Library 2.0.  6. ^ "Scala.collection.mutable.MultiMap". Scala stable API.  7. ^ "Multimap". Quiver API docs. 
__label__pos
0.923027
java 关注公众号 jb51net 关闭 首页 > 软件编程 > java > SpringBoot使用JSch SpringBoot使用JSch操作Linux的方法 作者:C_C_菜园 JSch是一个Java库,它提供了SSH(Secure Shell)的Java实现,允许Java程序通过SSH协议连接到远程系统(如Linux),这篇文章主要介绍了SpringBoot使用JSch操作Linux,需要的朋友可以参考下 推荐使用Hutool的Jsch工具包(它用的连接池的技术) 一、SSH远程连接服务器 SSH更多见:http://t.csdnimg.cn/PrsNv 推荐连接工具:FinalShell、Xshell、secureCRT、PuTTY https://www.jb51.net/article/232575.htm 1、SSH(Secure Shell)主要有两大功能 二、JNA、Process和JSchJNA JNA主要用于在Java程序中调用本地库的函数,而不是用于远程连接到其他系统。如果你的Java程序正在Linux系统上运行,你可以使用JNA来调用Linux的本地库函数,包括那些可以获取文件和文件夹路径的函数。然而,如果你的Java程序正在一个系统(如Windows)上运行,你不能使用JNA来连接到另一个系统(如Linux)。 如果你需要从一个运行在Windows上的Java程序连接到一个Linux系统,你可能需要使用其他的工具或库。例如,你可以使用SSH(安全壳层)来远程连接到Linux系统,然后执行命令来获取文件和文件夹的路径。在Java中,有一些库可以帮助你使用SSH,如JSch和Apache MINA SSHD。 简单理解JNA、Process和JSch (Java Native Access)和Process类都是Java中与本地系统交互的工具。JNA允许Java代码直接调用本地(C/C++)库的函数,而Process类则允许Java代码启动和控制操作系统的进程,例如执行shell命令。(JNA和Process是Java调用系统(Windows、Linux等)的本地函数,或者三方程序) JSch是一个Java库,它提供了SSH(Secure Shell)的Java实现,允许Java程序通过SSH协议连接到远程系统(如Linux)。一旦连接成功,你可以通过JSch执行远程命令,上传和下载文件,就像直接在远程系统上操作一样。(JSch则是Java连接系统(Windows、Linux等)的工具,比如连接上Linux后,相当于直接操作Linux一样) 三、Java使用SSH的包 3.1、JSch和Apache MINA SSHD JSch和Apache MINA SSHD都是优秀的SSH库,它们各有优点,选择哪一个主要取决于你的具体需求。 JSch是一个成熟且广泛使用的库,它提供了SSH2的完整实现,包括SFTP,SCP,端口转发等功能。JSch的API相对简单,易于使用,而且JSch的社区活跃,有大量的教程和示例代码可供参考。 Apache MINA SSHD则是一个更现代的库,它基于Apache MINA,一个高性能的网络应用框架。MINA SSHD提供了SSH2的完整实现,包括SFTP,SCP,端口转发等功能。MINA SSHD的API设计更现代,更符合Java的编程习惯,而且MINA SSHD支持异步非阻塞IO,对于需要处理大量并发连接的应用来说,可能会有更好的性能。 总的来说,如果你需要一个简单易用,社区支持好的SSH库,JSch可能是一个不错的选择。如果你需要一个设计现代,支持异步非阻塞IO的SSH库,或者你已经在使用Apache MINA,那么MINA SSHD可能更适合你。 3.2、JSch的四种认证机制: 四、JSch实现登录Linux,远程命令执行、SFTP下载和上传文件 4.1、导包Jsch官方的包上次更新18年(本文使用) // jsch包 implementation 'com.jcraft:jsch:0.1.55' 长期维护的jsch:https://github.com/mwiede/jsch 4.2、Jsch工具类 package com.cc.jschdemo.utils; import com.jcraft.jsch.*; import lombok.Data; import org.apache.commons.lang3.StringUtils; import org.springframework.http.HttpHeaders; import org.springframework.stereotype.Component; import javax.annotation.PreDestroy; import javax.servlet.ServletOutputStream; import javax.servlet.http.HttpServletResponse; import java.io.*; import java.net.URLEncoder; import java.nio.charset.StandardCharsets; import java.util.*; /** * <p>JSch工具类</p> * <li>交给spring管理:每个使用的地方都是单例,都是单独的这个类。(new 也可以)</li> * * <li>所有方法都没有关闭(连接、会话),需要使用方自己关闭</li> * * @author CC * @since 2023/11/8 */ @Data @Component public class JSchUtil { //缓存session会话 private Session session; //通道:执行命令 private ChannelExec channelExec; //通道:SFTP private ChannelSftp channelSftp; //通道:执行复杂Shell命令 private ChannelShell channelShell; //登陆Linux服务器 public void loginLinux(String username, String password, String host, Integer port) { try { //每次都会重新初始化session if (Objects.isNull(session) || !session.isConnected()) { JSch jsch = new JSch(); session = jsch.getSession(username, host, port); session.setPassword(password); // 配置Session参数 Properties config = new Properties(); // 不进行公钥的检查 config.put("StrictHostKeyChecking", "no"); session.setConfig(config); // 设置连接超时时间(s/秒) session.setTimeout(300); } if (!session.isConnected()) { // 连接到远程服务器 session.connect(); } }catch(Exception e){ throw new RuntimeException("连接Linux失败:" + e.getMessage()); } } //执行命令:可以多次执行,然后必须调用关闭接口 public String executeCommand(String command) { StringBuilder result = new StringBuilder(); BufferedReader buf = null; try { //每次执行都创建新的通道 channelExec = (ChannelExec) session.openChannel("exec"); channelExec.setCommand(command); //正确的流中没有数据就走错误流中去拿。 InputStream in = channelExec.getInputStream(); InputStream errStream = channelExec.getErrStream(); channelExec.connect(); buf = new BufferedReader(new InputStreamReader(in)); String msg; while ((msg = buf.readLine()) != null) { result.append(msg); } if (StringUtils.isBlank(result.toString())) { buf = new BufferedReader(new InputStreamReader(errStream)); String msgErr; while ((msgErr = buf.readLine()) != null) { result.append(msgErr); } } }catch(Exception e){ throw new RuntimeException("关闭连接失败(执行命令):" + e.getMessage()); }finally { if (Objects.nonNull(buf)) { try { buf.close(); }catch(Exception e){ e.printStackTrace(); } } } return result.toString(); } /** * 执行复杂shell命令 * * @param cmds 多条命令 * @return 执行结果 * @throws Exception 连接异常 */ public String execCmdByShell(List<String> cmds) { String result = ""; try { channelShell = (ChannelShell) session.openChannel("shell"); InputStream inputStream = channelShell.getInputStream(); channelShell.setPty(true); channelShell.connect(); OutputStream outputStream = channelShell.getOutputStream(); PrintWriter printWriter = new PrintWriter(outputStream); for (String cmd : cmds) { printWriter.println(cmd); } printWriter.flush(); byte[] tmp = new byte[1024]; while (true) { while (inputStream.available() > 0) { int i = inputStream.read(tmp, 0, 1024); if (i < 0) { break; } String s = new String(tmp, 0, i); if (s.contains("--More--")) { outputStream.write((" ").getBytes()); outputStream.flush(); } System.out.println(s); } if (channelShell.isClosed()) { System.out.println("exit-status:" + channelShell.getExitStatus()); break; } //间隔1s后再执行 try { Thread.sleep(1000); } catch (Exception e) { e.printStackTrace(); } } outputStream.close(); inputStream.close(); }catch(Exception e){ e.printStackTrace(); } return result; } //下载除了云服务器的文件(你自己的服务器):因为云服务器,像阿里云服务器下载文件好像是一段一段给你的,不是一起给你。 public void downloadOtherFile(String remoteFileAbsolutePath, String fileName, HttpServletResponse response) { try { channelSftp = (ChannelSftp) session.openChannel("sftp"); channelSftp.connect(); //获取输入流 InputStream inputStream = channelSftp.get(remoteFileAbsolutePath); //直接下载到本地文件 // channelSftp.get(remoteFileAbsolutePath, "D:\\Develop\\Test\\studio-3t-x64.zip"); response.setCharacterEncoding(StandardCharsets.UTF_8.name()); response.setContentType("application/octet-stream;charset=".concat(StandardCharsets.UTF_8.name())); response.setHeader(HttpHeaders.ACCESS_CONTROL_EXPOSE_HEADERS, HttpHeaders.CONTENT_DISPOSITION); response.setHeader(HttpHeaders.CONTENT_DISPOSITION, "attachment; filename=".concat( URLEncoder.encode(fileName, StandardCharsets.UTF_8.name()) )); ServletOutputStream out = response.getOutputStream(); // 从InputStream输入流读取数据 并写入到ServletOutputStream输出流 byte[] buffer = new byte[4096]; int bytesRead; while ((bytesRead = inputStream.read(buffer)) != -1) { out.write(buffer, 0, bytesRead); } out.flush(); out.close(); }catch(Exception e){ throw new RuntimeException("关闭连接失败(下载文件):" + e.getMessage()); } } //下载云服务器的文件(因为云服务器传文件是一段一段的,所以不能直接像操作我们的服务器一样直接下载)(阿里云为例) public void downloadCloudServerFile(String remoteFileAbsolutePath, String fileName, HttpServletResponse response) { try { channelSftp = (ChannelSftp) session.openChannel("sftp"); channelSftp.connect(); //获取输入流 InputStream inputStream = channelSftp.get(remoteFileAbsolutePath); //阿里云应该是断点续传,后面研究…… }catch(Exception e){ throw new RuntimeException("关闭连接失败(下载文件):" + e.getMessage()); } } //ls命令:获取文件夹的信息 public String ls(String path){ StringBuilder sb = new StringBuilder(); try { channelSftp = (ChannelSftp) session.openChannel("sftp"); channelSftp.connect(); Vector ls = channelSftp.ls(path); Iterator iterator = ls.iterator(); while (iterator.hasNext()) { Object next = iterator.next(); System.out.println(next); sb.append(next); } } catch (Exception e){ throw new RuntimeException(e.getMessage()); } return sb.toString(); } //关闭通道:释放资源 private void closeChannel(){ //不为空,且已经连接:关闭 if (Objects.nonNull(channelExec)) { channelExec.disconnect(); } if (Objects.nonNull(channelSftp)) { channelSftp.disconnect(); } if (Objects.nonNull(channelShell)) { channelShell.disconnect(); } } /** 关闭通道、关闭会话:释放资源 * spring销毁前,关闭 所有会话 及 所有通道 */ @PreDestroy public void closeAll(){ System.out.println("我被销毁了。。。。。。。。。。。。。。。。。。。。。。"); this.closeChannel(); if (Objects.nonNull(session) && session.isConnected()) { session.disconnect(); } } } 4.2、使用Jsch工具类:执行命令 4.2.1、执行简单命令 package com.cc.jschdemo.web.controller; import com.cc.jschdemo.utils.JSchUtil; import com.jcraft.jsch.ChannelExec; import org.springframework.web.bind.annotation.GetMapping; import org.springframework.web.bind.annotation.RequestMapping; import org.springframework.web.bind.annotation.RequestParam; import org.springframework.web.bind.annotation.RestController; import javax.annotation.Resource; import java.io.InputStream; import java.util.Arrays; /** * <p></p> * * @author CC * @since 2023/11/8 */ @RestController @RequestMapping("/jsch") public class JSchController { @Resource private JSchUtil jSchUtil; /** <p>执行命令<p> **/ @GetMapping public String executeCommand() { //登陆(默认只连接5分钟,5分钟后销毁) jSchUtil.loginLinux("服务器账号", "服务器密码", "服务器IP", 服务器端口); //一、执行命令 String mkdir = jSchUtil.executeCommand("mkdir ccc"); String docker = jSchUtil.executeCommand("docker"); String dockerPs = jSchUtil.executeCommand("docker ps"); System.out.println(mkdir); System.out.println(docker); System.out.println(dockerPs); //执行完,关闭连接 jSchUtil.closeAll(); return docker; } } 结果:多了一个文件夹 4.2.2、执行复杂的shell命令 /** <p>执行命令<p> **/ @PostMapping public String execCmdByShell() { //登陆(默认只连接5分钟,5分钟后销毁) jSchUtil.loginLinux("服务器账号", "服务器密码", "服务器IP", 服务器端口); //二、执行shell脚本(可以改造成传入的shell脚步) jSchUtil.execCmdByShell(Arrays.asList("cd /", "ll" , "cd cc/", "mkdir ccccc222", "ll")); //执行完,关闭连接 jSchUtil.closeAll(); return "docker"; } 结果 4.3、使用Jsch工具类:下载文件 4.3.1、普通服务器下载 //下载普通服务器的文件 @PutMapping public void downloadOtherFile(HttpServletResponse response) { //登陆(默认只连接5分钟,5分钟后销毁) jSchUtil.loginLinux("服务器账号", "服务器密码", "服务器IP", 服务器端口); //下载文件 jSchUtil.downloadOtherFile( "/dev/libbb/studio-3t-x64.zip", "studio-3t-x64.zip", response ); //执行完,关闭连接 jSchUtil.closeAll(); } 4.3.2、阿里云服务器下载 https://www.jb51.net/article/178581.htm 五、Hutool工具封装的JSch(推荐) Hutool使用的是JSch连接池,推荐使用…… 六、总结 参考: https://www.jb51.net/article/264152.htm https://www.jb51.net/article/264148.htm 到此这篇关于SpringBoot使用JSch操作Linux的文章就介绍到这了,更多相关SpringBoot使用JSch内容请搜索脚本之家以前的文章或继续浏览下面的相关文章希望大家以后多多支持脚本之家! 您可能感兴趣的文章: 阅读全文
__label__pos
0.973361
Paste the following code in the active theme’s functions.php: /* if the product quantity is between X and X, display "plus que X en stock" in cart and checkout */ add_filter( 'woocommerce_cart_item_name', 'showing_stock_in_cart_items', 99, 3 ); function showing_stock_in_cart_items( $item_name, $cart_item, $cart_item_key ) { // The WC_Product object $product = $cart_item['data']; if (empty($product)) { return $item_name; } // Get the stock $stock = $product->get_stock_quantity(); // When stock doesn't exist if (empty($stock)) { return $item_name; } // display the stock if ($stock >= '1' && $stock <= '3') { /* was ($stock <= '3') */ $item_name .=" × " . $cart_item['quantity'] . " " . ' '.__( "Plus que " .$stock. " en stock","woocommerce").' '; } else { $item_name .=" × " . $cart_item['quantity'] . " "; } return $item_name; } Edit quantity (range) and text as needed. Then, paste the following code in Appearance > Customize > Additional CSS: /* remove native item quantity in checkout */ .woocommerce-checkout #order_review .product-quantity { display: none; } /* remove stock function quantity in cart and mini-cart*/ .woocommerce-cart .cart_item .function-product-quantity, .woocommerce-mini-cart .function-product-quantity { display: none !important; } Source: https://stackoverflow.com/questions/59001052/woocommerce-show-stock-status-on-cart-page (+ additional edits made by myself)
__label__pos
0.988309
Take the 2-minute tour × Stack Overflow is a question and answer site for professional and enthusiast programmers. It's 100% free, no registration required. Here's the code on JSFiddle. I need a sentence to slide out to the left (out of view), change text to next line in story (next element in array) and slide back right, into place. Also, I need to be able to have the whole thing repeat, but I expect that's just a matter of resetting i in the loop. Can you help me? HTML: ​<p class="wrap"><span>Foo</span></p>​​​​​​​​​​​​​​​​​​​​​​​​​​​​​​​​​​​​​​​​​ JS: var stringArr = ['One','Two','Three'], i = 0, t = 2000; for ( i ; i < 3 ; i++) { $('.wrap span').delay(t).animate({'left': '+=200px'}, 'slow'); $('.wrap span').text(stringArr[i]).appendTo($('.wrap span')); $('.wrap span').animate({'right': '+=200px'}, 'slow'); }​ What am I doing wrong? share|improve this question 3 Answers 3 up vote 1 down vote accepted You can achieve it simply by animating the left property of a positioned container: <div class="wrapper"> <div id="slider"></div> </div>​ with the following CSS (note position: relative for the slider): .wrapper { background-color: #ffbbbb; width: 200px; } #slider { width: 100%; position: relative; left: -100%; } This is the script I used in my fiddle: var strings = "Lorem ipsum dolor sit amen".split(" "), offset = 0, delay = 1500, slider = $("#slider"); var swap = function(element, strings, offset) { element.text(strings[(offset++) % strings.length]); element.animate({"left": "0%"}); element.delay(delay).animate({"left": "-100%"}); element.queue(function(next) { swap(element, strings, offset); next(); }); } swap(slider, strings, offset); share|improve this answer      This ran smooth from the start. Well done. Thanks. I've transfered my animation strategy to this one instead. Care to share now what you see wrong with setInterval()? –  Ryan Dec 3 '12 at 20:42      Basically, whenever you pass a delay to an API, you should really consider it a hint: the environment (browser, virtual machine, or OS) will do its best to execute the callback as soon as the delay expires, but in a multiprocess context there can be no guarantee about when your function will actually be called - on a busy machine two or more calls may overlap and cause glitches. When this sort of synchronization is needed you'd better rely on a clock, or put your code in a loop. In this case I used a callback –  Raffaele Dec 3 '12 at 21:10 var stringArr = ['One','Two','Three'], i = 0, t = 2000, initialDelay = 150, container = $(".wrap span"), len = stringArr.length, changer = function() { var step = (++i)%len; container.animate({'left': '+=200px'}, 'slow', function(){ container.text(stringArr[step]).animate({left: '-=200px'}, 'slow'); }) }, interval, stop = function(){ interval = interval !== undefined && clearInterval(interval); }, start = function() { stop(); interval = setInterval(changer, t); }; setTimeout(function() { changer(); start(); }, initialDelay); 1. You need to specify position for span otherwise left property won't affect it. 2. You can't append a node to itself. $("selector").appendTo($("selector")) throws dom exception. 3. You shouldn't animate left and right properties at the same time. It will stretch your element. http://jsfiddle.net/tarabyte/hBSdf/10/ share|improve this answer      It doesn't really work as intended. Also, setInterval() is a wrong choice here because there is no guarantee on the function start/end times, so you'll end up with glitches when two or more animations overlap. You can easily experience it with a smaller t and a slower animation. The solution here is using callbacks or a queue –  Raffaele Dec 1 '12 at 11:49      I notice the first time the page loads the animation looks choppy, but then it's fine. Any way to delay the start so that it looks great the first time? Also, how can I turn off the loop? –  Ryan Dec 1 '12 at 12:09      @Ryan: check my edit. Use initalDelay for later start, and stop, start functions to relaunch changing. –  Yury Tarabanko Dec 1 '12 at 12:18      @Raffaele: Actually it works. setInterval is precise enough provided that you are not going to win an oscar for best animation. :) –  Yury Tarabanko Dec 1 '12 at 12:25      @YuryTarabanko don't know, maybe you improved the code but didn't update the fiddle - still doesn't work, neither on page load nor after you click "start": the slider doesn't even slide offscreen. If it worked, I'd be able to show you what's wrong with setInterval() even if you don't plan to win an Oscar :) –  Raffaele Dec 1 '12 at 12:49 Still playing around with it but couple of things to start you off. The animate() function in jQuery requires a CSS Map, specifically a position field or else it won't really work. You can see for yourself. I don't really know why exactly so if someone wants to jump in and explain why please do ><... to be more specific in the documentation it says: The only required parameter is a map of CSS properties. This map is similar to the one that can be sent to the .css() method, except that the range of properties is more restrictive. Another idea you want is use a callback so that once your animation completes, it can change the word, and then maybe move it back in place. I don't think the delay function is wrong but it is a lot less dynamic and if you are supposedly making a book, it'd be nice if the program just knew to do something after something was completed. Here is a Fiddle. Maybe when I wake up I'll add some more/edit my answer but just something to get you started/thinking :) share|improve this answer Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.915856
use strict; use warnings; package HTML::MasonX::Free::Component; { $HTML::MasonX::Free::Component::VERSION = '0.004'; } use parent 'HTML::Mason::Component::FileBased'; # ABSTRACT: a component with a "main" method, not just a bunch of text sub new { my ($class, %arg) = @_; my $default_method_to_call = delete $arg{default_method_to_call} || 'main'; my $self = $class->SUPER::new(%arg); $self->{default_method_to_call} = $default_method_to_call; return $self; } sub run { my $self = shift; $self->{mfu_count}++; $self->call_method($self->{default_method_to_call} => @_); } 1; __END__ =pod =head1 NAME HTML::MasonX::Free::Component - a component with a "main" method, not just a bunch of text =head1 VERSION version 0.004 =head1 OVERVIEW In concept, a Mason component is broken down into special blocks (like once, shared, init), methods, and subcomponents. When you render a Mason component, using it as a template, you aren't calling one of its methods or blocks. Instead, all the stray code and text that was found I all of those is concatenated together and run. This is sort of a mess. If you use HTML::MasonX::Free::Component as your component class instead, rendering the component will call its C method instead of all that other junk. This component class extends HTML::Mason::Component::FileBased. If this is a problem because of your esoteric Mason configuration, don't panic. Just read the source. Seriously, it's tiny. This component class is meant to work well with L, which will let you throw a syntax exception if there's any significant content outside of blocks, and which can apply C to calls found when compiling. You can pass a C argument to the constructor for this class, but it's not all that easy to get where you need it, so maybe you should stick with the default: C =head1 AUTHOR Ricardo Signes =head1 COPYRIGHT AND LICENSE This software is copyright (c) 2012 by Ricardo Signes. This is free software; you can redistribute it and/or modify it under the same terms as the Perl 5 programming language system itself. =cut
__label__pos
0.983965
Network and IoT security in a zero trust security model You can never be too careful when it comes to network and IoT security. With a rapidly growing number of disparate devices being connected to corporate and industrial infrastructures, it’s better to be safe than sorry. network IoT security For network administrators it is no longer only about protecting laptops and PCs, but rather about managing a network comprised of a colorful palette of connected hardware including mobile and low-cost IoT devices. But how can you possibly keep your network secure when every device plays by its own rules? The answer is (relatively) simple: TRUST NO ONE! This is where the concept of a “zero trust architecture” comes in, which is a security concept based on not trusting a device by default simply because it is part of your network. Instead, every device needs to authenticate itself for every connection it wants to establish. Given that any possible connection involves at least two parties, the authentication that is required here is called mutual authentication. There are different communication protocols that make use of mutual authentication, such as SSH and TLS. But what these protocols have in common is that the authentication is based on unique device certificates. Without such a certificate, a device cannot authenticate itself. How does a device get a certificate? It all starts with the device having its own unique public-private keypair. To generate a certificate, the first step is sharing the public key from this pair with a Certificate Authority (CA). The CA will verify that the public key belongs to this device, by sending it a challenge. Only the device that possesses the corresponding private key will be able to successfully respond to this challenge. Now the CA knows that the public key belongs to the device, and it will create a certificate for it. Once created, the certificate can be sent to the device, which is now able to use it in future authentication protocols on networks that consider the specific CA that has created the certificate a trusted source. This is something that makes the term “zero trust” a bit misleading. Even when you don’t trust the devices, there is something that you need to trust. In this case, the trust is based on the certificates and on the authority that provides them. But there is another aspect that is important to consider: the private key. The private key is the foundation on which all the security is built. It’s what ties the certificate to the device because anyone who wants to check the authenticity of the certificate can do this by challenging the private key. And because this private key is so important, it should always be stored securely inside the device. An attacker should never be able to read, alter or copy this private key, as that compromises the security of the entire network that the device is connected to. Keeping the private key private should be any device’s highest priority. And the network needs to trust (there’s that word again…) the device to be able to do so. How is the private key stored securely on the device? There are a few ways to do this, starting with the traditional use of secure hardware, like a Secure Element or a Trusted Platform Module. These are both secure chips that need to be added to the device and that take care of creating and securely storing keys. This is an acceptable solution for expensive devices such as phones and laptops but does not usually solve all security issues since limited parties have access to it. However, for low-cost IoT devices adding a secure chip to the bill of materials adds too much cost. A more affordable solution is storing the key pair in the memory of one of the chips the device already needs anyway, like a microcontroller. In this case the key pair can be externally provisioned during manufacturing, or it can be internally generated (if the chip has an internal random number generator). The major downside of this option is that chips for IoT devices are not designed for securely storing keys. This means that there is a serious risk of the private key being compromised by a determined attacker with access to the device. On top of this, when keys are injected from the outside, the party that injects those keys is another entity that needs to be trusted to keep the secrets secret. There is yet another an alternative to these traditional methods for generating and storing secret keys, and it’s based on physical unclonable functions (PUFs). PUFs use deep submicron variations in the manufacturing process of chips to create device unique identifiers. This means that PUFs can generate cryptographic keys (like the key pair we require) from the silicon of chips. These keys are unique for every chip, and they never have to be stored in memory – they are simply (re-)generated every time they are needed. This eliminates the need for externally provisioning keys as well as for using dedicated hardware for protecting stored keys. This is why the deployment of PUFs is gaining traction rapidly, especially for low-cost IoT devices. Using PUFs to create and protect the keys needed to generate device certificates provides the kind of trust that a zero-trust architecture requires. Conclusion Now that we have seen all different building blocks that are required to securely connect devices in a network, we can take a step back and see what we have learned. It all starts on the device level, where the foundation of a zero-trust architecture is established by selecting the right way to provision the device with the keys that are the basis for its unique certificate. The method of choice will vary depending on the hardware of individual devices. Different approaches provide different levels of security, but one thing that they all have in common is that they need to instill the appropriate level of trust that they will keep the private key private. When a device is equipped with a public-private key pair, a CA can provide the next piece of the puzzle by generating a certificate for the device. Once a device has this unique certificate, it is ready for the mutual authentication that is required to be allowed entrance in a secure way to a network built on a zero-trust architecture. Combining the assurance that is provided by the way cryptographic keys are stored with the trust that needs to be placed in the CA, it is safe to state that, at least in this case, there is no “zero trust” without trust. Share this
__label__pos
0.816484
Topics allow Varnish 5.1 always online Recently I posted about Varnish with secure AWS S3 bucket as backend and I wanted to have the “always online” enabled, meaning you can take the backend offline while serving from Varnish cache. Varnish with secure AWS S3 bucket as backend Serving static contents from S3 is common, but using Varnish in front is a bit tricky. Especially if you want to keep the bucket secure and only serve from Varnish, here is a simple Varnish file to solve this problem. First secure your bucket via IP policy: { "Version": "2012-10-17", "Id": "S3PolicyId1", "Statement": [ { "Sid": "IPAllow", "Effect": "Allow", "Principal": "*", "Action": "s3:GetObject", "Resource": "arn:aws:s3:::example.bucket/*", "Condition": { "IpAddress": { "aws:SourceIp": [ "5.6.7.8/32" //varnish ip ] } } }, { "Sid": "Explicit deny to ensure requests are allowed only from specific referer.", "Effect": "Deny", "Principal": "*", "Action": "s3:GetObject", "Resource": "arn:aws:s3:::example.bucket/*", "Condition": { "StringNotLike": { "aws:Referer": [ "https://example.com/*" ] } } } ] } Setup postgreSQL RDS using Ansible Setting up PostgreSQL on RDS using ansible is a bit tricky because the main user on RDS is not a SUPERUSER and roles membership is not automatically granted for ex: “ERROR: must be member of role ..” is quite common. Here is a working solution: Custom validation messages for sails js //in api/models/User.js function validationError(invalidAttributes, status, message) { var WLValidationError = require('../../node_modules/sails/node_modules/waterline/lib/waterline/error/WLValidationError.js'); return new WLValidationError({ invalidAttributes: invalidAttributes, status: status, message: message } ); } var User = { attributes: { //... }, ownValidate:: function (values, update, cb) { //example of not allowed param on update //if it is an update then do not allow email param if (update && values.email) { return cb(validationError({ email: [ { message: 'Email is not allowed for updates.' } ] }, 400 /*status*/)); } sails.models['user'].findOne(values.email).exec(function (err, user) { if (err) return cb(err); if (user) { return cb(validationError({ email: [ { value: values.email, rule: 'E_UNIQUE' /* unique validation message is left for the default one here */ } ] }, 409)); } }); }, beforeCreate: function (values, cb) { return sails.models['user'].ownValidate(values, false, cb); }, beforeUpdate: function (values, cb) { return sails.models['user'].ownValidate(values, true, cb); } } For blueprint custom messages validation Nginx Error Log Reader Nginx Error Log Reader is a php reader/parser/analyzer for Nginx error log file. the script is able to read error logs recursively then display them in a user friendly table. Script configuration includes the number of bytes to read per page and allow pagination through the error log . Additionally, table columns are sortable and full description of every error is displayed using MonnaTip. For banning Ips, please refer to this post Using iptables to block ips that spam or attack your server
__label__pos
0.997542
Generate a timestamp Common Date Formats Timestamp 2038957562 Atom 2034-08-12T01:06:02+00:00 W3C 2034-08-12T01:06:02+00:00 ISO 8601 2034-08-12T01:06:02+0000 RFC 2822 Sat, 12 Aug 2034 01:06:02 +0000 Future Dates + 1 Hour 2038961162 + 1 Day 2039043962 + 1 Month 2041635962 + 1 Year 2070493562 + 10 Years 2354576762 Past Dates - 1 Hour 2038953962 - 1 Day 2038871162 - 1 Month 2036279162 - 1 Year 2007421562 - 10 Years 1723424762 Why would I need to convert a date to a timestamp? Timestamps are commonly used in computer systems and databases to record and track events with a high degree of accuracy. A timestamp represents a specific point in time, typically measured as the number of seconds that have elapsed since a specific starting point (the "epoch"). Converting a date to a timestamp can be useful in various scenarios. For example, if you need to compare two events that occurred at different times, converting their respective dates to timestamps can help you determine which event occurred first. Timestamps can also be used to generate unique identifiers or to sort and filter data in a database based on time-related criteria. Additionally, timestamps are commonly used in programming and web development to record the time a specific action occurred or to track how long a process took to execute. Overall, converting a date to a timestamp is a useful tool that can help you accurately record and track events over time, making it an essential tool for developers, data analysts, and anyone who needs to work with time-related data.
__label__pos
0.909234
4. Find the square root of each of the following (1) 13.69 (11) 0.002025 (v) 6146.56 (vi) 1.024144 (iii) 1.5129 Question 4. Find the square root of each of the following (1) 13.69 (11) 0.002025 (v) 6146.56 (vi) 1.024144 (iii) 1.5129 (iv) 20.7936​ in progress 0 Kennedy 2 months 2021-11-10T20:59:14+00:00 1 Answer 0 views 0 Answers ( ) 0 2021-11-10T21:00:18+00:00 Answer: 1. 3.7 2. 0.045 3. 78.4 4. 1.012 5. 1.23 6. 4.56 this is the square root of the given squares Leave an answer Browse 14:4+1-6*5-7*14:3+5 = ? ( )
__label__pos
0.911783
Information Security Stack Exchange is a question and answer site for information security professionals. Join them; it only takes a minute: Sign up Here's how it works: 1. Anybody can ask a question 2. Anybody can answer 3. The best answers are voted up and rise to the top I'm tasked with creating database tables in Oracle which contain encrypted strings (i.e., the columns are RAW). The strings are encrypted by the application (using AES, 128-bit key) and stored in Oracle, then later retrieved from Oracle and decrypted (i.e., Oracle itself never sees the unencrypted strings). I've come across this one column that will be one of two strings. I'm worried that someone will notice and presumably figure out what those two values to figure out the AES key. For example, if someone sees that the column is either Ciphertext #1 or #2: • Ciphertext #1: BF,4F,8B,FE, 60,D8,33,56, 1B,F2,35,72, 49,20,DE,C6. • Ciphertext #2: BC,E8,54,BD, F4,B3,36,3B, DD,70,76,45, 29,28,50,07. and knows the corresponding Plaintexts: • Plaintext #1 ("Detroit"): 44,00,65,00, 74,00,72,00, 6F,00,69,00, 74,00,00,00. • Plaintext #2 ("Chicago"): 43,00,68,00, 69,00,63,00, 61,00,67,00, 6F,00,00,00. can he deduce that the encryption key is "Buffalo"? 42,00,75,00, 66,00,66,00, 61,00,6C,00, 6F,00,00,00. I'm thinking that there should be only one 128-bit key that could convert Plaintext #1 to Ciphertext #1. Does this mean I should go to a 192-bit or 256-bit key instead, or find some other solution? (As an aside, here are two other ciphertexts for the same plaintexts but with a different key.) • Ciphertext #1 A ("Detroit"): E4,28,29,E3, 6E,C2,64,FA, A1,F4,F4,96, FC,18,4A,C5. • Ciphertext #2 A ("Chicago"): EA,87,30,F0, AC,44,5D,ED, FD,EB,A8,79, 83,59,53,B7. [Related question: When using AES and CBC, can the IV be a hash of the plaintext?] share|improve this question migrated from programmers.stackexchange.com Jul 16 '11 at 17:35 This question came from our site for professional programmers interested in conceptual questions about software development.      What mode are you using? If a stream mode such as CTR, you could get yourself into a situation where the key is not recoverable, but the plain text is recoverable. – jww Aug 23 '13 at 21:53 I am adding an answer as a community wiki because I believe that the accepted answer is dangerously misleading. Here's my reasoning: The question is asking about being able to derive the AES keys. In that regard the accepted answer is correct: that is called a Known-plaintext Attack, and AES is resistant to that kind of attack. So an attacker will not be able to leverage this to derive the key and make off with the whole database. But there is another, potentially dangerous attack at play here: a Ciphertext Indistinguishablity Attack. From Wikipedia: Ciphertext indistinguishability is a property of many encryption schemes. Intuitively, if a cryptosystem possesses the property of indistinguishability, then an adversary will be unable to distinguish pairs of ciphertexts based on the message they encrypt. The OP showed us that this column holds one of two possible values, and since the encryption is deterministic (ie does not use a random IV), and attacker can see which rows have the same value as each other. All the attacker has to do is figure out the plaintext for that column for a single row, and they've cracked the encryption on the entire column. Bad news if you want that data to stay private - which I'm assuming is why you encrypted it in the first place. Mitigation: To protect against this, make your encryption non-deterministic (or at least appear non-deterministic to the attacker) so that repeated encryptions of the same plaintext yields different cipher texts. You can for example do this by using AES in Cipher Block Chaining (CBC) mode with a random Initialization Vector (IV). Use a secure random number generator to generate a new IV for each row and store the IV in the table. This way, without the key, the attacker can not tell which rows have matching plaintext. share|improve this answer      My own answer, currently accepted, is a tad old and I wasn't aware of Ciphertext Indistinguishablity Attack at the time. I'd delete my own but it looks like I can't delete an accepted answer. – Vitor Py Jun 4 at 12:04      @VitorPy Wow. I respect you for that. – Mike Ounsworth Jun 4 at 12:04 1   @VitorPy You can probably flag it for moderator attention to have it un-accepted (in which case, leave it up for historical reasons). – Mike Ounsworth Jun 4 at 12:05 2   @VitorPy actually moderators have no ability to unaccept an answer (and don't we often wish we did!) You can comment on the question to the OP, and get him to swap it. That said, I don't think your answer is wrong per se (the key is still unretrievable, which is what was asked), it is just incomplete wrt what Mike is raising here. – AviD Jun 4 at 19:58 For a block cipher with a n-bit key, if, given a plaintext block and the corresponding ciphertext, the key can be guessed in less than 2n-1 step on average, then that block cipher will be said to be "broken" and cryptographers will make a point of not using it. The AES is not broken (yet). So no worry. A few things may still be said, though: • Having a plaintext and the corresponding ciphertext allows an attacker to verify a potential key value. • The 2n-1 value is actually half the size of the key space. The idea is that the attacker can try all possible keys, until one matches. On average, he will have to try half the keys before hitting the right one. This assumes that the key space has size 2n. You still have the possibility to reduce the key space: e.g., if you decide that your key is the name of a US town, then the number of possible keys is very much lower (there must not be more than 100000 towns in the USA). Hence, you get the promised 128-bit security only if your key generation process may indeed produce any 128-bit key. • You apparently encrypt each value by stuffing it directly into an AES core. AES being deterministic, this means that two cells with the same value will yield the same encrypted block, and any attacker may notice that. In other words, you leak information about which cells are equal to each other. Depending on your situation, this may or may not be an issue; you should be aware of it. • You do not say how you handle values longer than 16 bytes. This is not a simple issue. In all generality, this requires a chaining mode such as CBC, and an Initialization Vector (it depends on the mode; for CBC, a 16-byte random value -- a new IV for each encrypted value)(this can also fix the information leakage from the previous point). share|improve this answer 1   According to your definition, AES is broken, since the computational complexity has been reduced by ~3 bits: schneier.com/blog/archives/2011/08/new_attack_on_a_1.html – deed02392 Mar 7 '13 at 9:58 The answer: No, the AES key cannot be recovered in this scenario. AES is secure against known-plaintext attack. This means that, even if an attacker knows the plaintext and its corresponding ciphertext (its encryption under some unknown AES key), then the attacker cannot recover the AES key. In particular, the attacker cannot recover the AES key any faster than simply trying possible keys one after another -- which is a process that will take longer than the lifetime of our civilization, assuming that the AES key is chosen randomly. P.S. I noticed that whatever you are using for encryption does not seem to use an IV. This is a security risk. I don't know what mode of operation you are using, but you should use a well-regarded mode of encryption (e.g., CBC mode encryption, CTR mode encryption) with a random IV. The fact that encrypting the same message multiple times always gives you the same ciphertext every time is a security leak which is better to avoid. You can avoid this leak by using a standard mode of operation with an appropriate IV. (You probably also should use a message authentication code (MAC) to authenticate the ciphertext and prevent modifications to it.) share|improve this answer Salt your encryption. That way there wont be any patters in your encryption. (There are other benefits too!) http://stackoverflow.com/questions/5051007/what-is-the-purpose-of-salt share|improve this answer 7   For encryption, the usual term is not Salt, but Initialization Vector or IV. en.wikipedia.org/wiki/Initialization_vector en.wikipedia.org/wiki/Block_cipher_modes_of_operation – Secure Mar 11 '11 at 9:18 AES is not as easy as just building a rainbow table. The first thing you have to realize is the table requires an initialization vector. As long as you're changing this on a semi regular basis building a rainbow table (which is not really realistic.) would take a very very long time. Orders of magnitude long. Since a typical rainbow table would be essentially 2 dimensions, you would essentially need a cube of result sets to figure out both the IV and key. If you read Thomas Pornin's post he goes into pretty great detail as to what this means, in terms of brute forcing the result. The realistic worry is that someone with access to the database would be able to inject a string from another field (presumably because you're not using a random padding value in the column per element. Or seeding. ) If you seed the value you won't have this issue, and the only (realistic) attack on the cipher-text itself is made much more difficult. share|improve this answer Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.761337
Java Spring Part 2: All about Bean This is not a cooking class. Like cooking a soup process, I will show you the bean containers, scope, lifecycles, postprocessor, inheritance, and template. You will not become a chef but you should have a simple idea of how beans work in the Spring Framework. 🍳1. Cooking pot: IoC Container From part 1, you shall have seen how Spring blends Plain Old Java Object (POJO) and XML bean together. The cookware is called Spring container. The cooking method is called Inversion of Control principle, init as IoC. In the last part, we have seen one of IoC pot, ApplicationContext. The other IoC pot is BeanFactory. The AppContext is a big pot. It contains BeanFactory to deal with beans ( Spring’s AOP); handles server messages (I18N); publishes events to its listeners, and works with web applications. Let’s declare a BeanFactory in Java: Resource resource = new ClassPathResource("soup.xml"); BeanFactory factory = new XmlBeanFactory(resource); The resource instance injects into XmlBeanFactory to create the instance of BeanFactory. And take a look of ApplicationContext of part 1: ApplicationContext context = new ClassPathXmlApplicationContext("com/homanspring/Beans.xml"); It’s much simpler. We can create a new instance of AppContext with the name of XML file. 🍇2. Bean and Scope Now, let’s study the worker in the factory, Bean. <!-- simple --> <bean id = "..." class = "..."> <!-- lazy init --> <bean id = "..." class = "..." lazy-init = "true"> <!-- init method --> <bean id = "..." class = "..." init-method = "..."> <!-- destruction method --> <bean id = "..." class = "..." destroy-method = "..."> The beans can serve in a defined kitchen, called scope. There five scopes: • Singleton => IoC container • Prototype => I can handle many objects. • Request => Http • Session => Http • Global-session => Global Http You see, the Spring Framework heavily serves on the Web application. Let’s attach a scope to a bean: <!-- singleton scope --> <bean id = "..." class = "..." scope = "singleton"> 🍝3. Test Scope Back to the SpringExample project; in this time, we need to research how scope works in the code. Singleton Beans.xml, without a default value. ... <bean id = "hello" class = "com.homanspring.Hello" scope="singleton"> </bean> SpringApp.java: Let’s get the chicken and display her feather. public static void main(String[] args) { ApplicationContext context = new ClassPathXmlApplicationContext( "com/homanspring/Beans.xml"); Hello helloA = (Hello) context.getBean("hello"); helloA.setChicken("Yellow Chicken!"); helloA.getChicken(); Hello helloB = (Hello) context.getBean("hello"); helloB.getChicken(); } Run, Message: Yellow Chicken! Message: Yellow Chicken! The singleton stores the data in a cache. Every POJO call for the message will come from the same cache. So singleton scope is the same setting for all users. Prototype Beans.xml, without a default value. ... <bean id = "hello" class = "com.homanspring.Hello" scope="prototype"> </bean> Run, Message: Yellow Chicken! Message: null Under the prototype scope, IoC creates a new bean instance for a specific object. So the prototype is one to one relationships. 💱4. Bean Lifecycle Initialization In XML, <bean id = "whiteBean" class = "examples.FishBean" init-method = "init"/> it’s call init() in FishBean class. public class FishBean { public void init() { // do some initialization work } } Destruction In XML, <bean id = "greenBean" class = "examples.FishBean" destroy-method = "destroy"/> It calls destroy() in FishBean class. public class FishBean { public void destroy() { // do some destruction work } } 📑5. Test Lifecycle Back to the SpringExample project, let’s add two lifecycle callings in Hello.java. public class Hello { ... public void init() { System.out.println("Initialize cooking. Prepare the chicken."); } public void destroy() { System.out.println("Chicken is destroyed."); } } Beans.xml, ... <bean id = "hello" class = "com.homanspring.Hello" init-method = "init" destroy-method = "destroy"> <property name = "chicken" value = "Chicken is cooking."/> </bean> SpringApp.java, public static void main(String[] args) { AbstractApplicationContext context = new ClassPathXmlApplicationContext( "com/homanspring/Beans.xml"); Hello mHello = (Hello) context.getBean("hello"); mHello.getChicken(); context.registerShutdownHook(); } The registerShutdownHook() of AbstractApplicationContext will shutdown and call destroy(). Run, Initialize cooking. Prepare the chicken. Message: Chicken is cooking. Chicken is destroyed. It’s successful to call init() and destroy() methods. You don’t need to set them for all of the beans with the same name. A default value can be set in the XML file. <beans xmlns = ... default-init-method = "init" default-destroy-method = "destroy"> ⏳6. BeanPostProcessor BeanPostProcessor can give you extra options at the init processing. Let’s add a new class, MyInitHello.java. public class MyInitHello implements BeanPostProcessor {} public class MyInitHello implements BeanPostProcessor { @Override public Object postProcessAfterInitialization(Object bean, String beanName) throws BeansException { System.out.println("After Initialization: chicken is cleaned <= "+beanName+" bean"); return BeanPostProcessor.super.postProcessAfterInitialization(bean, beanName); } @Override public Object postProcessBeforeInitialization(Object bean, String beanName) throws BeansException { System.out.println("Before Initialization: shop for the chicken <= "+beanName+" bean"); return BeanPostProcessor.super.postProcessBeforeInitialization(bean, beanName); } } I add two system printouts. Let’s call this class at XML. Beans.xml, <bean id = "hello" class = "com.homanspring.Hello" init-method = "init" destroy-method = "destroy"> <property name = "chicken" value = "Chicken is cooking."/> </bean> <bean class = "com.homanspring.MyInitHello" /> Run, Before Initialization: shop for the chicken <= hello bean Initialize cooking. Prepare the chicken. After Initialization: chicken is cleaned <= hello bean Message: Chicken is cooking. Chicken is destroyed. 🎎7. Bean Inheritance Let’s test the inheritance situation of two beans. This time, I want to compare two menus. Beans.xml, secondMenu inherits to the firstMenu. <?xml version="1.0" encoding="UTF-8"?> <beans xmlns="http://www.springframework.org/schema/beans" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.springframework.org/schema/beans https://www.springframework.org/schema/beans/spring-beans-4.3.xsd"> <bean id = "firstMenu" class = "com.homanspring.MenuFirst" > <property name = "chicken" value = "Chicken Noodle."/> <property name = "beef" value = "Beef Pho."/> </bean> <bean id ="secondMenu" class = "com.homanspring.MenuSecond" parent = "firstMenu"> <property name = "chicken" value = "Chicken Egg Soup."/> <property name = "pork" value = "BBQ Pork"/> </bean> </beans> MenuFirst.java has chicken and beef on the menu. public class MenuFirst { private String chicken; private String beef; public void getChicken() { System.out.println("Selected chicken: "+chicken); } public void setChicken(String chicken) { this.chicken = chicken; } public void getBeef() { System.out.println("Selected beef: "+beef); } public void setBeef(String beef) { this.beef = beef; } } MenuSecond.java, public class MenuSecond { private String pork; private String chicken; private String beef; public void getChicken() { System.out.println("2nd-Menu Chicken: "+chicken); } public void setChicken(String chicken) { this.chicken = chicken; } public void getBeef() { System.out.println("2nd-Menu Beef: "+beef); } public void setBeef(String beef) { this.beef = beef; } public void getPork() { System.out.println("2nd-Menu Pork: "+pork); } public void setPork(String pork) { this.pork = pork; } } The second menu has added pork. SpringApp.java, public class SpringApp { public static void main(String[] args) { ApplicationContext context = new ClassPathXmlApplicationContext( "com/homanspring/Beans.xml"); MenuFirst firstMenu = (MenuFirst) context.getBean("firstMenu"); firstMenu.getChicken(); firstMenu.getBeef(); MenuSecond secondMenu = (MenuSecond) context.getBean("secondMenu"); secondMenu.getChicken(); secondMenu.getBeef(); secondMenu.getPork(); } } Run, Menu1 Chicken: Chicken Noodle. Menu1 Beef: Beef Pho. 2nd-Menu Chicken: Chicken Egg Soup. 2nd-Menu Beef: Beef Pho. 2nd-Menu Pork: BBQ Pork The bean of secondMenu doesn’t define beef value, so the beef inherits the value from the bean of firstMenu. 📰8. Bean Template It’s easy to write a template for a bean. <bean id = "templateMenu" abstract = "true" > <property name = "chicken" value = "Chicken Noodle."/> <property name = "beef" value = "Beef Pho."/> <property name = "pork" value = "BBQ Pork"/> </bean> <bean id ="secondMenu" class = "com.homanspring.MenuSecond" parent = "templateMenu"> <property name = "chicken" value = "Chicken Egg Soup."/> <property name = "pork" value = "Pork Onion Soup"/> </bean> Characteristic of template: no class + abstract + property (value). SpringApp.java, public static void main(String[] args) { ApplicationContext context = new ClassPathXmlApplicationContext( "com/homanspring/Beans.xml"); MenuSecond secondMenu = (MenuSecond) context.getBean("secondMenu"); secondMenu.getChicken(); secondMenu.getBeef(); secondMenu.getPork(); } MenuSecond.java has no change. Let’s run, 2nd-Menu Chicken: Chicken Egg Soup. 2nd-Menu Beef: Beef Pho. 2nd-Menu Pork: Pork Onion Soup The secondMenu has no beef on the menu, so it inherits the value from the templateMenu. 🎴9. Inner Bean Like a Java class, the Bean can have inner Bean. <bean id = "outerBean" class = "..."> <property name = "target"> <bean class = "..."/> </property> </bean> Here is the example between Menu and SoupMaker. SoupMaker.java, this is the class of an inner bean. public class SoupMaker { String meat; String vegetable; public void getVegetable() { System.out.print("Vagetable: "+vegetable); } public void setVegetable(String vagetable) { this.vegetable = vagetable; } public SoupMaker() { System.out.println("SoupMaker constructor..."); } public void getMeat() { System.out.println("Soup kind: "+meat); } public void setMeat(String food) { this.meat = food; } public void checkSoup() { System.out.println("Check soup: still cooking..."); } public void detail() { System.out.println("Soup contains: "+meat+" and "+vegetable); } } Like regular POJO, you need setter and getter for each bean related property. In addition, you can add extra functions like checkSoup() and detail(). Let’s take a look at the Beans.xml. <bean id ="secondMenu" class = "com.homanspring.MenuSecond"> <property name = "makeSoup"> <bean class = "com.homanspring.SoupMaker"> <property name="meat" value="Chicken"></property> <property name="vegetable" value="Tomato"></property> </bean> </property> </bean> The makeSoup is the variable in MenuSecond.java. The meat and the vegetable are the variables of SoupMaker.java. MenuSecond.java, public class MenuSecond { private SoupMaker makeSoup; public void setMakeSoup(SoupMaker makeSoup) { this.makeSoup = makeSoup; } public SoupMaker getMakeSoup() { return makeSoup; } public void soupChecker() { makeSoup.checkSoup(); } public void soupDetail() { makeSoup.detail(); } } SpringApp.java, public static void main(String[] args) { ApplicationContext context = new ClassPathXmlApplicationContext( "com/homanspring/Beans.xml"); MenuSecond secondMenu = (MenuSecond) context.getBean("secondMenu"); secondMenu.soupChecker(); secondMenu.soupDetail(); } Run, SoupMaker constructor... Check soup: still cooking... Soup contains: Chicken and Tomato Enjoy the reading! 08/15/2020 Computer Science BS from SFSU. I studied and worked on Android system since 2017. If you are interesting in my past works, please go to my LinkedIn. Get the Medium app A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
__label__pos
0.961027
How to Generate Backlinks to Your Website by JC Burrows  - December 15, 2022 Do you ever wonder how to generate backlinks to your website? Well, fret not! I’m here to share some nifty tips with you. In this guide, we’ll explore a range of strategies that can help you build valuable backlinks and enhance your website’s visibility. From guest blogging to promoting your content on social media, reaching out to influencers, and engaging in content marketing, we’ve got you covered. We’ll also delve into link building strategies, directory submissions, and even the clever technique of broken link building. So, if you’re eager to elevate your website and boost your online presence, let’s dive right into the exciting world of generating those valuable backlinks! Guest Blogging Guest blogging is an effective way to generate backlinks to our website. It involves writing and publishing articles on other websites or blogs within our industry or niche. By doing this, we can reach a wider audience and gain exposure to potential readers and customers who might not have found us otherwise. To make the most of guest blogging, it’s important to find reputable websites and blogs with a strong following and authority in our industry. By contributing valuable and high-quality content to these platforms, we can establish ourselves as experts and gain credibility among our target audience. This not only helps generate backlinks but also boosts our brand visibility and reputation. Another benefit of guest blogging is the opportunity for niche-specific partnerships. By collaborating with other experts and influencers in our industry, we can leverage their audience and reach to promote our content and website. This can result in increased traffic, engagement, and ultimately, more backlinks. Social Media Promotion When it comes to promoting your website on social media, there are three important factors to consider: influencer collaborations, viral content strategies, and paid social advertising. Let’s dive into each of these to understand how they can benefit your website. Firstly, influencer collaborations are a powerful way to leverage the audience and credibility of individuals in your industry. By partnering with influencers, you can gain valuable backlinks that can improve your website’s visibility and authority. Secondly, creating viral content that resonates with your target audience is a key strategy to attract natural backlinks. When your content goes viral, it has the potential to spread rapidly and generate buzz, increasing the chances of other websites linking back to it. Lastly, utilizing paid social advertising can help you reach a wider audience and drive traffic to your website. By investing in targeted social media ads, you can increase the visibility of your website and attract potential backlinks from interested users. Influencer Collaborations To increase the visibility of your website, we actively collaborate with influencers for social media promotion. By partnering with influential individuals in your industry, you can tap into their large and engaged audience, gaining exposure and credibility for your brand. Influencer endorsements have the power to significantly boost your website traffic and attract valuable backlinks. When influencers share your content or mention your brand on social media, their followers are more likely to visit your website, resulting in increased organic traffic and potential customer conversions. These partnerships can also lead to valuable backlinks from the influencer’s website, which can further improve your website’s search engine rankings. Now, let’s move on to the next section and explore effective strategies for creating viral content to drive even more traffic to your website. Viral Content Strategies When it comes to making content go viral, we harness the power of social media to amplify our website’s reach and engagement. One effective method is using viral videos. These videos are specifically crafted to grab viewers’ attention and motivate them to share it with their own networks. By creating captivating and shareable videos, we increase our chances of going viral and attracting a large audience to our website. Another strategy we utilize is meme marketing. Memes are amusing or relatable images or videos that quickly spread across social media platforms. By crafting and sharing relevant memes that strike a chord with our target audience, we can create a buzz and drive traffic to our website. Paid Social Advertising We actively use paid social advertising to promote our website and generate backlinks. This strategy helps us reach a specific audience and drive traffic to our site, increasing our chances of getting valuable backlinks. Paid social advertising provides several advantages, one of which is ad targeting. This feature allows us to reach people who have specific demographics, interests, and behaviors that are relevant to our website’s niche. By reaching the right people, we can make our ad campaigns more effective and increase the likelihood of attracting quality backlinks. Another benefit is conversion tracking, which helps us measure the success of our paid social advertising efforts. This allows us to identify the strategies that generate the most backlinks. Now, let’s explore another effective method for generating backlinks: influencer outreach. Influencer Outreach When it comes to reaching out to influencers, we have a specific strategy in place. We focus on connecting with individuals who have a strong influence within our niche market. These are what we call micro influencers, who have a dedicated following in our industry. By partnering with them, we can effectively reach a targeted audience that is more likely to engage with our content and generate backlinks to our website. Why do we prioritize micro influencers? Well, it’s because they play a crucial role in our backlink generation strategy. These influencers have already built trust with their followers, who are genuinely interested in our products or services. When these influencers endorse our brand or share our content, it carries a lot of weight and can drive significant traffic to our website. To find the right influencers for our outreach campaign, we use tools that help us analyze their audience demographics, engagement rates, and content relevance. This ensures that we are connecting with individuals who align with our brand values and have a genuine connection with our target market. Once we have identified our target influencers, we reach out to them with personalized messages. We explain why we believe a partnership would be mutually beneficial and offer them incentives, such as exclusive content, discounts, or product samples. These incentives encourage them to share our content and include backlinks to our website. Content Marketing When it comes to generating backlinks to your website, content marketing plays a crucial role. By implementing strategies that make your content intriguing and unpredictable, you can captivate your intended audience and encourage them to share your content. This, in turn, leads to valuable backlinks. The benefits of this approach are twofold: it boosts your brand’s visibility and credibility, while also establishing your website as a trusted and authoritative source in your industry. Engaging Content Strategies To generate backlinks to your website effectively, implement engaging content strategies that captivate your target audience. One crucial aspect of engaging content is content optimization, which involves creating high-quality, relevant content optimized for search engines. Start by conducting thorough keyword research to identify the keywords and phrases your audience uses to search for industry-related information. Strategically incorporate these keywords into your content to increase the likelihood of your website appearing in search engine results, attracting organic traffic and potential backlinks. Moreover, engaging content strategies require creating valuable and shareable content that resonates with your audience, encouraging them to link back to your website and share your content. Consistently produce engaging content that provides value to your audience to enhance your chances of generating backlinks and improving your website’s visibility and authority in search engine rankings. Targeted Audience Engagement To effectively engage our target audience and generate valuable backlinks to our website, we can employ strategic content marketing techniques. One actionable tip is to utilize targeted email outreach. By sending personalized emails to individuals who are likely to have an interest in our content, we can capture their attention and encourage them to visit our website. Additionally, participating in community forums is another effective strategy. By actively engaging in relevant discussions within these forums, we can establish ourselves as knowledgeable industry experts and form connections with our intended audience. This can potentially lead to valuable backlinks as forum members may reference our content in their own discussions or share our links with others. By implementing these targeted audience engagement strategies, we can enhance our website’s visibility and attract high-quality backlinks. Brand Visibility and Credibility To boost your brand’s visibility and credibility in content marketing, focus on creating top-notch and informative content. This will not only position you as an industry authority but also attract and engage your target audience. However, it’s crucial to recognize that brand visibility and credibility extend beyond content creation. Managing your online reputation is key in building trust and credibility. By actively monitoring and handling your online reputation, you can ensure that your brand is viewed positively by your audience. Additionally, implementing trust-building tactics like customer testimonials, case studies, and social proof can further enhance your brand’s credibility. By consistently providing valuable content and proactively managing your online reputation, you can establish yourself as a trusted and credible expert in your field. Link Building Strategies When it comes to building links to our website, we have a preferred approach that involves using a variety of effective strategies. Two key strategies we focus on are link acquisition and natural link building. Link acquisition means actively seeking out opportunities to get backlinks from other websites. We can do this by reaching out to relevant websites and suggesting collaborations or guest posting. By providing valuable and relevant content, we can attract links from trusted sources, which helps boost the visibility and credibility of our website. In addition to link acquisition, we also prioritize natural link building. This means creating high-quality content that naturally attracts backlinks from other websites. By consistently producing valuable and shareable content, we increase the chances of other websites linking to our content organically. By combining these two strategies, we can build a diverse and strong backlink profile for our website. This not only improves our search engine rankings but also drives targeted traffic to our site. Now that we’ve discussed our link building strategies, let’s move on to the next section: directory submissions. Directory Submissions When it comes to expanding our backlink profile, we can use a strategy called directory submissions. This involves listing our website in online directories that categorize and organize websites based on their niche or location. It’s an effective way to gain backlinks and improve our website’s visibility in search engine results. There are two types of directories we can focus on: local directories and niche directories. Local directories list businesses and websites in a specific geographic area. By submitting our website to these directories, we increase our chances of being found by users in our target location. This is particularly important for businesses that rely on local customers. Niche directories, on the other hand, focus on specific industries or topics. By submitting our website to relevant niche directories, we not only gain backlinks but also reach a more targeted audience. This can result in increased traffic and potential customers who are already interested in what we have to offer. Broken Link Building Let’s dive into a technique called broken link building to further enhance our backlink profile and boost our website’s visibility. Broken link building involves finding broken links on other websites and reaching out to the website owners to suggest replacing the broken link with a link to our own content. This approach not only helps us build backlinks but also adds value to other website owners by assisting them in fixing broken links. To get started with broken link building, we need to analyze our competitors to identify websites in our niche that have broken links. By examining the backlinks of our competitors, we can uncover opportunities for broken link building. Once we identify these websites, we can utilize various tools to check for broken links on their pages. After identifying broken links, we can reach out to the website owners through content outreach. Our aim is to inform them about the broken link on their website and suggest our content as a replacement. It’s crucial to approach them in a polite and helpful manner, highlighting the benefits of fixing the broken link for their website’s user experience. Broken link building is a powerful strategy that not only helps us gain backlinks but also fosters relationships with other website owners. By leveraging competitor analysis and content outreach, we can effectively improve our backlink profile and enhance our website’s visibility in search engine results. Frequently Asked Questions What Are the Benefits of Guest Blogging for Generating Backlinks to My Website? Guest blogging is a powerful strategy for generating backlinks to your website. It may sound perplexing at first, but it’s actually quite simple. When you write valuable content and share it on other websites, you create an opportunity to reach a wider audience. This burst of exposure can significantly increase your online visibility, making it easier for people to find and visit your website. But the benefits don’t stop there. Guest blogging also helps you establish credibility in your industry. By showcasing your expertise and knowledge, you can position yourself as a trusted authority. This can lead to more opportunities for collaboration, partnerships, and even potential customers or clients. To make the most of guest blogging, it’s important to choose websites that are relevant to your niche and have a solid reputation. Focus on providing valuable and actionable insights in your guest posts, rather than simply promoting your own website. This will not only attract more readers but also increase the likelihood of other websites linking back to your content. How Can I Effectively Promote My Website Through Social Media to Generate Backlinks? To effectively promote your website through social media and generate backlinks, there are a few strategies you can employ. One approach is to create captivating content that sparks curiosity and encourages sharing. This could be in the form of informative articles, engaging videos, or eye-catching visuals that pique people’s interest. By providing valuable and shareable content, you increase the likelihood of others linking back to your website. Collaborating with influencers is another effective tactic. These individuals have a strong online presence and a dedicated following. By partnering with influencers in your niche or industry, you can tap into their audience and gain exposure for your website. This can be done through guest blogging, co-creating content, or having influencers share your website link with their followers. Active participation in relevant online communities is also crucial. Find social media groups, forums, or discussions where your target audience congregates and engage with them. Share your expertise, answer questions, and provide valuable insights. By establishing yourself as a helpful and knowledgeable contributor, you increase the likelihood of others checking out your website and potentially linking back to it. What Is Influencer Outreach and How Can It Help in Generating Backlinks? Influencer outreach is a strategic approach where you collaborate with influential individuals and bloggers to form partnerships that generate backlinks. This process can do wonders for your website by boosting its visibility and credibility. But what exactly is influencer outreach and how can it help you generate backlinks? Well, influencer outreach is all about building connections with people who have a significant online presence and a loyal following. These individuals, often known as influencers, have established themselves as experts or authorities in specific fields or industries. By partnering with them, you can tap into their audience and leverage their influence to promote your website. Now, you might be wondering how this collaboration leads to backlinks. When you collaborate with influencers, they can create content that features your website or products and share it with their audience. This content can take the form of blog posts, social media mentions, or even videos. By including links to your website within this content, you can generate valuable backlinks that direct traffic to your site. But why are backlinks so important? Well, search engines like Google use backlinks as a measure of a website’s credibility and authority. When your website has a high number of quality backlinks from reputable sources, it signals to search engines that your site is trustworthy and relevant. As a result, your website may rank higher in search engine results pages, ultimately leading to increased visibility and organic traffic. So, how can you harness the power of influencer outreach to generate backlinks? Here are some actionable tips: 1. Identify relevant influencers: Look for individuals who have a strong following in your industry or niche. They should align with your brand values and have an engaged audience that would be interested in your offerings. 2. Build genuine relationships: Reach out to influencers and start building authentic connections. Engage with their content, share their posts, and establish a rapport before proposing any collaborations. 3. Offer value: When approaching influencers, offer something of value in exchange for their support. This could be exclusive content, product samples, or even monetary compensation. Make sure to tailor your offer to their specific needs and interests. 4. Collaborate on content: Work together with influencers to create compelling and relevant content that showcases your website or products. This could be in the form of guest blog posts, sponsored social media posts, or collaborative videos. 5. Track and measure results: Monitor the performance of your influencer outreach campaigns. Keep track of the backlinks generated, website traffic, and any changes in search engine rankings. This data will help you refine your approach and optimize future collaborations. How Does Content Marketing Contribute to Generating Backlinks to My Website? Content marketing is a powerful strategy for generating backlinks to your website. When we create content that is valuable and of high quality, we attract other websites to link to us. This helps to boost our online visibility and establish our authority in the digital realm. It’s a proven technique for building links that can greatly benefit your website. Are There Any Specific Strategies or Techniques for Building High-Quality Backlinks to My Website? There are numerous techniques and strategies to build high-quality backlinks for your website. These methods can significantly enhance your website’s visibility and ranking on search engines. Let’s dive into specific strategies and techniques you can utilize to generate valuable backlinks. Conclusion Generating backlinks to your website can be a puzzling endeavor, but fear not! I’m here to guide you through the process with straightforward strategies that anyone can implement. No need for fancy jargon or complex techniques – I’ll break it down for you in simple terms. First and foremost, it’s important to understand the concept of backlinks. These are links from other websites that point to your own. They serve as a vote of confidence, indicating to search engines that your website is trustworthy and relevant. So, how can you go about getting these valuable backlinks? One effective approach is to create high-quality content that others naturally want to link to. This could be informative articles, engaging videos, or helpful resources. By offering something unique and valuable, you increase the chances of others linking back to your site. Another strategy is to reach out to relevant websites and ask for a backlink. This can be done by finding websites in your niche or industry and sending a polite email explaining why your content would be a valuable addition to their site. Remember, personalization and building relationships are key here. Additionally, participating in online communities and forums related to your industry can be a fruitful way to gain backlinks. By actively engaging in discussions and providing useful insights, you establish yourself as an authority in the field. This increases the likelihood of others linking to your website. Lastly, don’t underestimate the power of social media. Sharing your content on platforms like Twitter, Facebook, and LinkedIn can attract attention and potentially lead to backlinks from interested parties. Remember, the key to successful backlink generation lies in offering valuable content, building relationships, and promoting your website strategically. So, get started today and watch as your website gains visibility and traffic. Good luck! Backlinks Purchase: How to Buy High Quality Backlinks {"email":"Email address invalid","url":"Website address invalid","required":"Required field missing"} You may be interested in What Our Clients Say Absolutely thrilled with our results! These guys have been a game-changer for our online presence. Within just a few months, we've climbed up the Google ranks and the traffic's booming. Definitely more bang for my buck with the uptick in sales. Big shoutout to the Rank Higher crew – you rock! 🚀🌟 Jake Davidson Service Pros Online I've been working with this company to revamp our website, and wow, what a transformation! But the cherry on top? The SEO magic they've worked. We're ranking higher than ever, and I'm seeing a real boost in traffic and sales. Hats off to the team for their hard work and genius touch! If you're looking to spruce up your site and get seen, these are the go-to pros. Lacey Roberts Deals Direct Daily
__label__pos
0.819319
Chapter: Relational Modal 1. Relational Algebra is a __________ query language that takes two relations as input and produces another relation as an output of the query. A. Relational B. Structural C. Procedural D. Fundamental Answer» C. Procedural 2. For select operation the ________ appear in the subscript and the ___________ argument appears in the paranthesis after the sigmA:) A. Predicates, relation B. Relation, Predicates C. Operation, Predicates D. Relation, Operation Answer» A. Predicates, relation 3. The ___________ operation, denoted by −, allows us to find tuples that are in one relation but are not in another. A. Union B. Set-difference C. Difference D. Intersection Answer» B. Set-difference 4. In precedence of set operators, the expression is evaluated from A. Left to left B. Left to right C. Right to left D. From user specification Answer» B. Left to right 5. Which one of the following is a set of one or more attributes taken collectively to uniquely identify a record? A. Candidate key B. Sub key C. Super key D. Foreign key Answer» C. Super key 6. Consider attributes ID, CITY and NAME. Which one of this can be considered as a super key? A. NAME B. ID C. CITY D. CITY, ID Answer» B. ID 7. A _____ is a property of the entire relation, rather than of the individual tuples in which each tuple is unique. A. Rows B. Key C. Attribute D. Fields Answer» B. Key 8. An attribute in a relation is a foreign key if the _______ key from one relation is used as an attribute in that relation. A. Candidate B. Primary C. Super D. Sub Answer» B. Primary 9. The relation with the attribute which is the primary key is referenced in another relation. The relation which has the attribute as a primary key is called ______________ A. Referential relation B. Referencing relation C. Referenced relation D. Referred relation Answer» C. Referenced relation 10. The ______ is the one in which the primary key of one relation is used as a normal attribute in another relation. A. Referential relation B. Referencing relation C. Referenced relation D. Referred relation Answer» C. Referenced relation 11. A _________ integrity constraint requires that the values appearing in specified attributes of any tuple in the referencing relation also appear in specified attributes of at least one tuple in the referenced relation. A. Referential B. Referencing C. Specific D. Primary Answer» A. Referential 12. A relational database consists of a collection of A. Tables B. Fields C. Records D. Keys Answer» A. Tables 13. A ________ in a table represents a relationship among a set of values. A. Column B. Key C. Row D. Entry Answer» C. Row 14. The term _______ is used to refer to a row. A. Attribute B. Tuple C. Field D. Instance Answer» B. Tuple 15. The term attribute refers to a ___________ of a table. A. Record B. Column C. Tuple D. Key Answer» B. Column 16. For each attribute of a relation, there is a set of permitted values, called the ________ of that attribute. A. Domain B. Relation C. Set D. Schema Answer» A. Domain 17. Database __________ which is the logical design of the database, and the database _______ which is a snapshot of the data in the database at a given instant in time. A. Instance, Schema B. Relation, Schema C. Relation, Domain D. Schema, Instance Answer» D. Schema, Instance 18. Course(course_id,sec_id,semester) Here the course_id,sec_id and semester are __________ and course is a _________ A. Relations, Attribute B. Attributes, Relation C. Tuple, Relation D. Tuple, Attributes Answer» B. Attributes, Relation 19. Department (dept name, building, budget) and Employee (employee_id, name, dept name, salary) Here the dept_name attribute appears in both the relations. Here using common attributes in relation schema is one way of relating ___________ relations. A. Attributes of common B. Tuple of common C. Tuple of distinct D. Attributes of distinct Answer» C. Tuple of distinct 20. A domain is atomic if elements of the domain are considered to be ____________ units. A. Different B. Indivisbile C. Constant D. Divisible Answer» B. Indivisbile 21. The tuples of the relations can be of ________ order. A. Any B. Same C. Sorted D. Constant Answer» A. Any 22. Choose the correct statement regarding superkeys A. A superkey is an attribute or a group of multiple attributes that can uniquely identify a tuple B. A superkey is a tuple or a set of multiple tuples that can uniquely identify an attribute C. Every superkey is a candidate key D. A superkey is an attribute or a set of attributes that distinguish the relation from other relations Answer» A. A superkey is an attribute or a group of multiple attributes that can uniquely identify a tuple 23. What is an Instance of a Database? A. The logical design of the database system B. The entire set of attributes of the Database put together in a single relation C. The state of the database system at any given point of time D. The initial values inserted into the Database immediately after its creation Answer» C. The state of the database system at any given point of time 24. What is a foreign key? A. A foreign key is a primary key of a relation which is an attribute in another relation B. A foreign key is a superkey of a relation which is an attribute in more than one other relations C. A foreign key is an attribute of a relation that is a primary key of another relation D. A foreign key is the primary key of a relation that does not occur anywhere else in the schema Answer» C. A foreign key is an attribute of a relation that is a primary key of another relation 25. What action does ⋈ operator perform in relational algebra A. Output specified attributes from all rows of the input relation and remove duplicate tuples from the output B. Outputs pairs of rows from the two input relations that have the same value on all attributes that have the same name C. Output all pairs of rows from the two input relations (regardless of whether or not they have the same values on common attributes) D. Return rows of the input relation that satisfy the predicate Answer» A. Output specified attributes from all rows of the input relation and remove duplicate tuples from the output 26. What does the “x” operator do in relational algebra? A. Output specified attributes from all rows of the input relation. Remove duplicate tuples from the output B. Output pairs of rows from the two input relations that have the same value on all attributes that have the same name C. Output all pairs of rows from the two input relations (regardless of whether or not they have the same values on common attributes) D. Returns the rows of the input relation that satisfy the predicate Answer» C. Output all pairs of rows from the two input relations (regardless of whether or not they have the same values on common attributes) 27. An attribute is a __________ in a relation. A. Row B. Column C. Value D. Tuple Answer» B. Column 28. What is the method of specifying a primary key in a schema description? A. By writing it in bold letters B. By underlining it using a dashed line C. By writing it in capital letters D. By underlining it using a bold line Answer» D. By underlining it using a bold line 29. Statement 1: A tuple is a row in a relation Statement 2: Existence of multiple foreign keys in a same relation is possible A. Both the statements are true B. Statement 1 is correct but Statement 2 is false C. Statement 1 is false but Statement 2 is correct D. Both the statements are false Answer» A. Both the statements are true 30. Choose the option that correctly explains in words, the function of the following relational algebra expression σyear≥2009 (book ⋈ borrow) A. Selects all tuples from the Cartesian product of book and borrow B. Selects all the tuples from the natural join of book and borrow wherever the year is lesser than 2009 C. Selects all the tuples from the natural join of book and student wherever the year is greater than or equal to 2009 D. Selects all tuples from the Cartesian product of book and borrow wherever the year is greater than or equal to 2009 Answer» B. Selects all the tuples from the natural join of book and borrow wherever the year is lesser than 2009 31. State true or false: If a relation consists of a foreign key, then it is called a referenced relation of the foreign key dependency. A. True B. False C. none D. all Answer» B. False 32. Which of the following information does an SQL DDL not specify? A. The schema for each relation B. The integrity constraints C. The operations on the tuples D. The security and authorization information for each relation Answer» C. The operations on the tuples 33. Which of the following data types does the SQL standard not support? A. char(n) B. String(n) C. varchar(n) D. float(n) Answer» B. String(n) 34. Which command is used to create a new relation in SQL A. create table( , …) B. create relation( , …) C. new table( , …) D. new relation( , …) Answer» A. create table( , …) 35. If a1, a2, a3 are attributes in a relation and S is another relation, which of the following is an incorrect specification of an integrity constraint? A. primary key(a1, a2, a3) B. primary key(a1) C. foreign key(a1, a2) references S D. foreign key(a1, a2) Answer» D. foreign key(a1, a2) 36. What is the syntax to load data into the database? (Consider D as the database and a, b, c as datA:) A. enter into D (a, b, C:); B. insert into D values (a, b, C:); C. insert into D (a, b, C:); D. insert (a, b, C:) values into D; Answer» B. insert into D values (a, b, C:); 37. Which of the following commands do we use to delete a relation (R) from a database? A. drop table R B. drop relation R C. delete table R D. delete from R Answer» A. drop table R 38. Which of the following commands do we use to delete all the tuples from a relation (R)? A. delete table R B. drop table R C. delete from R D. drop from R Answer» C. delete from R 39. Choose the correct command to delete an attribute A from a relation R A. alter table R delete A B. alter table R drop A C. alter table drop A from R D. delete A from R Answer» B. alter table R drop A 40. create table apartment(ownerID varchar (5), ownername varchar(25), floor numeric(4,0), primary key (ownerID:)); Choose the correct option regarding the above statement A. The statement is syntactically wrong B. It creates a relation with three attributes ownerID, ownername, floor in which floor cannot be null. C. It creates a relation with three attributes ownerID, ownername, floor in which ownerID cannot be null. D. It creates a relation with three attributes ownerID, ownername, floor in which ownername must consist of at least 25 characters. Answer» C. It creates a relation with three attributes ownerID, ownername, floor in which ownerID cannot be null. 41. What does the notnull integrity constraint do? A. It ensures that at least one tuple is present in the relation B. It ensures that at least one foreign key is present in the relation C. It ensures that all tuples have a finite value on a specified attribute D. It ensures that all tuples have finite attributes on all the relations Answer» C. It ensures that all tuples have a finite value on a specified attribute Tags Question and answers in Relational Modal, Relational Modal multiple choice questions and answers, Relational Modal Important MCQs, Solved MCQs for Relational Modal, Relational Modal MCQs with answers PDF download
__label__pos
0.999469
37 I have a request to add in another URL parameter that directs to a state that I already have set up. For efficiency purposes, I'm trying to see if I can add multiple URLs to point to the same state, or should I just use the $UrlRouterProvider.when() method to re-direct to that state in this new case. Ex. this is what already exists .state('site.link1', { url: '/link1', templateUrl: '/views/link1.html', controller: 'link1Ctrl' }) and the request is to add www.site.com/newlink that points to the link1 page. Is there something like this; .state('site.link1', { url: '/link1, /newlink', ... 3 • 1 Currently you'd have to use either .when or define state multi times, with different url... there is some how to: stackoverflow.com/a/23853129/1679310 Jan 25, 2015 at 15:48 • 1 if the only change is the stateParams you can get away with the same route and just passing empty param, the only thing that is not so elegant is that you'll have a "//" in your url. if the spec is defining a completely different url then a new state is needed Jan 25, 2015 at 16:23 • if 2 different param value on the same state redirects to these different ursl then you can use templateurl as a function where you get first argument as routeparam and return the url accordingly. but it is little unclear on what exactly you are trying to do. – PSL Jan 25, 2015 at 16:44 5 Answers 5 16 Try using the Regex and a parameter in the url. It is not optimal but works. .state('site.link1', { url: '/{path:link1|newlink}', templateUrl: '/views/link1.html', controller: 'link1Ctrl' }) More information on regex in Urls. To generate links with ui-sref pass the same parameter with the state name as a function <a ui-sref="site.link1({path:'link1'})" >site link 1</a> <a ui-sref="site.link1({path:'newlink'})">site new link</a> 15 You use params: https://github.com/angular-ui/ui-router/wiki/URL-Routing .state('site.link', { url: '/{link}' .. } so when you use the same state like this $state.go('site.link', {link: 'link1'}) $state.go('site.link', {link: 'link2'}) 2 • 1 i'm curious, where did you see such an example in the link you provide ? – mpgn Jan 15, 2016 at 15:06 • The link show the documentation and it uses contacts.details. I just use this example to show how to change state using "go" function. – Valter Jan 17, 2016 at 16:06 11 you can used when() function .state('site.link1', { url: '/link1', templateUrl: '/views/link1.html', controller: 'link1Ctrl' }) then on root config angular.module('myApp', [...]) .config(function ($urlRouterProvider) { $urlRouterProvider.when(/newlink/, ['$state','$match', function ($state, $match) { $state.go('site.link1'); }]); }); 4 I found this approach to be quite simple and clean: create two equal states, just changing the url property //Both root and login are the same, but with different url's. var rootConfig = { url: '/', templateUrl:'html/authentication/login.html', controller: 'authCtrl', data: { requireLogin: false } } var loginConfig = Object.create(rootConfig) loginConfig.url = '/login' $stateProvider .state('root', rootConfig) .state('login', loginConfig) 1 • Well, isn't this example for two states? Would there be an instance where we can override the state's url option? So, at the end, there will be just one state. Thanks. Jul 20, 2018 at 15:53 0 I had almost the same problem, only with another constraint - I didn't want to use a redirect, since I wanted the url in the browser to stay the same, but display the same state. This was because I wanted the chrome saved passwords to work for users that already saved the previous url. In my case I wanted these two urls : /gilly and /new/gilly to both point to the same state. I solved this by having one state defined for /gilly, and for the second url, I defined an abstract state called /new. This should be set up like this : $stateProvider.state('new', { abstract: true, url: '/new' template: '', controller: function() { } }).state('gilly', { url: '/gilly', template: 'gilly.html', controller: 'GillyController' }).state('new.gilly', { url: '/gilly', // don't add the '/new' prefix here! template: 'gilly.html', controller: 'GillyController' }); Your Answer By clicking “Post Your Answer”, you agree to our terms of service, privacy policy and cookie policy Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.518235
Beefy Boxes and Bandwidth Generously Provided by pair Networks Pathologically Eclectic Rubbish Lister   PerlMonks   Comment on ( #3333=superdoc: print w/ replies, xml ) Need Help?? I am still a newbie, but / therefore - what is wrong with using hash here (with keys as line numbers and values as bases)? #!/usr/bin/perl -l use strict; use warnings; print "If an array element is a line number and a base together as a s +tring:"; my %positions; for my $line (split /\n/, <<'END') 1 ACAC 2 AGAC 3 AGTC 4 ACCA END { next if $line =~ /[GT]/; my ($number, $bases ) = split / /, $line; $positions{$number} = $bases; } for my $number ( sort keys %positions ) { print "$number => $positions{$number}"; print "Just print a number $number"; } print "If there is an array element for a line number and for a base:" +; my @array = ( qw (1 ACAC 2 AGAC 3 AGTC 4 ACCA) ); %positions = (); for (my $i = 0; $i < @array; $i +=2 ) { my ($number, $bases ) = @array[$i, $i+1]; next if $bases =~ /[GT]/; $positions{$number} = $bases; } for my $number (sort keys %positions) { print "$number => $positions{$number}"; print "Just print a number $number"; } It prints: If an array element is a line number and a base together as a string: 1 => ACAC Just print a number 1 4 => ACCA Just print a number 4 If there is an array element for a line number and for a base: 1 => ACAC Just print a number 1 4 => ACCA Just print a number 4 In reply to Re: filtering an array by vagabonding electron in thread filtering an array by prbndr Title: Use:  <p> text here (a paragraph) </p> and:  <code> code here </code> to format your post; it's "PerlMonks-approved HTML": • Posts are HTML formatted. Put <p> </p> tags around your paragraphs. Put <code> </code> tags around your code and data! • Read Where should I post X? if you're not absolutely sure you're posting in the right place. • Please read these before you post! — • Posts may use any of the Perl Monks Approved HTML tags: a, abbr, b, big, blockquote, br, caption, center, col, colgroup, dd, del, div, dl, dt, em, font, h1, h2, h3, h4, h5, h6, hr, i, ins, li, ol, p, pre, readmore, small, span, spoiler, strike, strong, sub, sup, table, tbody, td, tfoot, th, thead, tr, tt, u, ul, wbr • Outside of code tags, you may need to use entities for some characters:         For:     Use: & &amp; < &lt; > &gt; [ &#91; ] &#93; • Link using PerlMonks shortcuts! What shortcuts can I use for linking? • See Writeup Formatting Tips and other pages linked from there for more info. • Log In? Username: Password: What's my password? Create A New User Chatterbox? and the web crawler heard nothing... How do I use this? | Other CB clients Other Users? Others making s'mores by the fire in the courtyard of the Monastery: (7) As of 2014-07-26 08:07 GMT Sections? Information? Find Nodes? Leftovers? Voting Booth? My favorite superfluous repetitious redundant duplicative phrase is: Results (175 votes), past polls
__label__pos
0.702564
1 vote 1answer 671 views How do you install Adobe Production Premium CS5 on a different hard drive in windows? Is there a way of installing Adobe Production Premium CS5 on a hard drive other than C:\? I tried changing the install path in the installer but after installing parts of the program don't work and it ... 2 votes 1answer 364 views How to setup one instance of Kontakt to use more than one virtual instruments to their own outputs So as far as using one MIDI channel per instruments, I understand that (somewhat). But how to route their output to individual channels in Reaper is a bit beyond me. Is there any way to achieve this? ...
__label__pos
0.663243
Stack Overflow is a community of 4.7 million programmers, just like you, helping each other. Join them; it only takes a minute: Sign up Join the Stack Overflow community to: 1. Ask programming questions 2. Answer and help your peers 3. Get recognized for your expertise Is there an article/algorithm on how I can read a long file at a certain rate? Say I do not want to pass 10 KB/sec while issuing reads. share|improve this question 1   The question is why would you want to read a file at a certain rate? Sounds like you want to read data on demand, thus understanding your "demand" might allow us to point you to a better solution. – EFraim May 16 '09 at 14:13      I am going to download a large file off the internet but i don't want my application to pass the user set limit. – Hamza Yerlikaya May 16 '09 at 14:17      up vote 4 down vote accepted The crude solution is just to read a chunk at a time and then sleep eg 10k then sleep a second. But the first question I have to ask is: why? There are a couple of likely answers: 1. You don't want to create work faster than it can be done; or 2. You don't want to create too great a load on the system. My suggestion is not to control it at the read level. That's kind of messy and inaccurate. Instead control it at the work end. Java has lots of great concurrency tools to deal with this. There are a few alternative ways of doing this. I tend to like using a producer consumer pattern for soling this kind of problem. It gives you great options on being able to monitor progress by having a reporting thread and so on and it can be a really clean solution. Something like an ArrayBlockingQueue can be used for the kind of throttling needed for both (1) and (2). With a limited capacity the reader will eventually block when the queue is full so won't fill up too fast. The workers (consumers) can be controlled to only work so fast to also throttle the rate covering (2). share|improve this answer A simple solution, by creating a ThrottledInputStream. This should be used like this: final InputStream slowIS = new ThrottledInputStream(new BufferedInputStream(new FileInputStream("c:\\file.txt"),8000),300); 300 is the number of kilobytes per second. 8000 is the block size for BufferedInputStream. This should of course be generalized by implementing read(byte b[], int off, int len), which will spare you a ton of System.currentTimeMillis() calls. System.currentTimeMillis() is called once for each byte read, which can cause a bit of an overhead. It should also be possible to store the number of bytes that can savely be read without calling System.currentTimeMillis(). Be sure to put a BufferedInputStream in between, otherwise the FileInputStream will be polled in single bytes rather than blocks. This will reduce the CPU load form 10% to almost 0. You will risk to exceed the data rate by the number of bytes in the block size. import java.io.InputStream; import java.io.IOException; public class ThrottledInputStream extends InputStream { private final InputStream rawStream; private long totalBytesRead; private long startTimeMillis; private static final int BYTES_PER_KILOBYTE = 1024; private static final int MILLIS_PER_SECOND = 1000; private final int ratePerMillis; public ThrottledInputStream(InputStream rawStream, int kBytesPersecond) { this.rawStream = rawStream; ratePerMillis = kBytesPersecond * BYTES_PER_KILOBYTE / MILLIS_PER_SECOND; } @Override public int read() throws IOException { if (startTimeMillis == 0) { startTimeMillis = System.currentTimeMillis(); } long now = System.currentTimeMillis(); long interval = now - startTimeMillis; //see if we are too fast.. if (interval * ratePerMillis < totalBytesRead + 1) { //+1 because we are reading 1 byte try { final long sleepTime = ratePerMillis / (totalBytesRead + 1) - interval; // will most likely only be relevant on the first few passes Thread.sleep(Math.max(1, sleepTime)); } catch (InterruptedException e) {//never realized what that is good for :) } } totalBytesRead += 1; return rawStream.read(); } } share|improve this answer 1   FYI: The interrupted Exception is to ensure that the Thread can immediately response to a interrupt request, even if it is sleeping. – Simiil Oct 31 '13 at 20:10 • while !EOF • store System.currentTimeMillis() + 1000 (1 sec) in a long variable • read a 10K buffer • check if stored time has passed • if it isn't, Thread.sleep() for stored time - current time Creating ThrottledInputStream that takes another InputStream as suggested would be a nice solution. share|improve this answer It depends a little on whether you mean "don't exceed a certain rate" or "stay close to a certain rate." If you mean "don't exceed", you can guarantee that with a simple loop: while not EOF do read a buffer Thread.wait(time) write the buffer od The amount of time to wait is a simple function of the size of the buffer; if the buffer size is 10K bytes, you want to wait a second between reads. If you want to get closer than that, you probably need to use a timer. • create a Runnable to do the reading • create a Timer with a TimerTask to do the reading • schedule the TimerTask n times a second. If you're concerned about the speed at which you're passing the data on to something else, instead of controlling the read, put the data into a data structure like a queue or circular buffer, and control the other end; send data periodically. You need to be careful with that, though, depending on the data set size and such, because you can run into memory limitations if the reader is very much faster than the writer. share|improve this answer If you have used Java I/O then you should be familiar with decorating streams. I suggest an InputStream subclass that takes another InputStream and throttles the flow rate. (You could subclass FileInputStream but that approach is highly error-prone and inflexible.) Your exact implementation will depend upon your exact requirements. Generally you will want to note the time your last read returned (System.nanoTime). On the current read, after the underlying read, wait until sufficient time has passed for the amount of data transferred. A more sophisticated implementation may buffer and return (almost) immediately with only as much data as rate dictates (be careful that you should only return a read length of 0 if the buffer is of zero length). share|improve this answer Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.810339
5.6 C New York Friday, February 23, 2024 The Importance of Timely Submission of Software Engineering Assignments Software engineering is a complex field that requires meticulous planning, coding, and testing to ensure the development of reliable and efficient software systems. As part of their academic journey, software engineering students often face numerous assignments that test their understanding and application of various concepts. One crucial aspect of these assignments is their timely submission, which holds significant importance for the students’ academic progress and future careers. This article explores the reasons why timely submission of software engineering assignments is essential and highlights the benefits of seeking Software Engineering Assignment Help when needed. Academic Progress: Timely submission of assignments is vital for maintaining a good academic record. Assignments contribute to a significant portion of students’ grades, and submitting them on time demonstrates discipline, responsibility, and commitment to learning. Late submissions can result in penalties, reduced grades, or even academic probation. By submitting assignments promptly, students ensure that they make the most of their academic opportunities and maximize their chances of success. Professionalism: In the field of software engineering, meeting deadlines is a crucial aspect of professional life. Completing assignments on time helps students cultivate professionalism, an essential quality sought after by employers. Employers value professionals who can work efficiently under pressure and deliver results within stipulated timeframes. By practicing timely submission of assignments during their academic journey, students develop a habit that can benefit them in their future careers. Time Management: Timely submission of assignments requires effective time management skills. By planning and organizing their work, students learn to allocate sufficient time to complete their assignments, ensuring they meet the deadlines. These skills are valuable in software engineering, where projects often have strict timelines and deliverables. Learning to manage time effectively during academic assignments prepares students to handle real-world projects with professionalism and efficiency. Reduced Stress: Procrastination and last-minute submissions can lead to increased stress levels. When students delay their assignments, they often find themselves rushing to complete them, resulting in compromised quality. By submitting assignments on time, students can avoid unnecessary stress and dedicate ample time to understand concepts, conduct thorough research, and produce high-quality work. Timely submission fosters a sense of calm and confidence, enabling students to perform at their best. Feedback and Improvement: Submitting assignments on time allows students to receive timely feedback from their professors. Feedback plays a crucial role in understanding strengths and areas for improvement. It helps students identify their mistakes, rectify them, and enhance their knowledge and skills. By receiving feedback promptly, students can implement suggestions in subsequent assignments, leading to continuous improvement in their understanding of software engineering concepts. Seeking Software Engineering Assignment Help: While the importance of timely submission is clear, students may encounter difficulties that hinder their progress. This is where software Engineering Assignment Help services can be invaluable. When facing complex assignments or struggling with time constraints, students can seek assistance from professionals in the field. These services provide expert guidance, ensuring students meet deadlines and produce high-quality work. Seeking help not only aids in timely submission but also enhances students’ understanding of the subject matter. In conclusion, timely submission of software engineering assignments is crucial for academic progress, professionalism, time management, stress reduction, feedback, and improvement. By submitting assignments on time, students demonstrate their commitment to learning, cultivate essential professional qualities, and develop effective time management skills. Additionally, seeking software engineering assignment help when needed can further support timely submission and provide valuable guidance. By recognizing the importance of meeting deadlines, software engineering students can lay a strong foundation for their academic and professional success. Ahsan Khan Ahsan Khan Hi, I'm admin of techfily if you need any post and any information then kindly contact us! Mail: [email protected] WhatsApp: +923233319956 Best Regards, Related Articles Stay Connected 0FansLike 3,912FollowersFollow 0SubscribersSubscribe Latest Articles
__label__pos
0.998621
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Public WebGL] Feedback from HTML5 WG on WebGL | [Fwd: Re: WebGL | The 3D Canvas Context for HTML] Since our release yesterday, we've got some feedback from the W3C HTML5 Listserv. Most of the WHATWG post on the HTML5 listserv at W3C, which has the advantage of attracting Microsoft participation. The W3C HTML5 listserv is archived publicly here: http://lists.w3.org/Archives/Public/public-html/ --- Begin Message --- On Thu, 10 Dec 2009 23:47:07 +0100, Arun Ranganathan <[email protected]> wrote: [...] I've been trying to cut my mailing list subscriptions recently so hopefully someone can forward this as appropriate. https://cvs.khronos.org/svn/repos/registry/trunk/public/webgl/doc/spec/WebGL-spec.html The draft appears to use outdated Web IDL syntax. It also uses features, e.g. NameDeleter that are controversial with TC39 so you may want to reconsider using those. It also seems that a bunch of the new objects introduced should really be part of ECMAScript instead. E.g. all the new typed array interfaces. Some of the typedefs are scoped to WebGLRenderingContext but are in fact used all over the place. Also, why is it WebGLRenderingContext and not CanvasRenderingContextWebGL which would be more consistent with the existing CanvasRenderingContext2D? It does not seem to be defined when the event defined at the end of the specification is actually dispatched. The name of the event is also not defined, just the interface it implements. Also, if more than one resource is lost, which one does "resource" return? And why do you need the context attribute? Isn't it already clear that the event is for the WebGL context? -- Anne van Kesteren http://annevankesteren.nl/ --- End Message ---
__label__pos
0.797165
anonymous • anonymous Tim, Paco, Maria, and Jenny are standing in line for lunch. Jenny is standing between Paco and Maria, and Paco’s position in line is an odd number. Tim is not standing on either end of the line, and he is in front of Jenny. Which friend is standing fourth in line? Discrete Math • Stacey Warren - Expert brainly.com Hey! We 've verified this expert answer for you, click below to unlock the details :) SOLVED At vero eos et accusamus et iusto odio dignissimos ducimus qui blanditiis praesentium voluptatum deleniti atque corrupti quos dolores et quas molestias excepturi sint occaecati cupiditate non provident, similique sunt in culpa qui officia deserunt mollitia animi, id est laborum et dolorum fuga. Et harum quidem rerum facilis est et expedita distinctio. Nam libero tempore, cum soluta nobis est eligendi optio cumque nihil impedit quo minus id quod maxime placeat facere possimus, omnis voluptas assumenda est, omnis dolor repellendus. Itaque earum rerum hic tenetur a sapiente delectus, ut aut reiciendis voluptatibus maiores alias consequatur aut perferendis doloribus asperiores repellat. schrodinger • schrodinger I got my questions answered at brainly.com in under 10 minutes. Go to brainly.com now for free help! anonymous • anonymous Let T=Tim P=Paco M=Maria J=Jenny then, if T is not on either end but must be in front of J, we know the position of the 4 must be either P T J M OR M T J P. Given that, which one must it be? Looking for something else? Not the answer you are looking for? Search for more explanations.
__label__pos
0.51978
Let 11 divide! N = 100 a + 10 b + c \begin{aligned} \text{N}= {\color{#3D99F6}100a + 10b + c }\end{aligned} I have a three digits number as shown above which is always divisible by 11 where a , b a, b and c c are it's digits. ( 1 ) c a b , c a b + c = 0 ( 2 ) c a > b , c a + b + c = 11 ( 3 ) c a > c , c a c b = 0 ( 4 ) c c > a , c c a b = 0 \begin{aligned} & (1) \phantom{c}a \leq b , \phantom{c}a-b +c =0 \\& (2) \phantom{c} a >b, \phantom{c} a+b +c =11 \\& (3) \phantom{c} a > c ,\phantom{c} a-c -b =0 \\& (4) \phantom{c} c> a, \phantom{c} c-a-b = 0\\&\end{aligned} How many of the above statements are/is always true ? 4 1 All are false 3 All are true 2 This section requires Javascript. You are seeing this because something didn't load right. We suggest you, (a) try refreshing the page, (b) enabling javascript if it is disabled on your browser and, finally, (c) loading the non-javascript version of this page . We're sorry about the hassle. 1 solution Giorgos K. Feb 22, 2018 using Mathematica case 1 #[[1]] - #[[2]] + #[[3]] & /@ Select[IntegerDigits /@ Table[11 i, {i, 10, 90}], #[[1]] <= #[[2]] &] {0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0} True case 2 #[[1]] + #[[2]] + #[[3]] & /@ Select[IntegerDigits /@ Table[11 i, {i, 10, 90}], #[[1]] > #[[2]] &] {11, 11, 13, 11, 13, 15, 11, 13, 15, 17, 11, 13, 15, 17, 19, 11, 13, 15, 17, 19, 21, 11, 13, 15, 17, 19, 21, 23, 11, 13, 15, 17, 19, 21, 23, 25} False case 3 #[[1]] - #[[2]] - #[[3]] & /@ Select[IntegerDigits /@ Table[11 i, {i, 10, 90}], #[[1]] > #[[3]] &] {0, 0, -2, 0, -2, -4, 0, -2, -4, -6, 0, -2, -4, -6, -8, 1, 0, -2, -4, -6, 3, 1, -1, 0, -2, -4, 5, 3, 1, -1, -3, 0, -2, 7, 5, 3, 1, -1, -3, -5, 0} False case 4 #[[3]] - #[[1]] + #[[2]] & /@ Select[IntegerDigits /@ Table[11 i, {i, 10, 90}], #[[1]] < #[[3]] &] {4, 6, 8, 10, 12, 14, 16, 7, 6, 8, 10, 12, 14, 5, 7, 8, 10, 12, 3, 5, 7, 10, 1, 3, 5, 7, 3, 5, 7, 5, 7, 7} False 0 pending reports × Problem Loading... Note Loading... Set Loading...
__label__pos
0.973146
FileMaker cURL JSON zu eigenem Webserver senden Mit FileMaker 16 ist es möglich cURL-Optionen mit nativen Bordmitteln zu nutzen. Wie man ein JSON von einem Webserver abholt, haben wir schon in einem früheren Tutorial gezeigt. In diesem Artikel geht es darum, ein JSON zu einem Webserver zu schicken. Auf unserer Homepage findet Ihr unter https://fm-tutorial.de/api/json/json_check.php ein PHP-Script, an das Ihr euer JSON senden könnt. Wenn Ihr auf den Link klickt, erscheint ein JSON: { "code": 500, "msg": "JSON - Syntaxfehler, ung\u00fcltiges JSON" } Dieses PHP-Script dient als kleine API z.B. zur eurer Homepage, Webshop, etc. Hier in unserem Beispiel-PHP-Script wird ein Token “fm-tutorial.de” im JSON erwartet. Sollte kein gültiger Token im JSON enthalten sein, wird obige Meldung mit dem HTTP-Code 500 ausgegeben. Da ich über einen ganz gewöhnlichen Browseraufruf kein JSON mit senden kann, wird ein JSON-Syntax-Fehler ausgegeben. Das PHP-Script beinhaltet zwei Funktionen isJSONValid() und getCode($code), um das JSON zu prüfen und einen Antwort-Header zu setzen. function isJSONValid() { switch(json_last_error()) { case JSON_ERROR_NONE: return "ok"; case JSON_ERROR_DEPTH: return 'JSON - Maximale Stacktiefe überschritten'; case JSON_ERROR_STATE_MISMATCH: return 'JSON - Unterlauf oder Nichtübereinstimmung der Modi'; case JSON_ERROR_CTRL_CHAR: return 'JSON - Unerwartetes Steuerzeichen gefunden'; case JSON_ERROR_SYNTAX: return 'JSON - Syntaxfehler, ungültiges JSON'; case JSON_ERROR_UTF8: return 'JSON - Missgestaltete UTF-8 Zeichen, möglicherweise fehlerhaft kodiert'; default: return 'JSON - Unbekannter Fehler'; } } Wird ausgeführt nachdem versucht wurde, das gesendete JSON in ein Objekt zu parsen, um zu prüfen welcher Fehler beim parsen des JSON auftrat. function getCode($code) { switch ($code) { case 100: $text = 'Continue'; break; case 101: $text = 'Switching Protocols'; break; case 200: $text = 'OK'; break; case 201: $text = 'Created'; break; case 202: $text = 'Accepted'; break; case 203: $text = 'Non-Authoritative Information'; break; case 204: $text = 'No Content'; break; case 205: $text = 'Reset Content'; break; case 206: $text = 'Partial Content'; break; case 300: $text = 'Multiple Choices'; break; case 301: $text = 'Moved Permanently'; break; case 302: $text = 'Moved Temporarily'; break; case 303: $text = 'See Other'; break; case 304: $text = 'Not Modified'; break; case 305: $text = 'Use Proxy'; break; case 400: $text = 'Bad Request'; break; case 401: $text = 'Unauthorized'; break; case 402: $text = 'Payment Required'; break; case 403: $text = 'Forbidden'; break; case 404: $text = 'Not Found'; break; case 405: $text = 'Method Not Allowed'; break; case 406: $text = 'Not Acceptable'; break; case 407: $text = 'Proxy Authentication Required'; break; case 408: $text = 'Request Time-out'; break; case 409: $text = 'Conflict'; break; case 410: $text = 'Gone'; break; case 411: $text = 'Length Required'; break; case 412: $text = 'Precondition Failed'; break; case 413: $text = 'Request Entity Too Large'; break; case 414: $text = 'Request-URI Too Large'; break; case 415: $text = 'Unsupported Media Type'; break; case 500: $text = 'Internal Server Error'; break; case 501: $text = 'Not Implemented'; break; case 502: $text = 'Bad Gateway'; break; case 503: $text = 'Service Unavailable'; break; case 504: $text = 'Gateway Time-out'; break; case 505: $text = 'HTTP Version not supported'; break; default: exit('Unknown http status code "' . htmlentities($code) . '"'); break; } $protocol = (isset($_SERVER['SERVER_PROTOCOL']) ? $_SERVER['SERVER_PROTOCOL'] : 'HTTP/1.0'); header($protocol . ' ' . $code . ' ' . $text); } Setzt den HTTP-Antwort-Code und den entsprechenden Header. Auf diesen Header reagiert auch FileMaker. Senden wir z.B. den Code 401, gibt FileMaker den FileMaker-Error-Code 1627 aus. Hinweis: FileMaker zeigt nur die Response, wenn es sich um einen erfolgreichen Quittierungcode z.B. 200, 201 handelt. Damit immer eine Response an FileMaker zurück gegeben wird, sollte im Header immer Code 200 zurückgegeben werden. Die Überprüfung kann man über das JSON vornehmen, in dem der Parameter code auf den Wert 200, in FileMaker geprüft wird.   Da das JSON über keinen POST-Parameter z.B. data = {“json” : “wert”} gesendet wird, muss die Response über file_get_contents ermittelt werden. json_decode(file_get_contents('php://input')); Das JSON wird aus dem Request-Body in ein Objekt geparst. Anschließend wird die Funktion isJSONValid() ausgeführt, um zu überprüfen, ob das JSON der Norm entspricht. Anschließend erfolgt die Überprüfung des Tokens. Dieser kann z.B. mittels einer Datenbank geprüft bzw. gesperrt werden. Ist der Token nicht “fm-tutorial.de” wird der Code 401 zurückgegeben. #JSON validieren $isValid = isJSONValid(); if( $isValid != 'ok') { $r['code'] = 500; $r['msg'] = $isValid; } #Token prüfen elseif($ar->token != 'fm-tutorial.de') { $r['code'] = 401; $r['msg'] = 'Dein Token stimmt nicht!'; } else { $r['code'] = 200; $r['msg'] = 'Ok, dein JSON entspricht der Prüfung.'; } Nun muss die Antwort ($r), welche momentan noch ein PHP-Array ist, in ein JSON formatiert und ausgegeben werden. #Output getCode($r['code']); header('Content-Type: application/json'); echo json_encode($r, JSON_PRETTY_PRINT); exit; Wir übergeben den Parameter “code” in der Funktion getCode, dadurch wird der entsprechende HTTP-Status-Code und Header gesetzt. Zusätzlich geben wir im Header an, das wir ein JSON zurückgeben – gefolgt von der Ausgabe des $r-Arrays als formatiertes (JSON_PRETTY_PRINT) JSON. FileMaker Damit ein JSON an unseren kleinen Webservice gesendet werden kann, benötigt man mindestens FileMaker 16. Das JSON holen wir hier im Beispiel aus dem Feld “json” und speichern dies in der Variablen $data. -H 'Content-Type: application/json' --data @$data Wir setzen den Header ( -H ), dass wir ein JSON senden. Durch das –data Attribut übermitteln wir die Daten. Durch das vorangestellte @ weißen wir FileMaker an, den Wert für –data aus der Variabelen $data zu holen, welches unser JSON gespeichert hat. Zusätzlich können wir noch für das Debuggen die Trace und Header setzen. --show-error --dump-header $header --trace-ascii $trace -i -v -H 'Content-Type: application/json' --data @$data Diese cURL-Anweisung speichern wir in den FileMaker-Variablen $curl. Danach benötigen wir noch den Scriptschritt Aus URL einfügen , um den Request abzusetzen. Das PHP-Script und die FileMaker-Programmierung findet Ihr im Download. Falls Ihr dieses PHP-Script auf eurem eigenen Webserver betreiben wollt, benötigt ihr PHP ab Version 5.6.   Download   FM cURL JSON Webserver (212,0 KiB, 1.701 hits) Schreibe einen Kommentar Deine E-Mail-Adresse wird nicht veröffentlicht.
__label__pos
0.886103
Learn how easy it is to sync an existing GitHub or Google Code repo to a SourceForge project! See Demo Close #793 Japanese text input problem (patch available) None fixed nobody None 1 2013-08-26 2013-08-19 Yuuto Tokunaga No I'm an linux user. When I input Japanese text in the editor with ibus (input method framework for linux), I can't see any preedit text on screen. After I pressed enter to settle preedit text, I can see what I wrote. I tried to fix this problem and I found QEditor::inputMethodEvent function in qcodeedit/lib/qeditor.cpp. I found a part of this function isn't work due to #ifdef Q_WS_MAC. Codes between #ifdef Q_WS_MAC and #endif only works on Mac OS. After a bit coding and debugging, I fixed this problem. I checked that Japanese input works good on Kubuntu. But I can't understand why the code is disabled by #ifdef Q_WS_MAC. Is there any reason? patch is here : http://pastebin.ubuntu.com/6002384/ 1 Attachments Discussion • how does "kate" handle japanese inputs ? The issue is that i did not want txs to feel differently from standard programs. why is "e->accept()" necessary ? when is the extra if(...) necessary ?   • Yuuto Tokunaga Yuuto Tokunaga 2013-08-20 I wrote this patch following kate 4.10.5 source code. I found e->accept() at the end of KateViewInternal::inputMethodEvent(QInputMethodEvent* e) function, so I added it. Extra if(...) is needed when the text is deleted by Backspace key or input canceled by Esc key. Without this code, you find that text isn't deleted after you canceled input by pressing Esc key. When you press Esc key, IME send QInputMethodEvent e that includes empty preeditString and empty commitString, but inputMethodEvent function can't handle this so the text isn't deleted.   • okay, implemented in svn. Let's see if any problems arise ...   • status: open --> fixed • Group: -->   • seems to work  
__label__pos
0.767335
head 1.3; access; symbols pkgsrc-2020Q1:1.3.0.30 pkgsrc-2020Q1-base:1.3 pkgsrc-2019Q4:1.3.0.52 pkgsrc-2019Q4-base:1.3 pkgsrc-2019Q3:1.3.0.48 pkgsrc-2019Q3-base:1.3 pkgsrc-2019Q2:1.3.0.46 pkgsrc-2019Q2-base:1.3 pkgsrc-2019Q1:1.3.0.44 pkgsrc-2019Q1-base:1.3 pkgsrc-2018Q4:1.3.0.42 pkgsrc-2018Q4-base:1.3 pkgsrc-2018Q3:1.3.0.40 pkgsrc-2018Q3-base:1.3 pkgsrc-2018Q2:1.3.0.38 pkgsrc-2018Q2-base:1.3 pkgsrc-2018Q1:1.3.0.36 pkgsrc-2018Q1-base:1.3 pkgsrc-2017Q4:1.3.0.34 pkgsrc-2017Q4-base:1.3 pkgsrc-2017Q3:1.3.0.32 pkgsrc-2017Q3-base:1.3 pkgsrc-2017Q2:1.3.0.28 pkgsrc-2017Q2-base:1.3 pkgsrc-2017Q1:1.3.0.26 pkgsrc-2017Q1-base:1.3 pkgsrc-2016Q4:1.3.0.24 pkgsrc-2016Q4-base:1.3 pkgsrc-2016Q3:1.3.0.22 pkgsrc-2016Q3-base:1.3 pkgsrc-2016Q2:1.3.0.20 pkgsrc-2016Q2-base:1.3 pkgsrc-2016Q1:1.3.0.18 pkgsrc-2016Q1-base:1.3 pkgsrc-2015Q4:1.3.0.16 pkgsrc-2015Q4-base:1.3 pkgsrc-2015Q3:1.3.0.14 pkgsrc-2015Q3-base:1.3 pkgsrc-2015Q2:1.3.0.12 pkgsrc-2015Q2-base:1.3 pkgsrc-2015Q1:1.3.0.10 pkgsrc-2015Q1-base:1.3 pkgsrc-2014Q4:1.3.0.8 pkgsrc-2014Q4-base:1.3 pkgsrc-2014Q3:1.3.0.6 pkgsrc-2014Q3-base:1.3 pkgsrc-2014Q2:1.3.0.4 pkgsrc-2014Q2-base:1.3 pkgsrc-2014Q1:1.3.0.2 pkgsrc-2014Q1-base:1.3 pkgsrc-2013Q4:1.2.0.6 pkgsrc-2013Q4-base:1.2 pkgsrc-2013Q3:1.2.0.4 pkgsrc-2013Q3-base:1.2 pkgsrc-2013Q2:1.2.0.2 pkgsrc-2013Q2-base:1.2 pkgsrc-2013Q1:1.1.0.58 pkgsrc-2013Q1-base:1.1 pkgsrc-2012Q4:1.1.0.56 pkgsrc-2012Q4-base:1.1 pkgsrc-2012Q3:1.1.0.54 pkgsrc-2012Q3-base:1.1 pkgsrc-2012Q2:1.1.0.52 pkgsrc-2012Q2-base:1.1 pkgsrc-2012Q1:1.1.0.50 pkgsrc-2012Q1-base:1.1 pkgsrc-2011Q4:1.1.0.48 pkgsrc-2011Q4-base:1.1 pkgsrc-2011Q3:1.1.0.46 pkgsrc-2011Q3-base:1.1 pkgsrc-2011Q2:1.1.0.44 pkgsrc-2011Q2-base:1.1 pkgsrc-2011Q1:1.1.0.42 pkgsrc-2011Q1-base:1.1 pkgsrc-2010Q4:1.1.0.40 pkgsrc-2010Q4-base:1.1 pkgsrc-2010Q3:1.1.0.38 pkgsrc-2010Q3-base:1.1 pkgsrc-2010Q2:1.1.0.36 pkgsrc-2010Q2-base:1.1 pkgsrc-2010Q1:1.1.0.34 pkgsrc-2010Q1-base:1.1 pkgsrc-2009Q4:1.1.0.32 pkgsrc-2009Q4-base:1.1 pkgsrc-2009Q3:1.1.0.30 pkgsrc-2009Q3-base:1.1 pkgsrc-2009Q2:1.1.0.28 pkgsrc-2009Q2-base:1.1 pkgsrc-2009Q1:1.1.0.26 pkgsrc-2009Q1-base:1.1 pkgsrc-2008Q4:1.1.0.24 pkgsrc-2008Q4-base:1.1 pkgsrc-2008Q3:1.1.0.22 pkgsrc-2008Q3-base:1.1 cube-native-xorg:1.1.0.20 cube-native-xorg-base:1.1 pkgsrc-2008Q2:1.1.0.18 pkgsrc-2008Q2-base:1.1 cwrapper:1.1.0.16 pkgsrc-2008Q1:1.1.0.14 pkgsrc-2008Q1-base:1.1 pkgsrc-2007Q4:1.1.0.12 pkgsrc-2007Q4-base:1.1 pkgsrc-2007Q3:1.1.0.10 pkgsrc-2007Q3-base:1.1 pkgsrc-2007Q2:1.1.0.8 pkgsrc-2007Q2-base:1.1 pkgsrc-2007Q1:1.1.0.6 pkgsrc-2007Q1-base:1.1 pkgsrc-2006Q4:1.1.0.4 pkgsrc-2006Q4-base:1.1 pkgsrc-2006Q3:1.1.0.2 pkgsrc-2006Q3-base:1.1; locks; strict; comment @ * @; 1.3 date 2014.03.12.14.20.43; author ryoon; state Exp; branches; next 1.2; commitid AcBFUxrunFdPVpsx; 1.2 date 2013.06.14.14.46.37; author tron; state Exp; branches; next 1.1; commitid qfoEjDfSVq6OMATw; 1.1 date 2006.07.14.14.23.06; author jlam; state Exp; branches; next ; desc @@ 1.3 log @Update to 20121220 * Works fine under Debian GNU/Linux 7.4, NetBSD/amd64 6.99.36 * Merge pkgsrc specific changes Changelog: Dec 20, 2012: fiddled makefile to get correct yacc and bison flags. pick yacc (linux) or bison (mac) as necessary. added __attribute__((__noreturn__)) to a couple of lines in proto.h, to silence someone's enthusiastic checker. fixed obscure call by value bug in split(a[1],a) reported on 9fans. the management of temporary values is just a mess; i took a shortcut by making an extra string copy. thanks to paul patience and arnold robbins for passing it on and for proposed patches. tiny fiddle in setfval to eliminate -0 results in T.expr, which has irritated me for 20+ years. Aug 10, 2011: another fix to avoid core dump with delete(ARGV); again, many thanks to ruslan ermilov. Aug 7, 2011: split(s, a, //) now behaves the same as split(s, a, "") Jun 12, 2011: /pat/, \n /pat/ {...} is now legal, though bad style to use. added checks to new -v code that permits -vnospace; thanks to ruslan ermilov for spotting this and providing the patch. removed fixed limit on number of open files; thanks to aleksey cheusov and christos zoulos. fixed day 1 bug that resurrected deleted elements of ARGV when used as filenames (in lib.c). minor type fiddles to make gcc -Wall -pedantic happier (but not totally so); turned on -fno-strict-aliasing in makefile. May 6, 2011: added #ifdef for isblank. now allows -ffoo as well as -f foo arguments. (thanks, ruslan) May 1, 2011: after advice from todd miller, kevin lo, ruslan ermilov, and arnold robbins, changed srand() to return the previous seed (which is 1 on the first call of srand). the seed is an Awkfloat internally though converted to unsigned int to pass to the library srand(). thanks, everyone. fixed a subtle (and i hope low-probability) overflow error in fldbld, by adding space for one extra \0. thanks to robert bassett for spotting this one and providing a fix. removed the files related to compilation on windows. i no longer have anything like a current windows environment, so i can't test any of it. May 23, 2010: fixed long-standing overflow bug in run.c; many thanks to nelson beebe for spotting it and providing the fix. fixed bug that didn't parse -vd=1 properly; thanks to santiago vila for spotting it. Feb 8, 2010: i give up. replaced isblank with isspace in b.c; there are no consistent header files. Nov 26, 2009: fixed a long-standing issue with when FS takes effect. a change to FS is now noticed immediately for subsequent splits. changed the name getline() to awkgetline() to avoid yet another name conflict somewhere. Feb 11, 2009: temporarily for now defined HAS_ISBLANK, since that seems to be the best way through the thicket. isblank arrived in C99, but seems to be arriving at different systems at different times. Oct 8, 2008: fixed typo in b.c that set tmpvec wrongly. no one had ever run into the problem, apparently. thanks to alistair crooks. Oct 23, 2007: minor fix in lib.c: increase inputFS to 100, change malloc for fields to n+1. fixed memory fault caused by out of order test in setsval. thanks to david o'brien, freebsd, for both fixes. May 1, 2007: fiddle in makefile to fix for BSD make; thanks to igor sobrado. Mar 31, 2007: fixed some null pointer refs calling adjbuf. Feb 21, 2007: fixed a bug in matching the null RE in sub and gsub. thanks to al aho who actually did the fix (in b.c), and to wolfgang seeberg for finding it and providing a very compact test case. fixed quotation in b.c; thanks to Hal Pratt and the Princeton Dante Project. removed some no-effect asserts in run.c. fiddled maketab.c to not complain about bison-generated values. removed the obsolete -V argument; fixed --version to print the version and exit. fixed wording and an outright error in the usage message; thanks to igor sobrado and jason mcintyre. fixed a bug in -d that caused core dump if no program followed. Jan 1, 2007: dropped mac.code from makefile; there are few non-MacOSX mac's these days. Jan 17, 2006: system() not flagged as unsafe in the unadvertised -safe option. found it while enhancing tests before shipping the ;login: article. practice what you preach. removed the 9-years-obsolete -mr and -mf flags. added -version and --version options. core dump on linux with BEGIN {nextfile}, now fixed. removed some #ifdef's in run.c and lex.c that appear to no longer be necessary. @ text @/* $NetBSD: b.c,v 1.2 2013/06/14 14:46:37 tron Exp $ */ /**************************************************************** Copyright (C) Lucent Technologies 1997 All Rights Reserved Permission to use, copy, modify, and distribute this software and its documentation for any purpose and without fee is hereby granted, provided that the above copyright notice appear in all copies and that both that the copyright notice and this permission notice and warranty disclaimer appear in supporting documentation, and that the name Lucent Technologies or any of its entities not be used in advertising or publicity pertaining to distribution of the software without specific, written prior permission. LUCENT DISCLAIMS ALL WARRANTIES WITH REGARD TO THIS SOFTWARE, INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL LUCENT OR ANY OF ITS ENTITIES BE LIABLE FOR ANY SPECIAL, INDIRECT OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. ****************************************************************/ /* lasciate ogne speranza, voi ch'intrate. */ #define DEBUG #include #include #include #include #include "awk.h" #include "ytab.h" #define HAT (NCHARS+2) /* matches ^ in regular expr */ /* NCHARS is 2**n */ #define MAXLIN 22 #define type(v) (v)->nobj /* badly overloaded here */ #define info(v) (v)->ntype /* badly overloaded here */ #define left(v) (v)->narg[0] #define right(v) (v)->narg[1] #define parent(v) (v)->nnext #define LEAF case CCL: case NCCL: case CHAR: case DOT: case FINAL: case ALL: #define ELEAF case EMPTYRE: /* empty string in regexp */ #define UNARY case STAR: case PLUS: case QUEST: /* encoding in tree Nodes: leaf (CCL, NCCL, CHAR, DOT, FINAL, ALL, EMPTYRE): left is index, right contains value or pointer to value unary (STAR, PLUS, QUEST): left is child, right is null binary (CAT, OR): left and right are children parent contains pointer to parent */ int *setvec; int *tmpset; int maxsetvec = 0; int rtok; /* next token in current re */ int rlxval; static uschar *rlxstr; static uschar *prestr; /* current position in current re */ static uschar *lastre; /* origin of last re */ static int setcnt; static int poscnt; char *patbeg; int patlen; #define NFA 20 /* cache this many dynamic fa's */ fa *fatab[NFA]; int nfatab = 0; /* entries in fatab */ fa *makedfa(const char *s, int anchor) /* returns dfa for reg expr s */ { int i, use, nuse; fa *pfa; static int now = 1; if (setvec == 0) { /* first time through any RE */ maxsetvec = MAXLIN; setvec = (int *) malloc(maxsetvec * sizeof(int)); tmpset = (int *) malloc(maxsetvec * sizeof(int)); if (setvec == 0 || tmpset == 0) overflo("out of space initializing makedfa"); } if (compile_time) /* a constant for sure */ return mkdfa(s, anchor); for (i = 0; i < nfatab; i++) /* is it there already? */ if (fatab[i]->anchor == anchor && strcmp((const char *) fatab[i]->restr, s) == 0) { fatab[i]->use = now++; return fatab[i]; } pfa = mkdfa(s, anchor); if (nfatab < NFA) { /* room for another */ fatab[nfatab] = pfa; fatab[nfatab]->use = now++; nfatab++; return pfa; } use = fatab[0]->use; /* replace least-recently used */ nuse = 0; for (i = 1; i < nfatab; i++) if (fatab[i]->use < use) { use = fatab[i]->use; nuse = i; } freefa(fatab[nuse]); fatab[nuse] = pfa; pfa->use = now++; return pfa; } fa *mkdfa(const char *s, int anchor) /* does the real work of making a dfa */ /* anchor = 1 for anchored matches, else 0 */ { Node *p, *p1; fa *f; p = reparse(s); p1 = op2(CAT, op2(STAR, op2(ALL, NIL, NIL), NIL), p); /* put ALL STAR in front of reg. exp. */ p1 = op2(CAT, p1, op2(FINAL, NIL, NIL)); /* put FINAL after reg. exp. */ poscnt = 0; penter(p1); /* enter parent pointers and leaf indices */ if ((f = (fa *) calloc(1, sizeof(fa) + poscnt*sizeof(rrow))) == NULL) overflo("out of space for fa"); f->accept = poscnt-1; /* penter has computed number of positions in re */ cfoll(f, p1); /* set up follow sets */ freetr(p1); if ((f->posns[0] = (int *) calloc(1, *(f->re[0].lfollow)*sizeof(int))) == NULL) overflo("out of space in makedfa"); if ((f->posns[1] = (int *) calloc(1, sizeof(int))) == NULL) overflo("out of space in makedfa"); *f->posns[1] = 0; f->initstat = makeinit(f, anchor); f->anchor = anchor; f->restr = (uschar *) tostring(s); return f; } int makeinit(fa *f, int anchor) { int i, k; f->curstat = 2; f->out[2] = 0; f->reset = 0; k = *(f->re[0].lfollow); xfree(f->posns[2]); if ((f->posns[2] = (int *) calloc(1, (k+1)*sizeof(int))) == NULL) overflo("out of space in makeinit"); for (i=0; i <= k; i++) { (f->posns[2])[i] = (f->re[0].lfollow)[i]; } if ((f->posns[2])[1] == f->accept) f->out[2] = 1; for (i=0; i < NCHARS; i++) f->gototab[2][i] = 0; f->curstat = cgoto(f, 2, HAT); if (anchor) { *f->posns[2] = k-1; /* leave out position 0 */ for (i=0; i < k; i++) { (f->posns[0])[i] = (f->posns[2])[i]; } f->out[0] = f->out[2]; if (f->curstat != 2) --(*f->posns[f->curstat]); } return f->curstat; } void penter(Node *p) /* set up parent pointers and leaf indices */ { switch (type(p)) { ELEAF LEAF info(p) = poscnt; poscnt++; break; UNARY penter(left(p)); parent(left(p)) = p; break; case CAT: case OR: penter(left(p)); penter(right(p)); parent(left(p)) = p; parent(right(p)) = p; break; default: /* can't happen */ FATAL("can't happen: unknown type %d in penter", type(p)); break; } } void freetr(Node *p) /* free parse tree */ { switch (type(p)) { ELEAF LEAF xfree(p); break; UNARY freetr(left(p)); xfree(p); break; case CAT: case OR: freetr(left(p)); freetr(right(p)); xfree(p); break; default: /* can't happen */ FATAL("can't happen: unknown type %d in freetr", type(p)); break; } } /* in the parsing of regular expressions, metacharacters like . have */ /* to be seen literally; \056 is not a metacharacter. */ int hexstr(uschar **pp) /* find and eval hex string at pp, return new p */ { /* only pick up one 8-bit byte (2 chars) */ uschar *p; int n = 0; int i; for (i = 0, p = (uschar *) *pp; i < 2 && isxdigit(*p); i++, p++) { if (isdigit(*p)) n = 16 * n + *p - '0'; else if (*p >= 'a' && *p <= 'f') n = 16 * n + *p - 'a' + 10; else if (*p >= 'A' && *p <= 'F') n = 16 * n + *p - 'A' + 10; } *pp = (uschar *) p; return n; } #define isoctdigit(c) ((c) >= '0' && (c) <= '7') /* multiple use of arg */ int quoted(uschar **pp) /* pick up next thing after a \\ */ /* and increment *pp */ { uschar *p = *pp; int c; if ((c = *p++) == 't') c = '\t'; else if (c == 'n') c = '\n'; else if (c == 'f') c = '\f'; else if (c == 'r') c = '\r'; else if (c == 'b') c = '\b'; else if (c == '\\') c = '\\'; else if (c == 'x') { /* hexadecimal goo follows */ c = hexstr(&p); /* this adds a null if number is invalid */ } else if (isoctdigit(c)) { /* \d \dd \ddd */ int n = c - '0'; if (isoctdigit(*p)) { n = 8 * n + *p++ - '0'; if (isoctdigit(*p)) n = 8 * n + *p++ - '0'; } c = n; } /* else */ /* c = c; */ *pp = p; return c; } char *cclenter(const char *argp) /* add a character class */ { int i, c, c2; uschar *p = (uschar *) argp; uschar *op, *bp; static uschar *buf = 0; static int bufsz = 100; op = p; if (buf == 0 && (buf = (uschar *) malloc(bufsz)) == NULL) FATAL("out of space for character class [%.10s...] 1", p); bp = buf; for (i = 0; (c = *p++) != 0; ) { if (c == '\\') { c = quoted(&p); } else if (c == '-' && i > 0 && bp[-1] != 0) { if (*p != 0) { c = bp[-1]; c2 = *p++; if (c2 == '\\') c2 = quoted(&p); if (c > c2) { /* empty; ignore */ bp--; i--; continue; } while (c < c2) { if (!adjbuf((char **) &buf, &bufsz, bp-buf+2, 100, (char **) &bp, "cclenter1")) FATAL("out of space for character class [%.10s...] 2", p); *bp++ = ++c; i++; } continue; } } if (!adjbuf((char **) &buf, &bufsz, bp-buf+2, 100, (char **) &bp, "cclenter2")) FATAL("out of space for character class [%.10s...] 3", p); *bp++ = c; i++; } *bp = 0; dprintf( ("cclenter: in = |%s|, out = |%s|\n", op, buf) ); xfree(op); return (char *) tostring((char *) buf); } void overflo(const char *s) { FATAL("regular expression too big: %.30s...", s); } void cfoll(fa *f, Node *v) /* enter follow set of each leaf of vertex v into lfollow[leaf] */ { int i; int *p; switch (type(v)) { ELEAF LEAF f->re[info(v)].ltype = type(v); f->re[info(v)].lval.np = right(v); while (f->accept >= maxsetvec) { /* guessing here! */ maxsetvec *= 4; setvec = (int *) realloc(setvec, maxsetvec * sizeof(int)); tmpset = (int *) realloc(tmpset, maxsetvec * sizeof(int)); if (setvec == 0 || tmpset == 0) overflo("out of space in cfoll()"); } for (i = 0; i <= f->accept; i++) setvec[i] = 0; setcnt = 0; follow(v); /* computes setvec and setcnt */ if ((p = (int *) calloc(1, (setcnt+1)*sizeof(int))) == NULL) overflo("out of space building follow set"); f->re[info(v)].lfollow = p; *p = setcnt; for (i = f->accept; i >= 0; i--) if (setvec[i] == 1) *++p = i; break; UNARY cfoll(f,left(v)); break; case CAT: case OR: cfoll(f,left(v)); cfoll(f,right(v)); break; default: /* can't happen */ FATAL("can't happen: unknown type %d in cfoll", type(v)); } } int first(Node *p) /* collects initially active leaves of p into setvec */ /* returns 0 if p matches empty string */ { int b, lp; switch (type(p)) { ELEAF LEAF lp = info(p); /* look for high-water mark of subscripts */ while (setcnt >= maxsetvec || lp >= maxsetvec) { /* guessing here! */ maxsetvec *= 4; setvec = (int *) realloc(setvec, maxsetvec * sizeof(int)); tmpset = (int *) realloc(tmpset, maxsetvec * sizeof(int)); if (setvec == 0 || tmpset == 0) overflo("out of space in first()"); } if (type(p) == EMPTYRE) { setvec[lp] = 0; return(0); } if (setvec[lp] != 1) { setvec[lp] = 1; setcnt++; } if (type(p) == CCL && (*(char *) right(p)) == '\0') return(0); /* empty CCL */ else return(1); case PLUS: if (first(left(p)) == 0) return(0); return(1); case STAR: case QUEST: first(left(p)); return(0); case CAT: if (first(left(p)) == 0 && first(right(p)) == 0) return(0); return(1); case OR: b = first(right(p)); if (first(left(p)) == 0 || b == 0) return(0); return(1); } FATAL("can't happen: unknown type %d in first", type(p)); /* can't happen */ return(-1); } void follow(Node *v) /* collects leaves that can follow v into setvec */ { Node *p; if (type(v) == FINAL) return; p = parent(v); switch (type(p)) { case STAR: case PLUS: first(v); follow(p); return; case OR: case QUEST: follow(p); return; case CAT: if (v == left(p)) { /* v is left child of p */ if (first(right(p)) == 0) { follow(p); return; } } else /* v is right child */ follow(p); return; } } int member(int c, const char *sarg) /* is c in s? */ { uschar *s = (uschar *) sarg; while (*s) if (c == *s++) return(1); return(0); } int match(fa *f, const char *p0) /* shortest match ? */ { int s, ns; uschar *p = (uschar *) p0; s = f->reset ? makeinit(f,0) : f->initstat; if (f->out[s]) return(1); do { /* assert(*p < NCHARS); */ if ((ns = f->gototab[s][*p]) != 0) s = ns; else s = cgoto(f, s, *p); if (f->out[s]) return(1); } while (*p++ != 0); return(0); } int pmatch(fa *f, const char *p0) /* longest match, for sub */ { int s, ns; uschar *p = (uschar *) p0; uschar *q; int i, k; /* s = f->reset ? makeinit(f,1) : f->initstat; */ if (f->reset) { f->initstat = s = makeinit(f,1); } else { s = f->initstat; } patbeg = (char *) p; patlen = -1; do { q = p; do { if (f->out[s]) /* final state */ patlen = q-p; /* assert(*q < NCHARS); */ if ((ns = f->gototab[s][*q]) != 0) s = ns; else s = cgoto(f, s, *q); if (s == 1) { /* no transition */ if (patlen >= 0) { patbeg = (char *) p; return(1); } else goto nextin; /* no match */ } } while (*q++ != 0); if (f->out[s]) patlen = q-p-1; /* don't count $ */ if (patlen >= 0) { patbeg = (char *) p; return(1); } nextin: s = 2; if (f->reset) { for (i = 2; i <= f->curstat; i++) xfree(f->posns[i]); k = *f->posns[0]; if ((f->posns[2] = (int *) calloc(1, (k+1)*sizeof(int))) == NULL) overflo("out of space in pmatch"); for (i = 0; i <= k; i++) (f->posns[2])[i] = (f->posns[0])[i]; f->initstat = f->curstat = 2; f->out[2] = f->out[0]; for (i = 0; i < NCHARS; i++) f->gototab[2][i] = 0; } } while (*p++ != 0); return (0); } int nematch(fa *f, const char *p0) /* non-empty match, for sub */ { int s, ns; uschar *p = (uschar *) p0; uschar *q; int i, k; /* s = f->reset ? makeinit(f,1) : f->initstat; */ if (f->reset) { f->initstat = s = makeinit(f,1); } else { s = f->initstat; } patlen = -1; while (*p) { q = p; do { if (f->out[s]) /* final state */ patlen = q-p; /* assert(*q < NCHARS); */ if ((ns = f->gototab[s][*q]) != 0) s = ns; else s = cgoto(f, s, *q); if (s == 1) { /* no transition */ if (patlen > 0) { patbeg = (char *) p; return(1); } else goto nnextin; /* no nonempty match */ } } while (*q++ != 0); if (f->out[s]) patlen = q-p-1; /* don't count $ */ if (patlen > 0 ) { patbeg = (char *) p; return(1); } nnextin: s = 2; if (f->reset) { for (i = 2; i <= f->curstat; i++) xfree(f->posns[i]); k = *f->posns[0]; if ((f->posns[2] = (int *) calloc(1, (k+1)*sizeof(int))) == NULL) overflo("out of state space"); for (i = 0; i <= k; i++) (f->posns[2])[i] = (f->posns[0])[i]; f->initstat = f->curstat = 2; f->out[2] = f->out[0]; for (i = 0; i < NCHARS; i++) f->gototab[2][i] = 0; } p++; } return (0); } Node *reparse(const char *p) /* parses regular expression pointed to by p */ { /* uses relex() to scan regular expression */ Node *np; dprintf( ("reparse <%s>\n", p) ); lastre = prestr = (uschar *) p; /* prestr points to string to be parsed */ rtok = relex(); /* GNU compatibility: an empty regexp matches anything */ if (rtok == '\0') { /* FATAL("empty regular expression"); previous */ return(op2(EMPTYRE, NIL, NIL)); } np = regexp(); if (rtok != '\0') FATAL("syntax error in regular expression %s at %s", lastre, prestr); return(np); } Node *regexp(void) /* top-level parse of reg expr */ { return (alt(concat(primary()))); } Node *primary(void) { Node *np; switch (rtok) { case CHAR: np = op2(CHAR, NIL, itonp(rlxval)); rtok = relex(); return (unary(np)); case ALL: rtok = relex(); return (unary(op2(ALL, NIL, NIL))); case EMPTYRE: rtok = relex(); return (unary(op2(ALL, NIL, NIL))); case DOT: rtok = relex(); return (unary(op2(DOT, NIL, NIL))); case CCL: np = op2(CCL, NIL, (Node*) cclenter((char *) rlxstr)); rtok = relex(); return (unary(np)); case NCCL: np = op2(NCCL, NIL, (Node *) cclenter((char *) rlxstr)); rtok = relex(); return (unary(np)); case '^': rtok = relex(); return (unary(op2(CHAR, NIL, itonp(HAT)))); case '$': rtok = relex(); return (unary(op2(CHAR, NIL, NIL))); case '(': rtok = relex(); if (rtok == ')') { /* special pleading for () */ rtok = relex(); return unary(op2(CCL, NIL, (Node *) tostring(""))); } np = regexp(); if (rtok == ')') { rtok = relex(); return (unary(np)); } else FATAL("syntax error in regular expression %s at %s", lastre, prestr); default: FATAL("illegal primary in regular expression %s at %s", lastre, prestr); } return 0; /*NOTREACHED*/ } Node *concat(Node *np) { switch (rtok) { case CHAR: case DOT: case ALL: case EMPTYRE: case CCL: case NCCL: case '$': case '(': return (concat(op2(CAT, np, primary()))); } return (np); } Node *alt(Node *np) { if (rtok == OR) { rtok = relex(); return (alt(op2(OR, np, concat(primary())))); } return (np); } Node *unary(Node *np) { switch (rtok) { case STAR: rtok = relex(); return (unary(op2(STAR, np, NIL))); case PLUS: rtok = relex(); return (unary(op2(PLUS, np, NIL))); case QUEST: rtok = relex(); return (unary(op2(QUEST, np, NIL))); default: return (np); } } /* * Character class definitions conformant to the POSIX locale as * defined in IEEE P1003.1 draft 7 of June 2001, assuming the source * and operating character sets are both ASCII (ISO646) or supersets * thereof. * * Note that to avoid overflowing the temporary buffer used in * relex(), the expanded character class (prior to range expansion) * must be less than twice the size of their full name. */ /* Because isblank doesn't show up in any of the header files on any * system i use, it's defined here. if some other locale has a richer * definition of "blank", define HAS_ISBLANK and provide your own * version. * the parentheses here are an attempt to find a path through the maze * of macro definition and/or function and/or version provided. thanks * to nelson beebe for the suggestion; let's see if it works everywhere. */ #if !defined(HAS_ISBLANK) && !defined(__APPLE__) int (xisblank)(int c) { return c==' ' || c=='\t'; } #endif struct charclass { const char *cc_name; int cc_namelen; int (*cc_func)(int); } charclasses[] = { { "alnum", 5, isalnum }, { "alpha", 5, isalpha }, #ifndef HAS_ISBLANK { "blank", 5, isspace }, /* was isblank */ #else { "blank", 5, isblank }, #endif { "cntrl", 5, iscntrl }, { "digit", 5, isdigit }, { "graph", 5, isgraph }, { "lower", 5, islower }, { "print", 5, isprint }, { "punct", 5, ispunct }, { "space", 5, isspace }, { "upper", 5, isupper }, { "xdigit", 6, isxdigit }, { NULL, 0, NULL }, }; int relex(void) /* lexical analyzer for reparse */ { int c, n; int cflag; static uschar *buf = 0; static int bufsz = 100; uschar *bp; struct charclass *cc; int i; switch (c = *prestr++) { case '|': return OR; case '*': return STAR; case '+': return PLUS; case '?': return QUEST; case '.': return DOT; case '\0': prestr--; return '\0'; case '^': case '$': case '(': case ')': return c; case '\\': rlxval = quoted(&prestr); return CHAR; default: rlxval = c; return CHAR; case '[': if (buf == 0 && (buf = (uschar *) malloc(bufsz)) == NULL) FATAL("out of space in reg expr %.10s..", lastre); bp = buf; if (*prestr == '^') { cflag = 1; prestr++; } else cflag = 0; n = 2 * strlen((const char *) prestr)+1; if (!adjbuf((char **) &buf, &bufsz, n, n, (char **) &bp, "relex1")) FATAL("out of space for reg expr %.10s...", lastre); for (; ; ) { if ((c = *prestr++) == '\\') { *bp++ = '\\'; if ((c = *prestr++) == '\0') FATAL("nonterminated character class %.20s...", lastre); *bp++ = c; /* } else if (c == '\n') { */ /* FATAL("newline in character class %.20s...", lastre); */ } else if (c == '[' && *prestr == ':') { /* POSIX char class names, Dag-Erling Smorgrav, des@@ofug.org */ for (cc = charclasses; cc->cc_name; cc++) if (strncmp((const char *) prestr + 1, (const char *) cc->cc_name, cc->cc_namelen) == 0) break; if (cc->cc_name != NULL && prestr[1 + cc->cc_namelen] == ':' && prestr[2 + cc->cc_namelen] == ']') { prestr += cc->cc_namelen + 3; for (i = 0; i < NCHARS; i++) { if (!adjbuf((char **) &buf, &bufsz, bp-buf+1, 100, (char **) &bp, "relex2")) FATAL("out of space for reg expr %.10s...", lastre); if (cc->cc_func(i)) { *bp++ = i; n++; } } } else *bp++ = c; } else if (c == '\0') { FATAL("nonterminated character class %.20s", lastre); } else if (bp == buf) { /* 1st char is special */ *bp++ = c; } else if (c == ']') { *bp++ = 0; rlxstr = (uschar *) tostring((char *) buf); if (cflag == 0) return CCL; else return NCCL; } else *bp++ = c; } } } int cgoto(fa *f, int s, int c) { int i, j, k; int *p, *q; assert(c == HAT || c < NCHARS); while (f->accept >= maxsetvec) { /* guessing here! */ maxsetvec *= 4; setvec = (int *) realloc(setvec, maxsetvec * sizeof(int)); tmpset = (int *) realloc(tmpset, maxsetvec * sizeof(int)); if (setvec == 0 || tmpset == 0) overflo("out of space in cgoto()"); } for (i = 0; i <= f->accept; i++) setvec[i] = 0; setcnt = 0; /* compute positions of gototab[s,c] into setvec */ p = f->posns[s]; for (i = 1; i <= *p; i++) { if ((k = f->re[p[i]].ltype) != FINAL) { if ((k == CHAR && c == ptoi(f->re[p[i]].lval.np)) || (k == DOT && c != 0 && c != HAT) || (k == ALL && c != 0) || (k == EMPTYRE && c != 0) || (k == CCL && member(c, (char *) f->re[p[i]].lval.up)) || (k == NCCL && !member(c, (char *) f->re[p[i]].lval.up) && c != 0 && c != HAT)) { q = f->re[p[i]].lfollow; for (j = 1; j <= *q; j++) { if (q[j] >= maxsetvec) { maxsetvec *= 4; setvec = (int *) realloc(setvec, maxsetvec * sizeof(int)); tmpset = (int *) realloc(tmpset, maxsetvec * sizeof(int)); if (setvec == 0 || tmpset == 0) overflo("cgoto overflow"); } if (setvec[q[j]] == 0) { setcnt++; setvec[q[j]] = 1; } } } } } /* determine if setvec is a previous state */ tmpset[0] = setcnt; j = 1; for (i = f->accept; i >= 0; i--) if (setvec[i]) { tmpset[j++] = i; } /* tmpset == previous state? */ for (i = 1; i <= f->curstat; i++) { p = f->posns[i]; if ((k = tmpset[0]) != p[0]) goto different; for (j = 1; j <= k; j++) if (tmpset[j] != p[j]) goto different; /* setvec is state i */ f->gototab[s][c] = i; return i; different:; } /* add tmpset to current set of states */ if (f->curstat >= NSTATES-1) { f->curstat = 2; f->reset = 1; for (i = 2; i < NSTATES; i++) xfree(f->posns[i]); } else ++(f->curstat); for (i = 0; i < NCHARS; i++) f->gototab[f->curstat][i] = 0; xfree(f->posns[f->curstat]); if ((p = (int *) calloc(1, (setcnt+1)*sizeof(int))) == NULL) overflo("out of space in cgoto"); f->posns[f->curstat] = p; f->gototab[s][c] = f->curstat; for (i = 0; i <= setcnt; i++) p[i] = tmpset[i]; if (setvec[f->accept]) f->out[f->curstat] = 1; else f->out[f->curstat] = 0; return f->curstat; } void freefa(fa *f) /* free a finite automaton */ { int i; if (f == NULL) return; for (i = 0; i <= f->curstat; i++) xfree(f->posns[i]); for (i = 0; i <= f->accept; i++) { xfree(f->re[i].lfollow); if (f->re[i].ltype == CCL || f->re[i].ltype == NCCL) xfree((f->re[i].lval.np)); } xfree(f->restr); xfree(f); } @ 1.2 log @Fix build under Mac OS X (Mountain Lion). @ text @d1 1 a1 1 /* $NetBSD: b.c,v 1.1 2006/07/14 14:23:06 jlam Exp $ */ d27 1 a27 1 /* lasciate ogne speranza, voi ch'entrate. */ d49 1 d53 1 a53 1 leaf (CCL, NCCL, CHAR, DOT, FINAL, ALL): d188 1 d213 1 d236 1 a236 1 int hexstr(char **pp) /* find and eval hex string at pp, return new p */ d250 1 a250 1 *pp = (char *) p; d256 1 a256 1 int quoted(char **pp) /* pick up next thing after a \\ */ d259 1 a259 1 char *p = *pp; d304 1 a304 1 c = quoted((char **) &p); d310 1 a310 1 c2 = quoted((char **) &p); d317 1 a317 1 if (!adjbuf((char **) &buf, &bufsz, bp-buf+2, 100, (char **) &bp, 0)) d325 1 a325 1 if (!adjbuf((char **) &buf, &bufsz, bp-buf+2, 100, (char **) &bp, 0)) d347 1 d384 1 a384 1 /* returns 1 if p matches empty string */ d389 1 d399 4 d479 1 a479 1 assert(*p < NCHARS); d510 1 a510 1 assert(*q < NCHARS); d568 1 a568 1 assert(*q < NCHARS); d615 1 a615 1 if (rtok == '\0') d617 2 a618 1 return(op2(ALL, NIL, NIL)); d642 3 d684 1 a684 1 case CHAR: case DOT: case ALL: case CCL: case NCCL: case '$': case '(': d738 1 a738 1 int (isblank)(int c) d752 3 d756 1 d793 1 a793 1 rlxval = quoted((char **) &prestr); d809 1 a809 1 if (!adjbuf((char **) &buf, &bufsz, n, n, (char **) &bp, 0)) d828 1 a828 1 if (!adjbuf((char **) &buf, &bufsz, bp-buf+1, 100, (char **) &bp, 0)) d877 1 d885 1 a885 1 tmpset = (int *) realloc(setvec, maxsetvec * sizeof(int)); @ 1.1 log @Prepare nawk for use as a bootstrap tool by including the sources directly in pkgsrc/lang/nawk/files and adjusting the package Makefile accordingly. The README.pkgsrc file includes notes on how the source files were (minimally) adjusted. @ text @d1 1 a1 1 /* $NetBSD$ */ d723 1 a723 1 #ifndef HAS_ISBLANK @
__label__pos
0.931017
Talos Vulnerability Report TALOS-2016-0165 Adobe Flash Player Infinite Recursion Arbitrary Read Access Violation June 14, 2016 Report ID CVE-2016-4132 Summary A potentially exploitable read access violation vulnerability exists in the a way Adobe Flash Player handles infinitely recursive calls. A specially crafted ActionScript code can cause a read access violation which can potentially be further abused. To trriger this vulnerability user interaction is required in that the user needs to visit a webpage with embedded malicious SWF file. Tested Versions Adobe Flash Player 21.0 (latest at the time of writing) Product URLs https://get.adobe.com/flashplayer/ Details Vulnerability exists in a way Flash Player handles recursion when calling implicit functions such as “toString” or “valueOf”. It is best illustrated by an example: public class Test extends Sprite { public function Test() { function func7(){ try { obj_URIError0.toString ( ); } catch(e:Error){} try { obj_ByteArray8.writeDouble ( Math.PI); } catch(e:Error){} return ""; } var specObj2 = { toString:func7}; var obj_URIError0:URIError; var obj_ByteArray8:ByteArray; try { obj_URIError0 = new URIError( specObj2, null ); } catch(e:Error){} func7(); } } The above code is broken down in a few steps: 1. specObj2 has its toString method overloaded with func7. ..* toString() gets called implicitly when type coercion to string is required. 2. obj_URIError0 is constructed with specObj2 as first argument ..* Constructor is public function URIError(message:String = "") 3. func7() is called directly. 4. In func7() , obj_URIError0.toString() is called first, which ends up calling specObj2.toString() implicitly. ..* Since specObj2.toString() is actually func7() this would create infinite recursion and the rest of that function should never get executed. 5. After a number of recursive calls, an upper recursion limit will get hit ..* Then second line in func7() gets executed leading to a crash 6. Flash crashes when trying to dereference Math.PI. Above example is using objects of type UriError, ByteArray, and Math, but other combinations are possible. Most combinations end up crashing as straight forward null pointer dereferences but Math constants are somewhat special. We can observe the following crash when Math.PI is being dereferenced: eax=00000000 ebx=00002000 ecx=052d73a0 edx=07f3c240 esi=05312880 edi=000505b0 eip=05fa8589 esp=000503c8 ebp=000504e0 iopl=0 nv up ei pl nz na pe nc cs=001b ss=0023 ds=0023 es=0023 fs=003b gs=0000 efl=00010206 05fa8589 f30f7e4050 movq xmm0,mmword ptr [eax+50h] ds:0023:00000050=???????????????? Note that the crash happens in the JITed code which is generated based on the actionscript bytecode. Above crash is still a null pointer dereference but what makes it slightly more interesting is the fact that varying the constant in use gives us limited control over the offset in the JITed code. For example, using Math.SQRT2 instead of Math.PI gives: eax=00000000 ebx=00002000 ecx=052b63a0 edx=07d7a240 esi=05311880 edi=000505b0 eip=05fb1577 esp=000503c8 ebp=000504e0 iopl=0 nv up ei pl nz na pe nc cs=001b ss=0023 ds=0023 es=0023 fs=003b gs=0000 efl=00010206 05fb1577 f30f7e4060 movq xmm0,mmword ptr [eax+60h] ds:0023:00000060=???????????????? So, by varying the constant, we can vary the offset in the JITed code, but unfortunately as expected, in case of Math the highest offset is with SQRT2 (0x60) which is still far too low to be of any use. Slight variations of the crashing piece of actionscript code yield slightly different results. For example: try { obj_ByteArray8.writeDouble ( Math[0x41414141]); } catch(e:Error){} Even though not strictly valid, the above code produces slightly more interesting assembly when JITed and crashes here: 00a52a2d 8b44240c mov eax,dword ptr [esp+0Ch] 00a52a31 8bcb mov ecx,ebx 00a52a33 83e1f8 and ecx,0FFFFFFF8h 00a52a36 8b11 mov edx,dword ptr [ecx] <=== CRASH 00a52a38 8b5244 mov edx,dword ptr [edx+44h] 00a52a3b 50 push eax 00a52a3c ffd2 call edx 00a52a3e 5b pop ebx 00a52a3f c20800 ret 8 Still a null pointer dereference, but this time near a call instruction and with eax fully controlled: eax=41414141 ebx=00000001 ecx=00000000 edx=07ecb240 esi=05312880 edi=000505b0 eip=00a52a36 esp=000503b0 ebp=000504e0 iopl=0 nv up ei pl zr na pe nc cs=001b ss=0023 ds=0023 es=0023 fs=003b gs=0000 efl=00010246 flash!IAEModule_IAEKernel_UnloadModule+0x1ccfa6: 00a52a36 8b11 mov edx,dword ptr [ecx] ds:0023:00000000=???????? By digging into the preceding code, we can see the following: 05fb04d8 e833c6acfa call flash!IAEModule_IAEKernel_UnloadModule+0x1f7080 (00a7cb10) [1] 05fb04dd 8bc8 mov ecx,eax 05fb04df c745c437000000 mov dword ptr [ebp-3Ch],37h 05fb04e6 8b4110 mov eax,dword ptr [ecx+10h] [2] 05fb04e9 c745c439000000 mov dword ptr [ebp-3Ch],39h 05fb04f0 b919424063 mov ecx,63404219h 05fb04f5 81f158030122 xor ecx,22010358h 05fb04fb c745c43b000000 mov dword ptr [ebp-3Ch],3Bh 05fb0502 8d4001 lea eax,[eax+1] [3] 05fb0505 83ec08 sub esp,8 05fb0508 51 push ecx 05fb0509 50 push eax [4] 05fb050a 8b8d10ffffff mov ecx,dword ptr [ebp-0F0h] 05fb0510 e80b25aafa call flash!IAEModule_IAEKernel_UnloadModule+0x1ccf90 (00a52a20) We can see that NULL being dereferenced at the time of the crash comes indirectly from the first call above and second argument (0x41414141) is put into ecx directly and pushed to the stack. The part with the xor is just an artifact of “constant blinding” JIT spray mitigation. So, in the above code, function 00a7cb10, at [1], will return a pointer in eax, which gets read at [2] (this sets eax to NULL). Then it gets incremented at [3] and pushed to the stack at [4]. Later, in the function 00a52a20 this NULL pointer is dereferenced and the process crashes due to ReadAV. That being said, the working hypothesis is that hitting a recursion limit sets the process in an irregular state, then, since the exception is inhibited by the try/catch block, code continues with execution in this exceptional state which leads to an invalid dereference when retrieving the Math object further leading to a crash. Above examples use Math constants to demonstrate limited control over the dereference offset. It could be possible that a special object could be crafted which would allow for bigger offsets turning this into an arbitrary read access violation which could be further abused. Credit Discovered by Aleksandar Nikolic of Cisco Talos. Timeline 2016-04-28 - Vendor Disclosure 2016-06-14 - Public Release
__label__pos
0.822607
Skip to main content Working with plugins Pioreactor plugins are a way to distribute code to others (or yourself), and avoid having to use the command line each time you invoke your custom code - you should be able to just use the web interface. There are two ways to distribute a plugin: 1. Adding Python files to plugins folder On each Pioreactor's Raspberry Pi is a folder at /home/pioreactor/.pioreactor/plugins. When the Pioreactor software starts, any Python files in this folder are read and loaded into memory. If you were to include an automation in here, or a background job (with the click CLI component), they would be available globally. Why would you want to distribute code this way? It's a great way to test or develop your code instead of committing to other distribution methods: short iterations times, tight feedback loop, and code runs in the production environment. Two downsides are that it's harder to distribute your code to the rest of the community, and that it doesn't have the same deployment pipeline (adding configs, etc.) caution The Pioreactor software will load and run each Python file in /plugins on each invocation of pio. Don't have long-running or blocking code without wrapping it in a function or if __name__ == "__main__". For example: # my plugin code.py #❌ don't do this import time time.sleep(100) # ✅ this is okay import time def sleep(): time.sleep(100) # ✅ also is okay import time if __name__ == "__main__": time.sleep(100) Custom background jobs Here's an example: place the following code into the file /home/pioreactor/.pioreactor/plugins/demo_job.py # -*- coding: utf-8 -*- import click from pioreactor.whoami import get_unit_name, get_latest_experiment_name from pioreactor.background_jobs.base import BackgroundJob __plugin_summary__ = "Just a demo job" __plugin_version__ = "0.0.1" __plugin_name__ = "Demo job" __plugin_author__ = "Cam Davidson-Pilon" __plugin_homepage__ = "https://docs.pioreactor.com" class DemoJob(BackgroundJob): job_name="demo_job" def __init__(self, unit, experiment, **kwargs): super().__init__(unit=unit, experiment=experiment) def on_ready(self): self.logger.debug("Hello, world!") def on_disconnect(self): self.logger.debug("Goodbye, world!") @click.command(name="demo_job", help=__plugin_summary__) def click_demo_job(): job = DemoJob( unit=get_unit_name(), experiment=get_latest_experiment_name(), ) job.block_until_disconnected() You should be able to execute the following from the command line now: pio run demo_job. Finally, in your web interface under plugins, you should see "Demo Job" installed. info How do you add this to your /pioreactors page in the UI? See here. Custom scripts If you are interested in creating a Python script to control multiple jobs, like in a previous Python scripting example, you can create a file called example_script.py in the /home/pioreactor/.pioreactor/plugins/ folder: import time import click from pioreactor.background_jobs.stirring import start_stirring from pioreactor.background_jobs.od_reading import start_od_reading from pioreactor.actions.led_intensity import led_intensity from pioreactor.background_jobs.temperature_control import start_temperature_control __plugin_summary__ = "My example script to control stirring, OD and temperature" __plugin_version__ = "0.0.1" __plugin_name__ = "Example Script" __plugin_author__ = "Cam Davidson-Pilon" __plugin_homepage__ = "https://docs.pioreactor.com" @click.command(name="my_script", help=__plugin_summary__) # the name field is used in the invocation `pio run X` def click_my_script(): led_intensity({"B": 50}) stirrer = start_stirring(target_rpm=400) od_reader = start_od_reading("90", "REF") temp_controller = start_temperature_control("thermostat", target_temperature=32) time.sleep(10) stirrer.set_target_rpm(300) stirrer.block_until_disconnected() You should be able to execute the following from the command line now: pio run my_script. (The my_script is from the @click.command line, you can change it there). info The function that click.command wraps should have it's name prepended by click_. Ex: def click_my_script is okay, but def my_script is not. info How do you add this to your /pioreactors page in the UI? See here. Custom automations Here's an example of adding a custom automation: place the following code into the file /home/pioreactor/.pioreactor/plugins/demo_automation.py # -*- coding: utf-8 -*- from pioreactor.automations.dosing.base import DosingAutomationContrib __plugin_summary__ = "A demo dosing automation" __plugin_version__ = "0.0.1" __plugin_name__ = "Demo Dosing Automation" __plugin_author__ = "Cam Davidson-Pilon" __plugin_homepage__ = "https://docs.pioreactor.com" class DemoAutomation(DosingAutomationContrib): automation_name = "demo" def __init__(self, volume, **kwargs): super().__init__(**kwargs) self.volume = volume def execute(self): self.logger("Here I would execute...") You should be able to execute the following from the command line now: pio run dosing_control --automation-name demo --volume 10 info How do you add this to your /pioreactors page in the UI? See here. 2. pip-installable plugins An alternative to placing Python files in the plugins folder is to bundle your code into a Python package that can be installed over the internet. This is the best way to ship your code to many users, and is pretty easy! We have a full guide on how to create a Pioreactor Python package.
__label__pos
0.763262
Answers Solutions by everydaycalculation.com Answers.everydaycalculation.com » Multiply fractions Multiply 25/75 with 14/63 This multiplication involving fractions can also be rephrased as "What is 25/75 of 14/63?" 25/75 × 14/63 is 2/27. Steps for multiplying fractions 1. Simply multiply the numerators and denominators separately: 2. 25/75 × 14/63 = 25 × 14/75 × 63 = 350/4725 3. After reducing the fraction, the answer is 2/27 MathStep (Works offline) Download our mobile app and learn to work with fractions in your own time: Android and iPhone/ iPad Related: © everydaycalculation.com
__label__pos
0.745562
Menu Close Settings Close Language and Page Formatting Options Chapter 17. Using a Vault to Obtain Secrets Several fields in the administration support obtaining the value of a secret from an external vault. To obtain a secret from a vault instead of entering it directly, enter the following specially crafted string into the appropriate field: ${vault.key} where you replace the key with the name of the secret as recognized by the vault. In order to prevent secrets from leaking across realms, implementations may combine the realm name with the key obtained from the vault expression. This means that the key won’t directly map to an entry in the vault, but rather be used to create the final entry name according to the algorithm used to combine it with the realm name. Currently, the secret can be obtained from the vault in the following fields: SMTP password In realm SMTP settings LDAP bind credential In LDAP settings of LDAP-based user federation. OIDC identity provider secret In Client Secret inside identity provider OpenID Connect Config To use a vault, a vault provider must be registered within Red Hat Single Sign-On. It is possible to either use a built-in provider described below or implement your own provider. See the Server Developer Guide for more information. Note There is at most one vault provider active per Red Hat Single Sign-On instance at any given time, and the vault provider in each instance within the cluster has to be configured consistently. 17.1. Kubernetes / OpenShift Files Plaintext Vault Provider Red Hat Single Sign-On supports vault implementation for Kubernetes secrets. These secrets can be mounted as data volumes, and they appear as a directory with a flat file structure, where each secret is represented by a file whose name is the secret name, and contents of that file is the secret value. The files within this directory have to be named as secret name prefixed by realm name and an underscore. All underscores within the secret name or the realm name have to be doubled in the file name. For example, for a field within a realm called sso_realm, a reference to a secret with name secret-name would be written as ${vault.secret-name}, and the file name looked up would be sso__realm_secret-name (note the underscore doubled in realm name). To use this type of secret store, you have to declare the files-plaintext vault provider in standalone.xml, and set its parameter for the directory that contains the mounted volume. The following example shows the files-plaintext provider with the directory where vault files are searched for set to standalone/configuration/vault relative to Red Hat Single Sign-On base directory: <spi name="vault"> <default-provider>files-plaintext</default-provider> <provider name="files-plaintext" enabled="true"> <properties> <property name="dir" value="${jboss.home.dir}/standalone/configuration/vault/" /> </properties> </provider> </spi> Here is the equivalent configuration using CLI commands: /subsystem=keycloak-server/spi=vault/:add /subsystem=keycloak-server/spi=vault/provider=files-plaintext/:add(enabled=true,properties={dir => "${jboss.home.dir}/standalone/configuration/vault"}) 17.2. Elytron Credential Store Vault Provider Red Hat Single Sign-On also provides support for reading secrets stored in an Elytron credential store. The elytron-cs-keystore vault provider is capable of retrieving secrets from the keystore-based implementation of the credential store, which is also the default implementation provided by Elytron. This credential store is backed by a keystore (JCEKS is the default format, but it is possible to use other formats such as PKCS12) and users can create and manage the store contents using either the elytron subsystem in WildFly/JBoss EAP, or using the elytron-tool.sh script. To use this provider, you have to declare the elytron-cs-keystore in the keycloak-server subsystem and set the location and master secret of the keystore that was created by Elytron. An example of the minimal configuration for the provider follows: <spi name="vault"> <default-provider>elytron-cs-keystore</default-provider> <provider name="elytron-cs-keystore" enabled="true"> <properties> <property name="location" value="${jboss.home.dir}/standalone/configuration/vault/credential-store.jceks" /> <property name="secret" value="secretpw1!"/> </properties> </provider> </spi> If the underlying keystore has a format other than JCEKS, this format has to be informed using the keyStoreType: <spi name="vault"> <default-provider>elytron-cs-keystore</default-provider> <provider name="elytron-cs-keystore" enabled="true"> <properties> <property name="location" value="${jboss.home.dir}/standalone/configuration/vault/credential-store.p12" /> <property name="secret" value="secretpw1!"/> <property name="keyStoreType" value="PKCS12"/> </properties> </provider> </spi> For the secret, the elytron-cs-keystore provider supports both clear-text values (as shown above) and also values that were masked using the elytron-tool.sh script: <spi name="vault"> ... <property name="secret" value="MASK-3u2HNQaMogJJ8VP7J6gRIl;12345678;321"/> ... </spi> For more detailed information on how to create/manage elytron credential stores, as well as how to mask keystore secrets, please refer to the Elytron documentation. Note The elytron-cs-keystore vault provider has been implemented as a WildFly extension and as such is only available if the Red Hat Single Sign-On server runs on WildFly/JBoss EAP. 17.3. Key Resolvers All built-in providers support the configuration of one or more key resolvers. A key resolver essentially implements the algorithm or strategy for combining the realm name with the key (as obtained from the ${vault.key} expression} into the final entry name that will be used to retrieve the secret from the vault. The keyResolvers property is used to configure the resolvers that are to be used by the provider. The value is a comma-separated list of resolver names. An example of configuration for the files-plaintext provider follows: <spi name="vault"> <default-provider>files-plaintext</default-provider> <provider name="files-plaintext" enabled="true"> <properties> <property name="dir" value="${jboss.home.dir}/standalone/configuration/vault/" /> <property name="keyResolvers" value="REALM_UNDERSCORE_KEY, KEY_ONLY"/> </properties> </provider> </spi> The resolvers are executed in the same order that they are declared in the configuration. For each resolver, the final entry name produced by the resolver that combines the realm with the vault key is used to search for the secret in the vault. If a secret is found, it is immediately returned. If not, the next resolver is used and this continues until a non-empty secret is found or all resolvers have been tried, in which case an empty secret is returned. In the example above, first the REALM_UNDERSCORE_KEY resolver is used. If an entry is found in the vault with the name it produces, it is returned. If not, then the KEY_ONLY resolver is used. Again, if an entry is found in the vault with the name it produces, it is returned. If not, an empty secret is returned since there are no more resolvers to be used. A list of the currently available resolvers follows: • KEY_ONLY: the realm name is ignored and the key from the vault expression is used as is. • REALM_UNDERSCORE_KEY: the realm and key are combined using an underscore _ character. Occurrences of underscore in either the realm or key are escaped by another underscore character. So if the realm is called master_realm and the key is smtp_key, the combined key will be master__realm_smtp__key. • REALM_FILESEPARATOR_KEY: the realm and key are combined using the platform file separator character. This is useful in situations where the keys are grouped by realm using a directory structure. If no resolver is configured for the built-in providers, the REALM_UNDERSCORE_KEY is selected by default. 17.4. Sample Configuration The following is an example of configuring a vault and credential store. The procedure involves two parts: • Creating the credential store and a vault, where the credential store and vault passwords are in plain text. • Updating the credential store and vault to have the password use a mask provided by elytron-tool.sh. In this example, the test target used is an LDAP instance with BIND DN credential: secret12. The target is mapped using user federation in the realm ldaptest. 17.4.1. Configuring the credential store and vault without a mask You create the credential store and a vault where the credential store and vault passwords are in plain text. Prerequisites • A running LDAP instance has BIND DN credential: secret12. • The alias uses the format <realm-name>_< key-value> when using the default key resolver. In this case, the instance is running in the realm ldaptest and ldaptest_ldap_secret is the alias that corresponds to the value ldap_secret in that realm. Note The resolver replaces underscore characters with double underscore characters in the realm and key names. For example, for the key ldaptest_ldap_secret, the final key will be ldaptest_ldap__secret. Procedure 1. Create the Elytron credential store. [standalone@localhost:9990 /] /subsystem=elytron/credential-store=test-store:add(create=true, location=/home/test/test-store.p12, credential-reference={clear-text=testpwd1!},implementation-properties={keyStoreType=PKCS12}) 2. Add an alias to the credential store. /subsystem=elytron/credential-store=test-store:add-alias(alias=ldaptest_ldap__secret,secret-value=secret12) Notice how the resolver causes the key ldaptest_ldap__secret to use double underscores. 3. List the aliases from the credential store to inspect the contents of the keystore that is produced by Elytron. keytool -list -keystore /home/test/test-store.p12 -storetype PKCS12 -storepass testpwd1! Keystore type: PKCS12 Keystore provider: SUN Your keystore contains 1 entries ldaptest_ldap__secret/passwordcredential/clear/, Oct 12, 2020, SecretKeyEntry, 4. Configure the vault SPI in Red Hat Single Sign-On. /subsystem=keycloak-server/spi=vault:add(default-provider=elytron-cs-keystore) /subsystem=keycloak-server/spi=vault/provider=elytron-cs-keystore:add(enabled=true, properties={location=>/home/test/test-store.p12, secret=>testpwd1!, keyStoreType=>PKCS12}) At this point, the vault and credentials store passwords are not masked. <spi name="vault"> <default-provider>elytron-cs-keystore</default-provider> <provider name="elytron-cs-keystore" enabled="true"> <properties> <property name="location" value="/home/test/test-store.p12"/> <property name="secret" value="testpwd1!"/> <property name="keyStoreType" value="PKCS12"/> </properties> </provider> </spi> <credential-stores> <credential-store name="test-store" location="/home/test/test-store.p12" create="true"> <implementation-properties> <property name="keyStoreType" value="PKCS12"/> </implementation-properties> <credential-reference clear-text="testpwd1!"/> </credential-store> </credential-stores> 5. In the LDAP provider, replace binDN credential with ${vault.ldap_secret}. 6. Test your LDAP connection. LDAP Vault LDAP Vault 17.4.2. Masking the password in the credential store and vault You can now update the credential store and vault to have passwords that use a mask provided by elytron-tool.sh. 1. Create a masked password using values for the salt and the iteration parameters: $ EAP_HOME/bin/elytron-tool.sh mask --salt SALT --iteration ITERATION_COUNT --secret PASSWORD For example: elytron-tool.sh mask --salt 12345678 --iteration 123 --secret testpwd1! MASK-3BUbFEyWu0lRAu8.fCqyUk;12345678;123 2. Update the Elytron credential store configuration to use the masked password. /subsystem=elytron/credential-store=cs-store:write-attribute(name=credential-reference.clear-text,value="MASK-3BUbFEyWu0lRAu8.fCqyUk;12345678;123") 3. Update the Red Hat Single Sign-On vault configuration to use the masked password. /subsystem=keycloak-server/spi=vault/provider=elytron-cs-keystore:remove() /subsystem=keycloak-server/spi=vault/provider=elytron-cs-keystore:add(enabled=true, properties={location=>/home/test/test-store.p12, secret=>”MASK-3BUbFEyWu0lRAu8.fCqyUk;12345678;123”, keyStoreType=>PKCS12}) The vault and credential store are now masked: <spi name="vault"> <default-provider>elytron-cs-keystore</default-provider> <provider name="elytron-cs-keystore" enabled="true"> <properties> <property name="location" value="/home/test/test-store.p12"/> <property name="secret" value="MASK-3BUbFEyWu0lRAu8.fCqyUk;12345678;123"/> <property name="keyStoreType" value="PKCS12"/> </properties> </provider> </spi> .... ..... <credential-stores> <credential-store name="test-store" location="/home/test/test-store.p12" create="true"> <implementation-properties> <property name="keyStoreType" value="PKCS12"/> </implementation-properties> <credential-reference clear-text="MASK-3BUbFEyWu0lRAu8.fCqyUk;12345678;123"/> </credential-store> </credential-stores> 4. You can now test the connection to the LDAP using ${vault.ldap_secret}. Additional resources For more information about the Elytron tool, see Using Credential Stores with Elytron Client.
__label__pos
0.848141
answersLogoWhite 0 Best Answer 177.8 User Avatar Wiki User 14y ago This answer is: User Avatar Add your answer: Earn +20 pts Q: If a sand box is 70 inches long how long would it be in cm? Write your answer... Submit Still have questions? magnify glass imp Related questions How much sand is in a sand box? To find out how much sand to put into a sand box, or how much sand is in a sand box, measure length x width x height to get a cubic volume number. Lets say you have a sand box that measures, 41 inches by 64 inches, with 12 inches deep of sand, (The child likes to dig!!) you would have yourself there, 31,488 cubic inches of sand. However, you want to know what the volume is in cubic yards. Well then, divide 31,488 cubic inches by 46,656 cubic inches per yard, and you get, some number like, 0.6748971193 cubic yards. Lets just call it 0.70 cubic yards of sand. How many cubic inches of sand can a cardboard box measureing 12 inches long by 14 inches wide by 10 inches tall contain? 1680 cubic inches The sandbox at the local elemenatry school is 60 inches wide and 100 inches long the sand in the box is 6 inches deep how many cubic inches of sand are in the sandbox? (60" x 100" x 6") = 36,000 cubic inches The base of a sandbox at a home is 6 feet long and 4 and a half feet wide If eighteen cubic feet of sand is spread evenly in the sandbox how deep will the sand be? Assuming the box was empty to begin with there will be 8 inches of sand now. What is the volume of a sand box if it is 6 feet wide and 5 foot long and 6 inches deep? that's not geometry. 6*6*5=180 in^2 The box is 8 inches long 8 inches wide and 8 inches tall.what is the surface area of the box? 384 square inches The diagonal of a rectangular box is 12 inches long One side of the box is 8 inches long What is the length of the other side? 8.9 I want to have some sand dumped in our backyard but the local sand man will only deliver 6 yards which seems like a lot of sand how big is 6 yards os sand? Think of a box about 5ft 6 inches tall, 5ft 6 inches wide and 5ft 6 inches deep filled with sand. That's about 6 cubic yards. What is the measurement of a box that is 80 linear inches long? 80 inches If a box is 18 inches thick by 28 inches wide by 30 inches long what is the linear inches? It depends on the context, but for airlines luggage, it would count as 18+28+30 = 76 inches How many yards of sand do you need to fill a sand box 16ft wide by 40ft x 6 inches deep? 12 sq yds A box was 11 inches long and 10 inches wide it volume was 1320 cubic inches what was the height of the box? 1320/(110) = 12in
__label__pos
1
CBSE Class 8CBSE Share Notifications View all notifications RD Sharma solutions for Class 8 Mathematics chapter 5 - Playing with Numbers Login Create free account       Forgot password? Chapters RD Sharma Mathematics Class 8 by R D Sharma (2019-2020 Session) Mathematics for Class 8 by R D Sharma (2019-2020 Session) - Shaalaa.com Chapter 5: Playing with Numbers Ex. 5.10Ex. 5.20Ex. 5.30 Chapter 5: Playing with Numbers Exercise 5.10 solutions [Page 5] Ex. 5.10 | Q 1 | Page 5 Without performing actual addition and division write the quotient when the sum of 69 and 96 is divided by  (i) 11 (ii) 15 Ex. 5.10 | Q 2 | Page 5 Without performing actual computations, find the quotient when 94 − 49 is divided by (i) 9 (ii) 5 Ex. 5.10 | Q 3 | Page 5 If sum of the number 985 and two other numbers obtained by arranging the digits of 985 in cyclic order is divided by 111, 22 and 37 respectively. Find the quotient in each case. Ex. 5.10 | Q 4 | Page 5 Find the quotient when the difference of 985 and 958 is divided by 9.   Chapter 5: Playing with Numbers Exercise 5.20 solutions [Page 20] Ex. 5.20 | Q 1 | Page 20 Given that the number \[\overline{{35\alpha64}}\] is divisible by 3, where α is a digit, what are the possible values of α? Ex. 5.20 | Q 2 | Page 20 If x is a digit such that the number \[\overline{{18x71}}\] is divisible by 3, find possible values of x. Ex. 5.20 | Q 3 | Page 20 If x is a digit of the number \[\overline {{66784x}}\] such that it is divisible by 9, find possible values of x. Ex. 5.20 | Q 4 | Page 20 Given that the number \[\overline{{67y19}}\] is divisible by 9, where y is a digit, what are the possible values of y? Ex. 5.20 | Q 5 | Page 20 If \[\overline{{3x2}}\] is a multiple of 11, where x is a digit, what is the value of x? Ex. 5.20 | Q 6 | Page 20 If \[\overline{{98215x2}}\] is a number with x as its tens digit such that is is divisible by 4. Find all possible values of x.   Ex. 5.20 | Q 7 | Page 20 If x denotes the digit at hundreds place of the number \[\overline{{67x19}}\] such that the number is divisible by 11. Find all possible values of x. Ex. 5.20 | Q 8 | Page 20 Find the remainder when 51439786 is divided by 3. Do this without performing actual division.   Ex. 5.20 | Q 9 | Page 20 Find the remainder when 51439786 is divided by 3. Do this without performing actual division.   Ex. 5.20 | Q 10 | Page 20 Find the remainder, without performing actual division, when 798 is divided by 11.   Ex. 5.20 | Q 11 | Page 20 Without performing actual division, find the remainder when 928174653 is divided by 11. Ex. 5.20 | Q 12.1 | Page 20 Given an example of a number which is divisible by 2 but not by 4. Ex. 5.20 | Q 12.2 | Page 20 Given an example of a number which is divisible by  3 but not by 6. Ex. 5.20 | Q 12.3 | Page 20 Given an example of a number which is divisible by 4 but not by 8. Ex. 5.20 | Q 12.4 | Page 20 Given an example of a number which is divisible by both 4 and 8 but not by 32. Ex. 5.20 | Q 13.01 | Page 20 Which of the following statement is true? If a number is divisible by 3, it must be divisible by 9. • True • False Ex. 5.20 | Q 13.02 | Page 20 Which of the following statement is true? If a number is divisible by 9, it must be divisible by 3. • True • False Ex. 5.20 | Q 13.03 | Page 20 Which of the following statement is true? If a number is divisible by 4, it must be divisible by 8. • True • False Ex. 5.20 | Q 13.04 | Page 20 Which of the following statement is true? If a number is divisible by 8, it must be divisible by 4. • True • False Ex. 5.20 | Q 13.05 | Page 20 Which of the following statement is true? A number is divisible by 18, if it is divisible by both 3 and 6. • True • False Ex. 5.20 | Q 13.06 | Page 20 Which of the following statement is true? If a number is divisible by both 9 and 10, it must be divisible by 90. • True • False Ex. 5.20 | Q 13.07 | Page 20 Which of the following statement is true? If a number exactly divides the sum of two numbers, it must exactly divide the numbers separately. • True • False Ex. 5.20 | Q 13.08 | Page 20 Which of the following statement is true? If a number divides three numbers exactly, it must divide their sum exactly. • True • False Ex. 5.20 | Q 13.09 | Page 20 Which of the following statement is true? If two numbers are co-prime, at least one of them must be a prime number. • True • False Ex. 5.20 | Q 13.1 | Page 20 Which of the following statement is true? The sum of two consecutive odd numbers is always divisible by 4. • True • False Chapter 5: Playing with Numbers Exercise 5.30 solutions [Page 30] Ex. 5.30 | Q 1 | Page 30 Solve each of the following Cryptarithms: 3 7 + A B 9 A Ex. 5.30 | Q 2 | Page 30 Solve each of the following Cryptarithm: A B +3 7 9 A Ex. 5.30 | Q 3 | Page 30 Solve each of the following Cryptarithm: A 1 + 1 B B 0 Ex. 5.30 | Q 4 | Page 30 Solve each of the following Cryptarithm: 2 A B + A B 1 B 1 8 Ex. 5.30 | Q 5 | Page 30 Solve each of the following Cryptarithm: 1 2 A + 6 A B A 0 9 Ex. 5.30 | Q 6 | Page 30 Solve each of the following Cryptarithm: A B 7 + 7 A B 9 8 A Ex. 5.30 | Q 7 | Page 30 Show that the Cryptarithm \[4 \times \overline{{AB}} = \overline{{CAB}}\] does not have any solution. Chapter 5: Playing with Numbers Ex. 5.10Ex. 5.20Ex. 5.30 RD Sharma Mathematics Class 8 by R D Sharma (2019-2020 Session) Mathematics for Class 8 by R D Sharma (2019-2020 Session) - Shaalaa.com RD Sharma solutions for Class 8 Mathematics chapter 5 - Playing with Numbers RD Sharma solutions for Class 8 Maths chapter 5 (Playing with Numbers) include all questions with solution and detail explanation. This will clear students doubts about any question and improve application skills while preparing for board exams. The detailed, step-by-step solutions will help you understand the concepts better and clear your confusions, if any. Shaalaa.com has the CBSE Mathematics for Class 8 by R D Sharma (2019-2020 Session) solutions in a manner that help students grasp basic concepts better and faster. Further, we at Shaalaa.com provide such solutions so that students can prepare for written exams. RD Sharma textbook solutions can be a core help for self-study and acts as a perfect self-help guidance for students. Concepts covered in Class 8 Mathematics chapter 5 Playing with Numbers are Tests of Divisibility - Divisibility by 5, Numbers in General Form, Games with Numbers, Letters for Digits, Tests of Divisibility - Divisibility by 10, Tests of Divisibility - Divisibility by 2, Tests of Divisibility - Divisibility by 9 and 3. Using RD Sharma Class 8 solutions Playing with Numbers exercise by students are an easy way to prepare for the exams, as they involve solutions arranged chapter-wise also page wise. The questions involved in RD Sharma Solutions are important questions that can be asked in the final exam. Maximum students of CBSE Class 8 prefer RD Sharma Textbook Solutions to score more in exam. Get the free view of chapter 5 Playing with Numbers Class 8 extra questions for Maths and can use Shaalaa.com to keep it handy for your exam preparation S View in app×
__label__pos
0.75726
[PATCH v3 1/4] Fix loop split incorrect count and probability Xionghu Luo [email protected] Mon Nov 8 06:09:35 GMT 2021 On 2021/10/27 15:44, Jan Hubicka wrote: >> On Wed, 27 Oct 2021, Jan Hubicka wrote: >> >>>> >>>> gcc/ChangeLog: >>>> >>>> * tree-ssa-loop-split.c (split_loop): Fix incorrect probability. >>>> (do_split_loop_on_cond): Likewise. >>>> --- >>>> gcc/tree-ssa-loop-split.c | 25 ++++++++++++++++--------- >>>> 1 file changed, 16 insertions(+), 9 deletions(-) >>>> >>>> diff --git a/gcc/tree-ssa-loop-split.c b/gcc/tree-ssa-loop-split.c >>>> index 3f6ad046623..d30782888f3 100644 >>>> --- a/gcc/tree-ssa-loop-split.c >>>> +++ b/gcc/tree-ssa-loop-split.c >>>> @@ -575,7 +575,11 @@ split_loop (class loop *loop1) >>>> stmts2); >>>> tree cond = build2 (guard_code, boolean_type_node, guard_init, border); >>>> if (!initial_true) >>>> - cond = fold_build1 (TRUTH_NOT_EXPR, boolean_type_node, cond); >>>> + cond = fold_build1 (TRUTH_NOT_EXPR, boolean_type_node, cond); >>>> + >>>> + edge true_edge = EDGE_SUCC (bbs[i], 0)->flags & EDGE_TRUE_VALUE >>>> + ? EDGE_SUCC (bbs[i], 0) >>>> + : EDGE_SUCC (bbs[i], 1); >>>> >>>> /* Now version the loop, placing loop2 after loop1 connecting >>>> them, and fix up SSA form for that. */ >>>> @@ -583,10 +587,10 @@ split_loop (class loop *loop1) >>>> basic_block cond_bb; >>>> >>>> class loop *loop2 = loop_version (loop1, cond, &cond_bb, >>>> - profile_probability::always (), >>>> - profile_probability::always (), >>>> - profile_probability::always (), >>>> - profile_probability::always (), >>>> + true_edge->probability, >>>> + true_edge->probability.invert (), >>>> + true_edge->probability, >>>> + true_edge->probability.invert (), >>>> true); >>> >>> As discussed yesterday, for loop of form >>> >>> for (...) >>> if (cond) >>> cond = something(); >>> else >>> something2 >>> >>> Split as >> >> Note that you are missing to conditionalize loop1 execution >> on 'cond' (not sure if that makes a difference). > You are right - forgot to mention that. > > Entry conditional makes no difference on scaling stmts inside loop but > affects its header and expected trip count. We however need to set up > probability of this conditional (and preheader count if it exists) > There is no general way to read the probability of this initial > conditional from cfg profile. So I guess we are stuck with guessing > some arbitrary value. I guess common case is that cond is true first > iteration tough and often we can easily see that fromo PHI node > initializing the test variable. > > Other thing that changes is expected number of iterations of the split > loops, so we may want to update the exit conditinal probability > accordingly... > Sorry for the late reply. The below updated patch mainly solves the issues you pointed out: - profile count proportion for both original loop and copied loop without dropping down the true branch's count; - probability update in the two loops and between the two loops; - number of iterations update/check for split_loop. [PATCH v3] Fix loop split incorrect count and probability In tree-ssa-loop-split.c, split_loop and split_loop_on_cond does two kind of split. split_loop only works for single loop and insert edge at exit when split, while split_loop_on_cond is not limited to single loop and insert edge at latch when split. Both split behavior should consider loop count and probability update. For split_loop, loop split condition is moved in front of loop1 and loop2; But split_loop_on_cond moves the condition between loop1 and loop2, this patch does: 1) profile count proportion for both original loop and copied loop without dropping down the true branch's count; 2) probability update in and between the two loops; 3) number of iterations update for split_loop. Regression tested pass, OK for master? Changes diff for split_loop and split_loop_on_cond cases: 1) diff base/loop-split.c.151t.lsplit patched/loop-split.c.152t.lsplit ... <bb 2> [local count: 118111600]: if (beg_5(D) < end_8(D)) goto <bb 14>; [89.00%] else goto <bb 6>; [11.00%] <bb 14> [local count: 105119324]: if (beg2_6(D) < c_9(D)) - goto <bb 15>; [100.00%] + goto <bb 15>; [33.00%] else - goto <bb 16>; [100.00%] + goto <bb 16>; [67.00%] - <bb 15> [local count: 105119324]: + <bb 15> [local count: 34689377]: _25 = beg_5(D) + 1; _26 = end_8(D) - beg_5(D); _27 = beg2_6(D) + _26; _28 = MIN_EXPR <c_9(D), _27>; - <bb 3> [local count: 955630225]: + <bb 3> [local count: 315357973]: # i_16 = PHI <i_11(8), beg_5(D)(15)> # j_17 = PHI <j_12(8), beg2_6(D)(15)> printf ("a: %d %d\n", i_16, j_17); i_11 = i_16 + 1; j_12 = j_17 + 1; if (j_12 < _28) - goto <bb 8>; [89.00%] + goto <bb 8>; [29.37%] else - goto <bb 17>; [11.00%] + goto <bb 17>; [70.63%] - <bb 8> [local count: 850510901]: + <bb 8> [local count: 280668596]: goto <bb 3>; [100.00%] - <bb 16> [local count: 105119324]: + <bb 16> [local count: 70429947]: # i_22 = PHI <beg_5(D)(14), i_29(17)> # j_23 = PHI <beg2_6(D)(14), j_30(17)> <bb 10> [local count: 955630225]: # i_2 = PHI <i_22(16), i_20(13)> # j_1 = PHI <j_23(16), j_21(13)> i_20 = i_2 + 1; j_21 = j_1 + 1; if (end_8(D) > i_20) - goto <bb 13>; [89.00%] + goto <bb 13>; [59.63%] else - goto <bb 9>; [11.00%] + goto <bb 9>; [40.37%] - <bb 13> [local count: 850510901]: + <bb 13> [local count: 569842305]: goto <bb 10>; [100.00%] <bb 17> [local count: 105119324]: # i_29 = PHI <i_11(3)> # j_30 = PHI <j_12(3)> if (end_8(D) > i_29) goto <bb 16>; [80.00%] else goto <bb 9>; [20.00%] <bb 9> [local count: 105119324]: <bb 6> [local count: 118111600]: return 0; } 2) diff base/loop-cond-split-1.c.151t.lsplit patched/loop-cond-split-1.c.151t.lsplit: ... <bb 2> [local count: 118111600]: if (n_7(D) > 0) goto <bb 4>; [89.00%] else goto <bb 3>; [11.00%] <bb 3> [local count: 118111600]: return; <bb 4> [local count: 105119324]: pretmp_3 = ga; - <bb 5> [local count: 955630225]: + <bb 5> [local count: 315357973]: # i_13 = PHI <i_10(20), 0(4)> # prephitmp_12 = PHI <prephitmp_5(20), pretmp_3(4)> if (prephitmp_12 != 0) goto <bb 6>; [33.00%] else goto <bb 7>; [67.00%] <bb 6> [local count: 315357972]: _2 = do_something (); ga = _2; - <bb 7> [local count: 955630225]: + <bb 7> [local count: 315357973]: # prephitmp_5 = PHI <prephitmp_12(5), _2(6)> i_10 = inc (i_13); if (n_7(D) > i_10) goto <bb 21>; [89.00%] else goto <bb 11>; [11.00%] <bb 11> [local count: 105119324]: goto <bb 3>; [100.00%] - <bb 21> [local count: 850510901]: + <bb 21> [local count: 280668596]: if (prephitmp_12 != 0) - goto <bb 20>; [100.00%] + goto <bb 20>; [33.00%] else - goto <bb 19>; [INV] + goto <bb 19>; [67.00%] - <bb 20> [local count: 850510901]: + <bb 20> [local count: 280668596]: goto <bb 5>; [100.00%] - <bb 19> [count: 0]: + <bb 19> [local count: 70429947]: # i_23 = PHI <i_10(21)> # prephitmp_25 = PHI <prephitmp_5(21)> - <bb 12> [local count: 955630225]: + <bb 12> [local count: 640272252]: # i_15 = PHI <i_23(19), i_22(16)> # prephitmp_16 = PHI <prephitmp_25(19), prephitmp_16(16)> i_22 = inc (i_15); if (n_7(D) > i_22) - goto <bb 16>; [89.00%] + goto <bb 16>; [59.63%] else - goto <bb 11>; [11.00%] + goto <bb 11>; [40.37%] - <bb 16> [local count: 850510901]: + <bb 16> [local count: 569842305]: goto <bb 12>; [100.00%] } gcc/ChangeLog: * tree-ssa-loop-split.c (split_loop): Fix incorrect profile_count and probability. (do_split_loop_on_cond): Likewise. --- gcc/tree-ssa-loop-split.c | 110 +++++++++++++++++++++++++++++++++++--- 1 file changed, 102 insertions(+), 8 deletions(-) diff --git a/gcc/tree-ssa-loop-split.c b/gcc/tree-ssa-loop-split.c index 3f6ad046623..102766241fb 100644 --- a/gcc/tree-ssa-loop-split.c +++ b/gcc/tree-ssa-loop-split.c @@ -575,7 +575,10 @@ split_loop (class loop *loop1) stmts2); tree cond = build2 (guard_code, boolean_type_node, guard_init, border); if (!initial_true) - cond = fold_build1 (TRUTH_NOT_EXPR, boolean_type_node, cond); + cond = fold_build1 (TRUTH_NOT_EXPR, boolean_type_node, cond); + + edge true_edge, false_edge; + extract_true_false_edges_from_block (bbs[i], &true_edge, &false_edge); /* Now version the loop, placing loop2 after loop1 connecting them, and fix up SSA form for that. */ @@ -583,11 +586,11 @@ split_loop (class loop *loop1) basic_block cond_bb; class loop *loop2 = loop_version (loop1, cond, &cond_bb, - profile_probability::always (), - profile_probability::always (), - profile_probability::always (), - profile_probability::always (), - true); + true_edge->probability, + true_edge->probability.invert (), + profile_probability::always (), + profile_probability::always (), + true); gcc_assert (loop2); edge new_e = connect_loops (loop1, loop2); @@ -607,6 +610,53 @@ split_loop (class loop *loop1) tree guard_next = PHI_ARG_DEF_FROM_EDGE (phi, loop_latch_edge (loop1)); patch_loop_exit (loop1, guard_stmt, guard_next, newend, initial_true); + /* Check first loop's number of iterations. */ + update_ssa (TODO_update_ssa); + gcc_assert (number_of_iterations_exit (loop1, single_exit (loop1), + &niter, false, true)); + + /* Proportion first loop's bb counts except those dominated by true + branch to avoid drop 1s down. */ + basic_block *bbs1, *bbs2; + bbs1 = get_loop_body (loop1); + unsigned j; + for (j = 0; j < loop1->num_nodes; j++) + if (bbs1[j] == loop1->latch + || !dominated_by_p (CDI_DOMINATORS, bbs1[j], true_edge->dest)) + bbs1[j]->count + = bbs1[j]->count.apply_probability (true_edge->probability); + free (bbs1); + + /* Fix first loop's exit probability after scaling. */ + edge exit_to_latch1 = single_pred_edge (loop1->latch); + exit_to_latch1->probability = exit_to_latch1->probability.apply_scale ( + true_edge->probability.to_reg_br_prob_base (), REG_BR_PROB_BASE); + single_exit (loop1)->probability + = exit_to_latch1->probability.invert (); + + /* Check second loop's number of iterations. */ + class tree_niter_desc niter2; + gcc_assert (number_of_iterations_exit (loop2, single_exit (loop2), + &niter2, false, true)); + + /* Proportion second loop's bb counts except those dominated by false + branch to avoid drop 1s down. */ + basic_block bbi_copy = get_bb_copy (false_edge->dest); + bbs2 = get_loop_body (loop2); + for (j = 0; j < loop2->num_nodes; j++) + if (bbs2[j] == loop2->latch + || !dominated_by_p (CDI_DOMINATORS, bbs2[j], bbi_copy)) + bbs2[j]->count = bbs2[j]->count.apply_probability ( + true_edge->probability.invert ()); + free (bbs2); + + /* Fix second loop's exit probability after scaling. */ + edge exit_to_latch2 = single_pred_edge (loop2->latch); + exit_to_latch2->probability = exit_to_latch2->probability.apply_scale ( + false_edge->probability.to_reg_br_prob_base (), REG_BR_PROB_BASE); + single_exit (loop2)->probability + = exit_to_latch2->probability.invert (); + /* Finally patch out the two copies of the condition to be always true/false (or opposite). */ gcond *force_true = as_a<gcond *> (last_stmt (bbs[i])); @@ -1486,8 +1536,8 @@ do_split_loop_on_cond (struct loop *loop1, edge invar_branch) initialize_original_copy_tables (); struct loop *loop2 = loop_version (loop1, boolean_true_node, NULL, - profile_probability::always (), - profile_probability::never (), + invar_branch->probability.invert (), + invar_branch->probability, profile_probability::always (), profile_probability::always (), true); @@ -1535,6 +1585,50 @@ do_split_loop_on_cond (struct loop *loop1, edge invar_branch) between loop1 and loop2. */ connect_loop_phis (loop1, loop2, to_loop2); + update_ssa (TODO_update_ssa); + + edge true_edge, false_edge, skip_edge1, skip_edge2; + extract_true_false_edges_from_block (cond_bb, &true_edge, &false_edge); + + /* Proportion first loop's bb counts except those dominated by true + branch to avoid drop 1s down. */ + skip_edge1 = true_invar ? false_edge : true_edge; + skip_edge2 = true_invar ? true_edge : false_edge; + basic_block *bbs1, *bbs2; + bbs1 = get_loop_body (loop1); + unsigned j; + for (j = 0; j < loop1->num_nodes; j++) + if (bbs1[j] == loop1->latch + || !dominated_by_p (CDI_DOMINATORS, bbs1[j], skip_edge1->dest)) + bbs1[j]->count + = bbs1[j]->count.apply_probability (skip_edge1->probability); + free (bbs1); + + /* Fix first loop's exit probability after scaling. */ + to_loop1->probability = invar_branch->probability.invert (); + to_loop2->probability = invar_branch->probability; + + /* Proportion second loop's bb counts except those dominated by false + branch to avoid drop 1s down. */ + basic_block bbi_copy = get_bb_copy (skip_edge2->dest); + bbs2 = get_loop_body (loop2); + for (j = 0; j < loop2->num_nodes; j++) + if (bbs2[j] == loop2->latch + || !dominated_by_p (CDI_DOMINATORS, bbs2[j], bbi_copy)) + bbs2[j]->count + = bbs2[j]->count.apply_probability (skip_edge2->probability); + free (bbs2); + + /* Fix second loop's exit probability after scaling. */ + edge loop2_latch_exit; + edge exit_to_latch2 = single_pred_edge (loop2->latch); + exit_to_latch2->probability = exit_to_latch2->probability.apply_scale ( + skip_edge2->probability.to_reg_br_prob_base (), REG_BR_PROB_BASE); + loop2_latch_exit = EDGE_SUCC (exit_to_latch2->src, 0) == exit_to_latch2 + ? EDGE_SUCC (exit_to_latch2->src, 1) + : EDGE_SUCC (exit_to_latch2->src, 0); + loop2_latch_exit->probability = exit_to_latch2->probability.invert (); + free_original_copy_tables (); return true; -- 2.27.0.90.geebb51ba8c More information about the Gcc-patches mailing list
__label__pos
0.772904
開始使用 Web 上的雲端存儲 Cloud Storage for Firebase 允許您上傳和共享用戶生成的內容,例如圖像和視頻,從而允許您在應用中構建富媒體內容。您的數據存儲在Google Cloud Storage 存儲桶中,這是一種具有高可用性和全局冗餘的 EB 級對象存儲解決方案。 Cloud Storage for Firebase 可讓您直接從移動設備和 Web 瀏覽器安全地上傳這些文件,輕鬆處理不穩定的網絡。 先決條件 如果您尚未安裝 Firebase JS SDK 並初始化 Firebase 創建默認 Cloud Storage 存儲分區 1. Firebase 控制台的導航窗格中,選擇“存儲” ,然後單擊“開始” 2. 查看有關使用安全規則保護 Cloud Storage 數據的消息。在開發過程中,請考慮設置公共訪問規則 3. 選擇默認 Cloud Storage 存儲分區的位置 • 此位置設置是您項目的默認 Google Cloud Platform (GCP) 資源位置。請注意,此位置將用於項目中需要位置設置的 GCP 服務,特別是Cloud Firestore數據庫和App Engine應用(如果您使用 Cloud Scheduler,則需要)。 • 如果您無法選擇位置,則您的項目已有默認 GCP 資源位置。它是在項目創建期間或設置另一個需要位置設置的服務時設置的。 如果您使用 Blaze 計劃,則可以創建多個存儲桶,每個存儲桶都有自己的位置 4. 單擊“完成” 設置公共訪問 Cloud Storage for Firebase 提供了一種聲明性規則語言,可讓您定義數據的結構方式、索引方式以及數據的讀取和寫入時間。默認情況下,對 Cloud Storage 的讀寫訪問受到限制,因此只有經過身份驗證的用戶才能讀取或寫入數據。要在不設置身份驗證的情況下開始,您可以配置公共訪問規則 這確實使雲存儲向任何人開放,甚至是不使用您的應用程序的人,因此請務必在設置身份驗證時再次限制您的雲存儲。 將您的存儲桶 URL 添加到您的應用程序 如果尚未包含,您需要將 Cloud Storage 存儲分區 URL 添加到Firebase 應用的配置對象 1. 轉到Firebase 控制台中的存儲儀表板。 2. 單擊“文件”選項卡,然後查看文件查看器的標題。 3. 將 URL 複製到剪貼板。它通常採用project-id .appspot.com形式。 4. 在應用程序中的firebaseConfig對像中,添加storageBucket屬性和您的存儲桶 URL: 網絡模塊化API import { initializeApp } from "firebase/app"; import { getStorage } from "firebase/storage"; // TODO: Replace the following with your app's Firebase project configuration // See: https://firebase.google.com/docs/web/learn-more#config-object const firebaseConfig = { // ... storageBucket: '' }; // Initialize Firebase const app = initializeApp(firebaseConfig); // Initialize Cloud Storage and get a reference to the service const storage = getStorage(app); Web 命名空間 API import firebase from "firebase/app"; import "firebase/compat/storage"; // TODO: Replace the following with your app's Firebase project configuration // See: https://firebase.google.com/docs/web/learn-more#config-object const firebaseConfig = { // ... storageBucket: '[your-storage-bucket-url]' }; // Initialize Firebase firebase.initializeApp(firebaseConfig); // Initialize Cloud Storage and get a reference to the service const storage = firebase.storage(); 您已準備好開始使用雲存儲! 下一步?了解如何創建 Cloud Storage 參考 高級設置 有一些用例需要額外的設置: • 多個地理區域使用 Cloud Storage 存儲分區 • 不同存儲類別中使用 Cloud Storage 存儲分區 • 在同一應用中對多個經過身份驗證的用戶使用 Cloud Storage 存儲桶 如果您的用戶遍布世界各地,並且希望將他們的數據存儲在他們附近,那麼第一個用例是完美的。例如,您可以在美國、歐洲和亞洲創建存儲桶來存儲這些地區用戶的數據,以減少延遲。 如果您的數據具有不同的訪問模式,第二個用例會很有幫助。例如:您可以設置一個多區域或區域存儲桶來存儲圖片或其他經常訪問的內容,以及一個近線或冷線存儲桶來存儲用戶備份或其他不經常訪問的內容。 在這兩種用例中,您都需要使用多個 Cloud Storage 存儲分區 如果您正在構建一個應用程序(例如 Google Drive),第三個用例非常有用,它允許用戶擁有多個登錄帳戶(例如,個人帳戶和工作帳戶)。您可以使用自定義 Firebase 應用實例對每個附加帳戶進行身份驗證。 使用多個 Cloud Storage 存儲桶 如果您想要使用上面提供的默認存儲桶以外的 Cloud Storage 存儲桶,或者在單個應用中使用多個 Cloud Storage 存儲桶,您可以創建一個引用您的自定義存儲桶的firebase.storage實例: Web modular API import { getApp } from "firebase/app"; import { getStorage } from "firebase/storage"; // Get a non-default Storage bucket const firebaseApp = getApp(); const storage = getStorage(firebaseApp, "gs://my-custom-bucket"); Web namespaced API // Get a non-default Storage bucket var storage = firebase.app().storage("gs://my-custom-bucket"); 使用導入的存儲桶 將現有 Cloud Storage 存儲桶導入 Firebase 時,您必須授予 Firebase 使用gsutil工具(包含在Google Cloud SDK中)訪問這些文件的能力: gsutil -m acl ch -r -u service-<project number>@gcp-sa-firebasestorage.iam.gserviceaccount.com gs://<your-cloud-storage-bucket> 您可以按照Firebase 項目簡介中的說明找到您的項目編號。 這不會影響新創建的存儲桶,因為這些存儲桶的默認訪問控制設置為允許 Firebase。這是一項臨時措施,將來會自動執行。 使用自定義 Firebase 應用程序 如果您正在使用自定義firebase.app.App構建更複雜的應用程序,則可以創建使用該應用程序初始化的firebase.storage.Storage實例: Web modular API import { getStorage } from "firebase/storage"; // Get the default bucket from a custom firebase.app.App const storage1 = getStorage(customApp); // Get a non-default bucket from a custom firebase.app.App const storage2 = getStorage(customApp, "gs://my-custom-bucket"); Web namespaced API // Get the default bucket from a custom firebase.app.App var storage = customApp.storage(); // Get a non-default bucket from a custom firebase.app.App var storage = customApp.storage("gs://my-custom-bucket"); 下一步
__label__pos
0.557367
Advertisement stephanlinke [PRTG] Get Limits Aug 17th, 2015 776 0 Never Not a member of Pastebin yet? Sign Up, it unlocks many cool features! 1. $prtghost="<your-prtg-host>" 2. $username="<your-prtg-user>"; 3. $passhash="<your-prtg-users-passhash>" # can be obtained from the accoutn configuration 4. $Sensors = ((Invoke-WebRequest -URI "http://$($prtghost)/api/table.json?content=sensors&output=json&columns=objid,device,sensor&username=$($username)&passhash=$($passhash)").Content | ConvertFrom-Json) 5. $Devices = ((Invoke-WebRequest -URI "http://$($prtghost)/api/table.json?content=devices&output=json&columns=objid,device&username=$($username)&passhash=$($passhash)").Content | ConvertFrom-Json) 6.   7. Write-Output "sep=;" 8. Write-Output "DeviceID;Device;Sensor;SensorID;Channel;ChannelID;LowerWarningLimit;LowerErrorLimit;UpperWarningLimit;UpperErrorLimit" 9. Foreach($Device in $Devices.devices){ 10.   11.     $Sensors = ((Invoke-WebRequest -URI "http://$($prtghost)/api/table.json?content=sensors&output=json&id=$($Device.objid)&columns=objid,device,sensor&username=$($username)&passhash=$($passhash)").Content | ConvertFrom-Json) 12.   13.     Foreach($Sensor in $Sensors.sensors){ 14.   15.         $Channels = ((Invoke-WebRequest -URI "http://$($prtghost)/api/table.json?content=channels&output=json&columns=name,lastvalue_,objid&id=$($sensor.objid)&username=$($username)&passhash=$($passhash)").Content | ConvertFrom-Json); 16.   17.         Foreach($Channel in $Channels.channels){ 18.             $ChannelSettings = (Invoke-WebRequest -Uri "http://$($prtghost)/controls/channeledit.htm?_hjax=true&id=$($Sensor.objid)&username=$($username)&passhash=$($passhash)&channel=$($Channel.objid)"); 19.             if(($channel.objid -eq -4) -or $Channel.name -eq "Execution Time"){ continue;} 20.             $LowerWarningLimit = $ChannelSettings.InputFields.FindById("limitminwarning_$($Channel.objid)").value 21.             $LowerErrorLimit = $ChannelSettings.InputFields.FindById("limitminerror_$($Channel.objid)").value 22.             $UpperWarningLimit = $ChannelSettings.InputFields.FindById("limitmaxwarning_$($Channel.objid)").value 23.             $UpperErrorLimit = $ChannelSettings.InputFields.FindById("limitmaxerror_$($Channel.objid)").value 24.   25.             Write-Output ("{0};{1};{2};{3};{4};{5};{6};{7};{8};{9}" -f $Device.device,$Device.objid,$Sensor.sensor,$Sensor.objid,$Channel.name,$Channel.objid,$LowerWarningLimit,$LowerErrorLimit,$UpperWarningLimit,$UpperErrorLimit) 26.         } 27.     } 28. } Advertisement Add Comment Please, Sign In to add comment Advertisement
__label__pos
0.987757
Using J2EE to Drive J2ME Applications Introducing wireless devices into an architecture comes with some challenges that are not present in a wire-based architectures. These challenges must be managed from the server as well as the device. These issues are most critical when connecting to Internet-based services over a wireless connection. Introduction Wireless technology is evolving at a rapid pace. There is a lot of talk about mobile and wireless computing and there is also a fair amount of hype. However, the one thing that is conspicuously absent from much of these discussions on mobile and wireless computing is a discussion on what these devices are connecting to. The fact is, most of the value, in terms of content and capabilities of the device, is a result of interacting with a server of some type. This is true whether we are talking about microbrowsers such as WAP and iMode, J2ME clients, or short message service (SMS) and email. Behind the scenes these devices are interacting with services that reside somewhere on a network. These services handle much of the complex details of the features offered by wireless devices. Although there are complexities that the mobile device must deal with, a well-designed wireless architecture delegates as much complexity as possible to the server. This is desirable because servers have more processing capabilities and do not have the power restrictions of mobile devices (i.e., servers don't run on batteries that are worn down by the CPU). This article examines wireless computing from the server's perspective. First, the problems of wireless computing or the public Internet are discussed. Then various models that help address these problems are provided. Introducing wireless devices into an architecture comes with some challenges that are not present in a wire-based architectures. These challenges must be managed from the server as well as the device. Although there are a number of issues to consider, the three following issues are most critical when connecting to Internet-based services over a wireless connection. • Bandwidth: The wireless networks available today are much slower than wire-line services. Connection speeds range from 2 kbps to about 20 kbps depending on the type of network and signal strength. • Service availability: Wireless networks have dead spots and limited coverage outside of heavily populated areas. What's more, connections between a device and the server can drop unexpectedly if a user moves out of coverage or steps into a dead spot by walking into an elevator, entering a stairwell and so forth. • Cost of the wireless service: Some networks offer unlimited usage, others charge by the kilobyte while others charge by length of time connected. Furthermore, users may move in and out of roaming areas while operating a device. A wireless architecture needs to take these issues into consideration in order to optimize network usage and avoid large service bills. So, How Can the Server Help? The server plays an important role in wireless device communication, being one of the receivers and senders of information that the device talks to over the network. The content provided to the device needs to be optimized for the limited bandwidth situations. If the chosen protocol is verbose or the server sends a lot of unneeded information to the device, the usability of the device can become compromised. Large amounts of data take longer to download and can be expensive if the wireless service charges by the kilobyte. The server is a participant in providing concise and appropriate units of information to the device. Another area in which the server participates is the handshaking that must occur when exchanging data between the server and a device. Whenever data is sent from the server to the device, or from the device to the server, some handshaking must take place in order to make sure that the receiving party actually receives the data before the sending party deletes the data or makes some kind of state change as a result of the transmission. If the data transmission is not successful, some type of reconciliation must take place. In some cases, this may mean resending the data. In other cases the problems may be more interesting. For example, suppose that data was sent from the device to the server. The server receives the data and returns an acknowledgement to the device indicating a successful transmission. However, the connection unexpectedly drops before the device actually receives the confirmation. In this case, when the device resends the data, the server must realize that the transmission is a duplicate and handle it accordingly. Optimizing Communication In order to optimize how the server communicates with a wireless device, it is desirable to keep the communication to a minimum. If possible, a connection between the server and device should be opened just long enough to exchange some data and then close down. Modems on mobile devices tend to consume a lot of battery power, which is a precious resource for a mobile device. Furthermore, some network providers, such as those operating on CDMA networks, tend to charge by amount of time connected. Keeping the connection open longer than necessary becomes unnecessarily expensive. Ideally, the data transferred between the device and the server is in a compact format. On this note, XML is not always the most compact format. If the messages that flow between the server and device can be relatively simple, then some type of delimited string or name-value pairs may be more appropriate. However, if the data requires or involves a lot of nested relationships (e.g., customers, their orders and their payment information), XML starts to earn its place. In keeping transactions short, applications become less vulnerable to spotty network connections. A failure on a short transaction results in resending a small amount of data. A failure on a large transaction results in resending a large amount of data. Not only does the latter consume more network resources, but the reconciliation of a large transmission failure is more complex and increases the potential for state inconsistencies between device and server to develop. The Burden of Reconciliation When a network transmission fails, steps must be taken to remedy the problem and make sure that the device and server do not become out of sync. This responsibility could be placed on either the device or the server to fix the problem, or they could share portions of the cleanup tasks. However, it is a good idea to place as much of this responsibility on the server as possible. Servers have far more processing power than mobile devices and can perform these tasks more efficiently. Furthermore, CPU cycles on a mobile device consume battery power; something a server does not need to be concerned about. The Wireless Challenge Hopefully by now you are getting an idea of the challenges faced by wireless applications. In the sections that follow, a number of communication models are introduced that help manage these problems. As with most solutions, there are pros and cons with each approach. Therefore, it is important to understand the benefits and weaknesses of each model so the appropriate model can be applied for a given situation. Model-View-Control MVC has become a cornerstone pattern of J2EE development over the last several years. This pattern has surfaced as part of the Sun's Wireless BluePrints in the whitepaper ("Designing Wireless Enterprise Applications Using Java Technology" ). This whitepaper indicates that the device handles the view and the server handles the model and the controller resides on both the device and the server to keep the two in sync. While this pattern does have its place with wireless architectures, it may not be a good fit for many wireless applications. In a world of unreliable network connections, limited bandwidth and various billing rates and plans, relying on a constant network connection may not be desirable or realistic. MVC Alternatives In order for a wireless architecture to reliably exchange data and provide services to wireless devices, other models are worth exploring. The following section discusses alternative models where the underlying goal is to exchange data in a manner that does not require a constant network connection. These models focus on bringing the device and server data into sync periodically. Each model contains mechanisms allowing the device and server to recover in the event that the connection fails during a transmission. Furthermore, since a network connection is not required to operate the application, the user interface is never dependent on a server response to display data or interact with the user. Synchronization Models Synchronization models have the explicit goal of exchanging data using a request-response pattern of network communication. State-Oriented Synchronization Model Under this model, data is sent from the server to the device, or from the device to the server based on the data coming into a particular state. For example, consider an application that handles to-do lists for employees. Assume that there is a web-based user interface available for employees to create and assign to-do items to each other. A to-do item has four states: • Created • Assigned • In-progress • Completed Each state corresponds to how the system will communicate in order to handle the lifecycle of a to-do item. The first state, "Created," indicates that the to-do item has been created (on the server), but it may remain unassigned for a while. However, once the to-do is assigned to an employee the to-do information becomes available to be transmitted to the employee's device. Once on the device, the to-do remains in an "assigned" state until the employee starts the task. Once started, the to-do item moves to the "In-Progress" state. Depending on the business, the server may want to receive an update on this state change so that the server-side user interface can reflect that the employee is working on the task. Once the to-do item is completed, the updated information is sent back to the server, along with the status change. In order to make all of this happen without loosing data, some handshaking between the device and the server is necessary. Each time the server sends data to the device and visa-versa, the data transmission must be followed by a confirmation. Since the device is responsible for initiating all requests, a request for data involves two HTTP requests. In the first request the device asks for the data. In the second request, the device tells the server whether or not it successfully received and processed the data. The following figure illustrates how the device acquires data from the server. When data is transmitted in the other direction, from the device to the server, the communication only requires a single request, since the server's reply contains the confirmation. If the transmission is successful, the device may take action on the information sent to the server, such as remove it from its data store. The figure below illustrates sending data from the device to the server, where the device sends data to the server and the server responds with a success or a failure message. The state-oriented synchronization model fits well into systems that share information based on discrete states where the data is "owned" by either the device or the server based on these states. This model is not a good fit for applications that continuously share or co-own data between the device and server. Nor is this model good a fit when discrete states, which would indicate ownership of information, are not present. Data-Oriented Synchronization Model A data-oriented model synchronizes data between the device data repository and the server data repository. One of the biggest advantages to this model is that is can be performed without having to know or understand anything about the business domain. In the previous example the synchronization mechanism requires explicit knowledge of a to-do item and its various states so that the item can be exchanged between the device and the server. Under a data-oriented synchronization model, the device and server repositories are periodically brought into sync, independent of any business processes. To do this, the device and server periodically run a synchronization process. This process sends changes made on the device to the server. The server then records these changes and sends any changes made on the server to the device. As data is exchanged, handshaking occurs to ensure the data is received on either end. However, since a data-oriented model does not contain knowledge about the business a couple of problems surface. The first problem is data scope. A server must be prepared to handle data synchronization with many devices. How is the scope of data synchronization for a single device managed? In the previous example, a to-do item is sent to an employee's device when the item is assigned. Under a data-oriented model, we need a way to scope the server's data to prevent replicating the entire enterprise onto every device. This can be done by specifying rules that make subsets of the enterprise data available only to specific devices. In the case of the to-do system, there needs to be a mechanism in place that indicates what data is made available to a particular device. The second problem that surfaces has to do with how synchronization conflicts are handled. Suppose changes were made to the same data elements on both the server and the device. Is one set of changes ignored or lost? Can both sets of changes be merged? Under the state-oriented model this is less of a concern, since either the device or the server has ownership over the data at a given point in time. Using a data-oriented model, rules need to be implemented to resolve these conflicts. At a basic level, rules such as "last write wins", "device wins" or "server wins" can be used as a starting points. However, it is unlikely that these rules will be ideal in all situations. In fact, applying blanket rules such as these can be dangerous. For example, suppose an employee makes changes to to-do item A on their device at 10 AM. Then, at 11 AM, another employee modifies to-do item A from the web interface. At noon, the device and server are synchronized. If we are using a "device wins" scenario, the 11 AM changes are lost. Under a "last write wins" or "server wins" scenario, the 10 AM changes are lost. Although these problems are inherent in any synchronization model, they are most pronounced in data-oriented synchronization. Therefore, it is necessary to have extension points to attach rules so that more complex decisions can be made to resolve these conflicts. Asynchronous Model (A.K.A Messaging) Messaging techniques, such as JMS, are well suited for the wireless paradigm. This is because JMS, and other Message-Oriented-Middleware (MOM) data exchange mechanisms, are built around the idea of exchanging data between two systems over an unreliable network. Using asynchronous messaging provides a way for the server to queue up data for the device and for the device to queue up transmissions to the server independent of a physical connection between the two systems. Under the asynchronous model, where a connection is not always present, a communication channel is opened periodically between the server and the device to allow data to be exchanged. The goal is to empty the queues of data on each system. However, if the window of data exchange is too short to empty the queues, the unsent data remains in the queue until the next transmission window. In order to guarantee data is sent reliably from one system to another, most messaging systems, such as JMS, implement some kind of handshaking and error recovery processes. If either the server or the device does not receive a certain message, the message stays in the queue and subsequent attempts will be made to resend the data. If the failures reach a threshold, either a timeout or number of attempts, an error handling path may be taken. Store and Forward JMS is not required in order to exchange data using asynchronous techniques. They key principle that an asynchronous model employs is often referred to as a store and forward mechanism. Store and forward means that before transmitting a piece of data, a copy of the message is stored in a persistent manner on the sending system. The data remains stored until a confirmation from the receiver indicates that the message was received successfully. Once this confirmation is returned, the sender may delete the persisted data. If no confirmation is returned, or an error is returned, the sender attempts to resend the data. Usually resending is limited to a certain number of attempts or there is a time-based expiration associated with the data to be sent. If these limits are exceeded, the system must deal with the error condition. Using a store and forward, asynchronous model allows applications to immediately process data for transmission. The application behaves the same whether a connection is present or not. Behind the scenes, the store-forward mechanism is queuing the data and exchanging data in a network-optimized way. This allows the systems to communicate as frequently as necessary, possible or practical given the constraints of the network. The following diagrams illustrate the path of a store-forward model. Sending a Message Receiving a Message Using asynchronous messaging for exchanging data is attractive because of the handshaking that takes place and the ability for an application to queue a message for transmission in the same manner whether or not a connection is present. The down side of this model is that asynchronous transmission may not be feasible for certain applications. For example, if a device submits a request for movie tickets and the user is waiting on the response, the server should respond as soon as possible with the results. In this case, the two systems are not casually exchanging data but are requesting certain things of one another (i.e., sending commands to each other). In this particular example, an asynchronous model is less desirable. Conclusion Wireless communication posses different challenges that are not present in browser-based and wire-line client-server applications. There are problems with network reliability, low bandwidth and cost of service that must be addressed. In order to deal with these problems, several models for communication were discussed. Different applications have different needs and therefore it is important to understand what a communication model buys and what needs to be managed. Each of the models discussed focus on moving data between a device and a server while minimizing dependencies on the network and allowing the wireless application to function while disconnected. The server plays an important role in optimizing performance and how wireless applications use network resources. Resources Sun Microsystem's Wireless Blueprints: http://java.sun.com/blueprints/guidelines/designing_wireless_enterprise_applications/main4.html Sun Microsystem's J2ME site: http://java.sun.com/j2me/ Sun's wireless developer site: http://wireless.java.sun.com MicroJava Network: http://www.microjava.com About the Author David Hemphill is the Lead Architect for Gearworks, Inc., a company that creates mobile and wireless software for field service industries. David is the co-author of Java 2 Micro Edition from Manning Publications and is a frequent speaker on J2ME at industry conferences. He can be reached at [email protected]. Dig Deeper on Mobile platforms Start the conversation Send me notifications when other members comment. By submitting you agree to receive email from TechTarget and its partners. If you reside outside of the United States, you consent to having your personal data transferred to and processed in the United States. Privacy Please create a username to comment. -ADS BY GOOGLE SearchCloudApplications SearchSoftwareQuality SearchHRSoftware SearchSAP SearchERP DevOpsAgenda Close
__label__pos
0.742511
Default Palette The default palette is an array of color values identifying the colors that can be used with a device context by default. the system associates the default palette with a context whenever an application creates a context for a device that supports color palettes. The default palette ensures that colors are available for use by an application without any further action. The default palette typically has 20 entries (colors), but the exact number of entries may vary from device to device. This number is equal to the NUMCOLORS value returned by the GetDeviceCaps function. An application can retrieve the color values for colors in the default palette by enumerating solid pens, the same technique used to discover the colors available on nonpalette devices. The colors in the default palette depend on the device. Display devices, for example, often use the 16 standard colors of the VGA display and 4 other colors defined by Windows. Print devices may use other default colors. When using the default palette, applications use color values to specify pen and text colors. If the requested color is not in the palette, The system approximates the color by using the closest color in the palette. If an application requests a solid brush color that is not in the palette, the system simulates the color by dithering with colors that are in the palette. To avoid approximations and dithering, applications can also specify pen, brush, and text colors by using color palette indexes rather than color values. A color palette index is an integer value that identifies a specific palette entry. Applications can use color palette indexes in place of color values but must use the PALETTEINDEX macro to create the indexes. Color palette indexes are only useful for devices that support color palettes. To avoid this device dependence, applications that use the same code to draw to both palette and nonpalette devices should use palette-relative color values to specify pen, brush, and text colors. These values are identical to color values except when creating solid brushes. (On palette devices, a solid brush color specified by a palette-relative color value is subject to color approximation instead of dithering.) Applications must use the PALETTERGB macro to create palette-relative color values. The system does not allow an application to change the entries in the default palette. To use colors other than those in the default palette, an application must create its own logical palette and select the palette into the device context.    
__label__pos
0.835619
Dismiss Notice Join Physics Forums Today! The friendliest, high quality science and math community on the planet! Everyone who loves science is here! Even partial derivatives of a ratio 1. Feb 12, 2009 #1 Is there a general formula for the even partial derivatives of a ratio, where both A and B are functions of f? [tex]\frac{\partial ^{(2n)}}{\partial f^{(2n)}} \left( \frac{A}{B} \right)[/tex] Thanks   2. jcsd 3. Feb 13, 2009 #2 CompuChip User Avatar Science Advisor Homework Helper I think not... if we let A' and B' denote the partial derivatives of A and B w.r.t. f, respectively, then for n = 1 you would get [tex]\frac{\partial^2}{\partial f^2} \frac{A}{B} = \frac{\partial}{\partial f} \left( \frac{A' B - A B'}{B^2} \right) = \frac{ (A'' B - A B'') B^2 - 2 B B' (A' B - A B') }{ B^4 }[/tex] which already looks really messy (and that's just n = 1). But you could try writing out some more terms and see if there is a pattern.   4. Feb 13, 2009 #3 arildno User Avatar Science Advisor Homework Helper Gold Member Dearly Missed It is fairly trivial to find an IMPLICIT equation for the n'th derivative of a fraction: Set [tex]y(x)=\frac{A(x)}{B(x)}\to{A}(x)=B(x)y(x)[/tex] Thus, we have for the n'th derivative: [tex]A^{(n)}(x)=\sum_{i=0}^{i=n}\binom{n}{i}B^{(n-i)}(x)y^{(i)}(x)[/tex] This is then readily solved for the n'th derivative for y: [tex]y^{(n)}(x)=\frac{1}{B(x)}(A^{(n)}(x)-\sum_{i=0}^{i=n-1}\binom{n}{i}B^{(n-i))(x}y^{(i)}(x))[/tex]   5. Feb 13, 2009 #4 Arildno, I have to disagree: that's not trivial, it's awesome! Thanks for your response. A couple of thoughts: 1) How did you get from generic differentiation of a product to that binomial series? Simple pattern recognition? If not, could you suggest a specific reference so that I can follow your reasoning? 2) There seems to be a tiny glitch in your final answer. Here's what I think it should be: [tex] y(x) = \frac{A(x)}{B(x)} \qquad (1)[/tex] [tex] A(x) = B(x) \cdot y(x) \qquad (2)[/tex] [tex] A^{(n)} (x) = \sum _{i = 0} ^{n} \left[ \binom{n}{i} B^{(n-i)} (x) \cdot y^{(i)}(x) \right] \qquad (3)[/tex] [tex] A^{(n)} (x) = \sum _{i = 0} ^{n-1} \left[ \binom{n-1}{i} B^{(n-1-i)} (x) \cdot y^{(i)}(x) \right] + B(x) \cdot y^{(n)}(x) \qquad (4)[/tex] [tex] y^{(n)} (x) = \frac{1}{B(x)} \left\{ A^{(n)} (x) - \sum _{i = 0} ^{n-1} \left[ \binom{n-1}{i} B^{(n-1-i)} (x) \cdot y^{(i)}(x) \right] \right\} \qquad (5)[/tex] Thanks again   6. Feb 13, 2009 #5 One more point. Unless I made a mistake, the expression for the n-th derivative of y [tex] y^{(n)} (x) = \frac{1}{B(x)} \left\{ A^{(n)} (x) - \sum _{i = 0} ^{n-1} \left[ \binom{n-1}{i} B^{(n-1-i)} (x) \cdot y^{(i)}(x) \right] \right\} \qquad (5) [/tex] does not seem to work. I tested it on MATLAB for this function [tex] y(x) = \frac{(2+4i)^x - (3-i)^x}{(6-2i)^x - (5+7i)^x} [/tex] where [tex]i = \sqrt{-1}[/tex], A is the numerator, and B is the denominator. According to the formula above, the first derivative is [tex] y^{\prime}(x) = \frac{1}{B(x)}\left( A^{\prime}(x) - B(x) \cdot y(x) \right) [/tex] Here's what I found: Code (Text): >> syms x >> syms y A B >> y = ((2+4*i)^x - (3-i)^x)/((6-2*i)^x - (5+7*i)^x); >> pretty(y)                                         x          x                              (2 + 4 I)  - (3 - I)                             -----------------------                                      x            x                             (6 - 2 I)  - (5 + 7 I) >> A = ((2+4*i)^x - (3-i)^x)   A =   (2+4*i)^x-(3-i)^x >> B = (6-2*i)^x - (5+7*i)^x   B =   (6-2*i)^x-(5+7*i)^x >> (1/B)*(diff(A,x,1) - B*y) == diff(y,x,1) ans =      0   Last edited: Feb 13, 2009 7. Feb 14, 2009 #6 I finally found the correct answer. Thank you for the hint, arildno! [tex] y= \frac{A}{B} \Longleftrightarrow A^{(0)} = B^{(0)} y^{(0)} [/tex] [tex] A^{(1)} = B^{(0)} y^{(1)} + B^{(1)} y^{(0)} [/tex] [tex] A^{(2)} = B^{(0)} y^{(2)} + 2 B^{(1)} y^{(1)} + B^{(2)} y^{(0)} [/tex] [tex] A^{(3)} = B^{(0)} y^{(3)} + 3 B^{(1)} y^{(2)} +3 B^{(2)} y^{(1)} + B^{(3)} y^{(0)} [/tex] [tex] A^{(4)} = B^{(0)} y^{(4)} + 4 B^{(1)} y^{(3)} + 6 B^{(2)} y^{(2)} + 4 B^{(3)} y^{(1)} + B^{(4)} y^{(0)} [/tex] [tex] A^{(5)} = B^{(0)} y^{(5)} + 5 B^{(1)} y^{(4)} + 10 B^{(2)} y^{(3)} + 10 B^{(3)} y^{(2)} + 5 B^{(4)} y^{(1)} + B^{(5)} y^{(0)} [/tex] These derivatives can be rephrased as follows. [tex] A^{(0)} = \binom{0}{0} B^{(0)} y^{(0)} [/tex] [tex] A^{(1)} = \binom{1}{0} B^{(0)} y^{(1)} + \binom{1}{1} B^{(1)} y^{(0)} [/tex] [tex] A^{(2)} = \binom{2}{0} B^{(0)} y^{(2)} + \binom{2}{1} B^{(1)} y^{(1)} + \binom{2}{2} B^{(2)} y^{(0)} [/tex] [tex] A^{(3)} = \binom{3}{0} B^{(0)} y^{(3)} + \binom{3}{1} B^{(1)} y^{(2)} +\binom{3}{2} B^{(2)} y^{(1)} + \binom{3}{3} B^{(3)} y^{(0)} [/tex] [tex] A^{(4)} = \binom{4}{0} B^{(0)} y^{(4)} + \binom{4}{1} B^{(1)} y^{(3)} + \binom{4}{2} B^{(2)} y^{(2)} + \binom{4}{3} B^{(3)} y^{(1)} + \binom{4}{4} B^{(4)} y^{(0)} [/tex] [tex] A^{(5)} = \binom{5}{0} B^{(0)} y^{(5)} + \binom{5}{1} B^{(1)} y^{(4)} + \binom{5}{2} B^{(2)} y^{(3)} + \binom{5}{3} B^{(3)} y^{(2)} + \binom{5}{4} B^{(4)} y^{(1)} + \binom{5}{5} B^{(5)} y^{(0)} [/tex] Therefore, [tex] A^{(n)} = \sum _{i=0} ^{n} \binom{n}{i} B^{(i)} y^{(n-i)} = B^{(0)} y^{(n)} + \sum _{i=1} ^{n} \binom{n}{i} B^{(i)} y^{(n-i)} [/tex] which means [tex] y^{(n)} = \frac{1}{B ^{(0)} } \left[ A ^{(n)} - \sum _{i=1} ^{n} \binom{n}{i} B^{(i)} y^{(n-i)} \right] \qquad \mbox{for } n \geq 0 [/tex]   Last edited: Feb 14, 2009 8. Feb 14, 2009 #7 arildno User Avatar Science Advisor Homework Helper Gold Member Dearly Missed This is dead wrong. The upper argument in the binomial coefficient IS n, not n-1!   9. Feb 18, 2009 #8 Derivatives of a ratio of products I thought I was done with this thread, but I decided to expand it to a more general case: [tex]R = \frac{A\cdot C}{B\cdot D}[/tex] where R, A, B, C, and D are functions of x. I've found an expression for the n-th derivative, but I'm not sure it correct. If anyone finds a mistake, please let me know. Thanks. Here it is: [tex] R^{\displaystyle (n)} = \frac{1}{B^{\displaystyle (0)} \cdot C^{\displaystyle (0)}} \left\{ \sum _{\displaystyle i = 0} ^{\displaystyle n} \left[ \binom{n}{i} A^{\displaystyle (i)} C^{\displaystyle (n-i)} \right] + \mathop{\sum _{\displaystyle j=0} ^{\displaystyle j=n-1}} _{\displaystyle k = 0} ^{\displaystyle k = n} \left[ \binom{n}{j,k,n-j-k} R^{\displaystyle (j)}} B ^{\displaystyle (k)}} D^{\displaystyle (n-j-k)}} \right] \right\} [/tex] [tex]\mbox{for } n-(j+k) \geq 0[/tex] [tex]\mbox{where } \binom{n}{i} = \frac{n!}{i!(n-i)!} \mbox{ and } \binom{n}{j,k,n-j-k} = \frac{n!}{j! k! (n-j-k)!} [/tex] And this is how I got it: [tex] A^{(0)} C^{(0)} = R^{(0)} B^{(0)} D^{(0)}[/tex] [tex] A^{(1)} C^{(0)} + A^{(0)} C^{(1)} = R^{(1)} B^{(0)} D^{(0)} + R^{(0)} B^{(1)} D^{(0)} + R^{(0)} B^{(0)} D^{(1)}[/tex] [tex] A^{(2)} C^{(0)} + 2 A^{(1)} C^{(1)} + A^{(0)} C^{(2)} = R^{(2)} B^{(0)} D^{(0)} + 2\left( R^{(1)} B^{(1)} D^{(0)} + R^{(1)} B^{(0)} D^{(1)} + R^{(0)} B^{(1)} D^{(1)}\right) + \ldots[/tex] [tex]\ldots R^{(0)} B^{(2)} D^{(0)} + R^{(0)} B^{(0)} D^{(2)}[/tex] [tex] A^{(3)} C^{(0)} + 3\left( A^{(2)} C^{(1)} + A^{(1)} C^{(2)} \right) + A^{(0)} C^{(3)} = R^{(3)} B^{(0)} D^{(0)} + 3 \Big( R^{(2)} B^{(1)} D^{(0)} + R^{(2)} B^{(0)} D^{(1)} + \ldots [/tex] [tex]\ldots R^{(1)} B^{(2)} D^{(0)} + R^{(1)} B^{(0)} D^{(2)} + R^{(0)} B^{(2)} D^{(1)} + R^{(0)} B^{(1)} D^{(2)} \Big) + 6 R^{(1)} B^{(1)} D^{(1)} + R^{(0)} B^{(3)} D^{(0)} + R^{(0)} B^{(0)} D^{(3)}[/tex] and so forth. Any help is highly appreciated.   Last edited: Feb 19, 2009 Know someone interested in this topic? Share this thread via Reddit, Google+, Twitter, or Facebook Similar Discussions: Even partial derivatives of a ratio 1. Partial Derivative (Replies: 3) Loading...
__label__pos
0.999515
show/hide this revision's text 1 [made Community Wiki] The game of Mafia, also marketed as Werewolf, depends in practice mostly on how skillful the players are at lying, but there are some fascinating mathematical questions that arise when tries to devise optimal strategies for expert players. Let me describe one such expert strategy to give the flavor. In what follows, I will assume basic knowledge of the (simple) rules, which you can find at the above Wikipedia link. Suppose there is a detective, who secretly learns someone's identity each night. How can the detective communicate his knowledge without exposing himself to the Mafia? Each day, each townsperson claims to be the detective, and announces the piece of information he learned the previous night. The real detective tells the truth, but the Mafia will usually not be able to distinguish the real detective from all the impersonators. Of course, the townspeople will not know either—until the detective is killed. Then the townspeople, being expert players with excellent memories, will remember everything the detective said before being killed, and will therefore get a windfall of truthful information that they can they exploit to their advantage. Many questions arise naturally. What is the probability that the townspeople win if they use this strategy? The Mafia have some extra information (they know who they are) and hence if some townsperson makes a false statement while impersonating the detective, the Mafia will detect this and know that that townsperson is not the detective. So perhaps the detective should lie occasionally to counter this strategy? How should the townspeople lie? Should they attempt to give mutually consistent stories or not? As far as I know, these strategic issues have remained largely unexplored. See also this MO question that announces a mathematical paper on the Mafia game.
__label__pos
0.541558
How to Distribute Trinomials A trinomial, a polynomial with three terms, can be distributed over another expression. Each term in the first factor is distributed separately over the second factor, and then the entire expression is simplified, combining anything that can be combined. Example: Simplify the expression, image0.png 1. Break the trinomial into its three terms x, y, and 2, and distribute each term of the trinomial over the other terms. image1.png 2. Do the multiplication. image2.png 3. Simplify and combine any like terms. image3.png blog comments powered by Disqus Advertisement Inside Dummies.com
__label__pos
0.964723
Popular topics: How many characters can I send in an SMS? How to get started with ClickSend How to send an email to text message To send an email to text message is as simple as sending a normal email. Just select your recipients, write your message and send. To get started you can sign up for a free ClickSend account here: https://www.clicksend.com/signup To get started, follow these simple steps: 1. Compose an email. 2. Set the recipient as [email protected] (eg. [email protected]), you can send to multiple recipients. 3. Write your message. You can use the subject and body of the email.  4. Press send. All the replies that you receive will automatically return to the inbox who originally sent the text message. This default can be changed in settings area. Was this article helpful? Can’t find what you’re looking for? Our award-winning customer care team is here for you. Contact Support
__label__pos
0.721839
Software Tuning, Performance Optimization & Platform Monitoring Discussion around monitoring and software tuning methodologies, Performance Monitoring Unit (PMU) of Intel microprocessors, and platform monitoring Concurrency support of CLFLUSH Steven_H_2 Beginner 321 Views As far as i know, the new CLFLUSHOPT instruction is  non-blocking while CLFLUSH is blocking. As shown in Page 20 of Ref 1. But to my surprise is that, we design a micro benchmark as shown below. When i change the OMP_NUM_THREADS from 1 to 32, I did find performance gain. I just want to know if CLFLUSH is a blocking command. And in the case of OpenMP environment, do we need to replace clflush with clflushopt if we have instruction support. inline void cflush_yh(volatile int *p) { asm volatile ("clflush (%0)" :: "r"(p)); } static __inline__ unsigned long long rdtsc(void){ unsigned long long int x; __asm__ volatile (".byte 0x0f, 0x31" : "=A"(x)); return x; } void omp_flush(int *src, int *dst, size_t n) { int i; int * addr[900000]; unsigned long long cycle, elapsed; for (i=0; i<n; i++){ dst = src ; addr = &dst; } cycle = rdtsc(); #pragma omp parallel for for (i=0; i<n; i++){ cflush_yh(addr); //mfence();//Obsolete. } elapsed = rdtsc() - cycle; printf("flush cycles: %d\n", (unsigned)elapsed/MM); } int main() { int *a, *b; int i; clock_t start,end; double time_elapsed; a = (int *) malloc(MM*sizeof(int)); b = (int *) malloc(MM*sizeof(int)); for(i = 0; i < MM; ++i){ a = i; b = MM-i; } omp_flush(a,b,MM); free(a); free(b); return 0; }   [1]http://research.cs.wisc.edu/sonar/tutorial/03-hardware.pdf 0 Kudos 1 Reply McCalpinJohn Black Belt 321 Views I am not sure that this is measuring what you think it is measuring.... It is a bit hard to tell what you are doing -- the array being flushed "addr" is private to the "omp_flush()" function, and is not ever used again after the "clflush" instruction, so the compiler does not need to actually perform either the assignment to the function or the "clflush" operation. You don't provide any examples of the timings from the function, so it is not clear whether the reported time is significant compared to the overhead of the "omp parallel for" region. You don't need to flush every array element. Cache lines are 64 Bytes on every recent Intel processor, so for 64-bit pointers you only need to flush every 8th element of addr[] to ensure that you have referenced every cache line. The term "blocking" is not appropriate for describing the ordering of CLFLUSH.   As I explained a few months ago (https://software.intel.com/en-us/forums/software-tuning-performance-optimization-platform-monitoring/topic/699950), CLFLUSH is *ordered* with respect to other CLFLUSH operations, but not with respect to CLFLUSHOPT operations to different addresses.   The *ordering* property only applies within each logical processor's instruction stream, so OpenMP parallel loops will be able to run concurrently.  Within each OpenMP thread, the CLFLUSH instructions will execute in program order, but there may be a great deal of concurrency even within that single thread.   Again (as I also explained in October), CLFLUSH instructions on dirty data are expected to be lightweight -- if the data is dirty in the cache, then no other cache in the system can hold the data, and the execution of the CLFLUSH instruction simply requires initiation of a writeback operation from the cache to memory.  Since we know that the processor can execute stores (which are strongly ordered) with significant concurrency using a single logical processor, it should certainly be able to execute the writeback portion of the store operation with at least as much concurrency. Reply
__label__pos
0.863629
Vue.js Axios.post向php发送变量,但我无法访问这些变量 内容来源于 Stack Overflow,并遵循CC BY-SA 3.0许可协议进行翻译与使用 • 回答 (1) • 关注 (0) • 查看 (344) 我是新的前端开发,我面临着一个我无法解决的问题。 我想要做的是通过调用axios.post我的vue.js文件中的方法打开一个带参数的新弹出HTML页面(我可以作为数组传递)。 我搜索了可能的解决方案,但仍无法找到问题的根源。观察Chrome控制台,我怀疑它axios.post工作正常,因为它说: data:“Array([index_id] => 1 [log_id] => 63)... ow.print(); //});”,状态:200,statusText:“OK”,标题:{...},config: {...},...} 但是,在弹出窗口中,我无法访问变量,或者它们是null 以下代码是我的vue.js中的一个函数: printTransaction: function(index){ // I have tried this but could not figure out // var formData = new FormData(); // formData.append('index_id', index); // // axios.post('/popup/popup_output_print.php', { // index_id: index, // }) // .then(function(response){ // console.info(response); // }) // .catch(function(error) { // console.error(error); // }) // and this too :( const params = { index_id: index, log_id: logId, }; axios.post('/popup/popup_output_print.php', params, { headers: { 'content-type': 'application/json', }, }) .then(function(response) { console.info(response); }) .catch(function(error) { console.error(error); }) let newWin = window.open('/popup/popup_output_print.php'); setTimeout(function() { newWin.print(); //newWin.close(); }, 2000); } 这是popup_output_print.php的第一部分 <?php require("../_/inc/init.php"); $data = json_decode(file_get_contents("php://input"), TRUE); $index_id = $data['index_id']; $log_id = $data['log_id']; print_r($data); ?> ... //Trying to print the value <?php if($index_id == null) { echo "index_id is null"; } else { echo $index_id; } ?> 然后打印弹出窗口 index_id is null 我究竟做错了什么? 更新:在评论之后,我只是尝试了var_dump,变量都是null。:( 提问于 用户回答回答于 发生的事情是axios.postwindow.open完全没有影响: let newWin = window.open('/popup/popup_output_print.php'); 这使得GET请求没有参数。 从我的观点来看,更好的方法是使用类似Vue模态的东西,即: axios(...).then( response => openVueModal(response.data)) 扫码关注云+社区 领取腾讯云代金券
__label__pos
0.901987
How to delete a node if it has no parent node .net asp.net c# html-agility-pack Question I'm using the HTML agility pack to clean up input to a WYSIWYG. This might not be the best way to do this but I'm working with developers who explode on contact with regex so it will have to suffice. My WYSIWYG content looks something like this (for example): <p></p> <p></p> <p><span><input id="textbox" type="text" /></span></p> I need to strip the empty paragraph tags. Here's how I'm doing it at the moment: HtmlNodeCollection nodes = doc.DocumentNode.SelectNodes("//p"); if (nodes == null) return; foreach (HtmlNode node in nodes) { node.InnerHtml = node.InnerHtml.Trim(); if (node.InnerHtml == string.Empty) node.ParentNode.RemoveChild(node); } However, because the HTML is not a complete document the paragraph tags do not have a parent node and RemoveChild will therefore fail since ParentNode is null. I can't find another way to remove tag though, can anyone point me at an alternate method? Accepted Answer Technically, first-level elements are children of the document root, so the following code should work: if (node.InnerHtml == String.Empty) { HtmlNode parent = node.ParentNode; if (parent == null) { parent = doc.DocumentNode; } parent.RemoveChild(node); } Popular Answer You want to remove from the collection, right? HtmlNodeCollection nodes = doc.DocumentNode.SelectNodes("//p"); if (nodes == null) return; for (int i = 0; i < nodes.Count - 1; i++) { nodes[i].InnerHtml = nodes[i].InnerHtml.Trim(); if (nodes[i].InnerHtml == string.Empty) nodes.Remove(i); } Related Licensed under: CC-BY-SA with attribution Not affiliated with Stack Overflow Licensed under: CC-BY-SA with attribution Not affiliated with Stack Overflow
__label__pos
0.997517
Skip to main content CSS Horizontal & Vertical Align Center Align Elements To horizontally center a block element (like <div>), use margin: auto; Setting the width of the element will prevent it from stretching out to the edges of its container. The element will then take up the specified width, and the remaining space will be split equally between the two margins: .center { margin: auto; width: 50%; border: 3px solid green; padding: 10px; } Example centered element note Center aligning has no effect if the width property is not set (or set to 100%). Center Align Text To just center the text inside an element, use text-align: center; .center { text-align: center; border: 3px solid green; } Example centered text Center an Image To center an image, set left and right margin to auto and make it into a block element: img { display: block; margin-left: auto; margin-right: auto; width: 40%; } Example centered image Left and Right Align Using position One method for aligning elements is to use position: absolute;: .right { position: absolute; right: 0px; width: 300px; border: 3px solid #73AD21; padding: 10px; } Example align using position note Absolute positioned elements are removed from the normal flow, and can overlap elements. Using float Another method for aligning elements is to use the float property: .right { float: right; width: 300px; border: 3px solid #73AD21; padding: 10px; } The clearfix Trick note If an element is taller than the element containing it, and it is floated, it will overflow outside of its container. You can use the "clearfix hack" to fix this (see example below). Example align with clearfix Then we can add the clearfix hack to the containing element to fix this problem: .clearfix::after { content: ""; clear: both; display: table; } Center Vertically Using padding There are many ways to center an element vertically in CSS. A simple solution is to use top and bottom padding: .center { padding: 70px 0; border: 3px solid green; } Example align vertically with padding To center both vertically and horizontally, use padding and text-align: center: .center { padding: 70px 0; border: 3px solid green; text-align: center; } Example align vertically and horizontally with padding Using line-height Another trick is to use the line-height property with a value that is equal to the height property: .center { line-height: 200px; height: 200px; border: 3px solid green; text-align: center; } /* If the text has multiple lines, add the following: */ .center p { line-height: 1.5; display: inline-block; vertical-align: middle; } Example align vertically and horizontally with line-height Using position & transform If padding and line-height are not options, another solution is to use positioning and the transform property: .center { height: 200px; position: relative; border: 3px solid green; } .center p { margin: 0; position: absolute; top: 50%; left: 50%; transform: translate(-50%, -50%); } Example align vertically and horizontally with position and transform Using Flexbox You can also use flexbox to center things. Just note that flexbox is not supported in IE10 and earlier versions: .center { display: flex; justify-content: center; align-items: center; height: 200px; border: 3px solid green; } Example align vertically and horizontally with flexbox
__label__pos
0.982884
An Eager Avocado Eager Avocado I give myself very good advice, but I very seldom follow it. How much harm can nested if's do? , The definition of a(), b(), and sw() below achieve the same effect with different implementations: if else, nested if, and switch a <- function(x) { if (x == 'A') { paste('Apple') } else if (x == 'R') { paste('Ready') } else if (x == 'N') { paste('Novel') } else if (x == 'G') { paste('Ginger') } else { paste("Bingo") } } b <- function(x) { ifelse (x == 'A', paste('Apple'), ifelse(x == 'R', paste('Ready'), ifelse(x == 'N', paste('Novel'), ifelse(x == 'G', paste('Ginger'), paste("Bingo"))))) } sw <- function(x) { switch (x, A = paste('Apple'), R = paste('Ready'), N = paste('Novel'), G = paste('Ginger'), "Bingo" ) } Timing microbenchmark(a('R'), b('R'), sw('R'), times=1000) ## Unit: microseconds ## expr min lq mean median uq max neval cld ## a("R") 2.250 2.4375 2.877913 2.5685 2.8615 15.950 1000 b ## b("R") 4.891 5.2855 6.250598 5.5870 6.1225 22.822 1000 c ## sw("R") 1.796 1.9850 2.341278 2.1080 2.3530 16.123 1000 a
__label__pos
0.999383
What follows is the transcript for a GIIDE entry, audio storytelling softward that makes audio fully interactive. I\’ve seen so many articles and posts and pontifications and prognostications and just, well HYPE, about what the Metaverse means, since Mark Zuckerberg brought the concept into the public eye last fall with the Meta name change. Linkedin has caught FIRE with all of the experts. I\’m generally speaking, hype averse – so have avoided adding my voice to the melee until now. As they say about opinions…everyone has one. And I certainly have my own – and thought, maybe, after all the initial hype has died down a bit, this would be a good time to actually say what I think it means. But first, a short history. The concept of the Metaverse has been around for quite some time now, and the word itself was coined in the book \”snow crash\” by Neal Stephenson back in 1992 (30 years ago!). Basically he envisioned the future as a virtual reality-based successor to the internet, which in some ways spatial computing definitely is. In Snow Crash, people use digital avatars of themselves to explore the online world, often as a way of escaping a dystopian reality. I have to admit – and this is HERESY in my world – I was never a fan of the book and have never actually got all the way through, as I\’m not a fan of dystopianism. But it was a huge hit, and continues to be a name that\’s often brought up. So what is this \”metaverse\” we\’re now talking about? For me it\’s not about virtual reality per se. It\’s about a bunch of technologies, working together, to bring us another dimension of interaction about and beyond what our physical bodies can do. Virtual reality – or, complete immersion in an alternative environment – is one. And yes, with avatars that are extensions of ourselves. But there\’s also augmented reality: bringing things into the real world, and overlaying them onto it. Objects, sounds, information – all an extra layer or experience that enhances the physical. I\’m personally a big fan of Microsoft\’s term \”mixed reality\” because it really describes how it\’s a spectrum of realities, that go from complete immersion to lightly overlaying the physical world. Neither of these will seamlessly integrate into our day without artificial intelligence feeding how these things work in context; no one wants 1000 things jumping out at us as we walk down the street. There needs to be a layer that acts as an intelligent mediator, gatekeeper and general personal concierge; this is table stakes in this game. AI will also facilitate intelligent conversation with a brand\’s avatar ambassador, for example, who you will be able to chat with using natural language. And finally – blockchain technology adds a transaction layer over (or under) all of these things, enabling you to credibly identify who you\’re talking to, or let you buy and own things in the virtual world. Cryptocurrencies and NFT enabled digital items, built using blockchain technology, will allow you to personalize your virtual experience with items that are personal to you, whether it\’s in a completely immersed space, or one that\’s adding data to the world you see around you., That, to me, is where the metaverse lives. Not in one destination, a \”let\’s go to the metaverse\” kind of a thing, but in the sum total of all the ways you interact with your spatial data. Everyone\’s metaverse will be their own, personal one that they create by engaging with and permissioning various parts of this experience. Just as the virtual world is intangible, so will the magic web of additional data we weave around us be something amorphous – and personal. We are so. far. away from this vision technically. So many companies are trying to put up walled gardens, and interoperability will be key to making this work. But As Tony Parisi said in his epic \”Seven rules of the metaverse\” it cannot be owned. And that\’s because there is no one entity supplying it; it\’s your choices that make it real to you. It exists, and always will exist, as a concept that describes an intangible extra dimension of experiences; and I for one love the idea that we can create a life that\’s enriched with additional connection and experiences above and beyond the physical limitations that humanity has always had. It\’s time to fly. \"\" Leave a Comment Your email address will not be published. Required fields are marked * Scroll to Top
__label__pos
0.557983
How to gzip responses from 1 endpoint in Phoenix? Background We have an old app using Phoenix 1.2 and we have dozens of endpoints. We want 1 of these endpoints to have its responses compressed via gzip (or whatever the default is that Phoenix uses). Code We have all our endpoints defined in a file like this: defmodule MyApp.Router do use MyApp.Web, :router scope "/", MyApp do get("/bananas", BananasController, :bananas) end end However, after checking specifications for Phoenix 1.2, we didn’t find a way to compress only 1 endpoint. We found something in SO, but we don’t understand how we can apply this to our app: Question How do we adapt our code to compress responses from only the /bananas endpoint? There is no way to do such, as this is feature of the adapter (Cowboy in most common example), not endpoint per se. So unless you run separate adapter for that controller then there is no way to compress responses. 1 Like maybe use send_resp and do it manually eg use :zlib.gzip(response) and add the appropriate headers? 2 Likes That’s the only way we can do it, apparently, yes. 1 Like
__label__pos
0.99756
octavia/octavia/common/data_models.py Carlos Goncalves c4faac25de Add Python 3.7 support In order to support Python 3.7, pylint has to be updated to 2.0.0 minimum. Newer versions of Pylint enforce additional checkers which can be addressed with some code refactoring rather than silently ignoring them in pylintrc; except useless-object-inheritance which is required to be silented so that we stay compatible with Python 2.x. Story: 2004073 Task: 27434 Change-Id: I52301d763797d619f195bd8a1c32bc47f1e68420 2019-05-14 17:11:22 +00:00 770 lines 30 KiB Python # Copyright (c) 2014 Rackspace # Copyright (c) 2016 Blue Box, an IBM Company # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import re import six from sqlalchemy.orm import collections from octavia.common import constants class BaseDataModel(object): def to_dict(self, calling_classes=None, recurse=False, **kwargs): """Converts a data model to a dictionary.""" calling_classes = calling_classes or [] ret = {} for attr in self.__dict__: if attr.startswith('_') or not kwargs.get(attr, True): continue value = self.__dict__[attr] if attr == 'tags': # tags is a list, it doesn't need recurse ret[attr] = value continue if recurse: if isinstance(getattr(self, attr), list): ret[attr] = [] for item in value: if isinstance(item, BaseDataModel): if type(self) not in calling_classes: ret[attr].append( item.to_dict(calling_classes=( calling_classes + [type(self)]), recurse=recurse)) else: ret[attr] = None else: ret[attr] = item elif isinstance(getattr(self, attr), BaseDataModel): if type(self) not in calling_classes: ret[attr] = value.to_dict( calling_classes=calling_classes + [type(self)], recurse=recurse) else: ret[attr] = None elif six.PY2 and isinstance(value, six.text_type): ret[attr.encode('utf8')] = value.encode('utf8') else: ret[attr] = value else: if isinstance(getattr(self, attr), (BaseDataModel, list)): ret[attr] = None else: ret[attr] = value return ret def __eq__(self, other): if isinstance(other, self.__class__): return self.to_dict() == other.to_dict() return False def __ne__(self, other): return not self.__eq__(other) @classmethod def from_dict(cls, dict): return cls(**dict) @classmethod def _name(cls): """Returns class name in a more human readable form.""" # Split the class name up by capitalized words return ' '.join(re.findall('[A-Z][^A-Z]*', cls.__name__)) def _get_unique_key(self, obj=None): """Returns a unique key for passed object for data model building.""" obj = obj or self # First handle all objects with their own ID, then handle subordinate # objects. if obj.__class__.__name__ in ['Member', 'Pool', 'LoadBalancer', 'Listener', 'Amphora', 'L7Policy', 'L7Rule']: return obj.__class__.__name__ + obj.id if obj.__class__.__name__ in ['SessionPersistence', 'HealthMonitor']: return obj.__class__.__name__ + obj.pool_id if obj.__class__.__name__ in ['ListenerStatistics']: return obj.__class__.__name__ + obj.listener_id + obj.amphora_id if obj.__class__.__name__ in ['VRRPGroup', 'Vip']: return obj.__class__.__name__ + obj.load_balancer_id if obj.__class__.__name__ in ['AmphoraHealth']: return obj.__class__.__name__ + obj.amphora_id if obj.__class__.__name__ in ['SNI']: return (obj.__class__.__name__ + obj.listener_id + obj.tls_container_id) raise NotImplementedError def _find_in_graph(self, key, _visited_nodes=None): """Locates an object with the given unique key in the current object graph and returns a reference to it. """ _visited_nodes = _visited_nodes or [] mykey = self._get_unique_key() if mykey in _visited_nodes: # Seen this node already, don't traverse further return None if mykey == key: return self _visited_nodes.append(mykey) attr_names = [attr_name for attr_name in dir(self) if not attr_name.startswith('_')] for attr_name in attr_names: attr = getattr(self, attr_name) if isinstance(attr, BaseDataModel): result = attr._find_in_graph( key, _visited_nodes=_visited_nodes) if result is not None: return result elif isinstance(attr, (collections.InstrumentedList, list)): for item in attr: if isinstance(item, BaseDataModel): result = item._find_in_graph( key, _visited_nodes=_visited_nodes) if result is not None: return result # If we are here we didn't find it. return None def update(self, update_dict): """Generic update method which works for simple, non-relational attributes. """ for key, value in update_dict.items(): setattr(self, key, value) class SessionPersistence(BaseDataModel): def __init__(self, pool_id=None, type=None, cookie_name=None, pool=None, persistence_timeout=None, persistence_granularity=None): self.pool_id = pool_id self.type = type self.cookie_name = cookie_name self.pool = pool self.persistence_timeout = persistence_timeout self.persistence_granularity = persistence_granularity def delete(self): self.pool.session_persistence = None class ListenerStatistics(BaseDataModel): def __init__(self, listener_id=None, amphora_id=None, bytes_in=0, bytes_out=0, active_connections=0, total_connections=0, request_errors=0): self.listener_id = listener_id self.amphora_id = amphora_id self.bytes_in = bytes_in self.bytes_out = bytes_out self.active_connections = active_connections self.total_connections = total_connections self.request_errors = request_errors def get_stats(self): stats = { 'bytes_in': self.bytes_in, 'bytes_out': self.bytes_out, 'active_connections': self.active_connections, 'total_connections': self.total_connections, 'request_errors': self.request_errors, } return stats def __iadd__(self, other): if isinstance(other, ListenerStatistics): self.bytes_in += other.bytes_in self.bytes_out += other.bytes_out self.request_errors += other.request_errors self.total_connections += other.total_connections return self class LoadBalancerStatistics(BaseDataModel): def __init__(self, bytes_in=0, bytes_out=0, active_connections=0, total_connections=0, request_errors=0, listeners=None): self.bytes_in = bytes_in self.bytes_out = bytes_out self.active_connections = active_connections self.total_connections = total_connections self.request_errors = request_errors self.listeners = listeners or [] def get_stats(self): stats = { 'bytes_in': self.bytes_in, 'bytes_out': self.bytes_out, 'active_connections': self.active_connections, 'total_connections': self.total_connections, 'request_errors': self.request_errors, } return stats class HealthMonitor(BaseDataModel): def __init__(self, id=None, project_id=None, pool_id=None, type=None, delay=None, timeout=None, fall_threshold=None, rise_threshold=None, http_method=None, url_path=None, expected_codes=None, enabled=None, pool=None, name=None, provisioning_status=None, operating_status=None, created_at=None, updated_at=None, tags=None, http_version=None, domain_name=None): self.id = id self.project_id = project_id self.pool_id = pool_id self.type = type self.delay = delay self.timeout = timeout self.fall_threshold = fall_threshold self.rise_threshold = rise_threshold self.http_method = http_method self.url_path = url_path self.expected_codes = expected_codes self.enabled = enabled self.pool = pool self.provisioning_status = provisioning_status self.operating_status = operating_status self.name = name self.created_at = created_at self.updated_at = updated_at self.tags = tags self.http_version = http_version self.domain_name = domain_name def delete(self): self.pool.health_monitor = None class Pool(BaseDataModel): def __init__(self, id=None, project_id=None, name=None, description=None, protocol=None, lb_algorithm=None, enabled=None, operating_status=None, members=None, health_monitor=None, session_persistence=None, load_balancer_id=None, load_balancer=None, listeners=None, l7policies=None, created_at=None, updated_at=None, provisioning_status=None, tags=None, tls_certificate_id=None, ca_tls_certificate_id=None, crl_container_id=None, tls_enabled=None): self.id = id self.project_id = project_id self.name = name self.description = description self.load_balancer_id = load_balancer_id self.load_balancer = load_balancer self.protocol = protocol self.lb_algorithm = lb_algorithm self.enabled = enabled self.operating_status = operating_status self.members = members or [] self.health_monitor = health_monitor self.session_persistence = session_persistence self.listeners = listeners or [] self.l7policies = l7policies or [] self.created_at = created_at self.updated_at = updated_at self.provisioning_status = provisioning_status self.tags = tags self.tls_certificate_id = tls_certificate_id self.ca_tls_certificate_id = ca_tls_certificate_id self.crl_container_id = crl_container_id self.tls_enabled = tls_enabled def update(self, update_dict): for key, value in update_dict.items(): if key == 'session_persistence': if value is None or value == {}: if self.session_persistence is not None: self.session_persistence.delete() elif self.session_persistence is not None: self.session_persistence.update(value) else: value.update({'pool_id': self.id}) self.session_persistence = SessionPersistence(**value) else: setattr(self, key, value) def delete(self): for listener in self.listeners: if listener.default_pool_id == self.id: listener.default_pool = None listener.default_pool_id = None for pool in listener.pools: if pool.id == self.id: listener.pools.remove(pool) break for pool in self.load_balancer.pools: if pool.id == self.id: self.load_balancer.pools.remove(pool) break for l7policy in self.l7policies: if l7policy.redirect_pool_id == self.id: # Technically this should never happen, as we block deletion # of pools in use by L7Policies at the API. However, we should # probably keep this here in case the data model gets # manipulated in some other way in the future. l7policy.action = constants.L7POLICY_ACTION_REJECT l7policy.redirect_pool = None l7policy.redirect_pool_id = None class Member(BaseDataModel): def __init__(self, id=None, project_id=None, pool_id=None, ip_address=None, protocol_port=None, weight=None, backup=None, enabled=None, subnet_id=None, operating_status=None, pool=None, created_at=None, updated_at=None, provisioning_status=None, name=None, monitor_address=None, monitor_port=None, tags=None): self.id = id self.project_id = project_id self.pool_id = pool_id self.ip_address = ip_address self.protocol_port = protocol_port self.weight = weight self.backup = backup self.enabled = enabled self.subnet_id = subnet_id self.operating_status = operating_status self.pool = pool self.created_at = created_at self.updated_at = updated_at self.provisioning_status = provisioning_status self.name = name self.monitor_address = monitor_address self.monitor_port = monitor_port self.tags = tags def delete(self): for mem in self.pool.members: if mem.id == self.id: self.pool.members.remove(mem) break class Listener(BaseDataModel): def __init__(self, id=None, project_id=None, name=None, description=None, default_pool_id=None, load_balancer_id=None, protocol=None, protocol_port=None, connection_limit=None, enabled=None, provisioning_status=None, operating_status=None, tls_certificate_id=None, stats=None, default_pool=None, load_balancer=None, sni_containers=None, peer_port=None, l7policies=None, pools=None, insert_headers=None, created_at=None, updated_at=None, timeout_client_data=None, timeout_member_connect=None, timeout_member_data=None, timeout_tcp_inspect=None, tags=None, client_ca_tls_certificate_id=None, client_authentication=None, client_crl_container_id=None): self.id = id self.project_id = project_id self.name = name self.description = description self.default_pool_id = default_pool_id self.load_balancer_id = load_balancer_id self.protocol = protocol self.protocol_port = protocol_port self.connection_limit = connection_limit self.enabled = enabled self.provisioning_status = provisioning_status self.operating_status = operating_status self.tls_certificate_id = tls_certificate_id self.stats = stats self.default_pool = default_pool self.load_balancer = load_balancer self.sni_containers = sni_containers or [] self.peer_port = peer_port self.l7policies = l7policies or [] self.insert_headers = insert_headers or {} self.pools = pools or [] self.created_at = created_at self.updated_at = updated_at self.timeout_client_data = timeout_client_data self.timeout_member_connect = timeout_member_connect self.timeout_member_data = timeout_member_data self.timeout_tcp_inspect = timeout_tcp_inspect self.tags = tags self.client_ca_tls_certificate_id = client_ca_tls_certificate_id self.client_authentication = client_authentication self.client_crl_container_id = client_crl_container_id def update(self, update_dict): for key, value in update_dict.items(): setattr(self, key, value) if key == 'default_pool_id': if self.default_pool is not None: l7_pool_ids = [p.redirect_pool_id for p in self.l7policies if p.redirect_pool_id is not None and p.l7rules and p.enabled is True] old_pool = self.default_pool if old_pool.id not in l7_pool_ids: if old_pool in self.pools: self.pools.remove(old_pool) if self in old_pool.listeners: old_pool.listeners.remove(self) if value is not None: pool = self._find_in_graph('Pool' + value) if pool not in self.pools: self.pools.append(pool) if self not in pool.listeners: pool.listeners.append(self) else: pool = None setattr(self, 'default_pool', pool) def delete(self): for listener in self.load_balancer.listeners: if listener.id == self.id: self.load_balancer.listeners.remove(listener) break for pool in self.pools: pool.listeners.remove(self) class LoadBalancer(BaseDataModel): def __init__(self, id=None, project_id=None, name=None, description=None, provisioning_status=None, operating_status=None, enabled=None, topology=None, vip=None, listeners=None, amphorae=None, pools=None, vrrp_group=None, server_group_id=None, created_at=None, updated_at=None, provider=None, tags=None, flavor_id=None): self.id = id self.project_id = project_id self.name = name self.description = description self.provisioning_status = provisioning_status self.operating_status = operating_status self.enabled = enabled self.vip = vip self.vrrp_group = vrrp_group self.topology = topology self.listeners = listeners or [] self.amphorae = amphorae or [] self.pools = pools or [] self.server_group_id = server_group_id self.created_at = created_at self.updated_at = updated_at self.provider = provider self.tags = tags or [] self.flavor_id = flavor_id def update(self, update_dict): for key, value in update_dict.items(): if key == 'vip': if self.vip is not None: self.vip.update(value) else: value.update({'load_balancer_id': self.id}) self.vip = Vip(**value) else: setattr(self, key, value) class VRRPGroup(BaseDataModel): def __init__(self, load_balancer_id=None, vrrp_group_name=None, vrrp_auth_type=None, vrrp_auth_pass=None, advert_int=None, smtp_server=None, smtp_connect_timeout=None, load_balancer=None): self.load_balancer_id = load_balancer_id self.vrrp_group_name = vrrp_group_name self.vrrp_auth_type = vrrp_auth_type self.vrrp_auth_pass = vrrp_auth_pass self.advert_int = advert_int self.load_balancer = load_balancer class Vip(BaseDataModel): def __init__(self, load_balancer_id=None, ip_address=None, subnet_id=None, network_id=None, port_id=None, load_balancer=None, qos_policy_id=None, octavia_owned=None): self.load_balancer_id = load_balancer_id self.ip_address = ip_address self.subnet_id = subnet_id self.network_id = network_id self.port_id = port_id self.load_balancer = load_balancer self.qos_policy_id = qos_policy_id self.octavia_owned = octavia_owned class SNI(BaseDataModel): def __init__(self, listener_id=None, position=None, listener=None, tls_container_id=None): self.listener_id = listener_id self.position = position self.listener = listener self.tls_container_id = tls_container_id class TLSContainer(BaseDataModel): def __init__(self, id=None, primary_cn=None, certificate=None, private_key=None, passphrase=None, intermediates=None): self.id = id self.primary_cn = primary_cn self.certificate = certificate self.private_key = private_key self.passphrase = passphrase self.intermediates = intermediates or [] class Amphora(BaseDataModel): def __init__(self, id=None, load_balancer_id=None, compute_id=None, status=None, lb_network_ip=None, vrrp_ip=None, ha_ip=None, vrrp_port_id=None, ha_port_id=None, load_balancer=None, role=None, cert_expiration=None, cert_busy=False, vrrp_interface=None, vrrp_id=None, vrrp_priority=None, cached_zone=None, created_at=None, updated_at=None, image_id=None, compute_flavor=None): self.id = id self.load_balancer_id = load_balancer_id self.compute_id = compute_id self.status = status self.lb_network_ip = lb_network_ip self.vrrp_ip = vrrp_ip self.ha_ip = ha_ip self.vrrp_port_id = vrrp_port_id self.ha_port_id = ha_port_id self.role = role self.vrrp_interface = vrrp_interface self.vrrp_id = vrrp_id self.vrrp_priority = vrrp_priority self.load_balancer = load_balancer self.cert_expiration = cert_expiration self.cert_busy = cert_busy self.cached_zone = cached_zone self.created_at = created_at self.updated_at = updated_at self.image_id = image_id self.compute_flavor = compute_flavor def delete(self): for amphora in self.load_balancer.amphorae: if amphora.id == self.id: self.load_balancer.amphorae.remove(amphora) break class AmphoraHealth(BaseDataModel): def __init__(self, amphora_id=None, last_update=None, busy=False): self.amphora_id = amphora_id self.last_update = last_update self.busy = busy class L7Rule(BaseDataModel): def __init__(self, id=None, l7policy_id=None, type=None, enabled=None, compare_type=None, key=None, value=None, l7policy=None, invert=False, provisioning_status=None, operating_status=None, project_id=None, created_at=None, updated_at=None, tags=None): self.id = id self.l7policy_id = l7policy_id self.type = type self.compare_type = compare_type self.key = key self.value = value self.l7policy = l7policy self.invert = invert self.provisioning_status = provisioning_status self.operating_status = operating_status self.project_id = project_id self.created_at = created_at self.updated_at = updated_at self.enabled = enabled self.tags = tags def delete(self): if len(self.l7policy.l7rules) == 1: # l7policy should disappear from pool and listener lists. Since # we are operating only on the data model, we can fake this by # calling the policy's delete method. self.l7policy.delete() for r in self.l7policy.l7rules: if r.id == self.id: self.l7policy.l7rules.remove(r) break class L7Policy(BaseDataModel): def __init__(self, id=None, name=None, description=None, listener_id=None, action=None, redirect_pool_id=None, redirect_url=None, position=None, listener=None, redirect_pool=None, enabled=None, l7rules=None, provisioning_status=None, operating_status=None, project_id=None, created_at=None, updated_at=None, redirect_prefix=None, tags=None, redirect_http_code=None): self.id = id self.name = name self.description = description self.listener_id = listener_id self.action = action self.redirect_pool_id = redirect_pool_id self.redirect_url = redirect_url self.position = position self.listener = listener self.redirect_pool = redirect_pool self.enabled = enabled self.l7rules = l7rules or [] self.provisioning_status = provisioning_status self.operating_status = operating_status self.project_id = project_id self.created_at = created_at self.updated_at = updated_at self.redirect_prefix = redirect_prefix self.tags = tags self.redirect_http_code = redirect_http_code def _conditionally_remove_pool_links(self, pool): """Removes links to the given pool from parent objects. Note this only happens if our listener isn't referencing the pool via its default_pool or another active l7policy's redirect_pool_id. """ if (self.listener.default_pool is not None and pool is not None and pool.id != self.listener.default_pool.id and pool in self.listener.pools): listener_l7pools = [ p.redirect_pool for p in self.listener.l7policies if p.redirect_pool is not None and p.l7rules and p.enabled is True and p.id != self.id] if pool not in listener_l7pools: self.listener.pools.remove(pool) pool.listeners.remove(self.listener) def update(self, update_dict): for key, value in update_dict.items(): if key == 'redirect_pool_id' and value is not None: self._conditionally_remove_pool_links(self.redirect_pool) self.action = constants.L7POLICY_ACTION_REDIRECT_TO_POOL self.redirect_url = None self.redirect_http_code = None pool = self._find_in_graph('Pool' + value) self.redirect_pool = pool if self.l7rules and (self.enabled is True or ( 'enabled' in update_dict.keys() and update_dict['enabled'] is True)): if pool not in self.listener.pools: self.listener.pools.append(pool) if self.listener not in pool.listeners: pool.listeners.append(self.listener) elif key == 'redirect_url' and value is not None: self.action = constants.L7POLICY_ACTION_REDIRECT_TO_URL self._conditionally_remove_pool_links(self.redirect_pool) self.redirect_pool = None self.redirect_pool_id = None elif key == 'action' and value == constants.L7POLICY_ACTION_REJECT: self.redirect_url = None self._conditionally_remove_pool_links(self.redirect_pool) self.redirect_pool = None self.redirect_pool_id = None self.redirect_http_code = None elif key == 'position': self.listener.l7policies.remove(self) self.listener.l7policies.insert(value - 1, self) elif key == 'enabled': if (value is True and self.action == constants.L7POLICY_ACTION_REDIRECT_TO_POOL and self.redirect_pool is not None and self.l7rules and self.redirect_pool not in self.listener.pools): self.listener.pools.append(self.redirect_pool) self.redirect_pool.listeners.append(self.listener) elif (value is False and self.action == constants.L7POLICY_ACTION_REDIRECT_TO_POOL and self.redirect_pool is not None): self._conditionally_remove_pool_links( self.redirect_pool) setattr(self, key, value) def delete(self): self._conditionally_remove_pool_links(self.redirect_pool) if self.redirect_pool: for p in self.redirect_pool.l7policies: if p.id == self.id: self.redirect_pool.l7policies.remove(p) for p in self.listener.l7policies: if p.id == self.id: self.listener.l7policies.remove(p) break class Quotas(BaseDataModel): def __init__(self, project_id=None, load_balancer=None, listener=None, pool=None, health_monitor=None, member=None, in_use_health_monitor=None, in_use_listener=None, in_use_load_balancer=None, in_use_member=None, in_use_pool=None): self.project_id = project_id self.health_monitor = health_monitor self.listener = listener self.load_balancer = load_balancer self.pool = pool self.member = member self.in_use_health_monitor = in_use_health_monitor self.in_use_listener = in_use_listener self.in_use_load_balancer = in_use_load_balancer self.in_use_member = in_use_member self.in_use_pool = in_use_pool class Flavor(BaseDataModel): def __init__(self, id=None, name=None, description=None, enabled=None, flavor_profile_id=None): self.id = id self.name = name self.description = description self.enabled = enabled self.flavor_profile_id = flavor_profile_id class FlavorProfile(BaseDataModel): def __init__(self, id=None, name=None, provider_name=None, flavor_data=None): self.id = id self.name = name self.provider_name = provider_name self.flavor_data = flavor_data
__label__pos
0.999644
1 estoy tratando de eliminar datos de dos tablas pero no puedo lograrlo. Tengo una tabla llamada chofer que contiene una llave primaria llamada idchofer y la cual la llevo como llave foránea a otra tabla llamada camión. Mi idea era eliminar la fila camión donde la llave foránea sea igual que la de id chofer y, eliminar a la vez, la fila chofer donde se aloje esa idchofer. Mis intentos han sido en vano por estoy acudiendo a ustedes :) Gracias de ante mano. mysql> select * from chofer; +----------+-----------+----------------+-----------+ | idchofer | nombre | segundo_nombre | apellidos | +----------+-----------+----------------+-----------+ | 1 | sad | | asd | | 2 | alejandro | sdf | aaa | +----------+-----------+----------------+-----------+ 2 rows in set (0.00 sec) mysql> select * from camion; +----------+-------+-------+--------+----------+ | idcamion | placa | marca | modelo | idchofer | +----------+-------+-------+--------+----------+ | 1 | asd | asd | asd | 1 | | 2 | sdf | sdf | sdf | 2 | +----------+-------+-------+--------+----------+ 2 rows in set (0.00 sec) mysql> delete chofer,camion -> from chofer,camion -> where chofer.idchofer=camion.idchofer -> and camion.idchofer=2; ERROR 1451 (23000): Cannot delete or update a parent row: a foreign key constraint fails (`ingsoft`.`camion`, CONSTRAINT `fk_camion_chofer1` FOREIGN KEY (`idchofer`) REFERENCES `chofer` (`idchofer`) ON DELETE NO ACTION ON UPDATE NO ACTION) 1 0 Realmente es mucho más sencillo. Si a la hora de crear la base de datos configuraste correctamente las identidades referenciales bastaría con ejecutar el comando delete, que lo que hace es borrar en la tabla que le indicas y todo lo relacionado en otras tablas relacionales. En caso de que no configurases correctamente las relaciones en la base de datos, podrías utilizar la siguiente query para actualizar. ALTER TABLE `camion` ADD CONSTRAINT `chofer` FOREIGN KEY (`idchofer`) REFERENCES `chofer` (`idchofer`) ON DELETE CASCADE; Una vez hecho eso solo tendrás que eliminar una fila de chofer y automaticamente se eliminará su camion asignado. 1 • muchas gracias, me funciono :). – alekeene el 23 jun. 20 a las 18:14 Tu Respuesta Al pulsar en “Publica tu respuesta”, muestras tu consentimiento a nuestros términos de servicio, política de privacidad y política de cookies ¿No es la respuesta que buscas? Examina otras preguntas con la etiqueta o formula tu propia pregunta.
__label__pos
0.861264
How To Get An Elasticsearch Cluster's Information In Python Introduction If you’re running an Elasticsearch cluster, you’ll occasionally need to get information about the cluster. The Elasticsearch client for Python makes this task quick and simple– all it takes is a few lines of code in Python to connect to Elasticsearch and retrieve the information you need. In this step-by-step tutorial, we’ll explain how to retrieve Elasticsearch cluster information in Python. Prerequisites Before we can attempt to get cluster information in Python, it’s important to make sure certain prerequisites are in place. For this task, there are a few key system requirements: • The server or machine must have Elasticsearch installed and running. You can see if the Elasticsearch service is running by visiting https://{YOUR_DOMAIN}:9200 or localhost:9200 in your browser. • Python (version 2 or 3) must be installed on the server. • The Python low-level Elasticsearch client library must be installed. You can install the client library using the PIP package manager for Python: 1 2 3 4 # 'pip3' is for Python 3 packages: pip3 install elasticsearch # or use 'pip' to install Python 2 packages: pip install elasticsearch NOTE: Python 3 is soon to be the standard for Python, as Python 2 is being depreciated. Import the Python module for the Elasticsearch client Once you’ve confirmed all the system requirements, you’re ready to begin your Python script. Use the Python import statement to add the library for Elasticsearch to the script or Python shell environment: 1 from elasticsearch import Elasticsearch If you encounter an import error, it’s likely that the Elasticsearch client library wasn’t installed properly, or you installed the library for the wrong version of Python. Create a new client instance for the Elasticsearch Python library You’ll need to create a client instance in order to make requests and changes to the Elasticsearch cluster and its indexes; however, the client instance has other useful methods as well. The code shown below creates a client instance that we’ll use to call the info() method: 1 elastic = Elasticsearch() You may also pass a string into the hosts array as a parameter, specifying the domain name, or IP address, of the server running Elasticsearch; however, this parameter is optional: 1 2 3 4 5 6 7 8 # domain name, or server's IP address, goes in the 'hosts' array elastic = Elasticsearch(hosts=["localhost"]) """ OR USE ONE OF THESE: elastic = Elasticsearch('localhost') elastic = Elasticsearch(hosts=["http://www.example.com/"]) elastic = Elasticsearch(hosts=["123.456.789"]) """ Call the client’s info() method to get information on an Elasticsearch cluster Now that our client instance is set up, we can make our request for Elasticsearch cluster information. The client instance has a built-in method called info() that returns information about the cluster. Calling this method returns a nested Python dictionary– in other words, a dictionary inside of a dictionary: 1 elastic.info() The outer dictionary contains keys for the cluster name, UUID, logical name, and its version dictionary. Call the Python client’s info() method, and have it return a nested dictionary with information about the cluster Screenshot of Python's IDLE getting the indices in an Elasticsearch cluster Get details about the Elasticsearch cluster’s version: The value for the version key itself is a dictionary, nested inside the outer dictionary, that can provide more detailed information about the Elasticsearch cluster. You can access the version key inside the nested dictionary returned by the client’s info() method: 1 2 3 4 5 6 7 8 # store the method call's returned dictionary in a variable info = cluster.info() # access the different keys of the nested dictionary # and print out their respective value: print ("Cluster Name:", info['cluster_name']) print ("Cluster version dictionary:", info['version']) print ("Cluster version number:", info['version']['number']) Conclusion It’s important to know how to retrieve basic information about your Elasticsearch cluster, and the Elasticsearch client for Python makes this task a simple one. Just a few lines of code in your Python script allows you to get the essential information you need about your Elasticsearch cluster, such as the cluster name, the UUID and the version of Elasticsearch that’s running. With the instructions provided in this article, you’ll have no trouble using the Elasticsearch client for Python to get information about your cluster. Pilot the ObjectRocket Platform Free! Try Fully-Managed CockroachDB, Elasticsearch, MongoDB, PostgreSQL (Beta) or Redis. Get Started Keep in the know! Subscribe to our emails and we’ll let you know what’s going on at ObjectRocket. We hate spam and make it easy to unsubscribe.
__label__pos
0.934868
Testifier Testifier - 1 year ago 84 CSS Question Create a group box around certain controls on a web form using CSS I have three controls on my web form of three drop down lists. I want to create a graphical "box" around these controls. The reason for this is that selecting these controls would be "STEP 1" of my process. So I want to put a box around these controls and call it "Step 1" How would I go about doing this with CSS? Example: box around form elements Answer Source A fieldset with a legend provides a visual and semantic grouping for form controls. You can then style this as desired with CSS. A fieldset is somewhat unique in that the legend is capable of visually interrupting the border of its parent fieldset (possible with other elements, but difficult). Example: http://jsfiddle.net/NUMcr/1/ <fieldset> <legend>Group 1</legend> <input type="text" /> <asp:Textbox runat="Server" id="txt1" /> <!-- etc --> </fieldset> Including CSS in case link to fiddle is ever broken: FIELDSET { margin: 8px; border: 1px solid silver; padding: 8px; border-radius: 4px; } LEGEND{ padding: 2px; } Recommended from our users: Dynamic Network Monitoring from WhatsUp Gold from IPSwitch. Free Download
__label__pos
0.976234
top button Flag Notify Connect to us Facebook Login Site Registration Why to Join Get Free Puzzle Updates Facebook Login Site Registration A bat and ball cost Rs. 1.10 and the difference between them is Rs. 1. How much does each cost? +1 vote 960 views A bat and ball cost Rs. 1.10 and the difference between them is Rs. 1. How much does each cost? posted Apr 11, 2014 by Karan Joglekar Share this puzzle Facebook Share Button Twitter Share Button Google+ Share Button LinkedIn Share Button Multiple Social Share Button 4 Solutions +3 votes Eq1: Bat + Ball = Rs 1.1 Eq2: Bat - Ball = Rs 1 Adding Eq1 + Eq2 2 Bat = Rs 2.1 i.e. Bat is Rs1.05 which implies ball is Rs .05 solution Apr 22, 2014 by Shatark Bajpai +2 votes x+y=1.10------>eqn1 x-y=1.0------->eqn2 2x=2.10------>by summing up x=1.05---->cost of bat y=.05----->cost of ball solution May 26, 2014 by Joy Dutta +1 vote Let, x = Cost of Bat y = Cost of Ball 1. x + y = 1.10 2. x - y = 1.00 2y = 0.10 Therefore , y = 0.10/2 = 0.05 x = 1.01-0.05 = 1.05 Now , Bat (x)= 1.05 Ball(y) = 0.05 solution Mar 10, 2015 by Varghese Anthony 0 votes bat rs 1, ball 0.10 solution Oct 4, 2015 by Sasikumar Arumugam Similar Puzzles 0 votes The difference between simple interest and compound interest on a sum for 2 years at 8%, when the interest is compounded annually is Rs.16. If the interest were compounded half yearly, then what would be the difference between simple interest and compound interest for the same sum in 2 years? 0 votes Pete'sa Pete lists prices for his extra toppings in pairs. How much does each topping cost by itself? Each price is whole-dollar increments. Pineapple & Hot Peppers $7 Broccoli & Extra Cheese $6 Mushrooms & Onions $2 Black Olives & Hot Peppers $5 Pepperoni & Mushrooms $4 Onions & Black Olives $3 Extra Cheese & Mushrooms $5 +1 vote The cost of apple and grapes are between Rs 20 and 30 per kg. If total cost of 5 Kg apple and 7 Kg grapes is a number X having unit's and hundred's place equal and sum of cost of 1 Kg apple and 1 Kg grapes is Rs 50. Then what is the value of X? 0 votes If difference between simple interest and the compound interest on a certain amount of money for 3 years at 10% per annum is Rs. 46.50. Then what is the value of sum? +1 vote If difference between the compounded interest and the simple interest on a certain sum at 10% per annum for 2 years is Rs.52, then what would be the the sum? Contact Us +91 9880187415 [email protected] [email protected] #280, 3rd floor, 5th Main 6th Sector, HSR Layout Bangalore-560102 Karnataka INDIA. QUERY HOME ...
__label__pos
0.964454
Take the 2-minute tour × Stack Overflow is a question and answer site for professional and enthusiast programmers. It's 100% free, no registration required. In java, I want to do 102%100 and get the remainder as 02 and not as 2.. How do i do that? int a = (102%100); a returns 2 and not 02.. I want to get 02.. because originally, it's supposed to be 1.02, but mod returns 2. :( Anybody know how to do that? share|improve this question 1   You're confusing numbers with String representation of numbers -- a subtle but important difference. –  Hovercraft Full Of Eels Dec 12 '11 at 3:36 4 Answers 4 2 and 02 are the same number. Since you are using this to find the first two numbers after the decimal-point, just pad any one-digit numbers with a 0. If you want all the numbers after the decimal-point, the more usual way is to do this: //Strip all numbers before decimal-point double decimalDigits = myNumber - (int)myNumber; share|improve this answer      is there no automatic way to do that? T_T –  황현정 Dec 12 '11 at 2:56      @황현정: If you are only doing this for ease of display, then yes, there is; in fact, you can automatically format decimals without all this hoopla. Check out the string formatter classes. –  BlueRaja - Danny Pflughoeft Dec 12 '11 at 2:58      Yes, NumberFormat. I'll add a code snippet –  Chip McCormick Dec 12 '11 at 3:01 NumberFormat is probably the best way to go. String.format and doing calculations also work, but since you're formatting a number, might as well use a number formatter. An example. I'm guessing on your requirements here a bit (e.g. what should 110%100 return, or do you only expect a single-digit remainder?): NumberFormat formatter = new DecimalFormat("'0'#"); int remainder = 102%100; System.out.println(formatter.format(remainder)); share|improve this answer As integers, they represent the same number. Use the following to format the number as a string for display purposes: int a = 102 / 100; int r = 102 % 100; System.out.print(a + "." + String.format("%02d", r)); Output: 1.02 share|improve this answer Here lies the answer to padding numbers with zeroes: Add leading zeroes in Java Really though, thats it. share|improve this answer Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.92136
Windows下sbt安装配置 版权声明:转载原创博文请注明转载地址 https://blog.csdn.net/u014532217/article/details/78966807 1.下载sbt 2.配置环境 #SBT_HOME设置为sbt解压目录,例如: SBT_HOME=C:\Programs\sbt; #在path中添加: Path=%SBT_HOME%\bin; 3.sbt配置 在sbt\conf\sbtconfig.txt中末尾添加 -Dsbt.boot.directory=C:/Programs/sbt/data/.sbt/boot -Dsbt.global.base=C:/Programs/sbt/data/.sbt -Dsbt.ivy.home=C:/Programs/sbt/data/.ivy2 -Dsbt.repository.config=C:/Programs/sbt/conf/repo.properties -Dsbt.repository.secure=false # 设置代理 # -Dhttp.proxyHost=10.18.11.11 # -Dhttp.proxyPort=8080 # -Dhttp.proxyUser=xx # -Dhttp.proxyPassword=xx # -Dhttps.proxyHost=10.18.1111 # -Dhttps.proxyPort=8080 # -Dhttps.proxyUser=xx # -Dhttps.proxyPassword=xx 如果有需要设置代理的话,将“#设置代理”下面的#号去掉,并添加相应的代理设置,#是注释。 4.设置阿里云镜像 国内的网络环境复杂,阿里云还挺好用的,在sbt\conf\下新建repo.properties文件,内容为: [repositories] local aliyun: http://maven.aliyun.com/nexus/content/groups/public/ typesafe: http://repo.typesafe.com/typesafe/ivy-releases/, [organization]/[module]/(scala_[scalaVersion]/)(sbt_[sbtVersion]/)[revision]/[type]s/[artifact](-[classifier]).[ext], bootOnly sonatype-oss-releases maven-central sonatype-oss-snapshots 5.验证 打开cmd,输入sbt,如果出现类似下面的结果则表明安装成功: C:\Users\PetterChen>sbt "C:\Users\PetterChen\.sbt\preloaded\org.scala-sbt\sbt\"1.0.3"\jars\sbt.jar" Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=256m; support was removed in 8.0 [info] Loading project definition from C:\Users\PetterChen\project [info] Set current project to petterchen (in build file:/C:/Users/PetterChen/) [info] sbt server started at 127.0.0.1:4729 sbt:petterchen> 注意,第一次运行的时候,sbt需要下载一些东西,下载消耗时间看网络情况状态。sbt版本不同的话,最后的结果也可能不一样,但是只要配置和网络正常,最后都会进入sbt的控制台。 阅读更多 想对作者说点什么? 博主推荐 换一批 没有更多推荐了,返回首页
__label__pos
0.751253
Will I Am Will I Am - 1 year ago 57 Scala Question Sequence of actor ask calls Suppose I have the following actor: class A extends Actor { def receive = { case Unlock => sender ! UnlockResponse (Boolean_code) case Read => sender ! ReadResponse (somedata) case Lock => sender ! LockResponse (Boolean_code) } } I would like to write.a helper method where I can call unlock, read, and lock sequentially, and return the "somedata" from the ReadResponse to the method caller after last step (lock) has completed. I like the for-comprehension approach, but how do I write it with a bit more flexibility, for example to continue to Read only if Unlock returns UnlockResponse(true) in the first step, but abort if the return is UnlockResponse(false)? Answer Source For the sequential logic you could use Scala Promises api: val promise = Promise[ReadResponse]() val unlockFuture = anActorRef ? Unlock val readFuture = anActorRef ? Read unlockFuture.map{ case UnlockResponse (true) => promise.completeWith(readFuture) case UnlockResponse (false) => p.failure(new IllegalStateException) } and then depending on the outcome you can do something like p.future.onComplete { case Success(somedata) => { processData(somedata) anActorRef ! Lock } case Failure(t) => logger.error(t) } A more elegant solution with the same sequential effect could be achieved via the Actor become/unbecome and stash, where the transaction logic is handled by the actor itself instead of the sender.
__label__pos
0.985081
Teradata SQL Error and Failure Codes from Error 3612 To 3626 SQLServerF1_Header_Small 3612 Incorrect use of a subquery %VSTR. Explanation: A subquery occurred where an expression was expected. Generated By: OPT modules. For Whom: End User. Remedy: Correct the statement and resubmit request. 3613 Dump/Restore, no hashed nonfallback tables found. Explanation: Selected AMP Dump or Restore with the all tables option found no qualified tables. Generated By: OPT modules. For Whom: Dump and Restore Program. Remedy: Correct the statement and resubmit the request. 3614 Statement permitted only during FastLoad or MLoad. Explanation: CHECKPOINT LOADING or END LOADING was issued outside of a Fast Load or MLoad Session. Generated By: OPT modules. For Whom: Host Utility Fast Load or MLoad. Remedy: Correct the statement and resubmit the request. 3615 Statement not permitted during Fast Load. Explanation: Only BT, CHECKPOINT LOADING, END LOADING, ET, and INSERT are permitted during Fast Load. Generated By: OPT modules. For Whom: Host Utility Fast Load. Remedy: Correct the statement and resubmit the request. 3616 Wrong table referenced for INSERT during Fast Load. Explanation: INSERT referenced a table other than that specified by BEGIN LOADING during a Fast Load session. Generated By: OPT modules. For Whom: Host Utility Fast Load. Remedy: Correct the statement and resubmit the request. 3617 FORMAT ’%VSTR’ does not match the datatype. Explanation: The user tried to use a numeric/graphic format with a character type, a character/graphic format with a numeric type, or character/numeric format with a graphic type. The mismatch may be within a single data descriptor as in (CHAR(3), FORMAT’999’) or may be the result of applying a FORMAT clause without an explicit type declaration to an expression which has an inappropriate type, as in (Year – 1900)(FORMAT ’XX’). Note that the FORMAT displayed in the error message may not show in the users source statement if the FORMAT was an attribute of a VIEW definition. Generated By: SYN and OPT modules. For Whom: End User. Notes: “x(CHAR(3), FORMAT ’999’)” is invalid. If this is a type conversion, the user probably wanted “(x(FORMAT ’999’))(CHAR(3))”. Remedy: Re-specify the format. 3618 Expression not allowed in Fast Load Insert, column %VSTR. Explanation: FastLoad does not permit expressions in the INSERT during Fast Load. Generated By: OPT modules. For Whom: Host Utility Fast Load. Remedy: Correct the statement and resubmit the request. 3619 Only one level of data type conversion allowed, column %VSTR. Explanation: FastLoad does not permit USING data types, which require more than one level of conversion. Generated By: OPT modules. For Whom: Host Utility Fast Load. Remedy: Correct the statement and resubmit the request. 3620 Cannot create nonfallback table ’%VSTR’ with AMP down. Explanation: The user tried to create a nonfallback table while an AMP is down. Generated By: OPT modules. For Whom: End User. Remedy: Wait until all AMPs are up. 3621 Cannot load table %TVMID unless secondary indexes and join indexes are removed. Explanation: BEGIN LOADING referenced a table that has secondary indexes. They must be removed and, if desired, created after the table is loaded. Generated By: OPT modules. For Whom: Host Utility Fast Load. Remedy: DROP secondary indexes and join indexes on table to be loaded. 3622 The user cannot perform that operation on DATEs. Explanation: An operation on a DATE data type does not make sense, for example, trying to SUM a column of DATEs. Generated By: OPT modules. For Whom: End User. Remedy: Rephrase the request and try again. 3623 The user cannot use COMPRESS on a primary index column or partitioning expression column. Explanation: A primary index column or partitioning expression column specifies the COMPRESS option. Generated By: GEN modules. For Whom: End User. Remedy: Rephrase the request and try again. 3624 There are no statistics defined for the table. Explanation: There are no statistics defined for the table referenced in a COLLECT STATISTICS or HELP STATISTICS statement. Generated By: OPT modules. For Whom: End User. Notes: This error may be caused by a misspelled name of a table. Remedy: Resubmit the statement with a valid table name. 3625 GROUP BY and WITH…BY clauses may not contain aggregate functions. Explanation: The given SELECT operation contains a GROUP BY, or a WITH…BY clause that uses aggregate function references. This is not legal. Generated By: OPT modules. For Whom: End User. Remedy: Resubmit statement with a valid GROUP BY or BY clause. 3626 Internal error: WITH BY clause contains multiple tables. Explanation: The given SELECT operation contains a WITH…BY grouping clause that contains multiple table references. This is an internal check for the case — this should not be possible. Generated By: OPT modules. For Whom: System Support Representative. Remedy: Save a test case that causes the failure and contact your support representative. Above are list of Teradata Errors or Failure Codes from Error 3612 To 3626 received while performing certain operation against Teradata Database or related products. What are Teradata Database Errors? In general, each Teradata error message contains the following information: • The message number. • The message text. This text is usually returned with the message number. Some messages employ word substitution, where the word substituted represents the system-replacement of a term more specific to the occurrence. • An explanation of how the error may have occurred. • Generated-by text that indicates the software module which initiated the message. This field serves a diagnostic purpose for support and development personnel. • A remedy which suggests how to resolve the condition. Hope this was helpful. Thanks, SQLServerF1 Team Information about Teradata SQL Error and Failure Codes and Error Messages on Windows, Linux Operating Systems.   Leave a Reply Your email address will not be published. Required fields are marked *
__label__pos
0.503102
Copy Character Appearance Hello! I want to create a statue of a player from a leaderboard but I don’t know how it can load the appearance of that player in a dummy. Is there a way for a dummy (statue) to copy the appearance of a player through a script? You can do game:GetService("Players"):CreateHumanoidModelFromUserId() 3 Likes Maybe you can use this plugin: https://www.roblox.com/library/752585459/Load-Character-Lite?Category=Plugins&SortType=Bestselling&SortAggregation=AllTime&SearchKeyword=+load+character&CreatorId=0&Page=1&Position=6&SearchId=90dce16e-bb62-4248-997e-8cc81efb4e92 2 Likes I think this should help you copy a character appearance. 2 Likes Or just get the player from the leaderboard and clone it then resize and anchore all the part where the statue should stand 1 Like
__label__pos
0.814586
Actual source code: fwd.c petsc-3.5.3 2015-01-31 Report Typos and Errors 2: /* 3: Code for manipulating files. 4: */ 5: #include <petscsys.h> 6: #if defined(PETSC_HAVE_PWD_H) 7: #include <pwd.h> 8: #endif 9: #include <ctype.h> 10: #include <sys/stat.h> 11: #if defined(PETSC_HAVE_UNISTD_H) 12: #include <unistd.h> 13: #endif 14: #if defined(PETSC_HAVE_SYS_UTSNAME_H) 15: #include <sys/utsname.h> 16: #endif 17: #if defined(PETSC_HAVE_DIRECT_H) 18: #include <direct.h> 19: #endif 20: #if defined(PETSC_HAVE_SYS_SYSTEMINFO_H) 21: #include <sys/systeminfo.h> 22: #endif 26: /*@C 27: PetscGetWorkingDirectory - Gets the current working directory. 29: Not Collective 31: Input Parameters: 32: . len - maximum length of path 34: Output Parameter: 35: . path - use to hold the result value. The string should be long enough 36: to hold the path. 38: Level: developer 40: Concepts: working directory 42: @*/ 43: PetscErrorCode PetscGetWorkingDirectory(char path[],size_t len) 44: { 46: #if defined(PETSC_HAVE_GETCWD) 47: if (!getcwd(path,len)) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_LIB,"getcwd()"); 48: #elif defined(PETSC_HAVE__GETCWD) 49: _getcwd(path,len); 50: #elif defined(PETSC_HAVE_GETWD) 51: getwd(path); 52: #else 53: SETERRQ(PETSC_COMM_SELF,PETSC_ERR_SUP_SYS, "Could not find getcwd() or getwd()"); 54: #endif 55: return(0); 56: }
__label__pos
0.993918
2 Replies Latest reply on Sep 25, 2009 3:04 AM by Pawel R how to save the history of modifications of the table? Pawel R Newbie Hi, I'm developing an application in Seam. Not going into the details this is a CRUD app and has to keep track on every modification a user ever did in the DB. So I have a few tables and modifying them the app has to save the old values in the historical tables with the same structure (could be even the same tables). Thanks to this I'll know which user did what. The question is how to implement this. I'm using seam conversation, so the modifications in the DB are being done once a conversation has ended. There is an annotation @PrePersist...but from there I can't get the previous version of the data. Can I ? Any ideas ?
__label__pos
0.836161
using nested for loop, workin with displaying prime numbers Discussion in 'C++' started by triplejump24, Nov 7, 2006. 1. triplejump24 triplejump24 Guest Hey. Im trying to make program that basically displays all the prime numbers. I need to use bool and for but im not quite sure if i have this right. So far i have this bunch of a mess but can anyone point me in the right direction? Thanks! # include <iostream> # include <cmath> using namespace std; int i; //int sqrt; //int j; int x; int j; int limit; bool primeNumber(int); int main() { // cout << "Limit: "; // cin >> x; // if ((x % 2 == 0) & (x % x == 0)) // { // return true; // } // else // { // return false; // return 0; //} // //bool primeNumber; { // if ((x % 2 == 0) & (x % x == 0)) // { // return true; // } // else // { // return false; // } //} for (i = 2; i <= x; i++) cout << i; { for (j = 2; j <= sqrt(i); i++) cout << sqrt(100.0) << endl; cout << sqrt(static_cast<double>(100)) << endl; //static cast conversts from one type to another } return 0; }   triplejump24, Nov 7, 2006 #1 1. Advertising 2. triplejump24 Daniel T. Guest "triplejump24" <> wrote: > Hey. Im trying to make program that basically displays all the prime > numbers. I need to use bool and for but im not quite sure if i have this > right. So far i have this bunch of a mess but can anyone point me in the > right direction? Here's a pointer for you. Start with the code below. When you run it, it will *not* print "working so far". Insert code where it says "insert code here" until you can get "working so far" to print (It should be quite easy :) After you get that part done, un-comment the second assert and modify the code until 'working so far" prints again, then the next comment, &c. #include <iostream> #include <cassert> // other includes as necessary bool is_prime( int number ) { bool result = false; // insert code here return result; } int main() { assert( is_prime( 2 ) ); //assert( is_prime( 3 ) ); //assert( ! is_prime( 4 ) ); //assert( is_prime( 5 ) ); cout << "Working so far!"; } -- To send me email, put "sheltie" in the subject.   Daniel T., Nov 7, 2006 #2 1. Advertising 3. > Hey. Im trying to make program that basically displays all the prime > numbers. I need to use bool and for but im not quite sure if i have this > right. So far i have this bunch of a mess but can anyone point me in the > right direction? Thanks! First ask yourself what is a primenumber? "An integer p is called a prime number if the only positive integers that divide p are 1 and p itself." So you need to check your would-be primenumber against all other integers. Well you do not need to check againts all other integers. Any integer larger than p will not divide p! So we need to check all integers smaller than p (excluding 0 and 1). This is something you should definitely use a loop for. You can use a "while()" loop, but i would prefer a "for()" loop. something like for(int n= 2; n<p;n++) in side the loop you can then check if n divides p. If it does, p is not a prime. If the loop completes without n dividing p, p is a prime. If you put all this into a function called IsPrime(int p) which can return a bool. ( true if the argument was a primenumber, false if not) if you now want test a number you can call if(IsPrime(number_to_test)) { // This was a prime number } You can now use a loop to test all the numbers you would like... (hint "for loop") or you can write each test explicitly if(Isprime(2)) : : if(Isprime(34)) I know there is a lot of room for improvement in my is prime function. But start by focussing on getting the function to work correctly first, then you can always optimise the code if needed. One optimization is only to test integers smaller than or equal to the squareroot of p. another ( and more difficult) is to only test a against all primenumbers smaller than the square root of p, provided you know ALL of them!! Two things I think you should consider is what should happen is you call IsPrime(0) and IsPrime(1). You wil probably have to modify the IsPrime function to handle this in a sensible way. A few comments on the code: > using namespace std; See http://www.parashift.com/c -faq-lite/coding-standards.html section 27.5 for a discussion. You need to use more comments in your code! If you have a varible, you should make a comment why its there, and what is is used for. The same applies to functions and files. make it a good habbit to write many comments, and most importantly keep the comments in sync with the code! > cout << sqrt(static_cast<double>(100)) << endl; Using a cast operator is something that should make you think "Is this really right?". Casting is often a consecuence of a flawed design. In some cases there are needed, but avoid them as much as possible. I hope this get you closer to your program. Bo Møller -- Bo Møller Hobby-Programmer   =?ISO-8859-15?Q?Bo_M=F8ller?=, Nov 7, 2006 #3 4. triplejump24 Don Guest triplejump24 wrote: > Hey. Im trying to make program that basically displays all the prime > numbers. #include <stdio.h> /* This program implements a blindingly fast O(n^n) algorithm to find prime numbers, using an elegant recursive method. As required it uses bool and for. */ bool _(int n, int m, int d) { bool r = m != n; for(int i=0; d && (i<n); i++) r &= _(n,(m<=n)?i*m:0,d-1)|!_(i,1,i); return r; } /*------------------------------------------ Print primes up to the requested value --------------------------------------------*/ int main(int argc, char* argv[]) { for(int n = 2; n; n++) printf("%d is%s prime\n",n, _(n,1,n)?"" : " not"); return 0; }   Don, Nov 10, 2006 #4 1. Advertising Want to reply to this thread or ask your own question? It takes just 2 minutes to sign up (and it's free!). Just click the sign up button to choose a username and then you can ask your own questions on the forum. Similar Threads 1. =?Utf-8?B?SW5kZXB0aA==?= Replies: 1 Views: 479 Bruce Barker Apr 1, 2005 2. nyenyec Replies: 3 Views: 1,023 Duncan Booth Jan 7, 2007 3. Sukeerthi Adiga :true not workin! Sukeerthi Adiga, Feb 27, 2008, in forum: Ruby Replies: 1 Views: 80 Andreas Kleindiek Feb 27, 2008 4. Jeremy Fischer Replies: 0 Views: 188 Jeremy Fischer Jan 16, 2005 5. Isaac Won Replies: 9 Views: 388 Ulrich Eckhardt Mar 4, 2013 Loading... Share This Page
__label__pos
0.741844
How can I check a value entered in a field on my form is a number? How can I check a value entered in a field on my form is a number? You can do it client side with JavaScript or server side with php etc. Are you using php, ASP etc. with your code or just HTML? How do I make a website field mandatory? As I posted on your earlier thread it depends on how you are coding your site - you need to put more information into your posts if you want a useful reply. The best way is to make it mandatory with both JS client side and php server side. If you need code help then give a bit more info as to your requirements. Threads merged as they seem to be related but will both require the same information supplied anyway. in HTML5 you can do type=‘number’ in an input tag, keeping in mind not all browsers support html5, and form validation is VITAL you can ad javascriptt fall backs, but the SAFEST way is to check your data server side, if when the data is received and checks it is valid then you redirect the user to a new page… other wise you redisplay the form. How you do his last part sever side will vary depending upon which server side language you are using.
__label__pos
0.991934
Commit 5d8e8d52 authored by Alvaro Sanchez's avatar Alvaro Sanchez Browse files Added texture width checks in opacity and gradientOpacity tables. ColorTransferFunction and PiecewiseFunction estimate an appropriate number of samples based on a given range and minimum distance between nodes. This estimated value is used by the volume lookup tables to sample the functions into the texture. parent ddb5ebc0 Pipeline #23051 passed with stage ......@@ -682,6 +682,37 @@ int vtkPiecewiseFunction::AdjustRange(double range[2]) return 1; } //-------------------------------------------------------------------------- int vtkPiecewiseFunction::EstimateMinNumberOfSamples(double const & x1, double const & x2) { double const d = this->FindMinimumXDistance(); int idealWidth = static_cast<int>(ceil((x2 - x1) / d)); return idealWidth; } //---------------------------------------------------------------------------- double vtkPiecewiseFunction::FindMinimumXDistance() { std::vector<vtkPiecewiseFunctionNode*> const & nodes = this->Internal->Nodes; size_t const size = nodes.size(); if (size < 2) return -1.0; double distance = std::numeric_limits<double>::max(); for (size_t i = 0; i < size - 1; i++) { double const currentDist = nodes[i + 1]->X - nodes[i]->X; if (currentDist < distance) { distance = currentDist; } } return distance; } // Returns a table of function values evaluated at regular intervals void vtkPiecewiseFunction::GetTable( double xStart, double xEnd, int size, double* table, ...... ......@@ -165,10 +165,26 @@ public: vtkGetMacro(AllowDuplicateScalars, int); vtkBooleanMacro(AllowDuplicateScalars, int); // Description: // Estimates the minimum size of a table such that it would correctly sample this function. // The returned value should be passed as parameter 'n' when calling GetTable(). int EstimateMinNumberOfSamples(double const & x1, double const & x2); protected: vtkPiecewiseFunction(); ~vtkPiecewiseFunction(); // Internal method to sort the vector and update the // Range whenever a node is added, edited or removed. // It always calls Modified(). void SortAndUpdateRange(); // Returns true if the range has been updated and Modified() has been called bool UpdateRange(); // Description: // Traverses the nodes to find the minimum distance. Assumes nodes are sorted. double FindMinimumXDistance(); // The internal STL structures vtkPiecewiseFunctionInternals *Internal; ......@@ -184,13 +200,6 @@ protected: // Min and max range of function point locations double Range[2]; // Internal method to sort the vector and update the // Range whenever a node is added, edited or removed. // It always calls Modified(). void SortAndUpdateRange(); // Returns true if the range has been updated and Modified() has been called bool UpdateRange(); int AllowDuplicateScalars; private: ...... ......@@ -1835,6 +1835,16 @@ int vtkColorTransferFunction::AdjustRange(double range[2]) return 1; } //-------------------------------------------------------------------------- int vtkColorTransferFunction::EstimateMinNumberOfSamples(double const & x1, double const & x2) { double const d = this->FindMinimumXDistance(); int idealWidth = static_cast<int>(ceil((x2 - x1) / d)); return idealWidth; } //---------------------------------------------------------------------------- double vtkColorTransferFunction::FindMinimumXDistance() { ......@@ -1844,11 +1854,9 @@ double vtkColorTransferFunction::FindMinimumXDistance() return -1.0; double distance = std::numeric_limits<double>::max(); for (size_t i = 0; i < size - 1; i++) { double const currentDist = nodes.at(i + 1)->X - nodes.at(i)->X; double const currentDist = nodes[i + 1]->X - nodes[i]->X; if (currentDist < distance) { distance = currentDist; ...... ......@@ -247,8 +247,9 @@ public: virtual void GetIndexedColor(vtkIdType idx, double rgba[4]); // Description: // Traverses the nodes to find the minimum distance. Assumes nodes are sorted. double FindMinimumXDistance(); // Estimates the minimum size of a table such that it would correctly sample this function. // The returned value should be passed as parameter 'n' when calling GetTable(). int EstimateMinNumberOfSamples(double const & x1, double const & x2); protected: vtkColorTransferFunction(); ......@@ -341,6 +342,10 @@ protected: // any point existed at newX, it will also be removed. void MovePoint(double oldX, double newX); // Description: // Traverses the nodes to find the minimum distance. Assumes nodes are sorted. double FindMinimumXDistance(); private: vtkColorTransferFunction(const vtkColorTransferFunction&) VTK_DELETE_FUNCTION; void operator=(const vtkColorTransferFunction&) VTK_DELETE_FUNCTION; ...... ......@@ -18,6 +18,7 @@ #include "vtkCamera.h" #include "vtkColorTransferFunction.h" #include "vtkGPUVolumeRayCastMapper.h" #include "vtkFixedPointVolumeRayCastMapper.h" #include "vtkImageData.h" #include "vtkLookupTable.h" #include "vtkPiecewiseFunction.h" ......@@ -33,6 +34,9 @@ #include "vtkVolumeProperty.h" #include "vtkXMLImageDataReader.h" #define GPU_MAPPER //---------------------------------------------------------------------------- int TestGPURayCastLargeColorTransferFunction(int argc, char* argv[]) { ......@@ -43,8 +47,16 @@ int TestGPURayCastLargeColorTransferFunction(int argc, char* argv[]) // http://www.spl.harvard.edu/publications/item/view/2037 vtkSmartPointer<vtkLookupTable> lut = vtkSmartPointer<vtkLookupTable>::New(); lut->SetNumberOfTableValues(5023); lut->SetTableRange(0, 5022); // Required for vtkLookupTable initialization int const NumValues = 5023; lut->SetNumberOfTableValues(NumValues); lut->SetTableRange(0, NumValues-1); for (int i = 0; i < NumValues; i++) { lut->SetTableValue(i, 0.0, 0.0, 0.0, 0.0); } lut->SetTableValue(0, 0 / 255.0, 0 / 255.0, 0 / 255.0, 0 / 255.0); lut->SetTableValue(2, 250 / 255.0, 250 / 255.0, 225 / 255.0, 255 / 255.0); lut->SetTableValue(3, 225 / 255.0, 190 / 255.0, 150 / 255.0, 255 / 255.0); ......@@ -80,7 +92,8 @@ int TestGPURayCastLargeColorTransferFunction(int argc, char* argv[]) lut->SetTableValue(54, 98 / 255.0, 153 / 255.0, 112 / 255.0, 255 / 255.0); lut->SetTableValue(58, 165 / 255.0, 0 / 255.0, 255 / 255.0, 255 / 255.0); lut->SetTableValue(60, 165 / 255.0, 40 / 255.0, 40 / 255.0, 255 / 255.0); lut->SetTableValue(61, 135 / 255.0, 205 / 255.0, 235 / 255.0, 255 / 255.0); // lut->SetTableValue(61, 165 / 255.0, 40 / 255.0, 40 / 255.0, 255 / 255.0); lut->SetTableValue(61, 135 / 255.0, 205 / 255.0, 235 / 255.0, 255 / 255.0); //medulla oblongata lut->SetTableValue(63, 90 / 255.0, 105 / 255.0, 215 / 255.0, 255 / 255.0); lut->SetTableValue(66, 0 / 255.0, 108 / 255.0, 112 / 255.0, 255 / 255.0); lut->SetTableValue(71, 0 / 255.0, 108 / 255.0, 112 / 255.0, 255 / 255.0); ......@@ -393,6 +406,7 @@ int TestGPURayCastLargeColorTransferFunction(int argc, char* argv[]) const double midPoint = 0.5; const double sharpness = 1.0; for (int i = 0; i < numColors; i++, value += step) // for (int i = 0; i < numColors; i++, value += (50 * step) ) { lut->GetTableValue(i, color); ......@@ -412,8 +426,14 @@ int TestGPURayCastLargeColorTransferFunction(int argc, char* argv[]) delete [] filename; filename = NULL; #ifdef GPU_MAPPER vtkSmartPointer<vtkGPUVolumeRayCastMapper> volumeMapper = vtkSmartPointer<vtkGPUVolumeRayCastMapper>::New(); #else vtkSmartPointer<vtkFixedPointVolumeRayCastMapper> volumeMapper = vtkSmartPointer<vtkFixedPointVolumeRayCastMapper>::New(); #endif volumeMapper->SetInputData(reader->GetOutput()); vtkSmartPointer<vtkVolumeProperty> volumeProperty = ......@@ -421,7 +441,7 @@ int TestGPURayCastLargeColorTransferFunction(int argc, char* argv[]) volumeProperty->SetColor(colorTransferFunction); volumeProperty->SetScalarOpacity(opacity); volumeProperty->SetInterpolationTypeToNearest(); volumeProperty->ShadeOn(); // volumeProperty->ShadeOn(); volumeProperty->SetAmbient(0.3); volumeProperty->SetDiffuse(0.6); volumeProperty->SetSpecular(0.5); ......@@ -437,6 +457,7 @@ int TestGPURayCastLargeColorTransferFunction(int argc, char* argv[]) xf->RotateZ(180.0); xf->RotateY(25.0); xf->RotateX(-65.0); //xf->RotateY(-90.0); volume->SetUserTransform(xf); vtkSmartPointer<vtkRenderWindow> renderWindow = ......@@ -460,7 +481,11 @@ int TestGPURayCastLargeColorTransferFunction(int argc, char* argv[]) iren->SetRenderWindow(renderWindow); renderWindow->Render();// make sure we have an OpenGL context. #ifdef GPU_MAPPER int valid = volumeMapper->IsRenderSupported(renderWindow, volumeProperty); #else int valid = 1; #endif int retVal; if (valid) ...... ......@@ -248,28 +248,9 @@ vtkStandardNewMacro(vtkOpenGLGPUVolumeRayCastMapper); //----------------------------------------------------------------------------- // //----------------------------------------------------------------------------- class vtkTextureTable class vtkTextureTable : public vtkObject { public: vtkTextureTable() { this->TextureId=0; this->Table=0; this->Loaded=false; this->LastLinearInterpolation=false; this->LastRange[0] = this->LastRange[1] = 0.0; } virtual ~vtkTextureTable() { if(this->TextureId!=0) { glDeleteTextures(1,&this->TextureId); vtkOpenGLStaticCheckErrorMacro("failed at glDeleteTextures"); this->TextureId=0; } delete[] this->Table; this->Table = 0; } bool IsLoaded()const { return this->Loaded; ......@@ -282,11 +263,17 @@ public: } int ComputeTableSize(vtkPiecewiseFunction* function) { return this->ComputeTableSize(function->GetSize()); int const idealW = function->EstimateMinNumberOfSamples(this->LastRange[0], this->LastRange[1]); return this->ComputeTableSize(idealW); } int ComputeTableSize(vtkColorTransferFunction* function) { return this->ComputeTableSize(function->GetSize()); int const idealW = function->EstimateMinNumberOfSamples(this->LastRange[0], this->LastRange[1]); return this->ComputeTableSize(idealW); } int ComputeTableSize(vtkLookupTable* function) { ......@@ -294,6 +281,26 @@ public: } protected: vtkTextureTable() { this->TextureId=0; this->Table=0; this->Loaded=false; this->LastLinearInterpolation=false; this->LastRange[0] = this->LastRange[1] = 0.0; } virtual ~vtkTextureTable() { if(this->TextureId!=0) { glDeleteTextures(1,&this->TextureId); vtkOpenGLStaticCheckErrorMacro("failed at glDeleteTextures"); this->TextureId=0; } delete[] this->Table; this->Table = 0; } GLuint TextureId; vtkTimeStamp BuildTime; float *Table; ......@@ -302,14 +309,38 @@ protected: double LastRange[2]; private: vtkTextureTable(const vtkTextureTable&); vtkTextureTable& operator=(const vtkTextureTable&); vtkTextureTable(const vtkTextureTable&); VTK_DELETE_FUNCTION vtkTextureTable& operator=(const vtkTextureTable&); VTK_DELETE_FUNCTION // Description: // Queries the GL_MAX_TEXTURE_SIZE and returns either the requested idealWidth // or the maximum supported. // Warning: This method assumes there is an active GL context. int ComputeTableSize(int idealWidth) { idealWidth = vtkMath::NearestPowerOfTwo(idealWidth); GLint maxWidth = -1; glGetIntegerv(GL_MAX_TEXTURE_SIZE, &maxWidth); if (maxWidth < 0) { vtkErrorMacro("Failed to query max texture size! using default " << vtkOpenGLGPUVolumeRayCastMapperOpacityTableSize); return vtkOpenGLGPUVolumeRayCastMapperOpacityTableSize; } int ComputeTableSize(int functionSize) { functionSize = std::max(functionSize, vtkOpenGLGPUVolumeRayCastMapperOpacityTableSize); return vtkMath::NearestPowerOfTwo(functionSize); if (maxWidth >= idealWidth) { idealWidth = vtkMath::Max(vtkOpenGLGPUVolumeRayCastMapperOpacityTableSize, idealWidth); return idealWidth; } vtkWarningMacro("This OpenGL implementation does not support the required " "texture size of " << idealWidth << ", falling back to maximum allowed, " << maxWidth << "." << "This may cause an incorrect color table mapping."); return maxWidth; } }; ......@@ -319,15 +350,8 @@ private: class vtkOpacityTable: public vtkTextureTable { public: vtkOpacityTable() { this->LastBlendMode=vtkVolumeMapper::MAXIMUM_INTENSITY_BLEND; this->LastSampleDistance=1.0; } ~vtkOpacityTable() { } static vtkOpacityTable* New(); vtkTypeMacro(vtkOpacityTable, vtkTextureTable); // \pre the active texture is set to TEXTURE2 void Update(vtkPiecewiseFunction *scalarOpacity, ......@@ -439,13 +463,22 @@ public: vtkOpenGLStaticCheckErrorMacro("failed after Update"); } protected: vtkOpacityTable() { this->LastBlendMode=vtkVolumeMapper::MAXIMUM_INTENSITY_BLEND; this->LastSampleDistance=1.0; } ~vtkOpacityTable() {}; int LastBlendMode; double LastSampleDistance; private: vtkOpacityTable(const vtkOpacityTable&); vtkOpacityTable& operator=(const vtkOpacityTable&); vtkOpacityTable(const vtkOpacityTable&); VTK_DELETE_FUNCTION vtkOpacityTable& operator=(const vtkOpacityTable&); VTK_DELETE_FUNCTION }; vtkStandardNewMacro(vtkOpacityTable); //----------------------------------------------------------------------------- // //----------------------------------------------------------------------------- ......@@ -454,43 +487,41 @@ class vtkOpacityTables public: vtkOpacityTables(unsigned int numberOfTables) { this->Tables = new vtkOpacityTable[numberOfTables]; this->NumberOfTables = numberOfTables; this->Tables.reserve(static_cast<size_t>(numberOfTables)); for (unsigned int i = 0; i < numberOfTables; i++) { this->Tables.push_back(vtkOpacityTable::New()); } } ~vtkOpacityTables() { delete [] this->Tables; size_t const size = this->Tables.size(); for (size_t i = 0; i < size; i++) { this->Tables[i]->Delete(); } } vtkOpacityTable* GetTable(unsigned int i) vtkOpacityTable* GetTable(size_t const i) { return &this->Tables[i]; return this->Tables[i]; } unsigned int GetNumberOfTables() size_t GetNumberOfTables() { return this->NumberOfTables; return this->Tables.size(); } private: unsigned int NumberOfTables; vtkOpacityTable *Tables; // undefined default constructor. vtkOpacityTables(); // undefined copy constructor. vtkOpacityTables(const vtkOpacityTables &other); // undefined assignment operator. vtkOpacityTables &operator=(const vtkOpacityTables &other); vtkOpacityTables(); VTK_DELETE_FUNCTION vtkOpacityTables(const vtkOpacityTables &other); VTK_DELETE_FUNCTION vtkOpacityTables &operator=(const vtkOpacityTables &other); VTK_DELETE_FUNCTION std::vector<vtkOpacityTable*> Tables; }; //----------------------------------------------------------------------------- class vtkRGBTable: public vtkTextureTable { public: vtkRGBTable() { } ~vtkRGBTable() { } static vtkRGBTable* New(); // \pre the active texture is set properly. (default color, // mask1, mask2,..) ......@@ -510,6 +541,8 @@ public: if (range[0] != this->LastRange[0] || range[1] != this->LastRange[1]) { needUpdate=true; this->LastRange[0] = range[0]; this->LastRange[1] = range[1]; } glBindTexture(GL_TEXTURE_1D,this->TextureId); if(needUpdate) ......@@ -535,8 +568,6 @@ public: vtkOpenGLStaticCheckErrorMacro("1d RGB texture is too large"); this->Loaded=true; this->BuildTime.Modified(); this->LastRange[0] = range[0]; this->LastRange[1] = range[1]; } needUpdate=needUpdate || ......@@ -559,8 +590,16 @@ public: vtkOpenGLStaticCheckErrorMacro("failed after Update"); } protected: vtkRGBTable() {}; ~vtkRGBTable() {}; private: vtkRGBTable(const vtkRGBTable &other); VTK_DELETE_FUNCTION vtkRGBTable &operator=(const vtkRGBTable &other); VTK_DELETE_FUNCTION }; vtkStandardNewMacro(vtkRGBTable); //----------------------------------------------------------------------------- // //----------------------------------------------------------------------------- ......@@ -2535,14 +2574,23 @@ void vtkOpenGLGPUVolumeRayCastMapper::ReleaseGraphicsResources( } } delete this->RGBTable; this->RGBTable=0; if (this->RGBTable) { this->RGBTable->Delete(); this->RGBTable = NULL; } delete this->Mask1RGBTable; this->Mask1RGBTable=0; if (this->Mask1RGBTable) { this->Mask1RGBTable->Delete(); this->Mask1RGBTable = NULL; } delete this->Mask2RGBTable; this->Mask2RGBTable=0; if (this->Mask2RGBTable) { this->Mask2RGBTable->Delete(); this->Mask2RGBTable = NULL; } delete this->OpacityTables; this->OpacityTables=0; ......@@ -4067,20 +4115,20 @@ void vtkOpenGLGPUVolumeRayCastMapper::PreRender(vtkRenderer *ren, this->TableRange[1]=scalarRange[1]; if(this->RGBTable==0) if(this->RGBTable == NULL) { this->RGBTable=new vtkRGBTable; this->RGBTable = vtkRGBTable::New(); } if (this->MaskInput != 0 && this->MaskType == LabelMapMaskType) if (this->MaskInput != NULL && this->MaskType == LabelMapMaskType) { if(this->Mask1RGBTable==0) if(this->Mask1RGBTable == NULL) { this->Mask1RGBTable=new vtkRGBTable; this->Mask1RGBTable = vtkRGBTable::New(); } if(this->Mask2RGBTable==0) if(this->Mask2RGBTable == NULL) { this->Mask2RGBTable=new vtkRGBTable; this->Mask2RGBTable = vtkRGBTable::New(); } } ...... ......@@ -560,13 +560,13 @@ void vtkOpenGLGPUVolumeRayCastMapper::vtkInternal::Initialize( if (this->Parent->MaskInput != 0 && this->Parent->MaskType == LabelMapMaskType) { if(this->Mask1RGBTable == 0) if(this->Mask1RGBTable == NULL) { this->Mask1RGBTable = new vtkOpenGLVolumeRGBTable(); this->Mask1RGBTable = vtkOpenGLVolumeRGBTable::New(); } if(this->Mask2RGBTable == 0) if(this->Mask2RGBTable == NULL) { this->Mask2RGBTable = new vtkOpenGLVolumeRGBTable(); this->Mask2RGBTable = vtkOpenGLVolumeRGBTable::New(); } } ......@@ -985,21 +985,27 @@ bool vtkOpenGLGPUVolumeRayCastMapper::vtkInternal::LoadMask(vtkRenderer* ren, void vtkOpenGLGPUVolumeRayCastMapper::vtkInternal::DeleteTransferFunctions() { delete this->RGBTables; this->RGBTables = 0; this->RGBTables = NULL; delete this->Mask1RGBTable; this->Mask1RGBTable=0; if (this->Mask1RGBTable) { this->Mask1RGBTable->Delete(); this->Mask1RGBTable = NULL; } delete this->Mask2RGBTable; this->Mask2RGBTable=0; if (this->Mask2RGBTable) { this->Mask2RGBTable->Delete(); this->Mask2RGBTable = NULL; } delete this->OpacityTables; this->OpacityTables = 0; this->OpacityTables = NULL;
__label__pos
0.980647
0 I've been using Amazon RDS instances for Postgres databses for a while, but I haven't had to do a lot of upgrading on large databases until recently. I hit an issue where one database was hundreds of gigs large. The upgrade never completes because the instance is oom killed. I've been reading a lot on the AWS website about common practices for upgrading databases major versions , but I haven't seen anything about large databases oom killing the process and how to avoid. I believe at least one of our databases on this instance have the potential to eventually get up to 500 gigs or so. Obviously scaling instances vertically to that much memory isn't a great solution (and probably doesn't exist). Also, scaling horizontally JUST for memory to upgrade doesn't seem to be a good approach. What is the normal way to handle this? 3 • FWIW, I'm surprised that the size of the database comes into play with a database system version upgrade. I think that's hardly the case with SQL Server. How have you determined that the size of your database is relevant to the issue? – J.D. Commented Feb 7 at 19:03 • I guess it's possible that it's not actually the cause. But, I have multiple RDS instances of identical configuration. I've upgraded all of them through the same versions, and the only one that fails is the one that has a database that's very large. And using monitoring tools I can see that the upgrade tries, until the instance is oom killed. There's no differences between all of the instances except for the size of that one database. I've also ruled out several other things that I thought it was initially. Commented Feb 7 at 19:06 • I cannot imagine an upgrade procedure that needs lots of RAM. Since you cannot directly see what is going on, you might have to ask Amazon. Commented Feb 8 at 16:36 0 Your Answer By clicking “Post Your Answer”, you agree to our terms of service and acknowledge you have read our privacy policy.
__label__pos
0.566741
Author Topic: TT1 Problem 3 (morning)  (Read 1901 times) Victor Ivrii • Administrator • Elder Member • ***** • Posts: 2599 • Karma: 0 • View Profile • Personal website of Victor Ivrii TT1 Problem 3 (morning) « on: October 19, 2018, 03:55:03 AM » (a) Show that $u(x,y)= 8xy^3 -8x^3 y+ 5x$ is a harmonic function (b) Find the harmonic conjugate function $v(x,y)$. (c) Consider $u(x,y)+iv(x,y)$ and write it as a function $f(z)$ of $z=x+iy$. Vedant Shah • Jr. Member • ** • Posts: 13 • Karma: 8 • View Profile Re: TT1 Problem 3 (morning) « Reply #1 on: October 19, 2018, 09:23:34 AM » (a) $ U_{xx} = \frac{\partial}{\partial x} \frac{\partial}{\partial x} U \\ = \frac{\partial}{\partial x} 8y^3 -24x^2 y +5 \\ = -48xy \\ \\ U_{yy} = \frac{\partial}{\partial y} \frac{\partial}{\partial y} U \\ = \frac{\partial}{\partial y} 24x y^2 - 8x^3 \\ = 48xy \\ U_{xx} + U_{yy} = -48xy + 48xy = 0 $ Thus, U is harmonic. (b) By Cauchy Reimann: $ V_y = U_x = 8y^3 - 24x^2 y + 5 \Rightarrow V = 2y^4 - 12 x^2 y^2 +5y +h(x)\\ V_x = -U_y = -24x y^2 + 8x^3 \Rightarrow V = -12 x^2 y^2 + 2x^4 + g(y) \\ \Rightarrow V(x,y) = 2x^4 - 12 x^2 y^2 + 2y^4 + 5y $ (c) $$ f(x,y) = U(x,y) + iV(x,y) = 8xy^3 - 8x^3y+5x + i(2x^4 - 12 x^2 y^2 + 2y^4 + 5y) \\ f(x,y) = 2i(x^4 + 4ix^3y - 6x^2y^2 -4ixy^3 +y^4) + 5(x+iy) \\ f(x,y) = 2i(x+iy)^4 + 5(x+iy) \\ f(z) = 2i ({z}) ^4 + 5z\color{red}{+Ci}. $$ « Last Edit: October 20, 2018, 03:08:41 PM by Victor Ivrii »  
__label__pos
0.996319
Home Homepage What is Software Reporter Tool? What is Software Reporter Tool? 0 What is Software Reporter Tool? What is Software Reporter Tool Software Reporter Tool In this article you will get complete information regarding Chrome Software Tool, the issues it creates and how to solve them.  Quick-Navigation Links What is a Software Reporter Tool? Software reporter tool is a Google Chrome application that keeps an eye on your browser so that any malicious add-on or forced advertisement could be prevented. This application is also used to send feed-back and crash reports of websites to chrome. The main purpose of the software reporter tool is to make your experience better on Chrome browser. You can find this application with the name ‘software_reporter_tool.exe’ in your system and it automatically downloads when you install google chrome in your PC. That’s why, the user is unaware of this application. the main function of chrome software reporter tool is to send scheduled reports to chrome regarding any suspicious software present in your PC that may affect your browsing experience. The Software Reporter Tool is an executable file which is why it works automatically. What is Software Reporter Tool What is Software Reporter Tool | Issues and Solutions? Chrome software reporter tool helps to keep malicious programs away from your browser and keeps it smooth. It has some issues too, find the solution here.’ What is Reporter Tool High CPU issue? This software reporter tool also has some drawbacks, one of its main drawbacks is that it makes your CPU high because it is running multiple times in the background. To see this thing you need to open ‘Task Manager’ by pressing ctrl+alt+del. There you will notice ‘software_reporter_tool.exe’ running multiple instances. Users have reported that this thing can occupy more than 50% of system memory which certainly makes the CPU slower. The question is why this happens? It happens because the chrome software reporter tool is a self-executable application and while you are using the browser and any page crashes then the tool becomes active and if this process gets repeated then your computer becomes slow. And the second issue chrome software reporter tool makes is by getting hanged, the software won’t respond and your google chrome gets stuck. This also makes you equally annoyed.   Where to find Chrome Software Reporter Tool? Software reporter tool is installed through google chrome and it is located in the chrome application data folder inside the SwReporter folder with the name of ‘software_reporter_tool.exe’. This folder also contains other files regarding your chrome like saved passwords, bookmarks and settings. There are multiple ways to open the location of software_reporter_tool. What is Software Reporter Tool What is Software Reporter Tool Method 1: Open the task manager and right click on ‘software_reporter-tool.exe’ then click ‘Open File Location’. There you will find the ‘software_reporter_tool.exe’ file which is an executable application. Method 2: Open ‘Run Dialog Box’ by pressing ‘windows key + R’ at the same time. Then insert below given syntax and press enter Syntax: %localappdata%\Google\Chrome\User Data\SwReporter   Is Software Reporter Tool a virus? No, This isn’t a virus at all. The software reporter tool is a Google application that gets installed with google chrome browser and it is completely safe. This is only a Google certified and authentic Google app which helps to keep your browser safe. Another important thing for you to know is that chrome software reporter tool does not have network connection of any kind which clearly means that it only sends reports to google. No antivirus software has ever detected software reporter tool as a virus.   Diff. b/w Chrome Software Reporter Tool and Chrome Clean-up Tool. Both of the applications have the same purpose, the chrome software reporter tool and chrome cleanup tool both detect suspicious applications from your system and ask you to delete them. The software reporter tool is attached with the cleanup tool of google chrome and it sends the report to the cleanup tool and then the google chrome browser asks the user to delete those applications causing interference in smooth browsing with the help of cleanup tool. This is why both applications exist and also have their separate functionalities. This tool is very helpful in creating good user experience and keeping suspicious softwares away from your PC that would harm you in any possible way. Here the main concern of users is regarding the memory consumption of software_reporter_tool that it slows down the PC but this happens only in few cases mostly when some cheap third-party software is installed that causes problems again and again for the browser; otherwise, it only takes about twenty to  twenty-five minutes at max for completely scanning the PC.   Common Issues of Chrome Software Reporter Tool • It may run multiple times in the background resulting in a high CPU because it will occupy RAM and disk space, slowing down the processing. • Software_Reporter_Tool gets crashed which also makes your browser stuck and starts showing errors. • It may delete necessary chrome add-ons installed by you because they may seem to be suspicious. • It also scans your system and sends reports in form of metadata that may harm your privacy concerns.   Can I disable the Software Reporter tool? Yes, you can delete files of software_reporter_tool.exe or completely uninstall it from your pc, you can also rename it which will prevent it from executing automatically. But, keep in mind that this software will be installed again when you update your browser so you may need to remove it again after updating.   How to Disable Software_Reporter_Tool? You can disable the software reporter tool in multiple ways and, we have created a step-by-step guide to help doing this. This is completely your personal preference if you want to disable this tool. Overall it works fine and also prevents you from malware. Click on this link to learn multiple methods of disabling this tool. LEAVE A REPLY Please enter your comment! Please enter your name here
__label__pos
0.882368
Tooltip Text Hello, How do I put a tooltip text on my toolbar buttons? Any coding required? If yes, please show me how to do it. If not, then where do I set the text? The prompt comes in the status bar, which is far down on the screen. Too difficult to be noticed. Please help. Thank you. cheeonnAsked: Who is Participating?   Meir RivkinConnect With a Mentor Full stack Software EngineerCommented: in the properties of each toolbar button there is a place to add text for the statusbar and for the tooltip. look for "Prompt" edit box... the syntax is this: "statusbar text\ntooltip text" notice that there's \n in the middle, this is for VC to differ between the text for the statusbar and the text for the tooltip. good luck 0   tdaoudCommented: I assume your toolbat buttons are mapped to Menu commands that you also have on your menus. In that case go to the menu command and where you type text that appears in the status bar, add to it "\nTool tip text" at the end of it.   for example if you have a menu command that edit employees you might have in its status bar text "Manage employees database\nManage Employees" So the first part before "\n" is the status bar text and the part after is the tooltip for the button of the same function. No programming required. Good Luck, Tarik 0   cheeonnAuthor Commented: Thank you all. Anyway, I found the method already. Thanks 0 Question has a verified solution. Are you are experiencing a similar issue? Get a personalized answer when you ask a related question. Have a better answer? Share it in a comment. All Courses From novice to tech pro — start learning today.
__label__pos
0.913532
D3 Tips and Tricks v6 Wednesday, 11 February 2015 Raspberry Pi System Information Measurement: 2. Record The following post is a section of the book 'Raspberry Pi: Measure, Record, Explore'.  The entire book can be downloaded in pdf format for free from Leanpub or you can read it online here. Since this post is a snapshot in time. I recommend that you download a copy of the book which is updated frequently to improve and expand the content. --------------------------------------- This is the second of three posts working through a project looking at Measuring Recording and Exploring system information from and with the Raspberry Pi. Record To record this data we will use a Python program that checks all the the values and writes them into our MySQL database along with the current time stamp. Our Python program will only write a single group of readings to the database and we will execute the program at a regular interval using a cron job in the same way as the multiple temperature sensor project. Database preparation First we will set up our database table that will store our data. Using the phpMyAdmin web interface that we set up, log on using the administrator (root) account and select the ‘measurements’ database that we created as part of the initial set-up. Create the MySQL Table Create the MySQL Table Enter in the name of the table and the number of columns that we are going to use for our measured values. In the screenshot above we can see that the name of the table is ‘system_info’ and the number of columns is ‘5’. We will use five columns so that we can store our four readings (system load, used ram, used disk and CPU temperature). Once we click on ‘Go’ we are presented with a list of options to configure our table’s columns. Again, we are going to keep the process as simple as practical and while we could be more economical with our data type selections, we’ll err on the side of simplicity. Configure the MySQL Table Columns Configure the MySQL Table Columns For the first column we can enter the name of the ‘Column’ as ‘load’ with a type of ‘REAL’. Our second column is ‘ram’ and our third is ‘disk’. Both of these should have a type of ‘TINYINT’. Then we have a column for ‘temperature’ and the type is ‘FLOAT’. Lastly we include the column ‘dtg’ (short for date time group) the type as ‘TIMESTAMP’ and the ‘Default’ value as ‘CURRENT_TIMESTAMP’. Save the MySQL Table Columns Save the MySQL Table Columns Scroll down a little and click on the ‘Save’ button and we’re done. Why did we choose those particular settings for our table? Our system load readings will vary from 0 up and will hopefully spend most of their time as a decimal less than 1. However, sometimes it will exceed this. Our ‘temperature’ readings will also be a decimal value. Assigning them both a REAL data type allows us to recorded values where the number can have a maximum of 65 Digits, with 30 digits after decimal point (that should be plenty). Both ‘ram’ and ‘disk’ will be integers, but since they will both be percentage values which will not exceed 100% we can use the ‘TINYINT’ type which allows values from -128 to 127 (or 0 to 255 if they’re unsigned). Our ‘dtg’ column needs to store a value of time that includes the date and the time, so the advantage of selecting TIMESTAMP in this case is that we can select the default value to be the current time which means that when we write our data to the table we only need to write the ‘temperature’ and ‘sensor_id’ values and the ‘dtg’ will be entered automatically for us. The disadvantage of using ‘TIMESTAMP’ is that it has a more limited range than DATETIME. TIMESTAMP can only have a range between ‘1970-01-01 00:00:01’ to ‘2038-01-19 03:14:07’. Record the system information values The following Python code is a script which allows us to check the system readings from the Raspberry Pi and writes them to our database. The full code can be found in the code samples bundled with this book (system_info.py). #!/usr/bin/python # -*- coding: utf-8 -*- import subprocess import os import MySQLdb as mdb # Function for storing readings into MySQL def insertDB(system_load, ram, disk, temperature): try: con = mdb.connect('localhost', 'pi_insert', 'xxxxxxxxxx', 'measurements'); cursor = con.cursor() sql = "INSERT INTO system_info(`load`,`ram`,`disk`,`temperature`) \ VALUES ('%s', '%s', '%s', '%s')" % \ (system_load, ram, disk, temperature) cursor.execute(sql) con.commit() con.close() except mdb.Error, e: con.rollback() print "Error %d: %s" % (e.args[0],e.args[1]) sys.exit(1) # returns the system load over the past minute def get_load(): try: s = subprocess.check_output(["cat","/proc/loadavg"]) return float(s.split()[0]) except: return 0 # Returns the used ram as a percentage of the total available def get_ram(): try: s = subprocess.check_output(["free","-m"]) lines = s.split("\n") used_mem = float(lines[1].split()[2]) total_mem = float(lines[1].split()[1]) return (int((used_mem/total_mem)*100)) except: return 0 # Returns the percentage used disk space on the /dev/root partition def get_disk(): try: s = subprocess.check_output(["df","/dev/root"]) lines = s.split("\n") return int(lines[1].split("%")[0].split()[4]) except: return 0 # Returns the temperature in degrees C of the CPU def get_temperature(): try: dir_path="/opt/vc/bin/vcgencmd" s = subprocess.check_output([dir_path,"measure_temp"]) return float(s.split("=")[1][:-3]) except: return 0 got_load = str(get_load()) got_ram = str(get_ram()) got_disk = str(get_disk()) got_temperature = str(get_temperature()) insertDB(got_load, got_ram, got_disk, got_temperature) This script can be saved in our home directory (/home/pi) and can be run by typing; python system_info.py While we won’t see much happening at the command line, if we use our web browser to go to the phpMyAdmin interface and select the ‘measurements’ database and then the ‘system_info’ table we will see a range of information for the different system parameters and their associated time of reading. System info in MySQL System info in MySQL As with our previous project recording multiple temperature points, this script only records a single line of data whenever it is run. To make the collection more regular we will put in a cron job later to regularly check and record. Code Explanation The script starts by importing the modules that it’s going to use for the process of reading and recording the temperature measurements; import subprocess import os import MySQLdb as mdb Python code in one module gains access to the code in another module by the process of importing it. The import statement invokes the process and combines two operations; it searches for the named module, then it binds the results of that search to a name in the local scope. We then declare the function that will insert the readings into the MySQL database; def insertDB(system_load, ram, disk, temperature): try: con = mdb.connect('localhost', 'pi_insert', 'xxxxxxxxxx', 'measurements'); cursor = con.cursor() sql = "INSERT INTO system_info(`load`,`ram`,`disk`,`temperature`) \ VALUES ('%s', '%s', '%s', '%s')" % \ (system_load, ram, disk, temperature) cursor.execute(sql) con.commit() con.close() except mdb.Error, e: con.rollback() print "Error %d: %s" % (e.args[0],e.args[1]) sys.exit(1) This is a fairly simple insert of the values we will be collecting into the database. Then we have our four functions that collect our system information; load As described in the earlier section we will be extracting information out of the /proc/loadavg file. Specifically we will extract the average load for the last minute. We are going to accomplish this using the following code; def get_load(): try: s = subprocess.check_output(["cat","/proc/loadavg"]) return float(s.split()[0]) except: return 0 The code employs the subprocess module (which we loaded at the start of the script) which in this case is using the check_outputconvenience function which will run the command (cat) and the argument (/proc/loadavg). It will return the output as a string (0.14 0.11 0.13 1/196 16991) that we can then manipulate. This string is stored in the variable s. The following line returns the value from the function. The value is a floating number (decimal) and we are taking the first part (split()[0]) of the string (s) which by default is being split on any whitespace. In the case of our example string (0.14 0.11 0.13 1/196 16991) that would return the value 0.14. If there is a problem retrieving the number it will be set to 0 (except: return 0). ram As described in the earlier section we will be extracting information out of the results from running the free command with the -margument. Specifically we will extract the used memory and total memory values and convert them to a percentage. We are going to accomplish this using the following code; def get_ram(): try: s = subprocess.check_output(["free","-m"]) lines = s.split("\n") used_mem = float(lines[1].split()[2]) total_mem = float(lines[1].split()[1]) return (int((used_mem/total_mem)*100)) except: return 0 The code employs the subprocess module (which we loaded at the start of the script) which in this case is using the check_outputconvenience function which will run the command (free) and the argument (-m). It will store the returned multi-line output showing memory usage in the variable s. The output from this command (if it is run from the command line) would look a little like this; total used free shared buffers cached Mem: 437 385 51 0 85 197 -/+ buffers/cache: 102 335 Swap: 99 0 99 We then split that output string line by line and store the result in the array lines using lines = s.split("\n"). Then we find the used and total memory by looking at the second line down (lines[1]) and extracting the appropriate column (split()[2] for used and split()[1] for total). Then it’s just some simple math to turn the memory variables (used_mem and total_mem) into a percentage. If there is a problem retrieving the number it will be set to 0 (except: return 0). disk As described in the earlier section we will be extracting information out of the results from running the df command with the /dev/rootargument. Specifically we will extract the percentage used value. We will accomplish this using the following code; def get_disk(): try: s = subprocess.check_output(["df","/dev/root"]) lines = s.split("\n") return int(lines[1].split("%")[0].split()[4]) except: return 0 The code employs the subprocess module (which we loaded at the start of the script) which in this case is using the check_outputconvenience function which will run the command (df) and the argument (/dev/root). It will store the returned multi-line output showing disk partition usage data in the variable s. We then split that output string line by line and store the result in the array lines using lines = s.split("\n"). Then, using the second line down (lines[1])… /dev/root 7513804 2671756 4486552 38% / … we extracting the percentage column (split()[4]) and remove the percentage sign from the number (split("%")[0]). The final value is returned as an integer. If there is a problem retrieving the number it will be set to 0 (except: return 0). temperature As described in the earlier section we will be extracting the temperature of the Raspberry Pis CPU the vcgencmd command with themeasure_temp argument. We will accomplish this using the following code; def get_temperature(): try: dir_path="/opt/vc/bin/vcgencmd" s = subprocess.check_output([dir_path,"measure_temp"]) return float(s.split("=")[1][:-3]) except: return 0 The code employs the subprocess module (which we loaded at the start of the script) which in this case is using the check_outputconvenience function which will run the command (vcgencmd) and the argument (/dev/root) The vcgencmd command is referenced by its full path name which is initially stored as the variable dir_path and is then used in the subprocess command (this is only done for the convenience of not causing a line break in the code for the book by the way). The measure_temp argument returns the temperature in a human readable string (temp=39.0'C) which is stored in the variable s. We are extracting the percentage value by splitting the line on the equals sign (split("=")), taking the text after the equals sign and trimming off the extra that is not required ([1][:-3]). The final value is returned as an real number. If there is a problem retrieving the number it will be set to 0 (except: return 0). Main program The main part of the program (if you can call it that) consists of only the following lines; got_load = str(get_load()) got_ram = str(get_ram()) got_disk = str(get_disk()) got_temperature = str(get_temperature()) insertDB(got_load, got_ram, got_disk, got_temperature) They serve to retrieve the value from each of our measurement functions and to then send the results to the function that writes the values to the database. Recording data on a regular basis with cron As mentioned earlier, while our code is a thing of beauty, it only records a single entry for each sensor every time it is run. What we need to implement is a schedule so that at a regular time, the program is run. This is achieved using cron via the crontab. While we will cover the requirements for this project here, you can read more about the crontab in the Glossary. To set up our schedule we need to edit the crontab file. This is is done using the following command; crontab -e Once run it will open the crontab in the nano editor. We want to add in an entry at the end of the file that looks like the following; */1 * * * * /usr/bin/python /home/pi/system_info.py This instructs the computer that exert minute of every hour of every day of every month we run the command /usr/bin/python /home/pi/system_info.py (which if we were at the command line in the pi home directory we would run as python system_info.py, but since we can’t guarantee where we will be when running the script, we are supplying the full path to the python command and the system_info.pyscript. Save the file and the computer will start running the program on its designated schedule and we will have sensor entries written to our database every minute. The post above (and heaps of other stuff) is in the book 'Raspberry Pi: Measure, Record, Explore' that can be downloaded for free (or donate if you really want to :-)). No comments: Post a comment
__label__pos
0.848874
Take the 2-minute tour × Stack Overflow is a question and answer site for professional and enthusiast programmers. It's 100% free, no registration required. I have set the buffer to size 100. I display the buffer in the main function where the buffer is declared. However, when I pass the buffer to the function and get the sizeof '4', I was thinking it should be 100, as that is the size of the buffer that I created in main. output: buffer size: 100 sizeof(buffer): 4 #include <string.h> #include <stdio.h> void load_buffer(char *buffer); int main() { char buffer[100]; printf("buffer size: %d\n", sizeof(buffer)); load_buffer(buffer); return 0; } void load_buffer(char *buffer) { printf("sizeof(buffer): %d\n", sizeof(buffer)); } share|improve this question      In main() you explicitly declared an array, whilst in load_buffer() you declared a pointer. The difference is, in main() the compiler knows that buffer is an array, but in load_buffer() the compiler doesn't know whether buffer is from an array, or an allocated peace of memory in the heap, or a struct or whatever. The compiler only knows the memory address (thus a pointer), so sizeof() returns the size of the memory address, which is 4 bytes long (because it's a 32 bits machine, if it was a 64 bits machine, it would be 8 bytes long). Hope this helps to understand a bit more. –  AndaluZ Apr 1 at 12:15 4 Answers 4 up vote 16 down vote accepted You are using the size of the pointer to the buffer (4 bytes), rather than the size of the buffer. In C, you have to pass the size of the buffer separately, which is part of the reason buffer overruns happen so easily and frequently. void load_buffer(char * buffer, size_t bufSize) { ... } share|improve this answer The answers by Mitch Wheat and hhafez are completely right and to the point. I'm going to show some additional information which may prove useful sometimes. Note that the same happens if you tell the compiler that you have an array of the right size void load_buffer(char buffer[100]) { /* prints 4 too! */ printf("sizeof(buffer): %d\n", sizeof(buffer)); } An array as parameter is just declaring a pointer. The compiler automatically changes that to char *name even if it was declared as char name[N]. If you want to force callers to pass an array of size 100 only, you can accept the address of the array (and the type of that) instead: void load_buffer(char (*buffer)[100]) { /* prints 100 */ printf("sizeof(buffer): %d\n", sizeof(*buffer)); } It's a pointer to the array you have in main, so you need to dereference in the function to get the array. Indexing then is done by buffer[0][N] or (*buffer)[N] Nobody I know is doing that and I'm neither doing it myself, because it rather complicates passing of the argument. But it's good to know about it. You can call the function like this then load_buffer(&buffer) If you want to accept other sizes too, i would go with the passing-N option the other two answers recommend. share|improve this answer      +1 For doing a little extra effort –  Tom Apr 25 '09 at 2:21 From the OP void load_buffer(char *buffer) { printf("sizeof(buffer): %d\n", sizeof(buffer)); } Even though you can imagine that load_buffer() is passed the buffer by refrence, what is really happening is you are passing a pointer to char by value. The actual array is not passed so there is no way for the load_buffer() function to know the size of the buffer array So what is sizeof(buffer) doing? It is simply returning the size of a pointer to char. If load_buffer() needs the size of the buffer it needs to be passed speratly. Or you can create a new struct that contains both a char array and the size of the array, and pass a pointer to that struct instead, that way the buffer and it's size are always together ;) share|improve this answer What happens, is when you pass an array to a function, you only pass the address of the array in the memory, not the size of the array. What sizeof(buffer) is outputting in load_buffer() is the size of the pointer, which is four bytes. The best way to keep the size of the buffer in the function is to change the function to: void load_buffer(char *buffer, int length); and the call to: load_buffer(buffer, sizeof(buffer)); and then use length whenever you want the size of buffer. share|improve this answer Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.914066