content
stringlengths
228
999k
pred_label
stringclasses
1 value
pred_score
float64
0.5
1
Welcome to the Onshape forum! Ask questions and join in the discussions about everything Onshape. First time visiting? Here are some places to start: 1. Looking for a certain topic? Check out the categories filter or use Search (upper right). 2. Need support? Ask a question to our Community Support category. 3. Please submit support tickets for bugs but you can request improvements in the Product Feedback category. 4. Be respectful, on topic and if you see a problem, Flag it. If you would like to contact our Community Manager personally, feel free to send a private message or an email. Finding 90 degree angled vertices among selected lines eric_ma454eric_ma454 Member Posts: 14 Hi all, My name is Eric Ma, and myself and 3 others represent a subteam of AguaClara Cornell dedicated to using Featurescript to model our plants. We just switched to OnShape, and are new to Featurescript, so we'll probably be showing up on the forums much more frequently. We're excited to learn more about the language, and all help would be greatly appreciated! Our current project involves inserting elbows into a selection of created pipes. This is a modification off of the current Beam feature, which is being used to model pipes. The current feature takes an input of edges, and creates "beams" from the edges. Our goal is to find all the points within the given selection of edges that have a 90 degree vertex, and insert an elbow of a specific size there. Our current issue stems from finding the specific 90 degree points. We have found all adjacent points with qVertexAdjacent, but the issue is we're unsure how to find that the vertices that only have two lines stemming from it at 90 degrees. Any help would be appreciated, on this specific question or on Featurescript in general. Best, Eric Ma AguaClara Cornell  @ethan_keller924 Comments • konstantin_shiriazdanovkonstantin_shiriazdanov ✭✭✭✭✭ Member Posts: 918 ✭✭✭✭✭ As i see you need to iterate through a list of given vertices in the loop and for each vertex query its adjacent line edges and check if the number of edges is exactly 2 and if the vectors of edge lines are perpendicular. so if definition.vertices is a Query of verteces then you get: normalAdjacentEdges = []; for vertex in evaluateQuery(context, definition.vertices) { var lineEdges =qGeometry( qVertexAdjacent(vertex, EntityType.EDGE), GeometryType.LINE); lineEdges = evaluateQuery(context, lineEdges); if (size(lineEdges)==2) { var line1 = evLine(..., lineEdges[0]); var line2 = evLine(..., lineEdges[1]); if (perpendicularVectors(line1.direction, line2.direction)) normalAdjacentEdges =  append(normalAdjacentEdges, lineEdges); } } • MBartlett21MBartlett21 EDU Member Posts: 1,640 EDU @konstantin_shiriazdanov & @eric_ma454 You may want to use qWithinRadius if they are sketch lines to find adjacent ones. Example code (Replace variables in BOLD_ITALICS with your variables: var perpendicularPointsAndLines = []; var evaluatedVertices = evaluateQuery(context, VERTICES); for(var vertex in evaluatedVertices){ var vertexPoint = evVertexPoint(context, {"vertex" : vertex }); var lineEdges = evaluateQuery(context, qGeometry(qWithinRadius(LINES_FROM_INPUT, vertexPoint, TOLERANCE.zeroLength * meter), GeometryType.LINE)); if(size(lineEdges) == 2){ var d1 = evLine(context, {"edge" : lineEdges[0]}).direction; var d2 = evLine(context, {"edge" : lineEdges[1]}).direction; if(perpendicularVectors(d1,d2)) perpendicularPointsAndLines = append(perpendicularPointsAndLines, {"vertex" : vertexPoint,"edges" : lineEdges}); } } MB - I make FeatureScripts: view FS (My FS's have "Official" beside them) • ethan_keller924ethan_keller924 ✭✭ Member Posts: 34 ✭✭ Another option is to create an array of all the angles between adjacent lines within a constructed path. This feature should work for you!: FeatureScript 937; import(path : "onshape/std/geometry.fs", version : "937.0"); annotation { "Feature Type Name" : "FindAngles" } export const myFeature = defineFeature(function(context is Context, id is Id, definition is map) precondition { // Define the parameters of the feature type annotation { "Name" : "Lines or arcs", "Filter" : EntityType.EDGE && (GeometryType.LINE || GeometryType.ARC) } definition.edges is Query; } { var path = constructPath(context, definition.edges); var nLines = size(path.edges); var tangentLines = evPathTangentLines(context, path, range(0,1,nLines)); var angles = []; for (var i = 0; i<nLines-1; i=i+1){ var angle = angleBetween(tangentLines.tangentLines[i].direction, tangentLines.tangentLines[i+1].direction); angles = append(angles, angle); } print("An array of the angle at each of the points along the selected path: "); print(angles); }); Sign In or Register to comment.
__label__pos
0.948075
0 I have downloaded the netCDF plugin for GeoServer 2.15 and put it in its directory >> it had been shown on GeoServer as raster data >> but while uploading .nc file by using this plugin I still have this error Could not list layers for this store, an error occurred retrieving them: Failed to create reader from file:gwc/psl_RF_1980-2005_monmean.nc and hints Hints: REPOSITORY = org.geoserver.catalog.CatalogRepository@7062ed2a EXECUTOR_SERVICE = java.util.concurrent.ThreadPoolExecutor@960bf8a[Running, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 0] System defaults: FORCE_LONGITUDE_FIRST_AXIS_ORDER = true STYLE_FACTORY = StyleFactoryImpl FORCE_AXIS_ORDER_HONORING = http FILTER_FACTORY = FilterFactoryImpl GRID_COVERAGE_FACTORY = GridCoverageFactory TILE_ENCODING = null LENIENT_DATUM_SHIFT = true COMPARISON_TOLERANCE = 1.0E-8 FEATURE_FACTORY = org.geotools.feature.LenientFeatureFactoryImpl@11acdc30 that can`t enable me to upload any data. • does geoserver have write permissions for the directory? – Ian Turton Mar 20 at 13:03 • what does the log file say? after you turn logging up to GeoTools-Developer – Ian Turton Mar 20 at 14:52 Your Answer By clicking “Post Your Answer”, you agree to our terms of service, privacy policy and cookie policy Browse other questions tagged or ask your own question.
__label__pos
0.748582
使用 Terraform 配置 Azure 虚拟桌面 文章使用以下 Terraform 和 Terraform 提供程序版本进行测试: 使用 Terraform 可以定义、预览和部署云基础结构。 使用 Terraform 时,请使用 HCL 语法来创建配置文件。 利用 HCL 语法,可指定 Azure 这样的云提供程序和构成云基础结构的元素。 创建配置文件后,请创建一个执行计划,利用该计划,可在部署基础结构更改之前先预览这些更改。 验证了更改后,请应用该执行计划以部署基础结构。 本文概述了如何使用 Terraform 部署 ARM Azure 虚拟桌面环境,而不是 AVD 经典环境。 Azure 虚拟桌面有几种先决条件要求 不熟悉 Azure 虚拟桌面? 从 什么是 Azure 虚拟桌面开始? 假定已设置适当的平台基础,该基础可能是 企业规模登陆区域平台基础,也可能不是企业规模登陆区域平台基础。 在本文中,学习如何: • 使用 Terraform 创建 Azure 虚拟桌面工作区 • 使用 Terraform 创建 Azure 虚拟桌面主机池 • 使用 Terraform 创建 Azure 桌面应用程序组 • 关联工作区和桌面应用程序组 1.配置环境 • Azure 订阅:如果没有 Azure 订阅,请在开始之前创建一个免费帐户 2. 实现 Terraform 代码 1. 创建用于测试和运行示例 Terraform 代码的目录,并将其设为当前目录。 2. 创建名为 providers.tf 的文件并插入下列代码: terraform { required_providers { azurerm = { source = "hashicorp/azurerm" version = "~>2.0" } azuread = { source = "hashicorp/azuread" } } } provider "azurerm" { features {} } 3. 创建名为 main.tf 的文件并插入下列代码: # Resource group name is output when execution plan is applied. resource "azurerm_resource_group" "sh" { name = var.rg_name location = var.resource_group_location } # Create AVD workspace resource "azurerm_virtual_desktop_workspace" "workspace" { name = var.workspace resource_group_name = azurerm_resource_group.sh.name location = azurerm_resource_group.sh.location friendly_name = "${var.prefix} Workspace" description = "${var.prefix} Workspace" } # Create AVD host pool resource "azurerm_virtual_desktop_host_pool" "hostpool" { resource_group_name = azurerm_resource_group.sh.name location = azurerm_resource_group.sh.location name = var.hostpool friendly_name = var.hostpool validate_environment = true custom_rdp_properties = "audiocapturemode:i:1;audiomode:i:0;" description = "${var.prefix} Terraform HostPool" type = "Pooled" maximum_sessions_allowed = 16 load_balancer_type = "DepthFirst" #[BreadthFirst DepthFirst] } resource "azurerm_virtual_desktop_host_pool_registration_info" "registrationinfo" { hostpool_id = azurerm_virtual_desktop_host_pool.hostpool.id expiration_date = var.rfc3339 } # Create AVD DAG resource "azurerm_virtual_desktop_application_group" "dag" { resource_group_name = azurerm_resource_group.sh.name host_pool_id = azurerm_virtual_desktop_host_pool.hostpool.id location = azurerm_resource_group.sh.location type = "Desktop" name = "${var.prefix}-dag" friendly_name = "Desktop AppGroup" description = "AVD application group" depends_on = [azurerm_virtual_desktop_host_pool.hostpool, azurerm_virtual_desktop_workspace.workspace] } # Associate Workspace and DAG resource "azurerm_virtual_desktop_workspace_application_group_association" "ws-dag" { application_group_id = azurerm_virtual_desktop_application_group.dag.id workspace_id = azurerm_virtual_desktop_workspace.workspace.id } 4. 创建名为 variables.tf 的文件并插入下列代码: variable "resource_group_location" { default = "eastus" description = "Location of the resource group." } variable "rg_name" { type = string default = "rg-avd-resources" description = "Name of the Resource group in which to deploy service objects" } variable "workspace" { type = string description = "Name of the Azure Virtual Desktop workspace" default = "AVD TF Workspace" } variable "hostpool" { type = string description = "Name of the Azure Virtual Desktop host pool" default = "AVD-TF-HP" } variable "rfc3339" { type = string default = "2022-03-30T12:43:13Z" description = "Registration token expiration" } variable "prefix" { type = string default = "avdtf" description = "Prefix of the name of the AVD machine(s)" } 1. 创建名为 output.tf 的文件并插入下列代码: output "azure_virtual_desktop_compute_resource_group" { description = "Name of the Resource group in which to deploy session host" value = azurerm_resource_group.sh.name } output "azure_virtual_desktop_host_pool" { description = "Name of the Azure Virtual Desktop host pool" value = azurerm_virtual_desktop_host_pool.hostpool.name } output "azurerm_virtual_desktop_application_group" { description = "Name of the Azure Virtual Desktop DAG" value = azurerm_virtual_desktop_application_group.dag.name } output "azurerm_virtual_desktop_workspace" { description = "Name of the Azure Virtual Desktop workspace" value = azurerm_virtual_desktop_workspace.workspace.name } output "location" { description = "The Azure region" value = azurerm_resource_group.rg.location } output "AVD_user_groupname" { description = "Azure Active Directory Group for AVD users" value = azuread_group.aad_group.display_name } 3. 初始化 Terraform 运行 terraform init,将 Terraform 部署进行初始化。 此命令下载管理 Azure 资源所需的 Azure 模块。 terraform init 4. 创建 Terraform 执行计划 运行 terraform plan 以创建执行计划。 terraform plan -out main.tfplan 要点: • terraform plan 命令将创建一个执行计划,但不会执行它。 它会确定创建配置文件中指定的配置需要执行哪些操作。 此模式允许你在对实际资源进行任何更改之前验证执行计划是否符合预期。 • 使用可选 -out 参数可以为计划指定输出文件。 使用 -out 参数可以确保所查看的计划与所应用的计划完全一致。 • 若要详细了解如何使执行计划和安全性持久化,请参阅安全警告一节 5. 应用 Terraform 执行计划 运行 terraform apply,将执行计划应用到云基础结构。 terraform apply main.tfplan 要点: • 上面的 terraform apply 命令假设之前运行了 terraform plan -out main.tfplan • 如果为 -out 参数指定了不同的文件名,请在对 terraform apply 的调用中使用该相同文件名。 • 如果未使用 -out 参数,则无需任何参数即可调用 terraform apply 6. 验证结果 1. 在Azure 门户上,选择 Azure 虚拟桌面 2. 选择 主机池 ,然后选择 池创建资源的名称 3. 选择 “会话主机 ”,然后验证会话主机是否已列出。 7.清理资源 不再需要通过 Terraform 创建的资源时,请执行以下步骤: 1. 运行 terraform plan 并指定 destroy 标志。 terraform plan -destroy -out main.destroy.tfplan 要点: • terraform plan 命令将创建一个执行计划,但不会执行它。 它会确定创建配置文件中指定的配置需要执行哪些操作。 此模式允许你在对实际资源进行任何更改之前验证执行计划是否符合预期。 • 使用可选 -out 参数可以为计划指定输出文件。 使用 -out 参数可以确保所查看的计划与所应用的计划完全一致。 • 若要详细了解如何使执行计划和安全性持久化,请参阅安全警告一节 2. 运行 terraform apply 以应用执行计划。 terraform apply main.destroy.tfplan Azure 上的 Terraform 故障排除 排查在 Azure 上使用 Terraform 时遇到的常见问题 后续步骤
__label__pos
0.998788
Class Reference IRIS for UNIX 2019.2 InterSystems: The power behind what matters    Documentation  Search   [%SYS] >  [%Library] >  [JDBCCatalog] Private  Storage    JDBC Catalog Queries Inventory Parameters Properties Methods Queries Indices ForeignKeys Triggers 74 25 Summary Methods ATClose ATExecute ATFetch BRClose BRExecute BRFetch CAClose CAExecute CAFetch CFClose CFExecute CFFetch CHClose CHExecute CHFetch COClose COExecute COFetch CPClose CPExecute CPFetch CRClose CRExecute CRFetch EKClose EKExecute EKFetch FCClose FCExecute FCFetch FNClose FNExecute FNFetch IIClose IIExecute IIFetch IKClose IKExecute IKFetch MakePat PCClose PCExecute PCFetch PKClose PKExecute PKFetch PRClose PRExecute PRFetch SCClose SCExecute SCFetch SLClose SLExecute SLFetch SYClose SYExecute SYFetch TAClose TAExecute TAFetch TIClose TIExecute TIFetch TPClose TPExecute TPFetch TTExecute TTFetch UTExecute UTFetch VCClose VCExecute VCFetch Methods • classmethod ATClose(ByRef qh As %Binary) as %Status • classmethod ATExecute(ByRef qh As %Binary, catalog As %String, schemaPattern As %String, typeNamePattern As %String, attributeNamePattern As %String) as %Status • classmethod ATFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod BRClose(qh As %Binary) as %Status • classmethod BRExecute(ByRef qh As %Binary, table As %String = " ", schema As %String) as %Status • classmethod BRFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod CAClose(ByRef qh As %Binary) as %Status • classmethod CAExecute(ByRef qh As %Binary) as %Status • classmethod CAFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod CFClose(qh As %Binary) as %Status • classmethod CFExecute(ByRef qh As %Binary) as %Status • classmethod CFFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod CHClose(qh As %Binary) as %Status • classmethod CHExecute(ByRef qh As %Binary, table As %String, column As %String, schema As %String) as %Status • classmethod CHFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod COClose(qh As %Binary) as %Status • classmethod COExecute(ByRef qh As %Binary, table As %String, column As %String, schema As %String) as %Status • classmethod COFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod CPClose(qh As %Binary) as %Status • classmethod CPExecute(ByRef qh As %Binary, table As %String, column As %String, schema As %String) as %Status • classmethod CPFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod CRClose(qh As %Binary) as %Status • classmethod CRExecute(ByRef qh As %Binary, primary As %String, foreign As %String, pkschema As %String, fkschema As %String) as %Status • classmethod CRFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod EKClose(qh As %Binary) as %Status • classmethod EKExecute(ByRef qh As %Binary, table As %String, schema As %String) as %Status • classmethod EKFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod FCClose(qh As %Binary) as %Status • classmethod FCExecute(ByRef qh As %Binary, procedure As %String, column As %String, schema As %String) as %Status • classmethod FCFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod FNClose(qh As %Binary) as %Status • classmethod FNExecute(ByRef qh As %Binary, name As %String, schema As %String) as %Status • classmethod FNFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod IIClose(qh As %Binary) as %Status • classmethod IIExecute(ByRef qh As %Binary, table As %String, schema As %String, nonunique As %SmallInt) as %Status • classmethod IIFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod IKClose(qh As %Binary) as %Status • classmethod IKExecute(ByRef qh As %Binary, table As %String, schema As %String) as %Status • classmethod IKFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod MakePat(like As %String, esc As %String) as %String • classmethod PCClose(qh As %Binary) as %Status • classmethod PCExecute(ByRef qh As %Binary, procedure As %String, column As %String, schema As %String) as %Status • classmethod PCFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod PKClose(qh As %Binary) as %Status • classmethod PKExecute(ByRef qh As %Binary, table As %String, schema As %String) as %Status • classmethod PKFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod PRClose(qh As %Binary) as %Status • classmethod PRExecute(ByRef qh As %Binary, name As %String, schema As %String) as %Status • classmethod PRFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod SCClose(qh As %Binary) as %Status • classmethod SCExecute(ByRef qh As %Binary, schemaPattern As %String = "") as %Status • classmethod SCFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod SLClose(ByRef qh As %Binary) as %Status • classmethod SLExecute(ByRef qh As %Binary, catalog As %String, schemaPattern As %String, typeNamePattern As %String) as %Status • classmethod SLFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod SYClose(ByRef qh As %Binary) as %Status • classmethod SYExecute(ByRef qh As %Binary, catalog As %String, schemaPattern As %String, typeNamePattern As %String) as %Status • classmethod SYFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod TAClose(qh As %Binary) as %Status • classmethod TAExecute(ByRef qh As %Binary, name As %String, ttype As %String, schema As %String) as %Status • classmethod TAFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod TIClose(qh As %Binary) as %Status • classmethod TIExecute(ByRef qh As %Binary) as %Status • classmethod TIFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod TPClose(qh As %Binary) as %Status • classmethod TPExecute(ByRef qh As %Binary, name As %String, schema As %String) as %Status • classmethod TPFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod TTExecute(ByRef qh As %Binary) as %Status • classmethod TTFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod UTExecute(ByRef qh As %Binary) as %Status • classmethod UTFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status • classmethod VCClose(qh As %Binary) as %Status • classmethod VCExecute(ByRef qh As %Binary, table As %String, schema As %String) as %Status • classmethod VCFetch(ByRef qh As %Binary, ByRef Row As %List, ByRef AtEnd As %Integer) as %Status Queries • query AT(catalog As %String(MAXLEN=128), schemaPattern As %String(MAXLEN=128), typeNamePattern As %String(MAXLEN=128), attributeNamePattern As %String(MAXLEN=128)) Selects TYPE_CAT As %String(MAXLEN=128), TYPE_SCHEM As %String(MAXLEN=128), TYPE_NAME As %String(MAXLEN=128), ATTR_NAME As %String(MAXLEN=128), DATA_TYPE As %SmallInt, ATTR_TYPE_NAME As %String(MAXLEN=128), ATTR_SIZE As %Integer, DECIMAL_DIGITS As %Integer, NUM_PREC_RADIX As %Integer, NULLABLE As %Integer, REMARKS As %String(MAXLEN=128), ATTR_DEF As %String(MAXLEN=128), SQL_DATA_TYPE As %Integer, SQL_DATETIME_SUB As %Integer, CHAR_OCTET_LENGTH As %Integer, ORDINAL_POSITION As %Integer, IS_NULLABLE As %String(MAXLEN=128), SCOPE_CATALOG As %String(MAXLEN=128), SCOPE_SCHEMA As %String(MAXLEN=128), SCOPE_TABLE As %String(MAXLEN=128), SOURCE_DATA_TYPE As %SmallInt Retrieves a description of the given attribute of the given type for a user-defined type (UDT) that is available in the given schema and catalog *************************************************************************** %JDBCCatalog_AT Stored Procedure for (JDBC 3.x) getAttributes() (AND246) *************************************************************************** • query BR(table As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects SCOPE As %SmallInt, COLUMN_NAME As %String(MAXLEN=128), DATA_TYPE As %SmallInt, TYPE_NAME As %String(MAXLEN=128), COLUMN_SIZE As %Integer, BUFFER_LENGTH As %Integer, DECIMAL_DIGITS As %SmallInt, PSEUDO_COLUMN As %SmallInt Get a description of a table's optimal set of columns that uniquely identifies a row. *************************************************************************** %JDBCCatalog_BR Stored Procedure for getBestRowIdentifier() *************************************************************************** • query CA() Selects TABLE_CAT As %String(MAXLEN=128) Gets the catalog names available in the database. *************************************************************************** %JDBCCatalog_CA Stored Procedure for getCatalogs() *************************************************************************** • query CF() Selects NAME As %String(MAXLEN=128), MAXLEN As %Integer, DEFAULT_VALUE As %String(MAXLEN=128), DESCRIPTION As %String(MAXLEN=128) Retrieves a list of the client info properties that the driver supports. *************************************************************************** %JDBCCatalog_CF Stored Procedure for getClientInfoProperties() The result set contains the following columns 1. NAME String=> The name of the client info property 2. MAX_LEN int=> The maximum length of the value for the property 3. DEFAULT_VALUE String=> The default value of the property 4. DESCRIPTION String=> A description of the property. This will typically contain information as to where this property is stored in the database. The ResultSet is sorted by the NAME column *************************************************************************** • query CH(table As %String(MAXLEN=128), column As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects TABLE_CAT As %String(MAXLEN=128), TABLE_SCHEM As %String(MAXLEN=128), TABLE_NAME As %String(MAXLEN=128), COLUMN_NAME As %String(MAXLEN=128), DATA_TYPE As %SmallInt, COLUMN_SIZE As %Integer, DECIMAL_DIGITS As %Integer, NUM_PREC_RADIX As %Integer, COLUMN_USAGE As %String, REMARKS As %String(MAXLEN=254), CHAR_OCTET_LENGTH As %Integer, IS_NULLABLE As %String(MAXLEN=3) Get hidden columns info from the catalog, sorted by table name and column name position *************************************************************************** %JDBCCatalog_CH Stored Procedure for getPseudoColumns() Retrieves a description of the pseudo or hidden columns available in a given table within the specified catalog and schema. Pseudo or hidden columns may not always be stored within a table and are not visible in a ResultSet unless they are specified in the query's outermost SELECT list. Pseudo or hidden columns may not necessarily be able to be modified. If there are no pseudo or hidden columns, an empty ResultSet is returned. Only column descriptions matching the catalog, schema, table and column name criteria are returned. They are ordered by TABLE_CAT,TABLE_SCHEM, TABLE_NAME and COLUMN_NAME. Each column description has the following columns: *************************************************************************** • query CO(table As %String(MAXLEN=128), column As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects TABLE_CAT As %String(MAXLEN=128), TABLE_SCHEM As %String(MAXLEN=128), TABLE_NAME As %String(MAXLEN=128), COLUMN_NAME As %String(MAXLEN=128), DATA_TYPE As %SmallInt, TYPE_NAME As %String(MAXLEN=128), COLUMN_SIZE As %Integer, BUFFER_LENGTH As %Integer, DECIMAL_DIGITS As %Integer, NUM_PREC_RADIX As %Integer, NULLABLE As %Integer, REMARKS As %String(MAXLEN=254), COLUMN_DEF As %String(MAXLEN=4096), SQL_DATA_TYPE As %Integer, SQL_DATETIME_SUB As %Integer, CHAR_OCTET_LENGTH As %Integer, ORDINAL_POSITION As %Integer, IS_NULLABLE As %String(MAXLEN=3), SCOPE_CATALOG As %String(MAXLEN=128), SCOPE_SCHEMA As %String(MAXLEN=128), SCOPE_TABLE As %String(MAXLEN=128), SOURCE_DATA_TYPE As %SmallInt, IS_AUTOINCREMENT As %Library.String(MAXLEN=3), IS_GENERATEDCOLUMN As %Library.String(MAXLEN=3) Get columns info from the catalog, sorted by table name and ordinal position *************************************************************************** %JDBCCatalog_CO Stored Procedure for getColumns() ResultSet getColumns(String catalog, String schemaPattern, String tableNamePattern, String columnNamePattern) throws SQLException Retrieves a description of table columns available in the specified catalog. Only column descriptions matching the catalog, schema, table and column name criteria are returned. They are ordered by TABLE_CAT,TABLE_SCHEM, TABLE_NAME, and ORDINAL_POSITION. Each column description has the following columns: 1 - TABLE_CAT String => table catalog (may be null) 2 - TABLE_SCHEM String => table schema (may be null) 3 - TABLE_NAME String => table name 4 - COLUMN_NAME String => column name 5 - DATA_TYPE int => SQL type from java.sql.Types 6 - TYPE_NAME String => Data source dependent type name, for a UDT the type name is fully qualified 7 - COLUMN_SIZE int => column size. 8 - BUFFER_LENGTH is not used. 9 - DECIMAL_DIGITS int => the number of fractional digits. Null is returned for data types where DECIMAL_DIGITS is not applicable. 10 - NUM_PREC_RADIX int => Radix (typically either 10 or 2) 11 - NULLABLE int => is NULL allowed. columnNoNulls - might not allow NULL values columnNullable - definitely allows NULL values columnNullableUnknown - nullability unknown 12 - REMARKS String => comment describing column (may be null) 13 - COLUMN_DEF String => default value for the column, which should be interpreted as a string when the value is enclosed in single quotes (may be null) 14 - SQL_DATA_TYPE int => unused 15 - SQL_DATETIME_SUB int => unused 16 - CHAR_OCTET_LENGTH int => for char types the maximum number of bytes in the column 17 - ORDINAL_POSITION int => index of column in table (starting at 1) 18 - IS_NULLABLE String => ISO rules are used to determine the nullability for a column. YES --- if the column can include NULLs NO --- if the column cannot include NULLs empty string --- if the nullability for the column is unknown 19 - SCOPE_CATALOG String => catalog of table that is the scope of a reference attribute (null if DATA_TYPE isn't REF) 20 - SCOPE_SCHEMA String => schema of table that is the scope of a reference attribute (null if the DATA_TYPE isn't REF) 21 - SCOPE_TABLE String => table name that this the scope of a reference attribute (null if the DATA_TYPE isn't REF) 22 - SOURCE_DATA_TYPE short => source type of a distinct type or user-generated Ref type, SQL type from java.sql.Types (null if DATA_TYPE isn't DISTINCT or user-generated REF) 23 - IS_AUTOINCREMENT String => Indicates whether this column is auto incremented YES --- if the column is auto incremented NO --- if the column is not auto incremented empty string --- if it cannot be determined whether the column is auto incremented 24 - IS_GENERATEDCOLUMN String => Indicates whether this is a generated column YES --- if this a generated column NO --- if this not a generated column empty string --- if it cannot be determined whether this is a generated column The COLUMN_SIZE column specifies the column size for the given column. For numeric data, this is the maximum precision. For character data, this is the length in characters. For datetime datatypes, this is the length in characters of the String representation (assuming the maximum allowed precision of the fractional seconds component). For binary data, this is the length in bytes. For the ROWID datatype, this is the length in bytes. Null is returned for data types where the column size is not applicable. Parameters: catalog - a catalog name; must match the catalog name as it is stored in the database; "" retrieves those without a catalog; null means that the catalog name should not be used to narrow the search schemaPattern - a schema name pattern; must match the schema name as it is stored in the database; "" retrieves those without a schema; null means that the schema name should not be used to narrow the search tableNamePattern - a table name pattern; must match the table name as it is stored in the database columnNamePattern - a column name pattern; must match the column name as it is stored in the database *************************************************************************** • query CP(table As %String(MAXLEN=128), column As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects TABLE_CAT As %String(MAXLEN=128), TABLE_SCHEM As %String(MAXLEN=128), TABLE_NAME As %String(MAXLEN=128), COLUMN_NAME As %String(MAXLEN=128), GRANTOR As %String(MAXLEN=128), GRANTEE As %String(MAXLEN=128), PRIVILEGE As %String(MAXLEN=128), IS_GRANTABLE As %String(MAXLEN=3) Gets column privileges, sorted by column name and privilege *************************************************************************** %JDBCCatalog_CP Stored Procedure for getColumnPrivileges() *************************************************************************** • query CR(primary As %String(MAXLEN=128), foreign As %String(MAXLEN=128), pkschema As %String(MAXLEN=128), fkschema As %String(MAXLEN=128)) Selects PKTABLE_CAT As %String(MAXLEN=128), PKTABLE_SCHEM As %String(MAXLEN=128), PKTABLE_NAME As %String(MAXLEN=128), PKCOLUMN_NAME As %String(MAXLEN=128), FKTABLE_CAT As %String(MAXLEN=128), FKTABLE_SCHEM As %String(MAXLEN=128), FKTABLE_NAME As %String(MAXLEN=128), FKCOLUMN_NAME As %String(MAXLEN=128), KEY_SEQ As %SmallInt, UPDATE_RULE As %SmallInt, DELETE_RULE As %SmallInt, FK_NAME As %String(MAXLEN=128), PK_NAME As %String(MAXLEN=128), DEFERRABILITY As %SmallInt Describes how one table imports the keys of another table. *************************************************************************** %JDBCCatalog_CR Stored Procedure for getCrossReference() *************************************************************************** • query EK(table As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects PKTABLE_CAT As %String, PKTABLE_SCHEM As %String, PKTABLE_NAME As %String, PKCOLUMN_NAME As %String, FKTABLE_CAT As %String, FKTABLE_SCHEM As %String, FKTABLE_NAME As %String, FKCOLUMN_NAME As %String, KEY_SEQ As %SmallInt, UPDATE_RULE As %SmallInt, DELETE_RULE As %SmallInt, FK_NAME As %String, PK_NAME As %String, DEFERRABILITY As %SmallInt Gets a description of the foreign key columns that reference the primary key columns in table TABLE. *************************************************************************** %JDBCCatalog_EK Stored Procedure for getExportedKeys() *************************************************************************** • query FC(procedure As %String(MAXLEN=128), column As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects FUNCTION_CAT As %String(MAXLEN=128), FUNCTION_SCHEM As %String(MAXLEN=128), FUNCTION_NAME As %String(MAXLEN=128), COLUMN_NAME As %String(MAXLEN=128), COLUMN_TYPE As %SmallInt, DATA_TYPE As %SmallInt, TYPE_NAME As %String(MAXLEN=128), PRECISION As %Integer, LENGTH As %Integer, SCALE As %SmallInt, RADIX As %Integer, NULLABLE As %SmallInt, REMARKS As %String(MAXLEN=254), CHAR_OCTET_LENGTH As %Integer, ORDINAL_POSITION As %Integer, IS_NULLABLE As %Library.String(MAXLEN=4), SPECIFIC_NAME As %Library.String(MAXLEN=128) Retrieves a description of the given catalog's system or user function parameters and return type. Only descriptions matching the schema, function and parameter name criteria are returned. They are ordered by FUNCTION_CAT, FUNCTION_SCHEM, FUNCTION_NAME and SPECIFIC_ NAME. Within this, the return value, if any, is first. Next are the parameter descriptions in call order. The column descriptions follow in column number order. *************************************************************************** %JDBCCatalog_FC Stored Procedure for getFunctionColumns() (DPV2968) *************************************************************************** • query FN(name As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects FUNCTION_CAT As %String(MAXLEN=128), FUNCTION_SCHEM As %String(MAXLEN=128), FUNCTION_NAME As %String(MAXLEN=128), REMARKS As %String(MAXLEN=254), FUNCTION_TYPE As %SmallInt, SPECIFIC_NAME As %String(MAXLEN=128) Retrieves a description of the system and user functions available in the given catalog Only system and user function descriptions matching the schema and function name criteria are returned. They are ordered by FUNCTION_CAT, FUNCTION_SCHEM, FUNCTION_NAME and SPECIFIC_ NAME. *************************************************************************** %JDBCCatalog_FN Stored Procedure for getFunctions() (DPV2968) Each function description has the the following columns: 1. FUNCTION_CAT String => function catalog (may be null) 2. FUNCTION_SCHEM String => function schema (may be null) 3. FUNCTION_NAME String => function name. This is the name used to invoke the function 4. REMARKS String => explanatory comment on the function 5. FUNCTION_TYPE short => kind of function: * functionResultUnknown - Cannot determine if a return value or table will be returned * functionNoTable- Does not return a table * functionReturnsTable - Returns a table 6. SPECIFIC_NAME String => the name which uniquely identifies this function within its schema. This is a user specified, or DBMS generated, name that may be different then the FUNCTION_NAME for example with overload functions *************************************************************************** • query II(table As %String(MAXLEN=128), schema As %String(MAXLEN=128), nonunique As %SmallInt(MAXLEN=128)) Selects TABLE_CAT As %String(MAXLEN=128), TABLE_SCHEM As %String(MAXLEN=128), TABLE_NAME As %String(MAXLEN=128), NON_UNIQUE As %SmallInt, INDEX_QUALIFIER As %String(MAXLEN=128), INDEX_NAME As %String(MAXLEN=128), TYPE As %SmallInt, ORDINAL_POSITION As %SmallInt, COLUMN_NAME As %String(MAXLEN=128), ASC_OR_DESC As %String(MAXLEN=1), CARDINALITY As %Integer, PAGES As %Integer, FILTER_CONDITION As %String(MAXLEN=128) Get a description of table's indices and statistics *************************************************************************** %JDBCCatalog_II Stored Procedure for getIndexInfo() *************************************************************************** • query IK(table As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects PKTABLE_CAT As %String(MAXLEN=128), PKTABLE_SCHEM As %String(MAXLEN=128), PKTABLE_NAME As %String(MAXLEN=128), PKCOLUMN_NAME As %String(MAXLEN=128), FKTABLE_CAT As %String(MAXLEN=128), FKTABLE_SCHEM As %String(MAXLEN=128), FKTABLE_NAME As %String(MAXLEN=128), FKCOLUMN_NAME As %String(MAXLEN=128), KEY_SEQ As %SmallInt, UPDATE_RULE As %SmallInt, DELETE_RULE As %SmallInt, FK_NAME As %String(MAXLEN=128), PK_NAME As %String(MAXLEN=128), DEFERRABILITY As %SmallInt Gets a description of the primary key columns that are referenced by the foreign key columns in table TABLE. *************************************************************************** %JDBCCatalog_IK Stored Procedure for getImportedKeys() *************************************************************************** • query PC(procedure As %String(MAXLEN=128), column As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects PROCEDURE_CAT As %String(MAXLEN=128), PROCEDURE_SCHEM As %String(MAXLEN=128), PROCEDURE_NAME As %String(MAXLEN=128), COLUMN_NAME As %String(MAXLEN=128), COLUMN_TYPE As %SmallInt, DATA_TYPE As %SmallInt, TYPE_NAME As %String(MAXLEN=128), PRECISION As %Integer, LENGTH As %Integer, SCALE As %SmallInt, RADIX As %Integer, NULLABLE As %SmallInt, REMARKS As %String(MAXLEN=254), COLUMN_DEF As %String(MAXLEN=254), SQL_DATA_TYPE As %Integer, SQL_DATETIME_SUB As %Integer, CHAR_OCTET_LENGTH As %Integer, ORDINAL_POSITION As %Integer, IS_NULLABLE As %Library.String(MAXLEN=4), SPECIFIC_NAME As %Library.String(MAXLEN=128) Gets a description of the input, output and results associated with certain stored procedures available in the catalog. *************************************************************************** %JDBCCatalog_PC Stored Procedure for getProcedureColumns() *************************************************************************** • query PK(table As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects TABLE_CAT As %String(MAXLEN=128), TABLE_SCHEM As %String(MAXLEN=128), TABLE_NAME As %String(MAXLEN=128), COLUMN_NAME As %String(MAXLEN=128), KEY_SEQ As %SmallInt, PK_NAME As %String(MAXLEN=128) Gets a description of a table's primary key columns. *************************************************************************** %JDBCCatalog_PK Stored Procedure for getPrimaryKeys() *************************************************************************** • query PR(name As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects PROCEDURE_CAT As %String(MAXLEN=128), PROCEDURE_SCHEM As %String(MAXLEN=128), PROCEDURE_NAME As %String(MAXLEN=128), R1 As %String(MAXLEN=1), R2 As %String(MAXLEN=1), R3 As %String(MAXLEN=1), REMARKS As %String(MAXLEN=254), PROCEDURE_TYPE As %SmallInt, SPECIFIC_NAME As %String(MAXLEN=128) Gets a description of the stored procedures available in the catalog. *************************************************************************** %JDBCCatalog_PR Stored Procedure for getProcedures() *************************************************************************** • query SC(schemaPattern As %String(MAXLEN=128)) Selects TABLE_SCHEM As %String(MAXLEN=128), TABLE_CATALOG As %String(MAXLEN=128) Gets the schema names available in the database. *************************************************************************** %JDBCCatalog_SC Stored Procedure for getSchemas() *************************************************************************** • query SL(catalog As %String(MAXLEN=128), schemaPattern As %String(MAXLEN=128), typeNamePattern As %String(MAXLEN=128)) Selects TABLE_CAT As %String(MAXLEN=128), TABLE_SCHEM As %String(MAXLEN=128), TABLE_NAME As %String(MAXLEN=128), SUPERTABLE_NAME As %String(MAXLEN=128) Retrieves a description of the table hierarchies defined in a particular schema in this database *************************************************************************** %JDBCCatalog_SL Stored Procedure for (JDBC 3.x) getSuperTables() (AND246) *************************************************************************** • query SY(catalog As %String(MAXLEN=128), schemaPattern As %String(MAXLEN=128), typeNamePattern As %String(MAXLEN=128)) Selects TYPE_CAT As %String(MAXLEN=128), TYPE_SCHEM As %String(MAXLEN=128), TYPE_NAME As %String(MAXLEN=128), SUPERTYPE_CAT As %String(MAXLEN=128), SUPERTYPE_SCHEM As %String(MAXLEN=128), SUPERTYPE_NAME As %String(MAXLEN=128) Retrieves a description of the user-defined type (UDT) hierarchies defined in a particular schema in this database. Only the immediate super type/sub type relationship is modeled *************************************************************************** %JDBCCatalog_SY Stored Procedure for (JDBC 3.x) getSuperTypes() (AND246) *************************************************************************** • query TA(name As %String(MAXLEN=128), ttype As %String, schema As %String(MAXLEN=128)) Selects TABLE_CAT As %String(MAXLEN=128), TABLE_SCHEM As %String(MAXLEN=128), TABLE_NAME As %String(MAXLEN=128), TABLE_TYPE As %String(MAXLEN=128), REMARKS As %String(MAXLEN=254), TYPE_CAT As %String(MAXLEN=128), TYPE_SCHEM As %String(MAXLEN=128), TYPE_NAME As %String(MAXLEN=128), SELF_REFERENCING_COL_NAME As %String(MAXLEN=128), REF_GENERATION As %String(MAXLEN=7) Gets a description of the tables available in the catalog. *************************************************************************** %JDBCCatalog_TA Stored Procedure for getTables() *************************************************************************** • query TI() Selects TYPE_NAME As %String(MAXLEN=128), DATA_TYPE As %SmallInt, PRECISION As %Integer, LITERAL_PREFIX As %String(MAXLEN=128), LITERAL_SUFFIX As %String(MAXLEN=128), CREATE_PARAMS As %String(MAXLEN=128), NULLABLE As %Integer, CASE_SENSITIVE As %SmallInt, SEARCHABLE As %SmallInt, UNSIGNED_ATTRIBUTE As %SmallInt, FIXED_PREC_SCALE As %SmallInt, AUTO_INCREMENT As %SmallInt, LOCAL_TYPE_NAME As %String(MAXLEN=128), MINIMUM_SCALE As %SmallInt, MAXIMUM_SCALE As %SmallInt, SQL_DATA_TYPE As %Integer, SQL_DATETIME_SUB As %Integer, NUM_PREC_RADIX As %Integer Gets a description of all the datatypes supported by the database. *************************************************************************** %JDBCCatalog_TI Stored Procedure for getTypeInfo() *************************************************************************** • query TP(name As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects TABLE_CAT As %String(MAXLEN=128), TABLE_SCHEM As %String(MAXLEN=128), TABLE_NAME As %String(MAXLEN=128), GRANTOR As %String(MAXLEN=128), GRANTEE As %String(MAXLEN=128), PRIVILEGE As %String(MAXLEN=128), IS_GRANTABLE As %String(MAXLEN=3) Gets a description of the access rights for each table available in the catalog. *************************************************************************** %JDBCCatalog_TP Stored Procedure for getTablePrivileges() *************************************************************************** • query TT() Selects TABLE_TYPE As %String(MAXLEN=128) Gets the table types available in this database. *************************************************************************** %JDBCCatalog_TT Stored Procedure for TableTypes() *************************************************************************** • query UT() Selects TYPE_CAT As %String(MAXLEN=128), TYPE_SCHEM As %String(MAXLEN=128), TYPE_NAME As %String(MAXLEN=128), CLASS_NAME As %String(MAXLEN=128), DATA_TYPE As %SmallInt, REMARKS As %String(MAXLEN=254), BASE_TYPE As %SmallInt Gets a description of UDTs. *************************************************************************** %JDBCCatalog_UT Stored Procedure for getUDTs() *************************************************************************** • query VC(table As %String(MAXLEN=128), schema As %String(MAXLEN=128)) Selects SCOPE As %SmallInt, COLUMN_NAME As %String(MAXLEN=128), DATA_TYPE As %SmallInt, TYPE_NAME As %String(MAXLEN=128), COLUMN_SIZE As %Integer, BUFFER_LENGTH As %Integer, DECIMAL_DIGITS As %SmallInt, PSEUDO_COLUMN As %SmallInt Gets a description of the columns in a table that are automatically updated when any value in a row is updated. *************************************************************************** %JDBCCatalog_VC Stored Procedure for getVersionColumns() *************************************************************************** Copyright (c) 2019 by InterSystems Corporation. Cambridge, Massachusetts, U.S.A. All rights reserved. Confidential property of InterSystems Corporation.
__label__pos
0.835621
_n( string $single, string $plural, int $number, string $domain = 'default' ) Translates and retrieves the singular or plural form based on the supplied number. Description Used when you want to use the appropriate form of a string based on whether a number is singular or plural. Example: printf( _n( '%s person', '%s people', $count, 'text-domain' ), number_format_i18n( $count ) ); Top ↑ Parameters $single (string) (Required) The text to be used if the number is singular. $plural (string) (Required) The text to be used if the number is plural. $number (int) (Required) The number to compare against to use either the singular or plural form. $domain (string) (Optional) Text domain. Unique identifier for retrieving translated strings. Default value: 'default' Top ↑ Return (string) The translated singular or plural form. Top ↑ Source File: wp-includes/l10n.php function _n( $single, $plural, $number, $domain = 'default' ) { $translations = get_translations_for_domain( $domain ); $translation = $translations->translate_plural( $single, $plural, $number ); /** * Filters the singular or plural form of a string. * * @since 2.2.0 * * @param string $translation Translated text. * @param string $single The text to be used if the number is singular. * @param string $plural The text to be used if the number is plural. * @param string $number The number to compare against to use either the singular or plural form. * @param string $domain Text domain. Unique identifier for retrieving translated strings. */ $translation = apply_filters( 'ngettext', $translation, $single, $plural, $number, $domain ); /** * Filters the singular or plural form of a string for a domain. * * The dynamic portion of the hook name, `$domain`, refers to the text domain. * * @since 5.5.0 * * @param string $translation Translated text. * @param string $single The text to be used if the number is singular. * @param string $plural The text to be used if the number is plural. * @param string $number The number to compare against to use either the singular or plural form. * @param string $domain Text domain. Unique identifier for retrieving translated strings. */ $translation = apply_filters( "ngettext_{$domain}", $translation, $single, $plural, $number, $domain ); return $translation; } Top ↑ Changelog Changelog Version Description 5.5.0 Introduced ngettext-{$domain} filter. 2.8.0 Introduced. Top ↑ User Contributed Notes 1. Skip to note 1 content Contributed by Codex Display either “1 star” or “x stars” for a star rating plugin. $rating = '3'; $text = sprintf( _n( '%s star', '%s stars', $rating, 'wpdocs_textdomain' ), $rating ); // "3 stars" echo $text; Important: Never do a calculation inside the sprintf() function! The following won’t work: $text = sprintf( _n( '%s star', '%s stars', $rating, 'wpdocs_textdomain' ), 2 <= $rating ? $rating -1 : $rating ); 2. Skip to note 2 content Contributed by Felipe Elia As explained by @sergeybiryukov here, here, here and here, this functions should NOT be use in one item or more than one item scenarios, but for singular form and plural forms, which is not the same thing. Also both strings should have placeholders. As he stated: In languages that have complex plural structures the singular form can be used for numbers other than 1, so if the goal is to display a string for exactly 1 item, an explicit check like 1 === count( $items ) should be used instead of _n() This problem is also covered in Codex. Although it would be possible address this issue changing the plural form for Russian and other slavic languages it would bring a too big effort as explained here. You must log in before being able to contribute a note or feedback.
__label__pos
0.975953
The Birdfont Source Code All Repositories / birdfont.git / blob – RSS feed KerningDisplay.vala in libbirdfont This file is a part of the Birdfont project. Contributing Send patches or pull requests to [email protected]. Clone this repository: git clone https://github.com/johanmattssonm/birdfont.git Revisions View the latest version of libbirdfont/KerningDisplay.vala. Thread safety in text rendering 1 /* 2 Copyright (C) 2012, 2014 Johan Mattsson 3 4 This library is free software; you can redistribute it and/or modify 5 it under the terms of the GNU Lesser General Public License as 6 published by the Free Software Foundation; either version 3 of the 7 License, or (at your option) any later version. 8 9 This library is distributed in the hope that it will be useful, but 10 WITHOUT ANY WARRANTY; without even the implied warranty of 11 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 12 Lesser General Public License for more details. 13 */ 14 15 using Cairo; 16 17 namespace BirdFont { 18 19 /** Kerning context. */ 20 public class KerningDisplay : FontDisplay { 21 22 public bool suppress_input = false; 23 24 Gee.ArrayList <GlyphSequence> row; 25 int active_handle = -1; 26 int selected_handle = -1; 27 bool moving = false; 28 Glyph left_active_glyph = new Glyph ("null", '\0'); 29 Glyph right_active_glyph = new Glyph ("null", '\0'); 30 31 double begin_handle_x = 0; 32 double begin_handle_y = 0; 33 34 double last_handle_x = 0; 35 36 public bool text_input = false; 37 38 Gee.ArrayList<UndoItem> undo_items; 39 Gee.ArrayList<UndoItem> redo_items; 40 bool first_update = true; 41 42 Font current_font = new Font (); 43 Text kerning_label = new Text (); 44 45 public bool adjust_side_bearings = false; 46 bool right_side_bearing = true; 47 48 public KerningDisplay () { 49 GlyphSequence w = new GlyphSequence (); 50 row = new Gee.ArrayList <GlyphSequence> (); 51 undo_items = new Gee.ArrayList <UndoItem> (); 52 redo_items = new Gee.ArrayList <UndoItem> (); 53 row.add (w); 54 } 55 56 public GlyphSequence get_first_row () { 57 return row.size > 0 ? row.get (0) : new GlyphSequence (); 58 } 59 60 public override string get_label () { 61 return t_("Kerning"); 62 } 63 64 public override string get_name () { 65 return "Kerning"; 66 } 67 68 public void show_parse_error () { 69 string line1 = t_("The current kerning class is malformed."); 70 string line2 = t_("Add single characters separated by space and ranges on the form A-Z."); 71 string line3 = t_("Type “space” to kern the space character and “divis” to kern -."); 72 73 MainWindow.show_dialog (new MessageDialog (line1 + " " + line2 + " " + line3)); 74 } 75 76 public override void draw (WidgetAllocation allocation, Context cr) { 77 draw_kerning_pairs (allocation, cr); 78 } 79 80 public double get_row_height () { 81 return current_font.top_limit - current_font.bottom_limit; 82 } 83 84 public void draw_kerning_pairs (WidgetAllocation allocation, Context cr) { 85 Glyph glyph; 86 double x, y, w, kern, alpha; 87 double x2; 88 double caret_y; 89 int i, wi; 90 Glyph? prev; 91 GlyphSequence word_with_ligatures; 92 GlyphRange? gr_left, gr_right; 93 bool first_row = true; 94 double row_height; 95 Font font; 96 double item_size = 1.0 / KerningTools.font_size; 97 double item_size2 = 2.0 / KerningTools.font_size; 98 99 font = current_font; 100 i = 0; 101 102 // bg color 103 cr.save (); 104 cr.set_source_rgba (1, 1, 1, 1); 105 cr.rectangle (0, 0, allocation.width, allocation.height); 106 cr.fill (); 107 cr.restore (); 108 109 cr.save (); 110 cr.scale (KerningTools.font_size, KerningTools.font_size); 111 112 glyph = MainWindow.get_current_glyph (); 113 114 row_height = get_row_height (); 115 116 alpha = 1; 117 y = get_row_height () + font.base_line + 20; 118 x = 20; 119 w = 0; 120 prev = null; 121 kern = 0; 122 123 foreach (GlyphSequence word in row) { 124 wi = 0; 125 word_with_ligatures = word.process_ligatures (); 126 gr_left = null; 127 gr_right = null; 128 foreach (Glyph? g in word_with_ligatures.glyph) { 129 if (g == null) { 130 continue; 131 } 132 133 if (prev == null || wi == 0) { 134 kern = 0; 135 } else { 136 return_if_fail (wi < word_with_ligatures.ranges.size); 137 return_if_fail (wi - 1 >= 0); 138 139 gr_left = word_with_ligatures.ranges.get (wi - 1); 140 gr_right = word_with_ligatures.ranges.get (wi); 141 142 kern = get_kerning_for_pair (((!)prev).get_name (), ((!)g).get_name (), gr_left, gr_right); 143 } 144 145 // draw glyph 146 if (g == null) { 147 w = 50; 148 alpha = 1; 149 } else { 150 alpha = 0; 151 glyph = (!) g; 152 153 cr.save (); 154 glyph.add_help_lines (); 155 cr.translate (kern + x - glyph.get_lsb () - Glyph.xc (), glyph.get_baseline () + y - Glyph.yc ()); 156 glyph.draw_paths (cr); 157 cr.restore (); 158 159 w = glyph.get_width (); 160 } 161 162 // handle 163 if (first_row && (active_handle == i || selected_handle == i)) { 164 x2 = x + kern / 2.0; 165 166 cr.save (); 167 168 if (selected_handle == i) { 169 cr.set_source_rgba (0, 0, 0, 1); 170 } else { 171 cr.set_source_rgba (123/255.0, 123/255.0, 123/255.0, 1); 172 } 173 174 if (!adjust_side_bearings) { 175 cr.move_to (x2 - 5 * item_size, y + 20 * item_size); 176 cr.line_to (x2, y + 20 * item_size - 5 * item_size); 177 cr.line_to (x2 + 5 * item_size, y + 20* item_size); 178 cr.fill (); 179 180 if (gr_left != null || gr_right != null) { 181 cr.move_to (x2 - 5 * item_size, y + 20 * item_size); 182 cr.line_to (x2 + 5 * item_size, y + 20 * item_size); 183 cr.line_to (x2 + 5 * item_size, y + 24 * item_size); 184 cr.line_to (x2 - 5 * item_size, y + 24 * item_size); 185 cr.fill (); 186 } 187 } else { 188 if (right_side_bearing) { 189 cr.move_to (x2 - 5 * item_size2, y + 20 * item_size2); 190 cr.line_to (x2, y + 20 * item_size2 - 5 * item_size2); 191 cr.line_to (x2, y + 20* item_size2); 192 cr.fill (); 193 } else { 194 cr.move_to (x2, y + 20 * item_size2); 195 cr.line_to (x2, y + 20 * item_size2 - 5 * item_size2); 196 cr.line_to (x2 + 5 * item_size2, y + 20* item_size2); 197 cr.fill (); 198 } 199 } 200 201 if (active_handle == i && !adjust_side_bearings) { 202 cr.save (); 203 cr.scale (1 / KerningTools.font_size, 1 / KerningTools.font_size); 204 kerning_label.widget_x = x2 * KerningTools.font_size; 205 kerning_label.widget_y = y * KerningTools.font_size + 40; 206 kerning_label.draw (cr); 207 cr.fill (); 208 cr.restore (); 209 } 210 } 211 212 x += w + kern; 213 214 // caption 215 if (g == null || ((!)g).is_empty ()) { 216 cr.save (); 217 cr.set_source_rgba (153/255.0, 153/255.0, 153/255.0, alpha); 218 cr.move_to (x - w / 2.0 - 5, y + 20); 219 cr.set_font_size (10 * item_size); 220 cr.show_text ("?"); 221 cr.restore (); 222 } 223 224 prev = g; 225 226 wi++; 227 i++; 228 } 229 230 // draw caret 231 if (first_row) { 232 x2 = x; 233 caret_y = get_row_height () + font.base_line + 20; 234 cr.save (); 235 cr.set_line_width (1.0 / KerningTools.font_size); 236 cr.set_source_rgba (0, 0, 0, 0.5); 237 cr.move_to (x2, caret_y + 20); 238 cr.line_to (x2, 20); 239 cr.stroke (); 240 cr.restore (); 241 } 242 243 y += row_height + 20; 244 x = 20; 245 first_row = false; 246 247 if (y > allocation.height) { 248 break; 249 } 250 } 251 252 for (int j = row.size - 1; j > 30; j--) { 253 row.remove_at (j); 254 } 255 256 cr.restore (); 257 } 258 259 private void display_kerning_value (double k) { 260 string kerning = round (k); 261 kerning_label = new Text (@"$(kerning)", 17 * MainWindow.units); 262 } 263 264 private void set_active_handle_index (int h) { 265 double kern = get_kerning_for_handle (h); 266 active_handle = h; 267 268 if (1 <= active_handle < row.get (0).glyph.size) { 269 display_kerning_value (kern); 270 } 271 } 272 273 private double get_kerning_for_handle (int handle) { 274 string a, b; 275 GlyphRange? gr_left, gr_right; 276 bool got_pair; 277 278 got_pair = get_kerning_pair (handle, out a, out b, out gr_left, out gr_right); 279 280 if (got_pair) { 281 return get_kerning_for_pair (a, b, gr_left, gr_right); 282 } 283 284 return 0; 285 } 286 287 private bool get_kerning_pair (int handle, out string left, out string right, 288 out GlyphRange? range_left, out GlyphRange? range_right) { 289 string a, b; 290 Font font; 291 int wi = 0; 292 GlyphSequence word_with_ligatures; 293 int ranges_index = 0; 294 GlyphRange? gr_left, gr_right; 295 int row_index = 0; 296 297 font = current_font; 298 299 font.touch (); 300 301 a = ""; 302 b = ""; 303 304 left = ""; 305 right = ""; 306 range_left = null; 307 range_right = null; 308 309 if (handle <= 0) { 310 return false; 311 } 312 313 foreach (GlyphSequence word in row) { 314 word_with_ligatures = word.process_ligatures (); 315 ranges_index = 0; 316 foreach (Glyph? g in word_with_ligatures.glyph) { 317 318 if (g == null) { 319 continue; 320 } 321 322 b = ((!) g).get_name (); 323 324 if (handle == wi && row_index == 0) { 325 if (wi >= word_with_ligatures.ranges.size) { 326 return false; 327 } 328 return_val_if_fail (wi - 1 >= 0, false); 329 330 if (word_with_ligatures.ranges.size != word_with_ligatures.glyph.size) { 331 return false; 332 } 333 334 gr_left = word_with_ligatures.ranges.get (wi - 1); 335 gr_right = word_with_ligatures.ranges.get (wi); 336 337 left = a; 338 right = b; 339 range_left = gr_left; 340 range_right = gr_right; 341 342 return true; 343 } 344 345 wi++; 346 347 a = b; 348 } 349 350 row_index++; 351 } 352 353 return false; 354 } 355 356 public void set_absolute_kerning (int handle, double val) { 357 double kern; 358 359 if (MenuTab.suppress_event) { 360 return; 361 } 362 363 if (!adjust_side_bearings) { 364 kern = get_kerning_for_handle (handle); 365 set_space (handle, val - kern); 366 } 367 } 368 369 370 /** Adjust kerning or right side bearing. */ 371 private void set_space (int handle, double val) { 372 string a, b; 373 Font font; 374 GlyphRange? gr_left, gr_right; 375 376 font = current_font; 377 font.touch (); 378 379 if (!adjust_side_bearings) { 380 get_kerning_pair (handle, out a, out b, out gr_left, out gr_right); 381 set_kerning_pair (a, b, ref gr_left, ref gr_right, val); 382 } else { 383 if (right_side_bearing) { 384 left_active_glyph.right_limit += val; 385 left_active_glyph.remove_lines (); 386 left_active_glyph.add_help_lines (); 387 left_active_glyph.update_other_spacing_classes (); 388 } else { 389 right_active_glyph.left_limit -= val; 390 right_active_glyph.remove_lines (); 391 right_active_glyph.add_help_lines (); 392 right_active_glyph.update_other_spacing_classes (); 393 } 394 } 395 } 396 397 /** Class based gpos kerning. */ 398 public void set_kerning_pair (string a, string b, 399 ref GlyphRange? gr_left, ref GlyphRange? gr_right, 400 double val) { 401 double kern; 402 GlyphRange grl, grr; 403 KerningClasses classes; 404 string n, f; 405 bool has_kerning; 406 Font font; 407 408 font = current_font; 409 font.touch (); 410 classes = font.get_kerning_classes (); 411 412 kern = get_kerning_for_pair (a, b, gr_left, gr_right); 413 414 try { 415 if (gr_left == null) { 416 grl = new GlyphRange (); 417 grl.parse_ranges (a); 418 gr_left = grl; // update the range list 419 } else { 420 grl = (!) gr_left; 421 } 422 423 if (gr_right == null) { 424 grr = new GlyphRange (); 425 grr.parse_ranges (b); 426 gr_right = grr; 427 } else { 428 grr = (!) gr_right; 429 } 430 431 if (first_update) { 432 f = grl.get_all_ranges (); 433 n = grr.get_all_ranges (); 434 has_kerning = classes.has_kerning (f, n); 435 undo_items.add (new UndoItem (f, n, kern, has_kerning)); 436 redo_items.clear (); 437 first_update = false; 438 } 439 440 classes.set_kerning (grl, grr, kern + val); 441 display_kerning_value (kern + val); 442 } catch (MarkupError e) { 443 // FIXME: unassigned glyphs and ligatures 444 warning (e.message); 445 } 446 } 447 448 public static double get_kerning_for_pair (string a, string b, GlyphRange? gr_left, GlyphRange? gr_right) { 449 KerningClasses k = BirdFont.get_current_font ().get_kerning_classes (); 450 return k.get_kerning_for_pair (a, b, gr_left, gr_right); 451 } 452 453 public void set_current_font (Font f) { 454 current_font = f; 455 } 456 457 public override void selected_canvas () { 458 Glyph g; 459 GlyphSequence w; 460 StringBuilder s = new StringBuilder (); 461 bool append_char = false; 462 463 current_font = BirdFont.get_current_font (); 464 465 KeyBindings.set_require_modifier (true); 466 467 g = MainWindow.get_current_glyph (); 468 s.append_unichar (g.get_unichar ()); 469 470 if (row.size == 0) { 471 append_char = true; 472 } 473 474 if (append_char) { 475 w = new GlyphSequence (); 476 row.add (w); 477 w.glyph.insert (0, current_font.get_glyph (s.str)); 478 } 479 } 480 481 public void add_kerning_class (int index) { 482 add_range (KerningTools.get_kerning_class (index)); 483 } 484 485 public void add_range (GlyphRange range) { 486 Font font = current_font; 487 Glyph? glyph; 488 489 glyph = font.get_glyph_by_name (range.get_char (0)); 490 491 if (glyph == null) { 492 warning ("Kerning range is not represented by a valid glyph."); 493 return; 494 } 495 496 row.get (0).glyph.add ((!) glyph); 497 row.get (0).ranges.add (range); 498 499 GlyphCanvas.redraw (); 500 } 501 502 void set_selected_handle (int handle) { 503 Glyph? g; 504 selected_handle = handle; 505 GlyphSequence sequence_with_ligatures; 506 507 sequence_with_ligatures = row.get (0).process_ligatures (); 508 509 if (selected_handle <= 0) { 510 selected_handle = 1; 511 } 512 513 if (selected_handle >= sequence_with_ligatures.glyph.size) { 514 selected_handle = (int) sequence_with_ligatures.glyph.size - 1; 515 } 516 517 set_active_handle_index (handle); 518 519 if (0 <= selected_handle - 1 < sequence_with_ligatures.glyph.size) { 520 g = sequence_with_ligatures.glyph.get (selected_handle - 1); 521 if (g != null) { 522 left_active_glyph = (!) g; 523 } 524 } 525 526 if (0 <= selected_handle < sequence_with_ligatures.glyph.size) { 527 g = sequence_with_ligatures.glyph.get (selected_handle); 528 if (g != null) { 529 right_active_glyph = (!) g; 530 } 531 } 532 533 GlyphCanvas.redraw (); 534 } 535 536 public static void previous_pair () { 537 KerningDisplay d = MainWindow.get_kerning_display (); 538 d.set_selected_handle (d.selected_handle - 1); 539 } 540 541 public static void next_pair () { 542 KerningDisplay d = MainWindow.get_kerning_display (); 543 d.set_selected_handle (d.selected_handle + 1); 544 } 545 546 private static string round (double d) { 547 char[] b = new char [22]; 548 unowned string s = d.format (b, "%.2f"); 549 string n = s.dup (); 550 551 n = n.replace (",", "."); 552 553 if (n == "-0.00") { 554 n = "0.00"; 555 } 556 557 return n; 558 } 559 560 public override void key_press (uint keyval) { 561 unichar c; 562 563 if (MenuTab.suppress_event) { // don't update kerning while saving font 564 warning ("A background thread uses the current font."); 565 return; 566 } 567 568 c = (unichar) keyval; 569 570 if (suppress_input) { 571 return; 572 } 573 574 if ((keyval == 'u' || keyval == 'U') && KeyBindings.has_ctrl ()) { 575 insert_unichar (); 576 } else { 577 if (keyval == Key.LEFT && KeyBindings.modifier == NONE) { 578 first_update = true; 579 set_space (selected_handle, -1 / KerningTools.font_size); 580 } 581 582 if (keyval == Key.RIGHT && KeyBindings.modifier == NONE) { 583 first_update = true; 584 set_space (selected_handle, 1 / KerningTools.font_size); 585 } 586 587 if (KeyBindings.modifier == CTRL && (keyval == Key.LEFT || keyval == Key.RIGHT)) { 588 if (keyval == Key.LEFT) { 589 selected_handle--; 590 } 591 592 if (keyval == Key.RIGHT) { 593 selected_handle++; 594 } 595 596 set_selected_handle (selected_handle); 597 } 598 599 if (KeyBindings.modifier == NONE 600 || KeyBindings.modifier == SHIFT 601 || KeyBindings.modifier == ALT) { 602 603 if (keyval == Key.BACK_SPACE && row.size > 0 && row.get (0).glyph.size > 0) { 604 row.get (0).glyph.remove_at (row.get (0).glyph.size - 1); 605 row.get (0).ranges.remove_at (row.get (0).ranges.size - 1); 606 } 607 608 if (row.size == 0 || c == Key.ENTER) { 609 new_line (); 610 } 611 612 add_character (c); 613 } 614 } 615 616 GlyphCanvas.redraw (); 617 } 618 619 public void insert_unichar () { 620 TextListener listener; 621 string submitted_value = ""; 622 string unicodestart; 623 624 unicodestart = (KeyBindings.has_shift ()) ? "" : "U+"; 625 626 listener = new TextListener (t_("Unicode"), unicodestart, t_("Insert")); 627 628 listener.signal_text_input.connect ((text) => { 629 submitted_value = text; 630 631 if (MenuTab.suppress_event) { 632 return; 633 } 634 635 GlyphCanvas.redraw (); 636 }); 637 638 listener.signal_submit.connect (() => { 639 unichar c; 640 MainWindow.native_window.hide_text_input (); 641 642 text_input = false; 643 suppress_input = false; 644 645 if (submitted_value.has_prefix ("u+") || submitted_value.has_prefix ("U+")) { 646 c = Font.to_unichar (submitted_value); 647 add_character (c); 648 } else { 649 add_text (submitted_value); 650 } 651 }); 652 653 suppress_input = true; 654 text_input = true; 655 MainWindow.native_window.set_text_listener (listener); 656 } 657 658 public void new_line () { 659 row.insert (0, new GlyphSequence ()); 660 } 661 662 void add_character (unichar c) { 663 Glyph? g; 664 string name; 665 Font f; 666 667 if (MenuTab.suppress_event) { 668 return; 669 } 670 671 f = current_font; 672 673 if (!is_modifier_key (c) && c.validate ()) { 674 name = f.get_name_for_character (c); 675 g = f.get_glyph_by_name (name); 676 inser_glyph (g); 677 } 678 } 679 680 public void inser_glyph (Glyph? g) { 681 if (g != null) { 682 row.get (0).glyph.add (g); 683 row.get (0).ranges.add (null); 684 685 set_selected_handle ((int) row.get (0).glyph.size - 1); 686 set_active_handle_index (selected_handle); 687 } 688 } 689 690 public override void motion_notify (double ex, double ey) { 691 double k, y; 692 693 if (MenuTab.suppress_event) { 694 return; 695 } 696 697 if (!moving) { 698 set_active_handle (ex, ey); 699 } else { 700 y = 1; 701 702 if (Math.fabs (ey - begin_handle_y) > 20) { 703 y = ((Math.fabs (ey - begin_handle_y) / 100) + 1); 704 } 705 706 k = (ex - last_handle_x) / y; // y-axis is for variable precision 707 k /= KerningTools.font_size; 708 set_space (selected_handle, k); 709 GlyphCanvas.redraw (); 710 } 711 712 last_handle_x = ex; 713 } 714 715 public void set_active_handle (double ex, double ey) { 716 double item_size = 1.0 / KerningTools.font_size; 717 double y = 100 * item_size; 718 double x = 20; 719 double w = 0; 720 double d, kern; 721 double min = double.MAX; 722 int i = 0; 723 int row_index = 0; 724 int col_index = 0; 725 double fs = KerningTools.font_size; 726 Glyph glyph = new Glyph.no_lines (""); 727 728 GlyphRange? gr_left, gr_right; 729 730 Glyph? prev = null; 731 string gl_name = ""; 732 GlyphSequence word_with_ligatures; 733 734 foreach (GlyphSequence word in row) { 735 col_index = 0; 736 737 word_with_ligatures = word.process_ligatures (); 738 foreach (Glyph? g in word_with_ligatures.glyph) { 739 if (g == null) { 740 w = 50; 741 warning ("glyph does not exist"); 742 } else { 743 glyph = (!) g; 744 w = glyph.get_width (); 745 } 746 747 gl_name = glyph.get_name (); 748 749 if (prev == null && col_index != 0) { 750 warning (@"previous glyph does not exist row: $row_index column: $col_index"); 751 } 752 753 if (prev == null || col_index == 0) { 754 kern = 0; 755 } else { 756 return_if_fail (col_index < word_with_ligatures.ranges.size); 757 return_if_fail (col_index - 1 >= 0); 758 759 gr_left = word_with_ligatures.ranges.get (col_index - 1); 760 gr_right = word_with_ligatures.ranges.get (col_index); 761 762 kern = get_kerning_for_pair (((!)prev).get_name (), ((!)g).get_name (), gr_left, gr_right); 763 } 764 765 d = Math.pow (fs * (x + kern) - ex, 2) + Math.pow (fs * (y - ey), 2); 766 767 if (d < min) { 768 min = d; 769 770 if (ex != fs * (x + kern)) { // don't swap direction after button release 771 right_side_bearing = ex < fs * (x + kern); // right or left side bearing handle 772 } 773 774 if (active_handle != i - row_index) { 775 set_active_handle_index (i - row_index); 776 GlyphCanvas.redraw (); 777 } 778 779 if (col_index == word.glyph.size || col_index == 0) { 780 set_active_handle_index (-1); 781 } else { 782 set_active_handle_index (active_handle + row_index); 783 } 784 } 785 786 prev = g; 787 x += w + kern; 788 i++; 789 col_index++; 790 } 791 792 row_index++; 793 y += MainWindow.get_current_glyph ().get_height () + 20; 794 x = 20; 795 } 796 } 797 798 public override void button_release (int button, double ex, double ey) { 799 set_active_handle (ex, ey); 800 moving = false; 801 first_update = true; 802 803 if (button == 3 || text_input) { 804 set_kerning_by_text (); 805 } 806 } 807 808 public void set_kerning_by_text () { 809 TextListener listener; 810 string kerning = @"$(get_kerning_for_handle (selected_handle))"; 811 812 if (MenuTab.suppress_event) { 813 return; 814 } 815 816 if (selected_handle == -1) { 817 set_selected_handle (0); 818 } 819 820 listener = new TextListener (t_("Kerning"), kerning, t_("Close")); 821 822 listener.signal_text_input.connect ((text) => { 823 string submitted_value; 824 double parsed_value; 825 826 if (MenuTab.suppress_event) { 827 return; 828 } 829 830 submitted_value = text.replace (",", "."); 831 parsed_value = double.parse (submitted_value); 832 set_absolute_kerning (selected_handle, parsed_value); 833 GlyphCanvas.redraw (); 834 }); 835 836 listener.signal_submit.connect (() => { 837 MainWindow.native_window.hide_text_input (); 838 text_input = false; 839 suppress_input = false; 840 }); 841 842 suppress_input = true; 843 text_input = true; 844 MainWindow.native_window.set_text_listener (listener); 845 846 GlyphCanvas.redraw (); 847 } 848 849 public override void button_press (uint button, double ex, double ey) { 850 set_active_handle (ex, ey); 851 set_selected_handle (active_handle); 852 begin_handle_x = ex; 853 begin_handle_y = ey; 854 last_handle_x = ex; 855 moving = true; 856 } 857 858 /** Insert text form clipboard. */ 859 public void add_text (string t) { 860 int c; 861 862 if (MenuTab.suppress_event) { 863 return; 864 } 865 866 c = t.char_count (); 867 for (int i = 0; i <= c; i++) { 868 add_character (t.get_char (t.index_of_nth_char (i))); 869 } 870 871 GlyphCanvas.redraw (); 872 } 873 874 public override void undo () { 875 UndoItem ui; 876 UndoItem redo_state; 877 878 if (MenuTab.suppress_event) { 879 return; 880 } 881 882 if (undo_items.size == 0) { 883 return; 884 } 885 886 ui = undo_items.get (undo_items.size - 1); 887 888 redo_state = apply_undo (ui); 889 redo_items.add (redo_state); 890 891 undo_items.remove_at (undo_items.size - 1); 892 } 893 894 public override void redo () { 895 UndoItem ui; 896 897 if (MenuTab.suppress_event) { 898 return; 899 } 900 901 if (redo_items.size == 0) { 902 return; 903 } 904 905 ui = redo_items.get (redo_items.size - 1); 906 apply_undo (ui); 907 redo_items.remove_at (redo_items.size - 1); 908 } 909 910 /** @return redo state. */ 911 public UndoItem apply_undo (UndoItem ui) { 912 KerningClasses classes = BirdFont.get_current_font ().get_kerning_classes (); 913 GlyphRange glyph_range_first, glyph_range_next; 914 Font font = current_font; 915 string l, r; 916 UndoItem redo_state = new UndoItem ("", "", 0, false); 917 double? k; 918 919 l = GlyphRange.unserialize (ui.first); 920 r = GlyphRange.unserialize (ui.next); 921 922 try { 923 glyph_range_first = new GlyphRange (); 924 glyph_range_next = new GlyphRange (); 925 926 glyph_range_first.parse_ranges (ui.first); 927 glyph_range_next.parse_ranges (ui.next); 928 929 if (!ui.has_kerning) { 930 if (glyph_range_first.is_class () || glyph_range_next.is_class ()) { 931 redo_state.first = glyph_range_first.get_all_ranges (); 932 redo_state.next = glyph_range_next.get_all_ranges (); 933 redo_state.has_kerning = true; 934 redo_state.kerning = classes.get_kerning_for_range (glyph_range_first, glyph_range_next); 935 936 classes.delete_kerning_for_class (ui.first, ui.next); 937 } else { 938 939 redo_state.first = ui.first; 940 redo_state.next = ui.next; 941 redo_state.has_kerning = true; 942 k = classes.get_kerning_for_single_glyphs (ui.first, ui.next); 943 944 if (k != null) { 945 redo_state.kerning = (!) k; 946 } else { 947 warning ("No kerning"); 948 } 949 950 classes.delete_kerning_for_pair (ui.first, ui.next); 951 } 952 } else if (glyph_range_first.is_class () || glyph_range_next.is_class ()) { 953 glyph_range_first = new GlyphRange (); 954 glyph_range_next = new GlyphRange (); 955 956 glyph_range_first.parse_ranges (ui.first); 957 glyph_range_next.parse_ranges (ui.next); 958 959 redo_state.first = glyph_range_first.get_all_ranges (); 960 redo_state.next = glyph_range_next.get_all_ranges (); 961 k = classes.get_kerning_for_range (glyph_range_first, glyph_range_next); 962 963 if (k != null) { 964 redo_state.kerning = (!) k; 965 redo_state.has_kerning = true; 966 } else { 967 redo_state.has_kerning = false; 968 } 969 970 classes.set_kerning (glyph_range_first, glyph_range_next, ui.kerning); 971 } else { 972 redo_state.first = ui.first; 973 redo_state.next = ui.next; 974 redo_state.has_kerning = true; 975 k = classes.get_kerning_for_single_glyphs (ui.first, ui.next); 976 977 if (k != null) { 978 redo_state.kerning = (!) k; 979 redo_state.has_kerning = true; 980 } else { 981 redo_state.has_kerning = false; 982 } 983 984 classes.set_kerning_for_single_glyphs (ui.first, ui.next, ui.kerning); 985 } 986 } catch (MarkupError e) { 987 warning (e.message); 988 } 989 990 font.touch (); 991 GlyphCanvas.redraw (); 992 993 return redo_state; 994 } 995 996 public override void zoom_in () { 997 KerningTools.font_size += 0.1; 998 999 if (KerningTools.font_size > 3) { 1000 KerningTools.font_size = 3; 1001 } 1002 1003 KerningTools.zoom_bar.set_zoom (KerningTools.font_size / 3); 1004 GlyphCanvas.redraw (); 1005 } 1006 1007 public override void zoom_out () { 1008 KerningTools.font_size -= 0.1; 1009 1010 if (KerningTools.font_size < 0.3) { 1011 KerningTools.font_size = 0.3; 1012 } 1013 1014 KerningTools.zoom_bar.set_zoom (KerningTools.font_size / 3); 1015 GlyphCanvas.redraw (); 1016 } 1017 1018 public class UndoItem : GLib.Object { 1019 public string first; 1020 public string next; 1021 public double kerning; 1022 public bool has_kerning; 1023 1024 public UndoItem (string first, string next, double kerning, bool has_kerning) { 1025 this.first = first; 1026 this.next = next; 1027 this.kerning = kerning; 1028 this.has_kerning = has_kerning; 1029 } 1030 } 1031 } 1032 1033 } 1034
__label__pos
0.953595
Documentation and Tutorials Features overview... Django image upload Overview Cloudinary is a cloud-based service that provides an end-to-end image management solution including uploads, storage, administration, image manipulation, and delivery. As part of its service, Cloudinary provides an API for uploading images and any other kind of files to the cloud. Images uploaded to Cloudinary are stored safely in the cloud with secure backups and revision history, utilizing Amazon's S3 service. After you upload your images to Cloudinary, you can browse them using an API or an interactive web interface and manipulate them to reach the size and look & feel that best matches your graphic design. All uploaded images and dynamically transformed images are optimized and delivered by Cloudinary through a fast CDN with advanced caching for optimal user experience. Cloudinary's APIs allow secure uploading images from your servers, directly from your visitors browsers or mobile applications, or fetched via remote public URLs. Comprehensive image transformations can be applied on uploaded images and you can extract images' metadata and semantic information once their upload completes. Cloudinary's Python library wraps Cloudinary's upload API and simplifies the integration. Python methods are available for easily uploading images and raw files to the cloud. Django view helper methods are available for uploading images directly from a browser to Cloudinary. This page covers common usage patterns for Django image upload with Cloudinary. For a full list of Django image upload options, refer to All Upload Options. Server side upload You can upload images (or any other raw file) to Cloudinary from your Python code or Django server. Uploading is done over HTTPS using a secure protocol based on your account's api_key and api_secret parameters. The following Python method uploads an image to the cloud: def upload(file, **options) For example, uploading a local image file named my_image.jpg: cloudinary.uploader.upload("my_picture.jpg") Uploading is performed synchronously. Once finished, the uploaded image is immediately available for manipulation and delivery. An upload API call returns a Hash with content similar to that shown in the following example: { u'bytes': 29802, u'created_at': u'2013-06-25T17:20:30Z', u'format': u'jpg', u'height': 282, u'public_id': u'hl22acprlomnycgiudor', u'resource_type': u'image', u'secure_url': u'https://res.cloudinary.com/demo/image/upload/v1372180830/hl22acprlomnycgiudor.jpg', u'signature': u'10594f028dbc23e920fd084f8482394798edbc68', u'type': u'upload', u'url': u'http://res.cloudinary.com/demo/image/upload/v1372180830/hl22acprlomnycgiudor.jpg', u'version': 1372180830, u'width': 292 } The response includes HTTP and HTTPS URLs for accessing the uploaded image as well as additional information regarding the uploaded file: The Public ID of the image (used for building viewing URLs), resource type, width and height, image format, file size in bytes, a signature for verifying the response and more. Public ID Each uploaded image is assigned with a unique identifier called Public ID. It is a URL-safe string that is used to reference the uploaded resource as well as building dynamic delivery and transformation URLs. By default, Cloudinary generates a unique, random Public ID for each uploaded image. This identifier is returned in the public_id response parameter. In the example above, the assigned Public ID is 'hl22acprlomnycgiudor'. As a result, the URL for accessing this image via our 'demo' account is the following: http://res.cloudinary.com/demo/image/upload/hl22acprlomnycgiudor.jpg You can specify your own custom public ID instead of using the randomly generated one. The following example specifies 'sample_id' as the public ID: cloudinary.uploader.upload("my_picture.jpg", public_id = 'sample_id') Using a custom public ID is useful when you want your delivery URLs to be readable and refer to the associated entity. For example, setting the public ID to a normalized user name and identifier in your local system: cloudinary.uploader.upload("my_picture.jpg", public_id = 'john_doe_1001') Public IDs can be organized in folders for more structured delivery URLs. To use folders, simply separate elements in your public ID string with slashes ('/'). Here's an example: cloudinary.uploader.upload("my_picture.jpg", public_id = 'my_folder/my_name') As the example below shows, your public IDs can include multiple folders: cloudinary.uploader.upload("my_picture.jpg", public_id = 'my_folder/my_sub_folder/my_name') Set use_filename as true to tell Cloudinary to use the original name of the uploaded image file as its Public ID. Notice that the file name will be normalized and a set of random characters will be appended to it to ensure uniqueness. This is quite useful if you want to safely reuse the filenames of files uploaded directly by your users. cloudinary.uploader.upload('sample.jpg', use_filename = True) # Generated public ID for example: 'sample_apvz1t' Data uploading options Cloudinary's Python library supports uploading files from various sources. You can upload an image by specifying a local path of an image file. For example: cloudinary.uploader.upload('/home/my_image.jpg') You can provide an IO object that you created: cloudinary.uploader.upload(open('/tmp/image1.jpg', 'rb')) If your images are already publicly available online, you can specify their remote HTTP URLs instead of uploading the actual data. In this case, Cloudinary will fetch the image from its remote URL for you. This option allows for a much faster migration of your existing images. Here's an example: cloudinary.uploader.upload('http://www.example.com/image.jpg') If you have existing images in an Amazon S3 bucket, you can point Cloudinary to their S3 URLs. Note - this option requires a quick manual setup. Contact us and we'll guide you on how to allow Cloudinary access to your relevant S3 buckets. cloudinary.uploader.upload('s3://my-bucket/my-path/my-file.jpg') In cases where images are uploaded by users of your Django application through a web form, you can pass the parameter of your Django's request.FILES to the upload method: cloudinary.uploader.upload(request.FILES['file']) Django forms and models You can integrate Cloudinary's image uploading with your Django models and forms using Cloudinary's helper classes. As shown in the example below, you can define a model class Photo in your models.py file. This class has an image field of the CloudinaryField class. from django.db import models from cloudinary.models import CloudinaryField class Photo(models.Model): image = CloudinaryField('image') In the forms.py file we define a PhotoForm class that has a form field named image of the CloudinaryFileField class (in default). from django.forms import ModelForm from .models import Photo class PhotoForm(ModelForm): class Meta: model = Photo The views.py file defines a view named upload which displays a HTML upload form and also handles posting of image files. Such images are uploaded to Cloudinary from your Django server by the CloudinaryFileField class. from django import forms from django.http import HttpResponse from cloudinary.forms import cl_init_js_callbacks from .models import Photo from .forms import PhotoDirectForm def upload(request): context = dict( backend_form = PhotoForm()) if request.method == 'POST': form = PhotoForm(request.POST, request.FILES) context['posted'] = form.instance if form.is_valid(): form.save() return render(request, 'upload.html', context) The following HTML templates includes a form that uploads images to your server for uploading to Cloudinary: {% load cloudinary %} {% load url from future %} {% block body %} <div id='backend_upload'> <form action="{% url "photo_album.views.upload" %}" method="post" enctype="multipart/form-data"> {% csrf_token %} {{ backend_form }} <input type="submit" value="Upload"> </form> {% endblock %} Having stored the image ID, you can now embed the image or a transformed version of it using the cloudinary template tag: {% load cloudinary %} {% cloudinary photo.image format="jpg" width=120 height=80 crop="fill" %} In addition, you can assign tags, apply transformation or specify any Cloudinary's upload options when initializing the CloudinaryFileField class. from django.forms import ModelForm from cloudinary.forms import CloudinaryFileField from .models import Photo class PhotoForm(ModelForm): class Meta: model = Photo image = CloudinaryFileField( attrs = { 'style': "margin-top: 30px" }, options = { 'tags': "directly_uploaded", 'crop': 'limit', 'width': 1000, 'height': 1000, 'eager' [{ 'crop': 'fill', 'width': 150, 'height': 100 }] }) Direct uploading from the browser The upload samples mentioned above allows your server-side Django code to upload images to Cloudinary. In this flow, if you have a web form that allows your users to upload images, the image data is first sent to your server and only then uploaded to Cloudinary. A more efficient and powerful option is to allow your users to upload images directly from the browser to Cloudinary instead of going through your servers. This method allows for faster uploading and better user experience. It also reduces load from your servers and reduces the complexity of your Django applications. Uploading directly from the browser is done using Cloudinary's jQuery plugin. To ensure that all uploads were authorized by your application, a secure signature must first be generated in your server-side Python code. Direct uploading environment setup Start by including the required Javascript files - jQuery, Cloudinary's plugin and the jQuery-File-Upload plugin it depends on. These are located in the cloudinary/static folder of the Django library. In your Django template, load Cloudinary and include jQuery and the required jQuery plugins: {% load cloudinary %} <script src="//ajax.googleapis.com/ajax/libs/jquery/1.8.3/jquery.min.js"></script> {% cloudinary_includes %} Cloudinary's jQuery plugin requires your cloud_name and additional configuration parameters to be available. Note: never expose your api_secret in public client side code. To automatically set-up Cloudinary's configuration, include the following line in your view or layout: {% cloudinary_js_config %} Direct uploading from the browser is performed using XHR (Ajax XMLHttpRequest‎) CORS (Cross Origin Resource Sharing) requests. In order to support older browsers that do not support CORS, the jQuery plugin will gracefully degrade to an iframe based solution. The file cloudinary_cors.html in automatically used to enable cross browser uploads when CORS is not supported. See cl_init_js_callbacks below. Direct upload file tag Cloudinary's direct uploading can be integrated with your Django model. Assuming you have the following Photo class defined in your models.py file. This class has an image field of the CloudinaryField class. from django.db import models from cloudinary.models import CloudinaryField class Photo(models.Model): image = CloudinaryField('image') In the forms.py file we define a PhotoDirectForm class that has a form field named image of the CloudinaryJsFileField class. This class does all the behind-the-scenes magic for you. from django.forms import ModelForm from cloudinary.forms import CloudinaryJsFileField from .models import Photo class PhotoDirectForm(ModelForm): class Meta: model = Photo image = CloudinaryJsFileField() The views.py file defines a view named upload_prompt which initializes the direct form and defines the required callback URL: from django import forms from django.http import HttpResponse from cloudinary.forms import cl_init_js_callbacks from .models import Photo from .forms import PhotoDirectForm def upload_prompt(request): context = dict(direct_form = PhotoDirectForm()) cl_init_js_callbacks(context['direct_form'], request) return render(request, 'upload_prompt.html', context) Embed a file input tag in your HTML templates using the cloudinary_direct_upload_field template tag or the direct_form you defined. The following example adds a file input field to your form. Selecting or dragging a file to this input field will automatically initiate uploading from the browser to Cloudinary. <form action="{% url "photo_album.views.direct_upload_complete" %}" method="post"> {% csrf_token %} {{ direct_form }} {# altenatively, use: {% cloudinary_direct_upload_field "image", request=request %} #} <input type="submit" value="Submit" /> </form> When uploading is completed, the identifier of the uploaded image is set as the value of a hidden input field of your selected name. You can then process the identifier received by your Django code and store it in your model for future use, exactly as if you're using a standard server side uploading. The following Django code in views.py processes the received identifier, verifies the signature (concatenated to the identifier) and updates a model entity with the identifiers of the uploaded image (i.e., the Public ID and version of the image). import json from django import forms from django.http import HttpResponse from django.views.decorators.csrf import csrf_exempt from .models import Photo from .forms import PhotoDirectForm ... @csrf_exempt def direct_upload_complete(request): form = PhotoDirectForm(request.POST) if form.is_valid(): form.save() ret = dict(photo_id = form.instance.id) else: ret = dict(errors = form.errors) return HttpResponse(json.dumps(ret), content_type='application/json') Having stored the image ID, you can now display a directly uploaded image in the same way you would display any other Cloudinary hosted image: {% load cloudinary %} {% cloudinary photo.image format="jpg" width=120 height=80 crop="fill" %} Additional direct uploading options When uploading directly from the browser, you can still specify all the upload options available to server-side uploading. For example, the following call performs direct uploading that will also tag the uploaded image, limit its size to given dimensions and generate a thumbnail eagerly. Also notice the custom HTML attributes. from django.forms import ModelForm from cloudinary.forms import CloudinaryJsFileField from .models import Photo class PhotoDirectForm(ModelForm): class Meta: model = Photo image = CloudinaryJsFileField( attrs = { 'style': "margin-top: 30px" }, options = { 'tags': "directly_uploaded", 'crop': 'limit', 'width': 1000, 'height': 1000, 'eager' [{ 'crop': 'fill', 'width': 150, 'height': 100 }] }) Preview thumbnail, progress indication, multiple images Cloudinary's jQuery library also enables an enhanced uploading experience - show a progress bar, display a thumbnail of the uploaded image, drag & drop support, upload multiple files and more. Bind to Cloudinary's cloudinarydone event if you want to be notified when an upload to Cloudinary has completed. You will have access to the full details of the uploaded image and you can display a cloud-generated thumbnail of the uploaded images using Cloudinary's jQuery plugin. The following sample code creates a 150x100 thumbnail of an uploaded image and updates an input field with the public ID of this image. $('.cloudinary-fileupload').bind('cloudinarydone', function(e, data) { $('.preview').html( $.cloudinary.image(data.result.public_id, { format: data.result.format, version: data.result.version, crop: 'fill', width: 150, height: 100 }) ); $('.image_public_id').val(data.result.public_id); return true; }); You can track the upload progress by binding to the following events: fileuploadsend, fileuploadprogress, fileuploaddone and fileuploadfail. You can find more details and options in the documention of jQuery-File-Upload. The following Javascript code updates a progress bar according to the data of the fileuploadprogress event: $('.cloudinary-fileupload').bind('fileuploadprogress', function(e, data) { $('.progress_bar').css('width', Math.round((data.loaded * 100.0) / data.total) + '%'); }); You can find some more examples as well as an upload button style customization in our Photo Album sample project. The file input field can be configured to support simultaneous multiple file uploading. Setting the multiple HTML parameter to true allows uploading multiple files. Note - currently, only a single input field is updated with the identifier of the uploaded image. You should manually bind to the cloudinarydone event to handle results of multiple uploads. Here's an example: image = CloudinaryJsFileField(attrs = { 'multiple': 1 }) For a fully working example, check out our Django sample project. For more details about direct uploading, see this blog post: Direct image uploads from the browser to the cloud with jQuery. Incoming transformations By default, images uploaded to Cloudinary are stored in the cloud as-is. Once safely stored in the Cloud, you can generate derived images from these originals by asking Cloudinary to apply transformations and manipulations. Sometimes you may want to normalize and transform the original images before storing them in the cloud. You can do that by applying an incoming transformation as part of the upload request. Any image transformation parameter can be specified as options passed to the upload call. A common incoming transformation use-case is shown in the following example, namely, limit the dimensions of user uploaded images to 1000x1000: cloudinary.uploader.upload('/home/my_image.jpg', width = 1000, height = 1000, crop = 'limit') Another example, this time performing custom coordinates cropping and forcing format conversion to PNG: cloudinary.uploader.upload('/home/my_image.jpg', width = 400, height = 300, x = 50, y = 80, crop = 'crop', format = 'png') Named transformations as well as multiple chained transformations can be applied using the transformation parameter. The following example first limits the dimensions of an uploaded image and then adds a watermark as an overlay. cloudinary.uploader.upload('/home/my_image.jpg', transformation = [ { 'width': 1000, 'height': 1000, 'crop': 'limit' }, { 'overlay': "my_watermark", 'flags': 'relative', 'width': 0.5 } ]) Eager transformations Cloudinary can dynamically transform images using specially crafted transformation URLs. By default, the first time a user accesses a transformation URL, the transformed images is created on-the-fly, stored persistently in the cloud and delivered through a fast CDN with advanced caching. All subsequent accesses to the same transformation would quickly deliver a cached copy of the previously generated image via the CDN. You can tell Cloudinary to eagerly generate one or more derived images while uploading. This approach is useful when you know in advance that certain transformed versions of your uploaded images will be required. Eagerly generated derived images are available for fast access immediately when the upload call returns. Generating transformed images eagerly is done by specifying the `eager` upload option. This option accepts either a hash of transformation parameters or an array of transformations. The following example eagerly generates a 150x100 face detection based thumbnail: cloudinary.uploader.upload('/home/my_image.jpg', public_id = "eager_sample", eager = { 'width': 150, 'height': 100, 'crop': 'thumb', 'gravity': 'face' }) Now, if you embed the same transformed image in your web view, it will already be available for fast delivery. The following tag embeds the derived image from the example above: cloudinary.CloudinaryImage("eager_sample.jpg").image(width = 150, height = 100, crop = 'thumb', grvity = 'face') The following example generates two transformed versions eagerly, one fitting image into a 100x150 PNG and another applies a named transformation. cloudinary.uploader.upload('/home/my_image.jpg', eager = [ {'width': 100, 'height': 150, 'crop': 'fit', 'format': 'png'}, {'transformation': 'jpg_with_quality_30'} ]) You can apply an incoming transformation and generate derived images eagerly using the same upload call. The following example does that while also applying a chained transformation: cloudinary.uploader.upload("sample.jpg", transformation = [ {'width': 100, 'height': 120, 'crop': 'limit'}, {'crop': 'crop', 'x': 5, 'y': 10, 'width': 40, 'height': 10} ], eager = [ {'width': 0.2, 'crop': 'scale'}, {'effect': 'hue:30'} ]) For more image transformation options in Django, see Django image manipulation. Semantic data extraction When you upload a resource to Cloudinary, the API call will report information about the uploaded asset: width, height, number of bytes and image format. Cloudinary supports extracting additional information from the uploaded image: Exif and IPTC camera metadata, color histogram, predominant colors and coordinates of automatically detected faces. You can ask Cloudinary for this semantic data either during the upload API call for newly uploaded images, or using our Admin API for previously uploaded images. See Cloudinary's Admin API documentation for more details: Details of a single resource. You can tell Cloudinary to include relevant metadata in its upload API response by setting the faces, exif, colors and image_metadata boolean parameters to true while calling the upload API. The following example uploads an image via a remote URL, and requests the coordinates of all detected faces in the image: cloudinary.uploader.upload("http://res.cloudinary.com/demo/image/upload/couple.jpg", faces = True) Below you can see the faces coordinates that were included in the upload response: { u'public_id': u'b2sajypsvfcdnxk3licr', ... u'faces': [[98, 74, 61, 83], [139, 130, 52, 71]] } Another example, this time requesting details about the main colors of the uploaded image. Each pair in the returned colors array includes the color name or its RGB representation and the percentage of the image comprising it. cloudinary.uploader.upload("http://res.cloudinary.com/demo/image/upload/couple.jpg", colors = True) # Output: { u'public_id': u'ipnmreyakj0n0tdggfq4', ... u'colors': [ [u'#152E02', 7.9], [u'#2E4F06', 6.3], [u'#3A6604', 5.6], [u'#EEF2F3', 5.2], [u'#598504', 4.6], [u'#0D1903', 4.6], [u'#DCDFDB', 4.4], [u'#7AA403', 3.9], [u'#DBE98A', 3.8], [u'#4C7208', 3.5], [u'#8DAC30', 3.3], [u'#6C9406', 3.2], [u'#6E912F', 3.2], [u'#B0C95F', 3.1], [u'#89AF07', 3.1], [u'#9E744E', 2.9], [u'#CD9A6F', 2.8], [u'#D8B395', 2.5], [u'#719357', 2.4], [u'#ACCDBC', 2.3], [u'#8E6437', 2.3], [u'#E2E0A0', 2.3], [u'#2C4B16', 2.3], [u'#656950', 2.1], [u'#25370A', 2.0], [u'#73A092', 1.9], [u'#4E3721', 1.6], [u'#A0AD9A', 1.6], [u'#BBD258', 1.5], [u'#5B602B', 1.3], [u'#302B1D', 1.3], [u'#9CA25C', 1.2]], u'predominant': { u'google': [ [u'yellow', 40.1], [u'green', 24.6], [u'brown', 13.4], [u'black', 12.5], [u'teal', 9.4]]} } You can also request Exif, IPTC, colors and faces data in a single upload call: cloudinary.uploader.upload("/home/my_image.jpg", faces = True, exif = True, colors = True, image_metadata = True) See the following blog post for more details: API for extracting semantic image data - colors, faces, Exif data and more Raw file uploading Cloudinary's main strength is in managing images. However, you can still use Cloudinary to manage any other file format using the same simple APIs. You'll get your files stored in a highly available storage with automatic backups and revision control, and when accessed, have them delivered through a fast CDN. You can upload a raw file to Cloudinary by setting the resource_type parameter to raw. cloudinary.uploader.upload("sample_spreadsheet.xls", public_id = "sample_spreadsheet", resource_type = 'raw') Non-image raw files are stored in the cloud as-is. Note that while public IDs of image files do not include the file's extension, public IDs of raw files does include the original file's extension. Here's a sample response of a raw upload call, which is slightly different from an image upload response: { u'bytes': 6144, u'created_at': u'2013-06-23T14:55:21Z', u'public_id': u'sample_spreadsheet.xls', u'resource_type': u'raw', u'secure_url': u'https://res.cloudinary.com/demo/raw/upload/v1372186255/sample_spreadsheet.xls', u'signature': u'2985c846b8ea5b0a280910a5f3b9d1cfec685f25', u'type': u'upload', u'url': u'http://res.cloudinary.com/demo/raw/upload/v1372186255/sample_spreadsheet.xls', u'version': 1372186255 } Sometimes you don't know whether your users would upload image files or raw files. In order to support that, you can set the resource_type parameter to auto. Cloudinary will automatically detect whether the uploaded file is an image or a non-image raw file. When using direct image uploading from the browser, resource type is set to auto by default. cloudinary.uploader.upload("sample_spreadsheet.xls", resource_type = 'auto') Delivery URLs of raw files are built quite similarly to those of images. Just make sure to set resource_type to raw. Here's an example: cloudinary.utils.cloudinary_url("sample_spreadsheet.xls", resource_type = 'raw') Update and delete images Image uploaded to Cloudinary are stored persistently in the cloud. You can programatically delete an uploaded image using the following method: def destroy(public_id, **options) For example, the following code would delete the uploaded image assigned with the public ID 'zombie': cloudinary.uploader.destroy('zombie') See our Admin API documentation for more options of listing and deleting images and files. When you delete an uploaded image, all its derived images are deleted as well. However, images and derived images that were already delivered to your users might have cached copies at the CDN layer. If you now upload a new image with the same public ID, the cached copy might be returned instead. To avoid that, use different randomly generated public IDs for each upload or alternatively, add the version component to the delivery URLs. If none of these solutions work, you might want to force cache invalidation for each deleted image. Forcing cache invalidation is done by setting the invalidate parameter to true either when deleting an image or uploading a new one. Note that it usually takes up to one hour for the CDN invalidation to take effect. Here's are two examples: cloudinary.uploader.destroy('zombie', invalidate = True) cloudinary.uploader.upload('new_zombie.jpg', public_id = 'zombie', invalidate = True) Refresh images Cloudinary supports forcing the refresh of Facebook & Twitter profile pictures. You can use the explicit API method for that. The response of this method includes the image's version. Use this version to bypass previously cached CDN copies. cloudinary.uploader.explicit("zuck", type = "facebook") You can also use the explicit API call to generate transformed versions of an uploaded image. This is useful when Strict Transformations are allowed for your account and you wish to create custom derived images for already uploaded images. cloudinary.uploader.explicit("sample_id", type = 'upload', eager = { 'width': 150, 'height': 230, 'crop': 'fill' } ) Rename images You can rename images uploaded to Cloudinary. Renaming means changing the public ID of already uploaded images. The following method allows renaming a public ID: def rename(from_public_id, to_public_id, **options) For example, renaming an image with the public ID 'old_name' ro 'new_name': cloudinary.uploader.rename('old_name', 'new_name') By default, Cloudinary prevents renaming to an already taken public ID. You can set the overwrite option to true to delete the image that has the target public ID and replace it with the image being renamed: cloudinary.uploader.rename('old_name', 'new_name', overwrite = true) Manage tags Cloudinary supports assigning one or more tags to uploaded images. Tags allow you to better organize your media library. You can use our Admin API and media library web interface for searching, listing and deleting images by tags. In addition, you can merge multiple images that share the same tag into a sprite, a multi-page PDF or an animated GIF. See our Admin API documentation for more details regarding managing images by tags. The following example assigned two tags while uploading. cloudinary.uploader.upload('/home/my_image.jpg', public_id = "sample_id", tags = ['special', 'for_homepage']) You can modify the assigned tags of an already uploaded image. The following example assigns a tag to a list of images: cloudinary.uploader.add_tag('another_tag', ['sample_id', 'de9wjix4hhnqpxixq6cw']) The following example clears the given tag from a list of images: cloudinary.uploader.remove_tag('another_tag', ['sample_id', 'de9wjix4hhnqpxixq6cw']) This example clears all tags from a given list of images: cloudinary.uploader.replace_tag('another_tag', ['sample_id']) Text creation Cloudinary allows generating dynamic text overlays with your custom text. First you need to create a text style - font family, size, color, etc. The name of the text style is its public ID and it behaves like an image of the text type. The following command creates a text style named dark_name of a certain font, color and style: cloudinary.uploader.text("Sample Name", public_id = 'dark_name', font_family = 'Arial', font_size = 12, font_color = 'black', opacity = 90) The following image tag adds a text overlay using the created dark_name text style. cloudinary.CloudinaryImage("sample.jpg").image( overlay = "text:bold_dark:Hello+World", gravity = "south_east", x = 5, y = 5) For more text style options, see Text layers creation. More information about text overlays is available in our adding text overlays in Django documentation. Notifications and async transformations By default, Cloudinary's upload API works synchronously. Uploaded images processed and eager transformations are generated synchronously during the upload API call. You can tell Cloudinary to generate eager transformations in the background and send you an HTTP notification callback when the transformations are ready. You can do that by setting the eager_async parameter to true and optionally setting eager_notification_url to the URL Cloudinary should send the callback to. Here's an example: cloudinary.uploader.upload('/home/my_image.jpg', eager = { 'width': 150, 'height': 100, 'crop': 'thumb', 'gravity':'face' }, eager_async = True, eager_notification_url = "http://mysite/my_notification_endpoint") Cloudinary also supports webhooks. With webhooks enabled, you can get a notification to your server when an upload is completed. This is useful when you use direct image uploading from the browser and you want to be informed whenever an image is uploaded by your users. Setting the notification_url parameter tells Cloudinary to perform an asynchronous HTTP GET request to your server when the upload is complete. For example: cloudinary.uploader.upload('/home/my_image.jpg', notification_url = "http://mysite/my_notification_endpoint") See the following blog post for more details: Webhooks, upload notifications and background image processing. All upload options cloudinary.uploader.upload(file, options = {}) Cloudinary's upload API call accepts the following options: • file - The resource to upload. Can be one of the following: • A local path (e.g., '/home/my_image.jpg'). • An HTTP URL of a resource available on the Internet (e.g., 'http://www.example.com/image.jpg'). • A URL of a file in a private S3 bucket white-listed for your account (e.g., 's3://my-bucket/my-path/my-file.jpg') • An IO input stream of the data (e.g., open(file, 'rb')). • public_id (Optional) - Public ID to assign to the uploaded image. Random ID is generated otherwise and returned as a result for this call. The public ID may contain a full path including folders separated by '/'. • tags (Optional) - A tag name or an array with a list of tag names to assign to the uploaded image. • context - A map of key-value pairs of general textual context metadata to attach to an uploaded resource. The context values of uploaded files are available for fetching using the Admin API. For example: { "alt" => "My image", "caption" => "Profile Photo" }. • format (Optional) - A format to convert the uploaded image to before saving in the cloud. For example: 'jpg'. • allowed_formats - A format name of an array of file formats that are allowed for uploading. The default is any supported image kind and any type of raw file. Files of other types will be rejected. The formats can be image types or raw file extensions. For example: ['jpg', 'gif' , 'doc']. • Transformation parameters (Optional) - Any combination of transformation-related parameters for transforming the uploaded image before storing in the cloud. For example: 'width', 'height', 'crop', 'gravity', 'quality', 'transformation'. • eager (Optional) - A list of transformations to generate for the uploaded image during the upload process, instead of lazily creating these on-the-fly on access. • eager_async (Optional, Boolean) - Whether to generate the eager transformations asynchronously in the background after the upload request is completed rather than online as part of the upload call. Default: false. • resource_type (Optional) - Valid values: 'image', 'raw' and 'auto'. Default: 'image'. • type (Optional) - Allows uploading images as 'private' or 'authenticated'. Valid values: 'upload', 'private' and 'authenticated'. Default: 'upload'. • headers (Optional) - An HTTP header or an array of headers for returning as response HTTP headers when delivering the uploaded image to your users. Supported headers: 'Link', 'X-Robots-Tag'. For example 'X-Robots-Tag: noindex'. • callback (Optional) - An HTTP URL to redirect to instead of returning the upload response. Signed upload result parameters are added to the callback URL. Ignored if it is an XHR upload request (Ajax XMLHttpRequest). • notification_url (Optional) - An HTTP URL to send notification to (a webhook) when the upload is completed. • eager_notification_url (Optional) - An HTTP URL to send notification to (a webhook) when the generation of eager transformations is completed. • backup (Optional, Boolean) - Tell Cloudinary whether to back up the uploaded image. Overrides the default backup settings of your account. • return_delete_token (Boolean) - Whether to return a deletion token in the upload response. The token can be used to delete the uploaded image within 10 minutes using an unauthenticated API request. • faces (Optional, Boolean) - Whether to retrieve a list of coordinates of automatically detected faces in the uploaded photo. Default: false. • exif (Optional, Boolean) - Whether to retrieve the Exif metadata of the uploaded photo. Default: false. • colors (Optional, Boolean) - Whether to retrieve predominant colors & color histogram of the uploaded image. Default: false. • image_metadata (Optional, Boolean) - Whether to retrieve IPTC and detailed Exif metadata of the uploaded photo. Default: false. • phash (Optional, Boolean) - Whether to return the perceptual hash (pHash) on the uploaded image. The pHash acts as a fingerprint that allows checking image similarity. Default: false. • invalidate (Optional, Boolean) - Whether to invalidate CDN cache copies of a previously uploaded image that shares the same public ID. Default: false. • use_filename (Optional, Boolean) - Whether to use the original file name of the uploaded image if available for the public ID. The file name is normalized and random characters are appended to ensure uniqueness. Default: false. • unique_filename (Optional, Boolean) - Only relevant if use_filename is true. When set to false, should not add random characters at the end of the filename that guarantee its uniqueness. Default: true. • folder - An optional folder name where the uploaded resource will be stored. The public ID contains the full path of the uploaded resource, including the folder name. • overwrite (Optional, Boolean) - Whether to overwrite existing resources with the same public ID. When set to false, return immediately if a resource with the same public ID was found. Default: true. • discard_original_filename (Optional, Boolean) - Whether to discard the name of the original uploaded file. Relevant when delivering images as attachments (setting the 'flags' transformation parameter to 'attachment'). Default: false. • face_coordinates (Optional, Array) - List of coordinates of faces contained in an uploaded image. The given coordinates are used for cropping uploaded images using the face or faces gravity mode. The specified coordinates override the automatically detected faces. Each face is specified by the X & Y coordinates of the top left corner and the width & height of the face. For example: [[10,20,150,130], [213,345,82,61]]. • custom_coordinates (Optional, Array) - Coordinates of an interesting region contained in an uploaded image. The given coordinates are used for cropping uploaded images using the custom gravity mode. The region is specified by the X & Y coordinates of the top left corner and the width & height of the region. For example: [85,120,220,310]. • raw_convert - Set to 'aspose' to automatically convert Office documents to PDF files and other image formats using the Aspose Document Conversion add-on. • categorization - Set to 'rekognition_scene' to automatically detect scene categories of photos using the ReKognition Scene Categorization add-on. • auto_tagging (0.0 to 1.0 Decimal number) - Whether to assign tags to an image according to detected scene categories with confidence score higher than the given value. • detection - Set to 'rekognition_face' to automatically extract advanced face attributes of photos using the ReKognition Detect Face Attributes add-on. • moderation - Set to 'manual' to add the uploaded image to a queue of pending moderation images. Set to 'webpurify' to automatically moderate the uploaded image using the WebPurify Image Moderation add-on. • upload_preset - Name of an upload preset that you defined for your Cloudinary account. An upload preset consists of upload parameters centrally managed using the Admin API or from the settings page of the management console. An upload preset may be marked as 'unsigned', which allows unsigned uploading directly from the browser and restrict the directly allowed parameters to: public_id, folder, tags, context, face_coordinates and custom_coordinates.
__label__pos
0.702699
131 years old now in 2022 - what year born? Today, 5 December 2022 if a person is 131 years old now, what year was he born? See below: If 131 years old, year of birth is: 1890 or 1891* *It depends on whether he was born before or after 1 January 1891. If now 131 years and his year of birth 1890 if he was born after 5 December 1890 until 31 december 1890 inclusively, and his year of birth is 1891 if he was born from 1 January 1891 to 5 December 1891. Other examples: If now 12 years, year of birth 2009 or 2010 If now 16 years, year of birth 2005 or 2006 If now 28 years, year of birth 1993 or 1994 If now 85 years, year of birth 1936 or 1937 HTML code: BB code: More:
__label__pos
0.981338
Tell me more × Mathematics Stack Exchange is a question and answer site for people studying math at any level and professionals in related fields. It's 100% free, no registration required. I am studying the linear Independence and linear dependence of the vectors in my Linear algebra course and i am confused about the following theorem (The book is Linear Algebra and its applications by David C Lay) and i need some explanation of this theorem. enter image description here In this theorem i am confused about "If S is linearly dependent and V1 not equal to zero then some Vj (with j>1) is a linear combination of the preceeding vectors" share|improve this question 3 Answers What exactly is your problem; general confusion is not well-defined. It means that one of the vectors is a linear combination of all of those before, in other words, you can usually pick the last vector if the set is finite. If it is infinite, this theorem still holds true (as there cannot be more then n linear independant vectors in a n-dimensional vector space). An example: let our set be $S=\{(1,2),(3,4),(4,6),(1,3)\}$ now, both $(4,6)$ and $(1,3)$ are linearly dependant to their predecessors, which ar, in turn, $\{(1,2),(3,4)\}$ and $\{(1,2),(3,4),(4,6)\}$. In fact, $(1,3)$ is already linear dependant to the first 2 vectors, the third is just "overkill" share|improve this answer Recall that a vector $w$ is a linear combination of $w_1,\ldots,w_k$ if there exist scalars $\lambda_1,\ldots,\lambda_k$ such that $w = \lambda_1w_1 + \cdots + \lambda_kw_k.$ A set $\{w_1,\ldots,w_k\}$ is linearly dependent if there exists a non-trivial solution to $\lambda_1w_1+\cdots+\lambda_kw_k = 0$, i.e. not all $\lambda_i = 0.$ Assume that $v_1 \neq 0$ and that $\{v_1,\ldots,v_n\}$ is linearly dependent. Since $v_1 \neq 0$ it follows that $\{v_1\}$ is linearly independent. Next, consider $\{v_1,v_2\}$. Either $\{v_1,v_2\}$ is linearly independent (LI) or linearly dependent (LD). If $\{v_1,v_2\}$ is LD then you have your $v_j$, namely $j=2$, because $\{v_1\}$ was LI while $\{v_1,v_2\}$ is LD meaning that $v_2$ is a linear combination of $v_1$. If $\{v_1,v_2\}$ is LI then consider $\{v_1,v_2,v_3\}$. Either $\{v_1,v_2,v_3\}$ is LI or LD. If $\{v_1,v_2,v_3\}$ is LD then you have your $v_j$, namely $j = 3$, because $\{v_1,v_2\}$ was LI while $\{v_1,v_2,v_3\}$ is LD meaning that $v_3$ is a linear combination of $v_1$ and $v_2$. If $\{v_1,v_2,v_3\}$ is LI then consider $\{v_1,v_2,v_3,v_4\}$, etc. Continue this process. You will eventually find a $v_j$ for which $\{v_1,\ldots,v_{j-1}\}$ is LI and $\{v_1,\ldots,v_j\}$ is LD meaning that $v_j$ is a linear combination of $v_1,\ldots,v_{j-1}.$ If you don't find such a $v_j$ then you have a contradiction: you would have that $\{v_1,\ldots,v_n\}$ was LI, but you assumed it was LD! share|improve this answer If $S=\left\{v_1,v_2,\ldots,v_\rho\right\}$ is linearly depended then \begin{equation}\lambda_1v_1+\lambda_2v_2+\cdots+\lambda_\rho v_\rho=0\end{equation} for some $\lambda_i$ not all zero. Let $j\in \{1,2,\ldots\rho\}$ be the least integer s.t. $\lambda_j=\lambda_{j+1}=\cdots=\lambda_\rho=0$. If we suppose that $v_1\neq0$ then $j>1$. Therefore \begin{equation}\lambda_1v_1+\cdots+\lambda_{j-1}v_{j-1}=0\end{equation} and $\lambda_{j-1}\neq0$. We conclude that \begin{equation}v_{j-1}=-\dfrac{\lambda_1}{\lambda_{j-1}}v_1-\cdots-\dfrac{\lambda_{j-2}}{\lambda_{j-1}}v_{j-2}.\end{equation} share|improve this answer Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.988018
2 $\begingroup$ I am trying to prove Theorem 6.2 on page 127 of the book Real-Time Systems by Jane W. S. Liu: http://www.cse.hcmut.edu.vn/~thai/books/2000%20_%20Liu-%20Real%20Time%20Systems.pdf It is based on Early Deadline First(EDF) scheduling. It says the proof is similar to the proof for Theorem 6.1 on page 124-126. However, I am still stuck. Here is what I have so far: Density of task $T_j:\ \delta_j=\frac{e_j}{\min{\left(D_j,p_j\right)}}$ Density of the system: $∆=\sum_{j=1}^{n}\delta_j=\sum_{j=1}^{n}\frac{e_j}{\min{\left(D_j,p_j\right)}}$ $e_j$ is the execution time for task $T_j$ $D_j$ is the deadline for task $T_j$ $p_j$ is the period for task $T_j$ $\emptyset_j$ is the phase for task $T_j$ $r_{j,b}$ is the release time for task $T_j$ at period $b$ $J_{j,b}$ is the job for task $T_j$ at period $b$ $\emptyset_j=r_{j,1}$ $r_{j,b}+p_j=r_{j,b+1}$ THEOREM 6.2. A system $T$ of independent, preemptable tasks can be feasibly scheduled on one processor if its density is equal to or less than $1$. If $D_j<p_j$ for some $j$, then $∆=\sum_{j=1}^{n}\frac{e_j}{\min{\left(D_j,p_j\right)}}\le1$ is only a sufficient condition so we can only say that the system may not be schedulable when the condition is not satisfied. I try to prove the contrapositive just as Theorem 6.1 on page 124-126. So what I try to prove is that if according to an EDF schedule, the system fails to meet some deadlines, then its density is larger than $1$. Suppose that the system begins to execute at time $0$. And at time $t$, the job $J_{i,c}$ of task $T_i$ misses its deadline. Assume the case that the current period of every task begins at or after $r_{i,c}$, the release time of the job that misses its deadline. We use $t$ to divide two types of task besides task $T_i$ as in below figure: 1) tasks with deadline happens before $t$ in the current period just like $T_f$, 2) tasks with deadline happens after $t$ in the current period just like $T_k$. enter image description here $J_{i,c}$ misses its deadline at $t$ tells us that any current job whose deadline is after $t$ is not given any processor time to execute before $t$ and that the total processor time required to complete $J_{i,c}$ and all the jobs with deadlines at or before $t$ exceeds the total available time $t$. So we have $t<\left\lceil\frac{\left(t-\emptyset_i\right)}{p_i}\right\rceil e_i+\sum_{k\neq i,k\neq f}{\left\lfloor\frac{\left(t-\emptyset_k\right)}{p_k}\right\rfloor e}_k+\sum_{f\neq i,f\neq k}{\left\lceil\frac{\left(t-\emptyset_f\right)}{p_f}\right\rceil e}_f$ then $t<\left\lfloor\frac{\left(t-\emptyset_i\right)}{p_i}\right\rfloor e_i+e_i+\sum_{k\neq i,k\neq f}{\left\lfloor\frac{\left(t-\emptyset_k\right)}{p_k}\right\rfloor e}_k+\sum_{f\neq i,f\neq k}{{\left\lfloor\frac{\left(t-\emptyset_f\right)}{p_f}\right\rfloor e}_f+\sum_{f\neq i,f\neq k}\ e_f}$ And it is $=\left\lfloor\frac{\left(t-\emptyset_i\right)}{p_i}\right\rfloor e_i+\sum_{k\neq i,k\neq f}{\left\lfloor\frac{\left(t-\emptyset_k\right)}{p_k}\right\rfloor e}_k+\sum_{f\neq i,f\neq k}{{\left\lfloor\frac{\left(t-\emptyset_f\right)}{p_f}\right\rfloor e}_f+\sum_{f\neq k}\ e_f}$ $\le \left(\frac{t}{p_i}\right)e_i+\sum_{k\neq i,k\neq f}{\left(\frac{t}{p_k}\right)e}_k+\sum_{f\neq i,f\neq k}{{\left(\frac{t}{p_f}\right)e}_f+\sum_{f\neq k}\ e_f}$ $=t\sum_{h=1}^{n}\frac{e_h}{p_h}+\sum_{f\neq k}\ e_f$ $\le t\sum_{h=1}^{n}\frac{e_h}{\min{\left(D_h,p_h\right)}}+\sum_{f\neq k}\ e_f$ $=t∆+\sum_{f\neq k}\ e_f$ So I prove up to $t<t∆+\sum_{f\neq k}\ e_f$ but I cannot prove $∆>1$ $\endgroup$ 0 0 $\begingroup$ I might be too focus on the algebra. I believe I only need to argue or proof a scenario that would give the least density of the system that will fail to meet a deadline. As shown in the figure below, all tasks start at the same time with same period and deadline. For each of period, each task only have time slot D=D1=D2=...=Dn to execute. In other words, they share exactly the same time slot to execute. If density > 1 then obviously schedule will fail. If the below system has a density of $1^+$(greater than 1 by the smallest amount), then any changes on Di or Pi will only increase the density of the system or make the system feasible. enter image description here $\endgroup$ Your Answer By clicking “Post Your Answer”, you agree to our terms of service, privacy policy and cookie policy Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.854552
Configuring Static And Dynamic Ip Addresses On Linux Configuring Static IP Addresses on Linux To configure a static IP address on Linux, edit the network configuration file, typically /etc/network/interfaces. For example, to set a static IP address of 192.168.1.100 with a netmask of 255.255.255.0 and a gateway of 192.168.1.1, add the following lines: auto eth0 iface eth0 inet static address 192.168.1.100 netmask 255.255.255.0 gateway 192.168.1.1 Restart the networking service for the changes to take effect: sudo systemctl restart networking Configuring Dynamic IP Addresses on Linux To configure a dynamic IP address on Linux, remove any static IP settings from the network configuration file and restart the networking service: sudo nano /etc/network/interfaces Comment out the static IP settings (if present): #auto eth0 #iface eth0 inet static #address 192.168.1.100 #netmask 255.255.255.0 #gateway 192.168.1.1 Restart the networking service: sudo systemctl restart networking The system will now obtain an IP address dynamically via DHCP. To verify the assigned IP address, use the ip addr command: ip addr show eth0 ```## Configuring Static And Dynamic IP Addresses On Linux ### Executive Summary Properly managing IP addresses is a vital aspect of network administration, ensuring the seamless connectivity and functionality of network devices. This comprehensive guide delves into the intricacies of configuring static and dynamic IP addresses on Linux systems, empowering network administrators with the knowledge and expertise to manage their network infrastructure effectively. By understanding the concepts, implementation techniques, and troubleshooting strategies covered in this guide, readers will gain the skills necessary to establish reliable and efficient network configurations. ### Introduction In the realm of networking, IP addresses serve as the unique identifiers for devices connected to a network. Assigning IP addresses to network devices is a fundamental task in network administration, and understanding the difference between static and dynamic IP addressing is crucial. This guide will explore the intricacies of configuring both static and dynamic IP addresses on Linux systems, providing a comprehensive overview of the steps and considerations involved. ### Static IP Addresses Static IP addresses are manually assigned to network devices and remain constant over time. This type of IP addressing is typically preferred for devices that require a persistent and predictable IP address, such as servers and network appliances. * **Advantages:** * Provides a fixed IP address for devices, simplifying network management and troubleshooting. * Enhances security by preventing unauthorized devices from accessing the network. * Ensures consistent communication between devices, reducing network disruptions. ### Configuring Static IP Addresses 1. **Determine the network interface:** Identify the network interface you wish to configure. This can be done using the `ifconfig` command. 2. **Edit the network interface configuration file:** Locate the network interface configuration file, typically named `/etc/network/interfaces` or `/etc/sysconfig/network-scripts/ifcfg-eth0`. 3. **Add the static IP address:** Add the following lines to the configuration file: address <IP Address> netmask <Subnet Mask> gateway <Default Gateway> ``` 1. Restart the network service: Once the changes have been made, restart the network service to apply the new configuration. DHCP (Dynamic Host Configuration Protocol) DHCP is a network protocol that automatically assigns IP addresses to devices on a network. This simplifies the process of IP address management and ensures that devices can dynamically obtain IP addresses without manual configuration. • Advantages: • Automates IP address assignment, reducing administrative overhead. • Supports a large number of devices, making it suitable for large networks. • Ensures that devices always have valid IP addresses, minimizing network disruptions. Configuring DHCP 1. Install the DHCP server: Install a DHCP server software package on the Linux system. 2. Configure the DHCP server: Edit the DHCP server configuration file, typically located at /etc/dhcp/dhcpd.conf. 3. Define the DHCP range: Specify the range of IP addresses that will be assigned by the DHCP server. 4. Start the DHCP server: Start the DHCP server to begin assigning IP addresses to devices. Advantages Of Static Vs. Dynamic IP Addresses The choice between static and dynamic IP addresses depends on the specific requirements of the network. • Static IP Addresses: • Suitable for devices that require a fixed IP address, such as servers and network appliances. • Enhances security and simplifies network management. • Dynamic IP Addresses: • Simplifies IP address management and reduces administrative overhead. • Ideal for large networks with numerous devices. Troubleshooting IP Address Issues Common IP address problems and their troubleshooting steps include: • IP address conflict: Occurs when two devices on the network have the same IP address. Check the IP addresses assigned to devices and resolve any conflicts. • Incorrect IP address: Verify that the IP address configured on the device matches the network’s IP addressing scheme. • Gateway issues: Ensure that the default gateway is correctly configured and accessible. Conclusion Effectively configuring IP addresses on Linux systems is a critical aspect of network administration, ensuring reliable and efficient network connectivity. By understanding the concepts, implementation techniques, and troubleshooting strategies presented in this guide, administrators can confidently manage their network infrastructure, optimizing performance and minimizing disruptions. Whether choosing static or dynamic IP addresses, administrators must consider the specific requirements of their network to ensure optimal functionality. Share this article Shareable URL Prev Post Using Logical Volume Manager (lvm) For Flexible Disk Management Next Post Introduction To The Linux Filesystem Hierarchy Comments 10 1. I honestly belive this is one of the better posts I have read on this website. It explains everything very clearly and concisely, and I learned a lot from it. 2. This post provides a good starting point for understanding static and dynamic IP addresses on Linux. However, I would recommend supplementing it with other resources to get a more comprehensive understanding of the topic. 3. I disagree with the author’s conclusion that dynamic IP addresses are always better than static IP addresses. In some cases, static IP addresses can be more beneficial. 4. I find it ironic that the author is advocating for the use of static IP addresses when dynamic IP addresses are becoming increasingly common. 5. This post is full of sarcasim. The author clearly doesn’t believe that static IP addresses are worth using. 6. I LOL’d at the author’s suggestion that we should all use static IP addresses. It’s like they’re living in the past. 7. I found the post helpful. It explained how to configure static and dynamic IP addresses on Linux in a clear and concise way. Dodaj komentarz Twój adres e-mail nie zostanie opublikowany. Wymagane pola są oznaczone * Read next
__label__pos
0.994991
  Is there a script to remove a profile? Is there a script to remove a profile? I have a list of approx 150 devices that we need to remove a profile from. Is there a way to do this via script to avoid having to go to each device individually or searching for the device manually in the assign section of the profile?   Regards, Johnathon 2 Answers Order By:   Standard | Newest | Votes Matt Dermody | posted this 28 July 2019 There might be another way to accomplish the same goal as well. You could possibly create another group, maybe even a subfolder of their existing group, and then unassign the specific profile to the subgroup. Moving the group of devices into the subgroup will remove the assigned profile but will leave all the other profiles assigned as they should inherit automatically to the subgroup.  • 1 • 0 Raymond Chan | posted this 27 July 2019 There is no MobiControl specific server-side script that you can use.   However,  you can achieve what you want partially by using either 1. MobiControl REST API's 2. Profile filter • 0 • 0 Give us your feedback Give us your feedback Feedback
__label__pos
0.978872
Ratios, proportions, percents worksheets for 6th and 7th grade in the Common Core State Standard. You can also control the amount of workspace, the font, font size, the border around the problems, and additional instructions. Convert to decimals Fractions to decimals (143.9 KiB, 1,010 hits) Percents to decimals (104.0 KiB, 893 hits) Convert to fractions Decimals to fractions (120.9 KiB, 985 hits) Percents to fractions (125.9 KiB, 704 hits) Convert to percents Decimals to percents (100.1 KiB, 776 hits) There is nothing to worry about as decimals, percents, and fractions, as they are just the different types to show a same value. These grade 5 math worksheets give students practice in converting between fractions, decimals and mixed numbers.All worksheets are printable pdf files. Click on the images to view, download, or print them. The relationship between fractions, decimals and percents is clear when you look at the fraction 1/2. Mar 28, 2020 - Explore Carol Camp's board "Fractions/Decimals/Percents", followed by 2930 people on Pinterest. Some of the worksheets displayed are Fractions decimals and percents, Fractions and decimals, Solve each round to the nearest tenth or, Write the name of each decimal place, Fractions decimals percentages, Addsubtracting fractions and mixed numbers, Finding percent change, Multiplying decimals date period. Free trial available at KutaSoftware.com That is seen in the following: 1 divided by 2 equals .5. The worksheets are available both in PDF and html formats (both are easy to print; html format is editable). Read the following statements and answer the survey question. Studying and teaching decimals is done by integrating many core math topics such as: addition, subtraction, multiplication and division, fractions, place value, percents and word problems. You can use our decimal math resources and exercises by selecting and printing our math decimal worksheets. Fractions, Percentages and Decimals Games, Videos and Worksheets First, children learn to recognize fractions, percentages, and decimals. These worksheets show that fractions divided by their denominators equal percents and decimals. Or 3/5 is 3 divided by 5 which equals 0.6. 27% =.27 = Percent with more than two digits is a mixed number. Converting among fractions, decimals and percents. Decimals, Fractions and Percentages. Percent always has a denominator of hundredths. No simplification. Comparing Fractions Decimals And Percent - Displaying top 8 worksheets found for this concept.. 80% j. ... Introduction to Fractions Introduction to Decimals Introduction to Percentages Percentage Calculator Percentage Difference Fractions Index Decimals Index Percents Index. The general information mentioned below about fractions, decimals and percents is very helpful for young children. 15 100.15 b. Learn how to convert fraction to percent and more. Fractions decimals percents math centers, worksheets, and activities to make this concept fun! 73 100 73% c. 39% d. 4 100 e..77 f. 46% g. 50 100 h..06 i. When you divide the numerator by the common denominator you get a decimal. These free fractions to percents to decimal worksheets make teaching about decimals easy. An unlimited supply of printable & customizable worksheets for practicing the conversions between percents and decimals. In each case, a value under one of three topics is provided. You can decide whethe Click one of the buttons below to see all of the worksheets in each set. In the last worksheet, students convert the percent to a decimal as well as a fraction. This relates to a percentage in the sense that .5 is a half and half is represented in percents as 50%. 26 100 Super Teacher Worksheets - www.superteacherworksheets.com This math worksheet was created on 2019-07-05 and has been viewed 29 times this week and 162 times this month. Ratios, proportions, percents worksheets for 6th and 7th grade in the Common Core State Standard, calculating ratio between two given numbers using fractions, find out what is the proportion of a number within a number set, converting proportion into percentage worksheets grade 6 with answers. Percent may be written as a decimals and fractions. This set of worksheets will provide your child with plenty of practice and give her insight as to how math comes in handy during our day-to-day lives. FRACTIONS, DECIMALS, AND PERCENTS LESSON: Fractions, Decimals, and Percents Lesson *. Problem 1 : Look at the number line given below and write the missing decimals and fractions. Displaying top 8 worksheets found for - Ordering Fractions Decimals And Percents. Let's Roll is a great center game where students roll 2 dice to get the percent (ex: rolling a 3 and 4 would be 34%) and converting it to a fraction, a decimal, and filling in a decimal model. Converting fractions to/from decimals worksheets for Grade 5. Explore our fantastic range of fractions, decimals and percents worksheets which are aimed at KS2 children. Lesson excerpt - 100 students were surveyed about their favorite food in the lunch room at your school.The results were calculated. A wide range of colourful and engaging fractions, decimals and percentages teaching resources including unit and lesson plans, real-world maths investigations, worksheets, hands-on activities, PowerPoint presentations, posters and much more. Two worksheets covering converting percentages to fractions. Use these educational resources when teaching your students how to identify and work with fractions, decimals and percentages. Mutual conversion of decimals, fractions and percents worksheets. ID: 924392 Language: English School subject: Math Grade/level: 6 Age: 10-13 Main content: Fractions to Percents Other contents: Percents Add to my workbooks (6) Download file pdf Embed in my website or blog Add to Google Classroom Both progress from easier to harder questions- the harder worksheet moves through the questions quicker and ends with an extension to convert back from fractions to percentages. Once they’ve got the basics down, they’re then expected to add, subtract, multiply, and divide them in a variety of equations. In these worksheets students convert percents to fractions. In addition to the stations there is a student answer sheet as well as "hint cards." Fractions, Decimals, Percents Worksheets. Showing top 8 worksheets in the category - Kuta Software Fractions To Decimal To Percents. This page contains links to free math worksheets for Fractions as Decimals problems. Welcome to the Converting Between Fractions, Decimals and Percents Worksheets section at Tutorialspoint.com.On this page, you will find worksheets on converting a fraction with a denominator of 100 to a percentage, converting a percentage to a fraction with a denominator of 100, finding the percentage of a grid that is shaded, representing benchmark percentages on a grid, converting … Percentages: Part Of One Hundred. name date decimals to percents & fractions sheet 1 answers decimal percent fraction 0.6 60% 60∕ 100 = 3∕ 5 0.2 120% ∕ 5 0.5 50% ½ 0.25 25% ¼ To download worksheet on percents, decimals and fractions, Please click here. See more ideas about math decimals, decimals worksheets, decimals. You can control the workspace, font size, number of decimal digits in the percent, and more. Write each as a percent. 358% = 3.58 = Read these lessons on converting percents to decimals and converting percents to fractions if you need to learn more about changing percents to decimals and fractions. Here is a collection of our printable worksheets for topic Decimals, Fractions and Percents of chapter Percent in section Algebra and Percent.. A brief description of the worksheets is on each of the worksheet widgets. Use repeating decimals when necessary. This worksheet contains tables that suggest the relationships between the three topics listed herein. Fraction to Decimal Drills. Review multiplying decimals and use this method to find the percent of a number, worksheet #2. The worksheets focus on equivalent fractions, finding percentages and adding decimals together which will help Year 3, 4, 5 and 6 children develop their maths skills effectively. 27) 1 2 50 % 28) 1 8 12 .5% 29) 2 3 66 .6% 30) 1 100 1% 31) 2 1 10 210 % 32) 3 8 37 .5% 33) 1 10 10 % 34) 87 100 87 %-2-Create your own worksheets like this one with Infinite Pre-Algebra. Worksheets > Math > Grade 5 > Fractions vs decimals. You can also use the 'Worksheets' menu on the side of this page to find worksheets on other math topics. Fractions, Decimals, and Percents Help your child understand how fractions, decimals, and percentages are all connected. This is a 7 th grade worksheet on converting percents to decimals, fractions or ratios. That's all you need to know to convert the following worksheets on fractions to decimals! Fractions are mostly language-based rather than Math-based. Fractions Percents And Decimals - Displaying top 8 worksheets found for this concept.. Welcome to The Converting from Percents to Fractions, Decimals and Part-to-Part Ratios (Terminating Decimals Only) (A) Math Worksheet from the Fractions Worksheets Page at Math-Drills.com. Fractions: A fraction is part of a whole. Decimals, Fractions and Percentages are just different ways of showing the same value: A Half can be written... As a fraction: 1 / 2. Converting Fractions, Decimals, and Percents fraction decimal percent a. In the 2nd worksheet, they are also asked to simplify the resulting fraction. The worksheets are very customizable: you can choose the number of decimal digits used, the types of denominators (easy, powers of ten, or random), and whether to include improprer fractions and mixed numbers or not. Converting fractions to decimals is a common concept that is often taught in the fifth and sixth grades in most educational jurisdictions. Word Doc PDF Solve word problems about paying sales tax, computing tips, and finding the … About this resource : Included are 6 math stations that allow your students to practice converting between fractions, decimals and percents. For instance 1/2 means the same as 1 divided by 2 which equals 0.5. Mar 4, 2013 - www.worksheetfun.com/category/math-worksheetfunmenu/decimal/. Pacb Stock Forecast 2025, Silent Night, Deadly Night Kill Count, 747 Bus Price, How To Get Pokemon Sword And Shield On Ps4, Earthquake Pakenham Now, Glenn Maxwell T20 Centuries,
__label__pos
0.918907
Maker.js, a Microsoft Garage project, is a JavaScript library for creating and sharing modular line drawings for CNC and laser cutters. View project on GitHub Star Intermediate drawing Zeroing and Centering To move a model so that its bottom and/or left edges are on the x & y axes, use model.zero. This function accepts 2 boolean parameters: zeroOnXAxis, zeroOnYAxis. If you do not pass any parameters, it will zero on both axes. //zero a model var makerjs = require('makerjs'); var model = { models: { crosshairs: { paths: { h: new makerjs.paths.Line([-5, 0], [5, 0]), v: new makerjs.paths.Line([0, -5], [0, 5]) } }, nut: { models: { polygon: new makerjs.models.Polygon(6, 40) }, paths: { inner: new makerjs.paths.Circle(20) } } } }; makerjs.model.zero(model.models.nut); var svg = makerjs.exporter.toSVG(model); document.write(svg); To move a model so that it is centered on on the x & y axes, use model.center. This function accepts 2 boolean parameters: centerOnXAxis, centerOnYAxis. If you do not pass any parameters, it will center on both axes. //center a couple of models var makerjs = require('makerjs'); var model = { models: { crosshairs: { paths: { h: new makerjs.paths.Line([-5, 0], [5, 0]), v: new makerjs.paths.Line([0, -5], [0, 5]) } }, box: { models: { outer: new makerjs.models.Rectangle(60, 30), inner: new makerjs.models.Oval(45, 15) } } } }; var shortcut = model.models.box.models; makerjs.model.center(shortcut.outer); makerjs.model.center(shortcut.inner); var svg = makerjs.exporter.toSVG(model); document.write(svg); Originating A path within a model is referenced relatively to its parent model. There may be times when you want all objects to be within the same coordinate space. Let's create a simple demonstration model: //render a couple boxes in their own coordinate space var makerjs = require('makerjs'); function box(origin) { this.models = { outer: new makerjs.models.RoundRectangle(100, 100, 1) }; this.paths = { inner: new makerjs.paths.Circle([50, 50], 25) }; this.origin = origin; } var box1 = new box([0, 0]); var box2 = new box([150, 0]); var model = { models: { box1: box1, box2: box2 } }; var svg = makerjs.exporter.toSVG(model); document.write(svg); console.log(box1.paths.inner.origin); console.log(box2.paths.inner.origin); In this example, both box1.paths.inner.origin and box2.paths.inner.origin have an origin of [50, 50] even though they are not in the same place, because they are located relative to the model that contains them. To make all models and paths occupy a singular coordinate space, we can use makerjs.model.originate: //render a couple boxes in the same coordinate space var makerjs = require('makerjs'); function box(origin) { this.models = { outer: new makerjs.models.RoundRectangle(100, 100, 1) }; this.paths = { inner: new makerjs.paths.Circle([50, 50], 25) }; this.origin = origin; } var box1 = new box([0, 0]); var box2 = new box([150, 0]); var model = { models: { box1: box1, box2: box2 } }; //move all path origins into the same space makerjs.model.originate(model); var svg = makerjs.exporter.toSVG(model); document.write(svg); console.log(box1.paths.inner.origin); console.log(box2.paths.inner.origin); Now box1.paths.inner.origin and box2.paths.inner.origin have the origins [50, 50] and [200, 50]. Scaling To proportionately scale a simple point, use makerjs.point.scale. To proportionately scale paths and models, use these functions: Each of these functions return the original object, so that we can "chain" on the same line of code. Scale path example: //render a scaled arc var makerjs = require('makerjs'); var arc1 = new makerjs.paths.Arc([0, 0], 25, 0, 90); var arc2 = new makerjs.paths.Arc([0, 0], 25, 0, 90); arc2 = makerjs.path.scale(arc2, 2); var svg = makerjs.exporter.toSVG({ paths: { arc1: arc1, arc2: arc2 }}); document.write(svg); Scale model example: //render a scaled polygon var makerjs = require('makerjs'); var model = { models: { inner: new makerjs.models.Polygon(6, 40), outer: makerjs.model.scale(new makerjs.models.Polygon(6, 40), 1.7) } }; var svg = makerjs.exporter.toSVG(model); document.write(svg); Distorting To disproportionately scale a simple point, use makerjs.point.distort. To disproportionately scale a path, use makerjs.path.distort(path: object, scaleX: number, scaleY: number) which returns a new object and does not modify the original. The type of returned object is dependent on the type of path being distorted: • A line will return a line IPath object, since the distortion can be represented with another line. • An arc will return a BezierCurve IModel object, since the distortion is not circular. • A circle will return an Ellipse IModel object, since the distortion is not circular. Distort path example: //render distorted paths var makerjs = require('makerjs'); var circle = new makerjs.paths.Circle(50); var line = new makerjs.paths.Line([-50,-50], [50, 50]); //a distorted line is a path, so it should be added to paths var distortedLine = makerjs.path.distort(line, 4, 1.5); //a distorted circle is a model, so it should be added to models var ellipse = makerjs.path.distort(circle, 4, 1.5); var model = { paths: { circle: circle, line: line, distortedLine: distortedLine }, models: { ellipse: ellipse } }; var svg = makerjs.exporter.toSVG(model); document.write(svg); To disproportionately scale a model, use makerjs.model.distort(model: object, scaleX: number, scaleY: number) which returns a new IModel object and does not modify the original. Distort model example: //render a distorted star var makerjs = require('makerjs'); var star = new makerjs.models.Star(5, 100); makerjs.model.rotate(star, 18); //make the star 4 times wider, and 2 times taller var wideStar = makerjs.model.distort(star, 4, 2); var model = { models: { star: star, wideStar: wideStar } }; var svg = makerjs.exporter.toSVG(model); document.write(svg); Rotating To rotate a single point, see makerjs.point.fromPolar and makerjs.point.rotate depending on what you are trying to achieve. You can rotate paths and models with these functions: Each of these functions return the original object, so that we can "chain" on the same line of code. Rotate path example: //render a rotated line var makerjs = require('makerjs'); var line1 = new makerjs.paths.Line([0, 0], [100, 0]); var line2 = new makerjs.paths.Line([0, 0], [100, 0]); var paths = [line1, makerjs.path.rotate(line2, -30, [100, 0])]; var svg = makerjs.exporter.toSVG(paths); document.write(svg); Rotate model example: //render a rotated rectangle var makerjs = require('makerjs'); var rect1 = new makerjs.models.Rectangle(40, 80); makerjs.model.rotate(rect1, 45, [0, 0]); var svg = makerjs.exporter.toSVG(rect1); document.write(svg); Cloning Models and paths are simple JavaScript objects, so they are easy to clone in a way that is standard to JavaScript. Maker.js provides a few functions for cloning: Cloning is useful in many situations. For example, if you need many copies of a model for rotation: //clone and rotate var makerjs = require('makerjs'); function sawtooth(numTeeth, r1, rd, offset) { var a = 360 / numTeeth; var a1 = 90 - a / 2; var r2 = r1 + rd; var p1 = makerjs.point.fromPolar(makerjs.angle.toRadians(a1), r1); var p2 = makerjs.point.rotate(p1, a, [0, 0]); var p3 = [-offset, r2]; this.paths = { outer: new makerjs.paths.Arc(p1, p3, r2 / 4, false, false), inner: new makerjs.paths.Arc(p2, p3, r1 / 4, false, false) }; } var wheel = { models: {} }; var numTeeth = 30; var tooth = new sawtooth(numTeeth, 100, 20, 10); for (var i = 0; i < numTeeth; i++ ) { var clone = makerjs.cloneObject(tooth); var a = 360 / numTeeth; makerjs.model.rotate(clone, a * i, [0, 0]); wheel.models[i] = clone; } var svg = makerjs.exporter.toSVG(wheel); document.write(svg); Mirroring Use makerjs.angle.mirror to get a mirror of an angle, and makerjs.point.mirror to get a mirror of a simple point. You can create a mirrored copy of paths and models with the following functions. The mirroring can occur on the x axis, the y axis, or both. Each of these functions returns a new object and does not modify the original. Mirror path example: //render a line mirrored in the x dimension var makerjs = require('makerjs'); var line1 = new makerjs.paths.Line([0, 0], [100, 100]); var line2 = makerjs.path.mirror(line1, true, false); var paths = [line1, line2]; var svg = makerjs.exporter.toSVG(paths); document.write(svg); Mirror model example: //render a model mirrored in the y dimension var makerjs = require('makerjs'); var ovalArc1 = new makerjs.models.OvalArc(45, 135, 50, 10); var model = { models: { ovalArc1: ovalArc1, ovalArc2: makerjs.model.mirror(ovalArc1, false, true) } }; var svg = makerjs.exporter.toSVG(model); document.write(svg); Hint: When creating symmetrical models, it may be easier to create one half, and then use mirror to generate the other half. Repeating layouts Maker.js provides several functions which will clone your paths or models and repeat them in various layouts. Columns Call makerjs.layout.cloneToColumn(path or model, count, [optional] margin) to repeatedly clone and layout in a column. The interval will be the height of the path's or model's bounding box. Extra vertical margin is optional. //Grooves for a finger joint var m = require('makerjs'); var dogbone = new m.models.Dogbone(50, 20, 2, -1, false); var grooves = m.layout.cloneToColumn(dogbone, 5, 20); document.write(m.exporter.toSVG(grooves)); Rows Call makerjs.layout.cloneToRow(path or model, count, [optional] margin) to repeatedly clone and layout in a row. The interval will be the width of the path's or model's bounding box. Extra horizontal margin is optional. //grill of ovals var m = require('makerjs'); var oval = new m.models.Oval(20, 150); var grill = m.layout.cloneToRow(oval, 12, 20); document.write(m.exporter.toSVG(grill)); Grid Call makerjs.layout.cloneToGrid(path or model, xcount, ycount, [optional] margin) to repeatedly clone and layout in a grid. The interval will be the path's or model's bounding box. Extra margin is optional. //grill of rounded squares var m = require('makerjs'); var roundSquare = new m.models.RoundRectangle(20, 20, 4); var grid = m.layout.cloneToGrid(roundSquare, 11, 5, 5); document.write(m.exporter.toSVG(grid)); Brick Call makerjs.layout.cloneToBrick(path or model, xcount, ycount, [optional] margin) to repeatedly clone and layout in a brick wall format. The interval will be the path's or model's bounding box. Extra margin is optional. //brick wall var m = require('makerjs'); var brick = new m.models.Rectangle(20, 8); var wall = m.layout.cloneToBrick(brick, 8, 7, 2); document.write(m.exporter.toSVG(wall)); Honeycomb Call makerjs.layout.cloneToHoneycomb(path or model, xcount, ycount, [optional] margin) to repeatedly clone and layout in a honeycomb format. The interval will be the path's or model's bounding hexagon. Extra margin is optional. //Honeycomb var m = require('makerjs'); var star = m.model.rotate(new m.models.Star(6, 50, 0, 2), 30); var pattern = m.layout.cloneToHoneycomb(star, 8, 5, 30); document.write(m.exporter.toSVG(pattern)); Radial Call makerjs.layout.cloneToRadial(path or model, count, angleInDegrees, [optional] rotationOrigin) to repeatedly clone and layout in a radial format. //spinner var m = require('makerjs'); var rect = m.model.move(new m.models.Rectangle(30, 10), [40, -5]); var spinner = m.layout.cloneToRadial(rect, 16, 22.5); document.write(m.exporter.toSVG(spinner)); Intersection You can find the point(s) of intersection between two paths using makerjs.path.intersection. If the paths do not intersect, this function will return null. Otherwise, it will return an object with a property named intersectionPoints which is an array of points. Additionally, if either path was an arc or circle, this object will contain the angles at which an intersection occurred. Intersection examples: //line-line intersection var makerjs = require('makerjs'); var model = { paths: { line1: new makerjs.paths.Line([0, 0], [20, 10]), line2: new makerjs.paths.Line([2, 10], [50, 2]) } }; var int = makerjs.path.intersection(model.paths.line1, model.paths.line2); if (int) { var p = int.intersectionPoints[0]; var id = JSON.stringify(makerjs.point.rounded(p, 0.01)); model.paths[id] = new makerjs.paths.Circle(p, 1); } var svg = makerjs.exporter.toSVG(model); document.write(svg); //circle-circle intersection var makerjs = require('makerjs'); var model = { paths: { circle1: new makerjs.paths.Circle([0, 10], 20), circle2: new makerjs.paths.Circle([20, 0], 20) } }; var int = makerjs.path.intersection(model.paths.circle1, model.paths.circle2); if (int) { int.intersectionPoints.forEach( function(p, i) { var id = JSON.stringify(makerjs.point.rounded(p, 0.01)) + ' intersects circle1 at ' + makerjs.round(int.path1Angles[i], .01) + ' circle2 at ' + makerjs.round(int.path2Angles[i], .01); model.paths[id] = new makerjs.paths.Circle(p, 1); } ); } var svg = makerjs.exporter.toSVG(model); document.write(svg); //line-arc intersection var makerjs = require('makerjs'); var model = { paths: { line1: new makerjs.paths.Line([0, 0], [20, 10]), arc1: new makerjs.paths.Arc([12, 0], 10, 45,215) } }; var int = makerjs.path.intersection(model.paths.line1, model.paths.arc1); if (int) { int.intersectionPoints.forEach( function(p, i) { var id = JSON.stringify(makerjs.point.rounded(p, 0.01)) + ' arc1 at ' + makerjs.round(int.path2Angles[i], .01); model.paths[id] = new makerjs.paths.Circle(p, 1); } ); } var svg = makerjs.exporter.toSVG(model); document.write(svg); Converging lines To make lines meet at their slope intersection point, use makerjs.path.converge. This function will only work with lines, it will not work with arcs. The converge function will try to use the end of the line that is closest to the convergence point. If you need to specify which ends of your lines should be converged, pass two additional boolean values. The boolean value is true to use the line's origin, false to use the end. Converge example: //converge lines var makerjs = require('makerjs'); var model = { origin: [0, 0], paths: { line1: new makerjs.paths.Line([0, 0], [10, 5]), line2: new makerjs.paths.Line([0, 10], [10, 4]), line3: new makerjs.paths.Line([1, 0], [5, -2]) } }; var clone1 = makerjs.cloneObject(model); clone1.origin = [10, 0]; var clone2 = makerjs.cloneObject(model); clone2.origin = [20, 0]; makerjs.path.converge(clone1.paths.line1, clone1.paths.line2); makerjs.path.converge(clone1.paths.line1, clone1.paths.line3); makerjs.path.converge(clone2.paths.line1, clone2.paths.line2, false, true); makerjs.path.converge(clone2.paths.line1, clone2.paths.line3, true, false); var svg = makerjs.exporter.toSVG({ models: { before: clone1, after: model, x: clone2 } }); document.write(svg); Modifying models We know that models are relatively simple objects with a well known recursive structure. This allows us to modify them for different purposes. Let's modify and combine two different models in one drawing. For this example we will use ovals to make an oval L shape. We begin by creating a model function that has two ovals: //render two ovals which overlap var makerjs = require('makerjs'); function ovalL(width, height, thickness) { var ovalH = new makerjs.models.Oval(width, thickness); var ovalV = new makerjs.models.Oval(thickness, height); this.models = { h: ovalH, v: ovalV }; } var svg = makerjs.exporter.toSVG(new ovalL(100, 100, 37)); document.write(svg); There are overlapping arcs in the lower left corner. We can remove them if we know their id and position in the heirarchy. There are several ways we can inspect this model, here are a few: • Look at the code which created it. This may involve deep lookups. For example, the Oval source code references the RoundRectangle source code. • Use the browser's console, or JavaScript debugger to set a breakpoint in your model. • Use the browser's DOM inspector to traverse the rendered SVG. • Output the raw JSON or SVG on screen. • Use the Playground app and click "show path names". By looking at the source code we know that an Oval is a RoundRectangle and that the ids for arcs are BottomLeft, BottomRight, TopLeft and TopRight. The ids for the sides are Left, Right, Top and Bottom. Also, we need to note the orientation of these lines so we know which are origin and end points. To remove a path we use the JavaScript delete keyword: //render two ovals which overlap var makerjs = require('makerjs'); function ovalL(width, height, thickness) { var ovalH = new makerjs.models.Oval(width, thickness); var ovalV = new makerjs.models.Oval(thickness, height); //delete the lower arcs from the vertical oval delete ovalV.paths.BottomLeft; delete ovalV.paths.BottomRight; //delete the inside arc of the horizontal delete ovalH.paths.TopLeft; this.models = { h: ovalH, v: ovalV }; } var svg = makerjs.exporter.toSVG(new ovalL(100, 100, 37)); document.write(svg); The next step is to eliminate the overlap in the lines. Here are two approaches to do this: Adjust only the x or y component of the point: //render an L shape, modifying points by their x and y var makerjs = require('makerjs'); function ovalL(width, height, thickness) { var ovalH = new makerjs.models.Oval(width, thickness); var ovalV = new makerjs.models.Oval(thickness, height); delete ovalV.paths.BottomLeft; delete ovalV.paths.BottomRight; delete ovalH.paths.TopLeft; //move the x of the horizontal's top ovalH.paths.Top.end[0] = thickness; //move the y of the vertical's right ovalV.paths.Right.origin[1] = thickness; this.models = { h: ovalH, v: ovalV }; } var svg = makerjs.exporter.toSVG(new ovalL(100, 100, 37)); document.write(svg); Share a point on both lines: //render an L shape, sharing a point var makerjs = require('makerjs'); function ovalL(width, height, thickness) { var ovalH = new makerjs.models.Oval(width, thickness); var ovalV = new makerjs.models.Oval(thickness, height); delete ovalV.paths.BottomLeft; delete ovalV.paths.BottomRight; delete ovalH.paths.TopLeft; //set to the same point ovalH.paths.Top.end = ovalV.paths.Right.origin = [thickness, thickness]; this.models = { h: ovalH, v: ovalV }; } var svg = makerjs.exporter.toSVG(new ovalL(100, 100, 37)); document.write(svg); Let's progress this example further, by modifying an L shape into a C shape. Create a new model function for C, and immediately create an L within it. The C may create a new models object for itself, and nest the L inside; alternatively, C can just assume L's models object: //render an L with an oval over it var makerjs = require('makerjs'); function ovalL(width, height, thickness) { var ovalH = new makerjs.models.Oval(width, thickness); var ovalV = new makerjs.models.Oval(thickness, height); delete ovalV.paths.BottomLeft; delete ovalV.paths.BottomRight; delete ovalH.paths.TopLeft; ovalH.paths.Top.end = ovalV.paths.Right.origin = [thickness, thickness]; this.models = { h: ovalH, v: ovalV }; } function ovalC(width, height, thickness) { //assume the same models as L this.models = new ovalL(width, height, thickness).models; //add another oval this.models.h2 = new makerjs.models.Oval(width, thickness); //move it to the top this.models.h2.origin = [0, height - thickness]; } //using C instead of L var svg = makerjs.exporter.toSVG(new ovalC(100, 100, 37)); document.write(svg); Just as before, we need to delete the overlapping paths using the delete keyword. Let us also make a short alias for this.models to save us some keystrokes: //render an L and form a C var makerjs = require('makerjs'); function ovalL(width, height, thickness) { var ovalH = new makerjs.models.Oval(width, thickness); var ovalV = new makerjs.models.Oval(thickness, height); delete ovalV.paths.BottomLeft; delete ovalV.paths.BottomRight; delete ovalH.paths.TopLeft; ovalH.paths.Top.end = ovalV.paths.Right.origin = [thickness, thickness]; this.models = { h: ovalH, v: ovalV }; } function ovalC(width, height, thickness) { //set local var m for easy typing var m = this.models = new ovalL(width, height, thickness).models; m.h2 = new makerjs.models.Oval(width, thickness); m.h2.origin = [0, height - thickness]; //delete overlapping arcs again delete m.h2.paths.TopLeft; delete m.h2.paths.BottomLeft; delete m.v.paths.TopRight; } var svg = makerjs.exporter.toSVG(new ovalC(100, 100, 37)); document.write(svg); Lastly, we need our overlapping lines to meet at a common point. Notice that the new oval h2 has a different origin the the previous ovals. So, we must originate for all of the ovals to share the same coordinate space. Afterwards, we can assign the common point to both lines. In the Play editor, try removing the call to originate to see the results without it. //render a C shape var makerjs = require('makerjs'); function ovalL(width, height, thickness) { var ovalH = new makerjs.models.Oval(width, thickness); var ovalV = new makerjs.models.Oval(thickness, height); delete ovalV.paths.BottomLeft; delete ovalV.paths.BottomRight; delete ovalH.paths.TopLeft; ovalH.paths.Top.end = ovalV.paths.Right.origin = [thickness, thickness]; this.models = { h: ovalH, v: ovalV }; } function ovalC(width, height, thickness) { var m = this.models = new ovalL(width, height, thickness).models; m.h2 = new makerjs.models.Oval(width, thickness); m.h2.origin = [0, height - thickness]; delete m.h2.paths.TopLeft; delete m.h2.paths.BottomLeft; delete m.v.paths.TopRight; //h2 has paths relative to h2 origin, //we need to originate to share the point makerjs.model.originate(this); //share the point m.h2.paths.Bottom.origin = m.v.paths.Right.end = [thickness, height - thickness]; } var svg = makerjs.exporter.toSVG(new ovalC(100, 100, 37)); document.write(svg); Breaking paths You can break paths into two pieces if you have a point that lies on the path (from an intersection, for example) by using makerjs.path.breakAtPoint. This function will change the path that you pass it, so that it is broken at that point, and it will return a new path object which is the other broken piece: //break a path in two var makerjs = require('makerjs'); var model = { paths: { arc: new makerjs.paths.Arc([0, 0], 50, 0, 180) } }; var arc2 = makerjs.path.breakAtPoint(model.paths.arc, [0, 50]); makerjs.model.moveRelative(arc2, [-10, 0]); model.paths.arc2 = arc2; var svg = makerjs.exporter.toSVG(model); document.write(svg); For Circle, the original path will be converted in place to an Arc, and null is returned. Fillets Fillets are round corners where two paths meet. Maker.js provides two types of fillets: traditional fillets and dogbone fillets. Traditional fillet Rounding a corner can add strength to your part, as well as make it faster to print. Using makerjs.path.fillet you can round a corner at the junction between two lines, two arcs, or a line and an arc. This function will clip the two paths that you pass it, and will return a new arc path which fits between the clipped ends. The paths must meet at one point, this is how it determines which ends of the paths to clip. You also provide a radius of the fillet. If the fillet cannot be created this function will return null. //fillet between lines var makerjs = require('makerjs'); var model = { paths: { line1: new makerjs.paths.Line([0, 20], [30, 10]), line2: new makerjs.paths.Line([10, 0], [30, 10]) } }; //create a fillet var arc = makerjs.path.fillet(model.paths.line1, model.paths.line2, 2); //add the fillet to the model model.paths.arc = arc; var svg = makerjs.exporter.toSVG(model); document.write(svg); //fillet between arcs var makerjs = require('makerjs'); var model = { paths: { arc1: new makerjs.paths.Arc([0, 50], 50, 270, 0), arc2: new makerjs.paths.Arc([100, 50], 50, 180, 270) } }; //create a fillet var arc = makerjs.path.fillet(model.paths.arc1, model.paths.arc2, 2); //add the fillet to the model model.paths.arc = arc; var svg = makerjs.exporter.toSVG(model); document.write(svg); //fillet between line and arc (or arc and line!) var makerjs = require('makerjs'); var model = { paths: { arc: new makerjs.paths.Arc([0, 50], 50, 270, 0), line: new makerjs.paths.Line([50, 50], [50, 0]) } }; //create a fillet var arc2 = makerjs.path.fillet(model.paths.arc, model.paths.line, 2); //add the fillet to the model model.paths.arc2 = arc2; var svg = makerjs.exporter.toSVG(model); document.write(svg); Dogbone Fillets Many CNC tools are not able to cut a sharp interior corner. The way to clear the apex of an interior corner is by encompassing the corner with a circular cut known as a dogbone fillet. Use makerjs.path.dogbone to round a corner at the junction between two lines. This function will only work for two lines which must meet at one point. It will clip the two lines that you pass it, and will return a new arc path which clears the corner where the lines meet. It will return null if a dogbone fillet cannot be created at the radius you specify. //dogbone fillet between lines. var makerjs = require('makerjs'); var model = { paths: { line1: new makerjs.paths.Line([0, 0], [0, 5]), line2: new makerjs.paths.Line([0, 5], [10, 5]) } }; //create dogbone fillet var arc1 = makerjs.path.dogbone(model.paths.line1, model.paths.line2, 1); //add the fillet to the model model.paths.arc1 = arc1; var svg = makerjs.exporter.toSVG(model); document.write(svg); Dogbone models If you need a rectangle with dogbones at each corner, you can use makerjs.models.Dogbone(width, height, radius, optional style, optional bottomless). There are a 3 styles of a Dogbone model: • 0 : (default) rounded • -1 : horizontal • -1 : vertical Dogbone model corner styles: //dogbone corner styles. var makerjs = require('makerjs'); var dogbones = { models: { round: new makerjs.models.Dogbone(100,50, 5, 0), horizontal: new makerjs.models.Dogbone(100,50, 5, -1), vertical: new makerjs.models.Dogbone(100,50, 5, 1) } }; dogbones.models.horizontal.origin = [115, 0]; dogbones.models.vertical.origin = [230, 0]; var svg = makerjs.exporter.toSVG(dogbones); document.write(svg); Making them bottomless is useful for creating tongue-and-groove shapes: //bottomless dogbones. var makerjs = require('makerjs'); var dogbones = { models: { O: new makerjs.models.Dogbone(100,50, 5, 0, true), horizontal: new makerjs.models.Dogbone(100,50, 5, -1, true), vertical: new makerjs.models.Dogbone(100,50, 5, 1, true) } }; dogbones.models.horizontal.origin = [115, 0]; dogbones.models.vertical.origin = [230, 0]; var svg = makerjs.exporter.toSVG(dogbones); document.write(svg); Layers Layers are a way of logically grouping your paths or models as you see fit. Simply add a layer property to any path or model object, with the name of the layer. Every path within a model will automatically inherit its parent model's layer, unless it has its own layer property. As you can see in this example, a layer can transcend the logical grouping boundaries of models: //render a round rectangle with arcs in their own layer var makerjs = require('makerjs'); var roundRect = new makerjs.models.RoundRectangle(100, 50, 10); roundRect.layer = "layer1"; roundRect.paths.BottomLeft.layer = "layer2"; roundRect.paths.BottomRight.layer = "layer2"; roundRect.paths.TopRight.layer = "layer2"; roundRect.paths.TopLeft.layer = "layer2"; var svg = makerjs.exporter.toSVG(roundRect); document.write(svg); Layers are not visible in this example but they logically exist to separate arcs from straight lines. A layer name can be any string. Furthermore, you can use a reserved color name from this list to get an automatic stroke color when your model is exported in DXF or SVG: aqua, black, blue, fuchsia, green, gray, lime, maroon, navy, olive, orange, purple, red, silver, teal, white, yellow //render a round rectangle with arcs in their own color layer var makerjs = require('makerjs'); var roundRect = new makerjs.models.RoundRectangle(100, 50, 10); roundRect.layer = "layer1"; roundRect.paths.BottomLeft.layer = "red"; roundRect.paths.BottomRight.layer = "red"; roundRect.paths.TopRight.layer = "blue"; roundRect.paths.TopLeft.layer = "blue"; var svg = makerjs.exporter.toSVG(roundRect); document.write(svg); Layers will be output during the export process in these formats: • DXF - paths will be assigned to a DXF layer. • SVG - in continuous mode, a new <path> element will be created for each layer. Cascading functions When calling a function, you can pass its output directly into another function. This is called cascading. This lets you do multiple operations in one statement. Here we will center, rotate and move a square: //cascade functions var makerjs = require('makerjs'); //many operations in this one statement var square = makerjs.model.moveRelative( makerjs.model.rotate( makerjs.model.center( new makerjs.models.Square(10) ), 45), [0, 15]) ; var drawing = { models: { dome: new makerjs.models.Dome(30, 30), square: square } }; var svg = makerjs.exporter.toSVG(drawing); document.write(svg); This is convenient, but it also has the drawback of making the code less readable. As more function calls are added, the parameters associated with the call are separated outward. Also notice that the final operation (moveRelative) appears at the beginning of the statement. The $ function As an alternative to cascading functions, Maker.js offers a handy way to modify your drawing in an object-oriented style, inspired by the jQuery library. Call makerjs.$(x) to get a cascade container object returned. You can then invoke a series of cascading functions upon x. The output of each function becomes the input of the next. A cascade container will only work with functions that output the same type of object that they input as their first parameter, which must be one of these types: • Model • Path • Point Container operators A cascade container will have special properties that operate the container itself (as opposed to operating upon x). These are prefixed with $: • $initial: object Gets the original x object that you passed in. • $result: object Gets the final result of all cascaded function calls. • $reset(): function() - Resets the container to $initial. Cascadable functions Depending on the type of x, a cascade container will provide all of the functions from one of the corresponding modules. These are the same functions that we've covered in previous lessons. One difference is that you do not need to provide the first parameter, since it will either be x or the cascaded result of the previous function call. Example Let's rewrite the example from above to compare the readability of the code: //cascade functions var makerjs = require('makerjs'); //many operations in this one statement var square = makerjs.$(new makerjs.models.Square(10)) .center() .rotate(45) .moveRelative([0, 15]) .$result; var drawing = { models: { dome: new makerjs.models.Dome(30, 30), square: square } }; var svg = makerjs.exporter.toSVG(drawing); document.write(svg); This has saved us some typing - we didnt need to use makerjs.model... to access any functions. The order of operations makes more sense too: the first operation (center()) is at the top, the final operation (moveRelative([0, 15])) is at the bottom, and the function parameters are together with their call. Using addTo() instead of .$result In some cases, you can avoid using .$result and just add a path or add a model to a parent model by calling addTo(model, id). This is particularly useful prior to a call that creates a clone (such as mirror): //use addTo with mirror var makerjs = require('makerjs'); var starburst = {}; makerjs.$(new makerjs.paths.Arc([5, 5], 5, 180, 270)) .addTo(starburst, 'arc1') .mirror(true, false) .addTo(starburst, 'arc2') .mirror(false, true) .addTo(starburst, 'arc3') .mirror(true, false) .addTo(starburst, 'arc4'); var svg = makerjs.exporter.toSVG(starburst); document.write(svg); Captions Captions are fragments of text that can be positioned anywhere in your model, useful for adding documentation within your drawing. Captions are unlike the Text model, which is a line drawing of glyphs in a given font. A caption is aligned to an invisible line called an anchor. The caption text is centered both horizontally and vertically at the center point of the anchor line. The text in a caption will not wrap, it is a single line of text. The text and anchor line do not need to be the same length, the anchor line is only used to determine the center point and the slope. The anchor line may be rotated to angle the caption text. Anchor lines are moved, originated, scaled, distorted and rotated accoordingly within a model. The font size of caption text is determined when you export your model. Note: In the Playground, caption text does not scale when you zoom in or out. Creating a caption object A caption is an object with these two properties: • text - String • anchor - Line object Add this to a model via the caption property: //add a caption to a model var makerjs = require('makerjs'); var square = new makerjs.models.Square(100); square.caption = { text: "a square", anchor: new makerjs.paths.Line([0, 50], [100, 50]) }; var svg = makerjs.exporter.toSVG(square); document.write(svg); There is a helper function makerjs.model.addCaption(text, [optional] leftAnchorPoint, [optional] rightAnchorPoint) which lets you add a caption on one line of code: //add a caption to a model with the helper var makerjs = require('makerjs'); var square = new makerjs.models.Square(100); makerjs.model.addPath(square, new makerjs.paths.Line([10, 10], [90, 90])); makerjs.model.addCaption(square, 'fold here', [10, 20], [80, 90]); var svg = makerjs.exporter.toSVG(square); document.write(svg); If the anchor line is degenerate (meaning its origin and end point are the same), you can achieve text which will remain horizontally aligned regardless of model rotation: //add a caption that will not rotate var makerjs = require('makerjs'); var square = makerjs.$(new makerjs.models.Square(100)) .addCaption('always aligned', [50, 50], [50, 50]) .rotate(22) .$result; var svg = makerjs.exporter.toSVG(square); document.write(svg); Next: learn more in Advanced drawing.
__label__pos
0.96577
=head1 Customizing the Publish Process in Krang =over =item * Introduction =item * Why Customize Publish Behavior in the Element Library =item * Changing the data returned by an element =over =item * template_data() =item * Story and Media Links =back =item * Changing how an element populates a template =over =item * fill_template() =item * Sample Element =item * Sample Templates =item * Option 1 - Make Small Changes, Let Krang Finish Template Population =over =item * Example - Adding one variable =back =item * Option 2 - Populating the Template Manually =over =item * Example 1 - A Single Variable =item * Example 2 - Element Children =item * Example 3 - Adding Contributors =item * Example 4 - Passing Parameters to Child Elements =over =item * fill_template_args =item * Handling fill_template_args When Overriding fill_template() =item * A Note on Passing Parameters =back =back =back =item * Generating Additional Content =over =item * Example - Creating an RSS File =item * Example - Generating a Wall Page =back =item * Changing How an Element Chooses a Template =over =item * Loading a Template =back =item * Changing the Publish Process for an Element =over =item * Preventing an Element from Publishing =item * Forcing Publish Without a Template =back =item * Conclusion =back =head1 Introduction This document covers the concept of customizing the publish process by making changes to the element library. It assumes that you've already got an understanding of templates and element libraries in Krang. It's a good idea to have read the following documents before going any further: =over =item * HREF[Writing HTML::Template Templates in Krang|writing_htmltemplate.html] =item * Creating an Element Library (TBD - Sam) =item * The POD for the CPAN modules L and L =item * The API documentation (L, L, L, L, L) =back =head1 Why Customize Publish Behavior in the Element Library Out of the box, Krang populates element templates according to a fixed set of rules. The standard publish process should be sufficient for publication of most sites. That being said, choices in how data is returned and how data is organized in the templates have been made - if these choices don't work with what you're attempting to accomplish, your next step is to change the behavior of the elements themselves. =head1 Changing the data returned by the element The simplest thing to change in an element is the form the element's data takes when returned. With a few exceptions (see below) elements return data in the same form as it was when stored. The advantage to this technique is that the results will be seen, regardless of whether or not a template is used. =over =item * template_data() This returns the data stored in the element. In most cases, it's the actual data stored in the element, unformatted. In the case of L or L objects, C will return the fully-qualified URL of the object. =back Suppose you wanted all header elements to return their data in all-caps when published, regardless of how they were entered into the system. At the same time, you don't want to actually make that change to the content itself, in case you change your mind later. Overriding C in your element library's header.pm as follows will do the trick: sub template_data { my $self = shift; my %args = @_; my $element = $args{element}; return uc($element->data()); } =head2 Story and Media Links Elements that handle links to Stories and Media need to be handled a little bit differently - they need to return the URL of the object being pointed to, rather than the data itself, and they need to return a URL that's consistent with the current output mode - publish or preview. Keep this in mind if you consider changing the behavior for either of these two. Here is how template_data() currently works for elements using L - sub template_data { my $self = shift; my %args = @_; my $element = $args{element}; if ($args{publisher}->is_publish()) { return 'http://' . $element->data()->url(); } elsif ($args{publisher}->is_preview()) { return 'http://' . $element->data()->preview_url(); } else { croak (__PACKAGE__ . ': Not in publish or preview mode. Cannot return proper URL.'); } } In short, it queries the publisher (C<$args{publisher}>) to determine if the mode is publish or preview (returning an error if it's neither). C<$element->data()> returns a L object (if this was L, it would be a L object). Depending on the mode. the appropriate URL is returned. =head1 Changing how an element populates a template The next option is more ambitious - changing how an element goes about populating the variables in a template. At this point, you have two options - you can piggyback your changes on top of the work that Krang does, or you can choose to do it all yourself. =over =item * fill_template() fill_template() is responsible for filling the template object with data built from the element tree. Generally, it traverses the element tree, creating scalars and loops on an as-needed basis, populating the template objects with the results. The rules by which fill_template() operates can be found in the section HREF[How Krang Builds Template Data|writing_htmltemplate.html#how%20krang%20builds%20template%20data] in HREF[Writing HTML::Template Templates in Krang|writing_htmltemplate.html]. If you are familiar with Bricolage, fill_template() functions using the same rules as the autofill() functionality found in Bricolage. =back =head2 Option 1 - Make Small Changes, Let Krang Finish Template Population With the object hierarchy Krang provides, you can make small additions to fill_template() and then let Krang pick things up from there by calling the parent method's C. =head3 Sample Element For these examples, we will be using the following Story element: Story - Deck (subclass of Krang::ElementClass::Text) + Page (subclass of Krang::ElementClass) - Header (subclass of Krang::ElementClass::Text) - Paragraph (subclass of Krang::ElementClass::TextArea) - Pull Quote (subclass of Krang::ElementClass::Text) - Paragraph (subclass of Krang::ElementClass::TextArea) =head3 Sample Templates The story element will use the following templates: =head4 Story.tmpl <tmpl_var title> =head4 Page.tmpl =head3 Example - Adding one variable As a simple example, we want to add a variable C to the page template. This can be done by overriding C in the article element, and adding a variable C to the template. Other than this variable, the template should be populated as usual. =head4 The new method This is the new C that would be used in the Page element: sub fill_template { my $self = shift; my %args = @_; my $template = $args{tmpl}; $template->param(greeting => 'Hello World!'); return $self->SUPER::fill_template(@_); } In short, add the variable C to the template, and then call the parent C method that was overridden by this method (passing the original set of parameters along). The rest of the publish process is unaffected, and nothing will be noticed on output until the article template uses C. =head4 Page.tmpl This new Page template will display the greeting: The output for Page.tmpl (not the entire story, mind you) will look something like this: Hello World! Header Header paragraph1 paragraph1 paragraph1 Quote Quote Quote paragraph2 paragraph2 paragraph2 =head2 Option 2 - Populating the Template Manually If you choose to populate the template manually, all the variables that Krang builds no longer apply. Additionally, it will be up to you to build variables based on the child elements of the current element. =head3 Submitted Parameters C takes a set of named parameters - =over =item * publisher The Krang::Publisher object for this publish run. =item * tmpl The HTML::Template object for the template being with this element. =item * element The Krang::Element object currently being published. =back C is expected to return the HTML that results from populating the template. Read the L API documentation for further documentation on the actual interface. =head3 Example 1 - A single variable This example re-uses the example from Option 1 - adding a single variable C to the template. B sub fill_template { my $self = shift; my %args = @_; my $template = $args{tmpl}; $template->param(greeting => 'Hello World!'); return $template->output(); } Where the previous example in option 1 made a call to C<$self->SUPER::fill_template()>, this example simply calls C<$template->output()>. The result is that in this example, with no parent method to do additional work populating the template, the only variable available to the template is C. B The template from the previous example will still work: By: and , ( ) However, with Krang not providing any additional variables, the template output will look like this: Hello World! =head3 Example 2 - Element Children Clearly, the output for the above template isn't what we're looking for - the header is missing, along with the entire element loop. The next step is to add these: B sub fill_template { my $self = shift; my %args = @_; my @element_loop; my %params; my $template = $args{tmpl}; my $element = $args{element}; my $publisher = $args{publisher}; $params{greeting} = 'Hello World!'; # retrieve the list of child elements my @children = $element->children(); foreach my $child (@children) { my $name = $child->name(); my $html = $child->publish(publisher => $publisher); unless (exists($params{$name})) { $params{$name} = $html; } push @{$params{element_loop}}, { "is_$name" => 1, $name => $html }; } $template->param(%params); return $template->output(); } Make sense? Rather than make a lot of calls to C<$template->param()>, parameters are stored in %params until all work is finished. The loop at the bottom iterates over the list of children (C<@children>), building HTML for each child, and then placing the results in %params - you can see the element_loop being built there as well. The resulting output looks like what we want: Hello World! Header Header paragraph1 paragraph1 paragraph1 Quote Quote Quote paragraph2 paragraph2 paragraph2 =head3 Example 3 - Adding Contributors Adding contributors here is a straightforward process - a single method call makes it possible: $contrib_loop = $self->_build_contrib_loop(@_); This can be added to C as follows: B sub fill_template { my $self = shift; my %args = @_; my @element_loop; my %params; my $template = $args{tmpl}; my $element = $args{element}; my $publisher = $args{publisher}; $params{greeting} = 'Hello World!'; $params{contrib_loop} = $self->_build_contrib_loop(@_); # retrieve the list of child elements my @children = $element->children(); foreach my $child (@children) { my $name = $child->name(); my $html = $child->publish(publisher => $publisher); unless (exists($params{$name})) { $params{$name} = $html; } push @{$params{element_loop}}, { "is_$name" => 1, $name => $html }; } $template->param(%params); return $template->output(); } With the C now added to the template: B By: and , ( ) The resulting output will look something like this: Hello World! Header Header By: JR Bob Dobb (Writer, Photographer) and Venus Dee Milo (Illustrator) paragraph1 paragraph1 paragraph1 Quote Quote Quote paragraph2 paragraph2 paragraph2 Go back to the HREF[Contributors|writing_htmltemplate.html#contributors] section of HREF[Writing HTML::Template Templates in Krang|writing_htmltemplate.html] for further documentation on using Contributors. =head3 Example 4 - Passing Parameters to Child Elements It may come about that you want to pass information along to a child element for use when it goes through the publish process. This can be done by adding arguments to the named parameters passed to C<$child->publish()>. =head4 fill_template_args If the child element you are calling is still using the C method provided by Krang, you can use the parameter C in the fashion below: foreach my $child ($element->children()) { my %new_args = (greeting => 'Hello World!'); my $html = $child->publish(publisher => $publisher, fill_template_args => \%new_args); $params{$child->name} = $html; } When the child element goes through the publish process, its C method will add C to template, provided the template is looking for a variable C. Keep in mind, you don't need to override C in the child element - this functionality is supported out-of-the-box. Using the same Page.tmpl we started with at the beginning of these examples: B Rather than override the Page element's C method, we're going to use the one provided by Krang. Instead, we're going to override the C method for the Story element, and have it pass along the greeting to the page element. B sub fill_template { my $self = shift; my %args = @_; my @element_loop; my %params; my $template = $args{tmpl}; my $element = $args{element}; my $publisher = $args{publisher}; my $story = $publisher->story(); $params{title} = $story->title(); # retrieve the list of child elements my @children = $element->children(); foreach my $child (@children) { my $name = $child->name(); my $html = $child->publish(publisher => $publisher, fill_template_args => { greeting => 'Hello World!' }); unless (exists($params{$name})) { $params{$name} = $html; } if ($name eq 'page') { push @{$params{"$name_loop"}}, { $name => $html }; } } $template->param(%params); return $template->output(); } With the Story element passing the C parameter along, you don't need to override the C method in the Page element - the standard method used by Krang will suffice. =head4 Handling fill_template_args When Overriding fill_template() Continuing from the previous example, if we were to override the Page element C method, we'd need to handle the C parameter as well: B sub fill_template { my $self = shift; my %args = @_; my %params; my $template = $args{tmpl}; my $element = $args{element}; my $publisher = $args{publisher}; # Additional template params passed in by the parent element. if (exists($args{fill_template_args})) { foreach my $arg (keys %{$args{fill_template_args}}) { $params{$arg} = $args{fill_template_args}{$arg}; } } # retrieve the list of child elements my @children = $element->children(); foreach my $child (@children) { my $name = $child->name(); my $html = $child->publish(publisher => $publisher); unless (exists($params{$name})) { $params{$name} = $html; } push @{$params{element_loop}}, { "is_$name" => 1, $name => $html }; } $template->param(%params); return $template->output(); } The small block that consists of: if (exists($args{fill_template_args})) { foreach my $arg (keys %{$args{fill_template_args}}) { $params{$arg} = $args{fill_template_args}{$arg}; } } Is the extent of what's needed to handle C. =head4 A Note on Passing Parameters While we use C in the previous examples, you can call $child->publish(publisher => $publisher, any_param_name_you_want => $foo); And the additional parameter(s) will get passed along to C method of the child object. It is up to you, of course, to make use of it on that end - the standard Krang implementation only makes use of C. =head1 Generating Additional Content While publishing a given story, you may want to publish additional files containing data related to the story. For example, an RDF file for syndication, or an XML file containing keywords for search-engines, or an article preview page for subscription purposes. While Krang doesn't provide direct support for such things within the UI, it provides a framework for you to use within your element library, allowing you to build the content in any way you see fit. During the publish process, you can add additional content at any point that you have the C object available. For example: # Write out 'extra.html' in conjunction with this story. my $additional_content = create_sidebar_story(); $publisher->additional_content_block(content => $additional_content, filename => 'extra.html', use_category => 1); At the end of the publish process, Krang will handle the entry in C separately from the main story, and write it to disk as 'extra.html' (or whatever you set C to be). B =over =item * An arbitrary number of additional files can be added created - just be careful filenames do not overlap. =item * You have the option to add (or not add) the current Category templates (e.g. header/footer) to your output. Simply set C to 1 if you want to wrap C<$additional_content> in the category templates, or set it to 0 if you want it to be written to disk as-is. =back B =over =item * No pagination - you can only create single-page files. =item * The file will get written out to the same directory as the story itself - writing to other directories is not supported (L does a lot of work to protect against directory conflicts, don't want to interfere with that). =back For the following examples, we're going to use the following Story element tree: Story - Deck (subclass of Krang::ElementClass::Text) + Page (subclass of Krang::ElementClass) - Header (subclass of Krang::ElementClass::Text) - Paragraph (subclass of Krang::ElementClass::TextArea) - Pull Quote (subclass of Krang::ElementClass::Text) - Paragraph (subclass of Krang::ElementClass::TextArea) - Leadin (subclass of Krang::ElementClass::StoryLink) - Leadin (subclass of Krang::ElementClass::StoryLink) - Leadin (subclass of Krang::ElementClass::StoryLink) + Page (subclass of Krang::ElementClass) - Paragraph (subclass of Krang::ElementClass::TextArea) - Paragraph (subclass of Krang::ElementClass::TextArea) - Pull Quote (subclass of Krang::ElementClass::Text) - Paragraph (subclass of Krang::ElementClass::TextArea) + Page (subclass of Krang::ElementClass) - Paragraph (subclass of Krang::ElementClass::TextArea) - Paragraph (subclass of Krang::ElementClass::TextArea) - Paragraph (subclass of Krang::ElementClass::TextArea) =head2 Example - Creating an RSS File If you aren't familiar with RSS (RDF Site Summary), take a look here: HREF[RDF Site Summary 1.0|http://www.purl.org/rss/] This example uses L, which is not part of Krang - it would have to be installed separately. To generate the RSS file, we're going to override the C method for the Story element to generate the RSS file, and then continue the publish process, adding the RSS file to the final output. sub fill_template { my $self = shift; my %args = @_; my $rss = new XML::RSS; my $publisher = $args{publisher}; my $story = $publisher->story(); $rss->channel(title => $story->title(), link => 'http://' . $story->url(), description => $story->slug() ); foreach my $leadin ($story->linked_stories()) { $rss->add_item(title => $leadin->title(), link => 'http://' . $leadin->url()); } my $rss_output = $rss->as_string(); return $self->SUPER::fill_template(@_) . $publisher->additional_content_block(content => $rss_output, filename => 'rss.xml', use_category => 0); } As you can see, the regular publish work is still done by Krang, in the call at the bottom, C<< $self->SUPER::fill_template(@_) >>. That output is concatenated with the output generated by the XML::RSS module (after being tagged properly by C<< $publisher->additional_content_block() >>), and returned to the Publisher, which will then write the two files out. The C option to C<< Krang::Publisher->additional_content_block() >> tells the Publisher to not combine the output from XML::RSS with any template output from the categories (e.g. headers/footers). While this is a desireable feature sometimes, we don't want to mix HTML templates with XML output in this case. =head2 Example - Generating a Wall Page We want have two goals here - first, we want to build a three-page story out of this tree. Second, we want to create a wall page using the content of the first page, and a wall template. Again, the best angle of attack here will be to work from the Story element - we want to manipulate content depending on which page element we're on, and page elements don't know about eachother. Additionally, we need access to elements that won't be available to Page elements. =head3 fill_template() - Story Element In overriding the C method in the story element, we're going to maintain two separate hashes of parameters - one will be used for the regular story template, the other for the wall page template. sub fill_template { my $self = shift; my %args = @_; my %params; my %wall_params; my $wall_template = $self->_load_template(@_, filename => 'wall.tmpl', search_path => ['/foo/bar']); my $template = $args{tmpl}; my $element = $args{element}; my $publisher = $args{publisher}; my $story = $publisher->story(); $params{title} = $story->title(); $wall_params{title} = $story->title(); # retrieve the list of child elements my @children = $element->children(); foreach my $child (@children) { my $name = $child->name(); my $html = $child->publish(publisher => $publisher, fill_template_args => { greeting => 'Hello World!' }); unless (exists($params{$name})) { $params{$name} = $html; $wall_params{$name} = $html; } if ($name eq 'page') { push @{$params{"$name_loop"}}, { $name => $html }; unless (exists($wall_params{"$name_loop"})) { push @{$wall_params{"$name_loop"}}, { $name => $html }; } } } $template->param(%params); $wall_template->param(%wall_params); my $html = $template->output(); my $wall_html = $wall_template->output(); $html .= $publisher->additional_content_block(filename => 'wall.html', content => $wall_html, use_category => 1 ); return $html; } The end result is the original three-page story, along with a wall.html file. Note - using the approach of the first example, calling C<< $self->SUPER::fill_template() >> to generate the content for the story itself could have been done here as well, but it would have incurred additional overhead, as some elements would have been published multiple times (the first page and all its child elements), creating a performance penalty. While the penalty is negligible in this case, be careful. =head1 Changing How an Element Chooses a Template =over =item * find_template(publisher => $publisher, element => $element); Returns an HTML::Template::Expr object with the template to be used by the element. Follows a defined protocol for locating the template on the filesystem. =back The process by which Krang chooses a publish template for a given element is as follows: =over =item 1) Determine the template name - by default, C< element-name.tmpl >. =item 2) Given a category path of C, start by looking for the template in C. =item 3) If the template is found, attempt to load and parse it. If successful, return an instantiated L object. Otherwise, throw an error (C). =item 4) If the template is not found at C, try C, C, finally C. If the template is found at any point, load the template as seen in step 3. =item 5) If the template cannot be found, throw an error (C). =back You can make changes to this process by overriding C to change what template Krang will look for, where Krang will look for it, or even what kind of template will be loaded. =head2 Loading a Template Loading an L template requires two things - the template filename, and a list of directories to search for the template. sub find_template { my $self = shift; my %args = @_; my $tmpl; my $publisher = $args{publisher}; my $element = $args{element}; my @search_path = $publisher->template_search_path(); my $filename = $element->name() . '.tmpl'; my $template = $self->_load_template(publisher => $publisher, element => $element, filename => $filename, search_path => \@search_path); return $template; } By making changes to either C<$filename> or C<@search_path> (ordered from first to last in terms of directories to search), you can affect what template gets loaded. C<< $self->_load_template() >> handles the actual process of finding, loading, and throwing any required errors for L templates. If you want to change the type of template being loaded (e.g. you don't want to use L templates), you need to roll your own code to find, load and parse the templates, throwing appropriate errors as needed. Be aware that C and C are expecting an L template, so you will need to override C and C functionality as well. =head1 Changing the Publish Process for an Element =over =item * publish(story => $story, category => $category_id) Ties the C and C methods together. Returns publish output for the current element (and therefore, any children beneath it). See L for more info on publish(). =back C acts as the coordinator of the publish process for a given element, making sure that the entire process runs smoothly. It works as follows: =over =item 1) Find a template for the current element using C. =item 2) If no template is found, decide if this is a problem. By default, if an element has no children, C will simply return C<$element->template_data()>. If the element has children, however, it will propegate the C error thrown by C. =item 3) If the template is found, pass it to C. =item 4) Once C is finished, return $template->output(). =back =head2 Preventing an Element from Publishing While Krang, by default, does not call C on any element that is not explicitly included in a template, this may not offer enough protection for elements you don't want published. Overriding publish to return will make sure that an element (and all of its children) will never be published. sub publish { return; } =head2 Forcing Publish Without a Template On the other hand, if you know an element will never have a template (or want to make sure that noone goofs things up by creating a template for that element), you can simplify the publish process greatly (again, no children would get published): sub publish { my $self = shift; my %args = @_; my $element = $args{element}; return $element->template_data(); } =head1 Conclusion This covers the major aspects of customizing the Krang publish process. By overriding the three methods C, C and C, there's a lot that can be done to change how a story publishes itself. At this point, if you want to learn more about how the publish process works, and what can be done, read the POD and the code itself for L and L. Good luck!
__label__pos
0.99947
Friday, 9 Jun 2023 How To How to Find the Volume of a Sphere There are some basic ways to determine the volume of a sphere. To begin, you must find the area of the sphere. This can be done by finding the hemisphere or the fraction of the sphere. You can then use the volume of a sphere formula to calculate the volume. Calculate the surface area A sphere is a three-dimensional geometrical object that has no vertices, no faces and no edges. Spheres have a diameter and a height and a volume. Using these characteristics and a few other variables, you can calculate the surface area of a sphere. One of the most important factors in calculating the surface area of a sphere is a simple formula that tells you how much of the surface is occupied by curved surfaces. This formula can be found in any calculator and is considered the standard for determining the surface area of a solid. The same formula can be used to determine the volume of a sphere. The easiest way to determine the surface area of a sphere is to take the radius of the circle and multiply it by the surface area. This is an ancient mathematical concept and is still used today. For example, a candy ball has a r value of h, which is equal to the diameter of the ball. Another way to calculate the surface area of a sphere is by measuring its circumference. Since spheres are three-dimensional, a circumference is actually a distance between the center of the sphere and all points on the sphere’s surface. If you measure the radius of the sphere and the radius of a cylinder, you’ll find that the sphere has a smaller diameter than the cylinder. Read Also: How to Find the Surface Area of a Triangular Prism In addition to the curved surface area of a sphere, a hemisphere has a flat face. Using the hat-box theorem of Archimedes, you can calculate the total surface area of a sphere by multiplying the surface area of the sphere’s hemisphere by the area of the sphere’s hemisphere’s circular base. If you want to know the actual volume of a sphere, you can use the same formula as the sphere’s surface area, but you have to calculate the radius and the height of the cylinder. Once you have these two numbers, you can calculate the volume of a sphere by dividing the radius by the cylinder’s volume. It’s also possible to do this if the cylinder is a cube, but this method will require you to calculate the square root of both sides. Use the volume of a sphere formula If you have been asked to calculate the volume of a sphere, you have a few options. You can use a formula or integrate the solution. The equation to find the volume of a sphere is 4/3 p r3. This formula can be expressed in terms of pi, which is an integral number, and is also used in math. Many mathematicians will use this formula to get answers that are more theoretical. However, it is infinitely more accurate to use the symbol p. To calculate the volume of a sphere, the first step is to find the radius of the sphere. The radius of a sphere is equal to the diameter of the sphere. It is the distance from the center to any point on the surface of the sphere. Once you know the radius of the sphere, you can easily determine its surface area. For example, if you have a sphere with a diameter of 14 cm, its radius is 6 cm. Since its surface area is half of the diameter, its volume is 14 cm. Alternatively, you can use the formula for the volume of a hemisphere. The volume of a hemisphere is the same as the volume of a sphere. In the case of a hemisphere, the sphere’s volume is the volume of the sphere plus the volume of the cylinder that has the same radius. Usually, you will need to multiply the radius of the sphere with pi to get the surface area of the sphere. Therefore, you can use a formula such as V = (4/3)pr3. Calculate the volume of a hemisphere or other fractions of a sphere If you need to calculate the volume of a hemisphere or other fractions of a sphere, there are several mathematical techniques you can use to solve this problem. You can start by calculating the radius of the sphere, and then you can find the hemisphere’s volume using the formula: v = p r3. The volume of a hemisphere is measured in cubic units, and the most common are liters and cubic inches. Another way to calculate the volume of a hemisphere is to cut it in half. This is done by revolving a semicircle around the sphere’s diameter edge. When this is done, two lines cross the center of the semicircle, and the area of the semicircle is divided into two equal parts. One line is perpendicular to the diameter edge, and the other is parallel to it. Another way to calculate the volume of an object is to use the Pappus centroid theorem. Pappus’s centroid theorem states that the volume of a solid object is the product of its area and the distance the region is revolving when it is turned. In addition to the Pappus centroid theorem, you can also use the hemisphere volume equation: V = 2pr3/3. The hemisphere’s volume is the amount of cubic units that are able to fit into a sphere of a certain size. Using the hemisphere’s volume, you can then calculate the total volume of a sphere. It is important to remember that you should not round the volume until you have finished the calculation. By doing this, you can ensure that the volume is accurate to two decimal places.
__label__pos
0.995295
Blog Infos Author Published Topics , , , Published Unlock the potential of these advanced techniques to elevate your smartwatch app design and functionality to new heights. If you’re developing an application for WearOs you’ve surely come across the Scaffold showing the TimeText(): WearOs Time in scaffold What if you wanted to put custom text instead of time, and maybe not at the top but with a specific starting angle? Well, then in that case the guide is made just for you! and is made in Jetpack Compose! What you will archive today   To achieve the effect you see above, thus being able to create your custom curved text at the position of your choice, in Jetpack Compose, you can take advantage of the CurvedLayout with the curvedText. 1. As first make sure you have the dependencies in your module: implementation("androidx.wear.compose:compose-material:1.2.1") implementation("androidx.wear.compose:compose-foundation:1.2.1") 2. Then, create the composable function that allows you to write the text : @Composable private fun MyCoolCurvedText(anchor: Float, color: Color, text: String) { CurvedLayout( anchor = anchor, anchorType = AnchorType.Center, modifier = Modifier.fillMaxSize(), ) { curvedRow( ) { curvedText( text = text, style = CurvedTextStyle( fontSize = 14.sp, color = color ) ) } } } //required imports: //import androidx.wear.compose.foundation.AnchorType //import androidx.wear.compose.foundation.CurvedLayout //import androidx.wear.compose.foundation.CurvedTextStyle //import androidx.wear.compose.foundation.curvedRow //import androidx.wear.compose.material.curvedText Of course, you can customize it to your liking, and put in more items according to your needs 3. Use it inside your UI class MainActivity : ComponentActivity() { override fun onCreate(savedInstanceState: Bundle?) { super.onCreate(savedInstanceState) setContent { Box(Modifier.fillMaxSize()) { MyCoolCurvedText( anchor = 0f, color = Color.Red, text = "Rounded text!" ) MyCoolCurvedText( anchor = 90f, color = Color.Green, text = "WearOs today!" ) MyCoolCurvedText( anchor = 180f, color = Color.Blue, text = "I'm rounded!" ) MyCoolCurvedText( anchor = 270f, color = Color.Yellow, text = "Hello devs!" ) } } } } And there you have it! If you are interested in more WearOs Jetpack Compose Android tips let me know in the comments! My advice is to always follow the official guidelines and clean code writing guidelines. If you like my article, please don’t forget to click 👏👏👏 to recommend it to others 👏👏👏. Feel free to ask questions, make comments, or propose a better solution. Don’t forget to follow me on Twitter and GitHub! This article was previously posted on proaandroiddev.com OUR VIDEO RECOMMENDATION Jobs No results found. YOU MAY BE INTERESTED IN YOU MAY BE INTERESTED IN blog It’s one of the common UX across apps to provide swipe to dismiss so… READ MORE blog In this part of our series on introducing Jetpack Compose into an existing project,… READ MORE blog In the world of Jetpack Compose, where designing reusable and customizable UI components is… READ MORE blog Hi, today I come to you with a quick tip on how to update… READ MORE Leave a Reply Your email address will not be published. Required fields are marked * Fill out this field Fill out this field Please enter a valid email address. Menu
__label__pos
0.972025
How to increase sql performance?, DOT NET Programming How can you increase SQL performance? 1)Keep your indexes as narrow as possible. This reduces the size of the index and decrease the number of reads needed to read the index. 2)Try to create indexes on the columns that have integer values rather than the character values. 3)If you create a composite (multi-column) index, then the order of the columns in the key are much important. Try to order the columns in the key as to enhance the selectivity, with the most selective columns to the leftmost of the key. 4)If you want to join several tables, try to generate surrogate integer keys for this purpose and create indexes on their columns. 5)Create the surrogate integer primary key (identity for example) if your table will not have many insert operations. 6)The Clustered indexes are more preferable than nonclustered, if you require to select by a wide range of values or you need to sort results set with GROUP BY or ORDER BY. 7)If your application will be performing the same query over and over again on the similar table, consider creating a covering index on the table. 8)You can use the SQL Server Profiler Create Trace Wizard with "Identify the Scans of the Large Tables" trace to determine which tables in your database may need indexes. This trace will show which tables are being scanned by queries in spite of using an index. Posted Date: 9/24/2012 4:29:13 AM | Location : United States Related Discussions:- How to increase sql performance?, Assignment Help, Ask Question on How to increase sql performance?, Get Answer, Expert's Help, How to increase sql performance? Discussions Write discussion on How to increase sql performance? Your posts are moderated Related Questions What is DMAIC and DMADV? The Six Sigma has two key methodologies viz. DMAIC and DMADV. The DMAIC is used to improve an existing business process. While DMADV is used to create How to Enable ASP.NET polling? All our database side is configured in order to get the SQL Cache working in the ASP.NET side we require to do some configuration in the web.conf What is Extensible Markup Language (XML). XML is an easy and flexible mark-up language in the text format. Nowadays, it is widely used to exchange a large variety of data over I need to Develop a real state website We want a programmer who can create a website using html5, css3, jQuery, php, MySQL. All connected to a CRM to accomplish all the informat How can we identify that the Page is PostBack? The Page object has the "IsPostBack" property which can be checked to know that the page is posted back. Create a Calculator application using Web Form with the following functions: 1.  Add 2.  Minus 3.  Multiply 4.  Divide Each of the above has to be a Web Service Question: (a) Explain the Postback process as use with Web Forms. (b) Write extract codes to demonstrate how to store and retrieve values in: (i) a Session Object (ii) I need a project which shows Paypal integration with ASP.net website We have events website written in C# and ASP.net. We want to integrate Paypal using API and have a full inte What is the difference between "DataSet" and "DataReader"? The major differences between "DataSet" and "DataReader" are as follows:- 1)The "DataSet" is a disconnected archit What does Address Of operator do? The Address Of operator generates a delegate object for the BackgroundProcess method. A delegate within VB.NET is a object-oriented, type-safe
__label__pos
0.849579
multi_physics/biot_npbc_lagrange.py Description Biot problem - deformable porous medium with the no-penetration boundary condition on a boundary region enforced using Lagrange multipliers. The non-penetration condition is enforced weakly using the Lagrange multiplier \lambda. There is also a rigid body movement constraint imposed on the \Gamma_{outlet} region using the linear combination boundary conditions. Find \ul{u}, p and \lambda such that: \int_{\Omega} D_{ijkl}\ e_{ij}(\ul{v}) e_{kl}(\ul{u}) - \int_{\Omega} p\ \alpha_{ij} e_{ij}(\ul{v}) + \int_{\Gamma_{walls}} \lambda \ul{n} \cdot \ul{v} = 0 \;, \quad \forall \ul{v} \;, \int_{\Omega} q\ \alpha_{ij} e_{ij}(\ul{u}) + \int_{\Omega} K_{ij} \nabla_i q \nabla_j p = 0 \;, \quad \forall q \;, \int_{\Gamma_{walls}} \hat\lambda \ul{n} \cdot \ul{u} = 0 \;, \quad \forall \hat\lambda \;, \ul{u} \cdot \ul{n} = 0 \mbox{ on } \Gamma_{walls} \;, where D_{ijkl} = \mu (\delta_{ik} \delta_{jl}+\delta_{il} \delta_{jk}) + \lambda \ \delta_{ij} \delta_{kl} \;. ../../_images/multi_physics-biot_npbc_lagrange.png source code r""" Biot problem - deformable porous medium with the no-penetration boundary condition on a boundary region enforced using Lagrange multipliers. The non-penetration condition is enforced weakly using the Lagrange multiplier :math:`\lambda`. There is also a rigid body movement constraint imposed on the :math:`\Gamma_{outlet}` region using the linear combination boundary conditions. Find :math:`\ul{u}`, :math:`p` and :math:`\lambda` such that: .. math:: \int_{\Omega} D_{ijkl}\ e_{ij}(\ul{v}) e_{kl}(\ul{u}) - \int_{\Omega} p\ \alpha_{ij} e_{ij}(\ul{v}) + \int_{\Gamma_{walls}} \lambda \ul{n} \cdot \ul{v} = 0 \;, \quad \forall \ul{v} \;, \int_{\Omega} q\ \alpha_{ij} e_{ij}(\ul{u}) + \int_{\Omega} K_{ij} \nabla_i q \nabla_j p = 0 \;, \quad \forall q \;, \int_{\Gamma_{walls}} \hat\lambda \ul{n} \cdot \ul{u} = 0 \;, \quad \forall \hat\lambda \;, \ul{u} \cdot \ul{n} = 0 \mbox{ on } \Gamma_{walls} \;, where .. math:: D_{ijkl} = \mu (\delta_{ik} \delta_{jl}+\delta_{il} \delta_{jk}) + \lambda \ \delta_{ij} \delta_{kl} \;. """ from __future__ import absolute_import from examples.multi_physics.biot_npbc import (cinc_simple, define_regions, get_pars) def define(): from sfepy import data_dir filename = data_dir + '/meshes/3d/cylinder.mesh' output_dir = 'output' return define_input(filename, output_dir) def post_process(out, pb, state, extend=False): from sfepy.base.base import Struct dvel = pb.evaluate('ev_diffusion_velocity.2.Omega( m.K, p )', mode='el_avg') out['dvel'] = Struct(name='output_data', var_name='p', mode='cell', data=dvel, dofs=None) stress = pb.evaluate('ev_cauchy_stress.2.Omega( m.D, u )', mode='el_avg') out['cauchy_stress'] = Struct(name='output_data', var_name='u', mode='cell', data=stress, dofs=None) return out def define_input(filename, output_dir): filename_mesh = filename options = { 'output_dir' : output_dir, 'output_format' : 'vtk', 'post_process_hook' : 'post_process', ## 'file_per_var' : True, 'ls' : 'ls', 'nls' : 'newton', } functions = { 'cinc_simple0' : (lambda coors, domain: cinc_simple(coors, 0),), 'cinc_simple1' : (lambda coors, domain: cinc_simple(coors, 1),), 'cinc_simple2' : (lambda coors, domain: cinc_simple(coors, 2),), 'get_pars' : (lambda ts, coors, mode=None, **kwargs: get_pars(ts, coors, mode, output_dir=output_dir, **kwargs),), } regions, dim = define_regions(filename_mesh) fields = { 'displacement': ('real', 'vector', 'Omega', 1), 'pressure': ('real', 'scalar', 'Omega', 1), 'multiplier': ('real', 'scalar', 'Walls', 1), } variables = { 'u' : ('unknown field', 'displacement', 0), 'v' : ('test field', 'displacement', 'u'), 'p' : ('unknown field', 'pressure', 1), 'q' : ('test field', 'pressure', 'p'), 'ul' : ('unknown field', 'multiplier', 2), 'vl' : ('test field', 'multiplier', 'ul'), } ebcs = { 'inlet' : ('Inlet', {'p.0' : 1.0, 'u.all' : 0.0}), 'outlet' : ('Outlet', {'p.0' : -1.0}), } lcbcs = { 'rigid' : ('Outlet', {'u.all' : None}, None, 'rigid'), } materials = { 'm' : 'get_pars', } equations = { 'eq_1' : """dw_lin_elastic.2.Omega( m.D, v, u ) - dw_biot.2.Omega( m.alpha, v, p ) + dw_non_penetration.2.Walls( v, ul ) = 0""", 'eq_2' : """dw_biot.2.Omega( m.alpha, u, q ) + dw_diffusion.2.Omega( m.K, q, p ) = 0""", 'eq_3' : """dw_non_penetration.2.Walls( u, vl ) = 0""", } solvers = { 'ls' : ('ls.scipy_direct', {}), 'newton' : ('nls.newton', {}), } return locals()
__label__pos
0.999972
llvm.org GIT mirror llvm / cb92739 [X86][TableGen] Recommitting the X86 memory folding tables TableGen backend while disabling it by default. After the original commit ([[ https://reviews.llvm.org/rL304088 | rL304088 ]]) was reverted, a discussion in llvm-dev was opened on 'how to accomplish this task'. In the discussion we concluded that the best way to achieve our goal (which is to automate the folding tables and remove the manually maintained tables) is: # Commit the tablegen backend disabled by default. # Proceed with an incremental updating of the manual tables - while checking the validity of each added entry. # Repeat previous step until we reach a state where the generated and the manual tables are identical. Then we can safely remove the manual tables and include the generated tables instead. # Schedule periodical (1 week/2 weeks/1 month) runs of the pass: - if changes appear (new entries): - make sure the entries are legal - If they are not, mark them as illegal to folding - Commit the changes (if there are any). CMake flag added for this purpose is "X86_GEN_FOLD_TABLES". Building with this flags will run the pass and emit the X86GenFoldTables.inc file under build/lib/Target/X86/ directory which is a good reference for any developer who wants to take part in the effort of completing the current folding tables. Differential Revision: https://reviews.llvm.org/D38028 git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@315173 91177308-0d34-0410-b5e6-96231b3b80d8 Ayman Musa 3 years ago 5 changed file(s) with 675 addition(s) and 0 deletion(s). Raw diff Collapse all Expand all 1212 tablegen(LLVM X86GenEVEX2VEXTables.inc -gen-x86-EVEX2VEX-tables) 1313 tablegen(LLVM X86GenRegisterBank.inc -gen-register-bank) 1414 tablegen(LLVM X86GenGlobalISel.inc -gen-global-isel) 15 16 if (X86_GEN_FOLD_TABLES) 17 tablegen(LLVM X86GenFoldTables.inc -gen-x86-fold-tables) 18 endif() 1519 1620 add_public_tablegen_target(X86CommonTableGen) 1721 3737 Types.cpp 3838 X86DisassemblerTables.cpp 3939 X86EVEX2VEXTablesEmitter.cpp 40 X86FoldTablesEmitter.cpp 4041 X86ModRMFilters.cpp 4142 X86RecognizableInstr.cpp 4243 CTagsEmitter.cpp 4646 GenSearchableTables, 4747 GenGlobalISel, 4848 GenX86EVEX2VEXTables, 49 GenX86FoldTables, 4950 GenRegisterBank, 5051 }; 5152 9899 "Generate GlobalISel selector"), 99100 clEnumValN(GenX86EVEX2VEXTables, "gen-x86-EVEX2VEX-tables", 100101 "Generate X86 EVEX to VEX compress tables"), 102 clEnumValN(GenX86FoldTables, "gen-x86-fold-tables", 103 "Generate X86 fold tables"), 101104 clEnumValN(GenRegisterBank, "gen-register-bank", 102105 "Generate registers bank descriptions"))); 103106 195198 case GenX86EVEX2VEXTables: 196199 EmitX86EVEX2VEXTables(Records, OS); 197200 break; 201 case GenX86FoldTables: 202 EmitX86FoldTables(Records, OS); 203 break; 198204 } 199205 200206 return false; 8181 void EmitSearchableTables(RecordKeeper &RK, raw_ostream &OS); 8282 void EmitGlobalISel(RecordKeeper &RK, raw_ostream &OS); 8383 void EmitX86EVEX2VEXTables(RecordKeeper &RK, raw_ostream &OS); 84 void EmitX86FoldTables(RecordKeeper &RK, raw_ostream &OS); 8485 void EmitRegisterBank(RecordKeeper &RK, raw_ostream &OS); 8586 8687 } // End llvm namespace 0 //===- utils/TableGen/X86FoldTablesEmitter.cpp - X86 backend-*- C++ -*-===// 1 // 2 // The LLVM Compiler Infrastructure 3 // 4 // This file is distributed under the University of Illinois Open Source 5 // License. See LICENSE.TXT for details. 6 // 7 //===----------------------------------------------------------------------===// 8 // 9 // This tablegen backend is responsible for emitting the memory fold tables of 10 // the X86 backend instructions. 11 // 12 //===----------------------------------------------------------------------===// 13 14 #include "CodeGenDAGPatterns.h" 15 #include "CodeGenTarget.h" 16 #include "X86RecognizableInstr.h" 17 #include "llvm/TableGen/Error.h" 18 #include "llvm/TableGen/TableGenBackend.h" 19 20 using namespace llvm; 21 22 namespace { 23 24 // 3 possible strategies for the unfolding flag (TB_NO_REVERSE) of the 25 // manual added entries. 26 enum UnfoldStrategy { 27 UNFOLD, // Allow unfolding 28 NO_UNFOLD, // Prevent unfolding 29 NO_STRATEGY // Make decision according to operands' sizes 30 }; 31 32 // Represents an entry in the manual mapped instructions set. 33 struct ManualMapEntry { 34 const char *RegInstStr; 35 const char *MemInstStr; 36 UnfoldStrategy Strategy; 37 38 ManualMapEntry(const char *RegInstStr, const char *MemInstStr, 39 UnfoldStrategy Strategy = NO_STRATEGY) 40 : RegInstStr(RegInstStr), MemInstStr(MemInstStr), Strategy(Strategy) {} 41 }; 42 43 class IsMatch; 44 45 // List of instructions requiring explicitly aligned memory. 46 const char *ExplicitAlign[] = {"MOVDQA", "MOVAPS", "MOVAPD", "MOVNTPS", 47 "MOVNTPD", "MOVNTDQ", "MOVNTDQA"}; 48 49 // List of instructions NOT requiring explicit memory alignment. 50 const char *ExplicitUnalign[] = {"MOVDQU", "MOVUPS", "MOVUPD"}; 51 52 // For manually mapping instructions that do not match by their encoding. 53 const ManualMapEntry ManualMapSet[] = { 54 { "ADD16ri_DB", "ADD16mi", NO_UNFOLD }, 55 { "ADD16ri8_DB", "ADD16mi8", NO_UNFOLD }, 56 { "ADD16rr_DB", "ADD16mr", NO_UNFOLD }, 57 { "ADD32ri_DB", "ADD32mi", NO_UNFOLD }, 58 { "ADD32ri8_DB", "ADD32mi8", NO_UNFOLD }, 59 { "ADD32rr_DB", "ADD32mr", NO_UNFOLD }, 60 { "ADD64ri32_DB", "ADD64mi32", NO_UNFOLD }, 61 { "ADD64ri8_DB", "ADD64mi8", NO_UNFOLD }, 62 { "ADD64rr_DB", "ADD64mr", NO_UNFOLD }, 63 { "ADD16rr_DB", "ADD16rm", NO_UNFOLD }, 64 { "ADD32rr_DB", "ADD32rm", NO_UNFOLD }, 65 { "ADD64rr_DB", "ADD64rm", NO_UNFOLD }, 66 { "PUSH16r", "PUSH16rmm", NO_UNFOLD }, 67 { "PUSH32r", "PUSH32rmm", NO_UNFOLD }, 68 { "PUSH64r", "PUSH64rmm", NO_UNFOLD }, 69 { "TAILJMPr", "TAILJMPm", UNFOLD }, 70 { "TAILJMPr64", "TAILJMPm64", UNFOLD }, 71 { "TAILJMPr64_REX", "TAILJMPm64_REX", UNFOLD }, 72 }; 73 74 75 static bool isExplicitAlign(const CodeGenInstruction *Inst) { 76 return any_of(ExplicitAlign, [Inst](const char *InstStr) { 77 return Inst->TheDef->getName().find(InstStr) != StringRef::npos; 78 }); 79 } 80 81 static bool isExplicitUnalign(const CodeGenInstruction *Inst) { 82 return any_of(ExplicitUnalign, [Inst](const char *InstStr) { 83 return Inst->TheDef->getName().find(InstStr) != StringRef::npos; 84 }); 85 } 86 87 class X86FoldTablesEmitter { 88 RecordKeeper &Records; 89 CodeGenTarget Target; 90 91 // Represents an entry in the folding table 92 class X86FoldTableEntry { 93 const CodeGenInstruction *RegInst; 94 const CodeGenInstruction *MemInst; 95 96 public: 97 bool CannotUnfold = false; 98 bool IsLoad = false; 99 bool IsStore = false; 100 bool IsAligned = false; 101 unsigned int Alignment = 0; 102 103 X86FoldTableEntry(const CodeGenInstruction *RegInst, 104 const CodeGenInstruction *MemInst) 105 : RegInst(RegInst), MemInst(MemInst) {} 106 107 friend raw_ostream &operator<<(raw_ostream &OS, 108 const X86FoldTableEntry &E) { 109 OS << "{ X86::" << E.RegInst->TheDef->getName().str() 110 << ", X86::" << E.MemInst->TheDef->getName().str() << ", "; 111 112 if (E.IsLoad) 113 OS << "TB_FOLDED_LOAD | "; 114 if (E.IsStore) 115 OS << "TB_FOLDED_STORE | "; 116 if (E.CannotUnfold) 117 OS << "TB_NO_REVERSE | "; 118 if (E.IsAligned) 119 OS << "TB_ALIGN_" + std::to_string(E.Alignment) + " | "; 120 121 OS << "0 },\n"; 122 123 return OS; 124 } 125 }; 126 127 typedef std::vector FoldTable; 128 // std::vector for each folding table. 129 // Table2Addr - Holds instructions which their memory form performs load+store 130 // Table#i - Holds instructions which the their memory form perform a load OR 131 // a store, and their #i'th operand is folded. 132 FoldTable Table2Addr; 133 FoldTable Table0; 134 FoldTable Table1; 135 FoldTable Table2; 136 FoldTable Table3; 137 FoldTable Table4; 138 139 public: 140 X86FoldTablesEmitter(RecordKeeper &R) : Records(R), Target(R) {} 141 142 // run - Generate the 6 X86 memory fold tables. 143 void run(raw_ostream &OS); 144 145 private: 146 // Decides to which table to add the entry with the given instructions. 147 // S sets the strategy of adding the TB_NO_REVERSE flag. 148 void updateTables(const CodeGenInstruction *RegInstr, 149 const CodeGenInstruction *MemInstr, 150 const UnfoldStrategy S = NO_STRATEGY); 151 152 // Generates X86FoldTableEntry with the given instructions and fill it with 153 // the appropriate flags - then adds it to Table. 154 void addEntryWithFlags(FoldTable &Table, const CodeGenInstruction *RegInstr, 155 const CodeGenInstruction *MemInstr, 156 const UnfoldStrategy S, const unsigned int FoldedInd); 157 158 // Print the given table as a static const C++ array of type 159 // X86MemoryFoldTableEntry. 160 void printTable(const FoldTable &Table, std::string TableName, 161 raw_ostream &OS) { 162 OS << "static const X86MemoryFoldTableEntry MemoryFold" << TableName 163 << "[] = {\n"; 164 165 for (const X86FoldTableEntry &E : Table) 166 OS << E; 167 168 OS << "};\n"; 169 } 170 }; 171 172 // Return true if one of the instruction's operands is a RST register class 173 static bool hasRSTRegClass(const CodeGenInstruction *Inst) { 174 return any_of(Inst->Operands, [](const CGIOperandList::OperandInfo &OpIn) { 175 return OpIn.Rec->getName() == "RST"; 176 }); 177 } 178 179 // Return true if one of the instruction's operands is a ptr_rc_tailcall 180 static bool hasPtrTailcallRegClass(const CodeGenInstruction *Inst) { 181 return any_of(Inst->Operands, [](const CGIOperandList::OperandInfo &OpIn) { 182 return OpIn.Rec->getName() == "ptr_rc_tailcall"; 183 }); 184 } 185 186 // Calculates the integer value representing the BitsInit object 187 static inline uint64_t getValueFromBitsInit(const BitsInit *B) { 188 assert(B->getNumBits() <= sizeof(uint64_t) * 8 && "BitInits' too long!"); 189 190 uint64_t Value = 0; 191 for (unsigned i = 0, e = B->getNumBits(); i != e; ++i) { 192 BitInit *Bit = cast(B->getBit(i)); 193 Value |= uint64_t(Bit->getValue()) << i; 194 } 195 return Value; 196 } 197 198 // Returns true if the two given BitsInits represent the same integer value 199 static inline bool equalBitsInits(const BitsInit *B1, const BitsInit *B2) { 200 if (B1->getNumBits() != B2->getNumBits()) 201 PrintFatalError("Comparing two BitsInits with different sizes!"); 202 203 for (unsigned i = 0, e = B1->getNumBits(); i != e; ++i) { 204 BitInit *Bit1 = cast(B1->getBit(i)); 205 BitInit *Bit2 = cast(B2->getBit(i)); 206 if (Bit1->getValue() != Bit2->getValue()) 207 return false; 208 } 209 return true; 210 } 211 212 // Return the size of the register operand 213 static inline unsigned int getRegOperandSize(const Record *RegRec) { 214 if (RegRec->isSubClassOf("RegisterOperand")) 215 RegRec = RegRec->getValueAsDef("RegClass"); 216 if (RegRec->isSubClassOf("RegisterClass")) 217 return RegRec->getValueAsListOfDefs("RegTypes")[0]->getValueAsInt("Size"); 218 219 llvm_unreachable("Register operand's size not known!"); 220 } 221 222 // Return the size of the memory operand 223 static inline unsigned int 224 getMemOperandSize(const Record *MemRec, const bool IntrinsicSensitive = false) { 225 if (MemRec->isSubClassOf("Operand")) { 226 // Intrinsic memory instructions use ssmem/sdmem. 227 if (IntrinsicSensitive && 228 (MemRec->getName() == "sdmem" || MemRec->getName() == "ssmem")) 229 return 128; 230 231 StringRef Name = 232 MemRec->getValueAsDef("ParserMatchClass")->getValueAsString("Name"); 233 if (Name == "Mem8") 234 return 8; 235 if (Name == "Mem16") 236 return 16; 237 if (Name == "Mem32") 238 return 32; 239 if (Name == "Mem64") 240 return 64; 241 if (Name == "Mem80") 242 return 80; 243 if (Name == "Mem128") 244 return 128; 245 if (Name == "Mem256") 246 return 256; 247 if (Name == "Mem512") 248 return 512; 249 } 250 251 llvm_unreachable("Memory operand's size not known!"); 252 } 253 254 // Returns true if the record's list of defs includes the given def. 255 static inline bool hasDefInList(const Record *Rec, const StringRef List, 256 const StringRef Def) { 257 if (!Rec->isValueUnset(List)) { 258 return any_of(*(Rec->getValueAsListInit(List)), 259 [Def](const Init *I) { return I->getAsString() == Def; }); 260 } 261 return false; 262 } 263 264 // Return true if the instruction defined as a register flavor. 265 static inline bool hasRegisterFormat(const Record *Inst) { 266 const BitsInit *FormBits = Inst->getValueAsBitsInit("FormBits"); 267 uint64_t FormBitsNum = getValueFromBitsInit(FormBits); 268 269 // Values from X86Local namespace defined in X86RecognizableInstr.cpp 270 return FormBitsNum >= X86Local::MRMDestReg && FormBitsNum <= X86Local::MRM7r; 271 } 272 273 // Return true if the instruction defined as a memory flavor. 274 static inline bool hasMemoryFormat(const Record *Inst) { 275 const BitsInit *FormBits = Inst->getValueAsBitsInit("FormBits"); 276 uint64_t FormBitsNum = getValueFromBitsInit(FormBits); 277 278 // Values from X86Local namespace defined in X86RecognizableInstr.cpp 279 return FormBitsNum >= X86Local::MRMDestMem && FormBitsNum <= X86Local::MRM7m; 280 } 281 282 static inline bool isNOREXRegClass(const Record *Op) { 283 return Op->getName().find("_NOREX") != StringRef::npos; 284 } 285 286 static inline bool isRegisterOperand(const Record *Rec) { 287 return Rec->isSubClassOf("RegisterClass") || 288 Rec->isSubClassOf("RegisterOperand") || 289 Rec->isSubClassOf("PointerLikeRegClass"); 290 } 291 292 static inline bool isMemoryOperand(const Record *Rec) { 293 return Rec->isSubClassOf("Operand") && 294 Rec->getValueAsString("OperandType") == "OPERAND_MEMORY"; 295 } 296 297 static inline bool isImmediateOperand(const Record *Rec) { 298 return Rec->isSubClassOf("Operand") && 299 Rec->getValueAsString("OperandType") == "OPERAND_IMMEDIATE"; 300 } 301 302 // Get the alternative instruction pointed by "FoldGenRegForm" field. 303 static inline const CodeGenInstruction * 304 getAltRegInst(const CodeGenInstruction *I, const RecordKeeper &Records, 305 const CodeGenTarget &Target) { 306 307 StringRef AltRegInstStr = I->TheDef->getValueAsString("FoldGenRegForm"); 308 Record *AltRegInstRec = Records.getDef(AltRegInstStr); 309 assert(AltRegInstRec && 310 "Alternative register form instruction def not found"); 311 CodeGenInstruction &AltRegInst = Target.getInstruction(AltRegInstRec); 312 return &AltRegInst; 313 } 314 315 // Function object - Operator() returns true if the given VEX instruction 316 // matches the EVEX instruction of this object. 317 class IsMatch { 318 const CodeGenInstruction *MemInst; 319 const RecordKeeper &Records; 320 321 public: 322 IsMatch(const CodeGenInstruction *Inst, const RecordKeeper &Records) 323 : MemInst(Inst), Records(Records) {} 324 325 bool operator()(const CodeGenInstruction *RegInst) { 326 Record *MemRec = MemInst->TheDef; 327 Record *RegRec = RegInst->TheDef; 328 329 // Return false if one (at least) of the encoding fields of both 330 // instructions do not match. 331 if (RegRec->getValueAsDef("OpEnc") != MemRec->getValueAsDef("OpEnc") || 332 !equalBitsInits(RegRec->getValueAsBitsInit("Opcode"), 333 MemRec->getValueAsBitsInit("Opcode")) || 334 // VEX/EVEX fields 335 RegRec->getValueAsDef("OpPrefix") != 336 MemRec->getValueAsDef("OpPrefix") || 337 RegRec->getValueAsDef("OpMap") != MemRec->getValueAsDef("OpMap") || 338 RegRec->getValueAsDef("OpSize") != MemRec->getValueAsDef("OpSize") || 339 RegRec->getValueAsBit("hasVEX_4V") != 340 MemRec->getValueAsBit("hasVEX_4V") || 341 RegRec->getValueAsBit("hasEVEX_K") != 342 MemRec->getValueAsBit("hasEVEX_K") || 343 RegRec->getValueAsBit("hasEVEX_Z") != 344 MemRec->getValueAsBit("hasEVEX_Z") || 345 RegRec->getValueAsBit("hasEVEX_B") != 346 MemRec->getValueAsBit("hasEVEX_B") || 347 RegRec->getValueAsBit("hasEVEX_RC") != 348 MemRec->getValueAsBit("hasEVEX_RC") || 349 RegRec->getValueAsBit("hasREX_WPrefix") != 350 MemRec->getValueAsBit("hasREX_WPrefix") || 351 RegRec->getValueAsBit("hasLockPrefix") != 352 MemRec->getValueAsBit("hasLockPrefix") || 353 !equalBitsInits(RegRec->getValueAsBitsInit("EVEX_LL"), 354 MemRec->getValueAsBitsInit("EVEX_LL")) || 355 !equalBitsInits(RegRec->getValueAsBitsInit("VEX_WPrefix"), 356 MemRec->getValueAsBitsInit("VEX_WPrefix")) || 357 // Instruction's format - The register form's "Form" field should be 358 // the opposite of the memory form's "Form" field. 359 !areOppositeForms(RegRec->getValueAsBitsInit("FormBits"), 360 MemRec->getValueAsBitsInit("FormBits")) || 361 RegRec->getValueAsBit("isAsmParserOnly") != 362 MemRec->getValueAsBit("isAsmParserOnly")) 363 return false; 364 365 // Make sure the sizes of the operands of both instructions suit each other. 366 // This is needed for instructions with intrinsic version (_Int). 367 // Where the only difference is the size of the operands. 368 // For example: VUCOMISDZrm and Int_VUCOMISDrm 369 // Also for instructions that their EVEX version was upgraded to work with 370 // k-registers. For example VPCMPEQBrm (xmm output register) and 371 // VPCMPEQBZ128rm (k register output register). 372 bool ArgFolded = false; 373 unsigned MemOutSize = MemRec->getValueAsDag("OutOperandList")->getNumArgs(); 374 unsigned RegOutSize = RegRec->getValueAsDag("OutOperandList")->getNumArgs(); 375 unsigned MemInSize = MemRec->getValueAsDag("InOperandList")->getNumArgs(); 376 unsigned RegInSize = RegRec->getValueAsDag("InOperandList")->getNumArgs(); 377 378 // Instructions with one output in their memory form use the memory folded 379 // operand as source and destination (Read-Modify-Write). 380 unsigned RegStartIdx = 381 (MemOutSize + 1 == RegOutSize) && (MemInSize == RegInSize) ? 1 : 0; 382 383 for (unsigned i = 0, e = MemInst->Operands.size(); i < e; i++) { 384 Record *MemOpRec = MemInst->Operands[i].Rec; 385 Record *RegOpRec = RegInst->Operands[i + RegStartIdx].Rec; 386 387 if (MemOpRec == RegOpRec) 388 continue; 389 390 if (isRegisterOperand(MemOpRec) && isRegisterOperand(RegOpRec)) { 391 if (getRegOperandSize(MemOpRec) != getRegOperandSize(RegOpRec) || 392 isNOREXRegClass(MemOpRec) != isNOREXRegClass(RegOpRec)) 393 return false; 394 } else if (isMemoryOperand(MemOpRec) && isMemoryOperand(RegOpRec)) { 395 if (getMemOperandSize(MemOpRec) != getMemOperandSize(RegOpRec)) 396 return false; 397 } else if (isImmediateOperand(MemOpRec) && isImmediateOperand(RegOpRec)) { 398 if (MemOpRec->getValueAsDef("Type") != RegOpRec->getValueAsDef("Type")) 399 return false; 400 } else { 401 // Only one operand can be folded. 402 if (ArgFolded) 403 return false; 404 405 assert(isRegisterOperand(RegOpRec) && isMemoryOperand(MemOpRec)); 406 ArgFolded = true; 407 } 408 } 409 410 return true; 411 } 412 413 private: 414 // Return true of the 2 given forms are the opposite of each other. 415 bool areOppositeForms(const BitsInit *RegFormBits, 416 const BitsInit *MemFormBits) { 417 uint64_t MemFormNum = getValueFromBitsInit(MemFormBits); 418 uint64_t RegFormNum = getValueFromBitsInit(RegFormBits); 419 420 if ((MemFormNum == X86Local::MRM0m && RegFormNum == X86Local::MRM0r) || 421 (MemFormNum == X86Local::MRM1m && RegFormNum == X86Local::MRM1r) || 422 (MemFormNum == X86Local::MRM2m && RegFormNum == X86Local::MRM2r) || 423 (MemFormNum == X86Local::MRM3m && RegFormNum == X86Local::MRM3r) || 424 (MemFormNum == X86Local::MRM4m && RegFormNum == X86Local::MRM4r) || 425 (MemFormNum == X86Local::MRM5m && RegFormNum == X86Local::MRM5r) || 426 (MemFormNum == X86Local::MRM6m && RegFormNum == X86Local::MRM6r) || 427 (MemFormNum == X86Local::MRM7m && RegFormNum == X86Local::MRM7r) || 428 (MemFormNum == X86Local::MRMXm && RegFormNum == X86Local::MRMXr) || 429 (MemFormNum == X86Local::MRMDestMem && 430 RegFormNum == X86Local::MRMDestReg) || 431 (MemFormNum == X86Local::MRMSrcMem && 432 RegFormNum == X86Local::MRMSrcReg) || 433 (MemFormNum == X86Local::MRMSrcMem4VOp3 && 434 RegFormNum == X86Local::MRMSrcReg4VOp3) || 435 (MemFormNum == X86Local::MRMSrcMemOp4 && 436 RegFormNum == X86Local::MRMSrcRegOp4)) 437 return true; 438 439 return false; 440 } 441 }; 442 443 } // end anonymous namespace 444 445 void X86FoldTablesEmitter::addEntryWithFlags(FoldTable &Table, 446 const CodeGenInstruction *RegInstr, 447 const CodeGenInstruction *MemInstr, 448 const UnfoldStrategy S, 449 const unsigned int FoldedInd) { 450 451 X86FoldTableEntry Result = X86FoldTableEntry(RegInstr, MemInstr); 452 Record *RegRec = RegInstr->TheDef; 453 Record *MemRec = MemInstr->TheDef; 454 455 // Only table0 entries should explicitly specify a load or store flag. 456 if (&Table == &Table0) { 457 unsigned MemInOpsNum = MemRec->getValueAsDag("InOperandList")->getNumArgs(); 458 unsigned RegInOpsNum = RegRec->getValueAsDag("InOperandList")->getNumArgs(); 459 // If the instruction writes to the folded operand, it will appear as an 460 // output in the register form instruction and as an input in the memory 461 // form instruction. 462 // If the instruction reads from the folded operand, it well appear as in 463 // input in both forms. 464 if (MemInOpsNum == RegInOpsNum) 465 Result.IsLoad = true; 466 else 467 Result.IsStore = true; 468 } 469 470 Record *RegOpRec = RegInstr->Operands[FoldedInd].Rec; 471 Record *MemOpRec = MemInstr->Operands[FoldedInd].Rec; 472 473 // Unfolding code generates a load/store instruction according to the size of 474 // the register in the register form instruction. 475 // If the register's size is greater than the memory's operand size, do not 476 // allow unfolding. 477 if (S == UNFOLD) 478 Result.CannotUnfold = false; 479 else if (S == NO_UNFOLD) 480 Result.CannotUnfold = true; 481 else if (getRegOperandSize(RegOpRec) > getMemOperandSize(MemOpRec)) 482 Result.CannotUnfold = true; // S == NO_STRATEGY 483 484 uint64_t Enc = getValueFromBitsInit(RegRec->getValueAsBitsInit("OpEncBits")); 485 if (isExplicitAlign(RegInstr)) { 486 // The instruction require explicitly aligned memory. 487 BitsInit *VectSize = RegRec->getValueAsBitsInit("VectSize"); 488 uint64_t Value = getValueFromBitsInit(VectSize); 489 Result.IsAligned = true; 490 Result.Alignment = Value; 491 } else if (Enc != X86Local::XOP && Enc != X86Local::VEX && 492 Enc != X86Local::EVEX) { 493 // Instructions with VEX encoding do not require alignment. 494 if (!isExplicitUnalign(RegInstr) && getMemOperandSize(MemOpRec) > 64) { 495 // SSE packed vector instructions require a 16 byte alignment. 496 Result.IsAligned = true; 497 Result.Alignment = 16; 498 } 499 } 500 501 Table.push_back(Result); 502 } 503 504 void X86FoldTablesEmitter::updateTables(const CodeGenInstruction *RegInstr, 505 const CodeGenInstruction *MemInstr, 506 const UnfoldStrategy S) { 507 508 Record *RegRec = RegInstr->TheDef; 509 Record *MemRec = MemInstr->TheDef; 510 unsigned MemOutSize = MemRec->getValueAsDag("OutOperandList")->getNumArgs(); 511 unsigned RegOutSize = RegRec->getValueAsDag("OutOperandList")->getNumArgs(); 512 unsigned MemInSize = MemRec->getValueAsDag("InOperandList")->getNumArgs(); 513 unsigned RegInSize = RegRec->getValueAsDag("InOperandList")->getNumArgs(); 514 515 // Instructions which have the WriteRMW value (Read-Modify-Write) should be 516 // added to Table2Addr. 517 if (hasDefInList(MemRec, "SchedRW", "WriteRMW") && MemOutSize != RegOutSize && 518 MemInSize == RegInSize) { 519 addEntryWithFlags(Table2Addr, RegInstr, MemInstr, S, 0); 520 return; 521 } 522 523 if (MemInSize == RegInSize && MemOutSize == RegOutSize) { 524 // Load-Folding cases. 525 // If the i'th register form operand is a register and the i'th memory form 526 // operand is a memory operand, add instructions to Table#i. 527 for (unsigned i = RegOutSize, e = RegInstr->Operands.size(); i < e; i++) { 528 Record *RegOpRec = RegInstr->Operands[i].Rec; 529 Record *MemOpRec = MemInstr->Operands[i].Rec; 530 if (isRegisterOperand(RegOpRec) && isMemoryOperand(MemOpRec)) { 531 switch (i) { 532 case 0: 533 addEntryWithFlags(Table0, RegInstr, MemInstr, S, 0); 534 return; 535 case 1: 536 addEntryWithFlags(Table1, RegInstr, MemInstr, S, 1); 537 return; 538 case 2: 539 addEntryWithFlags(Table2, RegInstr, MemInstr, S, 2); 540 return; 541 case 3: 542 addEntryWithFlags(Table3, RegInstr, MemInstr, S, 3); 543 return; 544 case 4: 545 addEntryWithFlags(Table4, RegInstr, MemInstr, S, 4); 546 return; 547 } 548 } 549 } 550 } else if (MemInSize == RegInSize + 1 && MemOutSize + 1 == RegOutSize) { 551 // Store-Folding cases. 552 // If the memory form instruction performs performs a store, the *output* 553 // register of the register form instructions disappear and instead a 554 // memory *input* operand appears in the memory form instruction. 555 // For example: 556 // MOVAPSrr => (outs VR128:$dst), (ins VR128:$src) 557 // MOVAPSmr => (outs), (ins f128mem:$dst, VR128:$src) 558 Record *RegOpRec = RegInstr->Operands[RegOutSize - 1].Rec; 559 Record *MemOpRec = MemInstr->Operands[RegOutSize - 1].Rec; 560 if (isRegisterOperand(RegOpRec) && isMemoryOperand(MemOpRec)) 561 addEntryWithFlags(Table0, RegInstr, MemInstr, S, 0); 562 } 563 564 return; 565 } 566 567 void X86FoldTablesEmitter::run(raw_ostream &OS) { 568 emitSourceFileHeader("X86 fold tables", OS); 569 570 // Holds all memory instructions 571 std::vector MemInsts; 572 // Holds all register instructions - divided according to opcode. 573 std::map> RegInsts; 574 575 ArrayRef NumberedInstructions = 576 Target.getInstructionsByEnumValue(); 577 578 for (const CodeGenInstruction *Inst : NumberedInstructions) { 579 if (!Inst->TheDef->getNameInit() || !Inst->TheDef->isSubClassOf("X86Inst")) 580 continue; 581 582 const Record *Rec = Inst->TheDef; 583 584 // - Do not proceed if the instruction is marked as notMemoryFoldable. 585 // - Instructions including RST register class operands are not relevant 586 // for memory folding (for further details check the explanation in 587 // lib/Target/X86/X86InstrFPStack.td file). 588 // - Some instructions (listed in the manual map above) use the register 589 // class ptr_rc_tailcall, which can be of a size 32 or 64, to ensure 590 // safe mapping of these instruction we manually map them and exclude 591 // them from the automation. 592 if (Rec->getValueAsBit("isMemoryFoldable") == false || 593 hasRSTRegClass(Inst) || hasPtrTailcallRegClass(Inst)) 594 continue; 595 596 // Add all the memory form instructions to MemInsts, and all the register 597 // form instructions to RegInsts[Opc], where Opc in the opcode of each 598 // instructions. this helps reducing the runtime of the backend. 599 if (hasMemoryFormat(Rec)) 600 MemInsts.push_back(Inst); 601 else if (hasRegisterFormat(Rec)) { 602 uint8_t Opc = getValueFromBitsInit(Rec->getValueAsBitsInit("Opcode")); 603 RegInsts[Opc].push_back(Inst); 604 } 605 } 606 607 // For each memory form instruction, try to find its register form 608 // instruction. 609 for (const CodeGenInstruction *MemInst : MemInsts) { 610 uint8_t Opc = 611 getValueFromBitsInit(MemInst->TheDef->getValueAsBitsInit("Opcode")); 612 613 if (RegInsts.count(Opc) == 0) 614 continue; 615 616 // Two forms (memory & register) of the same instruction must have the same 617 // opcode. try matching only with register form instructions with the same 618 // opcode. 619 std::vector &OpcRegInsts = 620 RegInsts.find(Opc)->second; 621 622 auto Match = find_if(OpcRegInsts, IsMatch(MemInst, Records)); 623 if (Match != OpcRegInsts.end()) { 624 const CodeGenInstruction *RegInst = *Match; 625 // If the matched instruction has it's "FoldGenRegForm" set, map the 626 // memory form instruction to the register form instruction pointed by 627 // this field 628 if (RegInst->TheDef->isValueUnset("FoldGenRegForm")) { 629 updateTables(RegInst, MemInst); 630 } else { 631 const CodeGenInstruction *AltRegInst = 632 getAltRegInst(RegInst, Records, Target); 633 updateTables(AltRegInst, MemInst); 634 } 635 OpcRegInsts.erase(Match); 636 } 637 } 638 639 // Add the manually mapped instructions listed above. 640 for (const ManualMapEntry &Entry : ManualMapSet) { 641 Record *RegInstIter = Records.getDef(Entry.RegInstStr); 642 Record *MemInstIter = Records.getDef(Entry.MemInstStr); 643 644 updateTables(&(Target.getInstruction(RegInstIter)), 645 &(Target.getInstruction(MemInstIter)), Entry.Strategy); 646 } 647 648 // Print all tables to raw_ostream OS. 649 printTable(Table2Addr, "Table2Addr", OS); 650 printTable(Table0, "Table0", OS); 651 printTable(Table1, "Table1", OS); 652 printTable(Table2, "Table2", OS); 653 printTable(Table3, "Table3", OS); 654 printTable(Table4, "Table4", OS); 655 } 656 657 namespace llvm { 658 659 void EmitX86FoldTables(RecordKeeper &RK, raw_ostream &OS) { 660 X86FoldTablesEmitter(RK).run(OS); 661 } 662 } // namespace llvm
__label__pos
0.965186
llvm.org GIT mirror llvm / 504fa89 CodeGen support for x86_64 SEH catch handlers in LLVM This adds handling for ExceptionHandling::MSVC, used by the x86_64-pc-windows-msvc triple. It assumes that filter functions have already been outlined in either the frontend or the backend. Filter functions are used in place of the landingpad catch clause type info operands. In catch clause order, the first filter to return true will catch the exception. The C specific handler table expects the landing pad to be split into one block per handler, but LLVM IR uses a single landing pad for all possible unwind actions. This patch papers over the mismatch by synthesizing single instruction BBs for every catch clause to fill in the EH selector that the landing pad block expects. Missing functionality: - Accessing data in the parent frame from outlined filters - Cleanups (from __finally) are unsupported, as they will require outlining and parent frame access - Filter clauses are unsupported, as there's no clear analogue in SEH In other words, this is the minimal set of changes needed to write IR to catch arbitrary exceptions and resume normal execution. Reviewers: majnemer Differential Revision: http://reviews.llvm.org/D6300 git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@225904 91177308-0d34-0410-b5e6-96231b3b80d8 Reid Kleckner 5 years ago 12 changed file(s) with 653 addition(s) and 18 deletion(s). Raw diff Collapse all Expand all 6565 MachineBasicBlock *LandingPadBlock; // Landing pad block. 6666 SmallVector BeginLabels; // Labels prior to invoke. 6767 SmallVector EndLabels; // Labels after invoke. 68 SmallVector ClauseLabels; // Labels for each clause. 6869 MCSymbol *LandingPadLabel; // Label at beginning of landing pad. 6970 const Function *Personality; // Personality function. 7071 std::vector TypeIds; // List of type ids (filters negative) 329330 /// 330331 void addCleanup(MachineBasicBlock *LandingPad); 331332 333 /// Add a clause for a landing pad. Returns a new label for the clause. This 334 /// is used by EH schemes that have more than one landing pad. In this case, 335 /// each clause gets its own basic block. 336 MCSymbol *addClauseForLandingPad(MachineBasicBlock *LandingPad); 337 332338 /// getTypeIDFor - Return the type id for the specified typeinfo. This is 333339 /// function wide. 334340 unsigned getTypeIDFor(const GlobalValue *TI); 120120 for (unsigned J = NumShared, M = TypeIds.size(); J != M; ++J) { 121121 int TypeID = TypeIds[J]; 122122 assert(-1 - TypeID < (int)FilterOffsets.size() && "Unknown filter id!"); 123 int ValueForTypeID = TypeID < 0 ? FilterOffsets[-1 - TypeID] : TypeID; 123 int ValueForTypeID = 124 isFilterEHSelector(TypeID) ? FilterOffsets[-1 - TypeID] : TypeID; 124125 unsigned SizeTypeID = getSLEB128Size(ValueForTypeID); 125126 126127 int NextAction = SizeAction ? -(SizeAction + SizeTypeID) : 0; 268269 CallSiteEntry Site = { 269270 BeginLabel, 270271 LastLabel, 271 LandingPad->LandingPadLabel, 272 LandingPad, 272273 FirstActions[P.PadIndex] 273274 }; 274275 275276 // Try to merge with the previous call-site. SJLJ doesn't do this 276277 if (PreviousIsInvoke && !IsSJLJ) { 277278 CallSiteEntry &Prev = CallSites.back(); 278 if (Site.PadLabel == Prev.PadLabel && Site.Action == Prev.Action) { 279 if (Site.LPad == Prev.LPad && Site.Action == Prev.Action) { 279280 // Extend the range of the previous entry. 280281 Prev.EndLabel = Site.EndLabel; 281282 continue; 575576 576577 // Offset of the landing pad, counted in 16-byte bundles relative to the 577578 // @LPStart address. 578 if (!S.PadLabel) { 579 if (!S.LPad) { 579580 if (VerboseAsm) 580581 Asm->OutStreamer.AddComment(" has no landing pad"); 581582 Asm->OutStreamer.EmitIntValue(0, 4/*size*/); 582583 } else { 583584 if (VerboseAsm) 584585 Asm->OutStreamer.AddComment(Twine(" jumps to ") + 585 S.PadLabel->getName()); 586 Asm->EmitLabelDifference(S.PadLabel, EHFuncBeginSym, 4); 586 S.LPad->LandingPadLabel->getName()); 587 Asm->EmitLabelDifference(S.LPad->LandingPadLabel, EHFuncBeginSym, 4); 587588 } 588589 589590 // Offset of the first associated action record, relative to the start of 680681 unsigned TypeID = *I; 681682 if (VerboseAsm) { 682683 --Entry; 683 if (TypeID != 0) 684 if (isFilterEHSelector(TypeID)) 684685 Asm->OutStreamer.AddComment("FilterInfo " + Twine(Entry)); 685686 } 686687 2222 class MachineInstr; 2323 class MachineFunction; 2424 class AsmPrinter; 25 class MCSymbol; 26 class MCSymbolRefExpr; 2527 2628 template 2729 class SmallVectorImpl; 5961 /// Structure describing an entry in the call-site table. 6062 struct CallSiteEntry { 6163 // The 'try-range' is BeginLabel .. EndLabel. 62 MCSymbol *BeginLabel; // zero indicates the start of the function. 63 MCSymbol *EndLabel; // zero indicates the end of the function. 64 MCSymbol *BeginLabel; // Null indicates the start of the function. 65 MCSymbol *EndLabel; // Null indicates the end of the function. 6466 65 // The landing pad starts at PadLabel. 66 MCSymbol *PadLabel; // zero indicates that there is no landing pad. 67 // LPad contains the landing pad start labels. 68 const LandingPadInfo *LPad; // Null indicates that there is no landing pad. 6769 unsigned Action; 6870 }; 6971 111113 112114 virtual void emitTypeInfos(unsigned TTypeEncoding); 113115 116 // Helpers for for identifying what kind of clause an EH typeid or selector 117 // corresponds to. Negative selectors are for filter clauses, the zero 118 // selector is for cleanups, and positive selectors are for catch clauses. 119 static bool isFilterEHSelector(int Selector) { return Selector < 0; } 120 static bool isCleanupEHSelector(int Selector) { return Selector == 0; } 121 static bool isCatchEHSelector(int Selector) { return Selector > 0; } 122 114123 public: 115124 EHStreamer(AsmPrinter *A); 116125 virtual ~EHStreamer(); 9898 9999 if (shouldEmitPersonality) { 100100 Asm->OutStreamer.PushSection(); 101 102 // Emit an UNWIND_INFO struct describing the prologue. 101103 Asm->OutStreamer.EmitWinEHHandlerData(); 102 emitExceptionTable(); 104 105 // Emit either MSVC-compatible tables or the usual Itanium-style LSDA after 106 // the UNWIND_INFO struct. 107 if (Asm->MAI->getExceptionHandlingType() == ExceptionHandling::MSVC) { 108 const Function *Per = MMI->getPersonalities()[MMI->getPersonalityIndex()]; 109 if (Per->getName() == "__C_specific_handler") 110 emitCSpecificHandlerTable(); 111 else 112 report_fatal_error(Twine("unexpected personality function: ") + 113 Per->getName()); 114 } else { 115 emitExceptionTable(); 116 } 117 103118 Asm->OutStreamer.PopSection(); 104119 } 105120 Asm->OutStreamer.EmitWinCFIEndProc(); 106121 } 122 123 const MCSymbolRefExpr *Win64Exception::createImageRel32(const MCSymbol *Value) { 124 return MCSymbolRefExpr::Create(Value, MCSymbolRefExpr::VK_COFF_IMGREL32, 125 Asm->OutContext); 126 } 127 128 /// Emit the language-specific data that __C_specific_handler expects. This 129 /// handler lives in the x64 Microsoft C runtime and allows catching or cleaning 130 /// up after faults with __try, __except, and __finally. The typeinfo values 131 /// are not really RTTI data, but pointers to filter functions that return an 132 /// integer (1, 0, or -1) indicating how to handle the exception. For __finally 133 /// blocks and other cleanups, the landing pad label is zero, and the filter 134 /// function is actually a cleanup handler with the same prototype. A catch-all 135 /// entry is modeled with a null filter function field and a non-zero landing 136 /// pad label. 137 /// 138 /// Possible filter function return values: 139 /// EXCEPTION_EXECUTE_HANDLER (1): 140 /// Jump to the landing pad label after cleanups. 141 /// EXCEPTION_CONTINUE_SEARCH (0): 142 /// Continue searching this table or continue unwinding. 143 /// EXCEPTION_CONTINUE_EXECUTION (-1): 144 /// Resume execution at the trapping PC. 145 /// 146 /// Inferred table structure: 147 /// struct Table { 148 /// int NumEntries; 149 /// struct Entry { 150 /// imagerel32 LabelStart; 151 /// imagerel32 LabelEnd; 152 /// imagerel32 FilterOrFinally; // Zero means catch-all. 153 /// imagerel32 LabelLPad; // Zero means __finally. 154 /// } Entries[NumEntries]; 155 /// }; 156 void Win64Exception::emitCSpecificHandlerTable() { 157 const std::vector &PadInfos = MMI->getLandingPads(); 158 159 // Simplifying assumptions for first implementation: 160 // - Cleanups are not implemented. 161 // - Filters are not implemented. 162 163 // The Itanium LSDA table sorts similar landing pads together to simplify the 164 // actions table, but we don't need that. 165 SmallVector LandingPads; 166 LandingPads.reserve(PadInfos.size()); 167 for (const auto &LP : PadInfos) 168 LandingPads.push_back(&LP); 169 170 // Compute label ranges for call sites as we would for the Itanium LSDA, but 171 // use an all zero action table because we aren't using these actions. 172 SmallVector FirstActions; 173 FirstActions.resize(LandingPads.size()); 174 SmallVector CallSites; 175 computeCallSiteTable(CallSites, LandingPads, FirstActions); 176 177 MCSymbol *EHFuncBeginSym = 178 Asm->GetTempSymbol("eh_func_begin", Asm->getFunctionNumber()); 179 MCSymbol *EHFuncEndSym = 180 Asm->GetTempSymbol("eh_func_end", Asm->getFunctionNumber()); 181 182 // Emit the number of table entries. 183 unsigned NumEntries = 0; 184 for (const CallSiteEntry &CSE : CallSites) { 185 if (!CSE.LPad) 186 continue; // Ignore gaps. 187 for (int Selector : CSE.LPad->TypeIds) { 188 // Ignore C++ filter clauses in SEH. 189 // FIXME: Implement cleanup clauses. 190 if (isCatchEHSelector(Selector)) 191 ++NumEntries; 192 } 193 } 194 Asm->OutStreamer.EmitIntValue(NumEntries, 4); 195 196 // Emit the four-label records for each call site entry. The table has to be 197 // sorted in layout order, and the call sites should already be sorted. 198 for (const CallSiteEntry &CSE : CallSites) { 199 // Ignore gaps. Unlike the Itanium model, unwinding through a frame without 200 // an EH table entry will propagate the exception rather than terminating 201 // the program. 202 if (!CSE.LPad) 203 continue; 204 const LandingPadInfo *LPad = CSE.LPad; 205 206 // Compute the label range. We may reuse the function begin and end labels 207 // rather than forming new ones. 208 const MCExpr *Begin = 209 createImageRel32(CSE.BeginLabel ? CSE.BeginLabel : EHFuncBeginSym); 210 const MCExpr *End; 211 if (CSE.EndLabel) { 212 // The interval is half-open, so we have to add one to include the return 213 // address of the last invoke in the range. 214 End = MCBinaryExpr::CreateAdd(createImageRel32(CSE.EndLabel), 215 MCConstantExpr::Create(1, Asm->OutContext), 216 Asm->OutContext); 217 } else { 218 End = createImageRel32(EHFuncEndSym); 219 } 220 221 // These aren't really type info globals, they are actually pointers to 222 // filter functions ordered by selector. The zero selector is used for 223 // cleanups, so slot zero corresponds to selector 1. 224 const std::vector &SelectorToFilter = MMI->getTypeInfos(); 225 226 // Do a parallel iteration across typeids and clause labels, skipping filter 227 // clauses. 228 assert(LPad->TypeIds.size() == LPad->ClauseLabels.size()); 229 for (size_t I = 0, E = LPad->TypeIds.size(); I < E; ++I) { 230 // AddLandingPadInfo stores the clauses in reverse, but there is a FIXME 231 // to change that. 232 int Selector = LPad->TypeIds[E - I - 1]; 233 MCSymbol *ClauseLabel = LPad->ClauseLabels[I]; 234 235 // Ignore C++ filter clauses in SEH. 236 // FIXME: Implement cleanup clauses. 237 if (!isCatchEHSelector(Selector)) 238 continue; 239 240 Asm->OutStreamer.EmitValue(Begin, 4); 241 Asm->OutStreamer.EmitValue(End, 4); 242 if (isCatchEHSelector(Selector)) { 243 assert(unsigned(Selector - 1) < SelectorToFilter.size()); 244 const GlobalValue *TI = SelectorToFilter[Selector - 1]; 245 if (TI) // Emit the filter function pointer. 246 Asm->OutStreamer.EmitValue(createImageRel32(Asm->getSymbol(TI)), 4); 247 else // Otherwise, this is a "catch i8* null", or catch all. 248 Asm->OutStreamer.EmitIntValue(0, 4); 249 } 250 Asm->OutStreamer.EmitValue(createImageRel32(ClauseLabel), 4); 251 } 252 } 253 } 2828 /// Per-function flag to indicate if frame moves info should be emitted. 2929 bool shouldEmitMoves; 3030 31 void emitCSpecificHandlerTable(); 32 33 const MCSymbolRefExpr *createImageRel32(const MCSymbol *Value); 34 3135 public: 3236 //===--------------------------------------------------------------------===// 3337 // Main entry points. 451451 LP.TypeIds.push_back(0); 452452 } 453453 454 MCSymbol * 455 MachineModuleInfo::addClauseForLandingPad(MachineBasicBlock *LandingPad) { 456 MCSymbol *ClauseLabel = Context.CreateTempSymbol(); 457 LandingPadInfo &LP = getOrCreateLandingPadInfo(LandingPad); 458 LP.ClauseLabels.push_back(ClauseLabel); 459 return ClauseLabel; 460 } 461 454462 /// TidyLandingPads - Remap landing pad labels and remove any deleted landing 455463 /// pads. 456464 void MachineModuleInfo::TidyLandingPads(DenseMap *LPMap) { 448448 case ExceptionHandling::DwarfCFI: 449449 case ExceptionHandling::ARM: 450450 case ExceptionHandling::ItaniumWinEH: 451 case ExceptionHandling::MSVC: // FIXME: Needs preparation. 451452 addPass(createDwarfEHPass(TM)); 452453 break; 453 case ExceptionHandling::MSVC: // FIXME: Add preparation. 454454 case ExceptionHandling::None: 455455 addPass(createLowerInvokePass()); 456456 20702070 // Get the two live-in registers as SDValues. The physregs have already been 20712071 // copied into virtual registers. 20722072 SDValue Ops[2]; 2073 Ops[0] = DAG.getZExtOrTrunc( 2074 DAG.getCopyFromReg(DAG.getEntryNode(), getCurSDLoc(), 2075 FuncInfo.ExceptionPointerVirtReg, TLI.getPointerTy()), 2076 getCurSDLoc(), ValueVTs[0]); 2073 if (FuncInfo.ExceptionPointerVirtReg) { 2074 Ops[0] = DAG.getZExtOrTrunc( 2075 DAG.getCopyFromReg(DAG.getEntryNode(), getCurSDLoc(), 2076 FuncInfo.ExceptionPointerVirtReg, TLI.getPointerTy()), 2077 getCurSDLoc(), ValueVTs[0]); 2078 } else { 2079 Ops[0] = DAG.getConstant(0, TLI.getPointerTy()); 2080 } 20772081 Ops[1] = DAG.getZExtOrTrunc( 20782082 DAG.getCopyFromReg(DAG.getEntryNode(), getCurSDLoc(), 20792083 FuncInfo.ExceptionSelectorVirtReg, TLI.getPointerTy()), 20832087 SDValue Res = DAG.getNode(ISD::MERGE_VALUES, getCurSDLoc(), 20842088 DAG.getVTList(ValueVTs), Ops); 20852089 setValue(&LP, Res); 2090 } 2091 2092 unsigned 2093 SelectionDAGBuilder::visitLandingPadClauseBB(GlobalValue *ClauseGV, 2094 MachineBasicBlock *LPadBB) { 2095 SDValue Chain = getControlRoot(); 2096 2097 // Get the typeid that we will dispatch on later. 2098 const TargetLowering &TLI = DAG.getTargetLoweringInfo(); 2099 const TargetRegisterClass *RC = TLI.getRegClassFor(TLI.getPointerTy()); 2100 unsigned VReg = FuncInfo.MF->getRegInfo().createVirtualRegister(RC); 2101 unsigned TypeID = DAG.getMachineFunction().getMMI().getTypeIDFor(ClauseGV); 2102 SDValue Sel = DAG.getConstant(TypeID, TLI.getPointerTy()); 2103 Chain = DAG.getCopyToReg(Chain, getCurSDLoc(), VReg, Sel); 2104 2105 // Branch to the main landing pad block. 2106 MachineBasicBlock *ClauseMBB = FuncInfo.MBB; 2107 ClauseMBB->addSuccessor(LPadBB); 2108 DAG.setRoot(DAG.getNode(ISD::BR, getCurSDLoc(), MVT::Other, Chain, 2109 DAG.getBasicBlock(LPadBB))); 2110 return VReg; 20862111 } 20872112 20882113 /// handleSmallSwitchCaseRange - Emit a series of specific tests (suitable for 712712 void visitJumpTable(JumpTable &JT); 713713 void visitJumpTableHeader(JumpTable &JT, JumpTableHeader &JTH, 714714 MachineBasicBlock *SwitchBB); 715 unsigned visitLandingPadClauseBB(GlobalValue *ClauseGV, 716 MachineBasicBlock *LPadMBB); 715717 716718 private: 717719 // These all get lowered before this pass. 1818 #include "llvm/Analysis/AliasAnalysis.h" 1919 #include "llvm/Analysis/BranchProbabilityInfo.h" 2020 #include "llvm/Analysis/CFG.h" 21 #include "llvm/CodeGen/Analysis.h" 2122 #include "llvm/CodeGen/FastISel.h" 2223 #include "llvm/CodeGen/FunctionLoweringInfo.h" 2324 #include "llvm/CodeGen/GCMetadata.h" 3940 #include "llvm/IR/Intrinsics.h" 4041 #include "llvm/IR/LLVMContext.h" 4142 #include "llvm/IR/Module.h" 43 #include "llvm/MC/MCAsmInfo.h" 4244 #include "llvm/Support/Compiler.h" 4345 #include "llvm/Support/Debug.h" 4446 #include "llvm/Support/ErrorHandling.h" 891893 void SelectionDAGISel::PrepareEHLandingPad() { 892894 MachineBasicBlock *MBB = FuncInfo->MBB; 893895 896 const TargetRegisterClass *PtrRC = TLI->getRegClassFor(TLI->getPointerTy()); 897 894898 // Add a label to mark the beginning of the landing pad. Deletion of the 895899 // landing pad can thus be detected via the MachineModuleInfo. 896900 MCSymbol *Label = MF->getMMI().addLandingPad(MBB); 902906 BuildMI(*MBB, FuncInfo->InsertPt, SDB->getCurDebugLoc(), II) 903907 .addSym(Label); 904908 909 if (TM.getMCAsmInfo()->getExceptionHandlingType() == 910 ExceptionHandling::MSVC) { 911 // Make virtual registers and a series of labels that fill in values for the 912 // clauses. 913 auto &RI = MF->getRegInfo(); 914 FuncInfo->ExceptionSelectorVirtReg = RI.createVirtualRegister(PtrRC); 915 916 // Get all invoke BBs that will unwind into the clause BBs. 917 SmallVector InvokeBBs(MBB->pred_begin(), 918 MBB->pred_end()); 919 920 // Emit separate machine basic blocks with separate labels for each clause 921 // before the main landing pad block. 922 const BasicBlock *LLVMBB = MBB->getBasicBlock(); 923 const LandingPadInst *LPadInst = LLVMBB->getLandingPadInst(); 924 MachineInstrBuilder SelectorPHI = BuildMI( 925 *MBB, MBB->begin(), SDB->getCurDebugLoc(), TII->get(TargetOpcode::PHI), 926 FuncInfo->ExceptionSelectorVirtReg); 927 for (unsigned I = 0, E = LPadInst->getNumClauses(); I != E; ++I) { 928 MachineBasicBlock *ClauseBB = MF->CreateMachineBasicBlock(LLVMBB); 929 MF->insert(MBB, ClauseBB); 930 931 // Add the edge from the invoke to the clause. 932 for (MachineBasicBlock *InvokeBB : InvokeBBs) 933 InvokeBB->addSuccessor(ClauseBB); 934 935 // Mark the clause as a landing pad or MI passes will delete it. 936 ClauseBB->setIsLandingPad(); 937 938 GlobalValue *ClauseGV = ExtractTypeInfo(LPadInst->getClause(I)); 939 940 // Start the BB with a label. 941 MCSymbol *ClauseLabel = MF->getMMI().addClauseForLandingPad(MBB); 942 BuildMI(*ClauseBB, ClauseBB->begin(), SDB->getCurDebugLoc(), II) 943 .addSym(ClauseLabel); 944 945 // Construct a simple BB that defines a register with the typeid constant. 946 FuncInfo->MBB = ClauseBB; 947 FuncInfo->InsertPt = ClauseBB->end(); 948 unsigned VReg = SDB->visitLandingPadClauseBB(ClauseGV, MBB); 949 CurDAG->setRoot(SDB->getRoot()); 950 SDB->clear(); 951 CodeGenAndEmitDAG(); 952 953 // Add the typeid virtual register to the phi in the main landing pad. 954 SelectorPHI.addReg(VReg).addMBB(ClauseBB); 955 } 956 957 // Remove the edge from the invoke to the lpad. 958 for (MachineBasicBlock *InvokeBB : InvokeBBs) 959 InvokeBB->removeSuccessor(MBB); 960 961 // Restore FuncInfo back to its previous state and select the main landing 962 // pad block. 963 FuncInfo->MBB = MBB; 964 FuncInfo->InsertPt = MBB->end(); 965 return; 966 } 967 905968 // Mark exception register as live in. 906 const TargetRegisterClass *PtrRC = TLI->getRegClassFor(TLI->getPointerTy()); 907969 if (unsigned Reg = TLI->getExceptionPointerRegister()) 908970 FuncInfo->ExceptionPointerVirtReg = MBB->addLiveIn(Reg, PtrRC); 909971 0 ; RUN: llc -mtriple x86_64-pc-windows-msvc < %s | FileCheck %s 1 2 define void @two_invoke_merged() { 3 entry: 4 invoke void @try_body() 5 to label %again unwind label %lpad 6 7 again: 8 invoke void @try_body() 9 to label %done unwind label %lpad 10 11 done: 12 ret void 13 14 lpad: 15 %vals = landingpad { i8*, i32 } personality i8* bitcast (i32 (...)* @__C_specific_handler to i8*) 16 catch i8* bitcast (i32 (i8*, i8*)* @filt0 to i8*) 17 catch i8* bitcast (i32 (i8*, i8*)* @filt1 to i8*) 18 %sel = extractvalue { i8*, i32 } %vals, 1 19 call void @use_selector(i32 %sel) 20 ret void 21 } 22 23 ; Normal path code 24 25 ; CHECK-LABEL: {{^}}two_invoke_merged: 26 ; CHECK: .seh_proc two_invoke_merged 27 ; CHECK: .seh_handler __C_specific_handler, @unwind, @except 28 ; CHECK: .Ltmp0: 29 ; CHECK: callq try_body 30 ; CHECK-NEXT: .Ltmp1: 31 ; CHECK: .Ltmp2: 32 ; CHECK: callq try_body 33 ; CHECK-NEXT: .Ltmp3: 34 ; CHECK: retq 35 36 ; Landing pad code 37 38 ; CHECK: .Ltmp5: 39 ; CHECK: movl $1, %ecx 40 ; CHECK: jmp 41 ; CHECK: .Ltmp6: 42 ; CHECK: movl $2, %ecx 43 ; CHECK: callq use_selector 44 45 ; CHECK: .seh_handlerdata 46 ; CHECK-NEXT: .long 2 47 ; CHECK-NEXT: .long .Ltmp0@IMGREL 48 ; CHECK-NEXT: .long .Ltmp3@IMGREL+1 49 ; CHECK-NEXT: .long filt0@IMGREL 50 ; CHECK-NEXT: .long .Ltmp5@IMGREL 51 ; CHECK-NEXT: .long .Ltmp0@IMGREL 52 ; CHECK-NEXT: .long .Ltmp3@IMGREL+1 53 ; CHECK-NEXT: .long filt1@IMGREL 54 ; CHECK-NEXT: .long .Ltmp6@IMGREL 55 ; CHECK: .text 56 ; CHECK: .seh_endproc 57 58 define void @two_invoke_gap() { 59 entry: 60 invoke void @try_body() 61 to label %again unwind label %lpad 62 63 again: 64 call void @do_nothing_on_unwind() 65 invoke void @try_body() 66 to label %done unwind label %lpad 67 68 done: 69 ret void 70 71 lpad: 72 %vals = landingpad { i8*, i32 } personality i8* bitcast (i32 (...)* @__C_specific_handler to i8*) 73 catch i8* bitcast (i32 (i8*, i8*)* @filt0 to i8*) 74 %sel = extractvalue { i8*, i32 } %vals, 1 75 call void @use_selector(i32 %sel) 76 ret void 77 } 78 79 ; Normal path code 80 81 ; CHECK-LABEL: {{^}}two_invoke_gap: 82 ; CHECK: .seh_proc two_invoke_gap 83 ; CHECK: .seh_handler __C_specific_handler, @unwind, @except 84 ; CHECK: .Ltmp11: 85 ; CHECK: callq try_body 86 ; CHECK-NEXT: .Ltmp12: 87 ; CHECK: callq do_nothing_on_unwind 88 ; CHECK: .Ltmp13: 89 ; CHECK: callq try_body 90 ; CHECK-NEXT: .Ltmp14: 91 ; CHECK: retq 92 93 ; Landing pad code 94 95 ; CHECK: .Ltmp16: 96 ; CHECK: movl $1, %ecx 97 ; CHECK: callq use_selector 98 99 ; CHECK: .seh_handlerdata 100 ; CHECK-NEXT: .long 2 101 ; CHECK-NEXT: .long .Ltmp11@IMGREL 102 ; CHECK-NEXT: .long .Ltmp12@IMGREL+1 103 ; CHECK-NEXT: .long filt0@IMGREL 104 ; CHECK-NEXT: .long .Ltmp16@IMGREL 105 ; CHECK-NEXT: .long .Ltmp13@IMGREL 106 ; CHECK-NEXT: .long .Ltmp14@IMGREL+1 107 ; CHECK-NEXT: .long filt0@IMGREL 108 ; CHECK-NEXT: .long .Ltmp16@IMGREL 109 ; CHECK: .text 110 ; CHECK: .seh_endproc 111 112 define void @two_invoke_nounwind_gap() { 113 entry: 114 invoke void @try_body() 115 to label %again unwind label %lpad 116 117 again: 118 call void @cannot_unwind() 119 invoke void @try_body() 120 to label %done unwind label %lpad 121 122 done: 123 ret void 124 125 lpad: 126 %vals = landingpad { i8*, i32 } personality i8* bitcast (i32 (...)* @__C_specific_handler to i8*) 127 catch i8* bitcast (i32 (i8*, i8*)* @filt0 to i8*) 128 %sel = extractvalue { i8*, i32 } %vals, 1 129 call void @use_selector(i32 %sel) 130 ret void 131 } 132 133 ; Normal path code 134 135 ; CHECK-LABEL: {{^}}two_invoke_nounwind_gap: 136 ; CHECK: .seh_proc two_invoke_nounwind_gap 137 ; CHECK: .seh_handler __C_specific_handler, @unwind, @except 138 ; CHECK: .Ltmp21: 139 ; CHECK: callq try_body 140 ; CHECK-NEXT: .Ltmp22: 141 ; CHECK: callq cannot_unwind 142 ; CHECK: .Ltmp23: 143 ; CHECK: callq try_body 144 ; CHECK-NEXT: .Ltmp24: 145 ; CHECK: retq 146 147 ; Landing pad code 148 149 ; CHECK: .Ltmp26: 150 ; CHECK: movl $1, %ecx 151 ; CHECK: callq use_selector 152 153 ; CHECK: .seh_handlerdata 154 ; CHECK-NEXT: .long 1 155 ; CHECK-NEXT: .long .Ltmp21@IMGREL 156 ; CHECK-NEXT: .long .Ltmp24@IMGREL+1 157 ; CHECK-NEXT: .long filt0@IMGREL 158 ; CHECK-NEXT: .long .Ltmp26@IMGREL 159 ; CHECK: .text 160 ; CHECK: .seh_endproc 161 162 declare void @try_body() 163 declare void @do_nothing_on_unwind() 164 declare void @cannot_unwind() nounwind 165 declare void @use_selector(i32) 166 167 declare i32 @filt0(i8* %eh_info, i8* %rsp) 168 declare i32 @filt1(i8* %eh_info, i8* %rsp) 169 170 declare void @handler0() 171 declare void @handler1() 172 173 declare i32 @__C_specific_handler(...) 174 declare i32 @llvm.eh.typeid.for(i8*) readnone nounwind 0 ; RUN: llc -mtriple x86_64-pc-windows-msvc < %s | FileCheck %s 1 2 ; This test case is also intended to be run manually as a complete functional 3 ; test. It should link, print something, and exit zero rather than crashing. 4 ; It is the hypothetical lowering of a C source program that looks like: 5 ; 6 ; int safe_div(int *n, int *d) { 7 ; int r; 8 ; __try { 9 ; __try { 10 ; r = *n / *d; 11 ; } __except(GetExceptionCode() == EXCEPTION_ACCESS_VIOLATION) { 12 ; puts("EXCEPTION_ACCESS_VIOLATION"); 13 ; r = -1; 14 ; } 15 ; } __except(GetExceptionCode() == EXCEPTION_INT_DIVIDE_BY_ZERO) { 16 ; puts("EXCEPTION_INT_DIVIDE_BY_ZERO"); 17 ; r = -2; 18 ; } 19 ; return r; 20 ; } 21 22 @str1 = internal constant [27 x i8] c"EXCEPTION_ACCESS_VIOLATION\00" 23 @str2 = internal constant [29 x i8] c"EXCEPTION_INT_DIVIDE_BY_ZERO\00" 24 25 define i32 @safe_div(i32* %n, i32* %d) { 26 entry: 27 %r = alloca i32, align 4 28 invoke void @try_body(i32* %r, i32* %n, i32* %d) 29 to label %__try.cont unwind label %lpad 30 31 lpad: 32 %vals = landingpad { i8*, i32 } personality i8* bitcast (i32 (...)* @__C_specific_handler to i8*) 33 catch i8* bitcast (i32 (i8*, i8*)* @safe_div_filt0 to i8*) 34 catch i8* bitcast (i32 (i8*, i8*)* @safe_div_filt1 to i8*) 35 %ehptr = extractvalue { i8*, i32 } %vals, 0 36 %sel = extractvalue { i8*, i32 } %vals, 1 37 %filt0_val = call i32 @llvm.eh.typeid.for(i8* bitcast (i32 (i8*, i8*)* @safe_div_filt0 to i8*)) 38 %is_filt0 = icmp eq i32 %sel, %filt0_val 39 br i1 %is_filt0, label %handler0, label %eh.dispatch1 40 41 eh.dispatch1: 42 %filt1_val = call i32 @llvm.eh.typeid.for(i8* bitcast (i32 (i8*, i8*)* @safe_div_filt1 to i8*)) 43 %is_filt1 = icmp eq i32 %sel, %filt1_val 44 br i1 %is_filt1, label %handler1, label %eh.resume 45 46 handler0: 47 call void @puts(i8* getelementptr ([27 x i8]* @str1, i32 0, i32 0)) 48 store i32 -1, i32* %r, align 4 49 br label %__try.cont 50 51 handler1: 52 call void @puts(i8* getelementptr ([29 x i8]* @str2, i32 0, i32 0)) 53 store i32 -2, i32* %r, align 4 54 br label %__try.cont 55 56 eh.resume: 57 resume { i8*, i32 } %vals 58 59 __try.cont: 60 %safe_ret = load i32* %r, align 4 61 ret i32 %safe_ret 62 } 63 64 ; Normal path code 65 66 ; CHECK: {{^}}safe_div: 67 ; CHECK: .seh_proc safe_div 68 ; CHECK: .seh_handler __C_specific_handler, @unwind, @except 69 ; CHECK: .Ltmp0: 70 ; CHECK: leaq [[rloc:.*\(%rsp\)]], %rcx 71 ; CHECK: callq try_body 72 ; CHECK-NEXT: .Ltmp1 73 ; CHECK: .LBB0_7: 74 ; CHECK: movl [[rloc]], %eax 75 ; CHECK: retq 76 77 ; Landing pad code 78 79 ; CHECK: .Ltmp3: 80 ; CHECK: movl $1, %[[sel:[a-z]+]] 81 ; CHECK: .Ltmp4 82 ; CHECK: movl $2, %[[sel]] 83 ; CHECK: .L{{.*}}: 84 ; CHECK: cmpl $1, %[[sel]] 85 86 ; CHECK: # %handler0 87 ; CHECK: callq puts 88 ; CHECK: movl $-1, [[rloc]] 89 ; CHECK: jmp .LBB0_7 90 91 ; CHECK: cmpl $2, %[[sel]] 92 93 ; CHECK: # %handler1 94 ; CHECK: callq puts 95 ; CHECK: movl $-2, [[rloc]] 96 ; CHECK: jmp .LBB0_7 97 98 ; FIXME: EH preparation should not call _Unwind_Resume. 99 ; CHECK: callq _Unwind_Resume 100 ; CHECK: ud2 101 102 ; CHECK: .seh_handlerdata 103 ; CHECK: .long 2 104 ; CHECK: .long .Ltmp0@IMGREL 105 ; CHECK: .long .Ltmp1@IMGREL+1 106 ; CHECK: .long safe_div_filt0@IMGREL 107 ; CHECK: .long .Ltmp3@IMGREL 108 ; CHECK: .long .Ltmp0@IMGREL 109 ; CHECK: .long .Ltmp1@IMGREL+1 110 ; CHECK: .long safe_div_filt1@IMGREL 111 ; CHECK: .long .Ltmp4@IMGREL 112 ; CHECK: .text 113 ; CHECK: .seh_endproc 114 115 116 define void @try_body(i32* %r, i32* %n, i32* %d) { 117 entry: 118 %0 = load i32* %n, align 4 119 %1 = load i32* %d, align 4 120 %div = sdiv i32 %0, %1 121 store i32 %div, i32* %r, align 4 122 ret void 123 } 124 125 ; The prototype of these filter functions is: 126 ; int filter(EXCEPTION_POINTERS *eh_ptrs, void *rbp); 127 128 ; The definition of EXCEPTION_POINTERS is: 129 ; typedef struct _EXCEPTION_POINTERS { 130 ; EXCEPTION_RECORD *ExceptionRecord; 131 ; CONTEXT *ContextRecord; 132 ; } EXCEPTION_POINTERS; 133 134 ; The definition of EXCEPTION_RECORD is: 135 ; typedef struct _EXCEPTION_RECORD { 136 ; DWORD ExceptionCode; 137 ; ... 138 ; } EXCEPTION_RECORD; 139 140 ; The exception code can be retreived with two loads, one for the record 141 ; pointer and one for the code. The values of local variables can be 142 ; accessed via rbp, but that would require additional not yet implemented LLVM 143 ; support. 144 145 define i32 @safe_div_filt0(i8* %eh_ptrs, i8* %rbp) { 146 %eh_ptrs_c = bitcast i8* %eh_ptrs to i32** 147 %eh_rec = load i32** %eh_ptrs_c 148 %eh_code = load i32* %eh_rec 149 ; EXCEPTION_ACCESS_VIOLATION = 0xC0000005 150 %cmp = icmp eq i32 %eh_code, 3221225477 151 %filt.res = zext i1 %cmp to i32 152 ret i32 %filt.res 153 } 154 155 define i32 @safe_div_filt1(i8* %eh_ptrs, i8* %rbp) { 156 %eh_ptrs_c = bitcast i8* %eh_ptrs to i32** 157 %eh_rec = load i32** %eh_ptrs_c 158 %eh_code = load i32* %eh_rec 159 ; EXCEPTION_INT_DIVIDE_BY_ZERO = 0xC0000094 160 %cmp = icmp eq i32 %eh_code, 3221225620 161 %filt.res = zext i1 %cmp to i32 162 ret i32 %filt.res 163 } 164 165 @str_result = internal constant [21 x i8] c"safe_div result: %d\0A\00" 166 167 define i32 @main() { 168 %d.addr = alloca i32, align 4 169 %n.addr = alloca i32, align 4 170 171 store i32 10, i32* %n.addr, align 4 172 store i32 2, i32* %d.addr, align 4 173 %r1 = call i32 @safe_div(i32* %n.addr, i32* %d.addr) 174 call void (i8*, ...)* @printf(i8* getelementptr ([21 x i8]* @str_result, i32 0, i32 0), i32 %r1) 175 176 store i32 10, i32* %n.addr, align 4 177 store i32 0, i32* %d.addr, align 4 178 %r2 = call i32 @safe_div(i32* %n.addr, i32* %d.addr) 179 call void (i8*, ...)* @printf(i8* getelementptr ([21 x i8]* @str_result, i32 0, i32 0), i32 %r2) 180 181 %r3 = call i32 @safe_div(i32* %n.addr, i32* null) 182 call void (i8*, ...)* @printf(i8* getelementptr ([21 x i8]* @str_result, i32 0, i32 0), i32 %r3) 183 ret i32 0 184 } 185 186 define void @_Unwind_Resume() { 187 call void @abort() 188 unreachable 189 } 190 191 declare i32 @__C_specific_handler(...) 192 declare i32 @llvm.eh.typeid.for(i8*) readnone nounwind 193 declare void @puts(i8*) 194 declare void @printf(i8*, ...) 195 declare void @abort()
__label__pos
0.760666
 Reference documentation for deal.II version Git f70953c 2018-04-22 22:20:09 +0200 Classes | Public Member Functions | Static Public Member Functions | Static Public Attributes | Protected Member Functions | Static Protected Member Functions | Protected Attributes | Friends | List of all members FiniteElement< dim, spacedim > Class Template Referenceabstract #include <deal.II/fe/fe.h> Inheritance diagram for FiniteElement< dim, spacedim >: [legend] Classes class  InternalDataBase   Public Member Functions  FiniteElement (const FiniteElementData< dim > &fe_data, const std::vector< bool > &restriction_is_additive_flags, const std::vector< ComponentMask > &nonzero_components)    FiniteElement (FiniteElement< dim, spacedim > &&)=default    FiniteElement (const FiniteElement< dim, spacedim > &)=default   virtual ~FiniteElement ()=default   std::pair< std::unique_ptr< FiniteElement< dim, spacedim > >, unsigned int > operator^ (const unsigned int multiplicity) const   virtual std::unique_ptr< FiniteElement< dim, spacedim > > clone () const =0   virtual std::string get_name () const =0   const FiniteElement< dim, spacedim > & operator[] (const unsigned int fe_index) const   bool operator== (const FiniteElement< dim, spacedim > &) const   virtual std::size_t memory_consumption () const   Shape function access virtual double shape_value (const unsigned int i, const Point< dim > &p) const   virtual double shape_value_component (const unsigned int i, const Point< dim > &p, const unsigned int component) const   virtual Tensor< 1, dim > shape_grad (const unsigned int i, const Point< dim > &p) const   virtual Tensor< 1, dim > shape_grad_component (const unsigned int i, const Point< dim > &p, const unsigned int component) const   virtual Tensor< 2, dim > shape_grad_grad (const unsigned int i, const Point< dim > &p) const   virtual Tensor< 2, dim > shape_grad_grad_component (const unsigned int i, const Point< dim > &p, const unsigned int component) const   virtual Tensor< 3, dim > shape_3rd_derivative (const unsigned int i, const Point< dim > &p) const   virtual Tensor< 3, dim > shape_3rd_derivative_component (const unsigned int i, const Point< dim > &p, const unsigned int component) const   virtual Tensor< 4, dim > shape_4th_derivative (const unsigned int i, const Point< dim > &p) const   virtual Tensor< 4, dim > shape_4th_derivative_component (const unsigned int i, const Point< dim > &p, const unsigned int component) const   virtual bool has_support_on_face (const unsigned int shape_index, const unsigned int face_index) const   Transfer and constraint matrices virtual const FullMatrix< double > & get_restriction_matrix (const unsigned int child, const RefinementCase< dim > &refinement_case=RefinementCase< dim >::isotropic_refinement) const   virtual const FullMatrix< double > & get_prolongation_matrix (const unsigned int child, const RefinementCase< dim > &refinement_case=RefinementCase< dim >::isotropic_refinement) const   bool prolongation_is_implemented () const   bool isotropic_prolongation_is_implemented () const   bool restriction_is_implemented () const   bool isotropic_restriction_is_implemented () const   bool restriction_is_additive (const unsigned int index) const   const FullMatrix< double > & constraints (const ::internal::SubfaceCase< dim > &subface_case=::internal::SubfaceCase< dim >::case_isotropic) const   bool constraints_are_implemented (const ::internal::SubfaceCase< dim > &subface_case=::internal::SubfaceCase< dim >::case_isotropic) const   virtual bool hp_constraints_are_implemented () const   virtual void get_interpolation_matrix (const FiniteElement< dim, spacedim > &source, FullMatrix< double > &matrix) const   Functions to support hp virtual void get_face_interpolation_matrix (const FiniteElement< dim, spacedim > &source, FullMatrix< double > &matrix) const   virtual void get_subface_interpolation_matrix (const FiniteElement< dim, spacedim > &source, const unsigned int subface, FullMatrix< double > &matrix) const   virtual std::vector< std::pair< unsigned int, unsigned int > > hp_vertex_dof_identities (const FiniteElement< dim, spacedim > &fe_other) const   virtual std::vector< std::pair< unsigned int, unsigned int > > hp_line_dof_identities (const FiniteElement< dim, spacedim > &fe_other) const   virtual std::vector< std::pair< unsigned int, unsigned int > > hp_quad_dof_identities (const FiniteElement< dim, spacedim > &fe_other) const   virtual FiniteElementDomination::Domination compare_for_face_domination (const FiniteElement< dim, spacedim > &fe_other) const   Index computations std::pair< unsigned int, unsigned int > system_to_component_index (const unsigned int index) const   unsigned int component_to_system_index (const unsigned int component, const unsigned int index) const   std::pair< unsigned int, unsigned int > face_system_to_component_index (const unsigned int index) const   unsigned int adjust_quad_dof_index_for_face_orientation (const unsigned int index, const bool face_orientation, const bool face_flip, const bool face_rotation) const   virtual unsigned int face_to_cell_index (const unsigned int face_dof_index, const unsigned int face, const bool face_orientation=true, const bool face_flip=false, const bool face_rotation=false) const   unsigned int adjust_line_dof_index_for_line_orientation (const unsigned int index, const bool line_orientation) const   const ComponentMaskget_nonzero_components (const unsigned int i) const   unsigned int n_nonzero_components (const unsigned int i) const   bool is_primitive () const   bool is_primitive (const unsigned int i) const   unsigned int n_base_elements () const   virtual const FiniteElement< dim, spacedim > & base_element (const unsigned int index) const   unsigned int element_multiplicity (const unsigned int index) const   const FiniteElement< dim, spacedim > & get_sub_fe (const ComponentMask &mask) const   virtual const FiniteElement< dim, spacedim > & get_sub_fe (const unsigned int first_component, const unsigned int n_selected_components) const   std::pair< std::pair< unsigned int, unsigned int >, unsigned int > system_to_base_index (const unsigned int index) const   std::pair< std::pair< unsigned int, unsigned int >, unsigned int > face_system_to_base_index (const unsigned int index) const   types::global_dof_index first_block_of_base (const unsigned int b) const   std::pair< unsigned int, unsigned int > component_to_base_index (const unsigned int component) const   std::pair< unsigned int, unsigned int > block_to_base_index (const unsigned int block) const   std::pair< unsigned int, types::global_dof_indexsystem_to_block_index (const unsigned int component) const   unsigned int component_to_block_index (const unsigned int component) const   Component and block matrices ComponentMask component_mask (const FEValuesExtractors::Scalar &scalar) const   ComponentMask component_mask (const FEValuesExtractors::Vector &vector) const   ComponentMask component_mask (const FEValuesExtractors::SymmetricTensor< 2 > &sym_tensor) const   ComponentMask component_mask (const BlockMask &block_mask) const   BlockMask block_mask (const FEValuesExtractors::Scalar &scalar) const   BlockMask block_mask (const FEValuesExtractors::Vector &vector) const   BlockMask block_mask (const FEValuesExtractors::SymmetricTensor< 2 > &sym_tensor) const   BlockMask block_mask (const ComponentMask &component_mask) const   virtual std::pair< Table< 2, bool >, std::vector< unsigned int > > get_constant_modes () const   Support points and interpolation const std::vector< Point< dim > > & get_unit_support_points () const   bool has_support_points () const   virtual Point< dim > unit_support_point (const unsigned int index) const   const std::vector< Point< dim-1 > > & get_unit_face_support_points () const   bool has_face_support_points () const   virtual Point< dim-1 > unit_face_support_point (const unsigned int index) const   const std::vector< Point< dim > > & get_generalized_support_points () const   bool has_generalized_support_points () const   const std::vector< Point< dim-1 > > & get_generalized_face_support_points () const   bool has_generalized_face_support_points () const   GeometryPrimitive get_associated_geometry_primitive (const unsigned int cell_dof_index) const   virtual void convert_generalized_support_point_values_to_dof_values (const std::vector< Vector< double > > &support_point_values, std::vector< double > &nodal_values) const   - Public Member Functions inherited from Subscriptor  Subscriptor ()    Subscriptor (const Subscriptor &)    Subscriptor (Subscriptor &&) noexcept   virtual ~Subscriptor ()   Subscriptoroperator= (const Subscriptor &)   Subscriptoroperator= (Subscriptor &&) noexcept   void subscribe (const char *identifier=nullptr) const   void unsubscribe (const char *identifier=nullptr) const   unsigned int n_subscriptions () const   void list_subscribers () const   template<class Archive > void serialize (Archive &ar, const unsigned int version)   - Public Member Functions inherited from FiniteElementData< dim >  FiniteElementData (const std::vector< unsigned int > &dofs_per_object, const unsigned int n_components, const unsigned int degree, const Conformity conformity=unknown, const BlockIndices &block_indices=BlockIndices())   unsigned int n_dofs_per_vertex () const   unsigned int n_dofs_per_line () const   unsigned int n_dofs_per_quad () const   unsigned int n_dofs_per_hex () const   unsigned int n_dofs_per_face () const   unsigned int n_dofs_per_cell () const   template<int structdim> unsigned int n_dofs_per_object () const   unsigned int n_components () const   unsigned int n_blocks () const   const BlockIndicesblock_indices () const   unsigned int tensor_degree () const   bool conforms (const Conformity) const   bool operator== (const FiniteElementData &) const   Static Public Member Functions static::ExceptionBase & ExcShapeFunctionNotPrimitive (int arg1)   static::ExceptionBase & ExcFENotPrimitive ()   static::ExceptionBase & ExcUnitShapeValuesDoNotExist ()   static::ExceptionBase & ExcFEHasNoSupportPoints ()   static::ExceptionBase & ExcEmbeddingVoid ()   static::ExceptionBase & ExcProjectionVoid ()   static::ExceptionBase & ExcWrongInterfaceMatrixSize (int arg1, int arg2)   static::ExceptionBase & ExcInterpolationNotImplemented ()   - Static Public Member Functions inherited from Subscriptor static::ExceptionBase & ExcInUse (int arg1, char *arg2, std::string &arg3)   static::ExceptionBase & ExcNoSubscriber (char *arg1, char *arg2)   Static Public Attributes static const unsigned int space_dimension = spacedim   - Static Public Attributes inherited from FiniteElementData< dim > static const unsigned int dimension = dim   Protected Member Functions void reinit_restriction_and_prolongation_matrices (const bool isotropic_restriction_only=false, const bool isotropic_prolongation_only=false)   TableIndices< 2 > interface_constraints_size () const   virtual UpdateFlags requires_update_flags (const UpdateFlags update_flags) const =0   virtual std::unique_ptr< InternalDataBaseget_data (const UpdateFlags update_flags, const Mapping< dim, spacedim > &mapping, const Quadrature< dim > &quadrature,::internal::FEValuesImplementation::FiniteElementRelatedData< dim, spacedim > &output_data) const =0   virtual std::unique_ptr< InternalDataBaseget_face_data (const UpdateFlags update_flags, const Mapping< dim, spacedim > &mapping, const Quadrature< dim-1 > &quadrature,::internal::FEValuesImplementation::FiniteElementRelatedData< dim, spacedim > &output_data) const   virtual std::unique_ptr< InternalDataBaseget_subface_data (const UpdateFlags update_flags, const Mapping< dim, spacedim > &mapping, const Quadrature< dim-1 > &quadrature,::internal::FEValuesImplementation::FiniteElementRelatedData< dim, spacedim > &output_data) const   virtual void fill_fe_values (const typename Triangulation< dim, spacedim >::cell_iterator &cell, const CellSimilarity::Similarity cell_similarity, const Quadrature< dim > &quadrature, const Mapping< dim, spacedim > &mapping, const typename Mapping< dim, spacedim >::InternalDataBase &mapping_internal, const ::internal::FEValuesImplementation::MappingRelatedData< dim, spacedim > &mapping_data, const InternalDataBase &fe_internal,::internal::FEValuesImplementation::FiniteElementRelatedData< dim, spacedim > &output_data) const =0   virtual void fill_fe_face_values (const typename Triangulation< dim, spacedim >::cell_iterator &cell, const unsigned int face_no, const Quadrature< dim-1 > &quadrature, const Mapping< dim, spacedim > &mapping, const typename Mapping< dim, spacedim >::InternalDataBase &mapping_internal, const ::internal::FEValuesImplementation::MappingRelatedData< dim, spacedim > &mapping_data, const InternalDataBase &fe_internal,::internal::FEValuesImplementation::FiniteElementRelatedData< dim, spacedim > &output_data) const =0   virtual void fill_fe_subface_values (const typename Triangulation< dim, spacedim >::cell_iterator &cell, const unsigned int face_no, const unsigned int sub_no, const Quadrature< dim-1 > &quadrature, const Mapping< dim, spacedim > &mapping, const typename Mapping< dim, spacedim >::InternalDataBase &mapping_internal, const ::internal::FEValuesImplementation::MappingRelatedData< dim, spacedim > &mapping_data, const InternalDataBase &fe_internal,::internal::FEValuesImplementation::FiniteElementRelatedData< dim, spacedim > &output_data) const =0   Static Protected Member Functions static std::vector< unsigned int > compute_n_nonzero_components (const std::vector< ComponentMask > &nonzero_components)   Protected Attributes std::vector< std::vector< FullMatrix< double > > > restriction   std::vector< std::vector< FullMatrix< double > > > prolongation   FullMatrix< double > interface_constraints   std::vector< Point< dim > > unit_support_points   std::vector< Point< dim-1 > > unit_face_support_points   std::vector< Point< dim > > generalized_support_points   std::vector< Point< dim-1 > > generalized_face_support_points   Table< 2, int > adjust_quad_dof_index_for_face_orientation_table   std::vector< int > adjust_line_dof_index_for_line_orientation_table   std::vector< std::pair< unsigned int, unsigned int > > system_to_component_table   std::vector< std::pair< unsigned int, unsigned int > > face_system_to_component_table   std::vector< std::pair< std::pair< unsigned int, unsigned int >, unsigned int > > system_to_base_table   std::vector< std::pair< std::pair< unsigned int, unsigned int >, unsigned int > > face_system_to_base_table   BlockIndices base_to_block_indices   std::vector< std::pair< std::pair< unsigned int, unsigned int >, unsigned int > > component_to_base_table   const std::vector< bool > restriction_is_additive_flags   const std::vector< ComponentMasknonzero_components   const std::vector< unsigned int > n_nonzero_components_table   const bool cached_primitivity   Friends class FEValuesBase< dim, spacedim >   class FEValues< dim, spacedim >   class FEFaceValues< dim, spacedim >   class FESubfaceValues< dim, spacedim >   class FESystem< dim, spacedim >   Additional Inherited Members - Public Types inherited from FiniteElementData< dim > enum  Conformity {   unknown = 0x00, L2 = 0x01, Hcurl = 0x02, Hdiv = 0x04,   H1 = Hcurl | Hdiv, H2 = 0x0e }   - Public Attributes inherited from FiniteElementData< dim > const unsigned int dofs_per_vertex   const unsigned int dofs_per_line   const unsigned int dofs_per_quad   const unsigned int dofs_per_hex   const unsigned int first_line_index   const unsigned int first_quad_index   const unsigned int first_hex_index   const unsigned int first_face_line_index   const unsigned int first_face_quad_index   const unsigned int dofs_per_face   const unsigned int dofs_per_cell   const unsigned int components   const unsigned int degree   const Conformity conforming_space   const BlockIndices block_indices_data   Detailed Description template<int dim, int spacedim = dim> class FiniteElement< dim, spacedim > This is the base class for finite elements in arbitrary dimensions. It declares the interface both in terms of member variables and public member functions through which properties of a concrete implementation of a finite element can be accessed. This interface generally consists of a number of groups of variables and functions that can roughly be delineated as follows: The following sections discuss many of these concepts in more detail, and outline strategies by which concrete implementations of a finite element can provide the details necessary for a complete description of a finite element space. As a general rule, there are three ways by which derived classes provide this information: Nomenclature Finite element classes have to define a large number of different properties describing a finite element space. The following subsections describe some nomenclature that will be used in the documentation below. Components and blocks Vector-valued finite element are elements used for systems of partial differential equations. Oftentimes, they are composed via the FESystem class (which is itself derived from the current class), but there are also non-composed elements that have multiple components (for example the FE_Nedelec and FE_RaviartThomas classes, among others). For any of these vector valued elements, individual shape functions may be nonzero in one or several components of the vector valued function. If the element is primitive, there is indeed a single component with a nonzero entry for each shape function. This component can be determined using the FiniteElement::system_to_component_index() function. On the other hand, if there is at least one shape function that is nonzero in more than one vector component, then we call the entire element "non- primitive". The FiniteElement::get_nonzero_components() can then be used to determine which vector components of a shape function are nonzero. The number of nonzero components of a shape function is returned by FiniteElement::n_components(). Whether a shape function is non-primitive can be queried by FiniteElement::is_primitive(). Oftentimes, one may want to split linear system into blocks so that they reflect the structure of the underlying operator. This is typically not done based on vector components, but based on the use of blocks, and the result is then used to substructure objects of type BlockVector, BlockSparseMatrix, BlockMatrixArray, and so on. If you use non-primitive elements, you cannot determine the block number by FiniteElement::system_to_component_index(). Instead, you can use FiniteElement::system_to_block_index(). The number of blocks of a finite element can be determined by FiniteElement::n_blocks(). To better illustrate these concepts, let's consider the following example of the multi-component system FESystem<dim> fe_basis(FE_Q<dim>(2), dim, FE_Q<dim>(1),1); with dim=2. The resulting finite element has 3 components: two that come from the quadratic element and one from the linear element. If, for example, this system were used to discretize a problem in fluid dynamics then one could think of the first two components representing a vector-valued velocity field whereas the last one corresponds to the scalar pressure field. Without degree-of-freedom (DoF) renumbering this finite element will produce the following distribution of local DoFs: fe_system_example.png DoF indices Using the two functions FiniteElement::system_to_component_index() and FiniteElement::system_to_base_index() one can get the following information for each degree-of-freedom "i": const unsigned int component = fe_basis.system_to_component_index(i).first; const unsigned int within_base = fe_basis.system_to_component_index(i).second; const unsigned int base = fe_basis.system_to_base_index(i).first.first; const unsigned int multiplicity = fe_basis.system_to_base_index(i).first.second; const unsigned int within_base_ = fe_basis.system_to_base_index(i).second; // same as above which will result in: DoF Component Base element Shape function within base Multiplicity 0 0 0 0 0 1 1 0 0 1 2 2 1 0 0 3 0 0 1 0 4 1 0 1 1 5 2 1 1 0 6 0 0 2 0 7 1 0 2 1 8 2 1 2 0 9 0 0 3 0 10 1 0 3 1 11 2 1 3 0 12 0 0 4 0 13 1 0 4 1 14 0 0 5 0 15 1 0 5 1 16 0 0 6 0 17 1 0 6 1 18 0 0 7 0 19 1 0 7 1 20 0 0 8 0 21 1 0 8 1 What we see is the following: there are a total of 22 degrees-of-freedom on this element with components ranging from 0 to 2. Each DoF corresponds to one of the two base elements used to build FESystem : \(\mathbb Q_2\) or \(\mathbb Q_1\). Since FE_Q are primitive elements, we have a total of 9 distinct scalar-valued shape functions for the quadratic element and 4 for the linear element. Finally, for DoFs corresponding to the first base element multiplicity is either zero or one, meaning that we use the same scalar valued \(\mathbb Q_2\) for both \(x\) and \(y\) components of the velocity field \(\mathbb Q_2 \otimes \mathbb Q_2\). For DoFs corresponding to the second base element multiplicity is zero. Support points Finite elements are frequently defined by defining a polynomial space and a set of dual functionals. If these functionals involve point evaluations, then the element is "interpolatory" and it is possible to interpolate an arbitrary (but sufficiently smooth) function onto the finite element space by evaluating it at these points. We call these points "support points". Most finite elements are defined by mapping from the reference cell to a concrete cell. Consequently, the support points are then defined on the reference ("unit") cell, see this glossary entry. The support points on a concrete cell can then be computed by mapping the unit support points, using the Mapping class interface and derived classes, typically via the FEValues class. A typical code snippet to do so would look as follows: Quadrature<dim> dummy_quadrature (fe.get_unit_support_points()); FEValues<dim> fe_values (mapping, fe, dummy_quadrature, fe_values.reinit (cell); Point<dim> mapped_point = fe_values.quadrature_point (i); Alternatively, the points can be transformed one-by-one: const vector<Point<dim> > &unit_points = fe.get_unit_support_points(); Point<dim> mapped_point = mapping.transform_unit_to_real_cell (cell, unit_points[i]); Note Finite elements' implementation of the get_unit_support_points() function returns these points in the same order as shape functions. As a consequence, the quadrature points accessed above are also ordered in this way. The order of shape functions is typically documented in the class documentation of the various finite element classes. Implementing finite element spaces in derived classes The following sections provide some more guidance for implementing concrete finite element spaces in derived classes. This includes information that depends on the dimension for which you want to provide something, followed by a list of tools helping to generate information in concrete cases. It is important to note that there is a number of intermediate classes that can do a lot of what is necessary for a complete description of finite element spaces. For example, the FE_Poly, FE_PolyTensor, and FE_PolyFace classes in essence build a complete finite element space if you only provide them with an abstract description of the polynomial space upon which you want to build an element. Using these intermediate classes typically makes implementing finite element descriptions vastly simpler. As a general rule, if you want to implement an element, you will likely want to look at the implementation of other, similar elements first. Since many of the more complicated pieces of a finite element interface have to do with how they interact with mappings, quadrature, and the FEValues class, you will also want to read through the How Mapping, FiniteElement, and FEValues work together documentation module. Interpolation matrices in one dimension In one space dimension (i.e., for dim==1 and any value of spacedim), finite element classes implementing the interface of the current base class need only set the restriction and prolongation matrices that describe the interpolation of the finite element space on one cell to that of its parent cell, and to that on its children, respectively. The constructor of the current class in one dimension presets the interface_constraints matrix (used to describe hanging node constraints at the interface between cells of different refinement levels) to have size zero because there are no hanging nodes in 1d. Interpolation matrices in two dimensions In addition to the fields discussed above for 1D, a constraint matrix is needed to describe hanging node constraints if the finite element has degrees of freedom located on edges or vertices. These constraints are represented by an \(m\times n\)-matrix interface_constraints, where m is the number of degrees of freedom on the refined side without the corner vertices (those dofs on the middle vertex plus those on the two lines), and n is that of the unrefined side (those dofs on the two vertices plus those on the line). The matrix is thus a rectangular one. The \(m\times n\) size of the interface_constraints matrix can also be accessed through the interface_constraints_size() function. The mapping of the dofs onto the indices of the matrix on the unrefined side is as follows: let \(d_v\) be the number of dofs on a vertex, \(d_l\) that on a line, then \(n=0...d_v-1\) refers to the dofs on vertex zero of the unrefined line, \(n=d_v...2d_v-1\) to those on vertex one, \(n=2d_v...2d_v+d_l-1\) to those on the line. Similarly, \(m=0...d_v-1\) refers to the dofs on the middle vertex of the refined side (vertex one of child line zero, vertex zero of child line one), \(m=d_v...d_v+d_l-1\) refers to the dofs on child line zero, \(m=d_v+d_l...d_v+2d_l-1\) refers to the dofs on child line one. Please note that we do not need to reserve space for the dofs on the end vertices of the refined lines, since these must be mapped one-to-one to the appropriate dofs of the vertices of the unrefined line. Through this construction, the degrees of freedom on the child faces are constrained to the degrees of freedom on the parent face. The information so provided is typically consumed by the DoFTools::make_hanging_node_constraints() function. Note The hanging node constraints described by these matrices are only relevant to the case where the same finite element space is used on neighboring (but differently refined) cells. The case that the finite element spaces on different sides of a face are different, i.e., the \(hp\) case (see hp finite element support) is handled by separate functions. See the FiniteElement::get_face_interpolation_matrix() and FiniteElement::get_subface_interpolation_matrix() functions. Interpolation matrices in three dimensions For the interface constraints, the 3d case is similar to the 2d case. The numbering for the indices \(n\) on the mother face is obvious and keeps to the usual numbering of degrees of freedom on quadrilaterals. The numbering of the degrees of freedom on the interior of the refined faces for the index \(m\) is as follows: let \(d_v\) and \(d_l\) be as above, and \(d_q\) be the number of degrees of freedom per quadrilateral (and therefore per face), then \(m=0...d_v-1\) denote the dofs on the vertex at the center, \(m=d_v...5d_v-1\) for the dofs on the vertices at the center of the bounding lines of the quadrilateral, \(m=5d_v..5d_v+4*d_l-1\) are for the degrees of freedom on the four lines connecting the center vertex to the outer boundary of the mother face, \(m=5d_v+4*d_l...5d_v+4*d_l+8*d_l-1\) for the degrees of freedom on the small lines surrounding the quad, and \(m=5d_v+12*d_l...5d_v+12*d_l+4*d_q-1\) for the dofs on the four child faces. Note the direction of the lines at the boundary of the quads, as shown below. The order of the twelve lines and the four child faces can be extracted from the following sketch, where the overall order of the different dof groups is depicted: * *--15--4--16--* * | | | * 10 19 6 20 12 * | | | * 1--7---0--8---2 * | | | * 9 17 5 18 11 * | | | * *--13--3--14--* * The numbering of vertices and lines, as well as the numbering of children within a line is consistent with the one described in Triangulation. Therefore, this numbering is seen from the outside and inside, respectively, depending on the face. The three-dimensional case has a few pitfalls available for derived classes that want to implement constraint matrices. Consider the following case: * *-------* * / /| * / / | * / / | * *-------* | * | | *-------* * | | / /| * | 1 | / / | * | |/ / | * *-------*-------* | * | | | * * | | | / * | 2 | 3 | / * | | |/ * *-------*-------* * Now assume that we want to refine cell 2. We will end up with two faces with hanging nodes, namely the faces between cells 1 and 2, as well as between cells 2 and 3. Constraints have to be applied to the degrees of freedom on both these faces. The problem is that there is now an edge (the top right one of cell 2) which is part of both faces. The hanging node(s) on this edge are therefore constrained twice, once from both faces. To be meaningful, these constraints of course have to be consistent: both faces have to constrain the hanging nodes on the edge to the same nodes on the coarse edge (and only on the edge, as there can then be no constraints to nodes on the rest of the face), and they have to do so with the same weights. This is sometimes tricky since the nodes on the edge may have different local numbers. For the constraint matrix this means the following: if a degree of freedom on one edge of a face is constrained by some other nodes on the same edge with some weights, then the weights have to be exactly the same as those for constrained nodes on the three other edges with respect to the corresponding nodes on these edges. If this isn't the case, you will get into trouble with the ConstraintMatrix class that is the primary consumer of the constraint information: while that class is able to handle constraints that are entered more than once (as is necessary for the case above), it insists that the weights are exactly the same. Using this scheme, child face degrees of freedom are constrained against parent face degrees of freedom that contain those on the edges of the parent face; it is possible that some of them are in turn constrained themselves, leading to longer chains of constraints that the ConstraintMatrix class will eventually have to sort out. (The constraints described above are used by the DoFTools::make_hanging_node_constraints() function that constructs a ConstraintMatrix object.) However, this is of no concern for the FiniteElement and derived classes since they only act locally on one cell and its immediate neighbor, and do not see the bigger picture. The hp_paper details how such chains are handled in practice. Helper functions Construction of a finite element and computation of the matrices described above is often a tedious task, in particular if it has to be performed for several dimensions. Most of this work can be avoided by using the intermediate classes already mentioned above (e.g., FE_Poly, FE_PolyTensor, etc). Other tasks can be automated by some of the functions in namespace FETools. Computing the correct basis from a set of linearly independent functions First, it may already be difficult to compute the basis of shape functions for arbitrary order and dimension. On the other hand, if the node values are given, then the duality relation between node functionals and basis functions defines the basis. As a result, the shape function space may be defined from a set of linearly independent functions, such that the actual finite element basis is computed from linear combinations of them. The coefficients of these combinations are determined by the duality of node values and form a matrix. Using this matrix allows the construction of the basis of shape functions in two steps. 1. Define the space of shape functions using an arbitrary basis wj and compute the matrix M of node functionals Ni applied to these basis functions, such that its entries are mij = Ni(wj). 2. Compute the basis vj of the finite element shape function space by applying M-1 to the basis wj. The matrix M may be computed using FETools::compute_node_matrix(). This function relies on the existence of generalized_support_points and an implementation of the FiniteElement::interpolate() function with VectorSlice argument. (See the glossary entry on generalized support points for more information.) With this, one can then use the following piece of code in the constructor of a class derived from FiniteElement to compute the \(M\) matrix: this->inverse_node_matrix.reinit(this->dofs_per_cell, this->dofs_per_cell); this->inverse_node_matrix.invert(M); Don't forget to make sure that unit_support_points or generalized_support_points are initialized before this! Computing prolongation matrices Once you have shape functions, you can define matrices that transfer data from one cell to its children or the other way around. This is a common operation in multigrid, of course, but is also used when interpolating the solution from one mesh to another after mesh refinement, as well as in the definition of some error estimators. To define the prolongation matrices, i.e., those matrices that describe the transfer of a finite element field from one cell to its children, implementations of finite elements can either fill the prolongation array by hand, or can call FETools::compute_embedding_matrices(). In the latter case, all that is required is the following piece of code: for (unsigned int c=0; c<GeometryInfo<dim>::max_children_per_cell; ++c) this->prolongation[c].reinit (this->dofs_per_cell, this->dofs_per_cell); As in this example, prolongation is almost always implemented via embedding, i.e., the nodal values of the function on the children may be different from the nodal values of the function on the parent cell, but as a function of \(\mathbf x\in{\mathbb R}^\text{spacedim}\), the finite element field on the child is the same as on the parent. Computing restriction matrices The opposite operation, restricting a finite element function defined on the children to the parent cell is typically implemented by interpolating the finite element function on the children to the nodal values of the parent cell. In deal.II, the restriction operation is implemented as a loop over the children of a cell that each apply a matrix to the vector of unknowns on that child cell (these matrices are stored in restriction and are accessed by get_restriction_matrix()). The operation that then needs to be implemented turns out to be surprisingly difficult to describe, but is instructive to describe because it also defines the meaning of the restriction_is_additive_flags array (accessed via the restriction_is_additive() function). To give a concrete example, assume we use a \(Q_1\) element in 1d, and that on each of the parent and child cells degrees of freedom are (locally and globally) numbered as follows: meshes: *-------* *---*---* local DoF numbers: 0 1 0 1|0 1 global DoF numbers: 0 1 0 1 2 Then we want the restriction operation to take the value of the zeroth DoF on child 0 as the value of the zeroth DoF on the parent, and take the value of the first DoF on child 1 as the value of the first DoF on the parent. Ideally, we would like to write this follows \[ U^\text{coarse}|_\text{parent} = \sum_{\text{child}=0}^1 R_\text{child} U^\text{fine}|_\text{child} \] where \(U^\text{fine}|_\text{child=0}=(U^\text{fine}_0,U^\text{fine}_1)^T\) and \(U^\text{fine}|_\text{child=1}=(U^\text{fine}_1,U^\text{fine}_2)^T\). Writing the requested operation like this would here be possible by choosing \[ R_0 = \left(\begin{matrix}1 & 0 \\ 0 & 0\end{matrix}\right), \qquad\qquad R_1 = \left(\begin{matrix}0 & 0 \\ 0 & 1\end{matrix}\right). \] However, this approach already fails if we go to a \(Q_2\) element with the following degrees of freedom: meshes: *-------* *----*----* local DoF numbers: 0 2 1 0 2 1|0 2 1 global DoF numbers: 0 2 1 0 2 1 4 3 Writing things as the sum over matrix operations as above would not easily work because we have to add nonzero values to \(U^\text{coarse}_2\) twice, once for each child. Consequently, restriction is typically implemented as a concatenation operation. I.e., we first compute the individual restrictions from each child, \[ \tilde U^\text{coarse}_\text{child} = R_\text{child} U^\text{fine}|_\text{child}, \] and then compute the values of \(U^\text{coarse}|_\text{parent}\) with the following code: for (unsigned int child=0; child<cell->n_children(); ++child) for (unsigned int i=0; i<dofs_per_cell; ++i) if (U_tilde_coarse[child][i] != 0) U_coarse_on_parent[i] = U_tilde_coarse[child][i]; In other words, each nonzero element of \(\tilde U^\text{coarse}_\text{child}\) overwrites, rather than adds to the corresponding element of \(U^\text{coarse}|_\text{parent}\). This typically also implies that the restriction matrices from two different cells should agree on a value for coarse degrees of freedom that they both want to touch (otherwise the result would depend on the order in which we loop over children, which would be unreasonable because the order of children is an otherwise arbitrary convention). For example, in the example above, the restriction matrices will be \[ R_0 = \left(\begin{matrix}1 & 0 & 0 \\ 0 & 0 & 0 \\ 0 & 1 & 0 \end{matrix}\right), \qquad\qquad R_1 = \left(\begin{matrix}0 & 0 & 0 \\ 0 & 1 & 0 \\ 1 & 0 & 0 \end{matrix}\right), \] and the compatibility condition is the \(R_{0,21}=R_{1,20}\) because they both indicate that \(U^\text{coarse}|_\text{parent,2}\) should be set to one times \(U^\text{fine}|_\text{child=0,1}\) and \(U^\text{fine}|_\text{child=1,0}\). Unfortunately, not all finite elements allow to write the restriction operation in this way. For example, for the piecewise constant FE_DGQ(0) element, the value of the finite element field on the parent cell can not be determined by interpolation from the children. Rather, the only reasonable choice is to take it as the average value between the children – so we are back to the sum operation, rather than the concatenation. Further thought shows that whether restriction should be additive or not is a property of the individual shape function, not of the finite element as a whole. Consequently, the FiniteElement::restriction_is_additive() function returns whether a particular shape function should act via concatenation (a return value of false) or via addition (return value of true), and the correct code for the overall operation is then as follows (and as, in fact, implemented in DoFAccessor::get_interpolated_dof_values()): for (unsigned int child=0; child<cell->n_children(); ++child) for (unsigned int i=0; i<dofs_per_cell; ++i) if (fe.restriction_is_additive(i) == true) U_coarse_on_parent[i] += U_tilde_coarse[child][i]; else if (U_tilde_coarse[child][i] != 0) U_coarse_on_parent[i] = U_tilde_coarse[child][i]; Computing interface_constraints Constraint matrices can be computed semi-automatically using FETools::compute_face_embedding_matrices(). This function computes the representation of the coarse mesh functions by fine mesh functions for each child of a face separately. These matrices must be convoluted into a single rectangular constraint matrix, eliminating degrees of freedom on common vertices and edges as well as on the coarse grid vertices. See the discussion above for details of this numbering. Author Wolfgang Bangerth, Guido Kanschat, Ralf Hartmann, 1998, 2000, 2001, 2005, 2015 Definition at line 35 of file dof_accessor.h. Constructor & Destructor Documentation template<int dim, int spacedim> FiniteElement< dim, spacedim >::FiniteElement ( const FiniteElementData< dim > &  fe_data, const std::vector< bool > &  restriction_is_additive_flags, const std::vector< ComponentMask > &  nonzero_components  ) Constructor: initialize the fields of this base class of all finite elements. Parameters [in]fe_dataAn object that stores identifying (typically integral) information about the element to be constructed. In particular, this object will contain data such as the number of degrees of freedom per cell (and per vertex, line, etc), the number of vector components, etc. This argument is used to initialize the base class of the current object under construction. [in]restriction_is_additive_flagsA vector of size dofs_per_cell (or of size one, see below) that for each shape function states whether the shape function is additive or not. The meaning of these flags is described in the section on restriction matrices in the general documentation of this class. [in]nonzero_componentsA vector of size dofs_per_cell (or of size one, see below) that for each shape function provides a ComponentMask (of size fe_data.n_components()) that indicates in which vector components this shape function is nonzero (after mapping the shape function to the real cell). For "primitive" shape functions, this component mask will have a single entry (see GlossPrimitive for more information about primitive elements). On the other hand, for elements such as the Raviart-Thomas or Nedelec elements, shape functions are nonzero in more than one vector component (after mapping to the real cell) and the given component mask will contain more than one entry. (For these two elements, all entries will in fact be set, but this would not be the case if you couple a FE_RaviartThomas and a FE_Nedelec together into a FESystem.) Precondition restriction_is_additive_flags.size() == dofs_per_cell, or restriction_is_additive_flags.size() == 1. In the latter case, the array is simply interpreted as having size dofs_per_cell where each element has the same value as the single element given. nonzero_components.size() == dofs_per_cell, or nonzero_components.size() == 1. In the latter case, the array is simply interpreted as having size dofs_per_cell where each element equals the component mask provided in the single element given. Definition at line 55 of file fe.cc. template<int dim, int spacedim = dim> FiniteElement< dim, spacedim >::FiniteElement ( FiniteElement< dim, spacedim > &&  ) default Move constructor. template<int dim, int spacedim = dim> FiniteElement< dim, spacedim >::FiniteElement ( const FiniteElement< dim, spacedim > &  ) default Copy constructor. template<int dim, int spacedim = dim> virtual FiniteElement< dim, spacedim >::~FiniteElement ( ) virtualdefault Virtual destructor. Makes sure that pointers to this class are deleted properly. Member Function Documentation template<int dim, int spacedim> std::pair< std::unique_ptr< FiniteElement< dim, spacedim > >, unsigned int > FiniteElement< dim, spacedim >::operator^ ( const unsigned int  multiplicity) const Creates information for creating a FESystem with this class as base element and with multiplicity multiplicity. In particular, the return type of this function can be used in the constructor for a FESystem object. This function calls clone() and hence creates a copy of the current object. Definition at line 150 of file fe.cc. template<int dim, int spacedim = dim> virtual std::unique_ptr<FiniteElement<dim,spacedim> > FiniteElement< dim, spacedim >::clone ( ) const pure virtual template<int dim, int spacedim = dim> virtual std::string FiniteElement< dim, spacedim >::get_name ( ) const pure virtual Return a string that uniquely identifies a finite element. The general convention is that this is the class name, followed by the dimension in angle brackets, and the polynomial degree and whatever else is necessary in parentheses. For example, FE_Q<2>(3) is the value returned for a cubic element in 2d. Systems of elements have their own naming convention, see the FESystem class. Implemented in FE_Q< dim, spacedim >, FE_Q< dim >, FE_FaceP< 1, spacedim >, FE_Q_Hierarchical< dim >, FE_DGQHermite< dim, spacedim >, FESystem< dim, spacedim >, FE_FaceP< dim, spacedim >, FE_DGQLegendre< dim, spacedim >, FE_DGQArbitraryNodes< dim, spacedim >, FE_DGP< dim, spacedim >, FE_DGPMonomial< dim >, FE_Enriched< dim, spacedim >, FE_DGPNonparametric< dim, spacedim >, FE_P1NC, FE_Q_DG0< dim, spacedim >, FE_RaviartThomasNodal< dim >, FE_FaceQ< 1, spacedim >, FE_DGBDM< dim, spacedim >, FE_Bernstein< dim, spacedim >, FE_DGRaviartThomas< dim, spacedim >, FE_TraceQ< 1, spacedim >, FE_DGNedelec< dim, spacedim >, FE_Nedelec< dim >, FE_Q_iso_Q1< dim, spacedim >, FE_DGQ< dim, spacedim >, FE_RaviartThomas< dim >, FE_ABF< dim >, FE_Nothing< dim, spacedim >, FE_Nothing< dim >, FE_Q_Bubbles< dim, spacedim >, FE_RT_Bubbles< dim >, FE_FaceQ< dim, spacedim >, FE_BDM< dim >, FE_DGVector< PolynomialType, dim, spacedim >, FE_DGVector< PolynomialsRaviartThomas< dim >, dim, spacedim >, FE_DGVector< PolynomialsBDM< dim >, dim, spacedim >, FE_DGVector< PolynomialsNedelec< dim >, dim, spacedim >, FE_RannacherTurek< dim >, and FE_TraceQ< dim, spacedim >. template<int dim, int spacedim> const FiniteElement< dim, spacedim > & FiniteElement< dim, spacedim >::operator[] ( const unsigned int  fe_index) const inline This operator returns a reference to the present object if the argument given equals to zero. While this does not seem particularly useful, it is helpful in writing code that works with both DoFHandler and the hp version hp::DoFHandler, since one can then write code like this: dofs_per_cell = dof_handler->get_fe()[cell->active_fe_index()].dofs_per_cell; This code doesn't work in both situations without the present operator because DoFHandler::get_fe() returns a finite element, whereas hp::DoFHandler::get_fe() returns a collection of finite elements that doesn't offer a dofs_per_cell member variable: one first has to select which finite element to work on, which is done using the operator[]. Fortunately, cell->active_fe_index() also works for non-hp classes and simply returns zero in that case. The present operator[] accepts this zero argument, by returning the finite element with index zero within its collection (that, of course, consists only of the present finite element anyway). Definition at line 2990 of file fe.h. template<int dim, int spacedim> double FiniteElement< dim, spacedim >::shape_value ( const unsigned int  i, const Point< dim > &  p  ) const virtual Return the value of the ith shape function at the point p. p is a point on the reference element. If the finite element is vector-valued, then return the value of the only non-zero component of the vector value of this shape function. If the shape function has more than one non-zero component (which we refer to with the term non-primitive), then derived classes implementing this function should throw an exception of type ExcShapeFunctionNotPrimitive. In that case, use the shape_value_component() function. Implementations of this function should throw an exception of type ExcUnitShapeValuesDoNotExist if the shape functions of the FiniteElement under consideration depend on the shape of the cell in real space, i.e., if the shape functions are not defined by mapping from the reference cell. Some non-conforming elements are defined this way, as is the FE_DGPNonparametric class, to name just one example. The default implementation of this virtual function does exactly this, i.e., it simply throws an exception of type ExcUnitShapeValuesDoNotExist. Reimplemented in FESystem< dim, spacedim >, FE_Enriched< dim, spacedim >, FE_DGPNonparametric< dim, spacedim >, FE_PolyTensor< PolynomialType, dim, spacedim >, FE_PolyTensor< PolynomialsRaviartThomas< dim >, dim, spacedim >, FE_PolyTensor< PolynomialsNedelec< dim >, dim >, FE_PolyTensor< PolynomialsRaviartThomas< dim >, dim >, FE_PolyTensor< PolynomialsBDM< dim >, dim, spacedim >, FE_PolyTensor< PolynomialsBDM< dim >, dim >, FE_PolyTensor< PolynomialsABF< dim >, dim >, FE_PolyTensor< PolynomialsRT_Bubbles< dim >, dim >, FE_PolyTensor< PolynomialsNedelec< dim >, dim, spacedim >, FE_Nothing< dim, spacedim >, FE_Nothing< dim >, FE_Poly< PolynomialType, dim, spacedim >, FE_Poly< PolynomialSpace< dim >, dim, spacedim >, FE_Poly< PolynomialsP< dim >, dim >, FE_Poly< TensorProductPolynomials< dim >, dim >, FE_Poly< TensorProductPolynomials< dim >, dim, spacedim >, FE_Poly< PolynomialsRannacherTurek< dim >, dim >, FE_Poly< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Poly< TensorProductPolynomialsConst< dim >, dim, spacedim >, and FE_Poly< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >. Definition at line 159 of file fe.cc. template<int dim, int spacedim> double FiniteElement< dim, spacedim >::shape_value_component ( const unsigned int  i, const Point< dim > &  p, const unsigned int  component  ) const virtual template<int dim, int spacedim> Tensor< 1, dim > FiniteElement< dim, spacedim >::shape_grad ( const unsigned int  i, const Point< dim > &  p  ) const virtual Return the gradient of the ith shape function at the point p. p is a point on the reference element, and likewise the gradient is the gradient on the unit cell with respect to unit cell coordinates. If the finite element is vector-valued, then return the value of the only non- zero component of the vector value of this shape function. If the shape function has more than one non-zero component (which we refer to with the term non-primitive), then derived classes implementing this function should throw an exception of type ExcShapeFunctionNotPrimitive. In that case, use the shape_grad_component() function. Implementations of this function should throw an exception of type ExcUnitShapeValuesDoNotExist if the shape functions of the FiniteElement under consideration depend on the shape of the cell in real space, i.e., if the shape functions are not defined by mapping from the reference cell. Some non-conforming elements are defined this way, as is the FE_DGPNonparametric class, to name just one example. The default implementation of this virtual function does exactly this, i.e., it simply throws an exception of type ExcUnitShapeValuesDoNotExist. Reimplemented in FESystem< dim, spacedim >, FE_DGPNonparametric< dim, spacedim >, FE_PolyTensor< PolynomialType, dim, spacedim >, FE_PolyTensor< PolynomialsRaviartThomas< dim >, dim, spacedim >, FE_PolyTensor< PolynomialsNedelec< dim >, dim >, FE_PolyTensor< PolynomialsRaviartThomas< dim >, dim >, FE_PolyTensor< PolynomialsBDM< dim >, dim, spacedim >, FE_PolyTensor< PolynomialsBDM< dim >, dim >, FE_PolyTensor< PolynomialsABF< dim >, dim >, FE_PolyTensor< PolynomialsRT_Bubbles< dim >, dim >, FE_PolyTensor< PolynomialsNedelec< dim >, dim, spacedim >, FE_Poly< PolynomialType, dim, spacedim >, FE_Poly< PolynomialSpace< dim >, dim, spacedim >, FE_Poly< PolynomialsP< dim >, dim >, FE_Poly< TensorProductPolynomials< dim >, dim >, FE_Poly< TensorProductPolynomials< dim >, dim, spacedim >, FE_Poly< PolynomialsRannacherTurek< dim >, dim >, FE_Poly< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Poly< TensorProductPolynomialsConst< dim >, dim, spacedim >, and FE_Poly< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >. Definition at line 182 of file fe.cc. template<int dim, int spacedim> Tensor< 1, dim > FiniteElement< dim, spacedim >::shape_grad_component ( const unsigned int  i, const Point< dim > &  p, const unsigned int  component  ) const virtual template<int dim, int spacedim> Tensor< 2, dim > FiniteElement< dim, spacedim >::shape_grad_grad ( const unsigned int  i, const Point< dim > &  p  ) const virtual Return the tensor of second derivatives of the ith shape function at point p on the unit cell. The derivatives are derivatives on the unit cell with respect to unit cell coordinates. If the finite element is vector-valued, then return the value of the only non-zero component of the vector value of this shape function. If the shape function has more than one non-zero component (which we refer to with the term non- primitive), then derived classes implementing this function should throw an exception of type ExcShapeFunctionNotPrimitive. In that case, use the shape_grad_grad_component() function. Implementations of this function should throw an exception of type ExcUnitShapeValuesDoNotExist if the shape functions of the FiniteElement under consideration depend on the shape of the cell in real space, i.e., if the shape functions are not defined by mapping from the reference cell. Some non-conforming elements are defined this way, as is the FE_DGPNonparametric class, to name just one example. The default implementation of this virtual function does exactly this, i.e., it simply throws an exception of type ExcUnitShapeValuesDoNotExist. Reimplemented in FESystem< dim, spacedim >, FE_DGPNonparametric< dim, spacedim >, FE_PolyTensor< PolynomialType, dim, spacedim >, FE_PolyTensor< PolynomialsRaviartThomas< dim >, dim, spacedim >, FE_PolyTensor< PolynomialsNedelec< dim >, dim >, FE_PolyTensor< PolynomialsRaviartThomas< dim >, dim >, FE_PolyTensor< PolynomialsBDM< dim >, dim, spacedim >, FE_PolyTensor< PolynomialsBDM< dim >, dim >, FE_PolyTensor< PolynomialsABF< dim >, dim >, FE_PolyTensor< PolynomialsRT_Bubbles< dim >, dim >, FE_PolyTensor< PolynomialsNedelec< dim >, dim, spacedim >, FE_Poly< PolynomialType, dim, spacedim >, FE_Poly< PolynomialSpace< dim >, dim, spacedim >, FE_Poly< PolynomialsP< dim >, dim >, FE_Poly< TensorProductPolynomials< dim >, dim >, FE_Poly< TensorProductPolynomials< dim >, dim, spacedim >, FE_Poly< PolynomialsRannacherTurek< dim >, dim >, FE_Poly< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Poly< TensorProductPolynomialsConst< dim >, dim, spacedim >, and FE_Poly< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >. Definition at line 205 of file fe.cc. template<int dim, int spacedim> Tensor< 2, dim > FiniteElement< dim, spacedim >::shape_grad_grad_component ( const unsigned int  i, const Point< dim > &  p, const unsigned int  component  ) const virtual template<int dim, int spacedim> Tensor< 3, dim > FiniteElement< dim, spacedim >::shape_3rd_derivative ( const unsigned int  i, const Point< dim > &  p  ) const virtual Return the tensor of third derivatives of the ith shape function at point p on the unit cell. The derivatives are derivatives on the unit cell with respect to unit cell coordinates. If the finite element is vector-valued, then return the value of the only non-zero component of the vector value of this shape function. If the shape function has more than one non-zero component (which we refer to with the term non- primitive), then derived classes implementing this function should throw an exception of type ExcShapeFunctionNotPrimitive. In that case, use the shape_3rd_derivative_component() function. Implementations of this function should throw an exception of type ExcUnitShapeValuesDoNotExist if the shape functions of the FiniteElement under consideration depend on the shape of the cell in real space, i.e., if the shape functions are not defined by mapping from the reference cell. Some non-conforming elements are defined this way, as is the FE_DGPNonparametric class, to name just one example. The default implementation of this virtual function does exactly this, i.e., it simply throws an exception of type ExcUnitShapeValuesDoNotExist. Reimplemented in FESystem< dim, spacedim >, FE_Poly< PolynomialType, dim, spacedim >, FE_Poly< PolynomialSpace< dim >, dim, spacedim >, FE_Poly< PolynomialsP< dim >, dim >, FE_Poly< TensorProductPolynomials< dim >, dim >, FE_Poly< TensorProductPolynomials< dim >, dim, spacedim >, FE_Poly< PolynomialsRannacherTurek< dim >, dim >, FE_Poly< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Poly< TensorProductPolynomialsConst< dim >, dim, spacedim >, and FE_Poly< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >. Definition at line 228 of file fe.cc. template<int dim, int spacedim> Tensor< 3, dim > FiniteElement< dim, spacedim >::shape_3rd_derivative_component ( const unsigned int  i, const Point< dim > &  p, const unsigned int  component  ) const virtual template<int dim, int spacedim> Tensor< 4, dim > FiniteElement< dim, spacedim >::shape_4th_derivative ( const unsigned int  i, const Point< dim > &  p  ) const virtual Return the tensor of fourth derivatives of the ith shape function at point p on the unit cell. The derivatives are derivatives on the unit cell with respect to unit cell coordinates. If the finite element is vector-valued, then return the value of the only non-zero component of the vector value of this shape function. If the shape function has more than one non-zero component (which we refer to with the term non- primitive), then derived classes implementing this function should throw an exception of type ExcShapeFunctionNotPrimitive. In that case, use the shape_4th_derivative_component() function. Implementations of this function should throw an exception of type ExcUnitShapeValuesDoNotExist if the shape functions of the FiniteElement under consideration depend on the shape of the cell in real space, i.e., if the shape functions are not defined by mapping from the reference cell. Some non-conforming elements are defined this way, as is the FE_DGPNonparametric class, to name just one example. The default implementation of this virtual function does exactly this, i.e., it simply throws an exception of type ExcUnitShapeValuesDoNotExist. Reimplemented in FESystem< dim, spacedim >, FE_Poly< PolynomialType, dim, spacedim >, FE_Poly< PolynomialSpace< dim >, dim, spacedim >, FE_Poly< PolynomialsP< dim >, dim >, FE_Poly< TensorProductPolynomials< dim >, dim >, FE_Poly< TensorProductPolynomials< dim >, dim, spacedim >, FE_Poly< PolynomialsRannacherTurek< dim >, dim >, FE_Poly< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Poly< TensorProductPolynomialsConst< dim >, dim, spacedim >, and FE_Poly< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >. Definition at line 251 of file fe.cc. template<int dim, int spacedim> Tensor< 4, dim > FiniteElement< dim, spacedim >::shape_4th_derivative_component ( const unsigned int  i, const Point< dim > &  p, const unsigned int  component  ) const virtual template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::has_support_on_face ( const unsigned int  shape_index, const unsigned int  face_index  ) const virtual This function returns true, if the shape function shape_index has non-zero function values somewhere on the face face_index. The function is typically used to determine whether some matrix elements resulting from face integrals can be assumed to be zero and may therefore be omitted from integration. A default implementation is provided in this base class which always returns true. This is the safe way to go. Reimplemented in FE_Q_Hierarchical< dim >, FE_Q_Hierarchical< dim >, FESystem< dim, spacedim >, FE_Q_Hierarchical< dim >, FE_FaceP< dim, spacedim >, FE_DGPNonparametric< dim, spacedim >, FE_DGP< dim, spacedim >, FE_DGPMonomial< dim >, FE_DGPMonomial< dim >, FE_DGPMonomial< dim >, FE_DGPMonomial< dim >, FE_RaviartThomasNodal< dim >, FE_DGQ< dim, spacedim >, FE_Q_DG0< dim, spacedim >, FE_FaceQ< 1, spacedim >, FE_Q_Bubbles< dim, spacedim >, FE_Nedelec< dim >, FE_RaviartThomas< dim >, FE_ABF< dim >, FE_FaceQ< dim, spacedim >, FE_Q_Base< PolynomialType, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, dim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsConst< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >, FE_TraceQ< dim, spacedim >, FE_DGVector< PolynomialType, dim, spacedim >, FE_DGVector< PolynomialsRaviartThomas< dim >, dim, spacedim >, FE_DGVector< PolynomialsBDM< dim >, dim, spacedim >, and FE_DGVector< PolynomialsNedelec< dim >, dim, spacedim >. Definition at line 1071 of file fe.cc. template<int dim, int spacedim> const FullMatrix< double > & FiniteElement< dim, spacedim >::get_restriction_matrix ( const unsigned int  child, const RefinementCase< dim > &  refinement_case = RefinementCase<dim>::isotropic_refinement  ) const virtual Return the matrix that describes restricting a finite element field from the given child (as obtained by the given refinement_case) to the parent cell. The interpretation of the returned matrix depends on what restriction_is_additive() returns for each shape function. Row and column indices are related to coarse grid and fine grid spaces, respectively, consistent with the definition of the associated operator. If projection matrices are not implemented in the derived finite element class, this function aborts with an exception of type FiniteElement::ExcProjectionVoid. You can check whether this would happen by first calling the restriction_is_implemented() or the isotropic_restriction_is_implemented() function. Reimplemented in FESystem< dim, spacedim >, FE_Enriched< dim, spacedim >, FE_Nedelec< dim >, FE_DGQ< dim, spacedim >, FE_Q_Bubbles< dim, spacedim >, FE_Q_Base< PolynomialType, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, dim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsConst< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >, and FE_Bernstein< dim, spacedim >. Definition at line 301 of file fe.cc. template<int dim, int spacedim> const FullMatrix< double > & FiniteElement< dim, spacedim >::get_prolongation_matrix ( const unsigned int  child, const RefinementCase< dim > &  refinement_case = RefinementCase<dim>::isotropic_refinement  ) const virtual Prolongation/embedding matrix between grids. The identity operator from a coarse grid space into a fine grid space (where both spaces are identified as functions defined on the parent and child cells) is associated with a matrix P that maps the corresponding representations of these functions in terms of their nodal values. The restriction of this matrix P_i to a single child cell is returned here. The matrix P is the concatenation, not the sum of the cell matrices P_i. That is, if the same non-zero entry j,k exists in two different child matrices P_i, the value should be the same in both matrices and it is copied into the matrix P only once. Row and column indices are related to fine grid and coarse grid spaces, respectively, consistent with the definition of the associated operator. These matrices are used by routines assembling the prolongation matrix for multi-level methods. Upon assembling the transfer matrix between cells using this matrix array, zero elements in the prolongation matrix are discarded and will not fill up the transfer matrix. If prolongation matrices are not implemented in the derived finite element class, this function aborts with an exception of type FiniteElement::ExcEmbeddingVoid. You can check whether this would happen by first calling the prolongation_is_implemented() or the isotropic_prolongation_is_implemented() function. Reimplemented in FESystem< dim, spacedim >, FE_Q_Hierarchical< dim >, FE_Enriched< dim, spacedim >, FE_Nedelec< dim >, FE_DGQ< dim, spacedim >, FE_Q_Base< PolynomialType, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, dim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsConst< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >, FE_Q_Bubbles< dim, spacedim >, and FE_Bernstein< dim, spacedim >. Definition at line 321 of file fe.cc. template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::prolongation_is_implemented ( ) const Return whether this element implements its prolongation matrices. The return value also indicates whether a call to the get_prolongation_matrix() function will generate an error or not. Note, that this function returns true only if the prolongation matrices of the isotropic and all anisotropic refinement cases are implemented. If you are interested in the prolongation matrices for isotropic refinement only, use the isotropic_prolongation_is_implemented function instead. This function is mostly here in order to allow us to write more efficient test programs which we run on all kinds of weird elements, and for which we simply need to exclude certain tests in case something is not implemented. It will in general probably not be a great help in applications, since there is not much one can do if one needs these features and they are not implemented. This function could be used to check whether a call to get_prolongation_matrix() will succeed; however, one then still needs to cope with the lack of information this just expresses. Definition at line 668 of file fe.cc. template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::isotropic_prolongation_is_implemented ( ) const Return whether this element implements its prolongation matrices for isotropic children. The return value also indicates whether a call to the get_prolongation_matrix function will generate an error or not. This function is mostly here in order to allow us to write more efficient test programs which we run on all kinds of weird elements, and for which we simply need to exclude certain tests in case something is not implemented. It will in general probably not be a great help in applications, since there is not much one can do if one needs these features and they are not implemented. This function could be used to check whether a call to get_prolongation_matrix() will succeed; however, one then still needs to cope with the lack of information this just expresses. Definition at line 720 of file fe.cc. template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::restriction_is_implemented ( ) const Return whether this element implements its restriction matrices. The return value also indicates whether a call to the get_restriction_matrix() function will generate an error or not. Note, that this function returns true only if the restriction matrices of the isotropic and all anisotropic refinement cases are implemented. If you are interested in the restriction matrices for isotropic refinement only, use the isotropic_restriction_is_implemented() function instead. This function is mostly here in order to allow us to write more efficient test programs which we run on all kinds of weird elements, and for which we simply need to exclude certain tests in case something is not implemented. It will in general probably not be a great help in applications, since there is not much one can do if one needs these features and they are not implemented. This function could be used to check whether a call to get_restriction_matrix() will succeed; however, one then still needs to cope with the lack of information this just expresses. Definition at line 694 of file fe.cc. template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::isotropic_restriction_is_implemented ( ) const Return whether this element implements its restriction matrices for isotropic children. The return value also indicates whether a call to the get_restriction_matrix() function will generate an error or not. This function is mostly here in order to allow us to write more efficient test programs which we run on all kinds of weird elements, and for which we simply need to exclude certain tests in case something is not implemented. It will in general probably not be a great help in applications, since there is not much one can do if one needs these features and they are not implemented. This function could be used to check whether a call to get_restriction_matrix() will succeed; however, one then still needs to cope with the lack of information this just expresses. Definition at line 746 of file fe.cc. template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::restriction_is_additive ( const unsigned int  index) const inline Access the restriction_is_additive_flags field. See the discussion about restriction matrices in the general class documentation for more information. The index must be between zero and the number of shape functions of this element. Definition at line 3165 of file fe.h. template<int dim, int spacedim = dim> const FullMatrix< double > & FiniteElement< dim, spacedim >::constraints ( const ::internal::SubfaceCase< dim > &  subface_case = ::internal::SubfaceCase<dim>::case_isotropic) const Return a read only reference to the matrix that describes the constraints at the interface between a refined and an unrefined cell. Some finite elements do not (yet) implement hanging node constraints. If this is the case, then this function will generate an exception, since no useful return value can be generated. If you should have a way to live with this, then you might want to use the constraints_are_implemented() function to check up front whether this function will succeed or generate the exception. Definition at line 793 of file fe.cc. template<int dim, int spacedim = dim> bool FiniteElement< dim, spacedim >::constraints_are_implemented ( const ::internal::SubfaceCase< dim > &  subface_case = ::internal::SubfaceCase<dim>::case_isotropic) const Return whether this element implements its hanging node constraints. The return value also indicates whether a call to the constraints() function will generate an error or not. This function is mostly here in order to allow us to write more efficient test programs which we run on all kinds of weird elements, and for which we simply need to exclude certain tests in case hanging node constraints are not implemented. It will in general probably not be a great help in applications, since there is not much one can do if one needs hanging node constraints and they are not implemented. This function could be used to check whether a call to constraints() will succeed; however, one then still needs to cope with the lack of information this just expresses. Definition at line 772 of file fe.cc. template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::hp_constraints_are_implemented ( ) const virtual Return whether this element implements its hanging node constraints in the new way, which has to be used to make elements "hp compatible". That means, the element properly implements the get_face_interpolation_matrix and get_subface_interpolation_matrix methods. Therefore the return value also indicates whether a call to the get_face_interpolation_matrix() method and the get_subface_interpolation_matrix() method will generate an error or not. Currently the main purpose of this function is to allow the make_hanging_node_constraints method to decide whether the new procedures, which are supposed to work in the hp framework can be used, or if the old well verified but not hp capable functions should be used. Once the transition to the new scheme for computing the interface constraints is complete, this function will be superfluous and will probably go away. Derived classes should implement this function accordingly. The default assumption is that a finite element does not provide hp capable face interpolation, and the default implementation therefore returns false. Reimplemented in FESystem< dim, spacedim >, FE_Q_Hierarchical< dim >, FE_FaceP< dim, spacedim >, FE_DGPNonparametric< dim, spacedim >, FE_DGP< dim, spacedim >, FE_DGPMonomial< dim >, FE_Enriched< dim, spacedim >, FE_RaviartThomasNodal< dim >, FE_FaceQ< 1, spacedim >, FE_DGQ< dim, spacedim >, FE_Q_Base< PolynomialType, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, dim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsConst< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >, FE_Nothing< dim, spacedim >, FE_Nothing< dim >, FE_FaceQ< dim, spacedim >, FE_Nedelec< dim >, FE_Bernstein< dim, spacedim >, and FE_TraceQ< dim, spacedim >. Definition at line 784 of file fe.cc. template<int dim, int spacedim> void FiniteElement< dim, spacedim >::get_interpolation_matrix ( const FiniteElement< dim, spacedim > &  source, FullMatrix< double > &  matrix  ) const virtual Return the matrix interpolating from the given finite element to the present one. The size of the matrix is then dofs_per_cell times source.dofs_per_cell. Derived elements will have to implement this function. They may only provide interpolation matrices for certain source finite elements, for example those from the same family. If they don't implement interpolation from a given element, then they must throw an exception of type ExcInterpolationNotImplemented. Reimplemented in FESystem< dim, spacedim >, FE_Q_DG0< dim, spacedim >, FE_Nothing< dim, spacedim >, FE_DGQ< dim, spacedim >, FE_Q_Bubbles< dim, spacedim >, FE_Bernstein< dim, spacedim >, FE_Q_Base< PolynomialType, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, dim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsConst< dim >, dim, spacedim >, and FE_Q_Base< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >. Definition at line 846 of file fe.cc. template<int dim, int spacedim> void FiniteElement< dim, spacedim >::get_face_interpolation_matrix ( const FiniteElement< dim, spacedim > &  source, FullMatrix< double > &  matrix  ) const virtual Return the matrix interpolating from a face of one element to the face of the neighboring element. The size of the matrix is then source.dofs_per_face times this->dofs_per_face. Derived elements will have to implement this function. They may only provide interpolation matrices for certain source finite elements, for example those from the same family. If they don't implement interpolation from a given element, then they must throw an exception of type ExcInterpolationNotImplemented. Reimplemented in FESystem< dim, spacedim >, FE_FaceP< dim, spacedim >, FE_DGP< dim, spacedim >, FE_DGPNonparametric< dim, spacedim >, FE_Enriched< dim, spacedim >, FE_Nothing< dim, spacedim >, FE_DGQ< dim, spacedim >, FE_Bernstein< dim, spacedim >, FE_TraceQ< dim, spacedim >, FE_FaceQ< dim, spacedim >, FE_Q_Base< PolynomialType, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, dim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsConst< dim >, dim, spacedim >, and FE_Q_Base< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >. Definition at line 862 of file fe.cc. template<int dim, int spacedim> void FiniteElement< dim, spacedim >::get_subface_interpolation_matrix ( const FiniteElement< dim, spacedim > &  source, const unsigned int  subface, FullMatrix< double > &  matrix  ) const virtual Return the matrix interpolating from a face of one element to the subface of the neighboring element. The size of the matrix is then source.dofs_per_face times this->dofs_per_face. Derived elements will have to implement this function. They may only provide interpolation matrices for certain source finite elements, for example those from the same family. If they don't implement interpolation from a given element, then they must throw an exception of type ExcInterpolationNotImplemented. Reimplemented in FESystem< dim, spacedim >, FE_FaceP< dim, spacedim >, FE_DGP< dim, spacedim >, FE_DGPNonparametric< dim, spacedim >, FE_Enriched< dim, spacedim >, FE_Nothing< dim, spacedim >, FE_DGQ< dim, spacedim >, FE_Bernstein< dim, spacedim >, FE_TraceQ< dim, spacedim >, FE_FaceQ< dim, spacedim >, FE_Q_Base< PolynomialType, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, dim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsConst< dim >, dim, spacedim >, and FE_Q_Base< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >. Definition at line 878 of file fe.cc. template<int dim, int spacedim> std::vector< std::pair< unsigned int, unsigned int > > FiniteElement< dim, spacedim >::hp_vertex_dof_identities ( const FiniteElement< dim, spacedim > &  fe_other) const virtual If, on a vertex, several finite elements are active, the hp code first assigns the degrees of freedom of each of these FEs different global indices. It then calls this function to find out which of them should get identical values, and consequently can receive the same global DoF index. This function therefore returns a list of identities between DoFs of the present finite element object with the DoFs of fe_other, which is a reference to a finite element object representing one of the other finite elements active on this particular vertex. The function computes which of the degrees of freedom of the two finite element objects are equivalent, both numbered between zero and the corresponding value of dofs_per_vertex of the two finite elements. The first index of each pair denotes one of the vertex dofs of the present element, whereas the second is the corresponding index of the other finite element. Reimplemented in FESystem< dim, spacedim >, FE_DGPNonparametric< dim, spacedim >, FE_Enriched< dim, spacedim >, FE_DGP< dim, spacedim >, FE_DGQ< dim, spacedim >, FE_Q_Base< PolynomialType, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, dim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsConst< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >, FE_Nothing< dim, spacedim >, FE_Bernstein< dim, spacedim >, and FE_FaceQ< dim, spacedim >. Definition at line 895 of file fe.cc. template<int dim, int spacedim> std::vector< std::pair< unsigned int, unsigned int > > FiniteElement< dim, spacedim >::hp_line_dof_identities ( const FiniteElement< dim, spacedim > &  fe_other) const virtual template<int dim, int spacedim> std::vector< std::pair< unsigned int, unsigned int > > FiniteElement< dim, spacedim >::hp_quad_dof_identities ( const FiniteElement< dim, spacedim > &  fe_other) const virtual template<int dim, int spacedim> FiniteElementDomination::Domination FiniteElement< dim, spacedim >::compare_for_face_domination ( const FiniteElement< dim, spacedim > &  fe_other) const virtual template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::operator== ( const FiniteElement< dim, spacedim > &  f) const Comparison operator. We also check for equality of the constraint matrix, which is quite an expensive operation. Do therefore use this function with care, if possible only for debugging purposes. Since this function is not that important, we avoid an implementational question about comparing arrays and do not compare the matrix arrays restriction and prolongation. Definition at line 938 of file fe.cc. template<int dim, int spacedim> std::pair< unsigned int, unsigned int > FiniteElement< dim, spacedim >::system_to_component_index ( const unsigned int  index) const inline Compute vector component and index of this shape function within the shape functions corresponding to this component from the index of a shape function within this finite element. If the element is scalar, then the component is always zero, and the index within this component is equal to the overall index. If the shape function referenced has more than one non-zero component, then it cannot be associated with one vector component, and an exception of type ExcShapeFunctionNotPrimitive will be raised. Note that if the element is composed of other (base) elements, and a base element has more than one component but all its shape functions are primitive (i.e. are non-zero in only one component), then this mapping contains valid information. However, the index of a shape function of this element within one component (i.e. the second number of the respective entry of this array) does not indicate the index of the respective shape function within the base element (since that has more than one vector-component). For this information, refer to the system_to_base_table field and the system_to_base_index() function. See the class description above for an example of how this function is typically used. The use of this function is explained extensively in the step-8 and step-20 tutorial programs as well as in the Handling vector valued problems module. Definition at line 3003 of file fe.h. template<int dim, int spacedim> unsigned int FiniteElement< dim, spacedim >::component_to_system_index ( const unsigned int  component, const unsigned int  index  ) const inline Compute the shape function for the given vector component and index. If the element is scalar, then the component must be zero, and the index within this component is equal to the overall index. This is the opposite operation from the system_to_component_index() function. Definition at line 3037 of file fe.h. template<int dim, int spacedim> std::pair< unsigned int, unsigned int > FiniteElement< dim, spacedim >::face_system_to_component_index ( const unsigned int  index) const inline Same as system_to_component_index(), but do it for shape functions and their indices on a face. The range of allowed indices is therefore 0..dofs_per_face. You will rarely need this function in application programs, since almost all application codes only need to deal with cell indices, not face indices. The function is mainly there for use inside the library. Definition at line 3059 of file fe.h. template<int dim, int spacedim> unsigned int FiniteElement< dim, spacedim >::adjust_quad_dof_index_for_face_orientation ( const unsigned int  index, const bool  face_orientation, const bool  face_flip, const bool  face_rotation  ) const For faces with non-standard face_orientation in 3D, the dofs on faces (quads) have to be permuted in order to be combined with the correct shape functions. Given a local dof index on a quad, return the local index, if the face has non-standard face_orientation, face_flip or face_rotation. In 2D and 1D there is no need for permutation and consequently an exception is thrown. Definition at line 613 of file fe.cc. template<int dim, int spacedim> unsigned int FiniteElement< dim, spacedim >::face_to_cell_index ( const unsigned int  face_dof_index, const unsigned int  face, const bool  face_orientation = true, const bool  face_flip = false, const bool  face_rotation = false  ) const virtual Given an index in the natural ordering of indices on a face, return the index of the same degree of freedom on the cell. To explain the concept, consider the case where we would like to know whether a degree of freedom on a face, for example as part of an FESystem element, is primitive. Unfortunately, the is_primitive() function in the FiniteElement class takes a cell index, so we would need to find the cell index of the shape function that corresponds to the present face index. This function does that. Code implementing this would then look like this: for (i=0; i<dofs_per_face; ++i) if (fe.is_primitive(fe.face_to_cell_index(i, some_face_no))) ... do whatever The function takes additional arguments that account for the fact that actual faces can be in their standard ordering with respect to the cell under consideration, or can be flipped, oriented, etc. Parameters face_dof_indexThe index of the degree of freedom on a face. This index must be between zero and dofs_per_face. faceThe number of the face this degree of freedom lives on. This number must be between zero and GeometryInfo::faces_per_cell. face_orientationOne part of the description of the orientation of the face. See GlossFaceOrientation. face_flipOne part of the description of the orientation of the face. See GlossFaceOrientation. face_rotationOne part of the description of the orientation of the face. See GlossFaceOrientation. Returns The index of this degree of freedom within the set of degrees of freedom on the entire cell. The returned value will be between zero and dofs_per_cell. Note This function exists in this class because that is where it was first implemented. However, it can't really work in the most general case without knowing what element we have. The reason is that when a face is flipped or rotated, we also need to know whether we need to swap the degrees of freedom on this face, or whether they are immune from this. For this, consider the situation of a \(Q_3\) element in 2d. If face_flip is true, then we need to consider the two degrees of freedom on the edge in reverse order. On the other hand, if the element were a \(Q_1^2\), then because the two degrees of freedom on this edge belong to different vector components, they should not be considered in reverse order. What all of this shows is that the function can't work if there are more than one degree of freedom per line or quad, and that in these cases the function will throw an exception pointing out that this functionality will need to be provided by a derived class that knows what degrees of freedom actually represent. Reimplemented in FESystem< dim, spacedim >, FE_Q_Base< PolynomialType, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, dim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsConst< dim >, dim, spacedim >, and FE_Q_Base< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >. Definition at line 524 of file fe.cc. template<int dim, int spacedim> unsigned int FiniteElement< dim, spacedim >::adjust_line_dof_index_for_line_orientation ( const unsigned int  index, const bool  line_orientation  ) const For lines with non-standard line_orientation in 3D, the dofs on lines have to be permuted in order to be combined with the correct shape functions. Given a local dof index on a line, return the local index, if the line has non-standard line_orientation. In 2D and 1D there is no need for permutation, so the given index is simply returned. Definition at line 645 of file fe.cc. template<int dim, int spacedim> const ComponentMask & FiniteElement< dim, spacedim >::get_nonzero_components ( const unsigned int  i) const inline Return in which of the vector components of this finite element the ith shape function is non-zero. The length of the returned array is equal to the number of vector components of this element. For most finite element spaces, the result of this function will be a vector with exactly one element being true, since for most spaces the individual vector components are independent. In that case, the component with the single zero is also the first element of what system_to_component_index() returns. Only for those spaces that couple the components, for example to make a shape function divergence free, will there be more than one true entry. Elements for which this is true are called non-primitive (see GlossPrimitive). Definition at line 3177 of file fe.h. template<int dim, int spacedim> unsigned int FiniteElement< dim, spacedim >::n_nonzero_components ( const unsigned int  i) const inline Return in how many vector components the ith shape function is non- zero. This value equals the number of entries equal to true in the result of the get_nonzero_components() function. For most finite element spaces, the result will be equal to one. It is not equal to one only for those ansatz spaces for which vector-valued shape functions couple the individual components, for example in order to make them divergence-free. Definition at line 3188 of file fe.h. template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::is_primitive ( ) const inline Return whether the entire finite element is primitive, in the sense that all its shape functions are primitive. If the finite element is scalar, then this is always the case. Since this is an extremely common operation, the result is cached and returned by this function. Definition at line 3199 of file fe.h. template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::is_primitive ( const unsigned int  i) const inline Return whether the ith shape function is primitive in the sense that the shape function is non-zero in only one vector component. Non- primitive shape functions would then, for example, be those of divergence free ansatz spaces, in which the individual vector components are coupled. The result of the function is true if and only if the result of n_nonzero_components(i) is equal to one. Definition at line 3209 of file fe.h. template<int dim, int spacedim> unsigned int FiniteElement< dim, spacedim >::n_base_elements ( ) const inline Number of base elements in a mixed discretization. Note that even for vector valued finite elements, the number of components needs not coincide with the number of base elements, since they may be reused. For example, if you create a FESystem with three identical finite element classes by using the constructor that takes one finite element and a multiplicity, then the number of base elements is still one, although the number of components of the finite element is equal to the multiplicity. Definition at line 3017 of file fe.h. template<int dim, int spacedim> const FiniteElement< dim, spacedim > & FiniteElement< dim, spacedim >::base_element ( const unsigned int  index) const virtual Access to base element objects. If the element is atomic, then base_element(0) is this. Reimplemented in FESystem< dim, spacedim >, and FE_Enriched< dim, spacedim >. Definition at line 1225 of file fe.cc. template<int dim, int spacedim> unsigned int FiniteElement< dim, spacedim >::element_multiplicity ( const unsigned int  index) const inline This index denotes how often the base element index is used in a composed element. If the element is atomic, then the result is always equal to one. See the documentation for the n_base_elements() function for more details. Definition at line 3027 of file fe.h. template<int dim, int spacedim> const FiniteElement< dim, spacedim > & FiniteElement< dim, spacedim >::get_sub_fe ( const ComponentMask mask) const Return a reference to a contained finite element that matches the components selected by the given ComponentMask mask. For an arbitrarily nested FESystem, this function returns the inner-most FiniteElement that matches the given mask. The method fails if the mask does not exactly match one of the contained finite elements. It is most useful if the current object is an FESystem, as the return value can only be this in all other cases. Note that the returned object can be an FESystem if the mask matches it but not any of the contained objects. Let us illustrate the function with the an FESystem fe with 7 components: FESystem<2> fe_velocity(FE_Q<2>(2), 2); FE_Q<2> fe_pressure(1); FE_DGP<2> fe_dg(0); FE_BDM<2> fe_nonprim(1); FESystem<2> fe(fe_velocity, 1, fe_pressure, 1, fe_dg, 2, fe_nonprim, 1); The following table lists all possible component masks you can use: ComponentMask Result Description [true,true,true,true,true,true,true] FESystem<2>[FESystem<2>[FE_Q<2>(2)^2]-FE_Q<2>(1)-FE_DGP<2>(0)^2-FE_BDM<2>(1)] fe itself, the whole FESystem [true,true,false,false,false,false,false] FESystem<2>[FE_Q<2>(2)^2] just the fe_velocity [true,false,false,false,false,false,false] FE_Q<2>(2) The first component in fe_velocity [false,true,false,false,false,false,false] FE_Q<2>(2) The second component in fe_velocity [false,false,true,false,false,false,false] FE_Q<2>(1) fe_pressure [false,false,false,true,false,false,false] FE_DGP<2>(0) first copy of fe_dg [false,false,false,false,true,false,false] FE_DGP<2>(0) second copy of fe_dg [false,false,false,false,false,true,true] FE_BDM<2>(1) both components of fe_nonprim Definition at line 1083 of file fe.cc. template<int dim, int spacedim> const FiniteElement< dim, spacedim > & FiniteElement< dim, spacedim >::get_sub_fe ( const unsigned int  first_component, const unsigned int  n_selected_components  ) const virtual Return a reference to a contained finite element that matches the components n_selected_components components starting at component with index first_component. See the other get_sub_fe() function above for more details. Reimplemented in FESystem< dim, spacedim >. Definition at line 1116 of file fe.cc. template<int dim, int spacedim> std::pair< std::pair< unsigned int, unsigned int >, unsigned int > FiniteElement< dim, spacedim >::system_to_base_index ( const unsigned int  index) const inline Return for shape function index the base element it belongs to, the number of the copy of this base element (which is between zero and the multiplicity of this element), and the index of this shape function within this base element. If the element is not composed of others, then base and instance are always zero, and the index is equal to the number of the shape function. If the element is composed of single instances of other elements (i.e. all with multiplicity one) all of which are scalar, then base values and dof indices within this element are equal to the system_to_component_table. It differs only in case the element is composed of other elements and at least one of them is vector-valued itself. See the class documentation above for an example of how this function is typically used. This function returns valid values also in the case of vector-valued (i.e. non-primitive) shape functions, in contrast to the system_to_component_index() function. Definition at line 3089 of file fe.h. template<int dim, int spacedim> std::pair< std::pair< unsigned int, unsigned int >, unsigned int > FiniteElement< dim, spacedim >::face_system_to_base_index ( const unsigned int  index) const inline Same as system_to_base_index(), but for degrees of freedom located on a face. The range of allowed indices is therefore 0..dofs_per_face. You will rarely need this function in application programs, since almost all application codes only need to deal with cell indices, not face indices. The function is mainly there for use inside the library. Definition at line 3102 of file fe.h. template<int dim, int spacedim> types::global_dof_index FiniteElement< dim, spacedim >::first_block_of_base ( const unsigned int  b) const inline Given a base element number, return the first block of a BlockVector it would generate. Definition at line 3114 of file fe.h. template<int dim, int spacedim> std::pair< unsigned int, unsigned int > FiniteElement< dim, spacedim >::component_to_base_index ( const unsigned int  component) const inline For each vector component, return which base element implements this component and which vector component in this base element this is. This information is only of interest for vector-valued finite elements which are composed of several sub-elements. In that case, one may want to obtain information about the element implementing a certain vector component, which can be done using this function and the FESystem::base_element() function. If this is a scalar finite element, then the return value is always equal to a pair of zeros. Definition at line 3124 of file fe.h. template<int dim, int spacedim> std::pair< unsigned int, unsigned int > FiniteElement< dim, spacedim >::block_to_base_index ( const unsigned int  block) const inline Return the base element for this block and the number of the copy of the base element. Definition at line 3137 of file fe.h. template<int dim, int spacedim> std::pair< unsigned int, types::global_dof_index > FiniteElement< dim, spacedim >::system_to_block_index ( const unsigned int  component) const inline The vector block and the index inside the block for this shape function. Definition at line 3147 of file fe.h. template<int dim, int spacedim> unsigned int FiniteElement< dim, spacedim >::component_to_block_index ( const unsigned int  component) const The vector block for this component. Definition at line 343 of file fe.cc. template<int dim, int spacedim> ComponentMask FiniteElement< dim, spacedim >::component_mask ( const FEValuesExtractors::Scalar scalar) const Return a component mask with as many elements as this object has vector components and of which exactly the one component is true that corresponds to the given argument. See the glossary for more information. Parameters scalarAn object that represents a single scalar vector component of this finite element. Returns A component mask that is false in all components except for the one that corresponds to the argument. Definition at line 356 of file fe.cc. template<int dim, int spacedim> ComponentMask FiniteElement< dim, spacedim >::component_mask ( const FEValuesExtractors::Vector vector) const Return a component mask with as many elements as this object has vector components and of which exactly the dim components are true that correspond to the given argument. See the glossary for more information. Parameters vectorAn object that represents dim vector components of this finite element. Returns A component mask that is false in all components except for the ones that corresponds to the argument. Definition at line 374 of file fe.cc. template<int dim, int spacedim> ComponentMask FiniteElement< dim, spacedim >::component_mask ( const FEValuesExtractors::SymmetricTensor< 2 > &  sym_tensor) const Return a component mask with as many elements as this object has vector components and of which exactly the dim*(dim+1)/2 components are true that correspond to the given argument. See the glossary for more information. Parameters sym_tensorAn object that represents dim*(dim+1)/2 components of this finite element that are jointly to be interpreted as forming a symmetric tensor. Returns A component mask that is false in all components except for the ones that corresponds to the argument. Definition at line 393 of file fe.cc. template<int dim, int spacedim> ComponentMask FiniteElement< dim, spacedim >::component_mask ( const BlockMask block_mask) const Given a block mask (see this glossary entry), produce a component mask (see this glossary entry) that represents the components that correspond to the blocks selected in the input argument. This is essentially a conversion operator from BlockMask to ComponentMask. Parameters block_maskThe mask that selects individual blocks of the finite element Returns A mask that selects those components corresponding to the selected blocks of the input argument. Definition at line 416 of file fe.cc. template<int dim, int spacedim> BlockMask FiniteElement< dim, spacedim >::block_mask ( const FEValuesExtractors::Scalar scalar) const Return a block mask with as many elements as this object has blocks and of which exactly the one component is true that corresponds to the given argument. See the glossary for more information. Note This function will only succeed if the scalar referenced by the argument encompasses a complete block. In other words, if, for example, you pass an extractor for the single \(x\) velocity and this object represents an FE_RaviartThomas object, then the single scalar object you selected is part of a larger block and consequently there is no block mask that would represent it. The function will then produce an exception. Parameters scalarAn object that represents a single scalar vector component of this finite element. Returns A component mask that is false in all components except for the one that corresponds to the argument. Definition at line 438 of file fe.cc. template<int dim, int spacedim> BlockMask FiniteElement< dim, spacedim >::block_mask ( const FEValuesExtractors::Vector vector) const Return a component mask with as many elements as this object has vector components and of which exactly the dim components are true that correspond to the given argument. See the glossary for more information. Note The same caveat applies as to the version of the function above: The extractor object passed as argument must be so that it corresponds to full blocks and does not split blocks of this element. Parameters vectorAn object that represents dim vector components of this finite element. Returns A component mask that is false in all components except for the ones that corresponds to the argument. Definition at line 449 of file fe.cc. template<int dim, int spacedim> BlockMask FiniteElement< dim, spacedim >::block_mask ( const FEValuesExtractors::SymmetricTensor< 2 > &  sym_tensor) const Return a component mask with as many elements as this object has vector components and of which exactly the dim*(dim+1)/2 components are true that correspond to the given argument. See the glossary for more information. Note The same caveat applies as to the version of the function above: The extractor object passed as argument must be so that it corresponds to full blocks and does not split blocks of this element. Parameters sym_tensorAn object that represents dim*(dim+1)/2 components of this finite element that are jointly to be interpreted as forming a symmetric tensor. Returns A component mask that is false in all components except for the ones that corresponds to the argument. Definition at line 460 of file fe.cc. template<int dim, int spacedim> BlockMask FiniteElement< dim, spacedim >::block_mask ( const ComponentMask component_mask) const Given a component mask (see this glossary entry), produce a block mask (see this glossary entry) that represents the blocks that correspond to the components selected in the input argument. This is essentially a conversion operator from ComponentMask to BlockMask. Note This function will only succeed if the components referenced by the argument encompasses complete blocks. In other words, if, for example, you pass an component mask for the single \(x\) velocity and this object represents an FE_RaviartThomas object, then the single component you selected is part of a larger block and consequently there is no block mask that would represent it. The function will then produce an exception. Parameters component_maskThe mask that selects individual components of the finite element Returns A mask that selects those blocks corresponding to the selected blocks of the input argument. Definition at line 472 of file fe.cc. template<int dim, int spacedim> std::pair< Table< 2, bool >, std::vector< unsigned int > > FiniteElement< dim, spacedim >::get_constant_modes ( ) const virtual Return a list of constant modes of the element. The number of rows in the resulting table depends on the elements in use. For standard elements, the table has as many rows as there are components in the element and dofs_per_cell columns. To each component of the finite element, the row in the returned table contains a basis representation of the constant function 1 on the element. However, there are some scalar elements where there is more than one constant mode, e.g. the element FE_Q_DG0. In order to match the constant modes to the actual components in the element, the returned data structure also returns a vector with as many components as there are constant modes on the element that contains the component number. Reimplemented in FESystem< dim, spacedim >, FE_Q_Hierarchical< dim >, FE_FaceP< dim, spacedim >, FE_DGQLegendre< dim, spacedim >, FE_DGP< dim, spacedim >, FE_FaceQ< 1, spacedim >, FE_DGQ< dim, spacedim >, FE_Q_DG0< dim, spacedim >, FE_Nedelec< dim >, FE_Q_Base< PolynomialType, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, dim >, FE_Q_Base< TensorProductPolynomials< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomialsConst< dim >, dim, spacedim >, FE_Q_Base< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >, FE_FaceQ< dim, spacedim >, FE_RaviartThomas< dim >, and FE_TraceQ< dim, spacedim >. Definition at line 1133 of file fe.cc. template<int dim, int spacedim> const std::vector< Point< dim > > & FiniteElement< dim, spacedim >::get_unit_support_points ( ) const Return the support points of the trial functions on the unit cell, if the derived finite element defines them. Finite elements that allow some kind of interpolation operation usually have support points. On the other hand, elements that define their degrees of freedom by, for example, moments on faces, or as derivatives, don't have support points. In that case, the returned field is empty. If the finite element defines support points, then their number equals the number of degrees of freedom of the element. The order of points in the array matches that returned by the cell->get_dof_indices function. See the class documentation for details on support points. Note Finite elements' implementation of this function returns these points in the same order as shape functions. The order of shape functions is typically documented in the class documentation of the various finite element classes. In particular, shape functions (and consequently the mapped quadrature points discussed in the class documentation of this class) will then traverse first those shape functions located on vertices, then on lines, then on quads, etc. If this element implements support points, then it will return one such point per shape function. Since multiple shape functions may be defined at the same location, the support points returned here may be duplicated. An example would be an element of the kind FESystem(FE_Q(1),3) for which each support point would appear three times in the returned array. Definition at line 949 of file fe.cc. template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::has_support_points ( ) const Return whether a finite element has defined support points. If the result is true, then a call to the get_unit_support_points() yields a non-empty array. The result may be false if an element is not defined by interpolating shape functions, for example by P-elements on quadrilaterals. It will usually only be true if the element constructs its shape functions by the requirement that they be one at a certain point and zero at all the points associated with the other shape functions. In composed elements (i.e. for the FESystem class), the result will be true if all the base elements have defined support points. FE_Nothing is a special case in FESystems, because it has 0 support points and has_support_points() is false, but an FESystem containing an FE_Nothing among other elements will return true. Definition at line 965 of file fe.cc. template<int dim, int spacedim> Point< dim > FiniteElement< dim, spacedim >::unit_support_point ( const unsigned int  index) const virtual Return the position of the support point of the indexth shape function. If it does not exist, raise an exception. The default implementation simply returns the respective element from the array you get from get_unit_support_points(), but derived elements may overload this function. In particular, note that the FESystem class overloads it so that it can return the support points of individual base elements, if not all the base elements define support points. In this way, you can still ask for certain support points, even if get_unit_support_points() only returns an empty array. Reimplemented in FESystem< dim, spacedim >. Definition at line 996 of file fe.cc. template<int dim, int spacedim> const std::vector< Point< dim-1 > > & FiniteElement< dim, spacedim >::get_unit_face_support_points ( ) const Return the support points of the trial functions on the unit face, if the derived finite element defines some. Finite elements that allow some kind of interpolation operation usually have support points. On the other hand, elements that define their degrees of freedom by, for example, moments on faces, or as derivatives, don't have support points. In that case, the returned field is empty Note that elements that have support points need not necessarily have some on the faces, even if the interpolation points are located physically on a face. For example, the discontinuous elements have interpolation points on the vertices, and for higher degree elements also on the faces, but they are not defined to be on faces since in that case degrees of freedom from both sides of a face (or from all adjacent elements to a vertex) would be identified with each other, which is not what we would like to have). Logically, these degrees of freedom are therefore defined to belong to the cell, rather than the face or vertex. In that case, the returned element would therefore have length zero. If the finite element defines support points, then their number equals the number of degrees of freedom on the face (dofs_per_face). The order of points in the array matches that returned by the cell->face(face)->get_dof_indices function. See the class documentation for details on support points. Definition at line 1009 of file fe.cc. template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::has_face_support_points ( ) const Return whether a finite element has defined support points on faces. If the result is true, then a call to the get_unit_face_support_points() yields a non-empty vector. For more information, see the documentation for the has_support_points() function. Definition at line 1025 of file fe.cc. template<int dim, int spacedim> Point< dim-1 > FiniteElement< dim, spacedim >::unit_face_support_point ( const unsigned int  index) const virtual The function corresponding to the unit_support_point() function, but for faces. See there for more information. Reimplemented in FESystem< dim, spacedim >. Definition at line 1058 of file fe.cc. template<int dim, int spacedim> const std::vector< Point< dim > > & FiniteElement< dim, spacedim >::get_generalized_support_points ( ) const Return a vector of generalized support points. Note The vector returned by this function is always a minimal set of unique support points. This is in contrast to the behavior of get_unit_support_points() that returns a repeated list of unit support points for an FESystem of numerous (Lagrangian) base elements. See the glossary entry on generalized support points for more information. Definition at line 974 of file fe.cc. template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::has_generalized_support_points ( ) const Return whether a finite element has defined generalized support points. If the result is true, then a call to the get_generalized_support_points() yields a non-empty vector. See the glossary entry on generalized support points for more information. Definition at line 987 of file fe.cc. template<int dim, int spacedim> const std::vector< Point< dim-1 > > & FiniteElement< dim, spacedim >::get_generalized_face_support_points ( ) const Return the equivalent to get_generalized_support_points(), except for faces. Deprecated: In general, it is not possible to associate a unique subset of generalized support points describing degrees of freedom for a given face. Don't use this function Definition at line 1034 of file fe.cc. template<int dim, int spacedim> bool FiniteElement< dim, spacedim >::has_generalized_face_support_points ( ) const Return whether a finite element has defined generalized support points on faces. If the result is true, then a call to the get_generalized_face_support_points() function yields a non-empty array. For more information, see the documentation for the has_support_points() function. Deprecated: In general, it is not possible to associate a unique subset of generalized support points describing degrees of freedom for a given face. Don't use this function Definition at line 1049 of file fe.cc. template<int dim, int spacedim> GeometryPrimitive FiniteElement< dim, spacedim >::get_associated_geometry_primitive ( const unsigned int  cell_dof_index) const inline For a given degree of freedom, return whether it is logically associated with a vertex, line, quad or hex. For instance, for continuous finite elements this coincides with the lowest dimensional object the support point of the degree of freedom lies on. To give an example, for \(Q_1\) elements in 3d, every degree of freedom is defined by a shape function that we get by interpolating using support points that lie on the vertices of the cell. The support of these points of course extends to all edges connected to this vertex, as well as the adjacent faces and the cell interior, but we say that logically the degree of freedom is associated with the vertex as this is the lowest- dimensional object it is associated with. Likewise, for \(Q_2\) elements in 3d, the degrees of freedom with support points at edge midpoints would yield a value of GeometryPrimitive::line from this function, whereas those on the centers of faces in 3d would return GeometryPrimitive::quad. To make this more formal, the kind of object returned by this function represents the object so that the support of the shape function corresponding to the degree of freedom, (i.e., that part of the domain where the function "lives") is the union of all of the cells sharing this object. To return to the example above, for \(Q_2\) in 3d, the shape function with support point at an edge midpoint has support on all cells that share the edge and not only the cells that share the adjacent faces, and consequently the function will return GeometryPrimitive::line. On the other hand, for discontinuous elements of type \(DGQ_2\), a degree of freedom associated with an interpolation polynomial that has its support point physically located at a line bounding a cell, but is nonzero only on one cell. Consequently, it is logically associated with the interior of that cell (i.e., with a GeometryPrimitive::quad in 2d and a GeometryPrimitive::hex in 3d). Parameters [in]cell_dof_indexThe index of a shape function or degree of freedom. This index must be in the range [0,dofs_per_cell). Note The integer value of the object returned by this function equals the dimensionality of the object it describes, and can consequently be used in generic programming paradigms. For example, if a degree of freedom is associated with a vertex, then this function returns GeometryPrimitive::vertex, which has a numeric value of zero (the dimensionality of a vertex). Definition at line 3234 of file fe.h. template<int dim, int spacedim> void FiniteElement< dim, spacedim >::convert_generalized_support_point_values_to_dof_values ( const std::vector< Vector< double > > &  support_point_values, std::vector< double > &  nodal_values  ) const virtual Given the values of a function \(f(\mathbf x)\) at the (generalized) support points of the reference cell, this function then computes what the nodal values of the element are, i.e., \(\Psi_i[f]\), where \(\Psi_i\) are the node functionals of the element (see also Node values or node functionals). The values \(\Psi_i[f]\) are then the expansion coefficients for the shape functions of the finite element function that interpolates the given function \(f(x)\), i.e., \( f_h(\mathbf x) = \sum_i \Psi_i[f] \varphi_i(\mathbf x) \) is the finite element interpolant of \(f\) with the current element. The operation described here is used, for example, in the FETools::compute_node_matrix() function. In more detail, let us assume that the generalized support points (see this glossary entry ) of the current element are \(\hat{\mathbf x}_i\) and that the node functionals associated with the current element are \(\Psi_i[\cdot]\). Then, the fact that the element is based on generalized support points, implies that if we apply \(\Psi_i\) to a (possibly vector-valued) finite element function \(\varphi\), the result must have the form \(\Psi_i[\varphi] = f_i(\varphi(\hat{\mathbf x}_i))\) – in other words, the value of the node functional \(\Psi_i\) applied to \(\varphi\) only depends on the values of \(\varphi\) at \(\hat{\mathbf x}_i\) and not on values anywhere else, or integrals of \(\varphi\), or any other kind of information. The exact form of \(f_i\) depends on the element. For example, for scalar Lagrange elements, we have that in fact \(\Psi_i[\varphi] = \varphi(\hat{\mathbf x}_i)\). If you combine multiple scalar Lagrange elements via an FESystem object, then \(\Psi_i[\varphi] = \varphi(\hat{\mathbf x}_i)_{c(i)}\) where \(c(i)\) is the result of the FiniteElement::system_to_component_index() function's return value's first component. In these two cases, \(f_i\) is therefore simply the identity (in the scalar case) or a function that selects a particular vector component of its argument. On the other hand, for Raviart-Thomas elements, one would have that \(f_i(\mathbf y) = \mathbf y \cdot \mathbf n_i\) where \(\mathbf n_i\) is the normal vector of the face at which the shape function is defined. Given all of this, what this function does is the following: If you input a list of values of a function \(\varphi\) at all generalized support points (where each value is in fact a vector of values with as many components as the element has), then this function returns a vector of values obtained by applying the node functionals to these values. In other words, if you pass in \(\{\varphi(\hat{\mathbf x}_i)\}_{i=0}^{N-1}\) then you will get out a vector \(\{\Psi[\varphi]\}_{i=0}^{N-1}\) where \(N\) equals dofs_per_cell. Parameters [in]support_point_valuesAn array of size dofs_per_cell (which equals the number of points the get_generalized_support_points() function will return) where each element is a vector with as many entries as the element has vector components. This array should contain the values of a function at the generalized support points of the current element. [out]nodal_valuesAn array of size dofs_per_cell that contains the node functionals of the element applied to the given function. Note It is safe to call this function for (transformed) values on the real cell only for elements with trivial MappingType. For all other elements (for example for H(curl), or H(div) conforming elements) vector values have to be transformed to the reference cell first. Given what the function is supposed to do, the function clearly can only work for elements that actually implement (generalized) support points. Elements that do not have generalized support points – e.g., elements whose nodal functionals evaluate integrals or moments of functions (such as FE_Q_Hierarchical) – can in general not make sense of the operation that is required for this function. They consequently may not implement it. Reimplemented in FESystem< dim, spacedim >, FE_Q< dim, spacedim >, FE_Q< dim >, FE_DGQArbitraryNodes< dim, spacedim >, FE_DGQ< dim, spacedim >, FE_RaviartThomasNodal< dim >, FE_Q_DG0< dim, spacedim >, FE_Nedelec< dim >, FE_Q_iso_Q1< dim, spacedim >, FE_RaviartThomas< dim >, FE_ABF< dim >, FE_RT_Bubbles< dim >, FE_Q_Bubbles< dim, spacedim >, FE_FaceQ< dim, spacedim >, FE_BDM< dim >, FE_RannacherTurek< dim >, and FE_TraceQ< dim, spacedim >. Definition at line 1146 of file fe.cc. template<int dim, int spacedim> std::size_t FiniteElement< dim, spacedim >::memory_consumption ( ) const virtual Determine an estimate for the memory consumption (in bytes) of this object. This function is made virtual, since finite element objects are usually accessed through pointers to their base class, rather than the class itself. Reimplemented in FESystem< dim, spacedim >, FE_Q_Hierarchical< dim >, FE_DGPNonparametric< dim, spacedim >, FE_DGP< dim, spacedim >, FE_DGPMonomial< dim >, FE_DGQ< dim, spacedim >, FE_Nedelec< dim >, FE_RaviartThomas< dim >, FE_ABF< dim >, FE_DGVector< PolynomialType, dim, spacedim >, FE_DGVector< PolynomialsRaviartThomas< dim >, dim, spacedim >, FE_DGVector< PolynomialsBDM< dim >, dim, spacedim >, and FE_DGVector< PolynomialsNedelec< dim >, dim, spacedim >. Definition at line 1162 of file fe.cc. template<int dim, int spacedim> void FiniteElement< dim, spacedim >::reinit_restriction_and_prolongation_matrices ( const bool  isotropic_restriction_only = false, const bool  isotropic_prolongation_only = false  ) protected Reinit the vectors of restriction and prolongation matrices to the right sizes: For every refinement case, except for RefinementCase::no_refinement, and for every child of that refinement case the space of one restriction and prolongation matrix is allocated, see the documentation of the restriction and prolongation vectors for more detail on the actual vector sizes. Parameters isotropic_restriction_onlyonly the restriction matrices required for isotropic refinement are reinited to the right size. isotropic_prolongation_onlyonly the prolongation matrices required for isotropic refinement are reinited to the right size. Definition at line 273 of file fe.cc. template<int dim, int spacedim> TableIndices< 2 > FiniteElement< dim, spacedim >::interface_constraints_size ( ) const protected Return the size of interface constraint matrices. Since this is needed in every derived finite element class when initializing their size, it is placed into this function, to avoid having to recompute the dimension- dependent size of these matrices each time. Note that some elements do not implement the interface constraints for certain polynomial degrees. In this case, this function still returns the size these matrices should have when implemented, but the actual matrices are empty. Definition at line 819 of file fe.cc. template<int dim, int spacedim> std::vector< unsigned int > FiniteElement< dim, spacedim >::compute_n_nonzero_components ( const std::vector< ComponentMask > &  nonzero_components) staticprotected Given the pattern of nonzero components for each shape function, compute for each entry how many components are non-zero for each shape function. This function is used in the constructor of this class. Definition at line 1182 of file fe.cc. template<int dim, int spacedim = dim> virtual UpdateFlags FiniteElement< dim, spacedim >::requires_update_flags ( const UpdateFlags  update_flags) const protectedpure virtual Given a set of update flags, compute which other quantities also need to be computed in order to satisfy the request by the given flags. Then return the combination of the original set of flags and those just computed. As an example, if update_flags contains update_gradients a finite element class will typically require the computation of the inverse of the Jacobian matrix in order to rotate the gradient of shape functions on the reference cell to the real cell. It would then return not just update_gradients, but also update_covariant_transformation, the flag that makes the mapping class produce the inverse of the Jacobian matrix. An extensive discussion of the interaction between this function and FEValues can be found in the How Mapping, FiniteElement, and FEValues work together documentation module. See also UpdateFlags Implemented in FESystem< dim, spacedim >, FE_DGPNonparametric< dim, spacedim >, FE_Enriched< dim, spacedim >, FE_P1NC, FE_FaceQ< 1, spacedim >, FE_PolyTensor< PolynomialType, dim, spacedim >, FE_PolyTensor< PolynomialsRaviartThomas< dim >, dim, spacedim >, FE_PolyTensor< PolynomialsNedelec< dim >, dim >, FE_PolyTensor< PolynomialsRaviartThomas< dim >, dim >, FE_PolyTensor< PolynomialsBDM< dim >, dim, spacedim >, FE_PolyTensor< PolynomialsBDM< dim >, dim >, FE_PolyTensor< PolynomialsABF< dim >, dim >, FE_PolyTensor< PolynomialsRT_Bubbles< dim >, dim >, FE_PolyTensor< PolynomialsNedelec< dim >, dim, spacedim >, FE_Nothing< dim, spacedim >, FE_Nothing< dim >, FE_Poly< PolynomialType, dim, spacedim >, FE_Poly< PolynomialSpace< dim >, dim, spacedim >, FE_Poly< PolynomialsP< dim >, dim >, FE_Poly< TensorProductPolynomials< dim >, dim >, FE_Poly< TensorProductPolynomials< dim >, dim, spacedim >, FE_Poly< PolynomialsRannacherTurek< dim >, dim >, FE_Poly< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Poly< TensorProductPolynomialsConst< dim >, dim, spacedim >, FE_Poly< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >, FE_PolyFace< PolynomialType, dim, spacedim >, FE_PolyFace< PolynomialSpace< dim-1 >, dim, spacedim >, and FE_PolyFace< TensorProductPolynomials< dim-1 >, dim, spacedim >. template<int dim, int spacedim = dim> virtual std::unique_ptr<InternalDataBase> FiniteElement< dim, spacedim >::get_data ( const UpdateFlags  update_flags, const Mapping< dim, spacedim > &  mapping, const Quadrature< dim > &  quadrature, ::internal::FEValuesImplementation::FiniteElementRelatedData< dim, spacedim > &  output_data  ) const protectedpure virtual Create an internal data object and return a pointer to it of which the caller of this function then assumes ownership. This object will then be passed to the FiniteElement::fill_fe_values() every time the finite element shape functions and their derivatives are evaluated on a concrete cell. The object created here is therefore used by derived classes as a place for scratch objects that are used in evaluating shape functions, as well as to store information that can be pre-computed once and re-used on every cell (e.g., for evaluating the values and gradients of shape functions on the reference cell, for later re-use when transforming these values to a concrete cell). This function is the first one called in the process of initializing a FEValues object for a given mapping and finite element object. The returned object will later be passed to FiniteElement::fill_fe_values() for a concrete cell, which will itself place its output into an object of type internal::FEValuesImplementation::FiniteElementRelatedData. Since there may be data that can already be computed in its final form on the reference cell, this function also receives a reference to the internal::FEValuesImplementation::FiniteElementRelatedData object as its last argument. This output argument is guaranteed to always be the same one when used with the InternalDataBase object returned by this function. In other words, the subdivision of scratch data and final data in the returned object and the output_data object is as follows: If data can be pre- computed on the reference cell in the exact form in which it will later be needed on a concrete cell, then this function should already emplace it in the output_data object. An example are the values of shape functions at quadrature points for the usual Lagrange elements which on a concrete cell are identical to the ones on the reference cell. On the other hand, if some data can be pre-computed to make computations on a concrete cell cheaper, then it should be put into the returned object for later re-use in a derive class's implementation of FiniteElement::fill_fe_values(). An example are the gradients of shape functions on the reference cell for Lagrange elements: to compute the gradients of the shape functions on a concrete cell, one has to multiply the gradients on the reference cell by the inverse of the Jacobian of the mapping; consequently, we cannot already compute the gradients on a concrete cell at the time the current function is called, but we can at least pre-compute the gradients on the reference cell, and store it in the object returned. An extensive discussion of the interaction between this function and FEValues can be found in the How Mapping, FiniteElement, and FEValues work together documentation module. See also the documentation of the InternalDataBase class. Parameters [in]update_flagsA set of UpdateFlags values that describe what kind of information the FEValues object requests the finite element to compute. This set of flags may also include information that the finite element can not compute, e.g., flags that pertain to data produced by the mapping. An implementation of this function needs to set up all data fields in the returned object that are necessary to produce the finite- element related data specified by these flags, and may already pre- compute part of this information as discussed above. Elements may want to store these update flags (or a subset of these flags) in InternalDataBase::update_each so they know at the time when FiniteElement::fill_fe_values() is called what they are supposed to compute [in]mappingA reference to the mapping used for computing values and derivatives of shape functions. [in]quadratureA reference to the object that describes where the shape functions should be evaluated. [out]output_dataA reference to the object that FEValues will use in conjunction with the object returned here and where an implementation of FiniteElement::fill_fe_values() will place the requested information. This allows the current function to already pre-compute pieces of information that can be computed on the reference cell, as discussed above. FEValues guarantees that this output object and the object returned by the current function will always be used together. Returns A pointer to an object of a type derived from InternalDataBase and that derived classes can use to store scratch data that can be pre- computed, or for scratch arrays that then only need to be allocated once. The calling site assumes ownership of this object and will delete it when it is no longer necessary. Implemented in FESystem< dim, spacedim >, FE_Enriched< dim, spacedim >, FE_DGPNonparametric< dim, spacedim >, FE_Poly< PolynomialType, dim, spacedim >, FE_Poly< PolynomialSpace< dim >, dim, spacedim >, FE_Poly< TensorProductPolynomials< dim >, dim, spacedim >, FE_Poly< TensorProductPolynomialsBubbles< dim >, dim, spacedim >, FE_Poly< TensorProductPolynomialsConst< dim >, dim, spacedim >, FE_Poly< TensorProductPolynomials< dim, Polynomials::PiecewisePolynomial< double > >, dim, spacedim >, FE_PolyTensor< PolynomialType, dim, spacedim >, FE_PolyTensor< PolynomialsRaviartThomas< dim >, dim, spacedim >, FE_PolyTensor< PolynomialsBDM< dim >, dim, spacedim >, FE_PolyTensor< PolynomialsNedelec< dim >, dim, spacedim >, FE_Nothing< dim, spacedim >, FE_PolyFace< PolynomialType, dim, spacedim >, FE_PolyFace< PolynomialSpace< dim-1 >, dim, spacedim >, and FE_PolyFace< TensorProductPolynomials< dim-1 >, dim, spacedim >. template<int dim, int spacedim> std::unique_ptr< typename FiniteElement< dim, spacedim >::InternalDataBase > FiniteElement< dim, spacedim >::get_face_data ( const UpdateFlags  update_flags, const Mapping< dim, spacedim > &  mapping, const Quadrature< dim-1 > &  quadrature, ::internal::FEValuesImplementation::FiniteElementRelatedData< dim, spacedim > &  output_data  ) const protectedvirtual Like get_data(), but return an object that will later be used for evaluating shape function information at quadrature points on faces of cells. The object will then be used in calls to implementations of FiniteElement::fill_fe_face_values(). See the documentation of get_data() for more information. The default implementation of this function converts the face quadrature into a cell quadrature with appropriate quadrature point locations, and with that calls the get_data() function above that has to be implemented in derived classes. Parameters [in]update_flagsA set of UpdateFlags values that describe what kind of information the FEValues object requests the finite element to compute. This set of flags may also include information that the finite element can not compute, e.g., flags that pertain to data produced by the mapping. An implementation of this function needs to set up all data fields in the returned object that are necessary to produce the finite- element related data specified by these flags, and may already pre- compute part of this information as discussed above. Elements may want to store these update flags (or a subset of these flags) in InternalDataBase::update_each so they know at the time when FiniteElement::fill_fe_face_values() is called what they are supposed to compute [in]mappingA reference to the mapping used for computing values and derivatives of shape functions. [in]quadratureA reference to the object that describes where the shape functions should be evaluated. [out]output_dataA reference to the object that FEValues will use in conjunction with the object returned here and where an implementation of FiniteElement::fill_fe_face_values() will place the requested information. This allows the current function to already pre-compute pieces of information that can be computed on the reference cell, as discussed above. FEValues guarantees that this output object and the object returned by the current function will always be used together. Returns A pointer to an object of a type derived from InternalDataBase and that derived classes can use to store scratch data that can be pre- computed, or for scratch arrays that then only need to be allocated once. The calling site assumes ownership of this object and will delete it when it is no longer necessary. Reimplemented in FESystem< dim, spacedim >, FE_Enriched< dim, spacedim >, FE_PolyFace< PolynomialType, dim, spacedim >, FE_PolyFace< PolynomialSpace< dim-1 >, dim, spacedim >, and FE_PolyFace< TensorProductPolynomials< dim-1 >, dim, spacedim >. Definition at line 1197 of file fe.cc. template<int dim, int spacedim> std::unique_ptr< typename FiniteElement< dim, spacedim >::InternalDataBase > FiniteElement< dim, spacedim >::get_subface_data ( const UpdateFlags  update_flags, const Mapping< dim, spacedim > &  mapping, const Quadrature< dim-1 > &  quadrature, ::internal::FEValuesImplementation::FiniteElementRelatedData< dim, spacedim > &  output_data  ) const protectedvirtual Like get_data(), but return an object that will later be used for evaluating shape function information at quadrature points on children of faces of cells. The object will then be used in calls to implementations of FiniteElement::fill_fe_subface_values(). See the documentation of get_data() for more information. The default implementation of this function converts the face quadrature into a cell quadrature with appropriate quadrature point locations, and with that calls the get_data() function above that has to be implemented in derived classes. Parameters [in]update_flagsA set of UpdateFlags values that describe what kind of information the FEValues object requests the finite element to compute. This set of flags may also include information that the finite element can not compute, e.g., flags that pertain to data produced by the mapping. An implementation of this function needs to set up all data fields in the returned object that are necessary to produce the finite- element related data specified by these flags, and may already pre- compute part of this information as discussed above. Elements may want to store these update flags (or a subset of these flags) in InternalDataBase::update_each so they know at the time when FiniteElement::fill_fe_subface_values() is called what they are supposed to compute [in]mappingA reference to the mapping used for computing values and derivatives of shape functions. [in]quadratureA reference to the object that describes where the shape functions should be evaluated. [out]output_dataA reference to the object that FEValues will use in conjunction with the object returned here and where an implementation of FiniteElement::fill_fe_subface_values() will place the requested information. This allows the current function to already pre-compute pieces of information that can be computed on the reference cell, as discussed above. FEValues guarantees that this output object and the object returned by the current function will always be used together. Returns A pointer to an object of a type derived from InternalDataBase and that derived classes can use to store scratch data that can be pre- computed, or for scratch arrays that then only need to be allocated once. The calling site assumes ownership of this object and will delete it when it is no longer necessary. Reimplemented in FESystem< dim, spacedim >, FE_Enriched< dim, spacedim >, FE_PolyFace< PolynomialType, dim, spacedim >, FE_PolyFace< PolynomialSpace< dim-1 >, dim, spacedim >, and FE_PolyFace< TensorProductPolynomials< dim-1 >, dim, spacedim >. Definition at line 1211 of file fe.cc. template<int dim, int spacedim = dim> virtual void FiniteElement< dim, spacedim >::fill_fe_values ( const typename Triangulation< dim, spacedim >::cell_iterator &  cell, const CellSimilarity::Similarity  cell_similarity, const Quadrature< dim > &  quadrature, const Mapping< dim, spacedim > &  mapping, const typename Mapping< dim, spacedim >::InternalDataBase mapping_internal, const ::internal::FEValuesImplementation::MappingRelatedData< dim, spacedim > &  mapping_data, const InternalDataBase fe_internal, ::internal::FEValuesImplementation::FiniteElementRelatedData< dim, spacedim > &  output_data  ) const protectedpure virtual Compute information about the shape functions on the cell denoted by the first argument. Derived classes will have to implement this function based on the kind of element they represent. It is called by FEValues::reinit(). Conceptually, this function evaluates shape functions and their derivatives at the quadrature points represented by the mapped locations of those described by the quadrature argument to this function. In many cases, computing derivatives of shape functions (and in some cases also computing values of shape functions) requires making use of the mapping from the reference to the real cell; this information can either be taken from the mapping_data object that has been filled for the current cell before this function is called, or by calling the member functions of a Mapping object with the mapping_internal object that also corresponds to the current cell. The information computed by this function is used to fill the various member variables of the output argument of this function. Which of the member variables of that structure should be filled is determined by the update flags stored in the FiniteElement::InternalDataBase::update_each field of the object passed to this function. These flags are typically set by FiniteElement::get_data(), FiniteElement::get_face_date() and FiniteElement::get_subface_data() (or, more specifically, implementations of these functions in derived classes). An extensive discussion of the interaction between this function and FEValues can be found in the How Mapping, FiniteElement, and FEValues work together documentation module. Parameters [in]cellThe cell of the triangulation for which this function is to compute a mapping from the reference cell to. [in]cell_similarityWhether or not the cell given as first argument is simply a translation, rotation, etc of the cell for which this function was called the most recent time. This information is computed simply by matching the vertices (as stored by the Triangulation) between the previous and the current cell. The value passed here may be modified by implementations of this function and should then be returned (see the discussion of the return value of this function). [in]quadratureA reference to the quadrature formula in use for the current evaluation. This quadrature object is the same as the one used when creating the internal_data object. The current object is then responsible for evaluating shape functions at the mapped locations of the quadrature points represented by this object. [in]mappingA reference to the mapping object used to map from the reference cell to the current cell. This object was used to compute the information in the mapping_data object before the current function was called. It is also the mapping object that created the mapping_internal object via Mapping::get_data(). You will need the reference to this mapping object most often to call Mapping::transform() to transform gradients and higher derivatives from the reference to the current cell. [in]mapping_internalAn object specific to the mapping object. What the mapping chooses to store in there is of no relevance to the current function, but you may have to pass a reference to this object to certain functions of the Mapping class (e.g., Mapping::transform()) if you need to call them from the current function. [in]mapping_dataThe output object into which the Mapping::fill_fe_values() function wrote the mapping information corresponding to the current cell. This includes, for example, Jacobians of the mapping that may be of relevance to the current function, as well as other information that FEValues::reinit() requested from the mapping. [in]fe_internalA reference to an object previously created by get_data() and that may be used to store information the mapping can compute once on the reference cell. See the documentation of the FiniteElement::InternalDataBase class for an extensive description of the purpose of these objects. [out]output_dataA reference to an object whose member variables should be computed. Not all of the members of this argument need to be filled; which ones need to be filled is determined by the update flags stored inside the fe_internal object. Note FEValues ensures that this function is always called with the same pair of fe_internal and output_data objects. In other words, if an implementation of this function knows that it has written a piece of data into the output argument in a previous call, then there is no need to copy it there again in a later call if the implementation knows that this is the same value. template<int dim, int spacedim = dim> virtual void FiniteElement< dim, spacedim >::fill_fe_face_values ( const typename Triangulation< dim, spacedim >::cell_iterator &  cell, const unsigned int  face_no, const Quadrature< dim-1 > &  quadrature, const Mapping< dim, spacedim > &  mapping, const typename Mapping< dim, spacedim >::InternalDataBase mapping_internal, const ::internal::FEValuesImplementation::MappingRelatedData< dim, spacedim > &  mapping_data, const InternalDataBase fe_internal, ::internal::FEValuesImplementation::FiniteElementRelatedData< dim, spacedim > &  output_data  ) const protectedpure virtual This function is the equivalent to FiniteElement::fill_fe_values(), but for faces of cells. See there for an extensive discussion of its purpose. It is called by FEFaceValues::reinit(). Parameters [in]cellThe cell of the triangulation for which this function is to compute a mapping from the reference cell to. [in]face_noThe number of the face we are currently considering, indexed among the faces of the cell specified by the previous argument. [in]quadratureA reference to the quadrature formula in use for the current evaluation. This quadrature object is the same as the one used when creating the internal_data object. The current object is then responsible for evaluating shape functions at the mapped locations of the quadrature points represented by this object. [in]mappingA reference to the mapping object used to map from the reference cell to the current cell. This object was used to compute the information in the mapping_data object before the current function was called. It is also the mapping object that created the mapping_internal object via Mapping::get_data(). You will need the reference to this mapping object most often to call Mapping::transform() to transform gradients and higher derivatives from the reference to the current cell. [in]mapping_internalAn object specific to the mapping object. What the mapping chooses to store in there is of no relevance to the current function, but you may have to pass a reference to this object to certain functions of the Mapping class (e.g., Mapping::transform()) if you need to call them from the current function. [in]mapping_dataThe output object into which the Mapping::fill_fe_values() function wrote the mapping information corresponding to the current cell. This includes, for example, Jacobians of the mapping that may be of relevance to the current function, as well as other information that FEValues::reinit() requested from the mapping. [in]fe_internalA reference to an object previously created by get_data() and that may be used to store information the mapping can compute once on the reference cell. See the documentation of the FiniteElement::InternalDataBase class for an extensive description of the purpose of these objects. [out]output_dataA reference to an object whose member variables should be computed. Not all of the members of this argument need to be filled; which ones need to be filled is determined by the update flags stored inside the fe_internal object. template<int dim, int spacedim = dim> virtual void FiniteElement< dim, spacedim >::fill_fe_subface_values ( const typename Triangulation< dim, spacedim >::cell_iterator &  cell, const unsigned int  face_no, const unsigned int  sub_no, const Quadrature< dim-1 > &  quadrature, const Mapping< dim, spacedim > &  mapping, const typename Mapping< dim, spacedim >::InternalDataBase mapping_internal, const ::internal::FEValuesImplementation::MappingRelatedData< dim, spacedim > &  mapping_data, const InternalDataBase fe_internal, ::internal::FEValuesImplementation::FiniteElementRelatedData< dim, spacedim > &  output_data  ) const protectedpure virtual This function is the equivalent to FiniteElement::fill_fe_values(), but for the children of faces of cells. See there for an extensive discussion of its purpose. It is called by FESubfaceValues::reinit(). Parameters [in]cellThe cell of the triangulation for which this function is to compute a mapping from the reference cell to. [in]face_noThe number of the face we are currently considering, indexed among the faces of the cell specified by the previous argument. [in]sub_noThe number of the subface, i.e., the number of the child of a face, that we are currently considering, indexed among the children of the face specified by the previous argument. [in]quadratureA reference to the quadrature formula in use for the current evaluation. This quadrature object is the same as the one used when creating the internal_data object. The current object is then responsible for evaluating shape functions at the mapped locations of the quadrature points represented by this object. [in]mappingA reference to the mapping object used to map from the reference cell to the current cell. This object was used to compute the information in the mapping_data object before the current function was called. It is also the mapping object that created the mapping_internal object via Mapping::get_data(). You will need the reference to this mapping object most often to call Mapping::transform() to transform gradients and higher derivatives from the reference to the current cell. [in]mapping_internalAn object specific to the mapping object. What the mapping chooses to store in there is of no relevance to the current function, but you may have to pass a reference to this object to certain functions of the Mapping class (e.g., Mapping::transform()) if you need to call them from the current function. [in]mapping_dataThe output object into which the Mapping::fill_fe_values() function wrote the mapping information corresponding to the current cell. This includes, for example, Jacobians of the mapping that may be of relevance to the current function, as well as other information that FEValues::reinit() requested from the mapping. [in]fe_internalA reference to an object previously created by get_data() and that may be used to store information the mapping can compute once on the reference cell. See the documentation of the FiniteElement::InternalDataBase class for an extensive description of the purpose of these objects. [out]output_dataA reference to an object whose member variables should be computed. Not all of the members of this argument need to be filled; which ones need to be filled is determined by the update flags stored inside the fe_internal object. Member Data Documentation template<int dim, int spacedim = dim> const unsigned int FiniteElement< dim, spacedim >::space_dimension = spacedim static The dimension of the image space, corresponding to Triangulation. Definition at line 640 of file fe.h. template<int dim, int spacedim = dim> std::vector<std::vector<FullMatrix<double> > > FiniteElement< dim, spacedim >::restriction protected Vector of projection matrices. See get_restriction_matrix() above. The constructor initializes these matrices to zero dimensions, which can be changed by derived classes implementing them. Note, that restriction[refinement_case-1][child] includes the restriction matrix of child child for the RefinementCase refinement_case. Here, we use refinement_case-1 instead of refinement_case as for RefinementCase::no_refinement(=0) there are no restriction matrices available. Definition at line 2346 of file fe.h. template<int dim, int spacedim = dim> std::vector<std::vector<FullMatrix<double> > > FiniteElement< dim, spacedim >::prolongation protected Vector of embedding matrices. See get_prolongation_matrix() above. The constructor initializes these matrices to zero dimensions, which can be changed by derived classes implementing them. Note, that prolongation[refinement_case-1][child] includes the prolongation matrix of child child for the RefinementCase refinement_case. Here, we use refinement_case-1 instead of refinement_case as for RefinementCase::no_refinement(=0) there are no prolongation matrices available. Definition at line 2360 of file fe.h. template<int dim, int spacedim = dim> FullMatrix<double> FiniteElement< dim, spacedim >::interface_constraints protected Specify the constraints which the dofs on the two sides of a cell interface underlie if the line connects two cells of which one is refined once. For further details see the general description of the derived class. This field is obviously useless in one dimension and has there a zero size. Definition at line 2372 of file fe.h. template<int dim, int spacedim = dim> std::vector<Point<dim> > FiniteElement< dim, spacedim >::unit_support_points protected List of support points on the unit cell, in case the finite element has any. The constructor leaves this field empty, derived classes may write in some contents. Finite elements that allow some kind of interpolation operation usually have support points. On the other hand, elements that define their degrees of freedom by, for example, moments on faces, or as derivatives, don't have support points. In that case, this field remains empty. Definition at line 2384 of file fe.h. template<int dim, int spacedim = dim> std::vector<Point<dim-1> > FiniteElement< dim, spacedim >::unit_face_support_points protected Same for the faces. See the description of the get_unit_face_support_points() function for a discussion of what contributes a face support point. Definition at line 2391 of file fe.h. template<int dim, int spacedim = dim> std::vector<Point<dim> > FiniteElement< dim, spacedim >::generalized_support_points protected Support points used for interpolation functions of non-Lagrangian elements. Definition at line 2397 of file fe.h. template<int dim, int spacedim = dim> std::vector<Point<dim-1> > FiniteElement< dim, spacedim >::generalized_face_support_points protected Face support points used for interpolation functions of non-Lagrangian elements. Definition at line 2403 of file fe.h. template<int dim, int spacedim = dim> Table<2,int> FiniteElement< dim, spacedim >::adjust_quad_dof_index_for_face_orientation_table protected For faces with non-standard face_orientation in 3D, the dofs on faces (quads) have to be permuted in order to be combined with the correct shape functions. Given a local dof index on a quad, return the shift in the local index, if the face has non-standard face_orientation, i.e. old_index + shift = new_index. In 2D and 1D there is no need for permutation so the vector is empty. In 3D it has the size of dofs_per_quad * 8 , where 8 is the number of orientations, a face can be in (all combinations of the three bool flags face_orientation, face_flip and face_rotation). The constructor of this class fills this table with zeros, i.e., no permutation at all. Derived finite element classes have to fill this Table with the correct values. Definition at line 2420 of file fe.h. template<int dim, int spacedim = dim> std::vector<int> FiniteElement< dim, spacedim >::adjust_line_dof_index_for_line_orientation_table protected For lines with non-standard line_orientation in 3D, the dofs on lines have to be permuted in order to be combined with the correct shape functions. Given a local dof index on a line, return the shift in the local index, if the line has non-standard line_orientation, i.e. old_index + shift = new_index. In 2D and 1D there is no need for permutation so the vector is empty. In 3D it has the size of dofs_per_line. The constructor of this class fills this table with zeros, i.e., no permutation at all. Derived finite element classes have to fill this vector with the correct values. Definition at line 2435 of file fe.h. template<int dim, int spacedim = dim> std::vector<std::pair<unsigned int, unsigned int> > FiniteElement< dim, spacedim >::system_to_component_table protected Store what system_to_component_index() will return. Definition at line 2440 of file fe.h. template<int dim, int spacedim = dim> std::vector<std::pair<unsigned int, unsigned int> > FiniteElement< dim, spacedim >::face_system_to_component_table protected Map between linear dofs and component dofs on face. This is filled with default values in the constructor, but derived classes will have to overwrite the information if necessary. By component, we mean the vector component, not the base element. The information thus makes only sense if a shape function is non-zero in only one component. Definition at line 2451 of file fe.h. template<int dim, int spacedim = dim> std::vector<std::pair<std::pair<unsigned int,unsigned int>,unsigned int> > FiniteElement< dim, spacedim >::system_to_base_table protected For each shape function, store to which base element and which instance of this base element (in case its multiplicity is greater than one) it belongs, and its index within this base element. If the element is not composed of others, then base and instance are always zero, and the index is equal to the number of the shape function. If the element is composed of single instances of other elements (i.e. all with multiplicity one) all of which are scalar, then base values and dof indices within this element are equal to the system_to_component_table. It differs only in case the element is composed of other elements and at least one of them is vector-valued itself. This array has valid values also in the case of vector-valued (i.e. non- primitive) shape functions, in contrast to the system_to_component_table. Definition at line 2470 of file fe.h. template<int dim, int spacedim = dim> std::vector<std::pair<std::pair<unsigned int,unsigned int>,unsigned int> > FiniteElement< dim, spacedim >::face_system_to_base_table protected Likewise for the indices on faces. Definition at line 2476 of file fe.h. template<int dim, int spacedim = dim> BlockIndices FiniteElement< dim, spacedim >::base_to_block_indices protected For each base element, store the number of blocks generated by the base and the first block in a block vector it will generate. Definition at line 2482 of file fe.h. template<int dim, int spacedim = dim> std::vector<std::pair<std::pair<unsigned int, unsigned int>, unsigned int> > FiniteElement< dim, spacedim >::component_to_base_table protected The base element establishing a component. For each component number c, the entries have the following meaning: table[c].first.first Number of the base element for c. This is the index you can pass to base_element(). table[c].first.second Component within the base element for c. This value is between 0 and the n_components() of this base element. table[c].second Index of the multiple of the base element that contains c. This value is between 0 and the element_multiplicity() of this base element. This variable is set to the correct size by the constructor of this class, but needs to be initialized by derived classes, unless its size is one and the only entry is a zero, which is the case for scalar elements. In that case, the initialization by the base class is sufficient. Note This table is filled by FETools::Compositing::build_cell_tables(). Definition at line 2505 of file fe.h. template<int dim, int spacedim = dim> const std::vector<bool> FiniteElement< dim, spacedim >::restriction_is_additive_flags protected A flag determining whether restriction matrices are to be concatenated or summed up. See the discussion about restriction matrices in the general class documentation for more information. Definition at line 2512 of file fe.h. template<int dim, int spacedim = dim> const std::vector<ComponentMask> FiniteElement< dim, spacedim >::nonzero_components protected For each shape function, give a vector of bools (with size equal to the number of vector components which this finite element has) indicating in which component each of these shape functions is non-zero. For primitive elements, there is only one non-zero component. Definition at line 2521 of file fe.h. template<int dim, int spacedim = dim> const std::vector<unsigned int> FiniteElement< dim, spacedim >::n_nonzero_components_table protected This array holds how many values in the respective entry of the nonzero_components element are non-zero. The array is thus a short-cut to allow faster access to this information than if we had to count the non-zero entries upon each request for this information. The field is initialized in the constructor of this class. Definition at line 2530 of file fe.h. template<int dim, int spacedim = dim> const bool FiniteElement< dim, spacedim >::cached_primitivity protected Store whether all shape functions are primitive. Since finding this out is a very common operation, we cache the result, i.e. compute the value in the constructor for simpler access. Definition at line 2537 of file fe.h. The documentation for this class was generated from the following files:
__label__pos
0.771629
Solve the cubic equation: $$x^3-3x^2-3x+1=0 $$ Quick Answer Since the discriminant $$ \Delta <0$$, the cubic equation has three distinct real roots. $$ \Delta=-4$$ $$\begin{cases} x_1=2\sqrt{2}\cos \bigg[\dfrac{1}{3}\cdot \arccos\big(\sqrt{\dfrac{1}{2}}\big)\bigg]+1 \\ x_2=-1 \\ x_3=2\sqrt{2} \cos \bigg[ \dfrac{1}{3}\cdot \arccos\big(\sqrt{\dfrac{1}{2}}\big)+\dfrac{4\pi}{3} \bigg]+1 \end{cases}$$ In decimals, $$\begin{cases} x_1=3.7320508075689 \\ x_2=-1 \\ x_3=0.26794919243112 \end{cases}$$ Detailed Steps on Solution A cubic equation has at least one real root. If the coefficient of leading term is 1, one of solutions could be a factor of the constant term. 1. Factorization Method Find all possible factors for constant $$1$$ $$-1$$ Substitute the factors to the function $$f(x) = x³ - 3x² - 3x + 1$$ and find the one that makes $$f(x) = 0$$. According to factor theorem, $$f(n) = 0$$, if and only if the polynomial $$x³ - 3x² - 3x + 1$$ has a factor $$x-n$$, that is, $$x=n$$ is a root of the equation. Fortunetely, one of the numbers is found to make the equation equal. $$f(-1) = (-1)³ - 3(-1)² - 3(-1) + 1 = -4$$ then we get the first root $$x_1 = -1$$ And the cubic equation can be factored as $$(x +1)(ax^2+bx+c) = 0$$ Next we can use either long division or synthetic division to determine the expression of trinomial Long division Divide the polynomial $$x³ - 3x² - 3x + 1$$ by $$x + 1$$ -4x+1 x + 1-3x²-3x+1 +x² -4x²-3x -4x²-4x x+1 x+1 0 Now we get another factor of the cubic equation $$x² - 4x + 1$$ Solve the quadratic equation: $$x² - 4x + 1 = 0$$ Given $$a =1, b=-4, c=1$$, Use the root solution formula for a quadratic equation, the roots of the equation are given as $$\begin{aligned} \\t&=\dfrac{-b\pm\sqrt{b^2-4ac} }{2a}\\ & =\dfrac{-(-4)\pm\sqrt{(-4)^2-4\cdot 1\cdot 1}}{2 \cdot 1}\\ & =\dfrac{4\pm2\sqrt{3}}{2}\\ & =2\pm\sqrt{3}\\ \end{aligned}$$ Since the discriminat is greater than zero, we get another two real roots: That is, $$\begin{cases} t_2 =2+\sqrt{3} \\ t_3=2-\sqrt{3} \end{cases}$$ Another method to find the roots of a general cubic equation is simplifing it to a depressed form. This method is applicable to all cases, especially to those difficult to find factors. 2. Convert to depressed cubic equation The idea is to convert general form of cubic equation $$ax^3+bx^2+cx+d = 0$$ to the form without quadratic term. $$t^3+pt+q = 0$$ By substituting $$x$$ with $$t - \dfrac{b}{3a}$$, the general cubic equation could be transformed to $$t^3+\dfrac{3ac-b^2}{3a^2}t+\dfrac{2b^3-9abc+27a^2d}{27a^3} = 0 $$ Compare with the depressed cubic equation. Then, $$p = \dfrac{3ac-b^2}{3a^2}$$ $$q = \dfrac{2b^3-9abc+27a^2d}{27a^3} $$ Substitute the values of coefficients, $$p, q$$ is obtained as $$p = \dfrac{3\cdot 1\cdot (-3)-(-3)^2}{3\cdot 1^2}=-6$$ $$q = \dfrac{2\cdot (-3)^3-9\cdot1\cdot (-3)\cdot (-3)+27\cdot 1^2\cdot1}{27\cdot 1^3}=-4$$ Use the substitution to transform Let $$p$$ and $$q$$ being the coefficient of the linean and constant terms, the depressed cubic equation is expressed as. $$t^3 +pt+q=0$$ Let $$x=t+1$$ The cubic equation $$x³ - 3x² - 3x + 1=0$$ is transformed to $$t^3 -6t-4=0$$ 3. Cardano's solution Let $$t=u-v$$ Cube both sides and extract common factor from two middle terms after expanding the bracket. $$\begin{aligned} \\t^3&=(u-v)^3\\ & =u^3-3u^2v+3uv^2-v^3\\ & =-3uv(u-v)+u^3-v^3\\ \end{aligned}$$ Since $$u-v=t$$, substitution gives a linear term for the equation. Rearrange terms. $$x^3+3uvx-u^3+v^3=0$$ Compare the cubic equation with the original one (1) $$\begin{cases} 3uv=-6\quad\text{or}\quad v=-\dfrac{2}{u}\\ v^3-u^3=-4\\ \end{cases}$$ $$v=-\dfrac{2}{u}$$ gives relationship between the two variables. Substitute the value of $$v$$ to the second equation $$\Big(-\dfrac{2}{u}\Big)^3-u^3=-4$$ Simplifying gives, $$u^3+8\dfrac{1}{u^3}-4=0$$2 Let $$m=u^3$$, then the equation is transformed to a quadratic equation in terms of $$m$$. Once the value of $$m$$ is determined, $$v^3$$ could be determined by $$v^3=-4+u^3$$. $$m^2-4m+8=0$$ Sovling the quadratic euqation will give two roots (some may be equal). Here we only cosider one case with positive sign before the square root radical since the negative case will produce the same result. $$\begin{aligned} \\u^3=m&=2+\dfrac{1}{2}\sqrt{\Big(-4^2\Big)-4\cdot 8}\\ & =2+\dfrac{1}{2}\sqrt{16-32}\\ & =2+\dfrac{1}{2}\sqrt{16}i\\ & =2+2i\\ \end{aligned}$$ $$v^3$$ can be determined by the equation we deduced $$v^3-u^3=-4$$. Then, $$\begin{aligned} \\v^3&=-4+u^3\\ & =-4+2+2i\\ & =-2+2i\\ \end{aligned}$$ Now we have, $$u^3=2+2i$$ and $$v^3=-2+2i$$ Evaluating the simplest cubic equation $$x^3-A=0$$, it has 3 roots, in which the first root is a real number or a complex number. The second and third are expressed in the product of cubic root of unity and the first one. If $$ω = \dfrac{-1+i\sqrt{3}}{2}$$, then its reciprocal is equal to its conjugate, $$\dfrac{1}{ω}=\overline{ω}$$. $$\begin{cases} r_1=\sqrt[3]{A}\\ r_2=\dfrac{-1+i\sqrt{3}}{2}\cdot \sqrt[3]{A}\\ r_3=\dfrac{-1-i\sqrt{3}}{2}\cdot \sqrt[3]{A}\\ \end{cases}$$ Similary, taking cubic root for $$u^3$$ and $$v^3$$ also gives 3 roots. $$\begin{cases} u_1=\sqrt[3]{2+2i}\\ u_2=\dfrac{-1+i\sqrt{3}}{2}\cdot \sqrt[3]{2+2i}\\ u_3=\dfrac{-1-i\sqrt{3}}{2}\cdot \sqrt[3]{2+2i}\\ \end{cases}$$ For $$v_2$$ and $$v_3$$, the complex numbers before radicals are the conjugates of those for $$u_2$$ and $$u_3$$, which can be verified by the reciprocal property of the cubic root of unity from the equation $$v=-\dfrac{2}{u}$$. The radicand can be taken as the negative conjugate of that in $$u_1$$, $$u_2$$ and $$u_3$$, which is the same in value. $$\begin{cases} v_1=\sqrt[3]{-2+2i}\\ v_2=\dfrac{-1-i\sqrt{3}}{2}\cdot \sqrt[3]{-2+2i}\\ v_3=\dfrac{-1+i\sqrt{3}}{2}\cdot \sqrt[3]{-2+2i}\\ \end{cases}$$ Since $$t=u-v$$, the firt root $$t_1$$ can be expressed as the sum of cubic root of two conjugate complex numbers $$t_1=\sqrt[3]{2+2i}+\sqrt[3]{2-2i}$$ Let $$u^3_1=z=2+2i$$, and $$z$$ can be expressed in trigonomic form $$r(\cos θ + i \sin θ)$$, where $$r$$ and $$θ$$ are the modulus and principle argument of the complex number. Then $$-v^3_1$$ is the conjugate $$\overline{z}=2-2i$$, or $$\overline{z} = r(\cos θ - i \sin θ)$$ Now let's calculate the value of $$r$$ and $$ θ$$. $$\begin{aligned} \\r&=\sqrt{\Big(2\Big)^2+\Big(2\Big)^2}\\ & =2\sqrt{2}\\ \end{aligned}$$ $$\cosθ=\dfrac{2}{2\sqrt{2}}=\sqrt{\dfrac{1}{2}}$$ The argument is the inverse function of the cosθ $$θ=\arccos\Big(\sqrt{\dfrac{1}{2}}\Big)$$ Using de Moivre’s formula, the cubic root of $$z$$ could be determined. $$\begin{aligned} \\u_1=\sqrt[3]{z}&=\sqrt[3]{r}(\cos\dfrac{θ}{3}+i\sin\dfrac{θ}{3})\\ & =\sqrt[3]{2\sqrt{2}}\Big[\cos\dfrac{1}{3}\arccos\Big(\sqrt{\dfrac{1}{2}}\Big)+\sin\dfrac{1}{3}\arccos\Big(\sqrt{\dfrac{1}{2}}\Big)\Big]\\ & =\sqrt{2}\Big[\cos\dfrac{1}{3}\arccos\Big(\sqrt{\dfrac{1}{2}}\Big)+\sin\dfrac{1}{3}\arccos\Big(\sqrt{\dfrac{1}{2}}\Big)\Big]\\ \end{aligned}$$ Since $$-v^3$$ is the conjugate of $$z$$ as we mentioned above, $$\begin{aligned} \\v_1=-\sqrt[3]{\overline{z}}&=-\sqrt[3]{r}(\cos\dfrac{θ}{3}-i\sin\dfrac{θ}{3})\\ & =\sqrt{2}\Big[-\cos\dfrac{1}{3}\arccos\Big(\sqrt{\dfrac{1}{2}}\Big)+\sin\dfrac{1}{3}\arccos\Big(\sqrt{\dfrac{1}{2}}\Big)\Big]\\ \end{aligned}$$ The first root is the difference of $$u_1$$ and $$v_1$$. $$\begin{aligned} \\t_1&=u_1-v_1\\ & =2\cos\dfrac{1}{3}\arccos\Big(\sqrt{\dfrac{1}{2}}\Big)\\ & =2\sqrt{2}\cos\dfrac{1}{3}\arccos\Big(\sqrt{\dfrac{1}{2}}\Big)\\ \end{aligned}$$ The second root is the difference of $$u_2$$ and $$v_2$$, in which $$u_2$$ is the product of the cubic root of $$z$$ and the cubic root of unity, $$v_2$$ is the product of the negative of the conjugate of $$z$$ and the conjugate of the cubic root of unity. $$\begin{aligned} \\t_2&=u_2-v_2\\ & =\dfrac{-1+i\sqrt{3}}{2}\sqrt[3]{2+2i}+\dfrac{-1-i\sqrt{3}}{2}\sqrt[3]{2-2i}\\ & =\dfrac{-1+i\sqrt{3}}{2}\sqrt[3]{r}(\cos\dfrac{θ}{3}+i\sin\dfrac{θ}{3})+\dfrac{-1-i\sqrt{3}}{2}\sqrt[3]{r}(\cos\dfrac{θ}{3}-i\sin\dfrac{θ}{3})\\ & =-\sqrt[3]{r}(\cos\dfrac{ θ}{3} + \sqrt{3} \sin\dfrac{ θ}{3} ) \\ & =2\sqrt[3]{r}\cos\Big(\dfrac{θ}{3}+ \dfrac{4\pi}{3}\Big)\\ & =2\sqrt{2}\cos\Big(\dfrac{1}{3}\arccos\Big(\sqrt{\dfrac{1}{2}}\Big)+ \dfrac{4\pi}{3}\Big)\\ \end{aligned}$$ The third root is the difference of $$u_3$$ and $$v_3$$, in which $$u_3$$ is the product of the cubic root of $$z$$ and the conjugate of the cubic root of unity, $$v_3$$ is the product of the negative of the conjugate of $$z$$ and the cubic root of unity. $$\begin{aligned} \\t_3&=u_3-v_3\\ & =\dfrac{-1-i\sqrt{3}}{2}\sqrt[3]{2+2i}+\dfrac{-1+i\sqrt{3}}{2}\sqrt[3]{2-2i}\\ & =\dfrac{-1-i\sqrt{3}}{2}\sqrt[3]{r}(\cos\dfrac{θ}{3}+i\sin\dfrac{θ}{3})+\dfrac{-1+i\sqrt{3}}{2}\sqrt[3]{r}(\cos\dfrac{θ}{3}-i\sin\dfrac{θ}{3})\\ & =\sqrt[3]{r}(-\cos\dfrac{θ}{3}+\sqrt{3} \sin \dfrac{ θ}{3})\\ & =2\sqrt[3]{r}\cos\Big(\dfrac{θ}{3}+ \dfrac{2\pi}{3}\Big)\\ & =2\sqrt{2}\cos\Big(\dfrac{1}{3}\arccos\Big(\sqrt{\dfrac{1}{2}}\Big)+ \dfrac{2\pi}{3}\Big)\\ & = 2\sqrt{2}\cdot -\sqrt{\dfrac{1}{2}}\\ & = -2\\ \end{aligned}$$ 4. Vieta's Substitution In Cardano' solution, $$t$$ is defined as the difference of $$u$$ and $$v$$. If we substitute the value of $$v$$ (4) into (2), we get the equation. $$t=u+\dfrac{2}{u}$$. And then substitute the equation to the cubic equation $$t^3-6t-4=0$$. This method is called Vieta's Substitution for solving a cubic equation, which simplied the Cardano' solution. The substitution expression can be obtained by the following formula directly. $$t=u-\dfrac{p}{3u}$$ Substitute the expression $$t=u+\dfrac{2}{u}$$ to the cubic equation $$\Big(u+\dfrac{2}{u}\Big)^3-6\Big(u+\dfrac{2}{u}\Big)-4=0$$ Expand brackets and cancel the like terms $$u^3+\cancel{6u^2\dfrac{1}{u}}+\cancel{12u\dfrac{1}{u^2}}+8\dfrac{1}{u^3}-\cancel{6u}-\cancel{12\dfrac{1}{u}}-4=0$$ Then we get the same equation as (2) $$u^3+8\dfrac{1}{u^3}-4=0$$ The rest of the steps will be the same as those of Cardano's solution 5. Euler's Solution $$t^3-6t-4=0$$ Move the linear term and constant of (1) to its right hand side. We get the following form of the equation. $$t^3=6t+4 $$3 Let the root of the cubic equation be the sum of two cubic roots $$t=\sqrt[3]{r_1}+\sqrt[3]{r_2} $$4 in which $$r_1$$ and $$r_2$$ are two roots of a quadratic equation $$z^2-\alpha z+ β=0 $$5 Using Vieta's Formula, the following equations are established. $$r_1+r_2 = \alpha \quad \text{and} \quad r_1r_2 = β $$ To determine $$\alpha$$, $$β$$, cube both sides of the equation (4) $$t^3=3\sqrt[3]{r_1r_2}(\sqrt[3]{r_1}+\sqrt[3]{r_2})+r_1+r_2 $$ Substituting, the equation is simplified to $$t^3=3\sqrt[3]{β}t+\alpha $$ Compare the cubic equation with (3), the following equations are established $$\begin{cases} 3\sqrt[3]{β}=6\\ \alpha=4\\ \end{cases}$$ Solving for $$β$$ gives $$β=8 $$ So the quadratic equation (5) is determined as $$z^2-4z+8=0$$6 Solving the quadratic equation yields $$\begin{cases} r_1=2+2i\approx2+2i\\ r_2=2-2i\approx2-2i\\ \end{cases}$$ Therefore, one of the roots of the cubic equation could be obtained from (4). $$t_1=\sqrt[3]{2+2i}+\sqrt[3]{2-2i} $$ in decimals, $$t_1=2.7320508075689 $$ However, since the cube root of a quantity has triple values, The other two roots could be determined as, $$t_2=\dfrac{-1+i\sqrt{3}}{2}\sqrt[3]{2+2i}+\dfrac{-1-i\sqrt{3}}{2}\sqrt[3]{2-2i} $$ $$t_3=\dfrac{-1-i\sqrt{3}}{2}\sqrt[3]{2+2i}+\dfrac{-1+i\sqrt{3}}{2}\sqrt[3]{2-2i} $$ Since the expression involes cubic root of complex number, the final result can be deduced by using trigonometric method as shown in Cardano's solution. For the equation $$t^3 -6t-4$$, we have $$p=-6$$ and $$q = -4$$ Calculate the discriminant The nature of the roots are determined by the sign of the discriminant. Since $$p$$ is negative, the discriminant will be less than zero if the absolute value of $$p$$ is large enough. $$\begin{aligned} \\\Delta&=\dfrac{q^2}{4}+\dfrac{p^3}{27}\\ & =\dfrac{(-4)^2}{4}+\dfrac{(-6)^3}{27}\\ & =4-8\\ & =-4\\ \end{aligned}$$ 5.1 Use the root formula directly If $$\Delta < 0$$, then there are 3 distinct real roots for the cubic equation $$\begin{cases} t_1 &= 2\sqrt[3]{r} \cos\dfrac{ θ}{3} \\ t_2 & = 2\sqrt[3]{r}\cos\Big( \dfrac{ θ}{3}+\dfrac{2\pi}{3} \Big) \\ t_3&= 2\sqrt[3]{r}\cos\Big( \dfrac{ θ}{3}+\dfrac{4\pi}{3} \Big) \end{cases}$$ where $$\theta = \arccos(\dfrac{3q}{2p}\sqrt{-\dfrac{3}{p} } )$$ and $$\sqrt[3]{r} = \sqrt{\dfrac{-p}{3} } $$ Substitute the values of $$p$$ and $$q$$ to determine the value of $$\theta$$ $$\begin{aligned} \\\theta&= \arccos(\dfrac{3q}{2p}\sqrt{-\dfrac{3}{p} } )\\ & =\arccos\big(\dfrac{3\cdot -4}{2 \cdot -6}\sqrt{-\dfrac{3}{-6} }\big)\\ & =\arccos\big(\dfrac{1}{2}\cdot 3\cdot 4\cdot \dfrac{1}{6}\cdot \sqrt{\dfrac{1}{2}}\big)\\ & = \arccos\big(\sqrt{\dfrac{1}{2}}\big)\\ \end{aligned}$$ Then we can determine one third of $$\theta$$ $$\dfrac{\theta}{3} = \dfrac{1}{3} \arccos\big(\sqrt{\dfrac{1}{2}}\big) $$ Substitute the value of $$p$$ to determine the cubic root of $$r$$ $$\begin{aligned} \\\sqrt[3]{r}& =\sqrt{\dfrac{-p}{3}}\\ & =\sqrt{\frac{-(-6)}{3}} \\ & =\sqrt{2}\\ \end{aligned}$$ Substitute the value of $$\dfrac{\theta}{3}$$ and $$\sqrt[3]{r}$$ to the root formulas. Then we get $$\begin{aligned} \\t_1&= 2\sqrt[3]{r} \cos\dfrac{ θ}{3}\\ & =2\cdot \sqrt{2}\cdot \cos \bigg[ \dfrac{1}{3}\cdot\arccos\big(\sqrt{\dfrac{1}{2}}\big)\bigg]\\ & =2\sqrt{2}\cos \bigg[\dfrac{1}{3}\cdot \arccos\big(\sqrt{\dfrac{1}{2}}\big)\bigg]\\ & \approx 2\sqrt{2} \cos 0.26179938779915\\ & \approx 2.7320508075689\\ \end{aligned}$$ $$\begin{aligned} \\t_2&=2\sqrt{2} \cos \bigg[ \dfrac{1}{3}\cdot \arccos\big(\sqrt{\dfrac{1}{2}}\big)+\dfrac{2\pi}{3}\bigg]\\ & =2\sqrt{2}\cdot\Big(-\sqrt{\dfrac{1}{2}}\Big)\\ & =-2\\ \end{aligned}$$ A querstion arises. If $$\cos\theta = 1\sqrt{\dfrac{1}{2}}$$, show that $$\cos \Big(\dfrac{\theta}{3}+\dfrac{2\pi}{3}\Big) = -\sqrt{\dfrac{1}{2}}$$ $$\begin{aligned} \\t_3&=2\sqrt{2} \cos \bigg[ \dfrac{1}{3}\cdot \arccos\big(\sqrt{\dfrac{1}{2}}\big)+\dfrac{4\pi}{3} \bigg]\\ & \approx -0.73205080756888\\ \end{aligned}$$ 5.2 Trigonometric Method For a depressed cubic equation, another method is to convert the equation to the form of triple angle identity. By comparing the trigonometric value of the triple angle, the value of the single angle is determined. And then the roots of the cubic equation can be found. Compare the given equation with the standard depressed cubic equation $$t^3 +pt+q=0$$, we get $$p=-6$$ and $$q = -4$$ Using the formula $$t= 2\sqrt{\dfrac{-p}{3}}u $$ to introduce an intermediate variable $$u$$ so that the given equation is transformed to a new equation in terms of $$u$$, which is analogous to a trignometric triple angle identity. Then, $$t= 2\sqrt{2}u $$ Substitute to the cubic equation and simplify. $$\big(2\sqrt{2}u\big)^3 -6\big(2\sqrt{2}u\big)-4=0$$ Expand and simplify coefficients $$8\cdot2\sqrt{2}u^3 -\dfrac{12}{1}\sqrt{2}u-4=0 $$ Continute simplifying $$16\sqrt{2}u^3-12\sqrt{2}u-4=0$$ Cancel the common factors of coefficients of each term $$4\sqrt{2}u^3-3\sqrt{2}u-1=0$$ Dividing the equation by $$\sqrt{2}$$ gives $$ 4u^3-3u-\sqrt{\dfrac{1}{2}}=0$$ The equation becomes the form of a triple angle identity for cosine function. $$4\cos^3θ-3\cos θ-\cos3θ =0$$ Comparing the equation with triple angle identity gives $$u =\cos θ$$ and $$\cos3θ =\sqrt{\dfrac{1}{2}}$$ The fact that the cosine function is periodic with period 2π implies the following equation. $$\cos(3θ-2πk) =\sqrt{\dfrac{1}{2}}$$ Solving for θ yields $$ θ =\dfrac{1}{3}\cdot \arccos\Big(\sqrt{\dfrac{1}{2}}\Big)+\dfrac{2πk}{3} , $$ in witch $$k=0,1,2$$. Then we can determine the value of $$u$$, that is $$\cos θ$$ $$\begin{aligned} \\u&= \cos θ\\ & = \cos\Big[\dfrac{1}{3}\cdot \arccos\Big(\sqrt{\dfrac{1}{2}}\Big)+\dfrac{2πk}{3}\Big]\\ \end{aligned}$$ and subsiquently, $$\begin{aligned} \\t&= 2\sqrt{2}u\\ & = 2\sqrt{2} \cos\Big[\dfrac{1}{3}\cdot \arccos\Big(\sqrt{\dfrac{1}{2}}\Big)+\dfrac{2πk}{3}\Big]\\ \end{aligned}$$ Substituting $$k=0,1,2$$ gives the roots of solution set. $$\begin{aligned} \\t_1&= 2\sqrt{2}\cos\Big[\dfrac{1}{3}\cdot \arccos\Big(\sqrt{\dfrac{1}{2}}\Big)\Big]\\ & \approx2.7320508075689\\ \end{aligned}$$ Since $$\cos 3\theta = \sqrt{\dfrac{1}{2}}$$, by using triple angle identity we could derive $$\cos\Big[ \theta+\dfrac{2π}{3}\Big] = -\sqrt{\dfrac{1}{2}}$$ or it can be expressed as, $$\cos\Big[\dfrac{1}{3}\cdot \arccos\Big(\sqrt{\dfrac{1}{2}}\Big)+\dfrac{2π}{3}\Big]=-\sqrt{\dfrac{1}{2}}$$ $$\begin{aligned} \\t_2&= 2\sqrt{2}\cos\Big[\dfrac{1}{3}\cdot \arccos\Big(\sqrt{\dfrac{1}{2}}\Big)+\dfrac{2π}{3}\Big]\\ & = 2\sqrt{2}\cdot -\sqrt{\dfrac{1}{2}}\\ & = -2\\ \end{aligned}$$ $$\begin{aligned} \\t_3&= 2\sqrt{2}\cos\Big[\dfrac{1}{3}\cdot \arccos\Big(\sqrt{\dfrac{1}{2}}\Big)+\dfrac{4π}{3}\Big]\\ & \approx-0.73205080756888\\ \end{aligned}$$ which shows the trigonometric method gets the same solution set as that by using roots formula. Roots of the general cubic equation Since $$x = t - \dfrac{b}{3a}$$, substituting the values of $$t$$, $$a$$ and $$b$$ gives $$\begin{aligned} \\x_1&=t_1-\dfrac{b}{3a}\\ & =2\sqrt{2}\cos \bigg[\dfrac{1}{3}\cdot \arccos\big(\sqrt{\dfrac{1}{2}}\big)\bigg]+1\\ \end{aligned}$$ $$\begin{aligned} \\x_2&=t_2-\dfrac{b}{3a}\\ & =-2+1\\ & =-1\\ \end{aligned}$$ $$\begin{aligned} \\x_3&=t_3-\dfrac{b}{3a}\\ & =2\sqrt{2} \cos \bigg[ \dfrac{1}{3}\cdot \arccos\big(\sqrt{\dfrac{1}{2}}\big)+\dfrac{4\pi}{3} \bigg]+1\\ \end{aligned}$$ 6. Summary In summary, we have tried the method of factorization, cubic root formula, trigonometric to explore the solutions of the equation. The cubic equation $$x³ - 3x² - 3x + 1=0$$ is found to have three real roots . Exact values and approximations are given below. $$\begin{cases} x_1=2\sqrt{2}\cos \bigg[\dfrac{1}{3}\cdot \arccos\big(\sqrt{\dfrac{1}{2}}\big)\bigg]+1 \\ x_2=-1 \\ x_3=2\sqrt{2} \cos \bigg[ \dfrac{1}{3}\cdot \arccos\big(\sqrt{\dfrac{1}{2}}\big)+\dfrac{4\pi}{3} \bigg]+1 \end{cases}$$ Convert to decimals, $$\begin{cases} x_1=3.7320508075689 \\ x_2=-1 \\ x_3=0.26794919243112 \end{cases}$$ Using the method of factorization, the roots are derived to the following forms $$\begin{cases} x_1=-1 \\ x_2=2-\sqrt{3} \\ x_3=2+\sqrt{3} \end{cases}$$ 7. Math problems derived by the cubic equation In the course of solving the cubic equation, $$x³ - 3x² - 3x + 1$$, the following intesting math problem is discovered. If $$\cos\theta = 1\sqrt{\dfrac{1}{2}}$$, show that $$\cos \Big(\dfrac{\theta}{3}+\dfrac{2\pi}{3}\Big) = -\sqrt{\dfrac{1}{2}}$$ 8. Graph for the function $$f(x) = x³ - 3x² - 3x + 1$$ Since the discriminat is less than zero, the curve of the cubic function $$f(x) = x³ - 3x² - 3x + 1$$ has 3 intersection points with horizontal axis. More cubic equations Scroll to Top
__label__pos
1
Magento Commerce only Configure nginx and Elasticsearch Contents Overview of secure web server communication This topic discusses an example of securing communication between your web server and Elasticsearch using a combination of Transport Layer Security (TLS) encryption and HTTP Basic authentication. You can optionally configure other types of authentication as well; we provide references for that information. (An older term, Secure Sockets Layer (SSL), is frequently used interchangeably with TLS. In this topic, we refer to TLS.) Unless otherwise noted, all commands in this topic must be entered as a user with root privileges. Recommendations We recommend the following: • Your web server uses TLS. TLS is beyond the scope of this topic; however, we strongly recommend you use a real certificate in production and not a self-signed certificate. • Elasticsearch runs on the same host as a web server. Running Elasticsearch and the web server on different hosts is beyond the scope of this topic. The advantage of putting Elasticsearch and the web server on the same host is that it makes intercepting encrypted communication impossible. The Elasticsearch web server doesn’t have to be the same as the Magento web server; for example, Magento can run Apache and Elasticsearch can run nginx. More information about TLS See one of the following resources: Set up a proxy This section discusses how to configure nginx as an unsecure proxy so that Magento can use Elasticsearch running on this server. This section does not discuss setting up HTTP Basic authentication; that is discussed in Secure communication with nginx. The reason the proxy is not secured in this example is it's easier to set up and verify. You can use TLS with this proxy if you want; to do so, make sure you add the proxy information to your secure server block configuration. See one of the following sections for more information: Step 1: Specify additional configuration files in your global nginx.conf Make sure your global /etc/nginx/nginx.conf contains the following line so it loads the other configuration files discussed in the following sections: include /etc/nginx/conf.d/*.conf; Step 2: Set up nginx as a proxy This section discusses how to specify who can access the nginx server. 1. Use a text editor to create a new file /etc/nginx/conf.d/magento_es_auth.conf with the following contents: server { listen 8080; location / { proxy_pass http://localhost:9200; } } 2. Restart nginx: service nginx restart 3. Verify the proxy works by entering the following command: curl -i http://localhost:<proxy port>/_cluster/health For example, if your proxy uses port 8080: curl -i http://localhost:8080/_cluster/health Messages similar to the following display to indicate success: HTTP/1.1 200 OK Date: Tue, 23 Feb 2016 20:38:03 GMT Content-Type: application/json; charset=UTF-8 Content-Length: 389 Connection: keep-alive {"cluster_name":"elasticsearch","status":"yellow","timed_out":false,"number_of_nodes":1,"number_of_data_nodes":1,"active_primary_shards":5,"active_shards":5,"relocating_shards":0,"initializing_shards":0,"unassigned_shards":5,"delayed_unassigned_shards":0,"number_of_pending_tasks":0,"number_of_in_flight_fetch":0,"task_max_waiting_in_queue_millis":0,"active_shards_percent_as_number":50.0} 4. Continue with the next section. Configure Magento to use Elasticsearch This section discusses the minimum settings you must choose to test Elasticsearch with Magento 2. For additional details about configuring Elasticsearch, see the Magento Commerce User Guide. To configure Magento to use Elasticsearch: 1. Log in to the Magento Admin as an administrator. 2. Click Stores > Settings > Configuration > Catalog > Catalog > Catalog Search. 3. From the Search Engine list, click Elasticsearch or Elasticsearch 5.0+ as the following figure shows. (The Elasticsearch 5.0+ option is not available for Magento 2.1.) 4. The following table discusses only the configuration options required to test the connection with Magento. Unless you changed Elasticsearch server settings, the defaults should work. Skip to the next step. Option Description Elasticsearch Server Hostname Enter the fully qualified hostname or IP address of the machine running Elasticsearch. Magento Commerce (Cloud): Get this value from your integration system. Elasticsearch Server Port Enter the Elasticsearch web server proxy port. In our example, the port is 8080 but if you're using a secure proxy, it's typically 443. Magento Commerce (Cloud): Get this value from your integration system. Elasticsearch Index Prefix Enter the Elasticsearch index prefix. If you use a single Elasticsearch instance for more than one Magento installation (Staging and Production environments), you must specify a unique prefix for each installation. Otherwise, you can use the default prefix magento2. Enable Elasticsearch HTTP Auth Click Yes only if you enabled authentication for your Elasticsearch server. If so, provide a username and password in the provided fields. 5. Click Test Connection. One of the following displays: Result Meaning Magento successfully connected to the Elasticsearch server. Continue with Configure Apache and Elasticsearch or Configure nginx and Elasticsearch. Try the following: • Make sure the Elasticsearch server is running. • If the Elasticsearch server is on a different host from Magento, log in to the Magento server and ping the Elasticsearch host. Resolve network connectivity issues and test the connection again. • Examine the command window in which you started Elasticsearch for stack traces and exceptions. You must resolve those before you continue. In particular, make sure you started Elasticsearch as a user with root privileges. • Make sure that UNIX firewall and SELinux are both disabled, or set up rules to enable Elasticsearch and Magento to communicate with each other. • Verify the value of the Elasticsearch Server Hostname field. Make sure the server is available. You can try the server's IP address instead. • Use the command netstat -an | grep listen-port command to verify that the port specified in the Elasticsearch Server Port field is not being used by another process. For example, to see if Elasticsearch is running on its default port, use the following command: netstat -an | grep 9200 If Elasticsearch is running on port 9200, it displays similar to the following: tcp 0 0 :::9200 :::* LISTEN Reindexing catalog search and refreshing the full page cache After you change Magento’s Elasticsearch configuration, you must reindex the catalog search index and refresh the full page cache using the Admin or command line. To refresh the cache using the Admin: 1. In the Admin, click System > Cache Management. 2. Select the checkbox next to Page Cache. 3. From the Actions list in the upper right, click Refresh. The following figure shows an example. To clean the cache using the command line, use the magento cache:clean command. To reindex using the command line: 1. Log in to your Magento server as, or switch to, the Magento file system owner. 2. Enter any of the following commands: Enter the following command to reindex the catalog search index only: php <your Magento install dir>/bin magento indexer:reindex catalogsearch_fulltext Enter the following command to reindex all indexers: php <your Magento install dir>/bin magento indexer:reindex 3. Wait while the reindexing completes. Unlike the cache, indexers are updated by a cron job. Make sure cron is enabled before you start using Elasticsearch. Secure communication with nginx This section discusses how to set up HTTP Basic authentication with your secure proxy. Use of TLS and HTTP Basic authentication together prevents anyone from intercepting communication with Elasticsearch or with your Magento server. Because nginx natively supports HTTP Basic authentication, we recommend it over, for example, Digest authentication, which isn’t recommended in production. Additional resources: See the following sections for more information: Step 1: Create a password We recommend you use the Apache htpasswd command to encode passwords for a user with access to Elasticsearch (named magento_elasticsearch in this example). To create a password: 1. Enter the following command to determine if htpasswd is already installed: which htpasswd If a path displays, it is installed; if the command returns no output, htpasswd is not installed. 2. If necessary, install htpasswd: • Ubuntu: apt-get -y install apache2-utils • CentOS: yum -y install httpd-tools 3. Create a /etc/nginx/passwd directory to store passwords: mkdir -p /etc/nginx/passwd htpasswd -c /etc/nginx/passwd/.<filename> <username> For security reasons, <filename> should be hidden; that is, it must start with a period. An example follows. Example: mkdir -p /etc/nginx/passwd htpasswd -c /etc/nginx/passwd/.magento_elasticsearch magento_elasticsearch Follow the prompts on your screen to create the user’s password. 4. (Optional). To add another user to your password file, enter the same command without the -c (create) option: htpasswd /etc/nginx/passwd/.<filename> <username> 5. Verify that the contents of /etc/nginx/passwd is correct. Step 3: Set up access to nginx This section discusses how to specify who can access the nginx server. The example shown is for an unsecure proxy. To use a secure proxy, add the following contents (except the listen port) to your secure server block. Use a text editor to modify either /etc/nginx/conf.d/magento_es_auth.conf (unsecure) or your secure server block with the following contents: server { listen 8080; server_name 127.0.0.1; location / { limit_except HEAD { auth_basic "Restricted"; auth_basic_user_file /etc/nginx/passwd/.htpasswd_magento_elasticsearch; } proxy_pass http://127.0.0.1:9200; proxy_redirect off; proxy_set_header Host $host; proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; } location /_aliases { auth_basic "Restricted"; auth_basic_user_file /etc/nginx/passwd/.htpasswd_magento_elasticsearch; proxy_pass http://127.0.0.1:9200; proxy_redirect off; proxy_set_header Host $host; proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; } include /etc/nginx/auth/*.conf; } The Elasticsearch listen port shown in the preceding example are examples only. For security reasons, we recommend you use a non-default listen port for Elasticsearch. Step 4: Set up a restricted context for Elasticsearch This section discusses how to specify who can access the Elasticsearch server. 1. Enter the following command to create a new directory to store the authentication configuration: mkdir /etc/nginx/auth/ 2. Use a text editor to create a new file /etc/nginx/auth/magento_elasticsearch.conf with the following contents: location /elasticsearch { auth_basic "Restricted - elasticsearch"; auth_basic_user_file /etc/nginx/passwd/.htpasswd_magento_elasticsearch; proxy_pass http://127.0.0.1:9200; proxy_redirect off; proxy_set_header Host $host; proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; } 3. If you set up a secure proxy, delete /etc/nginx/conf.d/magento_es_auth.conf. 4. Restart nginx and continue with the next section: service nginx restart Verify communication is secure This section discusses two ways to verify that HTTP Basic authentication is working: • Using a curl command to verify you must enter a username and password to get cluster status • Configuring HTTP Basic authentication in the Magento Admin Use a curl command to verify cluster status Enter the following command: curl -i http://<hostname, ip, or localhost>:<proxy port>/_cluster/health For example, if you enter the command on the Elasticsearch server and your proxy uses port 8080: curl -i http://localhost:8080/_cluster/health The following message displays to indicate authentication failed: HTTP/1.1 401 Unauthorized Date: Tue, 23 Feb 2016 20:35:29 GMT Content-Type: text/html Content-Length: 194 Connection: keep-alive WWW-Authenticate: Basic realm="Restricted" <html> <head><title>401 Authorization Required</title></head> <body bgcolor="white"> <center><h1>401 Authorization Required</h1></center> </body> </html> Now try the following command: curl -i -u <username>:<password> http://<hostname, ip, or localhost>:<proxy port>/_cluster/health For example: curl -i -u magento_elasticsearch:mypassword http://localhost:8080/_cluster/health This time the command succeeds with a message similar to the following: HTTP/1.1 200 OK Date: Tue, 23 Feb 2016 20:38:03 GMT Content-Type: application/json; charset=UTF-8 Content-Length: 389 Connection: keep-alive {"cluster_name":"elasticsearch","status":"yellow","timed_out":false,"number_of_nodes":1,"number_of_data_nodes":1,"active_primary_shards":5,"active_shards":5,"relocating_shards":0,"initializing_shards":0,"unassigned_shards":5,"delayed_unassigned_shards":0,"number_of_pending_tasks":0,"number_of_in_flight_fetch":0,"task_max_waiting_in_queue_millis":0,"active_shards_percent_as_number":50.0} Configure HTTP Basic authentication in the Magento Admin Perform the same tasks as discussed in Configure Magento to use Elasticsearch except click Yes from the Enable Elasticsearch HTTP Auth list and enter your username and password in the provided fields. Click Test Connection to make sure it works and then click Save Config. You must flush the Magento cache and reindex before you continue. Next Configure Elasticsearch stopwords
__label__pos
0.924644
NetBurner 3.3 Creating a GET Request Handler There are two common methods for moving data from the client web browser to the web server on an embedded platform: HTML Forms using POST, and storing the data in the URL. For example, an e-commerce applications might store product information such as: http://www.store.com/orderform?type=order123 , which is storing the data type=order123 in the URL. Everything following the ‘?’ character is ignored by the browser, so your application can store whatever data it needs after the character. One advantage of this method is that the application is stateful, meaning multiple users can access the same application and each user’s session maintains its specific data in the URL. When a web browser requests something from a web server, such as an HTML page or image, it makes a GET request. The web server normally handles static web pages and dynamic web pages with the CPPCALL and VARIABLE tags, but your application can intercept the request and take control of the processing using a callback function object called CallBackFunctionPageHandler. When you declare the instance of the object you specify the name of the request to intercept and a pointer to the function in your application to process the request. For example, to take control of processing for a HTML page named setcookie.html: int setCookieGetReqCallback( int sock, HTTP_Request &httpRequestInfo ) { iprintf( "HTTP request: %s\r\n", httpRequestInfo.pURL ); // Set the cookie value to "MyCookie ". Note that you need a trailing space. SendHTMLHeaderWCookie(sock, "MyCookie "); // Since we are handling the GET request, we need to send the web page SendFileFragment("setcookie.html", sock); return 1; // Notify the system GET handler we handled the request } "setcookie.html", // Web page to intercept setCookieGetReqCallback, // Pointer to callback function tGet, // Type of request, GET 0, // Password level, none true ); // Take responsibility for entire response, the web // server will not send any data of its own A callback function can be created for any type of URL request. If you application has the ability to create a graph or image named TempGraph.gif, you could use a callback for TempGraph.gif instead of setcookie.html in the above example. There are examples in the \nburn\examples\web directory. You can also use the ‘*’ as a wild card for matching a number of pages to process in a single callback. For example, to process all requests that begin with “set”, specify “set*”. SendHTMLHeaderWCookie void SendHTMLHeaderWCookie(int sock, char *cookie) Send a HTML response header and cookie. Definition: httpinternal.cpp:266 tGet @ tGet GET request. Definition: http.h:27 CallBackFunctionPageHandler Implements the HtmlPageHandler class as a function pointer callback for GET requests. Definition: http.h:158 HTTP_Request::pURL PSTR pURL Request URL. Definition: http.h:61 HTTP_Request HTTP Request Structure. Definition: http.h:60 SendFileFragment int32_t SendFileFragment(char const *name, int32_t fd, PCSTR url=NULL) Send a file fragment without a header. Definition: htmldecomp.cpp:347
__label__pos
0.622804
JavaScript Clone or download Fetching latest commit… Cannot retrieve the latest commit at this time. Permalink Failed to load latest commit information. bench dist docs scripts src test .eslintrc .gitignore .npmignore .prettierrc .travis.yml CHANGELOG.md LICENSE.md README.md bower.json klipse-github-docs.config.js package-lock.json package.json rollup.config.js README.md Fastener · GitHub stars npm Zippers are a powerful abstraction for implementing arbitrary queries and transforms on immutable data structures and for step-by-step navigation and modification of data structures. This library implements a simple zipper designed for manipulating JSON data. npm version Bower version Build Status Code Coverage Contents Tutorial Playing with zippers in a REPL can be very instructive. First we require the libraries import * as F from "fastener" import * as R from "ramda" and define a little helper using reduce to perform a sequence of operations on a value: const seq = (x, ...fs) => R.reduce((x, f) => f(x), x, fs) Let's work with the following simple JSON object: const data = {contents: [{language: "en", text: "Title"}, {language: "sv", text: "Rubrik"}]} First we just create a zipper using F.toZipper: seq(F.toZipper(data)) // { focus: { contents: [ [Object], [Object] ] } } As can be seen, the zipper is just a simple JSON object and the focus is the data object that we gave to F.toZipper. As long the data structure being manipulated is JSON, you can serialize and deserialize zippers as JSON. However, it is recommended that you use the zipper combinators to operate on zippers rather than rely on their exact format. Let's then move into the contents property of the object using F.downTo: seq(F.toZipper(data), F.downTo('contents')) // { left: null, // focus: // [ { language: 'en', text: 'Title' }, // { language: 'sv', text: 'Rubrik' } ], // key: 'contents', // right: null } As seen above, the focus now has the contents array. We can use F.get to extract the value under focus: seq(F.toZipper(data), F.downTo('contents'), F.get) // [ { language: 'en', text: 'Title' }, // { language: 'sv', text: 'Rubrik' } ] Then we move into the first element of contents using F.downHead: seq(F.toZipper(data), F.downTo('contents'), F.downHead) // { left: null, // focus: { language: 'en', text: 'Title' }, // key: 0, // right: [ null, { language: 'sv', text: 'Rubrik' } ], // up: { left: null, key: 'contents', right: null } } And continue into the first property of that which happens to be the language: seq(F.toZipper(data), F.downTo('contents'), F.downHead, F.downHead) // { left: null, // focus: 'en', // key: 'language', // right: [ null, 'Title', 'text' ], // up: // { left: null, // key: 0, // right: [ null, [Object] ], // up: { left: null, key: 'contents', right: null } } } And to the next property, title, using F.right: seq(F.toZipper(data), F.downTo('contents'), F.downHead, F.downHead, F.right) // { left: [ null, 'en', 'language' ], // focus: 'Title', // key: 'text', // right: null, // up: // { left: null, // key: 0, // right: [ null, [Object] ], // up: { left: null, key: 'contents', right: null } } } Let's then use F.modify to modify the title: seq(F.toZipper(data), F.downTo('contents'), F.downHead, F.downHead, F.right, F.modify(t => "The " + t)) // { left: [ null, 'en', 'language' ], // focus: 'The Title', // key: 'text', // right: null, // up: // { left: null, // key: 0, // right: [ null, [Object] ], // up: { left: null, key: 'contents', right: null } } } When we now move outwards using F.up we can see the changed title become part of the data: seq(F.toZipper(data), F.downTo('contents'), F.downHead, F.downHead, F.right, F.modify(t => "The " + t), F.up) // { left: null, // key: 0, // right: [ null, { language: 'sv', text: 'Rubrik' } ], // up: { left: null, key: 'contents', right: null }, // focus: { language: 'en', text: 'The Title' } } We can also just move back to the root and get the updated data structure using F.fromZipper: seq(F.toZipper(data), F.downTo('contents'), F.downHead, F.downHead, F.right, F.modify(t => "The " + t), F.fromZipper) // { contents: // [ { language: 'en', text: 'The Title' }, // { language: 'sv', text: 'Rubrik' } ] } The above hopefully helped to understand how zippers work. However, it is important to realize that one typically does not use zipper combinators to create such a specific sequence of operations. One rather uses the zipper combinators to create new combinators that perform more complex operations directly. Let's first define a zipper combinator that, given a zipper focused on an array, tries to focus on an element inside the array that satisfies a given predicate: const find = R.curry((p, z) => F.downTo(R.findIndex(p, F.get(z)), z)) Like all the basic zipper movement combinators, F.downTo is a partial function that returns undefined in case the index is out of bounds. Let's define a simple function to compose partial functions: const pipePartial = (...fs) => z => { for (let i=0; z !== undefined && i<fs.length; ++i) z = fs[i](z) return z } We can now compose a zipper combinator that, given a zipper focused on an object like data, tries to focus on the text element of an object with the given language inside the contents: const textIn = language => pipePartial( F.downTo('contents'), find(R.whereEq({language})), F.downTo('text')) Now we can say: seq(data, F.toZipper, textIn("en"), F.modify(x => 'The ' + x), F.fromZipper) // { contents: // [ { language: 'en', text: 'The Title' }, // { language: 'sv', text: 'Rubrik' } ] } Of course, this just scratches the surface. Zippers are powerful enough to implement arbitrary transforms on data structures. This can also make them more difficult to compose and reason about than more limited approaches such as lenses. Reference The zipper combinators are available as named imports. Typically one just imports the library as: import * as F from "fastener" In the following examples we will make use of the function const seq = (x, ...fs) => R.reduce((x, f) => f(x), x, fs) written using reduce that allows one to express a sequence of operations to perform starting from a given value. Introduction and Elimination F.toZipper(json) ~> zipper F.toZipper(json) creates a new zipper that is focused on the root of the given JSON object. For example: seq(F.toZipper([1,2,3]), F.downHead, F.modify(x => x + 1), F.fromZipper) // [ 2, 2, 3 ] F.fromZipper(zipper) ~> json F.fromZipper(zipper) extracts the modified JSON object from the given zipper. For example: seq(F.toZipper([1,2,3]), F.downHead, F.modify(x => x + 1), F.fromZipper) // [ 2, 2, 3 ] Focus Focus combinators allow one to inspect and modify the element that a zipper is focused on. F.get(zipper) ~> json F.get(zipper) returns the element that the zipper is focused on. For example: seq(F.toZipper(1), F.get) // 1 seq(F.toZipper(["a","b","c"]), F.downTo(2), F.get) // 'c' F.modify(json => json, zipper) ~> zipper F.modify(fn, zipper) is equivalent to F.set(fn(F.get(zipper)), zipper) and replaces the element that the zipper is focused on with the value returned by the given function for the element. For example: seq(F.toZipper(["a","b","c"]), F.downTo(2), F.modify(x => x + x), F.fromZipper) // [ 'a', 'b', 'cc' ] F.set(json, zipper) ~> zipper F.set(json, zipper) replaces the element that the zipper is focused on with the given value. For example: seq(F.toZipper(["a","b","c"]), F.downTo(1), F.set('lol'), F.fromZipper) // [ 'a', 'lol', 'c' ] Movement Movement combinators can be applied to any zipper, but they return undefined in case of illegal moves. Parent-Child movement Parent-Child movement is moving the focus between a parent object or array and a child element of said parent. F.downHead(zipper) ~> maybeZipper F.downHead(zipper) moves the focus to the leftmost element of the object or array that the zipper is focused on. F.downLast(zipper) ~> maybeZipper F.downLast(zipper) moves the focus to the rightmost element of the object or array that the zipper is focused on. F.downTo(key, zipper) ~> maybeZipper F.downTo(key, zipper) moves the focus to the specified object property or array index of the object or array that the zipper is focused on. F.keyOf(zipper) ~> maybeKey F.keyOf(zipper) returns the object property name or the array index that the zipper is currently focused on. F.up(zipper) ~> maybeZipper F.up(zipper) moves the focus from an array element or object property to the containing array or object. Path movement Path movement is moving the focus along a path from a parent object or array to a nested child element. F.downPath([...keys], zipper) ~> maybeZipper F.downPath(path, zipper) moves the focus along the specified path of keys. F.pathOf(zipper) ~> [...keys] F.pathOf(zipper) returns the path from the root to the current element focused on by the zipper. Sibling movement Sibling movement is moving the focus between the elements of an array or an object. F.head(zipper) ~> maybeZipper F.head(zipper) moves the focus to the leftmost sibling of the current focus. F.last(zipper) ~> maybeZipper F.last(zipper) moves the focus to the rightmost sibling of the current focus. F.left(zipper) ~> maybeZipper F.left(zipper) moves the focus to the element on the left of the current focus. F.right(zipper) ~> maybeZipper F.right(zipper) moves the focus to the element on the right of the current focus. Queries F.queryMove(zipper => maybeZipper, value, zipper => value, zipper) ~> value F.queryMove(move, default, fn, zipper) applies the given function fn to the zipper focused on after the given movement and returns the result unless the move was illegal in which case the given default value is returned instead. For example: seq(F.toZipper({x: 1}), F.queryMove(F.downTo('y'), false, () => true)) // false seq(F.toZipper({y: 1}), F.queryMove(F.downTo('y'), false, () => true)) // true Transforms F.transformMove(move, zipper => zipper, zipper) ~> zipper F.transformMove(move, fn, zipper) applies the given function to the zipper focused on after the given movement. The movement move must be one of F.downHead, F.downLast, F.downTo(key), F.left, F.right, or F.up. The function fn must the return a zipper focused on the same element that it was given. Then the focus is moved back to the element that the zipper was originally focused on. Nothing is done in case of an illegal move. For example: seq(F.toZipper({y: 1}), F.transformMove(F.downTo('y'), F.modify(x => x + 1)), F.fromZipper) // { y: 2 } seq(F.toZipper({x: 1}), F.transformMove(F.downTo('y'), F.modify(x => x + 1)), F.fromZipper) // { x: 1 } F.everywhere(json => json, zipper) ~> zipper F.everywhere(fn, zipper) performs a transform of the focused element by modifying each possible focus of the element with a bottom-up traversal. For example: seq(F.toZipper({foo: 1, bar: [{lol: "bal", example: 2}]}), F.everywhere(x => typeof x === "number" ? x + 1 : x), F.fromZipper) // { foo: 2, bar: [ { lol: 'bal', example: 3 } ] } Related Work While the implementation is very different, the choice of combinators is based on Michael D. Adams' paper Scrap Your Zippers.
__label__pos
0.956518
summaryrefslogtreecommitdiff path: root/xmloff/source/text/txtparae.cxx blob: 9410c49ca4229e4171618505a1ec48c594dab2b5 (plain) 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505 1506 1507 1508 1509 1510 1511 1512 1513 1514 1515 1516 1517 1518 1519 1520 1521 1522 1523 1524 1525 1526 1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625 1626 1627 1628 1629 1630 1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671 1672 1673 1674 1675 1676 1677 1678 1679 1680 1681 1682 1683 1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728 1729 1730 1731 1732 1733 1734 1735 1736 1737 1738 1739 1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772 1773 1774 1775 1776 1777 1778 1779 1780 1781 1782 1783 1784 1785 1786 1787 1788 1789 1790 1791 1792 1793 1794 1795 1796 1797 1798 1799 1800 1801 1802 1803 1804 1805 1806 1807 1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819 1820 1821 1822 1823 1824 1825 1826 1827 1828 1829 1830 1831 1832 1833 1834 1835 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850 1851 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883 1884 1885 1886 1887 1888 1889 1890 1891 1892 1893 1894 1895 1896 1897 1898 1899 1900 1901 1902 1903 1904 1905 1906 1907 1908 1909 1910 1911 1912 1913 1914 1915 1916 1917 1918 1919 1920 1921 1922 1923 1924 1925 1926 1927 1928 1929 1930 1931 1932 1933 1934 1935 1936 1937 1938 1939 1940 1941 1942 1943 1944 1945 1946 1947 1948 1949 1950 1951 1952 1953 1954 1955 1956 1957 1958 1959 1960 1961 1962 1963 1964 1965 1966 1967 1968 1969 1970 1971 1972 1973 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 2017 2018 2019 2020 2021 2022 2023 2024 2025 2026 2027 2028 2029 2030 2031 2032 2033 2034 2035 2036 2037 2038 2039 2040 2041 2042 2043 2044 2045 2046 2047 2048 2049 2050 2051 2052 2053 2054 2055 2056 2057 2058 2059 2060 2061 2062 2063 2064 2065 2066 2067 2068 2069 2070 2071 2072 2073 2074 2075 2076 2077 2078 2079 2080 2081 2082 2083 2084 2085 2086 2087 2088 2089 2090 2091 2092 2093 2094 2095 2096 2097 2098 2099 2100 2101 2102 2103 2104 2105 2106 2107 2108 2109 2110 2111 2112 2113 2114 2115 2116 2117 2118 2119 2120 2121 2122 2123 2124 2125 2126 2127 2128 2129 2130 2131 2132 2133 2134 2135 2136 2137 2138 2139 2140 2141 2142 2143 2144 2145 2146 2147 2148 2149 2150 2151 2152 2153 2154 2155 2156 2157 2158 2159 2160 2161 2162 2163 2164 2165 2166 2167 2168 2169 2170 2171 2172 2173 2174 2175 2176 2177 2178 2179 2180 2181 2182 2183 2184 2185 2186 2187 2188 2189 2190 2191 2192 2193 2194 2195 2196 2197 2198 2199 2200 2201 2202 2203 2204 2205 2206 2207 2208 2209 2210 2211 2212 2213 2214 2215 2216 2217 2218 2219 2220 2221 2222 2223 2224 2225 2226 2227 2228 2229 2230 2231 2232 2233 2234 2235 2236 2237 2238 2239 2240 2241 2242 2243 2244 2245 2246 2247 2248 2249 2250 2251 2252 2253 2254 2255 2256 2257 2258 2259 2260 2261 2262 2263 2264 2265 2266 2267 2268 2269 2270 2271 2272 2273 2274 2275 2276 2277 2278 2279 2280 2281 2282 2283 2284 2285 2286 2287 2288 2289 2290 2291 2292 2293 2294 2295 2296 2297 2298 2299 2300 2301 2302 2303 2304 2305 2306 2307 2308 2309 2310 2311 2312 2313 2314 2315 2316 2317 2318 2319 2320 2321 2322 2323 2324 2325 2326 2327 2328 2329 2330 2331 2332 2333 2334 2335 2336 2337 2338 2339 2340 2341 2342 2343 2344 2345 2346 2347 2348 2349 2350 2351 2352 2353 2354 2355 2356 2357 2358 2359 2360 2361 2362 2363 2364 2365 2366 2367 2368 2369 2370 2371 2372 2373 2374 2375 2376 2377 2378 2379 2380 2381 2382 2383 2384 2385 2386 2387 2388 2389 2390 2391 2392 2393 2394 2395 2396 2397 2398 2399 2400 2401 2402 2403 2404 2405 2406 2407 2408 2409 2410 2411 2412 2413 2414 2415 2416 2417 2418 2419 2420 2421 2422 2423 2424 2425 2426 2427 2428 2429 2430 2431 2432 2433 2434 2435 2436 2437 2438 2439 2440 2441 2442 2443 2444 2445 2446 2447 2448 2449 2450 2451 2452 2453 2454 2455 2456 2457 2458 2459 2460 2461 2462 2463 2464 2465 2466 2467 2468 2469 2470 2471 2472 2473 2474 2475 2476 2477 2478 2479 2480 2481 2482 2483 2484 2485 2486 2487 2488 2489 2490 2491 2492 2493 2494 2495 2496 2497 2498 2499 2500 2501 2502 2503 2504 2505 2506 2507 2508 2509 2510 2511 2512 2513 2514 2515 2516 2517 2518 2519 2520 2521 2522 2523 2524 2525 2526 2527 2528 2529 2530 2531 2532 2533 2534 2535 2536 2537 2538 2539 2540 2541 2542 2543 2544 2545 2546 2547 2548 2549 2550 2551 2552 2553 2554 2555 2556 2557 2558 2559 2560 2561 2562 2563 2564 2565 2566 2567 2568 2569 2570 2571 2572 2573 2574 2575 2576 2577 2578 2579 2580 2581 2582 2583 2584 2585 2586 2587 2588 2589 2590 2591 2592 2593 2594 2595 2596 2597 2598 2599 2600 2601 2602 2603 2604 2605 2606 2607 2608 2609 2610 2611 2612 2613 2614 2615 2616 2617 2618 2619 2620 2621 2622 2623 2624 2625 2626 2627 2628 2629 2630 2631 2632 2633 2634 2635 2636 2637 2638 2639 2640 2641 2642 2643 2644 2645 2646 2647 2648 2649 2650 2651 2652 2653 2654 2655 2656 2657 2658 2659 2660 2661 2662 2663 2664 2665 2666 2667 2668 2669 2670 2671 2672 2673 2674 2675 2676 2677 2678 2679 2680 2681 2682 2683 2684 2685 2686 2687 2688 2689 2690 2691 2692 2693 2694 2695 2696 2697 2698 2699 2700 2701 2702 2703 2704 2705 2706 2707 2708 2709 2710 2711 2712 2713 2714 2715 2716 2717 2718 2719 2720 2721 2722 2723 2724 2725 2726 2727 2728 2729 2730 2731 2732 2733 2734 2735 2736 2737 2738 2739 2740 2741 2742 2743 2744 2745 2746 2747 2748 2749 2750 2751 2752 2753 2754 2755 2756 2757 2758 2759 2760 2761 2762 2763 2764 2765 2766 2767 2768 2769 2770 2771 2772 2773 2774 2775 2776 2777 2778 2779 2780 2781 2782 2783 2784 2785 2786 2787 2788 2789 2790 2791 2792 2793 2794 2795 2796 2797 2798 2799 2800 2801 2802 2803 2804 2805 2806 2807 2808 2809 2810 2811 2812 2813 2814 2815 2816 2817 2818 2819 2820 2821 2822 2823 2824 2825 2826 2827 2828 2829 2830 2831 2832 2833 2834 2835 2836 2837 2838 2839 2840 2841 2842 2843 2844 2845 2846 2847 2848 2849 2850 2851 2852 2853 2854 2855 2856 2857 2858 2859 2860 2861 2862 2863 2864 2865 2866 2867 2868 2869 2870 2871 2872 2873 2874 2875 2876 2877 2878 2879 2880 2881 2882 2883 2884 2885 2886 2887 2888 2889 2890 2891 2892 2893 2894 2895 2896 2897 2898 2899 2900 2901 2902 2903 2904 2905 2906 2907 2908 2909 2910 2911 2912 2913 2914 2915 2916 2917 2918 2919 2920 2921 2922 2923 2924 2925 2926 2927 2928 2929 2930 2931 2932 2933 2934 2935 2936 2937 2938 2939 2940 2941 2942 2943 2944 2945 2946 2947 2948 2949 2950 2951 2952 2953 2954 2955 2956 2957 2958 2959 2960 2961 2962 2963 2964 2965 2966 2967 2968 2969 2970 2971 2972 2973 2974 2975 2976 2977 2978 2979 2980 2981 2982 2983 2984 2985 2986 2987 2988 2989 2990 2991 2992 2993 2994 2995 2996 2997 2998 2999 3000 3001 3002 3003 3004 3005 3006 3007 3008 3009 3010 3011 3012 3013 3014 3015 3016 3017 3018 3019 3020 3021 3022 3023 3024 3025 3026 3027 3028 3029 3030 3031 3032 3033 3034 3035 3036 3037 3038 3039 3040 3041 3042 3043 3044 3045 3046 3047 3048 3049 3050 3051 3052 3053 3054 3055 3056 3057 3058 3059 3060 3061 3062 3063 3064 3065 3066 3067 3068 3069 3070 3071 3072 3073 3074 3075 3076 3077 3078 3079 3080 3081 3082 3083 3084 3085 3086 3087 3088 3089 3090 3091 3092 3093 3094 3095 3096 3097 3098 3099 3100 3101 3102 3103 3104 3105 3106 3107 3108 3109 3110 3111 3112 3113 3114 3115 3116 3117 3118 3119 3120 3121 3122 3123 3124 3125 3126 3127 3128 3129 3130 3131 3132 3133 3134 3135 3136 3137 3138 3139 3140 3141 3142 3143 3144 3145 3146 3147 3148 3149 3150 3151 3152 3153 3154 3155 3156 3157 3158 3159 3160 3161 3162 3163 3164 3165 3166 3167 3168 3169 3170 3171 3172 3173 3174 3175 3176 3177 3178 3179 3180 3181 3182 3183 3184 3185 3186 3187 3188 3189 3190 3191 3192 3193 3194 3195 3196 3197 3198 3199 3200 3201 3202 3203 3204 3205 3206 3207 3208 3209 3210 3211 3212 3213 3214 3215 3216 3217 3218 3219 3220 3221 3222 3223 3224 3225 3226 3227 3228 3229 3230 3231 3232 3233 3234 3235 3236 3237 3238 3239 3240 3241 3242 3243 3244 3245 3246 3247 3248 3249 3250 3251 3252 3253 3254 3255 3256 3257 3258 3259 3260 3261 3262 3263 3264 3265 3266 3267 3268 3269 3270 3271 3272 3273 3274 3275 3276 3277 3278 3279 3280 3281 3282 3283 3284 3285 3286 3287 3288 3289 3290 3291 3292 3293 3294 3295 3296 3297 3298 3299 3300 3301 3302 3303 3304 3305 3306 3307 3308 3309 3310 3311 3312 3313 3314 3315 3316 3317 3318 3319 3320 3321 3322 3323 3324 3325 3326 3327 3328 3329 3330 3331 3332 3333 3334 3335 3336 3337 3338 3339 3340 3341 3342 3343 3344 3345 3346 3347 3348 3349 3350 3351 3352 3353 3354 3355 3356 3357 3358 3359 3360 3361 3362 3363 3364 3365 3366 3367 3368 3369 3370 3371 3372 3373 3374 3375 3376 3377 3378 3379 3380 3381 3382 3383 3384 3385 3386 3387 3388 3389 3390 3391 3392 3393 3394 3395 3396 3397 3398 3399 3400 3401 3402 3403 3404 3405 3406 3407 3408 3409 3410 3411 3412 3413 3414 3415 3416 3417 3418 3419 3420 3421 3422 3423 3424 3425 3426 3427 3428 3429 3430 3431 3432 3433 3434 3435 3436 3437 3438 3439 3440 3441 3442 3443 3444 3445 3446 3447 3448 3449 3450 3451 3452 3453 3454 3455 3456 3457 3458 3459 3460 3461 3462 3463 3464 3465 3466 3467 3468 3469 3470 3471 3472 3473 3474 3475 3476 3477 3478 3479 3480 3481 3482 3483 3484 3485 3486 3487 3488 3489 3490 3491 3492 3493 3494 3495 3496 3497 3498 3499 3500 3501 3502 3503 3504 3505 3506 3507 3508 3509 3510 3511 3512 3513 3514 3515 3516 3517 3518 3519 3520 3521 3522 3523 3524 3525 3526 3527 3528 3529 3530 3531 3532 3533 3534 3535 3536 3537 3538 3539 3540 3541 3542 3543 3544 3545 3546 3547 3548 3549 3550 3551 3552 3553 3554 3555 3556 3557 3558 3559 3560 3561 3562 3563 3564 3565 3566 3567 3568 3569 3570 3571 3572 3573 3574 3575 3576 3577 3578 3579 3580 3581 3582 3583 3584 3585 3586 3587 3588 3589 3590 3591 3592 3593 3594 3595 3596 3597 3598 3599 3600 3601 3602 3603 3604 3605 3606 3607 3608 3609 3610 3611 3612 3613 3614 3615 3616 3617 3618 3619 3620 3621 3622 3623 3624 3625 3626 3627 3628 3629 3630 3631 3632 3633 3634 3635 3636 3637 3638 3639 3640 3641 3642 3643 3644 3645 3646 3647 3648 3649 3650 3651 3652 3653 3654 3655 3656 3657 3658 3659 3660 3661 3662 3663 3664 3665 3666 3667 3668 3669 3670 3671 3672 3673 3674 3675 3676 3677 3678 3679 3680 3681 3682 3683 3684 3685 3686 3687 3688 3689 3690 3691 3692 3693 3694 3695 3696 3697 3698 3699 3700 3701 3702 3703 3704 3705 3706 3707 3708 3709 3710 3711 3712 3713 3714 3715 3716 3717 3718 3719 3720 3721 3722 3723 3724 3725 3726 3727 3728 3729 3730 3731 3732 3733 3734 3735 3736 3737 3738 3739 3740 3741 3742 3743 3744 3745 3746 3747 3748 3749 3750 3751 3752 3753 3754 3755 3756 3757 3758 3759 3760 3761 3762 3763 3764 3765 3766 3767 3768 3769 3770 3771 3772 3773 3774 3775 3776 3777 3778 3779 3780 3781 3782 3783 3784 3785 3786 3787 3788 3789 3790 3791 3792 3793 3794 3795 3796 3797 3798 3799 3800 3801 3802 3803 3804 3805 3806 3807 3808 3809 3810 3811 3812 3813 3814 3815 3816 3817 3818 3819 3820 3821 3822 3823 3824 3825 3826 3827 3828 3829 3830 3831 3832 3833 3834 3835 3836 3837 3838 3839 3840 3841 3842 3843 3844 3845 3846 3847 3848 /* -*- Mode: C++; tab-width: 4; indent-tabs-mode: nil; c-basic-offset: 4 -*- */ /************************************************************************* * * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. * * Copyright 2000, 2010 Oracle and/or its affiliates. * * OpenOffice.org - a multi-platform office productivity suite * * This file is part of OpenOffice.org. * * OpenOffice.org is free software: you can redistribute it and/or modify * it under the terms of the GNU Lesser General Public License version 3 * only, as published by the Free Software Foundation. * * OpenOffice.org is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU Lesser General Public License version 3 for more details * (a copy is included in the LICENSE file that accompanied this code). * * You should have received a copy of the GNU Lesser General Public License * version 3 along with OpenOffice.org. If not, see * <http://www.openoffice.org/license.html> * for a copy of the LGPLv3 License. * ************************************************************************/ // MARKER(update_precomp.py): autogen include statement, do not remove #include "precompiled_xmloff.hxx" #include "unointerfacetouniqueidentifiermapper.hxx" #include <tools/debug.hxx> #ifndef _SVSTDARR_LONGS_DECL #define _SVSTDARR_LONGS #include <svl/svstdarr.hxx> #endif #include <svl/svarray.hxx> #include <rtl/ustrbuf.hxx> #include <sal/types.h> #include <vector> #include <list> #include <hash_map> #include <com/sun/star/lang/XServiceInfo.hpp> #include <com/sun/star/container/XEnumerationAccess.hpp> #include <com/sun/star/container/XEnumeration.hpp> #include <com/sun/star/container/XIndexReplace.hpp> #include <com/sun/star/beans/XPropertySet.hpp> #include <com/sun/star/beans/XMultiPropertySet.hpp> #include <com/sun/star/beans/XPropertyState.hpp> #include <com/sun/star/text/XTextDocument.hpp> #include <com/sun/star/text/XTextSectionsSupplier.hpp> #include <com/sun/star/text/XTextTablesSupplier.hpp> #include <com/sun/star/text/XNumberingRulesSupplier.hpp> #include <com/sun/star/text/XChapterNumberingSupplier.hpp> #include <com/sun/star/text/XTextTable.hpp> #include <com/sun/star/text/XText.hpp> #include <com/sun/star/text/XTextContent.hpp> #include <com/sun/star/text/XTextRange.hpp> #include <com/sun/star/text/XTextField.hpp> #include <com/sun/star/text/XFootnote.hpp> #include <com/sun/star/container/XNamed.hpp> #include <com/sun/star/container/XContentEnumerationAccess.hpp> #include <com/sun/star/text/XTextFrame.hpp> #include <com/sun/star/container/XNameAccess.hpp> #include <com/sun/star/text/SizeType.hpp> #include <com/sun/star/text/HoriOrientation.hpp> #include <com/sun/star/text/VertOrientation.hpp> #include <com/sun/star/text/TextContentAnchorType.hpp> #include <com/sun/star/text/XTextFramesSupplier.hpp> #include <com/sun/star/text/XTextGraphicObjectsSupplier.hpp> #include <com/sun/star/text/XTextEmbeddedObjectsSupplier.hpp> #include <com/sun/star/drawing/XDrawPageSupplier.hpp> #include <com/sun/star/document/XEmbeddedObjectSupplier.hpp> #include <com/sun/star/document/XEventsSupplier.hpp> #include <com/sun/star/document/XRedlinesSupplier.hpp> #include <com/sun/star/text/XBookmarksSupplier.hpp> #include <com/sun/star/text/XFormField.hpp> #include <com/sun/star/text/XTextSection.hpp> #include <com/sun/star/text/SectionFileLink.hpp> #include <com/sun/star/drawing/XShape.hpp> #include <com/sun/star/text/XTextShapesSupplier.hpp> #include <com/sun/star/style/XAutoStylesSupplier.hpp> #include <com/sun/star/style/XAutoStyleFamily.hpp> #include <com/sun/star/text/XTextFieldsSupplier.hpp> #include <com/sun/star/text/XFootnotesSupplier.hpp> #include <com/sun/star/text/XEndnotesSupplier.hpp> #include <com/sun/star/drawing/XControlShape.hpp> #include <com/sun/star/util/DateTime.hpp> #include "xmlnmspe.hxx" #include <xmloff/xmlaustp.hxx> #include <xmloff/families.hxx> #include "txtexppr.hxx" #include <xmloff/xmlnumfe.hxx> #include <xmloff/xmlnume.hxx> #include <xmloff/xmluconv.hxx> #include "XMLAnchorTypePropHdl.hxx" #include "xexptran.hxx" #include <xmloff/ProgressBarHelper.hxx> #include <xmloff/nmspmap.hxx> #include <xmloff/xmlexp.hxx> #include "txtflde.hxx" #include <xmloff/txtprmap.hxx> #include "XMLImageMapExport.hxx" #include "XMLTextNumRuleInfo.hxx" #include "XMLTextListAutoStylePool.hxx" #include <xmloff/txtparae.hxx> #include "XMLSectionExport.hxx" #include "XMLIndexMarkExport.hxx" #include <xmloff/XMLEventExport.hxx> #include "XMLRedlineExport.hxx" #include "MultiPropertySetHelper.hxx" #include <xmloff/formlayerexport.hxx> #include "XMLTextCharStyleNamesElementExport.hxx" #include <comphelper/stlunosequence.hxx> #include <xmloff/odffields.hxx> #include <com/sun/star/embed/ElementModes.hpp> #include <com/sun/star/embed/XTransactedObject.hpp> #include <com/sun/star/document/XStorageBasedDocument.hpp> #include <txtlists.hxx> #include <com/sun/star/rdf/XMetadatable.hpp> using ::rtl::OUString; using ::rtl::OUStringBuffer; using namespace ::std; using namespace ::com::sun::star; using namespace ::com::sun::star::uno; using namespace ::com::sun::star::lang; using namespace ::com::sun::star::beans; using namespace ::com::sun::star::container; using namespace ::com::sun::star::text; using namespace ::com::sun::star::style; using namespace ::com::sun::star::util; using namespace ::com::sun::star::drawing; using namespace ::com::sun::star::document; using namespace ::com::sun::star::frame; using namespace ::xmloff; using namespace ::xmloff::token; namespace { class TextContentSet { public: typedef Reference<XTextContent> text_content_ref_t; typedef list<text_content_ref_t> contents_t; typedef back_insert_iterator<contents_t> inserter_t; typedef contents_t::const_iterator const_iterator_t; inserter_t getInserter() { return back_insert_iterator<contents_t>(m_vTextContents); }; const_iterator_t getBegin() const { return m_vTextContents.begin(); }; const_iterator_t getEnd() const { return m_vTextContents.end(); }; private: contents_t m_vTextContents; }; struct FrameRefHash : public unary_function<Reference<XTextFrame>, size_t> { size_t operator()(const Reference<XTextFrame> xFrame) const { return sal::static_int_cast<size_t>(reinterpret_cast<sal_uIntPtr>(xFrame.get())); } }; static bool lcl_TextContentsUnfiltered(const Reference<XTextContent>&) { return true; }; static bool lcl_ShapeFilter(const Reference<XTextContent>& xTxtContent) { static const OUString sTextFrameService(RTL_CONSTASCII_USTRINGPARAM("com.sun.star.text.TextFrame")); static const OUString sTextGraphicService(RTL_CONSTASCII_USTRINGPARAM("com.sun.star.text.TextGraphicObject")); static const OUString sTextEmbeddedService(RTL_CONSTASCII_USTRINGPARAM("com.sun.star.text.TextEmbeddedObject")); Reference<XShape> xShape(xTxtContent, UNO_QUERY); if(!xShape.is()) return false; Reference<XServiceInfo> xServiceInfo(xTxtContent, UNO_QUERY); if(xServiceInfo->supportsService(sTextFrameService) || xServiceInfo->supportsService(sTextGraphicService) || xServiceInfo->supportsService(sTextEmbeddedService) ) return false; return true; }; class BoundFrames { public: typedef bool (*filter_t)(const Reference<XTextContent>&); BoundFrames( const Reference<XEnumerationAccess> xEnumAccess, const filter_t& rFilter) : m_xEnumAccess(xEnumAccess) { Fill(rFilter); }; BoundFrames() {}; const TextContentSet* GetPageBoundContents() const { return &m_vPageBounds; }; const TextContentSet* GetFrameBoundContents(const Reference<XTextFrame>& rParentFrame) const { framebound_map_t::const_iterator it = m_vFrameBoundsOf.find(rParentFrame); if(it == m_vFrameBoundsOf.end()) return NULL; return &(it->second); }; Reference<XEnumeration> createEnumeration() const { if(!m_xEnumAccess.is()) return Reference<XEnumeration>(); return m_xEnumAccess->createEnumeration(); }; private: typedef hash_map< Reference<XTextFrame>, TextContentSet, FrameRefHash> framebound_map_t; TextContentSet m_vPageBounds; framebound_map_t m_vFrameBoundsOf; const Reference<XEnumerationAccess> m_xEnumAccess; void Fill(const filter_t& rFilter); static const OUString our_sAnchorType; static const OUString our_sAnchorFrame; }; const OUString BoundFrames::our_sAnchorType(RTL_CONSTASCII_USTRINGPARAM("AnchorType")); const OUString BoundFrames::our_sAnchorFrame(RTL_CONSTASCII_USTRINGPARAM("AnchorFrame")); class FieldParamExporter { public: FieldParamExporter(SvXMLExport* const pExport, Reference<XNameContainer> xFieldParams) : m_pExport(pExport) , m_xFieldParams(xFieldParams) { }; void Export(); private: SvXMLExport* const m_pExport; const Reference<XNameContainer> m_xFieldParams; void ExportParameter(const OUString& sKey, const OUString& sValue); }; } namespace xmloff { class BoundFrameSets { public: BoundFrameSets(const Reference<XInterface> xModel); const BoundFrames* GetTexts() const { return m_pTexts.get(); }; const BoundFrames* GetGraphics() const { return m_pGraphics.get(); }; const BoundFrames* GetEmbeddeds() const { return m_pEmbeddeds.get(); }; const BoundFrames* GetShapes() const { return m_pShapes.get(); }; private: auto_ptr<BoundFrames> m_pTexts; auto_ptr<BoundFrames> m_pGraphics; auto_ptr<BoundFrames> m_pEmbeddeds; auto_ptr<BoundFrames> m_pShapes; }; } typedef OUString *OUStringPtr; SV_DECL_PTRARR_DEL( OUStrings_Impl, OUStringPtr, 20, 10 ) SV_IMPL_PTRARR( OUStrings_Impl, OUStringPtr ) SV_DECL_PTRARR_SORT_DEL( OUStringsSort_Impl, OUStringPtr, 20, 10 ) SV_IMPL_OP_PTRARR_SORT( OUStringsSort_Impl, OUStringPtr ) #ifdef DBG_UTIL static int txtparae_bContainsIllegalCharacters = sal_False; #endif // The following map shows which property values are required: // // property auto style pass export // -------------------------------------------------------- // ParaStyleName if style exists always // ParaConditionalStyleName if style exists always // NumberingRules if style exists always // TextSection always always // ParaChapterNumberingLevel never always // NumberingIsNumber never always // The conclusion is that for auto styles the first three properties // should be queried using a multi property set if, and only if, an // auto style needs to be exported. TextSection should be queried by // an individual call to getPropertyvalue, because this seems to be // less expensive than querying the first three properties if they aren't // required. // For the export pass all properties can be queried using a multi property // set. static const sal_Char* aParagraphPropertyNamesAuto[] = { "NumberingRules", "ParaConditionalStyleName", "ParaStyleName", NULL }; enum eParagraphPropertyNamesEnumAuto { NUMBERING_RULES_AUTO = 0, PARA_CONDITIONAL_STYLE_NAME_AUTO = 1, PARA_STYLE_NAME_AUTO = 2 }; static const sal_Char* aParagraphPropertyNames[] = { "NumberingIsNumber", "NumberingStyleName", "OutlineLevel", "ParaConditionalStyleName", "ParaStyleName", "TextSection", NULL }; enum eParagraphPropertyNamesEnum { NUMBERING_IS_NUMBER = 0, PARA_NUMBERING_STYLENAME = 1, PARA_OUTLINE_LEVEL=2, PARA_CONDITIONAL_STYLE_NAME = 3, PARA_STYLE_NAME = 4, TEXT_SECTION = 5 }; void BoundFrames::Fill(const filter_t& rFilter) { if(!m_xEnumAccess.is()) return; const Reference< XEnumeration > xEnum = m_xEnumAccess->createEnumeration(); if(!xEnum.is()) return; while(xEnum->hasMoreElements()) { Reference<XPropertySet> xPropSet(xEnum->nextElement(), UNO_QUERY); Reference<XTextContent> xTextContent(xPropSet, UNO_QUERY); if(!xPropSet.is() || !xTextContent.is()) continue; TextContentAnchorType eAnchor; xPropSet->getPropertyValue(our_sAnchorType) >>= eAnchor; if(TextContentAnchorType_AT_PAGE != eAnchor && TextContentAnchorType_AT_FRAME != eAnchor) continue; if(!rFilter(xTextContent)) continue; TextContentSet::inserter_t pInserter = m_vPageBounds.getInserter(); if(TextContentAnchorType_AT_FRAME == eAnchor) { Reference<XTextFrame> xAnchorTxtFrame( xPropSet->getPropertyValue(our_sAnchorFrame), uno::UNO_QUERY); pInserter = m_vFrameBoundsOf[xAnchorTxtFrame].getInserter(); } *pInserter++ = xTextContent; } } BoundFrameSets::BoundFrameSets(const Reference<XInterface> xModel) : m_pTexts(new BoundFrames()) , m_pGraphics(new BoundFrames()) , m_pEmbeddeds(new BoundFrames()) , m_pShapes(new BoundFrames()) { const Reference<XTextFramesSupplier> xTFS(xModel, UNO_QUERY); const Reference<XTextGraphicObjectsSupplier> xGOS(xModel, UNO_QUERY); const Reference<XTextEmbeddedObjectsSupplier> xEOS(xModel, UNO_QUERY); const Reference<XDrawPageSupplier> xDPS(xModel, UNO_QUERY); if(xTFS.is()) m_pTexts = auto_ptr<BoundFrames>(new BoundFrames( Reference<XEnumerationAccess>(xTFS->getTextFrames(), UNO_QUERY), &lcl_TextContentsUnfiltered)); if(xGOS.is()) m_pGraphics = auto_ptr<BoundFrames>(new BoundFrames( Reference<XEnumerationAccess>(xGOS->getGraphicObjects(), UNO_QUERY), &lcl_TextContentsUnfiltered)); if(xEOS.is()) m_pEmbeddeds = auto_ptr<BoundFrames>(new BoundFrames( Reference<XEnumerationAccess>(xEOS->getEmbeddedObjects(), UNO_QUERY), &lcl_TextContentsUnfiltered)); if(xDPS.is()) m_pShapes = auto_ptr<BoundFrames>(new BoundFrames( Reference<XEnumerationAccess>(xDPS->getDrawPage(), UNO_QUERY), &lcl_ShapeFilter)); }; void FieldParamExporter::Export() { static const Type aStringType = ::getCppuType((OUString*)0); static const Type aBoolType = ::getCppuType((sal_Bool*)0); static const Type aSeqType = ::getCppuType((Sequence<OUString>*)0); static const Type aIntType = ::getCppuType((sal_Int32*)0); Sequence<OUString> vParameters(m_xFieldParams->getElementNames()); for(const OUString* pCurrent=::comphelper::stl_begin(vParameters); pCurrent!=::comphelper::stl_end(vParameters); ++pCurrent) { const Any aValue = m_xFieldParams->getByName(*pCurrent); const Type aValueType = aValue.getValueType(); if(aValueType == aStringType) { OUString sValue; aValue >>= sValue; ExportParameter(*pCurrent,sValue); if ( pCurrent->equalsAscii( ODF_OLE_PARAM ) ) { // Save the OLE object Reference< embed::XStorage > xTargetStg = m_pExport->GetTargetStorage(); Reference< embed::XStorage > xDstStg = xTargetStg->openStorageElement( rtl::OUString(RTL_CONSTASCII_USTRINGPARAM("OLELinks")), embed::ElementModes::WRITE ); if ( !xDstStg->hasByName( sValue ) ) { Reference< XStorageBasedDocument > xStgDoc ( m_pExport->GetModel( ), UNO_QUERY ); Reference< embed::XStorage > xDocStg = xStgDoc->getDocumentStorage(); Reference< embed::XStorage > xOleStg = xDocStg->openStorageElement( rtl::OUString(RTL_CONSTASCII_USTRINGPARAM("OLELinks")), embed::ElementModes::READ ); xOleStg->copyElementTo( sValue, xDstStg, sValue ); Reference< embed::XTransactedObject > xTransact( xDstStg, UNO_QUERY ); if ( xTransact.is( ) ) xTransact->commit( ); } } } else if(aValueType == aBoolType) { sal_Bool bValue = false; aValue >>= bValue; ExportParameter(*pCurrent,OUString::createFromAscii(bValue ? "true" : "false")); ExportParameter(*pCurrent, (bValue ? OUString(RTL_CONSTASCII_USTRINGPARAM( "true" )) : OUString(RTL_CONSTASCII_USTRINGPARAM("false"))) ); } else if(aValueType == aSeqType) { Sequence<OUString> vValue; aValue >>= vValue; for(OUString* pSeqCurrent = ::comphelper::stl_begin(vValue); pSeqCurrent != ::comphelper::stl_end(vValue); ++pSeqCurrent) { ExportParameter(*pCurrent, *pSeqCurrent); } } else if(aValueType == aIntType) { sal_Int32 nValue = 0; aValue >>= nValue; ExportParameter(*pCurrent, OUStringBuffer().append(nValue).makeStringAndClear()); } } } void FieldParamExporter::ExportParameter(const OUString& sKey, const OUString& sValue) { m_pExport->AddAttribute(XML_NAMESPACE_FIELD, XML_NAME, sKey); m_pExport->AddAttribute(XML_NAMESPACE_FIELD, XML_VALUE, sValue); m_pExport->StartElement(XML_NAMESPACE_FIELD, XML_PARAM, sal_False); m_pExport->EndElement(XML_NAMESPACE_FIELD, XML_PARAM, sal_False); } void XMLTextParagraphExport::Add( sal_uInt16 nFamily, const Reference < XPropertySet > & rPropSet, const XMLPropertyState** ppAddStates, bool bDontSeek ) { UniReference < SvXMLExportPropertyMapper > xPropMapper; switch( nFamily ) { case XML_STYLE_FAMILY_TEXT_PARAGRAPH: xPropMapper = GetParaPropMapper(); break; case XML_STYLE_FAMILY_TEXT_TEXT: xPropMapper = GetTextPropMapper(); break; case XML_STYLE_FAMILY_TEXT_FRAME: xPropMapper = GetAutoFramePropMapper(); break; case XML_STYLE_FAMILY_TEXT_SECTION: xPropMapper = GetSectionPropMapper(); break; case XML_STYLE_FAMILY_TEXT_RUBY: xPropMapper = GetRubyPropMapper(); break; } DBG_ASSERT( xPropMapper.is(), "There is the property mapper?" ); vector< XMLPropertyState > xPropStates = xPropMapper->Filter( rPropSet ); if( ppAddStates ) { while( *ppAddStates ) { xPropStates.push_back( **ppAddStates ); ppAddStates++; } } if( !xPropStates.empty() ) { Reference< XPropertySetInfo > xPropSetInfo(rPropSet->getPropertySetInfo()); OUString sParent, sCondParent; sal_uInt16 nIgnoreProps = 0; switch( nFamily ) { case XML_STYLE_FAMILY_TEXT_PARAGRAPH: if( xPropSetInfo->hasPropertyByName( sParaStyleName ) ) { rPropSet->getPropertyValue( sParaStyleName ) >>= sParent; } if( xPropSetInfo->hasPropertyByName( sParaConditionalStyleName ) ) { rPropSet->getPropertyValue( sParaConditionalStyleName ) >>= sCondParent; } if( xPropSetInfo->hasPropertyByName( sNumberingRules ) ) { Reference < XIndexReplace > xNumRule(rPropSet->getPropertyValue( sNumberingRules ), uno::UNO_QUERY); if( xNumRule.is() && xNumRule->getCount() ) { Reference < XNamed > xNamed( xNumRule, UNO_QUERY ); OUString sName; if( xNamed.is() ) sName = xNamed->getName(); sal_Bool bAdd = !sName.getLength(); if( !bAdd ) { Reference < XPropertySet > xNumPropSet( xNumRule, UNO_QUERY ); const OUString sIsAutomatic( RTL_CONSTASCII_USTRINGPARAM( "IsAutomatic" ) ); if( xNumPropSet.is() && xNumPropSet->getPropertySetInfo() ->hasPropertyByName( sIsAutomatic ) ) { bAdd = *(sal_Bool *)xNumPropSet->getPropertyValue( sIsAutomatic ).getValue(); // Check on outline style (#i73361#) const OUString sNumberingIsOutline( RTL_CONSTASCII_USTRINGPARAM( "NumberingIsOutline" ) ); if ( bAdd && xNumPropSet->getPropertySetInfo() ->hasPropertyByName( sNumberingIsOutline ) ) { bAdd = !(*(sal_Bool *)xNumPropSet->getPropertyValue( sNumberingIsOutline ).getValue()); } } else { bAdd = sal_True; } } if( bAdd ) pListAutoPool->Add( xNumRule ); } } break; case XML_STYLE_FAMILY_TEXT_TEXT: { // Get parent and remove hyperlinks (they aren't of interest) UniReference< XMLPropertySetMapper > xPM(xPropMapper->getPropertySetMapper()); for( ::std::vector< XMLPropertyState >::iterator i(xPropStates.begin()); nIgnoreProps < 2 && i != xPropStates.end(); ) { if( i->mnIndex == -1 ) { ++i; continue; } switch( xPM->GetEntryContextId(i->mnIndex) ) { case CTF_CHAR_STYLE_NAME: case CTF_HYPERLINK_URL: i->mnIndex = -1; nIgnoreProps++; i = xPropStates.erase( i ); break; default: ++i; break; } } } break; case XML_STYLE_FAMILY_TEXT_FRAME: if( xPropSetInfo->hasPropertyByName( sFrameStyleName ) ) { rPropSet->getPropertyValue( sFrameStyleName ) >>= sParent; } break; case XML_STYLE_FAMILY_TEXT_SECTION: case XML_STYLE_FAMILY_TEXT_RUBY: ; // section styles have no parents break; } if( (xPropStates.size() - nIgnoreProps) > 0 ) { GetAutoStylePool().Add( nFamily, sParent, xPropStates, bDontSeek ); if( sCondParent.getLength() && sParent != sCondParent ) GetAutoStylePool().Add( nFamily, sCondParent, xPropStates ); } } } bool lcl_validPropState( const XMLPropertyState& rState ) { return rState.mnIndex != -1; } void XMLTextParagraphExport::Add( sal_uInt16 nFamily, MultiPropertySetHelper& rPropSetHelper, const Reference < XPropertySet > & rPropSet, const XMLPropertyState** ppAddStates) { UniReference < SvXMLExportPropertyMapper > xPropMapper; switch( nFamily ) { case XML_STYLE_FAMILY_TEXT_PARAGRAPH: xPropMapper = GetParaPropMapper(); break; } DBG_ASSERT( xPropMapper.is(), "There is the property mapper?" ); vector< XMLPropertyState > xPropStates(xPropMapper->Filter( rPropSet )); if( ppAddStates ) { while( *ppAddStates ) { xPropStates.push_back( **ppAddStates ); ++ppAddStates; } } if( rPropSetHelper.hasProperty( NUMBERING_RULES_AUTO ) ) { Reference < XIndexReplace > xNumRule(rPropSetHelper.getValue( NUMBERING_RULES_AUTO, rPropSet, sal_True ), uno::UNO_QUERY); if( xNumRule.is() && xNumRule->getCount() ) { Reference < XNamed > xNamed( xNumRule, UNO_QUERY ); OUString sName; if( xNamed.is() ) sName = xNamed->getName(); sal_Bool bAdd = !sName.getLength(); if( !bAdd ) { Reference < XPropertySet > xNumPropSet( xNumRule, UNO_QUERY ); const OUString sIsAutomatic( RTL_CONSTASCII_USTRINGPARAM( "IsAutomatic" ) ); if( xNumPropSet.is() && xNumPropSet->getPropertySetInfo() ->hasPropertyByName( sIsAutomatic ) ) { bAdd = *(sal_Bool *)xNumPropSet->getPropertyValue( sIsAutomatic ).getValue(); // Check on outline style (#i73361#) const OUString sNumberingIsOutline( RTL_CONSTASCII_USTRINGPARAM( "NumberingIsOutline" ) ); if ( bAdd && xNumPropSet->getPropertySetInfo() ->hasPropertyByName( sNumberingIsOutline ) ) { bAdd = !(*(sal_Bool *)xNumPropSet->getPropertyValue( sNumberingIsOutline ).getValue()); } } else { bAdd = sal_True; } } if( bAdd ) pListAutoPool->Add( xNumRule ); } } if( !xPropStates.empty() ) { OUString sParent, sCondParent; switch( nFamily ) { case XML_STYLE_FAMILY_TEXT_PARAGRAPH: if( rPropSetHelper.hasProperty( PARA_STYLE_NAME_AUTO ) ) { rPropSetHelper.getValue( PARA_STYLE_NAME_AUTO, rPropSet, sal_True ) >>= sParent; } if( rPropSetHelper.hasProperty( PARA_CONDITIONAL_STYLE_NAME_AUTO ) ) { rPropSetHelper.getValue( PARA_CONDITIONAL_STYLE_NAME_AUTO, rPropSet, sal_True ) >>= sCondParent; } break; } if( find_if( xPropStates.begin(), xPropStates.end(), lcl_validPropState ) != xPropStates.end() ) { GetAutoStylePool().Add( nFamily, sParent, xPropStates ); if( sCondParent.getLength() && sParent != sCondParent ) GetAutoStylePool().Add( nFamily, sCondParent, xPropStates ); } } } OUString XMLTextParagraphExport::Find( sal_uInt16 nFamily, const Reference < XPropertySet > & rPropSet, const OUString& rParent, const XMLPropertyState** ppAddStates) const { OUString sName( rParent ); UniReference < SvXMLExportPropertyMapper > xPropMapper; switch( nFamily ) { case XML_STYLE_FAMILY_TEXT_PARAGRAPH: xPropMapper = GetParaPropMapper(); break; case XML_STYLE_FAMILY_TEXT_FRAME: xPropMapper = GetAutoFramePropMapper(); break; case XML_STYLE_FAMILY_TEXT_SECTION: xPropMapper = GetSectionPropMapper(); break; case XML_STYLE_FAMILY_TEXT_RUBY: xPropMapper = GetRubyPropMapper(); break; } DBG_ASSERT( xPropMapper.is(), "There is the property mapper?" ); if( !xPropMapper.is() ) return sName; vector< XMLPropertyState > xPropStates(xPropMapper->Filter( rPropSet )); if( ppAddStates ) { while( *ppAddStates ) { xPropStates.push_back( **ppAddStates ); ++ppAddStates; } } if( find_if( xPropStates.begin(), xPropStates.end(), lcl_validPropState ) != xPropStates.end() ) sName = GetAutoStylePool().Find( nFamily, sName, xPropStates ); return sName; } OUString XMLTextParagraphExport::FindTextStyleAndHyperlink( const Reference < XPropertySet > & rPropSet, sal_Bool& rbHyperlink, sal_Bool& rbHasCharStyle, sal_Bool& rbHasAutoStyle, const XMLPropertyState** ppAddStates ) const { UniReference < SvXMLExportPropertyMapper > xPropMapper(GetTextPropMapper()); vector< XMLPropertyState > xPropStates(xPropMapper->Filter( rPropSet )); // Get parent and remove hyperlinks (they aren't of interest) OUString sName; rbHyperlink = rbHasCharStyle = rbHasAutoStyle = sal_False; sal_uInt16 nIgnoreProps = 0; UniReference< XMLPropertySetMapper > xPM(xPropMapper->getPropertySetMapper()); ::std::vector< XMLPropertyState >::iterator aFirstDel = xPropStates.end(); ::std::vector< XMLPropertyState >::iterator aSecondDel = xPropStates.end(); for( ::std::vector< XMLPropertyState >::iterator i = xPropStates.begin(); nIgnoreProps < 2 && i != xPropStates.end(); i++ ) { if( i->mnIndex == -1 ) continue; switch( xPM->GetEntryContextId(i->mnIndex) ) { case CTF_CHAR_STYLE_NAME: i->maValue >>= sName; i->mnIndex = -1; rbHasCharStyle = sName.getLength() > 0; if( nIgnoreProps ) aSecondDel = i; else aFirstDel = i; nIgnoreProps++; break; case CTF_HYPERLINK_URL: rbHyperlink = sal_True; i->mnIndex = -1; if( nIgnoreProps ) aSecondDel = i; else aFirstDel = i; nIgnoreProps++; break; } } if( ppAddStates ) { while( *ppAddStates ) { xPropStates.push_back( **ppAddStates ); ppAddStates++; } } if( (xPropStates.size() - nIgnoreProps) > 0L ) { // erase the character style, otherwise the autostyle cannot be found! // erase the hyperlink, otherwise the autostyle cannot be found! if ( nIgnoreProps ) { // If two elements of a vector have to be deleted, // we should delete the second one first. if( --nIgnoreProps ) xPropStates.erase( aSecondDel ); xPropStates.erase( aFirstDel ); } OUString sParent; // AutoStyles should not have parents! sName = GetAutoStylePool().Find( XML_STYLE_FAMILY_TEXT_TEXT, sParent, xPropStates ); DBG_ASSERT( sName.getLength(), "AutoStyle could not be found" ); rbHasAutoStyle = sal_True; } return sName; } OUString XMLTextParagraphExport::FindTextStyle( const Reference < XPropertySet > & rPropSet, sal_Bool& rHasCharStyle ) const { sal_Bool bDummy; sal_Bool bDummy2; return FindTextStyleAndHyperlink( rPropSet, bDummy, rHasCharStyle, bDummy2 ); } // adjustments to support lists independent from list style void XMLTextParagraphExport::exportListChange( const XMLTextNumRuleInfo& rPrevInfo, const XMLTextNumRuleInfo& rNextInfo ) { // end a list if ( rPrevInfo.GetLevel() > 0 ) { sal_Int16 nListLevelsToBeClosed = 0; if ( !rNextInfo.BelongsToSameList( rPrevInfo ) || rNextInfo.GetLevel() <= 0 ) { // close complete previous list nListLevelsToBeClosed = rPrevInfo.GetLevel(); } else if ( rPrevInfo.GetLevel() > rNextInfo.GetLevel() ) { // close corresponding sub lists DBG_ASSERT( rNextInfo.GetLevel() > 0, "<rPrevInfo.GetLevel() > 0> not hold. Serious defect -> please inform OD." ); nListLevelsToBeClosed = rPrevInfo.GetLevel() - rNextInfo.GetLevel(); } if ( nListLevelsToBeClosed > 0 && pListElements && pListElements->Count() >= ( 2 * nListLevelsToBeClosed ) ) { do { for( sal_uInt16 j = 0; j < 2; ++j ) { OUString *pElem = (*pListElements)[pListElements->Count()-1]; pListElements->Remove( pListElements->Count()-1 ); GetExport().EndElement( *pElem, sal_True ); delete pElem; } // remove closed list from list stack mpTextListsHelper->PopListFromStack(); --nListLevelsToBeClosed; } while ( nListLevelsToBeClosed > 0 ); } } const bool bExportODF = ( GetExport().getExportFlags() & EXPORT_OASIS ) != 0; const SvtSaveOptions::ODFDefaultVersion eODFDefaultVersion = GetExport().getDefaultVersion(); // start a new list if ( rNextInfo.GetLevel() > 0 ) { bool bRootListToBeStarted = false; sal_Int16 nListLevelsToBeOpened = 0; if ( !rPrevInfo.BelongsToSameList( rNextInfo ) || rPrevInfo.GetLevel() <= 0 ) { // new root list bRootListToBeStarted = true; nListLevelsToBeOpened = rNextInfo.GetLevel(); } else if ( rNextInfo.GetLevel() > rPrevInfo.GetLevel() ) { // open corresponding sub lists DBG_ASSERT( rPrevInfo.GetLevel() > 0, "<rPrevInfo.GetLevel() > 0> not hold. Serious defect -> please inform OD." ); nListLevelsToBeOpened = rNextInfo.GetLevel() - rPrevInfo.GetLevel(); } if ( nListLevelsToBeOpened > 0 ) { const ::rtl::OUString sListStyleName( rNextInfo.GetNumRulesName() ); // Currently only the text documents support <ListId>. // Thus, for other document types <sListId> is empty. const ::rtl::OUString sListId( rNextInfo.GetListId() ); bool bExportListStyle( true ); bool bRestartNumberingAtContinuedRootList( false ); sal_Int16 nRestartValueForContinuedRootList( -1 ); bool bContinueingPreviousSubList = !bRootListToBeStarted && rNextInfo.IsContinueingPreviousSubTree(); do { GetExport().CheckAttrList(); if ( bRootListToBeStarted ) { if ( !mpTextListsHelper->IsListProcessed( sListId ) ) { if ( bExportODF && eODFDefaultVersion >= SvtSaveOptions::ODFVER_012 && sListId.getLength() > 0 ) { /* Property text:id at element <text:list> has to be replaced by property xml:id (#i92221#) */ GetExport().AddAttribute( XML_NAMESPACE_XML, XML_ID, sListId ); } mpTextListsHelper->KeepListAsProcessed( sListId, sListStyleName, ::rtl::OUString() ); } else { const ::rtl::OUString sNewListId( mpTextListsHelper->GenerateNewListId() ); if ( bExportODF && eODFDefaultVersion >= SvtSaveOptions::ODFVER_012 && sListId.getLength() > 0 ) { /* Property text:id at element <text:list> has to be replaced by property xml:id (#i92221#) */ GetExport().AddAttribute( XML_NAMESPACE_XML, XML_ID, sNewListId ); } const ::rtl::OUString sContinueListId = mpTextListsHelper->GetLastContinuingListId( sListId ); // store that list with list id <sNewListId> is last list, // which has continued list with list id <sListId> mpTextListsHelper->StoreLastContinuingList( sListId, sNewListId ); if ( sListStyleName == mpTextListsHelper->GetListStyleOfLastProcessedList() && // Inconsistent behavior regarding lists (#i92811#) sContinueListId == mpTextListsHelper->GetLastProcessedListId() && !rNextInfo.IsRestart() ) { GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_CONTINUE_NUMBERING, XML_TRUE ); } else { if ( bExportODF && eODFDefaultVersion >= SvtSaveOptions::ODFVER_012 && sListId.getLength() > 0 ) { GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_CONTINUE_LIST, sContinueListId ); } if ( rNextInfo.IsRestart() && ( nListLevelsToBeOpened != 1 || !rNextInfo.HasStartValue() ) ) { bRestartNumberingAtContinuedRootList = true; nRestartValueForContinuedRootList = rNextInfo.GetListLevelStartValue(); } } mpTextListsHelper->KeepListAsProcessed( sNewListId, sListStyleName, sContinueListId ); } GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_STYLE_NAME, GetExport().EncodeStyleName( sListStyleName ) ); bExportListStyle = false; bRootListToBeStarted = false; } else if ( bExportListStyle && !mpTextListsHelper->EqualsToTopListStyleOnStack( sListStyleName ) ) { GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_STYLE_NAME, GetExport().EncodeStyleName( sListStyleName ) ); bExportListStyle = false; } if ( bContinueingPreviousSubList ) { GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_CONTINUE_NUMBERING, XML_TRUE ); bContinueingPreviousSubList = false; } enum XMLTokenEnum eLName = XML_LIST; OUString *pElem = new OUString( GetExport().GetNamespaceMap().GetQNameByKey( XML_NAMESPACE_TEXT, GetXMLToken(eLName) ) ); GetExport().IgnorableWhitespace(); GetExport().StartElement( *pElem, sal_False ); if( !pListElements ) pListElements = new OUStrings_Impl; pListElements->Insert( pElem, pListElements->Count() ); mpTextListsHelper->PushListOnStack( sListId, sListStyleName ); // <text:list-header> or <text:list-item> GetExport().CheckAttrList(); /* Export start value in case of <bRestartNumberingAtContinuedRootList> at correct list item (#i97309#) */ if ( nListLevelsToBeOpened == 1 ) { if ( rNextInfo.HasStartValue() ) { OUStringBuffer aBuffer; aBuffer.append( (sal_Int32)rNextInfo.GetStartValue() ); GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_START_VALUE, aBuffer.makeStringAndClear() ); } else if ( bRestartNumberingAtContinuedRootList ) { OUStringBuffer aBuffer; aBuffer.append( (sal_Int32)nRestartValueForContinuedRootList ); GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_START_VALUE, aBuffer.makeStringAndClear() ); bRestartNumberingAtContinuedRootList = false; } } eLName = ( rNextInfo.IsNumbered() || nListLevelsToBeOpened > 1 ) ? XML_LIST_ITEM : XML_LIST_HEADER; pElem = new OUString( GetExport().GetNamespaceMap().GetQNameByKey( XML_NAMESPACE_TEXT, GetXMLToken(eLName) ) ); GetExport().IgnorableWhitespace(); GetExport().StartElement( *pElem, sal_False ); pListElements->Insert( pElem, pListElements->Count() ); // export of <text:number> element for last opened <text:list-item>, if requested if ( GetExport().exportTextNumberElement() && eLName == XML_LIST_ITEM && nListLevelsToBeOpened == 1 && // last iteration --> last opened <text:list-item> rNextInfo.ListLabelString().getLength() > 0 ) { const ::rtl::OUString aTextNumberElem = OUString( GetExport().GetNamespaceMap().GetQNameByKey( XML_NAMESPACE_TEXT, GetXMLToken(XML_NUMBER) ) ); GetExport().IgnorableWhitespace(); GetExport().StartElement( aTextNumberElem, sal_False ); GetExport().Characters( rNextInfo.ListLabelString() ); GetExport().EndElement( aTextNumberElem, sal_True ); } --nListLevelsToBeOpened; } while ( nListLevelsToBeOpened > 0 ); } } if ( rNextInfo.GetLevel() > 0 && rNextInfo.IsNumbered() && rPrevInfo.BelongsToSameList( rNextInfo ) && rPrevInfo.GetLevel() >= rNextInfo.GetLevel() ) { // close previous list-item DBG_ASSERT( pListElements && pListElements->Count() >= 2, "SwXMLExport::ExportListChange: list elements missing" ); OUString *pElem = (*pListElements)[pListElements->Count()-1]; GetExport().EndElement( *pElem, sal_True ); pListElements->Remove( pListElements->Count()-1 ); delete pElem; // Only for sub lists (#i103745#) if ( rNextInfo.IsRestart() && !rNextInfo.HasStartValue() && rNextInfo.GetLevel() != 1 ) { // start new sub list respectively list on same list level pElem = (*pListElements)[pListElements->Count()-1]; GetExport().EndElement( *pElem, sal_True ); GetExport().IgnorableWhitespace(); GetExport().StartElement( *pElem, sal_False ); } // open new list-item GetExport().CheckAttrList(); if( rNextInfo.HasStartValue() ) { OUStringBuffer aBuffer; aBuffer.append( (sal_Int32)rNextInfo.GetStartValue() ); GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_START_VALUE, aBuffer.makeStringAndClear() ); } // Handle restart without start value on list level 1 (#i103745#) else if ( rNextInfo.IsRestart() && /*!rNextInfo.HasStartValue() &&*/ rNextInfo.GetLevel() == 1 ) { OUStringBuffer aBuffer; aBuffer.append( (sal_Int32)rNextInfo.GetListLevelStartValue() ); GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_START_VALUE, aBuffer.makeStringAndClear() ); } if ( ( GetExport().getExportFlags() & EXPORT_OASIS ) != 0 && GetExport().getDefaultVersion() >= SvtSaveOptions::ODFVER_012 ) { const ::rtl::OUString sListStyleName( rNextInfo.GetNumRulesName() ); if ( !mpTextListsHelper->EqualsToTopListStyleOnStack( sListStyleName ) ) { GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_STYLE_OVERRIDE, GetExport().EncodeStyleName( sListStyleName ) ); } } pElem = new OUString( GetExport().GetNamespaceMap().GetQNameByKey( XML_NAMESPACE_TEXT, GetXMLToken(XML_LIST_ITEM) ) ); GetExport().IgnorableWhitespace(); GetExport().StartElement( *pElem, sal_False ); pListElements->Insert( pElem, pListElements->Count() ); // export of <text:number> element for <text:list-item>, if requested if ( GetExport().exportTextNumberElement() && rNextInfo.ListLabelString().getLength() > 0 ) { const ::rtl::OUString aTextNumberElem = OUString( GetExport().GetNamespaceMap().GetQNameByKey( XML_NAMESPACE_TEXT, GetXMLToken(XML_NUMBER) ) ); GetExport().IgnorableWhitespace(); GetExport().StartElement( aTextNumberElem, sal_False ); GetExport().Characters( rNextInfo.ListLabelString() ); GetExport().EndElement( aTextNumberElem, sal_True ); } } } XMLTextParagraphExport::XMLTextParagraphExport( SvXMLExport& rExp, SvXMLAutoStylePoolP & rASP ) : XMLStyleExport( rExp, OUString(), &rASP ), rAutoStylePool( rASP ), pBoundFrameSets(new BoundFrameSets(GetExport().GetModel())), pFieldExport( 0 ), pListElements( 0 ), pListAutoPool( new XMLTextListAutoStylePool( this->GetExport() ) ), pSectionExport( NULL ), pIndexMarkExport( NULL ), pRedlineExport( NULL ), pHeadingStyles( NULL ), bProgress( sal_False ), bBlock( sal_False ), bOpenRuby( sal_False ), mpTextListsHelper( 0 ), maTextListsHelperStack(), sActualSize(RTL_CONSTASCII_USTRINGPARAM("ActualSize")), // Implement Title/Description Elements UI (#i73249#) sTitle(RTL_CONSTASCII_USTRINGPARAM("Title")), sDescription(RTL_CONSTASCII_USTRINGPARAM("Description")), sAnchorCharStyleName(RTL_CONSTASCII_USTRINGPARAM("AnchorCharStyleName")), sAnchorPageNo(RTL_CONSTASCII_USTRINGPARAM("AnchorPageNo")), sAnchorType(RTL_CONSTASCII_USTRINGPARAM("AnchorType")), sBeginNotice(RTL_CONSTASCII_USTRINGPARAM("BeginNotice")), sBookmark(RTL_CONSTASCII_USTRINGPARAM("Bookmark")), sCategory(RTL_CONSTASCII_USTRINGPARAM("Category")), sChainNextName(RTL_CONSTASCII_USTRINGPARAM("ChainNextName")), sCharStyleName(RTL_CONSTASCII_USTRINGPARAM("CharStyleName")), sCharStyleNames(RTL_CONSTASCII_USTRINGPARAM("CharStyleNames")), sContourPolyPolygon(RTL_CONSTASCII_USTRINGPARAM("ContourPolyPolygon")), sDocumentIndex(RTL_CONSTASCII_USTRINGPARAM("DocumentIndex")), sDocumentIndexMark(RTL_CONSTASCII_USTRINGPARAM("DocumentIndexMark")), sEndNotice(RTL_CONSTASCII_USTRINGPARAM("EndNotice")), sFootnote(RTL_CONSTASCII_USTRINGPARAM("Footnote")), sFootnoteCounting(RTL_CONSTASCII_USTRINGPARAM("FootnoteCounting")), sFrame(RTL_CONSTASCII_USTRINGPARAM("Frame")), sFrameHeightAbsolute(RTL_CONSTASCII_USTRINGPARAM("FrameHeightAbsolute")), sFrameHeightPercent(RTL_CONSTASCII_USTRINGPARAM("FrameHeightPercent")), sFrameStyleName(RTL_CONSTASCII_USTRINGPARAM("FrameStyleName")), sFrameWidthAbsolute(RTL_CONSTASCII_USTRINGPARAM("FrameWidthAbsolute")), sFrameWidthPercent(RTL_CONSTASCII_USTRINGPARAM("FrameWidthPercent")), sGraphicFilter(RTL_CONSTASCII_USTRINGPARAM("GraphicFilter")), sGraphicRotation(RTL_CONSTASCII_USTRINGPARAM("GraphicRotation")), sGraphicURL(RTL_CONSTASCII_USTRINGPARAM("GraphicURL")), sHeight(RTL_CONSTASCII_USTRINGPARAM("Height")), sHoriOrient(RTL_CONSTASCII_USTRINGPARAM("HoriOrient")), sHoriOrientPosition(RTL_CONSTASCII_USTRINGPARAM("HoriOrientPosition")), sHyperLinkName(RTL_CONSTASCII_USTRINGPARAM("HyperLinkName")), sHyperLinkTarget(RTL_CONSTASCII_USTRINGPARAM("HyperLinkTarget")), sHyperLinkURL(RTL_CONSTASCII_USTRINGPARAM("HyperLinkURL")), sIsAutomaticContour(RTL_CONSTASCII_USTRINGPARAM("IsAutomaticContour")), sIsCollapsed(RTL_CONSTASCII_USTRINGPARAM("IsCollapsed")), sIsPixelContour(RTL_CONSTASCII_USTRINGPARAM("IsPixelContour")), sIsStart(RTL_CONSTASCII_USTRINGPARAM("IsStart")), sIsSyncHeightToWidth(RTL_CONSTASCII_USTRINGPARAM("IsSyncHeightToWidth")), sIsSyncWidthToHeight(RTL_CONSTASCII_USTRINGPARAM("IsSyncWidthToHeight")), sNumberingRules(RTL_CONSTASCII_USTRINGPARAM("NumberingRules")), sNumberingType(RTL_CONSTASCII_USTRINGPARAM("NumberingType")), sPageDescName(RTL_CONSTASCII_USTRINGPARAM("PageDescName")), sPageStyleName(RTL_CONSTASCII_USTRINGPARAM("PageStyleName")), sParaChapterNumberingLevel(RTL_CONSTASCII_USTRINGPARAM("ParaChapterNumberingLevel")), sParaConditionalStyleName(RTL_CONSTASCII_USTRINGPARAM("ParaConditionalStyleName")), sParagraphService(RTL_CONSTASCII_USTRINGPARAM("com.sun.star.text.Paragraph")), sParaStyleName(RTL_CONSTASCII_USTRINGPARAM("ParaStyleName")), sPositionEndOfDoc(RTL_CONSTASCII_USTRINGPARAM("PositionEndOfDoc")), sPrefix(RTL_CONSTASCII_USTRINGPARAM("Prefix")), sRedline(RTL_CONSTASCII_USTRINGPARAM("Redline")), sReferenceId(RTL_CONSTASCII_USTRINGPARAM("ReferenceId")), sReferenceMark(RTL_CONSTASCII_USTRINGPARAM("ReferenceMark")), sRelativeHeight(RTL_CONSTASCII_USTRINGPARAM("RelativeHeight")), sRelativeWidth(RTL_CONSTASCII_USTRINGPARAM("RelativeWidth")), sRuby(RTL_CONSTASCII_USTRINGPARAM("Ruby")), sRubyAdjust(RTL_CONSTASCII_USTRINGPARAM("RubyAdjust")), sRubyCharStyleName(RTL_CONSTASCII_USTRINGPARAM("RubyCharStyleName")), sRubyText(RTL_CONSTASCII_USTRINGPARAM("RubyText")), sServerMap(RTL_CONSTASCII_USTRINGPARAM("ServerMap")), sShapeService(RTL_CONSTASCII_USTRINGPARAM("com.sun.star.drawing.Shape")), sSizeType(RTL_CONSTASCII_USTRINGPARAM("SizeType")), sSoftPageBreak( RTL_CONSTASCII_USTRINGPARAM( "SoftPageBreak" ) ), sStartAt(RTL_CONSTASCII_USTRINGPARAM("StartAt")), sSuffix(RTL_CONSTASCII_USTRINGPARAM("Suffix")), sTableService(RTL_CONSTASCII_USTRINGPARAM("com.sun.star.text.TextTable")), sText(RTL_CONSTASCII_USTRINGPARAM("Text")), sTextContentService(RTL_CONSTASCII_USTRINGPARAM("com.sun.star.text.TextContent")), sTextEmbeddedService(RTL_CONSTASCII_USTRINGPARAM("com.sun.star.text.TextEmbeddedObject")), sTextEndnoteService(RTL_CONSTASCII_USTRINGPARAM("com.sun.star.text.Endnote")), sTextField(RTL_CONSTASCII_USTRINGPARAM("TextField")), sTextFieldService(RTL_CONSTASCII_USTRINGPARAM("com.sun.star.text.TextField")), sTextFrameService(RTL_CONSTASCII_USTRINGPARAM("com.sun.star.text.TextFrame")), sTextGraphicService(RTL_CONSTASCII_USTRINGPARAM("com.sun.star.text.TextGraphicObject")), sTextPortionType(RTL_CONSTASCII_USTRINGPARAM("TextPortionType")), sTextSection(RTL_CONSTASCII_USTRINGPARAM("TextSection")), sUnvisitedCharStyleName(RTL_CONSTASCII_USTRINGPARAM("UnvisitedCharStyleName")), sVertOrient(RTL_CONSTASCII_USTRINGPARAM("VertOrient")), sVertOrientPosition(RTL_CONSTASCII_USTRINGPARAM("VertOrientPosition")), sVisitedCharStyleName(RTL_CONSTASCII_USTRINGPARAM("VisitedCharStyleName")), sWidth(RTL_CONSTASCII_USTRINGPARAM("Width")), sWidthType( RTL_CONSTASCII_USTRINGPARAM( "WidthType" ) ), sTextFieldStart( RTL_CONSTASCII_USTRINGPARAM( "TextFieldStart" ) ), sTextFieldEnd( RTL_CONSTASCII_USTRINGPARAM( "TextFieldEnd" ) ), sTextFieldStartEnd( RTL_CONSTASCII_USTRINGPARAM( "TextFieldStartEnd" ) ), aCharStyleNamesPropInfoCache( sCharStyleNames ) { UniReference < XMLPropertySetMapper > xPropMapper(new XMLTextPropertySetMapper( TEXT_PROP_MAP_PARA )); xParaPropMapper = new XMLTextExportPropertySetMapper( xPropMapper, GetExport() ); OUString sFamily( GetXMLToken(XML_PARAGRAPH) ); OUString aPrefix( String( 'P' ) ); rAutoStylePool.AddFamily( XML_STYLE_FAMILY_TEXT_PARAGRAPH, sFamily, xParaPropMapper, aPrefix ); xPropMapper = new XMLTextPropertySetMapper( TEXT_PROP_MAP_TEXT ); xTextPropMapper = new XMLTextExportPropertySetMapper( xPropMapper, GetExport() ); sFamily = OUString( GetXMLToken(XML_TEXT) ); aPrefix = OUString( String( 'T' ) ); rAutoStylePool.AddFamily( XML_STYLE_FAMILY_TEXT_TEXT, sFamily, xTextPropMapper, aPrefix ); xPropMapper = new XMLTextPropertySetMapper( TEXT_PROP_MAP_AUTO_FRAME ); xAutoFramePropMapper = new XMLTextExportPropertySetMapper( xPropMapper, GetExport() ); sFamily = OUString( RTL_CONSTASCII_USTRINGPARAM(XML_STYLE_FAMILY_SD_GRAPHICS_NAME) ); aPrefix = OUString( RTL_CONSTASCII_USTRINGPARAM( "fr" ) ); rAutoStylePool.AddFamily( XML_STYLE_FAMILY_TEXT_FRAME, sFamily, xAutoFramePropMapper, aPrefix ); xPropMapper = new XMLTextPropertySetMapper( TEXT_PROP_MAP_SECTION ); xSectionPropMapper = new XMLTextExportPropertySetMapper( xPropMapper, GetExport() ); sFamily = OUString( GetXMLToken( XML_SECTION ) ); aPrefix = OUString( RTL_CONSTASCII_USTRINGPARAM( "Sect" ) ); rAutoStylePool.AddFamily( XML_STYLE_FAMILY_TEXT_SECTION, sFamily, xSectionPropMapper, aPrefix ); xPropMapper = new XMLTextPropertySetMapper( TEXT_PROP_MAP_RUBY ); xRubyPropMapper = new SvXMLExportPropertyMapper( xPropMapper ); sFamily = OUString( GetXMLToken( XML_RUBY ) ); aPrefix = OUString( RTL_CONSTASCII_USTRINGPARAM( "Ru" ) ); rAutoStylePool.AddFamily( XML_STYLE_FAMILY_TEXT_RUBY, sFamily, xRubyPropMapper, aPrefix ); xPropMapper = new XMLTextPropertySetMapper( TEXT_PROP_MAP_FRAME ); xFramePropMapper = new XMLTextExportPropertySetMapper( xPropMapper, GetExport() ); pSectionExport = new XMLSectionExport( rExp, *this ); pIndexMarkExport = new XMLIndexMarkExport( rExp, *this ); if( ! IsBlockMode() && Reference<XRedlinesSupplier>( GetExport().GetModel(), UNO_QUERY ).is()) pRedlineExport = new XMLRedlineExport( rExp ); // The text field helper needs a pre-constructed XMLPropertyState // to export the combined characters field. We construct that // here, because we need the text property mapper to do it. // construct Any value, then find index sal_Int32 nIndex = xTextPropMapper->getPropertySetMapper()->FindEntryIndex( "", XML_NAMESPACE_STYLE, GetXMLToken(XML_TEXT_COMBINE)); pFieldExport = new XMLTextFieldExport( rExp, new XMLPropertyState( nIndex, uno::makeAny(sal_True) ) ); PushNewTextListsHelper(); } XMLTextParagraphExport::~XMLTextParagraphExport() { delete pHeadingStyles; delete pRedlineExport; delete pIndexMarkExport; delete pSectionExport; delete pFieldExport; delete pListElements; delete pListAutoPool; #ifdef DBG_UTIL txtparae_bContainsIllegalCharacters = sal_False; #endif PopTextListsHelper(); DBG_ASSERT( maTextListsHelperStack.size() == 0, "misusage of text lists helper stack - it is not empty. Serious defect - please inform OD" ); } SvXMLExportPropertyMapper *XMLTextParagraphExport::CreateShapeExtPropMapper( SvXMLExport& rExport ) { UniReference < XMLPropertySetMapper > xPropMapper = new XMLTextPropertySetMapper( TEXT_PROP_MAP_SHAPE ); return new XMLTextExportPropertySetMapper( xPropMapper, rExport ); } SvXMLExportPropertyMapper *XMLTextParagraphExport::CreateCharExtPropMapper( SvXMLExport& rExport) { XMLPropertySetMapper *pPropMapper = new XMLTextPropertySetMapper( TEXT_PROP_MAP_TEXT ); return new XMLTextExportPropertySetMapper( pPropMapper, rExport ); } SvXMLExportPropertyMapper *XMLTextParagraphExport::CreateParaExtPropMapper( SvXMLExport& rExport) { XMLPropertySetMapper *pPropMapper = new XMLTextPropertySetMapper( TEXT_PROP_MAP_SHAPE_PARA ); return new XMLTextExportPropertySetMapper( pPropMapper, rExport ); } SvXMLExportPropertyMapper *XMLTextParagraphExport::CreateParaDefaultExtPropMapper( SvXMLExport& rExport) { XMLPropertySetMapper *pPropMapper = new XMLTextPropertySetMapper( TEXT_PROP_MAP_TEXT_ADDITIONAL_DEFAULTS ); return new XMLTextExportPropertySetMapper( pPropMapper, rExport ); } void XMLTextParagraphExport::exportPageFrames( sal_Bool bAutoStyles, sal_Bool bIsProgress ) { const TextContentSet* const pTexts = pBoundFrameSets->GetTexts()->GetPageBoundContents(); const TextContentSet* const pGraphics = pBoundFrameSets->GetGraphics()->GetPageBoundContents(); const TextContentSet* const pEmbeddeds = pBoundFrameSets->GetEmbeddeds()->GetPageBoundContents(); const TextContentSet* const pShapes = pBoundFrameSets->GetShapes()->GetPageBoundContents(); for(TextContentSet::const_iterator_t it = pTexts->getBegin(); it != pTexts->getEnd(); ++it) exportTextFrame(*it, bAutoStyles, bIsProgress, sal_True); for(TextContentSet::const_iterator_t it = pGraphics->getBegin(); it != pGraphics->getEnd(); ++it) exportTextGraphic(*it, bAutoStyles); for(TextContentSet::const_iterator_t it = pEmbeddeds->getBegin(); it != pEmbeddeds->getEnd(); ++it) exportTextEmbedded(*it, bAutoStyles); for(TextContentSet::const_iterator_t it = pShapes->getBegin(); it != pShapes->getEnd(); ++it) exportShape(*it, bAutoStyles); } void XMLTextParagraphExport::exportFrameFrames( sal_Bool bAutoStyles, sal_Bool bIsProgress, const Reference < XTextFrame > *pParentTxtFrame ) { const TextContentSet* const pTexts = pBoundFrameSets->GetTexts()->GetFrameBoundContents(*pParentTxtFrame); if(pTexts) for(TextContentSet::const_iterator_t it = pTexts->getBegin(); it != pTexts->getEnd(); ++it) exportTextFrame(*it, bAutoStyles, bIsProgress, sal_True); const TextContentSet* const pGraphics = pBoundFrameSets->GetGraphics()->GetFrameBoundContents(*pParentTxtFrame); if(pGraphics) for(TextContentSet::const_iterator_t it = pGraphics->getBegin(); it != pGraphics->getEnd(); ++it) exportTextGraphic(*it, bAutoStyles); const TextContentSet* const pEmbeddeds = pBoundFrameSets->GetEmbeddeds()->GetFrameBoundContents(*pParentTxtFrame); if(pEmbeddeds) for(TextContentSet::const_iterator_t it = pEmbeddeds->getBegin(); it != pEmbeddeds->getEnd(); ++it) exportTextEmbedded(*it, bAutoStyles); const TextContentSet* const pShapes = pBoundFrameSets->GetShapes()->GetFrameBoundContents(*pParentTxtFrame); if(pShapes) for(TextContentSet::const_iterator_t it = pShapes->getBegin(); it != pShapes->getEnd(); ++it) exportShape(*it, bAutoStyles); } // bookmarks, reference marks (and TOC marks) are the same except for the // element names. We use the same method for export and it an array with // the proper element names static const enum XMLTokenEnum lcl_XmlReferenceElements[] = { XML_REFERENCE_MARK, XML_REFERENCE_MARK_START, XML_REFERENCE_MARK_END }; static const enum XMLTokenEnum lcl_XmlBookmarkElements[] = { XML_BOOKMARK, XML_BOOKMARK_START, XML_BOOKMARK_END }; // This function replaces the text portion iteration during auto style // collection. bool XMLTextParagraphExport::collectTextAutoStylesOptimized( sal_Bool bIsProgress ) { GetExport().GetShapeExport(); // make sure the graphics styles family is added const sal_Bool bAutoStyles = sal_True; const sal_Bool bExportContent = sal_False; // Export AutoStyles: Reference< XAutoStylesSupplier > xAutoStylesSupp( GetExport().GetModel(), UNO_QUERY ); if ( xAutoStylesSupp.is() ) { Reference< XAutoStyles > xAutoStyleFamilies = xAutoStylesSupp->getAutoStyles(); OUString sName; sal_uInt16 nFamily; for ( int i = 0; i < 3; ++i ) { if ( 0 == i ) { sName = OUString( RTL_CONSTASCII_USTRINGPARAM( "CharacterStyles" ) ); nFamily = XML_STYLE_FAMILY_TEXT_TEXT; } else if ( 1 == i ) { sName = OUString( RTL_CONSTASCII_USTRINGPARAM( "RubyStyles" ) ); nFamily = XML_STYLE_FAMILY_TEXT_RUBY; } else { sName = OUString( RTL_CONSTASCII_USTRINGPARAM( "ParagraphStyles" ) ); nFamily = XML_STYLE_FAMILY_TEXT_PARAGRAPH; } Any aAny = xAutoStyleFamilies->getByName( sName ); Reference< XAutoStyleFamily > xAutoStyles = *(Reference<XAutoStyleFamily>*)aAny.getValue(); Reference < XEnumeration > xAutoStylesEnum( xAutoStyles->createEnumeration() ); while ( xAutoStylesEnum->hasMoreElements() ) { aAny = xAutoStylesEnum->nextElement(); Reference< XAutoStyle > xAutoStyle = *(Reference<XAutoStyle>*)aAny.getValue(); Reference < XPropertySet > xPSet( xAutoStyle, uno::UNO_QUERY ); Add( nFamily, xPSet, 0, true ); } } } // Export Field AutoStyles: Reference< XTextFieldsSupplier > xTextFieldsSupp( GetExport().GetModel(), UNO_QUERY ); if ( xTextFieldsSupp.is() ) { Reference< XEnumerationAccess > xTextFields = xTextFieldsSupp->getTextFields(); Reference < XEnumeration > xTextFieldsEnum( xTextFields->createEnumeration() ); while ( xTextFieldsEnum->hasMoreElements() ) { Any aAny = xTextFieldsEnum->nextElement(); Reference< XTextField > xTextField = *(Reference<XTextField>*)aAny.getValue(); exportTextField( xTextField, bAutoStyles, bIsProgress, !xAutoStylesSupp.is() ); try { Reference < XPropertySet > xSet( xTextField, UNO_QUERY ); Reference < XText > xText; Any a = xSet->getPropertyValue( ::rtl::OUString(RTL_CONSTASCII_USTRINGPARAM("TextRange")) ); a >>= xText; if ( xText.is() ) { exportText( xText, sal_True, bIsProgress, bExportContent ); GetExport().GetTextParagraphExport() ->collectTextAutoStyles( xText ); } } catch (Exception&) { } } } // Export text frames: Reference<XEnumeration> xTextFramesEnum = pBoundFrameSets->GetTexts()->createEnumeration(); if(xTextFramesEnum.is()) while(xTextFramesEnum->hasMoreElements()) { Reference<XTextContent> xTxtCntnt(xTextFramesEnum->nextElement(), UNO_QUERY); if(xTxtCntnt.is()) exportTextFrame(xTxtCntnt, bAutoStyles, bIsProgress, bExportContent, 0); } // Export graphic objects: Reference<XEnumeration> xGraphicsEnum = pBoundFrameSets->GetGraphics()->createEnumeration(); if(xGraphicsEnum.is()) while(xGraphicsEnum->hasMoreElements()) { Reference<XTextContent> xTxtCntnt(xGraphicsEnum->nextElement(), UNO_QUERY); if(xTxtCntnt.is()) exportTextGraphic(xTxtCntnt, true, 0); } // Export embedded objects: Reference<XEnumeration> xEmbeddedsEnum = pBoundFrameSets->GetEmbeddeds()->createEnumeration(); if(xEmbeddedsEnum.is()) while(xEmbeddedsEnum->hasMoreElements()) { Reference<XTextContent> xTxtCntnt(xEmbeddedsEnum->nextElement(), UNO_QUERY); if(xTxtCntnt.is()) exportTextEmbedded(xTxtCntnt, true, 0); } // Export shapes: Reference<XEnumeration> xShapesEnum = pBoundFrameSets->GetShapes()->createEnumeration(); if(xShapesEnum.is()) while(xShapesEnum->hasMoreElements()) { Reference<XTextContent> xTxtCntnt(xShapesEnum->nextElement(), UNO_QUERY); if(xTxtCntnt.is()) { Reference<XServiceInfo> xServiceInfo(xTxtCntnt, UNO_QUERY); if( xServiceInfo->supportsService(sShapeService)) exportShape(xTxtCntnt, true, 0); } } sal_Int32 nCount; // AutoStyles for sections Reference< XTextSectionsSupplier > xSectionsSupp( GetExport().GetModel(), UNO_QUERY ); if ( xSectionsSupp.is() ) { Reference< XIndexAccess > xSections( xSectionsSupp->getTextSections(), UNO_QUERY ); if ( xSections.is() ) { nCount = xSections->getCount(); for( sal_Int32 i = 0; i < nCount; ++i ) { Any aAny = xSections->getByIndex( i ); Reference< XTextSection > xSection = *(Reference<XTextSection>*)aAny.getValue(); Reference < XPropertySet > xPSet( xSection, uno::UNO_QUERY ); Add( XML_STYLE_FAMILY_TEXT_SECTION, xPSet ); } } } // AutoStyles for tables (Note: suppress autostyle collection for paragraphs in exportTable) Reference< XTextTablesSupplier > xTablesSupp( GetExport().GetModel(), UNO_QUERY ); if ( xTablesSupp.is() ) { Reference< XIndexAccess > xTables( xTablesSupp->getTextTables(), UNO_QUERY ); if ( xTables.is() ) { nCount = xTables->getCount(); for( sal_Int32 i = 0; i < nCount; ++i ) { Any aAny = xTables->getByIndex( i ); Reference< XTextTable > xTable = *(Reference<XTextTable>*)aAny.getValue(); Reference < XTextContent > xTextContent( xTable, uno::UNO_QUERY ); exportTable( xTextContent, sal_True, sal_True ); } } } Reference< XNumberingRulesSupplier > xNumberingRulesSupp( GetExport().GetModel(), UNO_QUERY ); if ( xNumberingRulesSupp.is() ) { Reference< XIndexAccess > xNumberingRules = xNumberingRulesSupp->getNumberingRules(); nCount = xNumberingRules->getCount(); // Custom outline assignment lost after re-importing sxw (#i73361#) const OUString sNumberingIsOutline( RTL_CONSTASCII_USTRINGPARAM( "NumberingIsOutline" ) ); for( sal_Int32 i = 0; i < nCount; ++i ) { Reference< XIndexReplace > xNumRule( xNumberingRules->getByIndex( i ), UNO_QUERY ); if( xNumRule.is() && xNumRule->getCount() ) { Reference < XNamed > xNamed( xNumRule, UNO_QUERY ); OUString sName; if( xNamed.is() ) sName = xNamed->getName(); sal_Bool bAdd = !sName.getLength(); if( !bAdd ) { Reference < XPropertySet > xNumPropSet( xNumRule, UNO_QUERY ); const OUString sIsAutomatic( RTL_CONSTASCII_USTRINGPARAM( "IsAutomatic" ) ); if( xNumPropSet.is() && xNumPropSet->getPropertySetInfo() ->hasPropertyByName( sIsAutomatic ) ) { bAdd = *(sal_Bool *)xNumPropSet->getPropertyValue( sIsAutomatic ).getValue(); // Check on outline style (#i73361#) if ( bAdd && xNumPropSet->getPropertySetInfo() ->hasPropertyByName( sNumberingIsOutline ) ) { bAdd = !(*(sal_Bool *)xNumPropSet->getPropertyValue( sNumberingIsOutline ).getValue()); } } else { bAdd = sal_True; } } if( bAdd ) pListAutoPool->Add( xNumRule ); } } } return true; } void XMLTextParagraphExport::exportText( const Reference < XText > & rText, sal_Bool bAutoStyles, sal_Bool bIsProgress, sal_Bool bExportParagraph ) { if( bAutoStyles ) GetExport().GetShapeExport(); // make sure the graphics styles family // is added Reference < XEnumerationAccess > xEA( rText, UNO_QUERY ); Reference < XEnumeration > xParaEnum(xEA->createEnumeration()); Reference < XPropertySet > xPropertySet( rText, UNO_QUERY ); Reference < XTextSection > xBaseSection; // #97718# footnotes don't supply paragraph enumerations in some cases // This is always a bug, but at least we don't want to crash. DBG_ASSERT( xParaEnum.is(), "We need a paragraph enumeration" ); if( ! xParaEnum.is() ) return; sal_Bool bExportLevels = sal_True; if (xPropertySet.is()) { Reference < XPropertySetInfo > xInfo ( xPropertySet->getPropertySetInfo() ); if( xInfo.is() ) { if (xInfo->hasPropertyByName( sTextSection )) { xPropertySet->getPropertyValue(sTextSection) >>= xBaseSection ; } /* #i35937# // for applications that use the outliner we need to check if // the current text object needs the level information exported if( !bAutoStyles ) { // fixme: move string to class member, couldn't do now because // of no incompatible build OUString sHasLevels( RTL_CONSTASCII_USTRINGPARAM("HasLevels") ); if (xInfo->hasPropertyByName( sHasLevels ) ) { xPropertySet->getPropertyValue(sHasLevels) >>= bExportLevels; } } */ } } // #96530# Export redlines at start & end of XText before & after // exporting the text content enumeration if( !bAutoStyles && (pRedlineExport != NULL) ) pRedlineExport->ExportStartOrEndRedline( xPropertySet, sal_True ); exportTextContentEnumeration( xParaEnum, bAutoStyles, xBaseSection, bIsProgress, bExportParagraph, 0, bExportLevels ); if( !bAutoStyles && (pRedlineExport != NULL) ) pRedlineExport->ExportStartOrEndRedline( xPropertySet, sal_False ); } void XMLTextParagraphExport::exportText( const Reference < XText > & rText, const Reference < XTextSection > & rBaseSection, sal_Bool bAutoStyles, sal_Bool bIsProgress, sal_Bool bExportParagraph ) { if( bAutoStyles ) GetExport().GetShapeExport(); // make sure the graphics styles family // is added Reference < XEnumerationAccess > xEA( rText, UNO_QUERY ); Reference < XEnumeration > xParaEnum(xEA->createEnumeration()); // #98165# don't continue without a paragraph enumeration if( ! xParaEnum.is() ) return; // #96530# Export redlines at start & end of XText before & after // exporting the text content enumeration Reference<XPropertySet> xPropertySet; if( !bAutoStyles && (pRedlineExport != NULL) ) { xPropertySet.set(rText, uno::UNO_QUERY ); pRedlineExport->ExportStartOrEndRedline( xPropertySet, sal_True ); } exportTextContentEnumeration( xParaEnum, bAutoStyles, rBaseSection, bIsProgress, bExportParagraph ); if( !bAutoStyles && (pRedlineExport != NULL) ) pRedlineExport->ExportStartOrEndRedline( xPropertySet, sal_False ); } sal_Bool XMLTextParagraphExport::exportTextContentEnumeration( const Reference < XEnumeration > & rContEnum, sal_Bool bAutoStyles, const Reference < XTextSection > & rBaseSection, sal_Bool bIsProgress, sal_Bool bExportParagraph, const Reference < XPropertySet > *pRangePropSet, sal_Bool bExportLevels ) { DBG_ASSERT( rContEnum.is(), "No enumeration to export!" ); sal_Bool bHasMoreElements = rContEnum->hasMoreElements(); if( !bHasMoreElements ) return sal_False; XMLTextNumRuleInfo aPrevNumInfo; XMLTextNumRuleInfo aNextNumInfo; sal_Bool bHasContent = sal_False; Reference<XTextSection> xCurrentTextSection(rBaseSection); MultiPropertySetHelper aPropSetHelper( bAutoStyles ? aParagraphPropertyNamesAuto : aParagraphPropertyNames ); sal_Bool bHoldElement = sal_False; Reference < XTextContent > xTxtCntnt; while( bHoldElement || bHasMoreElements ) { if (bHoldElement) { bHoldElement = sal_False; } else { xTxtCntnt.set(rContEnum->nextElement(), uno::UNO_QUERY); aPropSetHelper.resetValues(); } Reference<XServiceInfo> xServiceInfo( xTxtCntnt, UNO_QUERY ); if( xServiceInfo->supportsService( sParagraphService ) ) { if( bExportLevels ) { if( bAutoStyles ) { exportListAndSectionChange( xCurrentTextSection, xTxtCntnt, aPrevNumInfo, aNextNumInfo, bAutoStyles ); } else { /* Pass list auto style pool to <XMLTextNumRuleInfo> instance Pass info about request to export <text:number> element to <XMLTextNumRuleInfo> instance (#i69627#) */ aNextNumInfo.Set( xTxtCntnt, GetExport().writeOutlineStyleAsNormalListStyle(), GetListAutoStylePool(), GetExport().exportTextNumberElement() ); exportListAndSectionChange( xCurrentTextSection, aPropSetHelper, TEXT_SECTION, xTxtCntnt, aPrevNumInfo, aNextNumInfo, bAutoStyles ); } } // if we found a mute section: skip all section content if (pSectionExport->IsMuteSection(xCurrentTextSection)) { // Make sure headings are exported anyway. if( !bAutoStyles ) pSectionExport->ExportMasterDocHeadingDummies(); while (rContEnum->hasMoreElements() && pSectionExport->IsInSection( xCurrentTextSection, xTxtCntnt, sal_True )) { xTxtCntnt.set(rContEnum->nextElement(), uno::UNO_QUERY); aPropSetHelper.resetValues(); aNextNumInfo.Reset(); } // the first non-mute element still needs to be processed bHoldElement = ! pSectionExport->IsInSection( xCurrentTextSection, xTxtCntnt, sal_False ); } else exportParagraph( xTxtCntnt, bAutoStyles, bIsProgress, bExportParagraph, aPropSetHelper ); bHasContent = sal_True; } else if( xServiceInfo->supportsService( sTableService ) ) { if( !bAutoStyles ) { aNextNumInfo.Reset(); } exportListAndSectionChange( xCurrentTextSection, xTxtCntnt, aPrevNumInfo, aNextNumInfo, bAutoStyles ); if (! pSectionExport->IsMuteSection(xCurrentTextSection)) { // export start + end redlines (for wholly redlined tables) if ((! bAutoStyles) && (NULL != pRedlineExport)) pRedlineExport->ExportStartOrEndRedline(xTxtCntnt, sal_True); exportTable( xTxtCntnt, bAutoStyles, bIsProgress ); if ((! bAutoStyles) && (NULL != pRedlineExport)) pRedlineExport->ExportStartOrEndRedline(xTxtCntnt, sal_False); } else if( !bAutoStyles ) { // Make sure headings are exported anyway. pSectionExport->ExportMasterDocHeadingDummies(); } bHasContent = sal_True; } else if( xServiceInfo->supportsService( sTextFrameService ) ) { exportTextFrame( xTxtCntnt, bAutoStyles, bIsProgress, sal_True, pRangePropSet ); } else if( xServiceInfo->supportsService( sTextGraphicService ) ) { exportTextGraphic( xTxtCntnt, bAutoStyles, pRangePropSet ); } else if( xServiceInfo->supportsService( sTextEmbeddedService ) ) { exportTextEmbedded( xTxtCntnt, bAutoStyles, pRangePropSet ); } else if( xServiceInfo->supportsService( sShapeService ) ) { exportShape( xTxtCntnt, bAutoStyles, pRangePropSet ); } else { DBG_ASSERT( !xTxtCntnt.is(), "unknown text content" ); } if( !bAutoStyles ) { aPrevNumInfo = aNextNumInfo; } bHasMoreElements = rContEnum->hasMoreElements(); } if( bExportLevels && bHasContent && !bAutoStyles ) { aNextNumInfo.Reset(); // close open lists and sections; no new styles exportListAndSectionChange( xCurrentTextSection, rBaseSection, aPrevNumInfo, aNextNumInfo, bAutoStyles ); } return sal_True; } void XMLTextParagraphExport::exportParagraph( const Reference < XTextContent > & rTextContent, sal_Bool bAutoStyles, sal_Bool bIsProgress, sal_Bool bExportParagraph, MultiPropertySetHelper& rPropSetHelper) { sal_Int16 nOutlineLevel = -1; if( bIsProgress ) { ProgressBarHelper *pProgress = GetExport().GetProgressBarHelper(); pProgress->SetValue( pProgress->GetValue()+1 ); } // get property set or multi property set and initialize helper Reference<XMultiPropertySet> xMultiPropSet( rTextContent, UNO_QUERY ); Reference<XPropertySet> xPropSet( rTextContent, UNO_QUERY ); // check for supported properties if( !rPropSetHelper.checkedProperties() ) rPropSetHelper.hasProperties( xPropSet->getPropertySetInfo() ); // if( xMultiPropSet.is() ) // rPropSetHelper.getValues( xMultiPropSet ); // else // rPropSetHelper.getValues( xPropSet ); if( bExportParagraph ) { if( bAutoStyles ) { Add( XML_STYLE_FAMILY_TEXT_PARAGRAPH, rPropSetHelper, xPropSet ); } else { // xml:id for RDF metadata GetExport().AddAttributeXmlId(rTextContent); GetExport().AddAttributesRDFa(rTextContent); OUString sStyle; if( rPropSetHelper.hasProperty( PARA_STYLE_NAME ) ) { if( xMultiPropSet.is() ) rPropSetHelper.getValue( PARA_STYLE_NAME, xMultiPropSet ) >>= sStyle; else rPropSetHelper.getValue( PARA_STYLE_NAME, xPropSet ) >>= sStyle; } Reference< XInterface > xRef( rTextContent, UNO_QUERY ); if( xRef.is() ) { const OUString& rIdentifier = GetExport().getInterfaceToIdentifierMapper().getIdentifier( xRef ); if( rIdentifier.getLength() ) { // FIXME: this is just temporary until EditEngine // paragraphs implement XMetadatable. // then that must be used and not the mapper, because // when both can be used we get two xml:id! uno::Reference<rdf::XMetadatable> const xMeta(xRef, uno::UNO_QUERY); OSL_ENSURE(!xMeta.is(), "paragraph that implements " "XMetadatable used in interfaceToIdentifierMapper?"); GetExport().AddAttributeIdLegacy(XML_NAMESPACE_TEXT, rIdentifier); } } OUString sAutoStyle( sStyle ); sAutoStyle = Find( XML_STYLE_FAMILY_TEXT_PARAGRAPH, xPropSet, sStyle ); if( sAutoStyle.getLength() ) GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_STYLE_NAME, GetExport().EncodeStyleName( sAutoStyle ) ); if( rPropSetHelper.hasProperty( PARA_CONDITIONAL_STYLE_NAME ) ) { OUString sCondStyle; if( xMultiPropSet.is() ) rPropSetHelper.getValue( PARA_CONDITIONAL_STYLE_NAME, xMultiPropSet ) >>= sCondStyle; else rPropSetHelper.getValue( PARA_CONDITIONAL_STYLE_NAME, xPropSet ) >>= sCondStyle; if( sCondStyle != sStyle ) { sCondStyle = Find( XML_STYLE_FAMILY_TEXT_PARAGRAPH, xPropSet, sCondStyle ); if( sCondStyle.getLength() ) GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_COND_STYLE_NAME, GetExport().EncodeStyleName( sCondStyle ) ); } } if( rPropSetHelper.hasProperty( PARA_OUTLINE_LEVEL ) ) { if( xMultiPropSet.is() ) rPropSetHelper.getValue( PARA_OUTLINE_LEVEL, xMultiPropSet ) >>= nOutlineLevel; else rPropSetHelper.getValue( PARA_OUTLINE_LEVEL, xPropSet ) >>= nOutlineLevel; if( 0 < nOutlineLevel ) { OUStringBuffer sTmp; sTmp.append( sal_Int32( nOutlineLevel) ); GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_OUTLINE_LEVEL, sTmp.makeStringAndClear() ); if( rPropSetHelper.hasProperty( NUMBERING_IS_NUMBER ) ) { bool bIsNumber = false; if( xMultiPropSet.is() ) rPropSetHelper.getValue( NUMBERING_IS_NUMBER, xMultiPropSet ) >>= bIsNumber; else rPropSetHelper.getValue( NUMBERING_IS_NUMBER, xPropSet ) >>= bIsNumber; OUString sListStyleName; if( xMultiPropSet.is() ) rPropSetHelper.getValue( PARA_NUMBERING_STYLENAME, xMultiPropSet ) >>= sListStyleName; else rPropSetHelper.getValue( PARA_NUMBERING_STYLENAME, xPropSet ) >>= sListStyleName; bool bAssignedtoOutlineStyle = false; { Reference< XChapterNumberingSupplier > xCNSupplier( GetExport().GetModel(), UNO_QUERY ); OUString sOutlineName; if (xCNSupplier.is()) { Reference< XIndexReplace > xNumRule ( xCNSupplier->getChapterNumberingRules() ); DBG_ASSERT( xNumRule.is(), "no chapter numbering rules" ); if (xNumRule.is()) { Reference< XPropertySet > xNumRulePropSet( xNumRule, UNO_QUERY ); xNumRulePropSet->getPropertyValue( OUString(RTL_CONSTASCII_USTRINGPARAM("Name")) ) >>= sOutlineName; bAssignedtoOutlineStyle = ( sListStyleName == sOutlineName ); } } } if( ! bIsNumber && bAssignedtoOutlineStyle ) GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_IS_LIST_HEADER, XML_TRUE ); } { String sParaIsNumberingRestart (RTL_CONSTASCII_USTRINGPARAM ("ParaIsNumberingRestart")); bool bIsRestartNumbering = false; Reference< XPropertySetInfo > xPropSetInfo(xMultiPropSet.is() ? xMultiPropSet->getPropertySetInfo(): xPropSet->getPropertySetInfo()); if (xPropSetInfo-> hasPropertyByName(sParaIsNumberingRestart)) { xPropSet->getPropertyValue(sParaIsNumberingRestart) >>= bIsRestartNumbering; } if (bIsRestartNumbering) { GetExport().AddAttribute(XML_NAMESPACE_TEXT, XML_RESTART_NUMBERING, XML_TRUE); String sNumberingStartValue (RTL_CONSTASCII_USTRINGPARAM ("NumberingStartValue")); sal_Int32 nStartValue = 0; if (xPropSetInfo-> hasPropertyByName(sNumberingStartValue)) { xPropSet->getPropertyValue(sNumberingStartValue) >>= nStartValue; OUStringBuffer sTmpStartValue; sTmpStartValue.append(nStartValue); GetExport(). AddAttribute(XML_NAMESPACE_TEXT, XML_START_VALUE, sTmpStartValue. makeStringAndClear()); } } } } } } } Reference < XEnumerationAccess > xEA( rTextContent, UNO_QUERY ); Reference < XEnumeration > xTextEnum; xTextEnum = xEA->createEnumeration(); const sal_Bool bHasPortions = xTextEnum.is(); Reference < XEnumeration> xContentEnum; Reference < XContentEnumerationAccess > xCEA( rTextContent, UNO_QUERY ); if( xCEA.is() ) xContentEnum.set(xCEA->createContentEnumeration( sTextContentService )); const sal_Bool bHasContentEnum = xContentEnum.is() && xContentEnum->hasMoreElements(); Reference < XTextSection > xSection; if( bHasContentEnum ) { // For the auto styles, the multi property set helper is only used // if hard attributes are existing. Therfor, it seems to be a better // strategy to have the TextSection property seperate, because otherwise // we always retrieve the style names even if they are not required. if( bAutoStyles ) { if( xPropSet->getPropertySetInfo()->hasPropertyByName( sTextSection ) ) { xSection.set(xPropSet->getPropertyValue( sTextSection ), uno::UNO_QUERY); } } else { if( rPropSetHelper.hasProperty( TEXT_SECTION ) ) { xSection.set(rPropSetHelper.getValue( TEXT_SECTION ), uno::UNO_QUERY); } } } if( bAutoStyles ) { sal_Bool bPrevCharIsSpace = sal_True; if( bHasContentEnum ) bPrevCharIsSpace = !exportTextContentEnumeration( xContentEnum, bAutoStyles, xSection, bIsProgress, sal_True, 0, sal_True ); if ( bHasPortions ) exportTextRangeEnumeration( xTextEnum, bAutoStyles, bIsProgress ); } else { sal_Bool bPrevCharIsSpace = sal_True; enum XMLTokenEnum eElem = 0 < nOutlineLevel ? XML_H : XML_P; SvXMLElementExport aElem( GetExport(), XML_NAMESPACE_TEXT, eElem, sal_True, sal_False ); if( bHasContentEnum ) bPrevCharIsSpace = !exportTextContentEnumeration( xContentEnum, bAutoStyles, xSection, bIsProgress ); exportTextRangeEnumeration( xTextEnum, bAutoStyles, bIsProgress, bPrevCharIsSpace ); } } void XMLTextParagraphExport::exportTextRangeEnumeration( const Reference < XEnumeration > & rTextEnum, sal_Bool bAutoStyles, sal_Bool bIsProgress, sal_Bool bPrvChrIsSpc ) { static OUString sMeta(RTL_CONSTASCII_USTRINGPARAM("InContentMetadata")); sal_Bool bPrevCharIsSpace = bPrvChrIsSpc; while( rTextEnum->hasMoreElements() ) { Reference<XPropertySet> xPropSet(rTextEnum->nextElement(), UNO_QUERY); Reference < XTextRange > xTxtRange(xPropSet, uno::UNO_QUERY); Reference<XPropertySetInfo> xPropInfo(xPropSet->getPropertySetInfo()); if (xPropInfo->hasPropertyByName(sTextPortionType)) { rtl::OUString sType; xPropSet->getPropertyValue(sTextPortionType) >>= sType; if( sType.equals(sText)) { exportTextRange( xTxtRange, bAutoStyles, bPrevCharIsSpace ); } else if( sType.equals(sTextField)) { exportTextField( xTxtRange, bAutoStyles, bIsProgress ); bPrevCharIsSpace = sal_False; } else if( sType.equals( sFrame ) ) { Reference < XEnumeration> xContentEnum; Reference < XContentEnumerationAccess > xCEA( xTxtRange, UNO_QUERY ); if( xCEA.is() ) xContentEnum.set(xCEA->createContentEnumeration( sTextContentService )); // frames are never in sections Reference<XTextSection> xSection; if( xContentEnum.is() ) exportTextContentEnumeration( xContentEnum, bAutoStyles, xSection, bIsProgress, sal_True, &xPropSet ); bPrevCharIsSpace = sal_False; } else if (sType.equals(sFootnote)) { exportTextFootnote(xPropSet, xTxtRange->getString(), bAutoStyles, bIsProgress ); bPrevCharIsSpace = sal_False; } else if (sType.equals(sBookmark)) { exportTextMark(xPropSet, sBookmark, lcl_XmlBookmarkElements, bAutoStyles); } else if (sType.equals(sReferenceMark)) { exportTextMark(xPropSet, sReferenceMark, lcl_XmlReferenceElements, bAutoStyles); } else if (sType.equals(sDocumentIndexMark)) { pIndexMarkExport->ExportIndexMark(xPropSet, bAutoStyles); } else if (sType.equals(sRedline)) { if (NULL != pRedlineExport) pRedlineExport->ExportChange(xPropSet, bAutoStyles); } else if (sType.equals(sRuby)) { exportRuby(xPropSet, bAutoStyles); } else if (sType.equals(sMeta)) { exportMeta(xPropSet, bAutoStyles, bIsProgress); } else if (sType.equals(sTextFieldStart)) { if ( GetExport().getDefaultVersion() == SvtSaveOptions::ODFVER_LATEST ) { Reference<XNamed> xBookmark(xPropSet->getPropertyValue(sBookmark), UNO_QUERY); if (xBookmark.is()) { GetExport().AddAttribute(XML_NAMESPACE_TEXT, XML_NAME, xBookmark->getName()); } Reference< ::com::sun::star::text::XFormField > xFormField(xPropSet->getPropertyValue(sBookmark), UNO_QUERY); if (xFormField.is()) { GetExport().AddAttribute(XML_NAMESPACE_FIELD, XML_TYPE, xFormField->getFieldType()); } GetExport().StartElement(XML_NAMESPACE_FIELD, XML_FIELDMARK_START, sal_False); if (xFormField.is()) { FieldParamExporter(&GetExport(), xFormField->getParameters()).Export(); } GetExport().EndElement(XML_NAMESPACE_FIELD, XML_FIELDMARK_START, sal_False); } } else if (sType.equals(sTextFieldEnd)) { if ( GetExport().getDefaultVersion() == SvtSaveOptions::ODFVER_LATEST ) { GetExport().StartElement(XML_NAMESPACE_FIELD, XML_FIELDMARK_END, sal_False); GetExport().EndElement(XML_NAMESPACE_FIELD, XML_FIELDMARK_END, sal_False); } } else if (sType.equals(sTextFieldStartEnd)) { if ( GetExport().getDefaultVersion() == SvtSaveOptions::ODFVER_LATEST ) { Reference<XNamed> xBookmark(xPropSet->getPropertyValue(sBookmark), UNO_QUERY); if (xBookmark.is()) { GetExport().AddAttribute(XML_NAMESPACE_TEXT, XML_NAME, xBookmark->getName()); } Reference< ::com::sun::star::text::XFormField > xFormField(xPropSet->getPropertyValue(sBookmark), UNO_QUERY); if (xFormField.is()) { GetExport().AddAttribute(XML_NAMESPACE_FIELD, XML_TYPE, xFormField->getFieldType()); } GetExport().StartElement(XML_NAMESPACE_FIELD, XML_FIELDMARK, sal_False); if (xFormField.is()) { FieldParamExporter(&GetExport(), xFormField->getParameters()).Export(); } GetExport().EndElement(XML_NAMESPACE_FIELD, XML_FIELDMARK, sal_False); } } else if (sType.equals(sSoftPageBreak)) { exportSoftPageBreak(xPropSet, bAutoStyles); } else { DBG_ERROR("unknown text portion type"); } } else { Reference<XServiceInfo> xServiceInfo( xTxtRange, UNO_QUERY ); if( xServiceInfo->supportsService( sTextFieldService ) ) { exportTextField( xTxtRange, bAutoStyles, bIsProgress ); bPrevCharIsSpace = sal_False; } else { // no TextPortionType property -> non-Writer app -> text exportTextRange( xTxtRange, bAutoStyles, bPrevCharIsSpace ); } } } // now that there are nested enumerations for meta(-field), this may be valid! // DBG_ASSERT( !bOpenRuby, "Red Alert: Ruby still open!" ); } void XMLTextParagraphExport::exportTable( const Reference < XTextContent > &, sal_Bool /*bAutoStyles*/, sal_Bool /*bIsProgress*/ ) { } void XMLTextParagraphExport::exportTextField( const Reference < XTextRange > & rTextRange, sal_Bool bAutoStyles, sal_Bool bIsProgress ) { Reference < XPropertySet > xPropSet( rTextRange, UNO_QUERY ); // non-Writer apps need not support Property TextField, so test first if (xPropSet->getPropertySetInfo()->hasPropertyByName( sTextField )) { Reference < XTextField > xTxtFld(xPropSet->getPropertyValue( sTextField ), uno::UNO_QUERY); DBG_ASSERT( xTxtFld.is(), "text field missing" ); if( xTxtFld.is() ) { exportTextField(xTxtFld, bAutoStyles, bIsProgress, sal_True); } else { // write only characters GetExport().Characters(rTextRange->getString()); } } } void XMLTextParagraphExport::exportTextField( const Reference < XTextField > & xTextField, const sal_Bool bAutoStyles, const sal_Bool bIsProgress, const sal_Bool bRecursive ) { if ( bAutoStyles ) { pFieldExport->ExportFieldAutoStyle( xTextField, bIsProgress, bRecursive ); } else { pFieldExport->ExportField( xTextField, bIsProgress ); } } void XMLTextParagraphExport::exportSoftPageBreak( const Reference<XPropertySet> & , sal_Bool ) { SvXMLElementExport aElem( GetExport(), XML_NAMESPACE_TEXT, XML_SOFT_PAGE_BREAK, sal_False, sal_False ); } void XMLTextParagraphExport::exportTextMark( const Reference<XPropertySet> & rPropSet, const OUString sProperty, const enum XMLTokenEnum pElements[], sal_Bool bAutoStyles) { // mib said: "Hau wech!" // // (Originally, I'd export a span element in case the (book|reference)mark // was formatted. This actually makes a difference in case some pervert // sets a point reference mark in the document and, say, formats it bold. // This basically meaningless formatting will now been thrown away // (aka cleaned up), since mib said: ... dvo if (!bAutoStyles) { // name element Reference<XNamed> xName(rPropSet->getPropertyValue(sProperty), UNO_QUERY); GetExport().AddAttribute(XML_NAMESPACE_TEXT, XML_NAME, xName->getName()); // start, end, or point-reference? sal_Int8 nElement; if( *(sal_Bool *)rPropSet->getPropertyValue(sIsCollapsed).getValue() ) { nElement = 0; } else { nElement = *(sal_Bool *)rPropSet->getPropertyValue(sIsStart).getValue() ? 1 : 2; } // bookmark, bookmark-start: xml:id and RDFa for RDF metadata if( nElement < 2 ) { GetExport().AddAttributeXmlId(xName); const uno::Reference<text::XTextContent> xTextContent( xName, uno::UNO_QUERY_THROW); GetExport().AddAttributesRDFa(xTextContent); } // export element DBG_ASSERT(pElements != NULL, "illegal element array"); DBG_ASSERT(nElement >= 0, "illegal element number"); DBG_ASSERT(nElement <= 2, "illegal element number"); SvXMLElementExport aElem(GetExport(), XML_NAMESPACE_TEXT, pElements[nElement], sal_False, sal_False); } // else: no styles. (see above) } sal_Bool lcl_txtpara_isBoundAsChar( const Reference < XPropertySet > & rPropSet, const Reference < XPropertySetInfo > & rPropSetInfo ) { sal_Bool bIsBoundAsChar = sal_False; OUString sAnchorType( RTL_CONSTASCII_USTRINGPARAM( "AnchorType" ) ); if( rPropSetInfo->hasPropertyByName( sAnchorType ) ) { TextContentAnchorType eAnchor; rPropSet->getPropertyValue( sAnchorType ) >>= eAnchor; bIsBoundAsChar = TextContentAnchorType_AS_CHARACTER == eAnchor; } return bIsBoundAsChar; } sal_Int32 XMLTextParagraphExport::addTextFrameAttributes( const Reference < XPropertySet >& rPropSet, sal_Bool bShape, OUString *pMinHeightValue ) { sal_Int32 nShapeFeatures = SEF_DEFAULT; // draw:name (#97662#: not for shapes, since those names will be // treated in the shape export) if( !bShape ) { Reference < XNamed > xNamed( rPropSet, UNO_QUERY ); if( xNamed.is() ) { OUString sName( xNamed->getName() ); if( sName.getLength() ) GetExport().AddAttribute( XML_NAMESPACE_DRAW, XML_NAME, xNamed->getName() ); } } OUStringBuffer sValue; // text:anchor-type TextContentAnchorType eAnchor = TextContentAnchorType_AT_PARAGRAPH; rPropSet->getPropertyValue( sAnchorType ) >>= eAnchor; { XMLAnchorTypePropHdl aAnchorTypeHdl; OUString sTmp; aAnchorTypeHdl.exportXML( sTmp, uno::makeAny(eAnchor), GetExport().GetMM100UnitConverter() ); GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_ANCHOR_TYPE, sTmp ); } // text:anchor-page-number if( TextContentAnchorType_AT_PAGE == eAnchor ) { sal_Int16 nPage = 0; rPropSet->getPropertyValue( sAnchorPageNo ) >>= nPage; GetExport().GetMM100UnitConverter().convertNumber( sValue, (sal_Int32)nPage ); GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_ANCHOR_PAGE_NUMBER, sValue.makeStringAndClear() ); } else { // #92210# nShapeFeatures |= SEF_EXPORT_NO_WS; } // OD 2004-06-01 #i27691# - correction: no export of svg:x, if object // is anchored as-character. if ( !bShape && eAnchor != TextContentAnchorType_AS_CHARACTER ) { // svg:x sal_Int16 nHoriOrient = HoriOrientation::NONE; rPropSet->getPropertyValue( sHoriOrient ) >>= nHoriOrient; if( HoriOrientation::NONE == nHoriOrient ) { sal_Int32 nPos = 0; rPropSet->getPropertyValue( sHoriOrientPosition ) >>= nPos; GetExport().GetMM100UnitConverter().convertMeasure( sValue, nPos ); GetExport().AddAttribute( XML_NAMESPACE_SVG, XML_X, sValue.makeStringAndClear() ); } } else if( TextContentAnchorType_AS_CHARACTER == eAnchor ) nShapeFeatures = (nShapeFeatures & ~SEF_EXPORT_X); if( !bShape || TextContentAnchorType_AS_CHARACTER == eAnchor ) { // svg:y sal_Int16 nVertOrient = VertOrientation::NONE; rPropSet->getPropertyValue( sVertOrient ) >>= nVertOrient; if( VertOrientation::NONE == nVertOrient ) { sal_Int32 nPos = 0; rPropSet->getPropertyValue( sVertOrientPosition ) >>= nPos; GetExport().GetMM100UnitConverter().convertMeasure( sValue, nPos ); GetExport().AddAttribute( XML_NAMESPACE_SVG, XML_Y, sValue.makeStringAndClear() ); } if( bShape ) nShapeFeatures = (nShapeFeatures & ~SEF_EXPORT_Y); } Reference< XPropertySetInfo > xPropSetInfo(rPropSet->getPropertySetInfo()); // svg:width sal_Int16 nWidthType = SizeType::FIX; if( xPropSetInfo->hasPropertyByName( sWidthType ) ) { rPropSet->getPropertyValue( sWidthType ) >>= nWidthType; } if( xPropSetInfo->hasPropertyByName( sWidth ) ) { sal_Int32 nWidth = 0; // VAR size will be written as zero min-size if( SizeType::VARIABLE != nWidthType ) { rPropSet->getPropertyValue( sWidth ) >>= nWidth; } GetExport().GetMM100UnitConverter().convertMeasure( sValue, nWidth ); if( SizeType::FIX != nWidthType ) GetExport().AddAttribute( XML_NAMESPACE_FO, XML_MIN_WIDTH, sValue.makeStringAndClear() ); else GetExport().AddAttribute( XML_NAMESPACE_SVG, XML_WIDTH, sValue.makeStringAndClear() ); } sal_Bool bSyncWidth = sal_False; if( xPropSetInfo->hasPropertyByName( sIsSyncWidthToHeight ) ) { bSyncWidth = *(sal_Bool *)rPropSet->getPropertyValue( sIsSyncWidthToHeight ).getValue(); if( bSyncWidth ) GetExport().AddAttribute( XML_NAMESPACE_STYLE, XML_REL_WIDTH, XML_SCALE ); } if( !bSyncWidth && xPropSetInfo->hasPropertyByName( sRelativeWidth ) ) { sal_Int16 nRelWidth = 0; rPropSet->getPropertyValue( sRelativeWidth ) >>= nRelWidth; DBG_ASSERT( nRelWidth >= 0 && nRelWidth <= 254, "Got illegal relative width from API" ); if( nRelWidth > 0 ) { GetExport().GetMM100UnitConverter().convertPercent( sValue, nRelWidth ); GetExport().AddAttribute( XML_NAMESPACE_STYLE, XML_REL_WIDTH, sValue.makeStringAndClear() ); } } // svg:height, fo:min-height or style:rel-height sal_Int16 nSizeType = SizeType::FIX; if( xPropSetInfo->hasPropertyByName( sSizeType ) ) { rPropSet->getPropertyValue( sSizeType ) >>= nSizeType; } sal_Bool bSyncHeight = sal_False; if( xPropSetInfo->hasPropertyByName( sIsSyncHeightToWidth ) ) { bSyncHeight = *(sal_Bool *)rPropSet->getPropertyValue( sIsSyncHeightToWidth ).getValue(); } sal_Int16 nRelHeight = 0; if( !bSyncHeight && xPropSetInfo->hasPropertyByName( sRelativeHeight ) ) { rPropSet->getPropertyValue( sRelativeHeight ) >>= nRelHeight; } if( xPropSetInfo->hasPropertyByName( sHeight ) ) { sal_Int32 nHeight = 0; if( SizeType::VARIABLE != nSizeType ) { rPropSet->getPropertyValue( sHeight ) >>= nHeight; } GetExport().GetMM100UnitConverter().convertMeasure( sValue, nHeight ); if( SizeType::FIX != nSizeType && 0==nRelHeight && !bSyncHeight && pMinHeightValue ) *pMinHeightValue = sValue.makeStringAndClear(); else GetExport().AddAttribute( XML_NAMESPACE_SVG, XML_HEIGHT, sValue.makeStringAndClear() ); } if( bSyncHeight ) { GetExport().AddAttribute( XML_NAMESPACE_STYLE, XML_REL_HEIGHT, SizeType::MIN == nSizeType ? XML_SCALE_MIN : XML_SCALE ); } else if( nRelHeight > 0 ) { GetExport().GetMM100UnitConverter().convertPercent( sValue, nRelHeight ); if( SizeType::MIN == nSizeType ) GetExport().AddAttribute( XML_NAMESPACE_FO, XML_MIN_HEIGHT, sValue.makeStringAndClear() ); else GetExport().AddAttribute( XML_NAMESPACE_STYLE, XML_REL_HEIGHT, sValue.makeStringAndClear() ); } OUString sZOrder( RTL_CONSTASCII_USTRINGPARAM( "ZOrder" ) ); if( xPropSetInfo->hasPropertyByName( sZOrder ) ) { sal_Int32 nZIndex = 0; rPropSet->getPropertyValue( sZOrder ) >>= nZIndex; if( -1 != nZIndex ) { GetExport().GetMM100UnitConverter().convertNumber( sValue, nZIndex ); GetExport().AddAttribute( XML_NAMESPACE_DRAW, XML_ZINDEX, sValue.makeStringAndClear() ); } } return nShapeFeatures; } void XMLTextParagraphExport::exportAnyTextFrame( const Reference < XTextContent > & rTxtCntnt, FrameType eType, sal_Bool bAutoStyles, sal_Bool bIsProgress, sal_Bool bExportContent, const Reference < XPropertySet > *pRangePropSet) { Reference < XPropertySet > xPropSet( rTxtCntnt, UNO_QUERY ); if( bAutoStyles ) { if( FT_EMBEDDED == eType ) _collectTextEmbeddedAutoStyles( xPropSet ); // No text frame style for shapes (#i28745#) else if ( FT_SHAPE != eType ) Add( XML_STYLE_FAMILY_TEXT_FRAME, xPropSet ); if( pRangePropSet && lcl_txtpara_isBoundAsChar( xPropSet, xPropSet->getPropertySetInfo() ) ) Add( XML_STYLE_FAMILY_TEXT_TEXT, *pRangePropSet ); switch( eType ) { case FT_TEXT: { // frame bound frames if ( bExportContent ) { Reference < XTextFrame > xTxtFrame( rTxtCntnt, UNO_QUERY ); Reference < XText > xTxt(xTxtFrame->getText()); exportFrameFrames( sal_True, bIsProgress, &xTxtFrame ); exportText( xTxt, bAutoStyles, bIsProgress, sal_True ); } } break; case FT_SHAPE: { Reference < XShape > xShape( rTxtCntnt, UNO_QUERY ); GetExport().GetShapeExport()->collectShapeAutoStyles( xShape ); } break; default: break; } } else { Reference< XPropertySetInfo > xPropSetInfo(xPropSet->getPropertySetInfo()); Reference< XPropertyState > xPropState( xPropSet, UNO_QUERY ); { sal_Bool bAddCharStyles = pRangePropSet && lcl_txtpara_isBoundAsChar( xPropSet, xPropSetInfo ); sal_Bool bIsUICharStyle; sal_Bool bHasAutoStyle = sal_False; sal_Bool bDummy; OUString sStyle; if( bAddCharStyles ) sStyle = FindTextStyleAndHyperlink( *pRangePropSet, bDummy, bIsUICharStyle, bHasAutoStyle ); else bIsUICharStyle = sal_False; XMLTextCharStyleNamesElementExport aCharStylesExport( GetExport(), bIsUICharStyle && aCharStyleNamesPropInfoCache.hasProperty( *pRangePropSet ), bHasAutoStyle, *pRangePropSet, sCharStyleNames ); if( sStyle.getLength() ) GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_STYLE_NAME, GetExport().EncodeStyleName( sStyle ) ); { SvXMLElementExport aElem( GetExport(), sStyle.getLength() > 0, XML_NAMESPACE_TEXT, XML_SPAN, sal_False, sal_False ); { SvXMLElementExport aElement( GetExport(), FT_SHAPE != eType && addHyperlinkAttributes( xPropSet, xPropState,xPropSetInfo ), XML_NAMESPACE_DRAW, XML_A, sal_False, sal_False ); switch( eType ) { case FT_TEXT: _exportTextFrame( xPropSet, xPropSetInfo, bIsProgress ); break; case FT_GRAPHIC: _exportTextGraphic( xPropSet, xPropSetInfo ); break; case FT_EMBEDDED: _exportTextEmbedded( xPropSet, xPropSetInfo ); break; case FT_SHAPE: { Reference < XShape > xShape( rTxtCntnt, UNO_QUERY ); sal_Int32 nFeatures = addTextFrameAttributes( xPropSet, sal_True ); GetExport().GetShapeExport() ->exportShape( xShape, nFeatures ); } break; } } } } } } void XMLTextParagraphExport::_exportTextFrame( const Reference < XPropertySet > & rPropSet, const Reference < XPropertySetInfo > & rPropSetInfo, sal_Bool bIsProgress ) { Reference < XTextFrame > xTxtFrame( rPropSet, UNO_QUERY ); Reference < XText > xTxt(xTxtFrame->getText()); OUString sStyle; if( rPropSetInfo->hasPropertyByName( sFrameStyleName ) ) { rPropSet->getPropertyValue( sFrameStyleName ) >>= sStyle; } OUString sAutoStyle( sStyle ); OUString aMinHeightValue; sAutoStyle = Find( XML_STYLE_FAMILY_TEXT_FRAME, rPropSet, sStyle ); if( sAutoStyle.getLength() ) GetExport().AddAttribute( XML_NAMESPACE_DRAW, XML_STYLE_NAME, GetExport().EncodeStyleName( sAutoStyle ) ); addTextFrameAttributes( rPropSet, sal_False, &aMinHeightValue ); SvXMLElementExport aElem( GetExport(), XML_NAMESPACE_DRAW, XML_FRAME, sal_False, sal_True ); if( aMinHeightValue.getLength() ) GetExport().AddAttribute( XML_NAMESPACE_FO, XML_MIN_HEIGHT, aMinHeightValue ); // draw:chain-next-name if( rPropSetInfo->hasPropertyByName( sChainNextName ) ) { OUString sNext; if( (rPropSet->getPropertyValue( sChainNextName ) >>= sNext) && sNext.getLength() > 0 ) GetExport().AddAttribute( XML_NAMESPACE_DRAW, XML_CHAIN_NEXT_NAME, sNext ); } { SvXMLElementExport aElement( GetExport(), XML_NAMESPACE_DRAW, XML_TEXT_BOX, sal_True, sal_True ); // frame bound frames exportFramesBoundToFrame( xTxtFrame, bIsProgress ); exportText( xTxt, sal_False, bIsProgress, sal_True ); } // script:events Reference<XEventsSupplier> xEventsSupp( xTxtFrame, UNO_QUERY ); GetExport().GetEventExport().Export(xEventsSupp); // image map GetExport().GetImageMapExport().Export( rPropSet ); // svg:title and svg:desc (#i73249#) exportTitleAndDescription( rPropSet, rPropSetInfo ); } void XMLTextParagraphExport::exportContour( const Reference < XPropertySet > & rPropSet, const Reference < XPropertySetInfo > & rPropSetInfo ) { if( !rPropSetInfo->hasPropertyByName( sContourPolyPolygon ) ) return; PointSequenceSequence aSourcePolyPolygon; rPropSet->getPropertyValue( sContourPolyPolygon ) >>= aSourcePolyPolygon; if( !aSourcePolyPolygon.getLength() ) return; awt::Point aPoint( 0, 0 ); awt::Size aSize( 0, 0 ); sal_Int32 nPolygons = aSourcePolyPolygon.getLength(); const PointSequence *pPolygons = aSourcePolyPolygon.getConstArray(); while( nPolygons-- ) { sal_Int32 nPoints = pPolygons->getLength(); const awt::Point *pPoints = pPolygons->getConstArray(); while( nPoints-- ) { if( aSize.Width < pPoints->X ) aSize.Width = pPoints->X; if( aSize.Height < pPoints->Y ) aSize.Height = pPoints->Y; pPoints++; } pPolygons++; } sal_Bool bPixel = sal_False; if( rPropSetInfo->hasPropertyByName( sIsPixelContour ) ) { bPixel = *(sal_Bool *)rPropSet->getPropertyValue( sIsPixelContour ).getValue(); } // svg: width OUStringBuffer aStringBuffer( 10 ); if( bPixel ) GetExport().GetMM100UnitConverter().convertMeasurePx(aStringBuffer, aSize.Width); else GetExport().GetMM100UnitConverter().convertMeasure(aStringBuffer, aSize.Width); GetExport().AddAttribute( XML_NAMESPACE_SVG, XML_WIDTH, aStringBuffer.makeStringAndClear() ); // svg: height if( bPixel ) GetExport().GetMM100UnitConverter().convertMeasurePx(aStringBuffer, aSize.Height); else GetExport().GetMM100UnitConverter().convertMeasure(aStringBuffer, aSize.Height); GetExport().AddAttribute( XML_NAMESPACE_SVG, XML_HEIGHT, aStringBuffer.makeStringAndClear() ); // svg:viewbox SdXMLImExViewBox aViewBox(0, 0, aSize.Width, aSize.Height); GetExport().AddAttribute(XML_NAMESPACE_SVG, XML_VIEWBOX, aViewBox.GetExportString()); sal_Int32 nOuterCnt( aSourcePolyPolygon.getLength() ); enum XMLTokenEnum eElem = XML_TOKEN_INVALID; if( 1L == nOuterCnt ) { // simple polygon shape, can be written as svg:points sequence /*const*/ PointSequence* pSequence = (PointSequence*)aSourcePolyPolygon.getConstArray(); SdXMLImExPointsElement aPoints( pSequence, aViewBox, aPoint, aSize ); // write point array GetExport().AddAttribute( XML_NAMESPACE_DRAW, XML_POINTS, aPoints.GetExportString()); eElem = XML_CONTOUR_POLYGON; } else { // polypolygon, needs to be written as a svg:path sequence /*const*/ PointSequence* pOuterSequence = (PointSequence*)aSourcePolyPolygon.getConstArray(); if(pOuterSequence) { // prepare svx:d element export SdXMLImExSvgDElement aSvgDElement( aViewBox ); for(sal_Int32 a(0L); a < nOuterCnt; a++) { /*const*/ PointSequence* pSequence = pOuterSequence++; if(pSequence) { aSvgDElement.AddPolygon(pSequence, 0L, aPoint, aSize, sal_True ); } } // write point array GetExport().AddAttribute( XML_NAMESPACE_SVG, XML_D, aSvgDElement.GetExportString()); eElem = XML_CONTOUR_PATH; } } if( rPropSetInfo->hasPropertyByName( sIsAutomaticContour ) ) { sal_Bool bTmp = *(sal_Bool *)rPropSet->getPropertyValue( sIsAutomaticContour ).getValue(); GetExport().AddAttribute( XML_NAMESPACE_DRAW, XML_RECREATE_ON_EDIT, bTmp ? XML_TRUE : XML_FALSE ); } // write object now SvXMLElementExport aElem( GetExport(), XML_NAMESPACE_DRAW, eElem, sal_True, sal_True ); } void XMLTextParagraphExport::_exportTextGraphic( const Reference < XPropertySet > & rPropSet, const Reference < XPropertySetInfo > & rPropSetInfo ) { OUString sStyle; if( rPropSetInfo->hasPropertyByName( sFrameStyleName ) ) { rPropSet->getPropertyValue( sFrameStyleName ) >>= sStyle; } OUString sAutoStyle( sStyle ); sAutoStyle = Find( XML_STYLE_FAMILY_TEXT_FRAME, rPropSet, sStyle ); if( sAutoStyle.getLength() ) GetExport().AddAttribute( XML_NAMESPACE_DRAW, XML_STYLE_NAME, GetExport().EncodeStyleName( sAutoStyle ) ); addTextFrameAttributes( rPropSet, sal_False ); // svg:transform sal_Int16 nVal = 0; rPropSet->getPropertyValue( sGraphicRotation ) >>= nVal; if( nVal != 0 ) { OUStringBuffer sRet( GetXMLToken(XML_ROTATE).getLength()+4 ); sRet.append( GetXMLToken(XML_ROTATE)); sRet.append( (sal_Unicode)'(' ); GetExport().GetMM100UnitConverter().convertNumber( sRet, (sal_Int32)nVal ); sRet.append( (sal_Unicode)')' ); GetExport().AddAttribute( XML_NAMESPACE_SVG, XML_TRANSFORM, sRet.makeStringAndClear() ); } SvXMLElementExport aElem( GetExport(), XML_NAMESPACE_DRAW, XML_FRAME, sal_False, sal_True ); // xlink:href OUString sOrigURL; rPropSet->getPropertyValue( sGraphicURL ) >>= sOrigURL; OUString sURL(GetExport().AddEmbeddedGraphicObject( sOrigURL )); setTextEmbeddedGraphicURL( rPropSet, sURL ); // If there still is no url, then then graphic is empty if( sURL.getLength() ) { GetExport().AddAttribute(XML_NAMESPACE_XLINK, XML_HREF, sURL ); GetExport().AddAttribute( XML_NAMESPACE_XLINK, XML_TYPE, XML_SIMPLE ); GetExport().AddAttribute( XML_NAMESPACE_XLINK, XML_SHOW, XML_EMBED ); GetExport().AddAttribute( XML_NAMESPACE_XLINK, XML_ACTUATE, XML_ONLOAD ); } // draw:filter-name OUString sGrfFilter; rPropSet->getPropertyValue( sGraphicFilter ) >>= sGrfFilter; if( sGrfFilter.getLength() ) GetExport().AddAttribute( XML_NAMESPACE_DRAW, XML_FILTER_NAME, sGrfFilter ); { SvXMLElementExport aElement( GetExport(), XML_NAMESPACE_DRAW, XML_IMAGE, sal_False, sal_True ); // optional office:binary-data GetExport().AddEmbeddedGraphicObjectAsBase64( sOrigURL ); } // script:events Reference<XEventsSupplier> xEventsSupp( rPropSet, UNO_QUERY ); GetExport().GetEventExport().Export(xEventsSupp); // image map GetExport().GetImageMapExport().Export( rPropSet ); // svg:title and svg:desc (#i73249#) exportTitleAndDescription( rPropSet, rPropSetInfo ); // draw:contour exportContour( rPropSet, rPropSetInfo ); } void XMLTextParagraphExport::_collectTextEmbeddedAutoStyles(const Reference < XPropertySet > & ) { DBG_ASSERT( !this, "no API implementation avialable" ); } void XMLTextParagraphExport::_exportTextEmbedded( const Reference < XPropertySet > &, const Reference < XPropertySetInfo > & ) { DBG_ASSERT( !this, "no API implementation avialable" ); } void XMLTextParagraphExport::exportEvents( const Reference < XPropertySet > & rPropSet ) { // script:events Reference<XEventsSupplier> xEventsSupp( rPropSet, UNO_QUERY ); GetExport().GetEventExport().Export(xEventsSupp); // image map OUString sImageMap(RTL_CONSTASCII_USTRINGPARAM("ImageMap")); if (rPropSet->getPropertySetInfo()->hasPropertyByName(sImageMap)) GetExport().GetImageMapExport().Export( rPropSet ); } // Implement Title/Description Elements UI (#i73249#) void XMLTextParagraphExport::exportTitleAndDescription( const Reference < XPropertySet > & rPropSet, const Reference < XPropertySetInfo > & rPropSetInfo ) { // svg:title if( rPropSetInfo->hasPropertyByName( sTitle ) ) { OUString sObjTitle; rPropSet->getPropertyValue( sTitle ) >>= sObjTitle; if( sObjTitle.getLength() ) { SvXMLElementExport aElem( GetExport(), XML_NAMESPACE_SVG, XML_TITLE, sal_True, sal_False ); GetExport().Characters( sObjTitle ); } } // svg:description if( rPropSetInfo->hasPropertyByName( sDescription ) ) { OUString sObjDesc; rPropSet->getPropertyValue( sDescription ) >>= sObjDesc; if( sObjDesc.getLength() ) { SvXMLElementExport aElem( GetExport(), XML_NAMESPACE_SVG, XML_DESC, sal_True, sal_False ); GetExport().Characters( sObjDesc ); } } } void XMLTextParagraphExport::setTextEmbeddedGraphicURL( const Reference < XPropertySet >&, OUString& /*rStreamName*/ ) const { } sal_Bool XMLTextParagraphExport::addHyperlinkAttributes( const Reference < XPropertySet > & rPropSet, const Reference < XPropertyState > & rPropState, const Reference < XPropertySetInfo > & rPropSetInfo ) { sal_Bool bExport = sal_False; OUString sHRef, sName, sTargetFrame, sUStyleName, sVStyleName; sal_Bool bServerMap = sal_False; /* bool bHyperLinkURL = false; bool bHyperLinkName = false; bool bHyperLinkTarget = false; bool bServer = false; bool bUnvisitedCharStyleName = false; bool bVisitedCharStyleName = false; const Reference< XMultiPropertySet > xMultiPropertySet( rPropSet, UNO_QUERY ); if ( xMultiPropertySet.is() ) { sal_uInt32 nCount = 0; Sequence< OUString > aPropertyNames( 6 ); OUString* pArray = aPropertyNames.getArray(); if ( rPropSetInfo->hasPropertyByName( sServerMap ) ) { bServer = true; pArray[ nCount++ ] = sServerMap; } if ( rPropSetInfo->hasPropertyByName( sHyperLinkName ) ) { bHyperLinkName = true; pArray[ nCount++ ] = sHyperLinkName; } if ( rPropSetInfo->hasPropertyByName( sHyperLinkTarget ) ) { bHyperLinkTarget = true; pArray[ nCount++ ] = sHyperLinkTarget; } if ( rPropSetInfo->hasPropertyByName( sHyperLinkURL ) ) { bHyperLinkURL = true; pArray[ nCount++ ] = sHyperLinkURL; } if ( rPropSetInfo->hasPropertyByName( sUnvisitedCharStyleName ) ) { bUnvisitedCharStyleName = true; pArray[ nCount++ ] = sUnvisitedCharStyleName; } if ( rPropSetInfo->hasPropertyByName( sVisitedCharStyleName ) ) { bVisitedCharStyleName = true; pArray[ nCount++ ] = sVisitedCharStyleName; } aPropertyNames.realloc( nCount ); if ( nCount ) { Sequence< PropertyState > aPropertyStates( nCount ); PropertyState* pStateArray = aPropertyStates.getArray(); if ( rPropState.is() ) aPropertyStates = rPropState->getPropertyStates( aPropertyNames ); Sequence< Any > aPropertyValues ( xMultiPropertySet->getPropertyValues( aPropertyNames ) ); Any* pValueArray = aPropertyValues.getArray(); sal_uInt32 nIdx = 0; if ( bServer ) { if ( !rPropState.is() || PropertyState_DIRECT_VALUE == pStateArray[ nIdx ] ) { bServerMap = *(sal_Bool *)pValueArray[ nIdx ].getValue(); if( bServerMap ) bExport = sal_True; } ++nIdx; } if ( bHyperLinkName ) { if ( !rPropState.is() || PropertyState_DIRECT_VALUE == pStateArray[ nIdx ] ) { pValueArray[ nIdx ] >>= sName; if( sName.getLength() > 0 ) bExport = sal_True; } ++nIdx; } if ( bHyperLinkTarget ) { if ( !rPropState.is() || PropertyState_DIRECT_VALUE == pStateArray[ nIdx ] ) { pValueArray[ nIdx ] >>= sTargetFrame; if( sTargetFrame.getLength() ) bExport = sal_True; } ++nIdx; } if ( bHyperLinkURL ) { if ( !rPropState.is() || PropertyState_DIRECT_VALUE == pStateArray[ nIdx ] ) { pValueArray[ nIdx ] >>= sHRef; if( sHRef.getLength() > 0 ) bExport = sal_True; } ++nIdx; } if ( bUnvisitedCharStyleName ) { if ( !rPropState.is() || PropertyState_DIRECT_VALUE == pStateArray[ nIdx ] ) { pValueArray[ nIdx ] >>= sUStyleName; if( sUStyleName.getLength() ) bExport = sal_True; } ++nIdx; } if ( bVisitedCharStyleName ) { if ( !rPropState.is() || PropertyState_DIRECT_VALUE == pStateArray[ nIdx ] ) { pValueArray[ nIdx ] >>= sVStyleName; if( sVStyleName.getLength() ) bExport = sal_True; } ++nIdx; } } } else {*/ if( rPropSetInfo->hasPropertyByName( sHyperLinkURL ) && ( !rPropState.is() || PropertyState_DIRECT_VALUE == rPropState->getPropertyState( sHyperLinkURL ) ) ) { rPropSet->getPropertyValue( sHyperLinkURL ) >>= sHRef; if( sHRef.getLength() > 0 ) bExport = sal_True; } if( rPropSetInfo->hasPropertyByName( sHyperLinkName ) && ( !rPropState.is() || PropertyState_DIRECT_VALUE == rPropState->getPropertyState( sHyperLinkName ) ) ) { rPropSet->getPropertyValue( sHyperLinkName ) >>= sName; if( sName.getLength() > 0 ) bExport = sal_True; } if( rPropSetInfo->hasPropertyByName( sHyperLinkTarget ) && ( !rPropState.is() || PropertyState_DIRECT_VALUE == rPropState->getPropertyState( sHyperLinkTarget ) ) ) { rPropSet->getPropertyValue( sHyperLinkTarget ) >>= sTargetFrame; if( sTargetFrame.getLength() ) bExport = sal_True; } if( rPropSetInfo->hasPropertyByName( sServerMap ) && ( !rPropState.is() || PropertyState_DIRECT_VALUE == rPropState->getPropertyState( sServerMap ) ) ) { bServerMap = *(sal_Bool *)rPropSet->getPropertyValue( sServerMap ).getValue(); if( bServerMap ) bExport = sal_True; } if( rPropSetInfo->hasPropertyByName( sUnvisitedCharStyleName ) && ( !rPropState.is() || PropertyState_DIRECT_VALUE == rPropState->getPropertyState( sUnvisitedCharStyleName ) ) ) { rPropSet->getPropertyValue( sUnvisitedCharStyleName ) >>= sUStyleName; if( sUStyleName.getLength() ) bExport = sal_True; } if( rPropSetInfo->hasPropertyByName( sVisitedCharStyleName ) && ( !rPropState.is() || PropertyState_DIRECT_VALUE == rPropState->getPropertyState( sVisitedCharStyleName ) ) ) { rPropSet->getPropertyValue( sVisitedCharStyleName ) >>= sVStyleName; if( sVStyleName.getLength() ) bExport = sal_True; } if( bExport ) { GetExport().AddAttribute( XML_NAMESPACE_XLINK, XML_TYPE, XML_SIMPLE ); GetExport().AddAttribute( XML_NAMESPACE_XLINK, XML_HREF, GetExport().GetRelativeReference( sHRef ) ); if( sName.getLength() > 0 ) GetExport().AddAttribute( XML_NAMESPACE_OFFICE, XML_NAME, sName ); if( sTargetFrame.getLength() ) { GetExport().AddAttribute( XML_NAMESPACE_OFFICE, XML_TARGET_FRAME_NAME, sTargetFrame ); enum XMLTokenEnum eTok = sTargetFrame.equalsAsciiL( "_blank", sizeof("_blank")-1 ) ? XML_NEW : XML_REPLACE; GetExport().AddAttribute( XML_NAMESPACE_XLINK, XML_SHOW, eTok ); } if( bServerMap ) GetExport().AddAttribute( XML_NAMESPACE_OFFICE, XML_SERVER_MAP, XML_TRUE ); if( sUStyleName.getLength() ) GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_STYLE_NAME, GetExport().EncodeStyleName( sUStyleName ) ); if( sVStyleName.getLength() ) GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_VISITED_STYLE_NAME, GetExport().EncodeStyleName( sVStyleName ) ); } return bExport; } void XMLTextParagraphExport::exportTextRange( const Reference < XTextRange > & rTextRange, sal_Bool bAutoStyles, sal_Bool& rPrevCharIsSpace ) { Reference < XPropertySet > xPropSet( rTextRange, UNO_QUERY ); if( bAutoStyles ) { Add( XML_STYLE_FAMILY_TEXT_TEXT, xPropSet ); } else { sal_Bool bHyperlink = sal_False; sal_Bool bIsUICharStyle = sal_False; sal_Bool bHasAutoStyle = sal_False; OUString sStyle(FindTextStyleAndHyperlink( xPropSet, bHyperlink, bIsUICharStyle, bHasAutoStyle )); Reference < XPropertySetInfo > xPropSetInfo; if( bHyperlink ) { Reference< XPropertyState > xPropState( xPropSet, UNO_QUERY ); xPropSetInfo.set(xPropSet->getPropertySetInfo()); bHyperlink = addHyperlinkAttributes( xPropSet, xPropState, xPropSetInfo ); } SvXMLElementExport aElem( GetExport(), bHyperlink, XML_NAMESPACE_TEXT, XML_A, sal_False, sal_False ); if( bHyperlink ) { // export events (if supported) OUString sHyperLinkEvents(RTL_CONSTASCII_USTRINGPARAM( "HyperLinkEvents")); if (xPropSetInfo->hasPropertyByName(sHyperLinkEvents)) { Reference<XNameReplace> xName(xPropSet->getPropertyValue(sHyperLinkEvents), uno::UNO_QUERY); GetExport().GetEventExport().Export(xName, sal_False); } } { XMLTextCharStyleNamesElementExport aCharStylesExport( GetExport(), bIsUICharStyle && aCharStyleNamesPropInfoCache.hasProperty( xPropSet, xPropSetInfo ), bHasAutoStyle, xPropSet, sCharStyleNames ); OUString aText(rTextRange->getString()); if( sStyle.getLength() ) GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_STYLE_NAME, GetExport().EncodeStyleName( sStyle ) ); { // in a block to make sure it is destroyed before the text:a element SvXMLElementExport aElement( GetExport(), sStyle.getLength() > 0, XML_NAMESPACE_TEXT, XML_SPAN, sal_False, sal_False ); exportText( aText, rPrevCharIsSpace ); } } } } void XMLTextParagraphExport::exportText( const OUString& rText, sal_Bool& rPrevCharIsSpace ) { sal_Int32 nExpStartPos = 0L; sal_Int32 nEndPos = rText.getLength(); sal_Int32 nSpaceChars = 0; for( sal_Int32 nPos = 0; nPos < nEndPos; nPos++ ) { sal_Unicode cChar = rText[nPos]; sal_Bool bExpCharAsText = sal_True; sal_Bool bExpCharAsElement = sal_False; sal_Bool bCurrCharIsSpace = sal_False; switch( cChar ) { case 0x0009: // Tab case 0x000A: // LF // These characters are exported as text. bExpCharAsElement = sal_True; bExpCharAsText = sal_False; break; case 0x000D: break; // legal character case 0x0020: // Blank if( rPrevCharIsSpace ) { // If the previous character is a space character, // too, export a special space element. bExpCharAsText = sal_False; } bCurrCharIsSpace = sal_True; break; default: if( cChar < 0x0020 ) { #ifdef DBG_UTIL OSL_ENSURE( txtparae_bContainsIllegalCharacters || cChar >= 0x0020, "illegal character in text content" ); txtparae_bContainsIllegalCharacters = sal_True; #endif bExpCharAsText = sal_False; } break; } // If the current character is not exported as text // the text that has not been exported by now has to be exported now. if( nPos > nExpStartPos && !bExpCharAsText ) { DBG_ASSERT( 0==nSpaceChars, "pending spaces" ); OUString sExp( rText.copy( nExpStartPos, nPos - nExpStartPos ) ); GetExport().Characters( sExp ); nExpStartPos = nPos; } // If there are spaces left that have not been exported and the // current chracter is not a space , the pending spaces have to be // exported now. if( nSpaceChars > 0 && !bCurrCharIsSpace ) { DBG_ASSERT( nExpStartPos == nPos, " pending characters" ); if( nSpaceChars > 1 ) { OUStringBuffer sTmp; sTmp.append( (sal_Int32)nSpaceChars ); GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_C, sTmp.makeStringAndClear() ); } SvXMLElementExport aElem( GetExport(), XML_NAMESPACE_TEXT, XML_S, sal_False, sal_False ); nSpaceChars = 0; } // If the current character has to be exported as a special // element, the elemnt will be exported now. if( bExpCharAsElement ) { switch( cChar ) { case 0x0009: // Tab { SvXMLElementExport aElem( GetExport(), XML_NAMESPACE_TEXT, XML_TAB, sal_False, sal_False ); } break; case 0x000A: // LF { SvXMLElementExport aElem( GetExport(), XML_NAMESPACE_TEXT, XML_LINE_BREAK, sal_False, sal_False ); } break; } } // If the current character is a space, and the previous one // is a space, too, the number of pending spaces is incremented // only. if( bCurrCharIsSpace && rPrevCharIsSpace ) nSpaceChars++; rPrevCharIsSpace = bCurrCharIsSpace; // If the currect character is not exported as text, the start // position for text is the position behind the current position. if( !bExpCharAsText ) { DBG_ASSERT( nExpStartPos == nPos, "wrong export start pos" ); nExpStartPos = nPos+1; } } if( nExpStartPos < nEndPos ) { DBG_ASSERT( 0==nSpaceChars, " pending spaces " ); OUString sExp( rText.copy( nExpStartPos, nEndPos - nExpStartPos ) ); GetExport().Characters( sExp ); } // If there are some spaces left, they have to be exported now. if( nSpaceChars > 0 ) { if( nSpaceChars > 1 ) { OUStringBuffer sTmp; sTmp.append( (sal_Int32)nSpaceChars ); GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_C, sTmp.makeStringAndClear() ); } SvXMLElementExport aElem( GetExport(), XML_NAMESPACE_TEXT, XML_S, sal_False, sal_False ); } } void XMLTextParagraphExport::exportTextDeclarations() { pFieldExport->ExportFieldDeclarations(); // get XPropertySet from the document and ask for AutoMarkFileURL. // If it exists, export the auto-mark-file element. Reference<XPropertySet> xPropertySet( GetExport().GetModel(), UNO_QUERY ); if (xPropertySet.is()) { OUString sUrl; OUString sIndexAutoMarkFileURL( RTL_CONSTASCII_USTRINGPARAM("IndexAutoMarkFileURL")); if (xPropertySet->getPropertySetInfo()->hasPropertyByName( sIndexAutoMarkFileURL)) { xPropertySet->getPropertyValue(sIndexAutoMarkFileURL) >>= sUrl; if (sUrl.getLength() > 0) { GetExport().AddAttribute( XML_NAMESPACE_XLINK, XML_HREF, GetExport().GetRelativeReference(sUrl) ); SvXMLElementExport aAutoMarkElement( GetExport(), XML_NAMESPACE_TEXT, XML_ALPHABETICAL_INDEX_AUTO_MARK_FILE, sal_True, sal_True ); } } } } void XMLTextParagraphExport::exportTextDeclarations( const Reference<XText> & rText ) { pFieldExport->ExportFieldDeclarations(rText); } void XMLTextParagraphExport::exportUsedDeclarations( sal_Bool bOnlyUsed ) { pFieldExport->SetExportOnlyUsedFieldDeclarations( bOnlyUsed ); } void XMLTextParagraphExport::exportTrackedChanges(sal_Bool bAutoStyles) { if (NULL != pRedlineExport) pRedlineExport->ExportChangesList( bAutoStyles ); } void XMLTextParagraphExport::exportTrackedChanges( const Reference<XText> & rText, sal_Bool bAutoStyle) { if (NULL != pRedlineExport) pRedlineExport->ExportChangesList(rText, bAutoStyle); } void XMLTextParagraphExport::recordTrackedChangesForXText( const Reference<XText> & rText ) { if (NULL != pRedlineExport) pRedlineExport->SetCurrentXText(rText); } void XMLTextParagraphExport::recordTrackedChangesNoXText() { if (NULL != pRedlineExport) pRedlineExport->SetCurrentXText(); } void XMLTextParagraphExport::exportTextAutoStyles() { GetAutoStylePool().exportXML( XML_STYLE_FAMILY_TEXT_PARAGRAPH, GetExport().GetDocHandler(), GetExport().GetMM100UnitConverter(), GetExport().GetNamespaceMap() ); GetAutoStylePool().exportXML( XML_STYLE_FAMILY_TEXT_TEXT, GetExport().GetDocHandler(), GetExport().GetMM100UnitConverter(), GetExport().GetNamespaceMap() ); GetAutoStylePool().exportXML( XML_STYLE_FAMILY_TEXT_FRAME, GetExport().GetDocHandler(), GetExport().GetMM100UnitConverter(), GetExport().GetNamespaceMap() ); GetAutoStylePool().exportXML( XML_STYLE_FAMILY_TEXT_SECTION, GetExport().GetDocHandler(), GetExport().GetMM100UnitConverter(), GetExport().GetNamespaceMap() ); GetAutoStylePool().exportXML( XML_STYLE_FAMILY_TEXT_RUBY, GetExport().GetDocHandler(), GetExport().GetMM100UnitConverter(), GetExport().GetNamespaceMap() ); pListAutoPool->exportXML(); } void XMLTextParagraphExport::exportRuby( const Reference<XPropertySet> & rPropSet, sal_Bool bAutoStyles ) { // early out: a collapsed ruby makes no sense if (*(sal_Bool*)rPropSet->getPropertyValue(sIsCollapsed).getValue()) return; // start value ? sal_Bool bStart = (*(sal_Bool*)rPropSet->getPropertyValue(sIsStart).getValue()); if (bAutoStyles) { // ruby auto styles if (bStart) Add( XML_STYLE_FAMILY_TEXT_RUBY, rPropSet ); } else { // prepare element names OUString aRuby(GetXMLToken(XML_RUBY)); OUString sTextRuby(GetExport().GetNamespaceMap(). GetQNameByKey(XML_NAMESPACE_TEXT, aRuby)); OUString sRubyBase(GetXMLToken(XML_RUBY_BASE)); OUString sTextRubyBase(GetExport().GetNamespaceMap(). GetQNameByKey(XML_NAMESPACE_TEXT, sRubyBase)); if (bStart) { // ruby start // we can only start a ruby if none is open DBG_ASSERT(! bOpenRuby, "Can't open a ruby inside of ruby!"); if( bOpenRuby ) return; // save ruby text + ruby char style rPropSet->getPropertyValue(sRubyText) >>= sOpenRubyText; rPropSet->getPropertyValue(sRubyCharStyleName) >>= sOpenRubyCharStyle; // ruby style GetExport().CheckAttrList(); OUString sEmpty; OUString sStyleName(Find( XML_STYLE_FAMILY_TEXT_RUBY, rPropSet, sEmpty )); DBG_ASSERT(sStyleName.getLength() > 0, "I can't find the style!"); GetExport().AddAttribute(XML_NAMESPACE_TEXT, XML_STYLE_NAME, sStyleName); // export <text:ruby> and <text:ruby-base> start elements GetExport().StartElement( XML_NAMESPACE_TEXT, XML_RUBY, sal_False); GetExport().ClearAttrList(); GetExport().StartElement( XML_NAMESPACE_TEXT, XML_RUBY_BASE, sal_False ); bOpenRuby = sal_True; } else { // ruby end // check for an open ruby DBG_ASSERT(bOpenRuby, "Can't close a ruby if none is open!"); if( !bOpenRuby ) return; // close <text:ruby-base> GetExport().EndElement(XML_NAMESPACE_TEXT, XML_RUBY_BASE, sal_False); // write the ruby text (with char style) { if (sOpenRubyCharStyle.getLength() > 0) GetExport().AddAttribute( XML_NAMESPACE_TEXT, XML_STYLE_NAME, GetExport().EncodeStyleName( sOpenRubyCharStyle) ); SvXMLElementExport aRubyElement( GetExport(), XML_NAMESPACE_TEXT, XML_RUBY_TEXT, sal_False, sal_False); GetExport().Characters(sOpenRubyText); } // and finally, close the ruby GetExport().EndElement(XML_NAMESPACE_TEXT, XML_RUBY, sal_False); bOpenRuby = sal_False; } } } void XMLTextParagraphExport::exportMeta( const Reference<XPropertySet> & i_xPortion, sal_Bool i_bAutoStyles, sal_Bool i_isProgress) { static OUString sMeta(RTL_CONSTASCII_USTRINGPARAM("InContentMetadata")); bool doExport(!i_bAutoStyles); // do not export element if autostyles // check version >= 1.2 switch (GetExport().getDefaultVersion()) { case SvtSaveOptions::ODFVER_011: // fall thru case SvtSaveOptions::ODFVER_010: doExport = false; break; default: break; } const Reference< XTextContent > xTextContent( i_xPortion->getPropertyValue(sMeta), UNO_QUERY_THROW); const Reference< XEnumerationAccess > xEA( xTextContent, UNO_QUERY_THROW ); const Reference< XEnumeration > xTextEnum( xEA->createEnumeration() ); if (doExport) { const Reference<rdf::XMetadatable> xMeta(xTextContent, UNO_QUERY_THROW); // text:meta with neither xml:id nor RDFa is invalid xMeta->ensureMetadataReference(); // xml:id and RDFa for RDF metadata GetExport().AddAttributeXmlId(xMeta); GetExport().AddAttributesRDFa(xTextContent); } SvXMLElementExport aElem( GetExport(), doExport, XML_NAMESPACE_TEXT, XML_META, sal_False, sal_False ); // recurse to export content exportTextRangeEnumeration( xTextEnum, i_bAutoStyles, i_isProgress ); } void XMLTextParagraphExport::PreventExportOfControlsInMuteSections( const Reference<XIndexAccess> & rShapes, UniReference<xmloff::OFormLayerXMLExport> xFormExport ) { // check parameters ad pre-conditions if( ( ! rShapes.is() ) || ( ! xFormExport.is() ) ) { // if we don't have shapes or a form export, there's nothing to do return; } DBG_ASSERT( pSectionExport != NULL, "We need the section export." ); Reference<XEnumeration> xShapesEnum = pBoundFrameSets->GetShapes()->createEnumeration(); if(!xShapesEnum.is()) return; while( xShapesEnum->hasMoreElements() ) { // now we need to check // 1) if this is a control shape, and // 2) if it's in a mute section // if both answers are 'yes', notify the form layer export // we join accessing the shape and testing for control Reference<XControlShape> xControlShape(xShapesEnum->nextElement(), UNO_QUERY); if( xControlShape.is() ) { // Reference<XPropertySet> xPropSet( xControlShape, UNO_QUERY ); // Reference<XTextContent> xTextContent; // xPropSet->getPropertyValue( OUString( RTL_CONSTASCII_USTRINGPARAM( "TextRange" ) ) ) >>= xTextContent; Reference<XTextContent> xTextContent( xControlShape, UNO_QUERY ); if( xTextContent.is() ) { if( pSectionExport->IsMuteSection( xTextContent, sal_False ) ) { // Ah, we've found a shape that // 1) is a control shape // 2) is anchored in a mute section // so: don't export it! xFormExport->excludeFromExport( xControlShape->getControl() ); } // else: not in mute section -> should be exported -> nothing // to do } // else: no anchor -> ignore } // else: no control shape -> nothing to do } } sal_Int32 XMLTextParagraphExport::GetHeadingLevel( const OUString& rStyleName ) { if( !pHeadingStyles ) { pHeadingStyles = new XMLStringVector; SvxXMLNumRuleExport::GetOutlineStyles( *pHeadingStyles, GetExport().GetModel() ); } for( XMLStringVector::size_type i=0; i < pHeadingStyles->size(); ++i ) { if( (*pHeadingStyles)[i] == rStyleName ) return static_cast < sal_Int32 >( i ); } return -1; } void XMLTextParagraphExport::PushNewTextListsHelper() { mpTextListsHelper = new XMLTextListsHelper(); maTextListsHelperStack.push_back( mpTextListsHelper ); } void XMLTextParagraphExport::PopTextListsHelper() { delete mpTextListsHelper; mpTextListsHelper = 0; maTextListsHelperStack.pop_back(); if ( !maTextListsHelperStack.empty() ) { mpTextListsHelper = maTextListsHelperStack.back(); } } /* vim:set shiftwidth=4 softtabstop=4 expandtab: */
__label__pos
0.950492
Passability Mini Map [Mini-Mapa Detalhado] 8 de outubro de 2013 Passability Mini Map é um script desenvolvido por Squall para o RPG Maker XP que implementa um mini-mapa em um dos cantos da tela principal do game/projeto, desenvolvido nesta ferramenta, relativamente detalhado. O script é de fácil instalação e pode ser personalizado, dependendo onde você quer que o mini-mapa apareça, bastando editar a linha: @corner = 4 # 1 or 2 or 3 or 4 do script (veja as instruções nos comentários do código). E você pode atribuir cores diferentes para cada tipo de evento, bastando inserir um comentário em cada (veja na demo para ter uma ideia melhor): • -npc: marrom • -enemy: vermelho • -savepoint: azul claro • -teleport: azul • -chest: laranja • -event: amarelo Para instalar o script, basta adicionar o código abaixo acima do Main: #============================================================================== # ■ Passability Mini Map #------------------------------------------------------------------------------ # made by squall // [email protected] # released the 30th of May 2006 #============================================================================== #============================================================================== # ■ Scene_Map #------------------------------------------------------------------------------ # draw the mini map # @corner is the corner you want the mini map to be displayed in. # 1 is upper left, 2 is upper right, 3 is bottom left and 4 is bottom right #============================================================================== class Scene_Map alias main_passminimap main alias update_passminimap update alias transfer_passminimap transfer_player #-------------------------------------------------------------------------- # ● initialize #-------------------------------------------------------------------------- def initialize @corner = 4 # 1 or 2 or 3 or 4 end #-------------------------------------------------------------------------- # ● main #-------------------------------------------------------------------------- def main @mini_map = Map_Event.new(@corner) main_passminimap @mini_map.dispose end #-------------------------------------------------------------------------- # ● update #-------------------------------------------------------------------------- def update @mini_map.update if $game_system.map_interpreter.running? @mini_map.visible = false elsif not $game_system.map_interpreter.running? and @mini_map.on? @mini_map.visible = true end update_passminimap end #-------------------------------------------------------------------------- # ● transfer_player #-------------------------------------------------------------------------- def transfer_player transfer_passminimap @mini_map.dispose @mini_map = Map_Event.new(@corner) end end #============================================================================== # ■ Map_Base #------------------------------------------------------------------------------ #  Base class for mini maps #============================================================================== class Map_Base < Sprite #-------------------------------------------------------------------------- # ● constants and instances #-------------------------------------------------------------------------- PMP_VERSION = 6 ACTIVATED_ID = 1 # set the switch id for the minimap display (on/off) attr_reader :event #-------------------------------------------------------------------------- # ● initialize #-------------------------------------------------------------------------- def initialize(corner) super(Viewport.new(16, 16, width, height)) viewport.z = 8000 @border = Sprite.new @border.x = viewport.rect.x - 6 @border.y = viewport.rect.y - 6 @border.z = viewport.z - 1 @border.bitmap = RPG::Cache.picture("mapback") self.visible = on? self.opacity = 180 case corner when 1 self.x = 16 self.y = 16 when 2 self.x = 640 - width - 16 self.y = 16 when 3 self.x = 16 self.y = 480 - height - 16 when 4 self.x = 640 - width - 16 self.y = 480 - height - 16 else self.x = 16 self.y = 16 end self.visible = on? end #-------------------------------------------------------------------------- # ● dispose #-------------------------------------------------------------------------- def dispose @border.dispose super end #-------------------------------------------------------------------------- # ● x= #-------------------------------------------------------------------------- def x=(x) self.viewport.rect.x = x @border.x = x - 6 end #-------------------------------------------------------------------------- # ● y= #-------------------------------------------------------------------------- def y=(y) self.viewport.rect.y = y @border.y = y - 6 end #-------------------------------------------------------------------------- # ● visible= #-------------------------------------------------------------------------- def visible=(bool) super self.viewport.visible = bool @border.visible = bool end #-------------------------------------------------------------------------- # ● minimap_on? #-------------------------------------------------------------------------- def on? return $game_switches[ACTIVATED_ID] end #-------------------------------------------------------------------------- # ● update #-------------------------------------------------------------------------- def update super self.visible = on? if viewport.ox < display_x viewport.ox += 1 elsif viewport.ox > display_x viewport.ox -= 1 end if viewport.oy < display_y viewport.oy += 1 elsif viewport.oy > display_y viewport.oy -= 1 end end #-------------------------------------------------------------------------- # ● width #-------------------------------------------------------------------------- def width return 120 end #-------------------------------------------------------------------------- # ● height #-------------------------------------------------------------------------- def height return 90 end #-------------------------------------------------------------------------- # ● display_x #-------------------------------------------------------------------------- def display_x return $game_map.display_x * 3 / 64 end #-------------------------------------------------------------------------- # ● display_y #-------------------------------------------------------------------------- def display_y return $game_map.display_y * 3 / 64 end end #============================================================================== # ■ Map_Passability #------------------------------------------------------------------------------ # draws the mini map # #  thanks to Fanha Giang (aka fanha99) for the autotile drawing method #============================================================================== class Map_Passability < Map_Base #-------------------------------------------------------------------------- # ● constants #-------------------------------------------------------------------------- INDEX = [ 26, 27, 32, 33, 4, 27, 32, 33, 26, 5, 32, 33, 4, 5, 32, 33, 26, 27, 32, 11, 4, 27, 32, 11, 26, 5, 32, 11, 4, 5, 32, 11, 26, 27, 10, 33, 4, 27, 10, 33, 26, 5, 10, 33, 4, 5, 10, 33, 26, 27, 10, 11, 4, 27, 10, 11, 26, 5, 10, 11, 4, 5, 10, 11, 24, 25, 30, 31, 24, 5, 30, 31, 24, 25, 30, 11, 24, 5, 30, 11, 14, 15, 20, 21, 14, 15, 20, 11, 14, 15, 10, 21, 14, 15, 10, 11, 28, 29, 34, 35, 28, 29, 10, 35, 4, 29, 34, 35, 4, 29, 10, 35, 38, 39, 44, 45, 4, 39, 44, 45, 38, 5, 44, 45, 4, 5, 44, 45, 24, 29, 30, 35, 14, 15, 44, 45, 12, 13, 18, 19, 12, 13, 18, 11, 16, 17, 22, 23, 16, 17, 10, 23, 40, 41, 46, 47, 4, 41, 46, 47, 36, 37, 42, 43, 36, 5, 42, 43, 12, 17, 18, 23, 12, 13, 42, 43, 36, 41, 42, 47, 16, 17, 46, 47, 12, 17, 42, 47, 0, 1, 6, 7 ] X = [0, 1, 0, 1] Y = [0, 0, 1, 1] #-------------------------------------------------------------------------- # ● initialize #-------------------------------------------------------------------------- def initialize(corner) super(corner) @autotile = RPG::Cache.picture("minimap_tiles") setup() end #-------------------------------------------------------------------------- # ● setup #-------------------------------------------------------------------------- def setup() @map = load_data(sprintf("Data/Map%03d.rxdata", $game_map.map_id)) tileset = $data_tilesets[@map.tileset_id] @passages = tileset.passages @priorities = tileset.priorities redefine_tiles refresh end #-------------------------------------------------------------------------- # ● pass #-------------------------------------------------------------------------- def pass(tile_id) return 15 if tile_id == nil return @passages[tile_id] != nil ? @passages[tile_id] : 15 end #-------------------------------------------------------------------------- # ● passable #-------------------------------------------------------------------------- def passable(tile_id) return pass(tile_id) < 15 end #-------------------------------------------------------------------------- # ● redefine_tile #-------------------------------------------------------------------------- def redefine_tiles width = @map.width height = @map.height map = RPG::Map.new(width, height) map.data = @map.data.dup for x in 0...width for y in 0...height for level in [1, 2] id = @map.data[x, y, level] if id != 0 and @priorities[id] == 0 @map.data[x, y, 0] = id @passages[@map.data[x, y, 0]] = @passages[id] end end end end for x in 0...width for y in 0...height for level in [0] tile = @map.data[x, y, level] u = @map.data[x, y-1, level] l = @map.data[x-1, y, level] r = @map.data[x+1, y, level] d = @map.data[x, y+1, level] if !passable(tile) map.data[x, y] = 0 else if tile == 0 map.data[x, y, level] = 0 next end if pass(tile) < 15 if !passable(u) and !passable(l) and !passable(r) and !passable(d) map.data[x, y, level] = 0 elsif !passable(u) and !passable(l) and !passable(r) and passable(d) map.data[x, y, level] = 90 elsif !passable(u) and !passable(l) and !passable(d) and passable(r) map.data[x, y, level] = 91 elsif !passable(u) and !passable(r) and !passable(d) and passable(l) map.data[x, y, level] = 93 elsif !passable(l) and !passable(r) and !passable(d) and passable(u) map.data[x, y, level] = 92 elsif !passable(u) and !passable(d) and passable(r) and passable(l) map.data[x, y, level] = 81 elsif !passable(u) and !passable(r) and passable(d) and passable(l) map.data[x, y, level] = 84 elsif !passable(u) and !passable(l) and passable(d) and passable(r) map.data[x, y, level] = 82 elsif !passable(d) and !passable(r) and passable(l) and passable(u) map.data[x, y, level] = 86 elsif !passable(d) and !passable(l) and passable(r) and passable(u) map.data[x, y, level] = 88 elsif !passable(r) and !passable(l) and passable(d) and passable(u) map.data[x, y, level] = 80 elsif !passable(u) and passable(d) and passable(r) and passable(l) map.data[x, y, level] = 68 elsif !passable(d) and passable(u) and passable(r) and passable(l) map.data[x, y, level] = 76 elsif !passable(r) and passable(d) and passable(u) and passable(l) map.data[x, y, level] = 72 elsif !passable(l) and passable(d) and passable(u) and passable(r) map.data[x, y, level] = 64 else map.data[x, y, level] = 48 end else map.data[x, y, level] = 0 end end end end end @map = map.dup map = nil end #-------------------------------------------------------------------------- # ● refresh #-------------------------------------------------------------------------- def refresh self.visible = false self.bitmap = Bitmap.new(@map.width * 6, @map.height * 6) bitmap = Bitmap.new(@map.width * 6, @map.height * 6) rect1 = Rect.new(6, 0, 6, 6) for y in [email protected] for x in [email protected] for level in [0] tile_id = @map.data[x, y, level] next if tile_id == 0 id = tile_id / 48 - 1 tile_id %= 48 for g in 0..3 h = 4 * tile_id + g y1 = INDEX[h] / 6 x1 = INDEX[h] % 6 rect2 = Rect.new(x1 * 3, y1 * 3, 3, 3) bitmap.blt(x * 6 + X[g] * 3, y * 6 + Y[g] * 3, @autotile, rect2) end end end end d_rect = Rect.new(0, 0, @map.width * 6, @map.height * 6) s_rect = Rect.new(0, 0, bitmap.width, bitmap.height) self.bitmap.stretch_blt(d_rect, bitmap, s_rect) self.viewport.ox = display_x self.viewport.oy = display_y bitmap.clear bitmap.dispose end end #============================================================================== # ■ Map_Event #------------------------------------------------------------------------------ #  draw the events and hero position #============================================================================== class Map_Event < Map_Passability #-------------------------------------------------------------------------- # ● initialize #-------------------------------------------------------------------------- def initialize(corner = 4) super(corner) @dots = [] @player = Sprite.new(self.viewport) @player.bitmap = RPG::Cache.picture("mm cursors") @player.src_rect = Rect.new(0, 0, 15, 15) @player.z = self.z + 3 @events = {} for key in $game_map.events.keys event = $game_map.events[key] next if event.list == nil for i in 0...event.list.size next if event.list[i].code != 108 @events[key] = Sprite.new(self.viewport) @events[key].z = self.z + 2 if event.list[i].parameters[0].include?("event") @events[key].bitmap = RPG::Cache.picture("event") elsif event.list[i].parameters[0].include?("enemy") @events[key].bitmap = RPG::Cache.picture("enemy") elsif event.list[i].parameters[0].include?("teleport") @events[key].bitmap = RPG::Cache.picture("teleport") elsif event.list[i].parameters[0].include?("chest") @events[key].bitmap = RPG::Cache.picture("chest") elsif event.list[i].parameters[0].include?("npc") @events[key].bitmap = RPG::Cache.picture("npc") elsif event.list[i].parameters[0].include?("savepoint") @events[key].bitmap = RPG::Cache.picture("savepoint") end end end end #-------------------------------------------------------------------------- # ● dispose #-------------------------------------------------------------------------- def dispose @player.dispose for event in @events.values event.dispose end super end #-------------------------------------------------------------------------- # ● update #-------------------------------------------------------------------------- def update super @player.x = $game_player.real_x * 3 / 64 - 5 @player.y = $game_player.real_y * 3 / 64 - 4 @player.src_rect.x = ($game_player.direction / 2 - 1) * 15 for key in @events.keys event = @events[key] mapevent = $game_map.events[key] event.x = mapevent.real_x * 3 / 64 event.y = mapevent.real_y * 3 / 64 end end end autor, site, canal ou publisher Squall tamanho 860KB licençaGrátis sistemas operacionais compativeisWindows 98/98SE/Me/2000/XP/Vista/7 download link DOWNLOAD Observação: se você gostou deste post ou ele lhe foi útil de alguma forma, por favor considere apoiar financeiramente a Gaming Room. Fico feliz só de ajudar, mas a contribuição do visitante é muito importante para que este site continua existindo e para que eu possa continuar provendo este tipo de conteúdo e melhorar cada vez mais. Clique aqui e saiba como. Obrigado! Deixe um comentário Inscreva-se na nossa newsletter!
__label__pos
0.730769
Commits Reimar Bauer committed 70c2406 Draft Merge merged main Comments (0) Files changed (11) MoinMoin/apps/frontend/views.py File contents unchanged. MoinMoin/constants/contenttypes.py File contents unchanged. MoinMoin/items/__init__.py File contents unchanged. MoinMoin/items/content.py @register -class TWikiDraw(Draw): - """ - drawings by TWikiDraw applet. It creates three files which are stored as tar file. - """ - contenttype = 'application/x-twikidraw' - display_name = 'TDRAW' - - class ModifyForm(Draw.ModifyForm): - template = "modify_twikidraw.html" - help = "" - - def handle_post(self): - # called from modify UI/POST - file_upload = request.files.get('filepath') - filename = request.form['filename'] - basepath, basename = os.path.split(filename) - basename, ext = os.path.splitext(basename) - - filecontent = file_upload.stream - content_length = None - if ext == '.draw': # TWikiDraw POSTs this first - filecontent = filecontent.read() # read file completely into memory - filecontent = filecontent.replace("\r", "") - elif ext == '.map': - filecontent = filecontent.read() # read file completely into memory - filecontent = filecontent.strip() - elif ext == '.png': - #content_length = file_upload.content_length - # XXX gives -1 for wsgiref, gives 0 for werkzeug :( - # If this is fixed, we could use the file obj, without reading it into memory completely: - filecontent = filecontent.read() - - self.put_member('drawing' + ext, filecontent, content_length, - expected_members=set(['drawing.draw', 'drawing.map', 'drawing.png'])) - - def _render_data(self): - # TODO: this could be a converter -> dom, then transcluding this kind - # of items and also rendering them with the code in base class could work - item_name = self.name - drawing_url = url_for('frontend.get_item', item_name=item_name, member='drawing.draw', rev=self.rev.revid) - png_url = url_for('frontend.get_item', item_name=item_name, member='drawing.png', rev=self.rev.revid) - title = _('Edit drawing %(filename)s (opens in new window)', filename=item_name) - - mapfile = self.get_member('drawing.map') - try: - image_map = mapfile.read() - mapfile.close() - except (IOError, OSError): - image_map = '' - if image_map: - # we have a image map. inline it and add a map ref to the img tag - mapid = 'ImageMapOf' + item_name - image_map = image_map.replace('%MAPNAME%', mapid) - # add alt and title tags to areas - image_map = re.sub(r'href\s*=\s*"((?!%TWIKIDRAW%).+?)"', r'href="\1" alt="\1" title="\1"', image_map) - image_map = image_map.replace('%TWIKIDRAW%"', '{0}" alt="{1}" title="{2}"'.format((drawing_url, title, title))) - title = _('Clickable drawing: %(filename)s', filename=item_name) - - return Markup(image_map + u'<img src="{0}" alt="{1}" usemap="#{2}" />'.format(png_url, title, mapid)) - else: - return Markup(u'<img src="{0}" alt="{1}" />'.format(png_url, title)) - - -@register class AnyWikiDraw(Draw): """ drawings by AnyWikiDraw applet. It creates three files which are stored as tar file. MoinMoin/templates/base.html File contents unchanged. MoinMoin/templates/modify_select_contenttype.html File contents unchanged. MoinMoin/templates/modify_twikidraw.html -{% macro data_editor(form, item_name) %} -<p> -<applet code="CH.ifa.draw.twiki.TWikiDraw.class" - archive="{{ url_for('serve.files', name='twikidraw_moin', filename='twikidraw_moin.jar') }}" - width="800" height="620"> - <param name="drawpath" value="{{ url_for('frontend.get_item', item_name=item_name, member='drawing.draw') }}" /> - <param name="pngpath" value="{{ url_for('frontend.get_item', item_name=item_name, member='drawing.png') }}" /> -<param name="savepath" value="{{ url_for('frontend.modify_item', item_name=item_name, contenttype='application/x-twikidraw') }}" /> -<param name="basename" value="drawing" /> -<param name="viewpath" value="{{ url_for('frontend.show_item', item_name=item_name) }}" /> -<param name="helppath" value="" /> -<strong>{{ _("NOTE:") }}</strong> {{ _("You need a Java enabled browser to edit the drawing.") }} -</applet> -</p> -<br /> -{% endmacro %} MoinMoin/templates/show.html {% extends theme("layout.html") %} {% import "utils.html" as utils %} - +{% block header %} + {{ super() }} + {{ JS_SCRIPTS["show.header"] }} +{%- endblock %} {% block head_links %} {{ super() }} <link rel="alternate" title="{{ item_name }} changes" href="{{ url_for('feed.atom', item_name=item_name) }}" type="application/atom+xml" /> MoinMoin/themes/__init__.py app.jinja_env.filters['json_dumps'] = dumps # please note that these filters are installed by flask-babel: # datetimeformat, dateformat, timeformat, timedeltaformat - + app.jinja_env.globals['JS_SCRIPTS'] = app.cfg.js_scripts app.jinja_env.globals.update({ # please note that flask-babel/jinja2.ext installs: # _, gettext, ngettext 'pdfminer', # pdf -> text/plain conversion 'XStatic>=0.0.2', # support for static file pypi packages 'XStatic-CKEditor>=3.6.1.2', - 'XStatic-jQuery>=1.8.2', 'XStatic-jQuery-File-Upload>=4.4.2', 'XStatic-JSON-js', 'XStatic-svgweb>=2011.2.3.2', - 'XStatic-TWikiDraw-moin>=2004.10.23.2', 'XStatic-AnyWikiDraw>=0.14.2', 'XStatic-svg-edit-moin>=2011.07.07.2', ], # see https://bitbucket.org/thomaswaldmann/xstatic for infos about xstatic: from xstatic.main import XStatic # names below must be package names - mod_names = ['jquery', 'jquery_file_upload', + mod_names = ['jquery_file_upload', 'json_js', 'ckeditor', 'svgweb', - 'svgedit_moin', 'twikidraw_moin', 'anywikidraw', + 'svgedit_moin', 'anywikidraw', ] pkg = __import__('xstatic.pkg', fromlist=mod_names) for mod_name in mod_names: mod = getattr(pkg, mod_name) xs = XStatic(mod, root_url='/static', provider='local', protocol='http') serve_files.update([(xs.name, xs.base_dir)]) - + # highly experimental + # if we do that we need for all of our xstatic mod_names a similar mplugin module + # + import sys + try: + mplugins = sys.modules['moinplugin.pkg'] + except KeyError: + mplugins = None + if mplugins: + plugin_packages = ["mathjax", "twikidraw", "jquery", ] + pkg = __import__('moinplugin.pkg', fromlist=plugin_packages) + template_dirs = [] + js_scripts = {} + for mod_name in plugin_packages: + mod = getattr(pkg, mod_name) + # plugin can add a template dir + try: + template_dirs.append(mod.TEMPLATE_DIR) + except AttributeError: + pass + # plugin need to serve something + try: + serve_files.update([(mod.STATIC_NAME, mod.STATIC_PATH)]) + except AttributeError: + pass + # mathjax plugin adds only js + try: + for key in mod.JS_SCRIPTS.keys(): + if key not in js_scripts: + js_scripts[key] = "" + js_scripts[key] += mod.JS_SCRIPTS[key] + except AttributeError: + pass MOINCFG = Config # Flask only likes uppercase stuff # Flask settings - see the flask documentation about their meaning
__label__pos
0.993217
Skip navigation links Oracle® Coherence Java API Reference Release 3.7.1.0 E22843-01 com.tangosol.io.pof Class ThrowablePofSerializer java.lang.Object   extended by com.tangosol.io.pof.ThrowablePofSerializer All Implemented Interfaces: PofSerializer public class ThrowablePofSerializer extends java.lang.Object implements PofSerializer PofSerializer implementation that can serialize and deserialize a Throwable to/from a POF stream. This serializer is provides a catch-all mechanism for serializing exceptions. Any deserialized exception will loose type information, and simply be represented as a PortableException. The basic detail information of the exception is retained. PortableException and this class work asymmetrically to provide the serialization routines for exceptions. Author: mf 2008.08.25 Constructor Summary ThrowablePofSerializer()           Default constructor.   Method Summary  java.lang.Object deserialize(PofReader in)           Deserialize a user type instance from a POF stream by reading its state using the specified PofReader object.  void serialize(PofWriter out, java.lang.Object o)           Serialize a user type instance to a POF stream by writing its state using the specified PofWriter object.   Constructor Detail ThrowablePofSerializer public ThrowablePofSerializer() Default constructor. Method Detail serialize public void serialize(PofWriter out, java.lang.Object o) throws java.io.IOException Serialize a user type instance to a POF stream by writing its state using the specified PofWriter object. An implementation of PofSerializer is required to follow the following steps in sequence for writing out an object of a user type: 1. If the object is evolvable, the implementation must set the version by calling PofWriter.setVersionId(int). 2. The implementation may write any combination of the properties of the user type by using the "write" methods of the PofWriter, but it must do so in the order of the property indexes. 3. After all desired properties of the user type have been written, the implementation must terminate the writing of the user type by calling PofWriter.writeRemainder(com.tangosol.util.Binary). Specified by: serialize in interface PofSerializer Parameters: out - the PofWriter with which to write the object's state o - the object to serialize Throws: java.io.IOException - if an I/O error occurs deserialize public java.lang.Object deserialize(PofReader in) throws java.io.IOException Deserialize a user type instance from a POF stream by reading its state using the specified PofReader object. An implementation of PofSerializer is required to follow the following steps in sequence for reading in an object of a user type: 1. If the object is evolvable, the implementation must get the version by calling PofReader.getVersionId(). 2. The implementation may read any combination of the properties of the user type by using "read" methods of the PofReader, but it must do so in the order of the property indexes. Additionally, the implementation must call PofReader.registerIdentity(java.lang.Object) with the new instance prior to reading any properties which are user type instances themselves. 3. After all desired properties of the user type have been read, the implementation must terminate the reading of the user type by calling PofReader.readRemainder(). Specified by: deserialize in interface PofSerializer Parameters: in - the PofReader with which to read the object's state Returns: the deserialized user type instance Throws: java.io.IOException - if an I/O error occurs Skip navigation links Oracle® Coherence Java API Reference Release 3.7.1.0 E22843-01 Copyright © 2000, 2011, Oracle and/or its affiliates. All rights reserved.
__label__pos
0.617912
LeetCode Add Two Numbers: Simulate Addition With Linked List Overview LeetCode Add Two Numbers is a direct simulation of addition, be careful about the carry which could be added into the more significant digit. LeetCode Add Two Numbers You are given two linked lists representing two non-negative numbers. The digits are stored in reverse order and each of their nodes contain a single digit. Add the two numbers and return it as a linked list. Input: (2 -> 4 -> 3) + (5 -> 6 -> 4) Output: 7 -> 0 -> 8 Analysis: Simulate Addition With Linked List There is no tricky algorithm involved in this problem. The solution is straightforward by directly simulation of addition. The only things need to be kept in our mind is the following cases: 1.   4->NULL + 6->NULL == 0->1->NULL 2.   1-NULL + 9->9->9->NULL == 0->0->0->1->NULL The following code is accepted by LeetCode OJ to pass this Add Two Numbers problem: ListNode *addTwoNumbers(ListNode *l1, ListNode *l2) { int sum = 0; int carry = 0; int lValue = 0; int rValue = 0; ListNode s(0); ListNode *p = &s; while (l1 || l2) { lValue = (NULL != l1) ? l1->val : 0; rValue = (NULL != l2) ? l2->val : 0; sum = lValue + rValue + carry; carry = sum / 10; sum %= 10; p->next = new ListNode(sum); p = p->next; if (l1) l1 = l1->next; if (l2) l2 = l2->next; } if (carry) p->next = new ListNode(carry); return s.next; } Summary LeetCode Add Two Numbers is a direct simulation of addition, be careful about the carry which could be added into the more significant digit. Written on January 16, 2015
__label__pos
0.994494
Sign up × MathOverflow is a question and answer site for professional mathematicians. It's 100% free, no registration required. Long time ago there was a question on whether a polynomial bijection $\mathbb Q^2\to\mathbb Q$ exists. Only one attempt of answering it has been given, highly downvoted by the way. But this answer isn't obviously unsuccessful, because the following problem (for case $n=2$) remains open. Problem. Let $f$ be a polynomial with rational (or even integer!) coefficients in $n$ variables $x_1,\dots,x_n$. Suppose there exist two distinct points $\boldsymbol a=(a_1,\dots,a_n)$ and $\boldsymbol b=(b_1,\dots,b_n)$ from $\mathbb R^n$ such that $f(\boldsymbol a)=f(\boldsymbol b)$. Does this imply the existence of two points $\boldsymbol a'$ and $\boldsymbol b'$ from $\mathbb Q^n$ satisfying $f(\boldsymbol a')=f(\boldsymbol b')$? Even case $n=1$ seems to be non-obvious. EDIT. Just because we have a very nice counter example (immediately highly rated by the MO community) by Hailong Dao in case $n=1$ and because for $n>1$ there are always points $\boldsymbol a,\boldsymbol b\in\mathbb R^n$ with the above property, the problem can be "simplified" as follows. Is it true for a polynomial $f\in\mathbb Q[\boldsymbol x]$ in $n>1$ variables that there exist two points $\boldsymbol a,\boldsymbol b\in\mathbb Q^n$ such that $f(\boldsymbol a)=f(\boldsymbol b)$? The existence of injective polynomials $\mathbb Q^2\to\mathbb Q$ is discussed in B. Poonen's preprint (and in comments to this question). What can be said for $n>2$? FURTHER EDIT. The expected answer to the problem is in negative. In other words, there exist injective polynomials $\mathbb Q^n\to\mathbb Q$ for any $n$. Thanks to the comments of Harry Altman and Will Jagy, case $n>1$ is now fully reduced to $n=2$. Namely, any injective polynomial $F(x_1,x_2)$ gives rise to the injective polynomial $F(F(x_1,x_2),x_3)$, and so on; in the other direction, any $F(x_1,\dots,x_n)$ in more than 2 variables can be specialized to $F(x_1,x_2,0,\dots,0)$. In spite of Bjorn Poonen's verdict that case $n=2$ can be resolved by an appeal to the Bombieri--Lang conjecture for $k$-rational points on surfaces of general type (or even to the 4-variable version of the $abc$ conjecture), I remain with a hope that this can be done by simpler means. My vague attempt (for which I search in the literature) is to start with a homogeneous form $F(x,y)=ax^n+by^n$, or any other homogeneous form of odd degree $n$, which has the property that only finitely many integers are represented by $F(x,y)$ with $x,y\in\mathbb Z$ relatively prime. In order to avoid this finite set of "unpleasant" pairs $x,y$, one can replace them by other homogeneous forms $x=AX^m+BY^m$ and $y=CX^m+DY^m$ (again, for $m$ odd and sufficiently large, say), so that $x$ and $y$ escape the unpleasant values. Then the newer homogeneous form $G(X,Y)=F(AX^m+BY^m,CX^m+DY^m)$ will give the desired polynomial injection. So, can one suggest a homogeneous form $F(x,y)$ with the above property? share|cite|improve this question      Well, it's clear that this won't work if you replace $\mathbb{R}$ with $\mathbb{C}$ from considering $x^n$ when $n$ is odd. That also shows it won't work if you replace it with general $\mathbb{Q}_p$, either... – Harry Altman Jun 6 '10 at 9:02 2   For any polynomial in two variables there exist distinct $a,b$ so that $f(a)=f(b)$, so the condition you put on $f$ is always fulfilled. – Guy Katriel Jun 6 '10 at 9:18 4   As regards the new question, if there's a counterexample $f$ for $n=2$, there's a counterexample for any $n$, as you can just take $f(f(x,y),z)$ when $n=3$, etc. So if we expect there is a counterexample for $n=2$ then we shouldn't be able to prove this at all; I guess considering $n>2$ might still be helpful if that makes finding counterexamples easier? – Harry Altman Jun 6 '10 at 17:58 4   @Harry, on the other hand, if there is an injective example in $ n \geq 3$ variables, by setting $n-2$ of them to $0$ we get an injective example in dimension $2.$ So you have shown that there is an injective polynomial in dimension 2 if and only if there is an example for every $n \geq 2.$ – Will Jagy Jun 6 '10 at 18:55 1   Harry and Will, thank you for these comments. So, the problem is reduced to finding just one counter example for some $n>1$. (On the other hand, I guess that your comments have resulted in somebody's downvote.) – Wadim Zudilin Jun 6 '10 at 21:53 1 Answer 1 up vote 27 down vote accepted Let $f(x)=x^3-5x/4$. Then for $x\neq y$, $f(x)=f(y)$ iff $x^2+xy+y^2=5/4$ or $(2x+y)^2+3y^2=5$. The last equation clealy have real solutions. But if there are rational solutions, then there are integers $X,Y,N$ such that $(2X+Y)^2+3Y^2=5N^2$. This shows $X,Y,N$ all divisible by $5$, ... share|cite|improve this answer      This is a very nice counter example for $n=1$! – Wadim Zudilin Jun 6 '10 at 9:52      I'm surprised this was possible with just a cubic. – Harry Altman Jun 6 '10 at 10:15 1   Also, this solution can be quickly tweaked to have integer coefficients; if there are no rational solutions to $x^2+xy+y^2=5/4$, then there can't be any to $x^2+xy+y^2=5$, either. Hence $x^3-5x$ also works. – Harry Altman Jun 6 '10 at 10:35 1   Hailong, I used your solution as the hardest problem in my number theory class (to show that your $f(x)$ is injective). Two students (of 16) could do it. – Wadim Zudilin Dec 9 '10 at 4:40 2   @Wadim: (-: That's a good use of MO. – Hailong Dao Dec 9 '10 at 22:15 Your Answer   discard By posting your answer, you agree to the privacy policy and terms of service. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.973481
What is Vietnam domain propagation? November 22, 2017 Vietnam domain propagation is the progress of transferring a domain to a new owner, and it is a progress that does not happen instantly. One can expect to wait for anywhere between 1-2 days before the full switch has been made and your site starts showing up at its new domain. So, why does this happen? Well, in this article we will look into all things related to Domain Propagation – so read on to find out more! 1. Domain propagation or DNS propagation Domain propagation, also sometimes referred to as DNS propagation, is the progress of updating every server across the web with brand new information. Now, think about it – there are millions upon millions of service across the entire web, and all of them need to be updated. It is no wonder why there is a lag between when changes are made and when all the servers have officially registered it. Domain vietnam 1. How does it work? DNS servers have the job of translating IP addresses into domain. Even if your site has a short and simple name, the fact remains that it is also located at a not so simple numerical IP address. When you visit a site, a DNS server receives the VN domain that you type in and starts processing which IP address it has in order to send you to correct place. This is a progress that happens seamlessly, why is why you do not ever have to think too much into it. Actually, the only time you ever have to think about this progress is when you make a DNS change. When you change domain or move to a new hosting provider, every single DNS server on the globe needs to register this change of information before they know how to translate your domain to the correct IP address. To further complicate things, different servers will receive updated information at different types, which is why you may be able to see new and updated information but your friend across the street may not. Because DNS changes are quite rare, most DNS servers will cache information that they have gathered in previous searches. Therefore, if you searched one domain .com.vn last week and the DNS server was indeed able to translate the domain to a particular IP address, it will go into default and direct you to that same IP again. After a while, it will learn that changed have been made and that it should now send people who are looking for this the domain .vn referred to the new IP address where the site lives. It is also worth keeping note that browsers often cache the information received from specific sites, so even after a DNS server has updated, you may still need to clear your browser’s cache in order to see the new and updated site. 1. When is this progress complete? This can be tricky as your DNS server can register the update before someone else’s does. This is why you cannot assume that once you see the new site at its domain – that everyone else will as well. You can get a good idea of when your domain propagation is complete by using this tool. The results will not give you a 100 percent guarantee that every single person on the planet can see your new site, but it can assist you in confirming when the change has been completed for the majority of people. tags: register domain vn
__label__pos
0.98498
How can I export a string to a .txt file? 1 ビュー (過去 30 日間) Pablo Fernández Pablo Fernández 2022 年 5 月 21 日 コメント済み: Pablo Fernández 2022 年 5 月 21 日 Hello, I'm working on a project where I need to export some text to a .txt file. Everything I find on the Internet is about exporting data (numbers) but I need to export some text. I can do this using numbers but I can't do it if it's a string or any kind of text. function creararchivo A = 5 ; save Prueba1.txt A -ascii end This code doesn't work as it did when A was 5: function creararchivo A = "B" ; save Prueba1.txt A -ascii end Thanks in advance 採用された回答 dpb dpb 2022 年 5 月 21 日 編集済み: dpb 2022 年 5 月 21 日 As documented, save translates character data to ASCII codes with the 'ascii' option. It means it literally! Use writematrix(A,'yourfilename.txt') instead   1 件のコメント Pablo Fernández Pablo Fernández 2022 年 5 月 21 日 It worked as I wanted to, thank you so much. サインインしてコメントする。 その他の回答 (1 件) Voss Voss 2022 年 5 月 21 日 From the documentation for save: "Use one of the text formats to save MATLAB numeric values to text files. In this case: • Each variable must be a two-dimensional double array. [...] If you specify a text format and any variable is a two-dimensional character array, then MATLAB translates characters to their corresponding internal ASCII codes. For example, 'abc' appears in a text file as: 9.7000000e+001 9.8000000e+001 9.9000000e+001" (That's talking about character arrays, and you showed code where you tried to save a string, but you also said it doesn't work if the variable is any kind of text, which would include character arrays.) So that explains what's happening because that's essentially the situation you have here. A = 'B' ; save Prueba1_char.txt A -ascii type Prueba1_char.txt % character 'B' written as its ASCII code, 66 6.6000000e+01 A = "B" ; save Prueba1_string.txt A -ascii Warning: Attempt to write an unsupported data type to an ASCII file. Variable 'A' not written to file. type Prueba1_string.txt % nothing There are functions you can use to write text to a text file. You might look into fprintf fid = fopen('Prueba1.txt','w'); fprintf(fid,'%s',A); fclose(fid); type Prueba1.txt B   1 件のコメント Pablo Fernández Pablo Fernández 2022 年 5 月 21 日 Thanks for your reply, this works perfectly and it allowed me to understand where was the problem. サインインしてコメントする。 Community Treasure Hunt Find the treasures in MATLAB Central and discover how the community can help you! Start Hunting! Translated by
__label__pos
0.837852
Como transformar uma consulta CTE do MSSQL para o MySQL? no meu esquema MySQL, eu tenho a tabela category(id, parentid, name) No MSSQL, eu tenho essa consulta CTE (para construir uma tree de categoria de baixo para cima para um ID de categoria fornecido: with CTE (id, pid, name) as ( select id, parentid as pid,name from category where id = 197 union all select CTE.pid as id , category.parentid as pid, category.name from CTE inner join category on category.id = CTE.pid ) select * from CTE Como “transformar” essa consulta no MySQL? Infelizmente o MySQL não suporta CTE (Common Table Expressions). Este é o IMO muito atrasado. Geralmente, você pode simplesmente usar uma subconsulta, mas esse CTE específico é recursivo : ele se refere a si mesmo dentro da consulta. Os CTEs recursivos são extremamente úteis para dados hierárquicos, mas novamente: o MySql não os suporta de forma alguma. Você precisa implementar um procedimento armazenado para obter os mesmos resultados. Uma resposta anterior minha deve fornecer um bom ponto de partida: Gerando Árvore Baseada em Profundidade a partir de Dados Hierárquicos no MySQL (sem CTEs) infelizmente MYSQL ou XAMPP (MARIADB) mysql não suporta CTEs (COMMON TABLE EXPRESSIONS), para o mesmo você terá que usar consultas aninhadas. Para mais informações clique no link abaixo: – https://mariadb.com/kb/en/library/with/ Felizmente não é mais necessário, pois o MySQL a partir do 8.0.1 suporta o CTE .
__label__pos
0.894958
Why do Tinyslideshow thumbnails disappear when page is zoomed smaller? I set up TinySlideshow to have 5 vertical thumbnails at the bottom, which looks fine. But when the page is zoomed-out in FireFox 8, Internet Explorer 8, and Chrome to make the page smaller in the browser window, there is a problem. (However, the problem doesn't happen in Opera). The last thumbnail at the far right edge disappears (turns black) when the page size is zoomed smaller. If the page is zoomed  back to normal size again, the thumbnail re-appears again. Scrolling the screen larger doesn't make the thumbnail disappear. It seems like the thumbnails are spread across 2 rows (the 2nd row being invisible, positioned directly beneath the first) and the thumbnail on the far right edge is always being droppped down onto this second line when the page gets zoomed smaller. Any solutions to this? Thanks! asked Nov 16, 2011 by anonymous Please log in or register to answer this question. Web Analytics
__label__pos
0.982765
1. Join Now AVForums.com uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More. Why??? Discussion in 'Nintendo Forums' started by hornydragon, Aug 14, 2004. 1. hornydragon hornydragon Well-known Member Joined: Dec 19, 2001 Messages: 28,293 Products Owned: 0 Products Wanted: 0 Trophy Points: 136 Location: Somewhere near the M4 most of the time...... Ratings: +1,215 Why does a none modified PAL XBOX with Component kit (Gamester) ouput 480? why not 576p???? Anyone tell me?   2. Pooon Pooon Well-known Member Joined: Dec 26, 2002 Messages: 7,118 Products Owned: 0 Products Wanted: 0 Trophy Points: 136 Location: Canterbury, Kent Ratings: +722 The PAL xbox doew not support any form of High-Def which i believe 576p to be, so thats why.   3. hornydragon hornydragon Well-known Member Joined: Dec 19, 2001 Messages: 28,293 Products Owned: 0 Products Wanted: 0 Trophy Points: 136 Location: Somewhere near the M4 most of the time...... Ratings: +1,215 576p ia Pal prog same as you get from PAL prog DVD, 480p is NTSC prog scan same as you get from an NTSC prog scan DVD. So why does a PAL XBOX output 480P instead of 576p or even 576i.   4. CAS FAN CAS FAN Moderator Joined: Nov 24, 2003 Messages: 34,512 Products Owned: 1 Products Wanted: 1 Trophy Points: 167 Location: The Wheldon Road end! Ratings: +7,730 I presume it's because the high def is not meant for Pal regions. When you mod the machine it must give you access to a function that was designed for NSTC regions, hence 480p. It could also be because games are originally written in NSTC and then our games are encoded to PAL. To view the game full screen therefore it must be run in NSTC (or 60 htz) mode.   5. hornydragon hornydragon Well-known Member Joined: Dec 19, 2001 Messages: 28,293 Products Owned: 0 Products Wanted: 0 Trophy Points: 136 Location: Somewhere near the M4 most of the time...... Ratings: +1,215 but this is a bog standard uk Xbox, its just a normal unit, but outputs 480p (why not 576p) there is no other option other than 480p very odd.   6. amordue amordue Standard Member Joined: Nov 13, 2002 Messages: 101 Products Owned: 0 Products Wanted: 0 Trophy Points: 21 Location: Edinburgh Ratings: +1 I'm guessing its because the PAL progressive standard was not fixed until after the XBox launch in this country. Alex   Share This Page Loading...
__label__pos
0.994865
blob: b91b0f165498676db1e5fa8f9718a075219a10de [file] [log] [blame] /* * This module provides an interface to trigger and test firmware loading. * * It is designed to be used for basic evaluation of the firmware loading * subsystem (for example when validating firmware verification). It lacks * any extra dependencies, and will not normally be loaded by the system * unless explicitly requested by name. */ #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt #include <linux/init.h> #include <linux/module.h> #include <linux/printk.h> #include <linux/completion.h> #include <linux/firmware.h> #include <linux/device.h> #include <linux/fs.h> #include <linux/miscdevice.h> #include <linux/slab.h> #include <linux/uaccess.h> #include <linux/delay.h> #include <linux/kthread.h> #define TEST_FIRMWARE_NAME "test-firmware.bin" #define TEST_FIRMWARE_NUM_REQS 4 static DEFINE_MUTEX(test_fw_mutex); static const struct firmware *test_firmware; struct test_batched_req { u8 idx; int rc; bool sent; const struct firmware *fw; const char *name; struct completion completion; struct task_struct *task; struct device *dev; }; /** * test_config - represents configuration for the test for different triggers * * @name: the name of the firmware file to look for * @sync_direct: when the sync trigger is used if this is true * request_firmware_direct() will be used instead. * @send_uevent: whether or not to send a uevent for async requests * @num_requests: number of requests to try per test case. This is trigger * specific. * @reqs: stores all requests information * @read_fw_idx: index of thread from which we want to read firmware results * from through the read_fw trigger. * @test_result: a test may use this to collect the result from the call * of the request_firmware*() calls used in their tests. In order of * priority we always keep first any setup error. If no setup errors were * found then we move on to the first error encountered while running the * API. Note that for async calls this typically will be a successful * result (0) unless of course you've used bogus parameters, or the system * is out of memory. In the async case the callback is expected to do a * bit more homework to figure out what happened, unfortunately the only * information passed today on error is the fact that no firmware was * found so we can only assume -ENOENT on async calls if the firmware is * NULL. * * Errors you can expect: * * API specific: * * 0: success for sync, for async it means request was sent * -EINVAL: invalid parameters or request * -ENOENT: files not found * * System environment: * * -ENOMEM: memory pressure on system * -ENODEV: out of number of devices to test * -EINVAL: an unexpected error has occurred * @req_firmware: if @sync_direct is true this is set to * request_firmware_direct(), otherwise request_firmware() */ struct test_config { char *name; bool sync_direct; bool send_uevent; u8 num_requests; u8 read_fw_idx; /* * These below don't belong her but we'll move them once we create * a struct fw_test_device and stuff the misc_dev under there later. */ struct test_batched_req *reqs; int test_result; int (*req_firmware)(const struct firmware **fw, const char *name, struct device *device); }; struct test_config *test_fw_config; static ssize_t test_fw_misc_read(struct file *f, char __user *buf, size_t size, loff_t *offset) { ssize_t rc = 0; mutex_lock(&test_fw_mutex); if (test_firmware) rc = simple_read_from_buffer(buf, size, offset, test_firmware->data, test_firmware->size); mutex_unlock(&test_fw_mutex); return rc; } static const struct file_operations test_fw_fops = { .owner = THIS_MODULE, .read = test_fw_misc_read, }; static void __test_release_all_firmware(void) { struct test_batched_req *req; u8 i; if (!test_fw_config->reqs) return; for (i = 0; i < test_fw_config->num_requests; i++) { req = &test_fw_config->reqs[i]; if (req->fw) release_firmware(req->fw); } vfree(test_fw_config->reqs); test_fw_config->reqs = NULL; } static void test_release_all_firmware(void) { mutex_lock(&test_fw_mutex); __test_release_all_firmware(); mutex_unlock(&test_fw_mutex); } static void __test_firmware_config_free(void) { __test_release_all_firmware(); kfree_const(test_fw_config->name); test_fw_config->name = NULL; } /* * XXX: move to kstrncpy() once merged. * * Users should use kfree_const() when freeing these. */ static int __kstrncpy(char **dst, const char *name, size_t count, gfp_t gfp) { *dst = kstrndup(name, count, gfp); if (!*dst) return -ENOSPC; return count; } static int __test_firmware_config_init(void) { int ret; ret = __kstrncpy(&test_fw_config->name, TEST_FIRMWARE_NAME, strlen(TEST_FIRMWARE_NAME), GFP_KERNEL); if (ret < 0) goto out; test_fw_config->num_requests = TEST_FIRMWARE_NUM_REQS; test_fw_config->send_uevent = true; test_fw_config->sync_direct = false; test_fw_config->req_firmware = request_firmware; test_fw_config->test_result = 0; test_fw_config->reqs = NULL; return 0; out: __test_firmware_config_free(); return ret; } static ssize_t reset_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { int ret; mutex_lock(&test_fw_mutex); __test_firmware_config_free(); ret = __test_firmware_config_init(); if (ret < 0) { ret = -ENOMEM; pr_err("could not alloc settings for config trigger: %d\n", ret); goto out; } pr_info("reset\n"); ret = count; out: mutex_unlock(&test_fw_mutex); return ret; } static DEVICE_ATTR_WO(reset); static ssize_t config_show(struct device *dev, struct device_attribute *attr, char *buf) { int len = 0; mutex_lock(&test_fw_mutex); len += scnprintf(buf, PAGE_SIZE - len, "Custom trigger configuration for: %s\n", dev_name(dev)); if (test_fw_config->name) len += scnprintf(buf+len, PAGE_SIZE - len, "name:\t%s\n", test_fw_config->name); else len += scnprintf(buf+len, PAGE_SIZE - len, "name:\tEMTPY\n"); len += scnprintf(buf+len, PAGE_SIZE - len, "num_requests:\t%u\n", test_fw_config->num_requests); len += scnprintf(buf+len, PAGE_SIZE - len, "send_uevent:\t\t%s\n", test_fw_config->send_uevent ? "FW_ACTION_HOTPLUG" : "FW_ACTION_NOHOTPLUG"); len += scnprintf(buf+len, PAGE_SIZE - len, "sync_direct:\t\t%s\n", test_fw_config->sync_direct ? "true" : "false"); len += scnprintf(buf+len, PAGE_SIZE - len, "read_fw_idx:\t%u\n", test_fw_config->read_fw_idx); mutex_unlock(&test_fw_mutex); return len; } static DEVICE_ATTR_RO(config); static ssize_t config_name_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { int ret; mutex_lock(&test_fw_mutex); kfree_const(test_fw_config->name); ret = __kstrncpy(&test_fw_config->name, buf, count, GFP_KERNEL); mutex_unlock(&test_fw_mutex); return ret; } /* * As per sysfs_kf_seq_show() the buf is max PAGE_SIZE. */ static ssize_t config_test_show_str(char *dst, char *src) { int len; mutex_lock(&test_fw_mutex); len = snprintf(dst, PAGE_SIZE, "%s\n", src); mutex_unlock(&test_fw_mutex); return len; } static int test_dev_config_update_bool(const char *buf, size_t size, bool *cfg) { int ret; mutex_lock(&test_fw_mutex); if (strtobool(buf, cfg) < 0) ret = -EINVAL; else ret = size; mutex_unlock(&test_fw_mutex); return ret; } static ssize_t test_dev_config_show_bool(char *buf, bool config) { bool val; mutex_lock(&test_fw_mutex); val = config; mutex_unlock(&test_fw_mutex); return snprintf(buf, PAGE_SIZE, "%d\n", val); } static ssize_t test_dev_config_show_int(char *buf, int cfg) { int val; mutex_lock(&test_fw_mutex); val = cfg; mutex_unlock(&test_fw_mutex); return snprintf(buf, PAGE_SIZE, "%d\n", val); } static int test_dev_config_update_u8(const char *buf, size_t size, u8 *cfg) { int ret; long new; ret = kstrtol(buf, 10, &new); if (ret) return ret; if (new > U8_MAX) return -EINVAL; mutex_lock(&test_fw_mutex); *(u8 *)cfg = new; mutex_unlock(&test_fw_mutex); /* Always return full write size even if we didn't consume all */ return size; } static ssize_t test_dev_config_show_u8(char *buf, u8 cfg) { u8 val; mutex_lock(&test_fw_mutex); val = cfg; mutex_unlock(&test_fw_mutex); return snprintf(buf, PAGE_SIZE, "%u\n", val); } static ssize_t config_name_show(struct device *dev, struct device_attribute *attr, char *buf) { return config_test_show_str(buf, test_fw_config->name); } static DEVICE_ATTR_RW(config_name); static ssize_t config_num_requests_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { int rc; mutex_lock(&test_fw_mutex); if (test_fw_config->reqs) { pr_err("Must call release_all_firmware prior to changing config\n"); rc = -EINVAL; mutex_unlock(&test_fw_mutex); goto out; } mutex_unlock(&test_fw_mutex); rc = test_dev_config_update_u8(buf, count, &test_fw_config->num_requests); out: return rc; } static ssize_t config_num_requests_show(struct device *dev, struct device_attribute *attr, char *buf) { return test_dev_config_show_u8(buf, test_fw_config->num_requests); } static DEVICE_ATTR_RW(config_num_requests); static ssize_t config_sync_direct_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { int rc = test_dev_config_update_bool(buf, count, &test_fw_config->sync_direct); if (rc == count) test_fw_config->req_firmware = test_fw_config->sync_direct ? request_firmware_direct : request_firmware; return rc; } static ssize_t config_sync_direct_show(struct device *dev, struct device_attribute *attr, char *buf) { return test_dev_config_show_bool(buf, test_fw_config->sync_direct); } static DEVICE_ATTR_RW(config_sync_direct); static ssize_t config_send_uevent_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { return test_dev_config_update_bool(buf, count, &test_fw_config->send_uevent); } static ssize_t config_send_uevent_show(struct device *dev, struct device_attribute *attr, char *buf) { return test_dev_config_show_bool(buf, test_fw_config->send_uevent); } static DEVICE_ATTR_RW(config_send_uevent); static ssize_t config_read_fw_idx_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { return test_dev_config_update_u8(buf, count, &test_fw_config->read_fw_idx); } static ssize_t config_read_fw_idx_show(struct device *dev, struct device_attribute *attr, char *buf) { return test_dev_config_show_u8(buf, test_fw_config->read_fw_idx); } static DEVICE_ATTR_RW(config_read_fw_idx); static ssize_t trigger_request_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { int rc; char *name; name = kstrndup(buf, count, GFP_KERNEL); if (!name) return -ENOSPC; pr_info("loading '%s'\n", name); mutex_lock(&test_fw_mutex); release_firmware(test_firmware); test_firmware = NULL; rc = request_firmware(&test_firmware, name, dev); if (rc) { pr_info("load of '%s' failed: %d\n", name, rc); goto out; } pr_info("loaded: %zu\n", test_firmware->size); rc = count; out: mutex_unlock(&test_fw_mutex); kfree(name); return rc; } static DEVICE_ATTR_WO(trigger_request); static DECLARE_COMPLETION(async_fw_done); static void trigger_async_request_cb(const struct firmware *fw, void *context) { test_firmware = fw; complete(&async_fw_done); } static ssize_t trigger_async_request_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { int rc; char *name; name = kstrndup(buf, count, GFP_KERNEL); if (!name) return -ENOSPC; pr_info("loading '%s'\n", name); mutex_lock(&test_fw_mutex); release_firmware(test_firmware); test_firmware = NULL; rc = request_firmware_nowait(THIS_MODULE, 1, name, dev, GFP_KERNEL, NULL, trigger_async_request_cb); if (rc) { pr_info("async load of '%s' failed: %d\n", name, rc); kfree(name); goto out; } /* Free 'name' ASAP, to test for race conditions */ kfree(name); wait_for_completion(&async_fw_done); if (test_firmware) { pr_info("loaded: %zu\n", test_firmware->size); rc = count; } else { pr_err("failed to async load firmware\n"); rc = -ENODEV; } out: mutex_unlock(&test_fw_mutex); return rc; } static DEVICE_ATTR_WO(trigger_async_request); static ssize_t trigger_custom_fallback_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { int rc; char *name; name = kstrndup(buf, count, GFP_KERNEL); if (!name) return -ENOSPC; pr_info("loading '%s' using custom fallback mechanism\n", name); mutex_lock(&test_fw_mutex); release_firmware(test_firmware); test_firmware = NULL; rc = request_firmware_nowait(THIS_MODULE, FW_ACTION_NOHOTPLUG, name, dev, GFP_KERNEL, NULL, trigger_async_request_cb); if (rc) { pr_info("async load of '%s' failed: %d\n", name, rc); kfree(name); goto out; } /* Free 'name' ASAP, to test for race conditions */ kfree(name); wait_for_completion(&async_fw_done); if (test_firmware) { pr_info("loaded: %zu\n", test_firmware->size); rc = count; } else { pr_err("failed to async load firmware\n"); rc = -ENODEV; } out: mutex_unlock(&test_fw_mutex); return rc; } static DEVICE_ATTR_WO(trigger_custom_fallback); static int test_fw_run_batch_request(void *data) { struct test_batched_req *req = data; if (!req) { test_fw_config->test_result = -EINVAL; return -EINVAL; } req->rc = test_fw_config->req_firmware(&req->fw, req->name, req->dev); if (req->rc) { pr_info("#%u: batched sync load failed: %d\n", req->idx, req->rc); if (!test_fw_config->test_result) test_fw_config->test_result = req->rc; } else if (req->fw) { req->sent = true; pr_info("#%u: batched sync loaded %zu\n", req->idx, req->fw->size); } complete(&req->completion); req->task = NULL; return 0; } /* * We use a kthread as otherwise the kernel serializes all our sync requests * and we would not be able to mimic batched requests on a sync call. Batched * requests on a sync call can for instance happen on a device driver when * multiple cards are used and firmware loading happens outside of probe. */ static ssize_t trigger_batched_requests_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { struct test_batched_req *req; int rc; u8 i; mutex_lock(&test_fw_mutex); test_fw_config->reqs = vzalloc(sizeof(struct test_batched_req) * test_fw_config->num_requests * 2); if (!test_fw_config->reqs) { rc = -ENOMEM; goto out_unlock; } pr_info("batched sync firmware loading '%s' %u times\n", test_fw_config->name, test_fw_config->num_requests); for (i = 0; i < test_fw_config->num_requests; i++) { req = &test_fw_config->reqs[i]; if (!req) { WARN_ON(1); rc = -ENOMEM; goto out_bail; } req->fw = NULL; req->idx = i; req->name = test_fw_config->name; req->dev = dev; init_completion(&req->completion); req->task = kthread_run(test_fw_run_batch_request, req, "%s-%u", KBUILD_MODNAME, req->idx); if (!req->task || IS_ERR(req->task)) { pr_err("Setting up thread %u failed\n", req->idx); req->task = NULL; rc = -ENOMEM; goto out_bail; } } rc = count; /* * We require an explicit release to enable more time and delay of * calling release_firmware() to improve our chances of forcing a * batched request. If we instead called release_firmware() right away * then we might miss on an opportunity of having a successful firmware * request pass on the opportunity to be come a batched request. */ out_bail: for (i = 0; i < test_fw_config->num_requests; i++) { req = &test_fw_config->reqs[i]; if (req->task || req->sent) wait_for_completion(&req->completion); } /* Override any worker error if we had a general setup error */ if (rc < 0) test_fw_config->test_result = rc; out_unlock: mutex_unlock(&test_fw_mutex); return rc; } static DEVICE_ATTR_WO(trigger_batched_requests); /* * We wait for each callback to return with the lock held, no need to lock here */ static void trigger_batched_cb(const struct firmware *fw, void *context) { struct test_batched_req *req = context; if (!req) { test_fw_config->test_result = -EINVAL; return; } /* forces *some* batched requests to queue up */ if (!req->idx) ssleep(2); req->fw = fw; /* * Unfortunately the firmware API gives us nothing other than a null FW * if the firmware was not found on async requests. Best we can do is * just assume -ENOENT. A better API would pass the actual return * value to the callback. */ if (!fw && !test_fw_config->test_result) test_fw_config->test_result = -ENOENT; complete(&req->completion); } static ssize_t trigger_batched_requests_async_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { struct test_batched_req *req; bool send_uevent; int rc; u8 i; mutex_lock(&test_fw_mutex); test_fw_config->reqs = vzalloc(sizeof(struct test_batched_req) * test_fw_config->num_requests * 2); if (!test_fw_config->reqs) { rc = -ENOMEM; goto out; } pr_info("batched loading '%s' custom fallback mechanism %u times\n", test_fw_config->name, test_fw_config->num_requests); send_uevent = test_fw_config->send_uevent ? FW_ACTION_HOTPLUG : FW_ACTION_NOHOTPLUG; for (i = 0; i < test_fw_config->num_requests; i++) { req = &test_fw_config->reqs[i]; if (!req) { WARN_ON(1); goto out_bail; } req->name = test_fw_config->name; req->fw = NULL; req->idx = i; init_completion(&req->completion); rc = request_firmware_nowait(THIS_MODULE, send_uevent, req->name, dev, GFP_KERNEL, req, trigger_batched_cb); if (rc) { pr_info("#%u: batched async load failed setup: %d\n", i, rc); req->rc = rc; goto out_bail; } else req->sent = true; } rc = count; out_bail: /* * We require an explicit release to enable more time and delay of * calling release_firmware() to improve our chances of forcing a * batched request. If we instead called release_firmware() right away * then we might miss on an opportunity of having a successful firmware * request pass on the opportunity to be come a batched request. */ for (i = 0; i < test_fw_config->num_requests; i++) { req = &test_fw_config->reqs[i]; if (req->sent) wait_for_completion(&req->completion); } /* Override any worker error if we had a general setup error */ if (rc < 0) test_fw_config->test_result = rc; out: mutex_unlock(&test_fw_mutex); return rc; } static DEVICE_ATTR_WO(trigger_batched_requests_async); static ssize_t test_result_show(struct device *dev, struct device_attribute *attr, char *buf) { return test_dev_config_show_int(buf, test_fw_config->test_result); } static DEVICE_ATTR_RO(test_result); static ssize_t release_all_firmware_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { test_release_all_firmware(); return count; } static DEVICE_ATTR_WO(release_all_firmware); static ssize_t read_firmware_show(struct device *dev, struct device_attribute *attr, char *buf) { struct test_batched_req *req; u8 idx; ssize_t rc = 0; mutex_lock(&test_fw_mutex); idx = test_fw_config->read_fw_idx; if (idx >= test_fw_config->num_requests) { rc = -ERANGE; goto out; } if (!test_fw_config->reqs) { rc = -EINVAL; goto out; } req = &test_fw_config->reqs[idx]; if (!req->fw) { pr_err("#%u: failed to async load firmware\n", idx); rc = -ENOENT; goto out; } pr_info("#%u: loaded %zu\n", idx, req->fw->size); if (req->fw->size > PAGE_SIZE) { pr_err("Testing interface must use PAGE_SIZE firmware for now\n"); rc = -EINVAL; goto out; } memcpy(buf, req->fw->data, req->fw->size); rc = req->fw->size; out: mutex_unlock(&test_fw_mutex); return rc; } static DEVICE_ATTR_RO(read_firmware); #define TEST_FW_DEV_ATTR(name) &dev_attr_##name.attr static struct attribute *test_dev_attrs[] = { TEST_FW_DEV_ATTR(reset), TEST_FW_DEV_ATTR(config), TEST_FW_DEV_ATTR(config_name), TEST_FW_DEV_ATTR(config_num_requests), TEST_FW_DEV_ATTR(config_sync_direct), TEST_FW_DEV_ATTR(config_send_uevent), TEST_FW_DEV_ATTR(config_read_fw_idx), /* These don't use the config at all - they could be ported! */ TEST_FW_DEV_ATTR(trigger_request), TEST_FW_DEV_ATTR(trigger_async_request), TEST_FW_DEV_ATTR(trigger_custom_fallback), /* These use the config and can use the test_result */ TEST_FW_DEV_ATTR(trigger_batched_requests), TEST_FW_DEV_ATTR(trigger_batched_requests_async), TEST_FW_DEV_ATTR(release_all_firmware), TEST_FW_DEV_ATTR(test_result), TEST_FW_DEV_ATTR(read_firmware), NULL, }; ATTRIBUTE_GROUPS(test_dev); static struct miscdevice test_fw_misc_device = { .minor = MISC_DYNAMIC_MINOR, .name = "test_firmware", .fops = &test_fw_fops, .groups = test_dev_groups, }; static int __init test_firmware_init(void) { int rc; test_fw_config = kzalloc(sizeof(struct test_config), GFP_KERNEL); if (!test_fw_config) return -ENOMEM; rc = __test_firmware_config_init(); if (rc) return rc; rc = misc_register(&test_fw_misc_device); if (rc) { kfree(test_fw_config); pr_err("could not register misc device: %d\n", rc); return rc; } pr_warn("interface ready\n"); return 0; } module_init(test_firmware_init); static void __exit test_firmware_exit(void) { mutex_lock(&test_fw_mutex); release_firmware(test_firmware); misc_deregister(&test_fw_misc_device); __test_firmware_config_free(); kfree(test_fw_config); mutex_unlock(&test_fw_mutex); pr_warn("removed interface\n"); } module_exit(test_firmware_exit); MODULE_AUTHOR("Kees Cook <[email protected]>"); MODULE_LICENSE("GPL");
__label__pos
0.998656
Briefly explained – Robotic Process Automation (RPA) https://open.spotify.com/episode/2LIdxXNbxoEPr1lYuenV2M Overview: Robotic Process Automation Robotic Process Automation often leads to confusion because there is no visible robot to perform actions. It is, on the other hand, software that takes over tasks previously performed by humans. So it is not to be confused with physical robots, as is known from production. Nevertheless, Robotic Process Automation offers many business starting points to automate tasks, save costs, and become more efficient. Here we show you what Robotic Process Automation is, what you should consider before implementation, where there are frequent applications and whether it is just a short-term trend. What is Robotic Process Automation? Robotic Process Automation is also often abbreviated with RPA or robot-controlled process automation. It is defined as automated processing of structured processes in the company by digital software bots. The RPA dates back to the 2000s and was linked to the evolution of three key technologies required for this purpose. Screen scraping, workflow automation and artificial intelligence form the basis for the development of RPA technology. The technology was also inspired by the robots from the industry, which were already known at the time, and can carry out production tasks independently and completely and not only semi-automated. Artificial intelligence and machine learning now allow you to automate routine tasks that previously only could be done by humans. Even very large amounts of data can be managed by RPA. The software bots or robots mimic the work of a human employee and can log in and out of applications on their own, enter and process data, calculate tasks and make transactions. The RPA software is over-stored in a company’s IT infrastructure and can therefore be implemented without having to make a change in the existing systems. This is fast and efficient. The software can take over almost any process and thus offers numerous and creative applications. Robotic Process Automation offers companies many advantages and is a central component of digitalization. Among other things, customer service can be improved and accelerated, processes become more efficient, costs for manual and repetitive tasks are saved, and overall productivity can be increased. Quality also remains uniformly high and is not subject to day-to-day fluctuations or human error, and the control and compliance of standards can be automated and thus increased. All work steps can be traced back by the software, should there be a problem, and be carried out around the clock, as no working hours have to be observed. RPA software is implemented relatively quickly and therefore saves resources over other automation strategies. Another advantage of RPA over other IT automation solutions is the skilful response to exceptions and changed circumstances. While IT solutions usually cannot handle this, Robotic Process Automation is trained in such cases. The algorithm is constantly learning and this makes it possible to react correctly to new situations. It is also possible to communicate with other systems without the need to interconnect an employee. Exceptions are therefore not simply marked and assigned to an employee for further processing, but the RPA technology searches independently for the missing information, even across systems. In which areas is Robotic Process Automation used and what do I have to consider before integration? The RPA software is most commonly used for manual, time-consuming and repetitive work. This can be the case, for example, in the office or in production. However, in order to operate Robotic Process Automation, strict rules must be in place for the operation. If this is not yet the case, processes should first be defined as routine processes and the exact steps should be defined. This does not have to be done centrally for the entire company, but offers individual departments individual opportunities to carry out their processes more efficiently. Due to individuality, flexibility does not have to suffer and can be maintained despite fixed routine processes. The objective of integrating RPA technology should also be maintained before it is put into practice. Typical applications of Robotic Process Automation are customer service, accounting, healthcare, human resources departments, financial services and supply chain management. In customer service, the technology automates the tasks of the call center. For example, documents can be uploaded automatically, email signatures can be checked and information submitted by the customer can be checked automatically for completeness and how to proceed with it. In accounting, Robotic Process Automation can handle general and operational accounting, budgeting, and transaction reporting. In healthcare, medical records, reports, and billing can be managed by RPA. The technology is also an efficient solution for HR departments and can take on time-consuming tasks such as time recording and management of employee information. In the area of financial services, for example, account openings and closures are carried out by Robotic Process Automation. Another important area of application is supply chain management. Here, inventories can be monitored, shipments can be tracked and payments and orders can be processed automatically. So there are already numerous effective applications for robotic process automation and in the future many more will probably be added with a further development of the technology. Conclusion on Robotic Process Automation Robotic Process Automation can make companies more successful and simplify their day-to-day work. However, the software should not be rushed and integrated headless just to be there. First, we really should take the time to analyze resource-intensive processes and consider where automation would make sense. The required workflows must be defined as fixed routines and rules in order for RPA technology to adopt them. You should also answer the question of whether you want fully automated or only semi-automated processes for your company. If the corresponding preparatory work is carried out, Robotic Process Automation can then lead to great cost savings, efficiency and productivity. What is certain is that the RBA is not only a trend, but will continue to accompany us for a long time to come. By 2025, 140 million full-time positions worldwide are to be replaced by the corresponding software, and by 2024 the RPA market is already expected to reach five million dollars. As a result, companies and employees will have many new opportunities, and tasks and responsibilities of previous job descriptions are likely to adapt and change across all levels of the hierarchy. So it‘s worth keeping up to date and thinking early on in which areas you could use this technology in a meaningful way. Related Articles LEAVE A REPLY Please enter your comment! Please enter your name here Stay Connected 22,952FansLike 3,332FollowersFollow 0SubscribersSubscribe - Advertisement -rec300 Latest Articles
__label__pos
0.555987
dongzhang2150 2016-09-25 15:51 浏览 60 已采纳 使用PHP Simple HTML DOM Parser提取HTML纯 Can you with "php simple html dom" take a piece of HTML and NOT only the content? I have trie with: foreach ($html->find('div.info-bar') as $infop) { $info = $infop->plaintext; } Unfortunately, the output is: Level 24 Trophies 4201 Bronze 2725 Silver 1057 Gold 341 Platinum 78 Level 24 66 66 % While I would like to extract the pure HTML.. This is my code: include('simple_html_dom.php'); $html = file_get_html('https://my.playstation.com/obaidmiz04'); foreach ($html->find('div.info-bar') as $infop) { $info = $infop->plaintext; } echo $info; • 写回答 1条回答 默认 最新 • doujianglin6704 2016-09-25 17:25 关注 $escapedHtmlChars = ""; $htmlElements = ""; $html = file_get_html('https://my.playstation.com/obaidmiz04'); foreach ($html->find('div.info-bar') as $infop) { //You can see html characters in text //This shows you html codes, not for using $escapedHtmlChars .= htmlspecialchars($infop); //You can see html elements $htmlElements .= $infop; //You can see only text in selected element $plainText .= $infop->plaintext; } echo $plainText; echo "<br /> <br />"; echo $escapedHtmlChars; echo "<br /> <br />"; echo $htmlElements; plaintext method returns only text in selected element. 本回答被题主选为最佳回答 , 对您是否有帮助呢? 评论 报告相同问题? 悬赏问题 • ¥15 streamingtool • ¥15 MATLAB图像问题 • ¥20 树莓派5做人脸情感识别与反馈系统 • ¥15 selenium 控制 chrome-for-testing 在 Linux 环境下报错 SessionNotCreatedException • ¥15 使用pyodbc操作SQL数据库 • ¥15 MATLAB实现下列 • ¥30 mininet可视化打不开.mn文件 • ¥50 C# 全屏打开Edge浏览器 • ¥80 WEBPACK性能优化 • ¥30 python拟合回归分析
__label__pos
0.829702
Probability and statistics questions Probability and statistics questions, This quiz will review the fundamentals of probability and statistics you will be asked to find the mean, median, mode, and range of a set of data you will be asked. Probability and statistics questions, This quiz will review the fundamentals of probability and statistics you will be asked to find the mean, median, mode, and range of a set of data you will be asked. Statistics & probability a course in statistics and probability beyond s39 use simulation methods to answer questions about probability models that are too. Mit introduction to probability and statistics practice tests with solutions emory univ math107 introduction to probability and statistics - exams and solutions. Sample questions on probability and statistics next: question 1 sample questions on probability and statistics. This section provides the course exams with introduction to probability and statistics exam questions and. Statistics & probability ccssmathcontent6spa1 recognize a statistical question as one that anticipates variability in the data related to the question and. Test and improve your knowledge of probability and statistics with fun multiple choice exams you can take online with studycom. Originally posted by that'sallfolks but again at the point of the choice, it is either your card or his card that is 50/50 he necessarily was going. Free tutorials cover statistics, probability, and matrix algebra strong focus on ap statistics written and video lessons online calculators. Probability questions with solutions solution to question 3: a probability is always greater than or equal to 0 more references on elementary statistics and. Statistics 8: chapters 7 to 10, sample multiple choice questions 1 if two events (both with probability greater than 0) are mutually exclusive, then. High school statistics and probability common core sample test version 1 q5 jill had to do 100 questions in 120 minutes after 60 minutes, she was. Elementary probability and statistics student name and id number final exam june 6, 2011 instructor: bj˝rn kjos-hanssen in the questions below. The revised probability and statistics for elementary and middle school probability background the big ideas of statistics posing questions session 1 topic. Sat math skill review: probability & statistics here, we’ll review a few of the concepts that you will need to be successful on these questions probability. Learn statistics and probability for free—everything you'd want to know about descriptive and inferential statistics full curriculum of exercises and videos. Math1005 quizzes you are here: one outcome has not been observed then the probability that it will occur in the next for questions or comments please contact. This course provides an elementary introduction to probability and statistics with applications topics include: basic combinatorics, random variables, probability. Probability and statistics problems completely solved in detail indexed to find topics easily based on ap statistics exam questions. Probability and statistics questions - instead of concerning about term paper writing find the needed help here quick and trustworthy writings from industry leading. Data analysis, statistics, and probability mastery 400 the powerscore sat math bible data analysis questions use diagrams, figures, tables, or graphs in conjunction with. Find the probability of correctly answering the first 4 questions on a multiple choice test using random guessing each question has 3 possible answersexplain. Introduction to probability and statistics introduction to probability and statistics the presumed probability of success in the experiments in question. This is the aptitude questions and answers section on probability with explanation for various interview, competitive examination and entrance test solved examples. Probability interview questions actually this question was about the statistics there is a constant probability of 08 of having a successful. Most popular questions you can find on probability questions what's a good way to study probability and statistics for a google product manager onsite interview. Probability how likely something is to happen many events can't be predicted with total certainty the best we can say is how likely they are to happen, using the. Probability and statistics questions Rated 3/5 based on 24 review xehomeworkesna.tlwsd.info
__label__pos
0.538276
Ticket #2834: dllMain.c File dllMain.c, 505 bytes (added by lewissandy, 9 years ago) Line  1#include <windows.h> 2#include <Rts.h> 3 4extern void __stginit_Adder(void); 5 6static char* args[] = { "ghcDll", NULL }; 7                       /* N.B. argv arrays must end with NULL */ 8BOOL 9STDCALL 10DllMain 11   ( HANDLE hModule 12   , DWORD reason 13   , void* reserved 14   ) 15{ 16  if (reason == DLL_PROCESS_ATTACH) { 17      /* By now, the RTS DLL should have been hoisted in, but we need to start it up. */ 18      startupHaskell(1, args, __stginit_Adder); 19      return TRUE; 20  } 21  return TRUE; 22}
__label__pos
0.987137
How to use LocatorLocatorOptions class of Microsoft.Playwright package Best Playwright-dotnet code snippet using Microsoft.Playwright.LocatorLocatorOptions ILocator.cs Source:ILocator.cs Github copy Full Screen ...428 /// A selector to use when resolving DOM element. See <a href="https://playwright.dev/dotnet/docs/selectors">working429 /// with selectors</a> for more details.430 /// </param>431 /// <param name="options">Call options</param>432 ILocator Locator(string selector, LocatorLocatorOptions? options = default);433 /// <summary>434 /// <para>435 /// Returns locator to the n-th matching element. It's zero based, <c>nth(0)</c> selects436 /// the first element.437 /// </para>438 /// </summary>439 /// <param name="index">440 /// </param>441 ILocator Nth(int index);442 /// <summary><para>A page this locator belongs to.</para></summary>443 IPage Page { get; }444 /// <summary>445 /// <para>Focuses the element, and then uses <see cref="IKeyboard.DownAsync"/> and <see cref="IKeyboard.UpAsync"/>.</para>446 /// <para>... Full Screen Full Screen Locator.cs Source:Locator.cs Github copy Full Screen ...34 internal class Locator : ILocator35 {36 internal readonly Frame _frame;37 internal readonly string _selector;38 private readonly LocatorLocatorOptions _options;39 public Locator(Frame parent, string selector, LocatorLocatorOptions options = null)40 {41 _frame = parent;42 _selector = selector;43 _options = options;44 if (options?.HasTextRegex != null)45 {46 _selector += $" >> :scope:text-matches({options.HasTextRegex.ToString().EscapeWithQuotes("\"")}, {options.HasTextRegex.Options.GetInlineFlags().EscapeWithQuotes("\"")})";47 }48 if (options?.HasTextString != null)49 {50 _selector += $" >> :scope:has-text({options.HasTextString.EscapeWithQuotes("\"")})";51 }52 if (options?.Has != null)53 {54 var has = (Locator)options.Has;55 if (has._frame != _frame)56 {57 throw new ArgumentException("Inner \"has\" locator must belong to the same frame.");58 }59 _selector += " >> has=" + JsonSerializer.Serialize(has._selector);60 }61 }62 public ILocator First => new Locator(_frame, $"{_selector} >> nth=0");63 public ILocator Last => new Locator(_frame, $"{_selector} >> nth=-1");64 IPage ILocator.Page => _frame.Page;65 public async Task<IReadOnlyList<string>> AllInnerTextsAsync()66 => await EvaluateAllAsync<string[]>("ee => ee.map(e => e.innerText)").ConfigureAwait(false);67 public async Task<IReadOnlyList<string>> AllTextContentsAsync()68 => await EvaluateAllAsync<string[]>("ee => ee.map(e => e.textContent || '')").ConfigureAwait(false);69 public async Task<LocatorBoundingBoxResult> BoundingBoxAsync(LocatorBoundingBoxOptions options = null)70 => await WithElementAsync(71 async (h, _) =>72 {73 var bb = await h.BoundingBoxAsync().ConfigureAwait(false);74 if (bb == null)75 {76 return null;77 }78 return new LocatorBoundingBoxResult()79 {80 Height = bb.Height,81 Width = bb.Width,82 X = bb.X,83 Y = bb.Y,84 };85 },86 options).ConfigureAwait(false);87 public Task CheckAsync(LocatorCheckOptions options = null)88 => _frame.CheckAsync(89 _selector,90 ConvertOptions<FrameCheckOptions>(options));91 public Task ClickAsync(LocatorClickOptions options = null)92 => _frame.ClickAsync(93 _selector,94 ConvertOptions<FrameClickOptions>(options));95 public Task SetCheckedAsync(bool checkedState, LocatorSetCheckedOptions options = null)96 => checkedState ?97 CheckAsync(ConvertOptions<LocatorCheckOptions>(options))98 : UncheckAsync(ConvertOptions<LocatorUncheckOptions>(options));99 public Task<int> CountAsync()100 => _frame.QueryCountAsync(_selector);101 public Task DblClickAsync(LocatorDblClickOptions options = null)102 => _frame.DblClickAsync(_selector, ConvertOptions<FrameDblClickOptions>(options));103 public Task DispatchEventAsync(string type, object eventInit = null, LocatorDispatchEventOptions options = null)104 => _frame.DispatchEventAsync(_selector, type, eventInit, ConvertOptions<FrameDispatchEventOptions>(options));105 public Task DragToAsync(ILocator target, LocatorDragToOptions options = null)106 => _frame.DragAndDropAsync(_selector, ((Locator)target)._selector, ConvertOptions<FrameDragAndDropOptions>(options));107 public async Task<IElementHandle> ElementHandleAsync(LocatorElementHandleOptions options = null)108 => await _frame.WaitForSelectorAsync(109 _selector,110 ConvertOptions<FrameWaitForSelectorOptions>(options)).ConfigureAwait(false);111 public Task<IReadOnlyList<IElementHandle>> ElementHandlesAsync()112 => _frame.QuerySelectorAllAsync(_selector);113 public Task<T> EvaluateAllAsync<T>(string expression, object arg = null)114 => _frame.EvalOnSelectorAllAsync<T>(_selector, expression, arg);115 public Task<JsonElement?> EvaluateAsync(string expression, object arg = null, LocatorEvaluateOptions options = null)116 => EvaluateAsync<JsonElement?>(expression, arg, options);117 public Task<T> EvaluateAsync<T>(string expression, object arg = null, LocatorEvaluateOptions options = null)118 => _frame.EvalOnSelectorAsync<T>(_selector, expression, arg, ConvertOptions<FrameEvalOnSelectorOptions>(options));119 public async Task<IJSHandle> EvaluateHandleAsync(string expression, object arg = null, LocatorEvaluateHandleOptions options = null)120 => await WithElementAsync(async (e, _) => await e.EvaluateHandleAsync(expression, arg).ConfigureAwait(false), options).ConfigureAwait(false);121 public async Task FillAsync(string value, LocatorFillOptions options = null)122 => await _frame.FillAsync(_selector, value, ConvertOptions<FrameFillOptions>(options)).ConfigureAwait(false);123 public Task FocusAsync(LocatorFocusOptions options = null)124 => _frame.FocusAsync(_selector, ConvertOptions<FrameFocusOptions>(options));125 IFrameLocator ILocator.FrameLocator(string selector) =>126 new FrameLocator(_frame, $"{_selector} >> {selector}");127 public Task<string> GetAttributeAsync(string name, LocatorGetAttributeOptions options = null)128 => _frame.GetAttributeAsync(_selector, name, ConvertOptions<FrameGetAttributeOptions>(options));129 public Task HoverAsync(LocatorHoverOptions options = null)130 => _frame.HoverAsync(_selector, ConvertOptions<FrameHoverOptions>(options));131 public Task<string> InnerHTMLAsync(LocatorInnerHTMLOptions options = null)132 => _frame.InnerHTMLAsync(_selector, ConvertOptions<FrameInnerHTMLOptions>(options));133 public Task<string> InnerTextAsync(LocatorInnerTextOptions options = null)134 => _frame.InnerTextAsync(_selector, ConvertOptions<FrameInnerTextOptions>(options));135 public Task<string> InputValueAsync(LocatorInputValueOptions options = null)136 => _frame.InputValueAsync(_selector, ConvertOptions<FrameInputValueOptions>(options));137 public Task<bool> IsCheckedAsync(LocatorIsCheckedOptions options = null)138 => _frame.IsCheckedAsync(_selector, ConvertOptions<FrameIsCheckedOptions>(options));139 public Task<bool> IsDisabledAsync(LocatorIsDisabledOptions options = null)140 => _frame.IsDisabledAsync(_selector, ConvertOptions<FrameIsDisabledOptions>(options));141 public Task<bool> IsEditableAsync(LocatorIsEditableOptions options = null)142 => _frame.IsEditableAsync(_selector, ConvertOptions<FrameIsEditableOptions>(options));143 public Task<bool> IsEnabledAsync(LocatorIsEnabledOptions options = null)144 => _frame.IsEnabledAsync(_selector, ConvertOptions<FrameIsEnabledOptions>(options));145 public Task<bool> IsHiddenAsync(LocatorIsHiddenOptions options = null)146 => _frame.IsHiddenAsync(_selector, ConvertOptions<FrameIsHiddenOptions>(options));147 public Task<bool> IsVisibleAsync(LocatorIsVisibleOptions options = null)148 => _frame.IsVisibleAsync(_selector, ConvertOptions<FrameIsVisibleOptions>(options));149 public ILocator Nth(int index)150 => new Locator(_frame, $"{_selector} >> nth={index}");151 public Task PressAsync(string key, LocatorPressOptions options = null)152 => _frame.PressAsync(_selector, key, ConvertOptions<FramePressOptions>(options));153 public Task<byte[]> ScreenshotAsync(LocatorScreenshotOptions options = null)154 => WithElementAsync(async (h, o) => await h.ScreenshotAsync(ConvertOptions<ElementHandleScreenshotOptions>(o)).ConfigureAwait(false), options);155 public Task ScrollIntoViewIfNeededAsync(LocatorScrollIntoViewIfNeededOptions options = null)156 => WithElementAsync(async (h, o) => await h.ScrollIntoViewIfNeededAsync(ConvertOptions<ElementHandleScrollIntoViewIfNeededOptions>(o)).ConfigureAwait(false), options);157 public Task<IReadOnlyList<string>> SelectOptionAsync(string values, LocatorSelectOptionOptions options = null)158 => _frame.SelectOptionAsync(_selector, values, ConvertOptions<FrameSelectOptionOptions>(options));159 public Task<IReadOnlyList<string>> SelectOptionAsync(IElementHandle values, LocatorSelectOptionOptions options = null)160 => _frame.SelectOptionAsync(_selector, values, ConvertOptions<FrameSelectOptionOptions>(options));161 public Task<IReadOnlyList<string>> SelectOptionAsync(IEnumerable<string> values, LocatorSelectOptionOptions options = null)162 => _frame.SelectOptionAsync(_selector, values, ConvertOptions<FrameSelectOptionOptions>(options));163 public Task<IReadOnlyList<string>> SelectOptionAsync(SelectOptionValue values, LocatorSelectOptionOptions options = null)164 => _frame.SelectOptionAsync(_selector, values, ConvertOptions<FrameSelectOptionOptions>(options));165 public Task<IReadOnlyList<string>> SelectOptionAsync(IEnumerable<IElementHandle> values, LocatorSelectOptionOptions options = null)166 => _frame.SelectOptionAsync(_selector, values, ConvertOptions<FrameSelectOptionOptions>(options));167 public Task<IReadOnlyList<string>> SelectOptionAsync(IEnumerable<SelectOptionValue> values, LocatorSelectOptionOptions options = null)168 => _frame.SelectOptionAsync(_selector, values, ConvertOptions<FrameSelectOptionOptions>(options));169 public Task SelectTextAsync(LocatorSelectTextOptions options = null)170 => WithElementAsync((h, o) => h.SelectTextAsync(ConvertOptions<ElementHandleSelectTextOptions>(o)), options);171 public Task SetInputFilesAsync(string files, LocatorSetInputFilesOptions options = null)172 => _frame.SetInputFilesAsync(_selector, files, ConvertOptions<FrameSetInputFilesOptions>(options));173 public Task SetInputFilesAsync(IEnumerable<string> files, LocatorSetInputFilesOptions options = null)174 => _frame.SetInputFilesAsync(_selector, files, ConvertOptions<FrameSetInputFilesOptions>(options));175 public Task SetInputFilesAsync(FilePayload files, LocatorSetInputFilesOptions options = null)176 => _frame.SetInputFilesAsync(_selector, files, ConvertOptions<FrameSetInputFilesOptions>(options));177 public Task SetInputFilesAsync(IEnumerable<FilePayload> files, LocatorSetInputFilesOptions options = null)178 => _frame.SetInputFilesAsync(_selector, files, ConvertOptions<FrameSetInputFilesOptions>(options));179 public Task TapAsync(LocatorTapOptions options = null)180 => _frame.TapAsync(_selector, ConvertOptions<FrameTapOptions>(options));181 public Task<string> TextContentAsync(LocatorTextContentOptions options = null)182 => _frame.TextContentAsync(_selector, ConvertOptions<FrameTextContentOptions>(options));183 public Task TypeAsync(string text, LocatorTypeOptions options = null)184 => _frame.TypeAsync(_selector, text, ConvertOptions<FrameTypeOptions>(options));185 public Task UncheckAsync(LocatorUncheckOptions options = null)186 => _frame.UncheckAsync(_selector, ConvertOptions<FrameUncheckOptions>(options));187 ILocator ILocator.Locator(string selector, LocatorLocatorOptions options)188 => new Locator(_frame, $"{_selector} >> {selector}", options);189 public Task WaitForAsync(LocatorWaitForOptions options = null)190 => _frame.LocatorWaitForAsync(_selector, ConvertOptions<LocatorWaitForOptions>(options));191 internal Task<FrameExpectResult> ExpectAsync(string expression, FrameExpectOptions options = null)192 => _frame.ExpectAsync(193 _selector,194 expression,195 options);196 public override string ToString() => "Locator@" + _selector;197 private T ConvertOptions<T>(object source)198 where T : class, new()199 {200 T target = new();201 var targetType = target.GetType();... Full Screen Full Screen PlaywrightSyncElement.cs Source:PlaywrightSyncElement.cs Github copy Full Screen ...31 /// </summary>32 /// <param name="parent">The parent playwright element</param>33 /// <param name="selector">Sub element selector</param>34 /// <param name="options">Advanced locator options</param>35 public PlaywrightSyncElement(PlaywrightSyncElement parent, string selector, LocatorLocatorOptions? options = null)36 {37 this.ParentLocator = parent.ElementLocator();38 this.Selector = selector;39 this.LocatorOptions = options;40 }41 /// <summary>42 /// Initializes a new instance of the <see cref="PlaywrightSyncElement" /> class43 /// </summary>44 /// <param name="frame">The assoicated playwright frame locator</param>45 /// <param name="selector">Element selector</param>46 public PlaywrightSyncElement(IFrameLocator frame, string selector)47 {48 this.ParentFrameLocator = frame;49 this.Selector = selector;50 }51 /// <summary>52 /// Initializes a new instance of the <see cref="PlaywrightSyncElement" /> class53 /// </summary>54 /// <param name="testObject">The assoicated playwright test object</param>55 /// <param name="selector">Element selector</param>56 /// <param name="options">Advanced locator options</param>57 public PlaywrightSyncElement(IPlaywrightTestObject testObject, string selector, PageLocatorOptions? options = null) : this(testObject.PageDriver.AsyncPage, selector, options)58 {59 }60 /// <summary>61 /// Initializes a new instance of the <see cref="PlaywrightSyncElement" /> class62 /// </summary>63 /// <param name="driver">The assoicated playwright page driver</param>64 /// <param name="selector">Element selector</param>65 /// <param name="options">Advanced locator options</param>66 public PlaywrightSyncElement(PageDriver driver, string selector, PageLocatorOptions? options = null) : this(driver.AsyncPage, selector, options)67 {68 }69 /// <summary>70 /// Gets the parent async page71 /// </summary>72 public IPage? ParentPage { get; private set; }73 /// <summary>74 /// Gets the parent locator75 /// </summary>76 public ILocator? ParentLocator { get; private set; }77 /// <summary>78 /// Gets the parent frame locator79 /// </summary>80 public IFrameLocator? ParentFrameLocator { get; private set; }81 /// <summary>82 /// Gets the page locator options83 /// </summary>84 public PageLocatorOptions? PageOptions { get; private set; }85 /// <summary>86 /// Gets the page locator options87 /// </summary>88 public LocatorLocatorOptions? LocatorOptions { get; private set; }89 /// <summary>90 /// Gets the selector string91 /// </summary>92 public string Selector { get; private set; }93 /// <summary>94 /// ILocator for this element95 /// </summary>96 /// <returns></returns>97 public ILocator ElementLocator()98 {99 if (this.ParentPage != null)100 {101 return this.ParentPage.Locator(Selector, PageOptions);102 }... Full Screen Full Screen LocatorLocatorOptions.cs Source:LocatorLocatorOptions.cs Github copy Full Screen ...35using System.Threading.Tasks;36#nullable enable37namespace Microsoft.Playwright38{39 public class LocatorLocatorOptions40 {41 public LocatorLocatorOptions() { }42 public LocatorLocatorOptions(LocatorLocatorOptions clone)43 {44 if (clone == null)45 {46 return;47 }48 Has = clone.Has;49 HasTextString = clone.HasTextString;50 HasTextRegex = clone.HasTextRegex;51 }52 /// <summary>53 /// <para>54 /// Matches elements containing an element that matches an inner locator. Inner locator55 /// is queried against the outer one. For example, <c>article</c> that has <c>text=Playwright</c>56 /// matches <c>&lt;article&gt;&lt;div&gt;Playwright&lt;/div&gt;&lt;/article&gt;</c>.... Full Screen Full Screen LocatorLocatorOptions Using AI Code Generation copy Full Screen 1using Microsoft.Playwright;2using Microsoft.Playwright.Core;3using Microsoft.Playwright.Helpers;4using Microsoft.Playwright.Transport;5using Microsoft.Playwright.Transport.Channels;6using Microsoft.Playwright.Transport.Protocol;7using System;8using System.Collections.Generic;9using System.Linq;10using System.Text;11using System.Threading;12using System.Threading.Tasks;13{14 {15 static async Task Main(string[] args)16 {17 var playwright = await Playwright.CreateAsync();18 var browser = await playwright.Chromium.LaunchAsync(new BrowserTypeLaunchOptions { Headless = false });19 var page = await browser.NewPageAsync();20 await page.TypeAsync("input[name='q']", "Playwright");21 await page.ClickAsync("input[value='Google Search']");22 {23 {24 }25 };26 var locator = await page.LocatorAsync(locatorOptions);27 await locator.ClickAsync();28 await page.ScreenshotAsync("result.png");29 await browser.CloseAsync();30 }31 }32} Full Screen Full Screen LocatorLocatorOptions Using AI Code Generation copy Full Screen 1using Microsoft.Playwright;2using System;3using System.Threading.Tasks;4{5 {6 static async Task Main(string[] args)7 {8 Console.WriteLine("Hello World!");9 using var playwright = await Playwright.CreateAsync();10 var browser = await playwright.Chromium.LaunchAsync(new BrowserTypeLaunchOptions11 {12 Args = new string[] { "--start-maximized" }13 });14 var context = await browser.NewContextAsync();15 var page = await context.NewPageAsync();16 await page.TypeAsync("input[name=q]", "Playwright");17 await page.PressAsync("input[name=q]", "Enter");18 await page.ScreenshotAsync(new PageScreenshotOptions { Path = "screenshot.png" });19 await browser.CloseAsync();20 }21 }22}23using Microsoft.Playwright;24using System;25using System.Threading.Tasks;26{27 {28 static async Task Main(string[] args)29 {30 Console.WriteLine("Hello World!");31 using var playwright = await Playwright.CreateAsync();32 var browser = await playwright.Chromium.LaunchAsync(new BrowserTypeLaunchOptions33 {34 Args = new string[] { "--start-maximized" }35 });36 var context = await browser.NewContextAsync();37 var page = await context.NewPageAsync();38 await page.TypeAsync("input[name=q]", "Playwright");39 await page.PressAsync("input[name=q]", "Enter");40 await page.ScreenshotAsync(new PageScreenshotOptions { Path = "screenshot.png" });41 await browser.CloseAsync();42 }43 }44}45using Microsoft.Playwright;46using System;47using System.Threading.Tasks;48{49 {50 static async Task Main(string[] args)51 {52 Console.WriteLine("Hello World!");53 using var playwright = await Playwright.CreateAsync();54 var browser = await playwright.Chromium.LaunchAsync(new BrowserTypeLaunchOptions Full Screen Full Screen LocatorLocatorOptions Using AI Code Generation copy Full Screen 1using Microsoft.Playwright;2using Microsoft.Playwright.NUnit;3using NUnit.Framework;4{5 {6 private IPage _page;7 public async Task SetUp()8 {9 var playwright = await Playwright.CreateAsync();10 var browser = await playwright.Chromium.LaunchAsync(new BrowserTypeLaunchOptions11 {12 });13 _page = await browser.NewPageAsync();14 }15 public async Task LocatorLocatorOptionsTest()16 {17 await _page.WaitForLoadStateAsync(LoadState.DOMContentLoaded);18 var locator = _page.Locator("input", new LocatorLocatorOptions19 {20 });21 Assert.NotNull(locator);22 }23 public async Task TearDown()24 {25 await _page.CloseAsync();26 }27 }28}29using Microsoft.Playwright;30using Microsoft.Playwright.NUnit;31using NUnit.Framework;32{33 {34 private IPage _page;35 public async Task SetUp()36 {37 var playwright = await Playwright.CreateAsync();38 var browser = await playwright.Chromium.LaunchAsync(new BrowserTypeLaunchOptions39 {40 });41 _page = await browser.NewPageAsync();42 }43 public async Task LocatorLocatorOptionsTest()44 {45 await _page.WaitForLoadStateAsync(LoadState.DOMContentLoaded);46 var locator = _page.Locator("input", new LocatorLocatorOptions47 {48 Text = new Regex("Google Search")49 });50 Assert.NotNull(locator);51 }52 public async Task TearDown()53 {54 await _page.CloseAsync();55 }56 }57}58using Microsoft.Playwright;59using Microsoft.Playwright.NUnit;60using NUnit.Framework; Full Screen Full Screen Playwright tutorial LambdaTest’s Playwright tutorial will give you a broader idea about the Playwright automation framework, its unique features, and use cases with examples to exceed your understanding of Playwright testing. This tutorial will give A to Z guidance, from installing the Playwright framework to some best practices and advanced concepts. Chapters: 1. What is Playwright : Playwright is comparatively new but has gained good popularity. Get to know some history of the Playwright with some interesting facts connected with it. 2. How To Install Playwright : Learn in detail about what basic configuration and dependencies are required for installing Playwright and run a test. Get a step-by-step direction for installing the Playwright automation framework. 3. Playwright Futuristic Features: Launched in 2020, Playwright gained huge popularity quickly because of some obliging features such as Playwright Test Generator and Inspector, Playwright Reporter, Playwright auto-waiting mechanism and etc. Read up on those features to master Playwright testing. 4. What is Component Testing: Component testing in Playwright is a unique feature that allows a tester to test a single component of a web application without integrating them with other elements. Learn how to perform Component testing on the Playwright automation framework. 5. Inputs And Buttons In Playwright: Every website has Input boxes and buttons; learn about testing inputs and buttons with different scenarios and examples. 6. Functions and Selectors in Playwright: Learn how to launch the Chromium browser with Playwright. Also, gain a better understanding of some important functions like “BrowserContext,” which allows you to run multiple browser sessions, and “newPage” which interacts with a page. 7. Handling Alerts and Dropdowns in Playwright : Playwright interact with different types of alerts and pop-ups, such as simple, confirmation, and prompt, and different types of dropdowns, such as single selector and multi-selector get your hands-on with handling alerts and dropdown in Playright testing. 8. Playwright vs Puppeteer: Get to know about the difference between two testing frameworks and how they are different than one another, which browsers they support, and what features they provide. 9. Run Playwright Tests on LambdaTest: Playwright testing with LambdaTest leverages test performance to the utmost. You can run multiple Playwright tests in Parallel with the LammbdaTest test cloud. Get a step-by-step guide to run your Playwright test on the LambdaTest platform. 10. Playwright Python Tutorial: Playwright automation framework support all major languages such as Python, JavaScript, TypeScript, .NET and etc. However, there are various advantages to Python end-to-end testing with Playwright because of its versatile utility. Get the hang of Playwright python testing with this chapter. 11. Playwright End To End Testing Tutorial: Get your hands on with Playwright end-to-end testing and learn to use some exciting features such as TraceViewer, Debugging, Networking, Component testing, Visual testing, and many more. 12. Playwright Video Tutorial: Watch the video tutorials on Playwright testing from experts and get a consecutive in-depth explanation of Playwright automation testing. Run Playwright-dotnet automation tests on LambdaTest cloud grid Perform automation testing on 3000+ real desktop and mobile devices online. Most used methods in LocatorLocatorOptions Try LambdaTest Now !! Get 100 minutes of automation test minutes FREE!! Next-Gen App & Browser Testing Cloud Was this article helpful? Helpful NotHelpful
__label__pos
0.678233
iolog: fix --bandwidth-log segfaults [fio.git] / init.c CommitLineData 906c8d75 1/* cb2c86fd 2 * This file contains job initialization and setup functions. 906c8d75 3 */ ebac4655 JA 4#include <stdio.h> 5#include <stdlib.h> 6#include <unistd.h> 7#include <fcntl.h> 8#include <ctype.h> 9#include <string.h> 10#include <errno.h> 11#include <sys/ipc.h> ebac4655 JA 12#include <sys/types.h> 13#include <sys/stat.h> 14 15#include "fio.h" a5e0ee11 O 16#ifndef FIO_NO_HAVE_SHM_H 17#include <sys/shm.h> 18#endif 19 cb2c86fd 20#include "parse.h" 2e5cdb11 21#include "smalloc.h" 380065aa 22#include "filehash.h" 4f5af7b2 23#include "verify.h" 79d16311 24#include "profile.h" 50d16976 25#include "server.h" f2a2ce0e 26#include "idletime.h" 243bfe19 27#include "filelock.h" ebac4655 28 f87cc561 29#include "oslib/getopt.h" 984f30c9 30#include "oslib/strcasestr.h" bf2e821a 31 10aa136b JA 32#include "crc/test.h" 33 3d43382c 34const char fio_version_string[] = FIO_VERSION; 214e1eca 35 ee738499 36#define FIO_RANDSEED (0xb1899bedUL) ebac4655 37 214e1eca 38static char **ini_file; fca70358 39static int max_jobs = FIO_MAX_JOBS; cca73aa7 40static int dump_cmdline; 25bd16ce 41static long long def_timeout; 111e032d 42static int parse_only; e1f36503 43 be4ecfdf 44static struct thread_data def_thread; 214e1eca 45struct thread_data *threads = NULL; 10aa136b JA 46static char **job_sections; 47static int nr_job_sections; e1f36503 48 214e1eca 49int exitall_on_terminate = 0; f3afa57e 50int output_format = FIO_OUTPUT_NORMAL; 5be4c944 51int eta_print = FIO_ETA_AUTO; e382e661 52int eta_new_line = 0; 214e1eca JA 53FILE *f_out = NULL; 54FILE *f_err = NULL; 07b3232d 55char *exec_profile = NULL; a9523c6f 56int warnings_fatal = 0; 4d658652 57int terse_version = 3; 50d16976 58int is_backend = 0; a37f69b7 59int nr_clients = 0; e46d8091 60int log_syslog = 0; ee738499 61 214e1eca 62int write_bw_log = 0; 4241ea8f 63int read_only = 0; 06464907 64int status_interval = 0; 214e1eca 65 ca09be4b 66char *trigger_file = NULL; ca09be4b 67long long trigger_timeout = 0; b63efd30 JA 68char *trigger_cmd = NULL; 69char *trigger_remote_cmd = NULL; ca09be4b 70 d264264a JA 71char *aux_path = NULL; 72 214e1eca 73static int prev_group_jobs; b4692828 74 ee56ad50 75unsigned long fio_debug = 0; 5e1d306e JA 76unsigned int fio_debug_jobno = -1; 77unsigned int *fio_debug_jobp = NULL; ee56ad50 78 4c6107ff 79static char cmd_optstr[256]; 085399db 80static int did_arg; 4c6107ff 81 7a4b8240 JA 82#define FIO_CLIENT_FLAG (1 << 16) 83 b4692828 JA 84/* 85 * Command line options. These will contain the above, plus a few 86 * extra that only pertain to fio itself and not jobs. 87 */ 5ec10eaa 88static struct option l_opts[FIO_NR_OPTIONS] = { b4692828 89 { 08d2a19c 90 .name = (char *) "output", b4692828 91 .has_arg = required_argument, 7a4b8240 92 .val = 'o' | FIO_CLIENT_FLAG, b4692828 JA 93 }, 94 { 08d2a19c 95 .name = (char *) "timeout", b4692828 96 .has_arg = required_argument, 7a4b8240 97 .val = 't' | FIO_CLIENT_FLAG, b4692828 JA 98 }, 99 { 08d2a19c 100 .name = (char *) "latency-log", b4692828 101 .has_arg = required_argument, 7a4b8240 102 .val = 'l' | FIO_CLIENT_FLAG, b4692828 JA 103 }, 104 { 08d2a19c 105 .name = (char *) "bandwidth-log", 2fd6fad5 106 .has_arg = no_argument, 7a4b8240 107 .val = 'b' | FIO_CLIENT_FLAG, b4692828 JA 108 }, 109 { 08d2a19c 110 .name = (char *) "minimal", c4ec0a1d 111 .has_arg = no_argument, 7a4b8240 112 .val = 'm' | FIO_CLIENT_FLAG, b4692828 113 }, cc372b17 114 { f3afa57e 115 .name = (char *) "output-format", d2c87a78 116 .has_arg = required_argument, f3afa57e 117 .val = 'F' | FIO_CLIENT_FLAG, cc372b17 118 }, f6a7df53 JA 119 { 120 .name = (char *) "append-terse", 121 .has_arg = optional_argument, 122 .val = 'f', 123 }, b4692828 124 { 08d2a19c 125 .name = (char *) "version", b4692828 126 .has_arg = no_argument, 7a4b8240 127 .val = 'v' | FIO_CLIENT_FLAG, b4692828 128 }, fd28ca49 129 { 08d2a19c 130 .name = (char *) "help", fd28ca49 131 .has_arg = no_argument, 7a4b8240 132 .val = 'h' | FIO_CLIENT_FLAG, fd28ca49 JA 133 }, 134 { 08d2a19c 135 .name = (char *) "cmdhelp", 320beefe 136 .has_arg = optional_argument, 7a4b8240 137 .val = 'c' | FIO_CLIENT_FLAG, fd28ca49 138 }, de890a1e 139 { 06464907 140 .name = (char *) "enghelp", de890a1e 141 .has_arg = optional_argument, 06464907 142 .val = 'i' | FIO_CLIENT_FLAG, de890a1e 143 }, cca73aa7 144 { 08d2a19c 145 .name = (char *) "showcmd", cca73aa7 146 .has_arg = no_argument, 7a4b8240 147 .val = 's' | FIO_CLIENT_FLAG, 724e4435 JA 148 }, 149 { 08d2a19c 150 .name = (char *) "readonly", 724e4435 151 .has_arg = no_argument, 7a4b8240 152 .val = 'r' | FIO_CLIENT_FLAG, cca73aa7 153 }, e592a06b 154 { 08d2a19c 155 .name = (char *) "eta", e592a06b 156 .has_arg = required_argument, 7a4b8240 157 .val = 'e' | FIO_CLIENT_FLAG, e592a06b 158 }, e382e661 JA 159 { 160 .name = (char *) "eta-newline", 161 .has_arg = required_argument, 162 .val = 'E' | FIO_CLIENT_FLAG, 163 }, ee56ad50 164 { 08d2a19c 165 .name = (char *) "debug", ee56ad50 166 .has_arg = required_argument, 7a4b8240 167 .val = 'd' | FIO_CLIENT_FLAG, ee56ad50 168 }, 111e032d JA 169 { 170 .name = (char *) "parse-only", 171 .has_arg = no_argument, 172 .val = 'P' | FIO_CLIENT_FLAG, 173 }, 01f06b63 174 { 08d2a19c 175 .name = (char *) "section", 01f06b63 176 .has_arg = required_argument, 7a4b8240 177 .val = 'x' | FIO_CLIENT_FLAG, 01f06b63 178 }, b26317c9 JA 179#ifdef CONFIG_ZLIB 180 { 181 .name = (char *) "inflate-log", 182 .has_arg = required_argument, 183 .val = 'X' | FIO_CLIENT_FLAG, 184 }, 185#endif 2b386d25 186 { 08d2a19c 187 .name = (char *) "alloc-size", 2b386d25 188 .has_arg = required_argument, 7a4b8240 189 .val = 'a' | FIO_CLIENT_FLAG, 2b386d25 190 }, 9ac8a797 191 { 08d2a19c 192 .name = (char *) "profile", 9ac8a797 193 .has_arg = required_argument, 7a4b8240 194 .val = 'p' | FIO_CLIENT_FLAG, 9ac8a797 195 }, a9523c6f 196 { 08d2a19c 197 .name = (char *) "warnings-fatal", a9523c6f 198 .has_arg = no_argument, 7a4b8240 199 .val = 'w' | FIO_CLIENT_FLAG, a9523c6f 200 }, fca70358 JA 201 { 202 .name = (char *) "max-jobs", 203 .has_arg = required_argument, 7a4b8240 204 .val = 'j' | FIO_CLIENT_FLAG, fca70358 205 }, f57a9c59 JA 206 { 207 .name = (char *) "terse-version", 208 .has_arg = required_argument, 7a4b8240 209 .val = 'V' | FIO_CLIENT_FLAG, f57a9c59 210 }, 50d16976 JA 211 { 212 .name = (char *) "server", 87aa8f19 213 .has_arg = optional_argument, 50d16976 JA 214 .val = 'S', 215 }, e46d8091 216 { .name = (char *) "daemonize", 402668f3 217 .has_arg = required_argument, e46d8091 JA 218 .val = 'D', 219 }, 132159a5 JA 220 { 221 .name = (char *) "client", 222 .has_arg = required_argument, 223 .val = 'C', 224 }, 323255cc JA 225 { 226 .name = (char *) "remote-config", 227 .has_arg = required_argument, 228 .val = 'R', 229 }, 7d11f871 JA 230 { 231 .name = (char *) "cpuclock-test", 232 .has_arg = no_argument, 233 .val = 'T', 234 }, fec0f21c JA 235 { 236 .name = (char *) "crctest", 237 .has_arg = optional_argument, 238 .val = 'G', 239 }, f2a2ce0e HL 240 { 241 .name = (char *) "idle-prof", 242 .has_arg = required_argument, 243 .val = 'I', 244 }, 06464907 JA 245 { 246 .name = (char *) "status-interval", 247 .has_arg = required_argument, 248 .val = 'L', 249 }, ca09be4b 250 { b63efd30 251 .name = (char *) "trigger-file", ca09be4b JA 252 .has_arg = required_argument, 253 .val = 'W', 254 }, 255 { 256 .name = (char *) "trigger-timeout", 257 .has_arg = required_argument, 258 .val = 'B', 259 }, b63efd30 JA 260 { 261 .name = (char *) "trigger", 262 .has_arg = required_argument, 263 .val = 'H', 264 }, 265 { 266 .name = (char *) "trigger-remote", 267 .has_arg = required_argument, 268 .val = 'J', 269 }, d264264a JA 270 { 271 .name = (char *) "aux-path", 272 .has_arg = required_argument, 273 .val = 'K', 274 }, b4692828 JA 275 { 276 .name = NULL, 277 }, 278}; 279 2bb3f0a7 280void free_threads_shm(void) 9d9eb2e7 281{ 9d9eb2e7 JA 282 if (threads) { 283 void *tp = threads; c8931876 JA 284#ifndef CONFIG_NO_SHM 285 struct shmid_ds sbuf; 9d9eb2e7 JA 286 287 threads = NULL; 2bb3f0a7 JA 288 shmdt(tp); 289 shmctl(shm_id, IPC_RMID, &sbuf); 290 shm_id = -1; c8931876 JA 291#else 292 threads = NULL; 293 free(tp); 294#endif 2bb3f0a7 JA 295 } 296} 297 10aa136b 298static void free_shm(void) 2bb3f0a7 JA 299{ 300 if (threads) { 9e684a49 301 flow_exit(); 9d9eb2e7 302 fio_debug_jobp = NULL; 2bb3f0a7 303 free_threads_shm(); 9d9eb2e7 JA 304 } 305 b63efd30 JA 306 free(trigger_file); 307 free(trigger_cmd); 308 free(trigger_remote_cmd); 309 trigger_file = trigger_cmd = trigger_remote_cmd = NULL; 310 5e333ada 311 options_free(fio_options, &def_thread.o); 243bfe19 312 fio_filelock_exit(); d619275c 313 file_hash_exit(); 9d9eb2e7 JA 314 scleanup(); 315} 316 317/* 318 * The thread area is shared between the main process and the job 319 * threads/processes. So setup a shared memory segment that will hold 320 * all the job info. We use the end of the region for keeping track of 321 * open files across jobs, for file sharing. 322 */ 323static int setup_thread_area(void) 324{ 9d9eb2e7 JA 325 if (threads) 326 return 0; 327 328 /* 329 * 1024 is too much on some machines, scale max_jobs if 330 * we get a failure that looks like too large a shm segment 331 */ 332 do { 333 size_t size = max_jobs * sizeof(struct thread_data); 334 9d9eb2e7 JA 335 size += sizeof(unsigned int); 336 c8931876 337#ifndef CONFIG_NO_SHM 9d9eb2e7 JA 338 shm_id = shmget(0, size, IPC_CREAT | 0600); 339 if (shm_id != -1) 340 break; f0c77f03 341 if (errno != EINVAL && errno != ENOMEM && errno != ENOSPC) { 9d9eb2e7 JA 342 perror("shmget"); 343 break; 344 } c8931876 JA 345#else 346 threads = malloc(size); 347 if (threads) 348 break; 349#endif 9d9eb2e7 JA 350 351 max_jobs >>= 1; 352 } while (max_jobs); 353 c8931876 354#ifndef CONFIG_NO_SHM 9d9eb2e7 JA 355 if (shm_id == -1) 356 return 1; 357 358 threads = shmat(shm_id, NULL, 0); 359 if (threads == (void *) -1) { 360 perror("shmat"); 361 return 1; 362 } c8931876 363#endif 9d9eb2e7 JA 364 365 memset(threads, 0, max_jobs * sizeof(struct thread_data)); 63a26e05 366 fio_debug_jobp = (void *) threads + max_jobs * sizeof(struct thread_data); 9d9eb2e7 367 *fio_debug_jobp = -1; 9e684a49 DE 368 369 flow_init(); 370 9d9eb2e7 JA 371 return 0; 372} 373 25bd16ce JA 374static void set_cmd_options(struct thread_data *td) 375{ 376 struct thread_options *o = &td->o; 377 378 if (!o->timeout) 379 o->timeout = def_timeout; 380} 381 c2292325 JA 382static void dump_print_option(struct print_option *p) 383{ 384 const char *delim; 385 386 if (!strcmp("description", p->name)) 387 delim = "\""; 388 else 389 delim = ""; 390 391 log_info("--%s%s", p->name, p->value ? "" : " "); 392 if (p->value) 393 log_info("=%s%s%s ", delim, p->value, delim); 394} 395 396static void dump_opt_list(struct thread_data *td) 397{ 398 struct flist_head *entry; 399 struct print_option *p; 400 401 if (flist_empty(&td->opt_list)) 402 return; 403 404 flist_for_each(entry, &td->opt_list) { 405 p = flist_entry(entry, struct print_option, list); 406 dump_print_option(p); 407 } 408} 409 410static void fio_dump_options_free(struct thread_data *td) 411{ 412 while (!flist_empty(&td->opt_list)) { 413 struct print_option *p; 414 415 p = flist_first_entry(&td->opt_list, struct print_option, list); 416 flist_del_init(&p->list); 417 free(p->name); 418 free(p->value); 419 free(p); 420 } 421} 422 876e1001 JA 423static void copy_opt_list(struct thread_data *dst, struct thread_data *src) 424{ 425 struct flist_head *entry; 426 427 if (flist_empty(&src->opt_list)) 428 return; 429 430 flist_for_each(entry, &src->opt_list) { 431 struct print_option *srcp, *dstp; 432 433 srcp = flist_entry(entry, struct print_option, list); 434 dstp = malloc(sizeof(*dstp)); 435 dstp->name = strdup(srcp->name); 436 if (srcp->value) 437 dstp->value = strdup(srcp->value); 438 else 439 dstp->value = NULL; 440 flist_add_tail(&dstp->list, &dst->opt_list); 441 } 442} 443 906c8d75 JA 444/* 445 * Return a free job structure. 446 */ de890a1e 447static struct thread_data *get_new_job(int global, struct thread_data *parent, ef3d8e53 448 int preserve_eo, const char *jobname) ebac4655 JA 449{ 450 struct thread_data *td; 451 25bd16ce JA 452 if (global) { 453 set_cmd_options(&def_thread); ebac4655 454 return &def_thread; 25bd16ce 455 } 9d9eb2e7 JA 456 if (setup_thread_area()) { 457 log_err("error: failed to setup shm segment\n"); 458 return NULL; 459 } e61f1ec8 BC 460 if (thread_number >= max_jobs) { 461 log_err("error: maximum number of jobs (%d) reached.\n", 462 max_jobs); ebac4655 463 return NULL; e61f1ec8 464 } ebac4655 JA 465 466 td = &threads[thread_number++]; ddaeaa5a 467 *td = *parent; ebac4655 468 c2292325 469 INIT_FLIST_HEAD(&td->opt_list); 876e1001 JA 470 if (parent != &def_thread) 471 copy_opt_list(td, parent); c2292325 472 de890a1e SL 473 td->io_ops = NULL; 474 if (!preserve_eo) 475 td->eo = NULL; 476 e0b0d892 JA 477 td->o.uid = td->o.gid = -1U; 478 cade3ef4 479 dup_files(td, parent); de890a1e 480 fio_options_mem_dupe(td); cade3ef4 481 15dc1934 JA 482 profile_add_hooks(td); 483 ebac4655 484 td->thread_number = thread_number; 69bdd6ba 485 td->subjob_number = 0; 108fea77 486 ef3d8e53 JA 487 if (jobname) 488 td->o.name = strdup(jobname); 489 13978e86 490 if (!parent->o.group_reporting || parent == &def_thread) 108fea77 JA 491 stat_number++; 492 25bd16ce 493 set_cmd_options(td); ebac4655 JA 494 return td; 495} 496 497static void put_job(struct thread_data *td) 498{ 549577a7 JA 499 if (td == &def_thread) 500 return; 84dd1886 501 58c55ba0 502 profile_td_exit(td); 9e684a49 503 flow_exit_job(td); 549577a7 504 16edf25d 505 if (td->error) 6d86144d 506 log_info("fio: %s\n", td->verror); 16edf25d 507 7e356b2d 508 fio_options_free(td); c2292325 509 fio_dump_options_free(td); de890a1e SL 510 if (td->io_ops) 511 free_ioengine(td); 7e356b2d 512 ef3d8e53 JA 513 if (td->o.name) 514 free(td->o.name); 515 ebac4655 JA 516 memset(&threads[td->thread_number - 1], 0, sizeof(*td)); 517 thread_number--; 518} 519 581e7141 520static int __setup_rate(struct thread_data *td, enum fio_ddir ddir) 127f6865 521{ 581e7141 522 unsigned int bs = td->o.min_bs[ddir]; 127f6865 523 ff58fced JA 524 assert(ddir_rw(ddir)); 525 ba3e4e0c 526 if (td->o.rate[ddir]) 1b8dbf25 527 td->rate_bps[ddir] = td->o.rate[ddir]; ba3e4e0c 528 else 194fffd0 529 td->rate_bps[ddir] = (uint64_t) td->o.rate_iops[ddir] * bs; 127f6865 530 1b8dbf25 531 if (!td->rate_bps[ddir]) { 127f6865 JA 532 log_err("rate lower than supported\n"); 533 return -1; 534 } 535 50a8ce86 536 td->rate_next_io_time[ddir] = 0; 89cdea5e 537 td->rate_io_issue_bytes[ddir] = 0; ff6bb260 538 td->last_usec = 0; 127f6865 JA 539 return 0; 540} 541 581e7141 JA 542static int setup_rate(struct thread_data *td) 543{ 544 int ret = 0; 545 546 if (td->o.rate[DDIR_READ] || td->o.rate_iops[DDIR_READ]) 547 ret = __setup_rate(td, DDIR_READ); 548 if (td->o.rate[DDIR_WRITE] || td->o.rate_iops[DDIR_WRITE]) 549 ret |= __setup_rate(td, DDIR_WRITE); 6eaf09d6 SL 550 if (td->o.rate[DDIR_TRIM] || td->o.rate_iops[DDIR_TRIM]) 551 ret |= __setup_rate(td, DDIR_TRIM); 581e7141 JA 552 553 return ret; 554} 555 8347239a JA 556static int fixed_block_size(struct thread_options *o) 557{ 558 return o->min_bs[DDIR_READ] == o->max_bs[DDIR_READ] && 559 o->min_bs[DDIR_WRITE] == o->max_bs[DDIR_WRITE] && 6eaf09d6 SL 560 o->min_bs[DDIR_TRIM] == o->max_bs[DDIR_TRIM] && 561 o->min_bs[DDIR_READ] == o->min_bs[DDIR_WRITE] && 562 o->min_bs[DDIR_READ] == o->min_bs[DDIR_TRIM]; 8347239a JA 563} 564 23ed19b0 CE 565 566static unsigned long long get_rand_start_delay(struct thread_data *td) 567{ 568 unsigned long long delayrange; c3546b53 569 uint64_t frand_max; 23ed19b0 CE 570 unsigned long r; 571 572 delayrange = td->o.start_delay_high - td->o.start_delay; 573 c3546b53 574 frand_max = rand_max(&td->delay_state); d6b72507 575 r = __rand(&td->delay_state); c3546b53 576 delayrange = (unsigned long long) ((double) delayrange * (r / (frand_max + 1.0))); 23ed19b0 CE 577 578 delayrange += td->o.start_delay; 579 return delayrange; 580} 581 dad915e3 JA 582/* 583 * Lazy way of fixing up options that depend on each other. We could also 584 * define option callback handlers, but this is easier. 585 */ 4e991c23 586static int fixup_options(struct thread_data *td) e1f36503 587{ 2dc1bbeb 588 struct thread_options *o = &td->o; eefd98b1 589 int ret = 0; dad915e3 590 f356d01d 591#ifndef FIO_HAVE_PSHARED_MUTEX 9bbf57cc 592 if (!o->use_thread) { f356d01d JA 593 log_info("fio: this platform does not support process shared" 594 " mutexes, forcing use of threads. Use the 'thread'" 595 " option to get rid of this warning.\n"); 9bbf57cc 596 o->use_thread = 1; a9523c6f 597 ret = warnings_fatal; f356d01d JA 598 } 599#endif 600 2dc1bbeb 601 if (o->write_iolog_file && o->read_iolog_file) { 076efc7c 602 log_err("fio: read iolog overrides write_iolog\n"); 2dc1bbeb JA 603 free(o->write_iolog_file); 604 o->write_iolog_file = NULL; a9523c6f 605 ret = warnings_fatal; 076efc7c 606 } 16b462ae 607 16b462ae 608 /* ed335855 609 * only really works with 1 file 16b462ae 610 */ 627aa1a8 611 if (o->zone_size && o->open_files > 1) 2dc1bbeb 612 o->zone_size = 0; 16b462ae 613 ed335855 SN 614 /* 615 * If zone_range isn't specified, backward compatibility dictates it 616 * should be made equal to zone_size. 617 */ 618 if (o->zone_size && !o->zone_range) 619 o->zone_range = o->zone_size; 620 16b462ae JA 621 /* 622 * Reads can do overwrites, we always need to pre-create the file 623 */ 624 if (td_read(td) || td_rw(td)) 2dc1bbeb 625 o->overwrite = 1; 16b462ae 626 2dc1bbeb 627 if (!o->min_bs[DDIR_READ]) 5ec10eaa 628 o->min_bs[DDIR_READ] = o->bs[DDIR_READ]; 2dc1bbeb JA 629 if (!o->max_bs[DDIR_READ]) 630 o->max_bs[DDIR_READ] = o->bs[DDIR_READ]; 631 if (!o->min_bs[DDIR_WRITE]) 5ec10eaa 632 o->min_bs[DDIR_WRITE] = o->bs[DDIR_WRITE]; 2dc1bbeb JA 633 if (!o->max_bs[DDIR_WRITE]) 634 o->max_bs[DDIR_WRITE] = o->bs[DDIR_WRITE]; 6eaf09d6 SL 635 if (!o->min_bs[DDIR_TRIM]) 636 o->min_bs[DDIR_TRIM] = o->bs[DDIR_TRIM]; 637 if (!o->max_bs[DDIR_TRIM]) 638 o->max_bs[DDIR_TRIM] = o->bs[DDIR_TRIM]; 639 2dc1bbeb 640 o->rw_min_bs = min(o->min_bs[DDIR_READ], o->min_bs[DDIR_WRITE]); 6eaf09d6 641 o->rw_min_bs = min(o->min_bs[DDIR_TRIM], o->rw_min_bs); a00735e6 642 2b7a01d0 JA 643 /* 644 * For random IO, allow blockalign offset other than min_bs. 645 */ 646 if (!o->ba[DDIR_READ] || !td_random(td)) 647 o->ba[DDIR_READ] = o->min_bs[DDIR_READ]; 648 if (!o->ba[DDIR_WRITE] || !td_random(td)) 649 o->ba[DDIR_WRITE] = o->min_bs[DDIR_WRITE]; 6eaf09d6 SL 650 if (!o->ba[DDIR_TRIM] || !td_random(td)) 651 o->ba[DDIR_TRIM] = o->min_bs[DDIR_TRIM]; 2b7a01d0 JA 652 653 if ((o->ba[DDIR_READ] != o->min_bs[DDIR_READ] || 6eaf09d6 SL 654 o->ba[DDIR_WRITE] != o->min_bs[DDIR_WRITE] || 655 o->ba[DDIR_TRIM] != o->min_bs[DDIR_TRIM]) && 9bbf57cc 656 !o->norandommap) { 2b7a01d0 657 log_err("fio: Any use of blockalign= turns off randommap\n"); 9bbf57cc 658 o->norandommap = 1; a9523c6f 659 ret = warnings_fatal; 2b7a01d0 JA 660 } 661 2dc1bbeb JA 662 if (!o->file_size_high) 663 o->file_size_high = o->file_size_low; 9c60ce64 664 23ed19b0 CE 665 if (o->start_delay_high) 666 o->start_delay = get_rand_start_delay(td); 667 8347239a JA 668 if (o->norandommap && o->verify != VERIFY_NONE 669 && !fixed_block_size(o)) { 670 log_err("fio: norandommap given for variable block sizes, " 83da8fbf 671 "verify limited\n"); a9523c6f 672 ret = warnings_fatal; bb8895e0 673 } 9b87f09b 674 if (o->bs_unaligned && (o->odirect || td_ioengine_flagged(td, FIO_RAWIO))) 690adba3 675 log_err("fio: bs_unaligned may not work with raw io\n"); e0a22335 676 48097d5c JA 677 /* 678 * thinktime_spin must be less than thinktime 679 */ 2dc1bbeb JA 680 if (o->thinktime_spin > o->thinktime) 681 o->thinktime_spin = o->thinktime; e916b390 JA 682 683 /* 684 * The low water mark cannot be bigger than the iodepth 685 */ 67bf9823 JA 686 if (o->iodepth_low > o->iodepth || !o->iodepth_low) 687 o->iodepth_low = o->iodepth; cb5ab512 JA 688 689 /* 690 * If batch number isn't set, default to the same as iodepth 691 */ 2dc1bbeb JA 692 if (o->iodepth_batch > o->iodepth || !o->iodepth_batch) 693 o->iodepth_batch = o->iodepth; b5af8293 694 82407585 RP 695 /* 696 * If max batch complete number isn't set or set incorrectly, 697 * default to the same as iodepth_batch_complete_min 698 */ 699 if (o->iodepth_batch_complete_min > o->iodepth_batch_complete_max) 700 o->iodepth_batch_complete_max = o->iodepth_batch_complete_min; 701 2dc1bbeb JA 702 if (o->nr_files > td->files_index) 703 o->nr_files = td->files_index; 9f9214f2 704 2dc1bbeb JA 705 if (o->open_files > o->nr_files || !o->open_files) 706 o->open_files = o->nr_files; 4e991c23 707 6eaf09d6 SL 708 if (((o->rate[DDIR_READ] + o->rate[DDIR_WRITE] + o->rate[DDIR_TRIM]) && 709 (o->rate_iops[DDIR_READ] + o->rate_iops[DDIR_WRITE] + o->rate_iops[DDIR_TRIM])) || 710 ((o->ratemin[DDIR_READ] + o->ratemin[DDIR_WRITE] + o->ratemin[DDIR_TRIM]) && 711 (o->rate_iops_min[DDIR_READ] + o->rate_iops_min[DDIR_WRITE] + o->rate_iops_min[DDIR_TRIM]))) { 4e991c23 712 log_err("fio: rate and rate_iops are mutually exclusive\n"); a9523c6f 713 ret = 1; 4e991c23 714 } 2af0ad5e D 715 if ((o->rate[DDIR_READ] && (o->rate[DDIR_READ] < o->ratemin[DDIR_READ])) || 716 (o->rate[DDIR_WRITE] && (o->rate[DDIR_WRITE] < o->ratemin[DDIR_WRITE])) || 717 (o->rate[DDIR_TRIM] && (o->rate[DDIR_TRIM] < o->ratemin[DDIR_TRIM])) || 718 (o->rate_iops[DDIR_READ] && (o->rate_iops[DDIR_READ] < o->rate_iops_min[DDIR_READ])) || 719 (o->rate_iops[DDIR_WRITE] && (o->rate_iops[DDIR_WRITE] < o->rate_iops_min[DDIR_WRITE])) || 720 (o->rate_iops[DDIR_TRIM] && (o->rate_iops[DDIR_TRIM] < o->rate_iops_min[DDIR_TRIM]))) { 4e991c23 721 log_err("fio: minimum rate exceeds rate\n"); a9523c6f 722 ret = 1; 4e991c23 JA 723 } 724 cf4464ca JA 725 if (!o->timeout && o->time_based) { 726 log_err("fio: time_based requires a runtime/timeout setting\n"); 727 o->time_based = 0; a9523c6f 728 ret = warnings_fatal; cf4464ca JA 729 } 730 aa31f1f1 731 if (o->fill_device && !o->size) 5921e80c 732 o->size = -1ULL; 5ec10eaa 733 9bbf57cc 734 if (o->verify != VERIFY_NONE) { 996936f9 735 if (td_write(td) && o->do_verify && o->numjobs > 1) { a9523c6f JA 736 log_info("Multiple writers may overwrite blocks that " 737 "belong to other jobs. This can cause " 738 "verification failures.\n"); 739 ret = warnings_fatal; 740 } 741 7627557b JA 742 if (!fio_option_is_set(o, refill_buffers)) 743 o->refill_buffers = 1; 744 2f1b8e8b JA 745 if (o->max_bs[DDIR_WRITE] != o->min_bs[DDIR_WRITE] && 746 !o->verify_interval) 747 o->verify_interval = o->min_bs[DDIR_WRITE]; 7ffa8930 JA 748 749 /* 750 * Verify interval must be smaller or equal to the 751 * write size. 752 */ 753 if (o->verify_interval > o->min_bs[DDIR_WRITE]) 754 o->verify_interval = o->min_bs[DDIR_WRITE]; 755 else if (td_read(td) && o->verify_interval > o->min_bs[DDIR_READ]) 756 o->verify_interval = o->min_bs[DDIR_READ]; d990588e 757 } 41ccd845 758 9bbf57cc JA 759 if (o->pre_read) { 760 o->invalidate_cache = 0; 9b87f09b 761 if (td_ioengine_flagged(td, FIO_PIPEIO)) { 9c0d2241 JA 762 log_info("fio: cannot pre-read files with an IO engine" 763 " that isn't seekable. Pre-read disabled.\n"); a9523c6f JA 764 ret = warnings_fatal; 765 } 9c0d2241 766 } 34f1c044 767 ad705bcb 768 if (!o->unit_base) { 9b87f09b 769 if (td_ioengine_flagged(td, FIO_BIT_BASED)) ad705bcb SN 770 o->unit_base = 1; 771 else 772 o->unit_base = 8; 773 } 774 67bf9823 775#ifndef CONFIG_FDATASYNC 9bbf57cc 776 if (o->fdatasync_blocks) { e72fa4d4 JA 777 log_info("fio: this platform does not support fdatasync()" 778 " falling back to using fsync(). Use the 'fsync'" 779 " option instead of 'fdatasync' to get rid of" 780 " this warning\n"); 9bbf57cc JA 781 o->fsync_blocks = o->fdatasync_blocks; 782 o->fdatasync_blocks = 0; a9523c6f 783 ret = warnings_fatal; e72fa4d4 JA 784 } 785#endif 786 93bcfd20 BC 787#ifdef WIN32 788 /* 789 * Windows doesn't support O_DIRECT or O_SYNC with the _open interface, 790 * so fail if we're passed those flags 791 */ 9b87f09b 792 if (td_ioengine_flagged(td, FIO_SYNCIO) && (td->o.odirect || td->o.sync_io)) { 93bcfd20 BC 793 log_err("fio: Windows does not support direct or non-buffered io with" 794 " the synchronous ioengines. Use the 'windowsaio' ioengine" 795 " with 'direct=1' and 'iodepth=1' instead.\n"); 796 ret = 1; 797 } 798#endif 799 811ac503 JA 800 /* 801 * For fully compressible data, just zero them at init time. 6269123c JA 802 * It's faster than repeatedly filling it. For non-zero 803 * compression, we should have refill_buffers set. Set it, unless 804 * the job file already changed it. 811ac503 805 */ 6269123c JA 806 if (o->compress_percentage) { 807 if (o->compress_percentage == 100) { 808 o->zero_buffers = 1; 809 o->compress_percentage = 0; 810 } else if (!fio_option_is_set(o, refill_buffers)) 811 o->refill_buffers = 1; 811ac503 JA 812 } 813 5881fda4 JA 814 /* 815 * Using a non-uniform random distribution excludes usage of 816 * a random map 817 */ 818 if (td->o.random_distribution != FIO_RAND_DIST_RANDOM) 819 td->o.norandommap = 1; 820 8d916c94 JA 821 /* 822 * If size is set but less than the min block size, complain 823 */ 824 if (o->size && o->size < td_min_bs(td)) { 825 log_err("fio: size too small, must be larger than the IO size: %llu\n", (unsigned long long) o->size); 826 ret = 1; 827 } 828 d01612f3 CM 829 /* 830 * O_ATOMIC implies O_DIRECT 831 */ 832 if (td->o.oatomic) 833 td->o.odirect = 1; 834 04778baf JA 835 /* 836 * If randseed is set, that overrides randrepeat 837 */ 40fe5e7b 838 if (fio_option_is_set(&td->o, rand_seed)) 04778baf JA 839 td->o.rand_repeatable = 0; 840 9b87f09b 841 if (td_ioengine_flagged(td, FIO_NOEXTEND) && td->o.file_append) { b1bebc32 JA 842 log_err("fio: can't append/extent with IO engine %s\n", td->io_ops->name); 843 ret = 1; 844 } 845 79c896a1 JA 846 if (fio_option_is_set(o, gtod_cpu)) { 847 fio_gtod_init(); 848 fio_gtod_set_cpu(o->gtod_cpu); 849 fio_gtod_offload = 1; 850 } 851 cf8a46a7 JA 852 td->loops = o->loops; 853 if (!td->loops) 854 td->loops = 1; 855 66347cfa 856 if (td->o.block_error_hist && td->o.nr_files != 1) { b3ec877c 857 log_err("fio: block error histogram only available " 66347cfa DE 858 "with a single file per job, but %d files " 859 "provided\n", td->o.nr_files); 860 ret = 1; 861 } 862 a9523c6f 863 return ret; e1f36503 JA 864} 865 f8977ee6 JA 866/* 867 * This function leaks the buffer 868 */ 99d633af 869char *fio_uint_to_kmg(unsigned int val) f8977ee6 JA 870{ 871 char *buf = malloc(32); f3502ba2 872 char post[] = { 0, 'K', 'M', 'G', 'P', 'E', 0 }; f8977ee6 JA 873 char *p = post; 874 245142ff 875 do { f8977ee6 JA 876 if (val & 1023) 877 break; 878 879 val >>= 10; 880 p++; 245142ff 881 } while (*p); f8977ee6 882 98ffb8f3 883 snprintf(buf, 32, "%u%c", val, *p); f8977ee6 JA 884 return buf; 885} 886 09629a90 JA 887/* External engines are specified by "external:name.o") */ 888static const char *get_engine_name(const char *str) 889{ 890 char *p = strstr(str, ":"); 891 892 if (!p) 893 return str; 894 895 p++; 896 strip_blank_front(&p); 897 strip_blank_end(p); 898 return p; 899} 900 99c94a6b JA 901static void init_rand_file_service(struct thread_data *td) 902{ 903 unsigned long nranges = td->o.nr_files << FIO_FSERVICE_SHIFT; 904 const unsigned int seed = td->rand_seeds[FIO_RAND_FILE_OFF]; 905 906 if (td->o.file_service_type == FIO_FSERVICE_ZIPF) { 907 zipf_init(&td->next_file_zipf, nranges, td->zipf_theta, seed); 908 zipf_disable_hash(&td->next_file_zipf); 909 } else if (td->o.file_service_type == FIO_FSERVICE_PARETO) { 910 pareto_init(&td->next_file_zipf, nranges, td->pareto_h, seed); 911 zipf_disable_hash(&td->next_file_zipf); 912 } else if (td->o.file_service_type == FIO_FSERVICE_GAUSS) { 913 gauss_init(&td->next_file_gauss, nranges, td->gauss_dev, seed); 914 gauss_disable_hash(&td->next_file_gauss); 915 } 916} 917 8f380de2 CJ 918void td_fill_verify_state_seed(struct thread_data *td) 919{ 920 bool use64; 921 922 if (td->o.random_generator == FIO_RAND_GEN_TAUSWORTHE64) 923 use64 = 1; 924 else 925 use64 = 0; 926 927 init_rand_seed(&td->verify_state, td->rand_seeds[FIO_RAND_VER_OFF], 928 use64); 929} 930 36dd3379 931static void td_fill_rand_seeds_internal(struct thread_data *td, bool use64) 4c07ad86 932{ 36dd3379 JA 933 int i; 934 c3546b53 935 init_rand_seed(&td->bsrange_state, td->rand_seeds[FIO_RAND_BS_OFF], use64); 8f380de2 936 td_fill_verify_state_seed(td); 36dd3379 937 init_rand_seed(&td->rwmix_state, td->rand_seeds[FIO_RAND_MIX_OFF], false); 4c07ad86 JA 938 939 if (td->o.file_service_type == FIO_FSERVICE_RANDOM) c3546b53 940 init_rand_seed(&td->next_file_state, td->rand_seeds[FIO_RAND_FILE_OFF], use64); 99c94a6b JA 941 else if (td->o.file_service_type & __FIO_FSERVICE_NONUNIFORM) 942 init_rand_file_service(td); 4c07ad86 943 c3546b53 JA 944 init_rand_seed(&td->file_size_state, td->rand_seeds[FIO_RAND_FILE_SIZE_OFF], use64); 945 init_rand_seed(&td->trim_state, td->rand_seeds[FIO_RAND_TRIM_OFF], use64); 946 init_rand_seed(&td->delay_state, td->rand_seeds[FIO_RAND_START_DELAY], use64); e7b24047 947 init_rand_seed(&td->poisson_state, td->rand_seeds[FIO_RAND_POISSON_OFF], 0); 36dd3379 JA 948 init_rand_seed(&td->dedupe_state, td->rand_seeds[FIO_DEDUPE_OFF], false); 949 init_rand_seed(&td->zone_state, td->rand_seeds[FIO_RAND_ZONE_OFF], false); 4c07ad86 JA 950 951 if (!td_random(td)) 952 return; 953 954 if (td->o.rand_repeatable) d24945f0 955 td->rand_seeds[FIO_RAND_BLOCK_OFF] = FIO_RANDSEED * td->thread_number; 4c07ad86 956 c3546b53 957 init_rand_seed(&td->random_state, td->rand_seeds[FIO_RAND_BLOCK_OFF], use64); 36dd3379 JA 958 959 for (i = 0; i < DDIR_RWDIR_CNT; i++) { 960 struct frand_state *s = &td->seq_rand_state[i]; 961 962 init_rand_seed(s, td->rand_seeds[FIO_RAND_SEQ_RAND_READ_OFF], false); 963 } 5bfc35d7 JA 964} 965 4c07ad86 JA 966void td_fill_rand_seeds(struct thread_data *td) 967{ 36dd3379 968 bool use64; c3546b53 969 56e2a5fc 970 if (td->o.allrand_repeatable) { 5c94b008 JA 971 unsigned int i; 972 973 for (i = 0; i < FIO_RAND_NR_OFFS; i++) 56e2a5fc CE 974 td->rand_seeds[i] = FIO_RANDSEED * td->thread_number 975 + i; 976 } 977 c3546b53 JA 978 if (td->o.random_generator == FIO_RAND_GEN_TAUSWORTHE64) 979 use64 = 1; 980 else 981 use64 = 0; 982 983 td_fill_rand_seeds_internal(td, use64); 3545a109 984 c3546b53 985 init_rand_seed(&td->buf_state, td->rand_seeds[FIO_RAND_BUF_OFF], use64); 5c94b008 986 frand_copy(&td->buf_state_prev, &td->buf_state); 4c07ad86 JA 987} 988 de890a1e SL 989/* 990 * Initializes the ioengine configured for a job, if it has not been done so 991 * already. 992 */ 993int ioengine_load(struct thread_data *td) 994{ 995 const char *engine; 996 997 /* 998 * Engine has already been loaded. 999 */ 1000 if (td->io_ops) 1001 return 0; 6e5c4b8e JA 1002 if (!td->o.ioengine) { 1003 log_err("fio: internal fault, no IO engine specified\n"); 1004 return 1; 1005 } de890a1e SL 1006 1007 engine = get_engine_name(td->o.ioengine); 1008 td->io_ops = load_ioengine(td, engine); 1009 if (!td->io_ops) { 1010 log_err("fio: failed to load engine %s\n", engine); 1011 return 1; 1012 } 1013 1014 if (td->io_ops->option_struct_size && td->io_ops->options) { 1015 /* 1016 * In cases where td->eo is set, clone it for a child thread. 1017 * This requires that the parent thread has the same ioengine, 1018 * but that requirement must be enforced by the code which 1019 * cloned the thread. 1020 */ 1021 void *origeo = td->eo; 1022 /* 1023 * Otherwise use the default thread options. 1024 */ 1025 if (!origeo && td != &def_thread && def_thread.eo && 1026 def_thread.io_ops->options == td->io_ops->options) 1027 origeo = def_thread.eo; 1028 1029 options_init(td->io_ops->options); 1030 td->eo = malloc(td->io_ops->option_struct_size); 1031 /* 1032 * Use the default thread as an option template if this uses the 1033 * same options structure and there are non-default options 1034 * used. 1035 */ 1036 if (origeo) { 1037 memcpy(td->eo, origeo, td->io_ops->option_struct_size); 1038 options_mem_dupe(td->eo, td->io_ops->options); 1039 } else { 1040 memset(td->eo, 0, td->io_ops->option_struct_size); 1041 fill_default_options(td->eo, td->io_ops->options); 1042 } 1043 *(struct thread_data **)td->eo = td; 1044 } 1045 9b87f09b JA 1046 if (td->o.odirect) 1047 td->io_ops->flags |= FIO_RAWIO; 1048 1049 td_set_ioengine_flags(td); de890a1e SL 1050 return 0; 1051} 1052 d72be545 JA 1053static void init_flags(struct thread_data *td) 1054{ 1055 struct thread_options *o = &td->o; 1056 1057 if (o->verify_backlog) 1058 td->flags |= TD_F_VER_BACKLOG; 1059 if (o->trim_backlog) 1060 td->flags |= TD_F_TRIM_BACKLOG; 1061 if (o->read_iolog_file) 1062 td->flags |= TD_F_READ_IOLOG; 1063 if (o->refill_buffers) 1064 td->flags |= TD_F_REFILL_BUFFERS; 1bf2498d 1065 /* b04f5905 1066 * Always scramble buffers if asked to 1bf2498d 1067 */ b04f5905 JA 1068 if (o->scramble_buffers && fio_option_is_set(o, scramble_buffers)) 1069 td->flags |= TD_F_SCRAMBLE_BUFFERS; 1070 /* 1071 * But also scramble buffers, unless we were explicitly asked 1072 * to zero them. 1073 */ 1074 if (o->scramble_buffers && !(o->zero_buffers && 1075 fio_option_is_set(o, zero_buffers))) d72be545 JA 1076 td->flags |= TD_F_SCRAMBLE_BUFFERS; 1077 if (o->verify != VERIFY_NONE) 1078 td->flags |= TD_F_VER_NONE; a9da8ab2 JA 1079 1080 if (o->verify_async || o->io_submit_mode == IO_MODE_OFFLOAD) 1081 td->flags |= TD_F_NEED_LOCK; d72be545 JA 1082} 1083 9dbc7bfe JA 1084static int setup_random_seeds(struct thread_data *td) 1085{ 1086 unsigned long seed; 1087 unsigned int i; 1088 40fe5e7b 1089 if (!td->o.rand_repeatable && !fio_option_is_set(&td->o, rand_seed)) 9dbc7bfe JA 1090 return init_random_state(td, td->rand_seeds, sizeof(td->rand_seeds)); 1091 40fe5e7b 1092 seed = td->o.rand_seed; 04778baf 1093 for (i = 0; i < 4; i++) 9dbc7bfe JA 1094 seed *= 0x9e370001UL; 1095 1096 for (i = 0; i < FIO_RAND_NR_OFFS; i++) { 87a0ea3b 1097 td->rand_seeds[i] = seed * td->thread_number + i; 9dbc7bfe JA 1098 seed *= 0x9e370001UL; 1099 } 1100 1101 td_fill_rand_seeds(td); 1102 return 0; 1103} 1104 de98bd30 JA 1105enum { 1106 FPRE_NONE = 0, 1107 FPRE_JOBNAME, 1108 FPRE_JOBNUM, 1109 FPRE_FILENUM 1110}; 1111 1112static struct fpre_keyword { 1113 const char *keyword; 1114 size_t strlen; 1115 int key; 1116} fpre_keywords[] = { 1117 { .keyword = "$jobname", .key = FPRE_JOBNAME, }, 1118 { .keyword = "$jobnum", .key = FPRE_JOBNUM, }, 1119 { .keyword = "$filenum", .key = FPRE_FILENUM, }, 1120 { .keyword = NULL, }, 1121 }; 1122 3660ceae 1123static char *make_filename(char *buf, size_t buf_size,struct thread_options *o, de98bd30 JA 1124 const char *jobname, int jobnum, int filenum) 1125{ 1126 struct fpre_keyword *f; 1127 char copy[PATH_MAX]; b400a20e 1128 size_t dst_left = PATH_MAX - 1; de98bd30 JA 1129 1130 if (!o->filename_format || !strlen(o->filename_format)) { 1131 sprintf(buf, "%s.%d.%d", jobname, jobnum, filenum); 1132 return NULL; 1133 } 1134 1135 for (f = &fpre_keywords[0]; f->keyword; f++) 1136 f->strlen = strlen(f->keyword); 1137 3660ceae JA 1138 buf[buf_size - 1] = '\0'; 1139 strncpy(buf, o->filename_format, buf_size - 1); 1140 de98bd30 JA 1141 memset(copy, 0, sizeof(copy)); 1142 for (f = &fpre_keywords[0]; f->keyword; f++) { 1143 do { 1144 size_t pre_len, post_start = 0; 1145 char *str, *dst = copy; 1146 e25b6d5a 1147 str = strcasestr(buf, f->keyword); de98bd30 JA 1148 if (!str) 1149 break; 1150 1151 pre_len = str - buf; 1152 if (strlen(str) != f->strlen) 1153 post_start = pre_len + f->strlen; 1154 1155 if (pre_len) { 1156 strncpy(dst, buf, pre_len); 1157 dst += pre_len; 73a467e6 1158 dst_left -= pre_len; de98bd30 JA 1159 } 1160 1161 switch (f->key) { 73a467e6 JA 1162 case FPRE_JOBNAME: { 1163 int ret; 1164 1165 ret = snprintf(dst, dst_left, "%s", jobname); 1166 if (ret < 0) 1167 break; 17a2be59 JA 1168 else if (ret > dst_left) { 1169 log_err("fio: truncated filename\n"); 1170 dst += dst_left; 1171 dst_left = 0; 1172 } else { 1173 dst += ret; 1174 dst_left -= ret; 1175 } de98bd30 1176 break; 73a467e6 JA 1177 } 1178 case FPRE_JOBNUM: { 1179 int ret; 1180 1181 ret = snprintf(dst, dst_left, "%d", jobnum); 1182 if (ret < 0) 1183 break; 17a2be59 JA 1184 else if (ret > dst_left) { 1185 log_err("fio: truncated filename\n"); 1186 dst += dst_left; 1187 dst_left = 0; 1188 } else { 1189 dst += ret; 1190 dst_left -= ret; 1191 } de98bd30 1192 break; 73a467e6 JA 1193 } 1194 case FPRE_FILENUM: { 1195 int ret; 1196 1197 ret = snprintf(dst, dst_left, "%d", filenum); 1198 if (ret < 0) 1199 break; 17a2be59 JA 1200 else if (ret > dst_left) { 1201 log_err("fio: truncated filename\n"); 1202 dst += dst_left; 1203 dst_left = 0; 1204 } else { 1205 dst += ret; 1206 dst_left -= ret; 1207 } de98bd30 1208 break; 73a467e6 1209 } de98bd30 JA 1210 default: 1211 assert(0); 1212 break; 1213 } 1214 1215 if (post_start) 73a467e6 1216 strncpy(dst, buf + post_start, dst_left); de98bd30 1217 3660ceae 1218 strncpy(buf, copy, buf_size - 1); de98bd30 JA 1219 } while (1); 1220 } 1221 1222 return buf; 1223} f9633d72 1224 8aa89d70 1225bool parse_dryrun(void) f9633d72 JA 1226{ 1227 return dump_cmdline || parse_only; 1228} 1229 3a5db920 JA 1230static void gen_log_name(char *name, size_t size, const char *logtype, 1231 const char *logname, unsigned int num, 1232 const char *suf, int per_job) 1233{ 1234 if (per_job) 1235 snprintf(name, size, "%s_%s.%d.%s", logname, logtype, num, suf); 1236 else 1237 snprintf(name, size, "%s_%s.%s", logname, logtype, suf); 1238} 1239 9cc8cb91 AK 1240static int check_waitees(char *waitee) 1241{ 1242 struct thread_data *td; 1243 int i, ret = 0; 1244 1245 for_each_td(td, i) { 1246 if (td->subjob_number) 1247 continue; 1248 1249 ret += !strcmp(td->o.name, waitee); 1250 } 1251 1252 return ret; 1253} 1254 1255static bool wait_for_ok(const char *jobname, struct thread_options *o) 1256{ 1257 int nw; 1258 1259 if (!o->wait_for) 1260 return true; 1261 1262 if (!strcmp(jobname, o->wait_for)) { 1263 log_err("%s: a job cannot wait for itself (wait_for=%s).\n", 1264 jobname, o->wait_for); 1265 return false; 1266 } 1267 1268 if (!(nw = check_waitees(o->wait_for))) { 1269 log_err("%s: waitee job %s unknown.\n", jobname, o->wait_for); 1270 return false; 1271 } 1272 1273 if (nw > 1) { 1274 log_err("%s: multiple waitees %s found,\n" 1275 "please avoid duplicates when using wait_for option.\n", 1276 jobname, o->wait_for); 1277 return false; 1278 } 1279 1280 return true; 1281} 1282 906c8d75 JA 1283/* 1284 * Adds a job to the list of things todo. Sanitizes the various options 1285 * to make sure we don't have conflicts, and initializes various 1286 * members of td. 1287 */ b7cfb2e1 1288static int add_job(struct thread_data *td, const char *jobname, int job_add_num, 46bcd498 1289 int recursed, int client_type) ebac4655 1290{ af52b345 1291 unsigned int i; af52b345 1292 char fname[PATH_MAX]; e132cbae 1293 int numjobs, file_alloced; de98bd30 1294 struct thread_options *o = &td->o; cb7e0ace 1295 char logname[PATH_MAX + 32]; ebac4655 1296 ebac4655 JA 1297 /* 1298 * the def_thread is just for options, it's not a real job 1299 */ 1300 if (td == &def_thread) 1301 return 0; 1302 d72be545 JA 1303 init_flags(td); 1304 cca73aa7 JA 1305 /* 1306 * if we are just dumping the output command line, don't add the job 1307 */ f9633d72 1308 if (parse_dryrun()) { cca73aa7 JA 1309 put_job(td); 1310 return 0; 1311 } 1312 46bcd498 JA 1313 td->client_type = client_type; 1314 58c55ba0 1315 if (profile_td_init(td)) 66251554 1316 goto err; 58c55ba0 1317 de890a1e 1318 if (ioengine_load(td)) 205927a3 1319 goto err; df64119d 1320 e132cbae 1321 file_alloced = 0; de98bd30 1322 if (!o->filename && !td->files_index && !o->read_iolog_file) { e132cbae 1323 file_alloced = 1; 80be24f4 1324 5e62565b 1325 if (o->nr_files == 1 && exists_and_not_regfile(jobname)) 5903e7b7 1326 add_file(td, jobname, job_add_num, 0); 7b05a215 1327 else { de98bd30 1328 for (i = 0; i < o->nr_files; i++) 3660ceae 1329 add_file(td, make_filename(fname, sizeof(fname), o, jobname, job_add_num, i), job_add_num, 0); af52b345 1330 } 0af7b542 1331 } ebac4655 1332 4e991c23 JA 1333 if (fixup_options(td)) 1334 goto err; e0a22335 1335 9cc8cb91 AK 1336 /* 1337 * Belongs to fixup_options, but o->name is not necessarily set as yet 1338 */ 1339 if (!wait_for_ok(jobname, o)) 1340 goto err; 1341 9e684a49 DE 1342 flow_init_job(td); 1343 de890a1e SL 1344 /* 1345 * IO engines only need this for option callbacks, and the address may 1346 * change in subprocesses. 1347 */ 1348 if (td->eo) 1349 *(struct thread_data **)td->eo = NULL; 1350 9b87f09b 1351 if (td_ioengine_flagged(td, FIO_DISKLESSIO)) { 07eb79df JA 1352 struct fio_file *f; 1353 1354 for_each_file(td, f, i) 1355 f->real_file_size = -1ULL; 1356 } 1357 521da527 1358 td->mutex = fio_mutex_init(FIO_MUTEX_LOCKED); ebac4655 1359 de98bd30 JA 1360 td->ts.clat_percentiles = o->clat_percentiles; 1361 td->ts.percentile_precision = o->percentile_precision; 1362 memcpy(td->ts.percentile_list, o->percentile_list, sizeof(o->percentile_list)); 83349190 1363 6eaf09d6 SL 1364 for (i = 0; i < DDIR_RWDIR_CNT; i++) { 1365 td->ts.clat_stat[i].min_val = ULONG_MAX; 1366 td->ts.slat_stat[i].min_val = ULONG_MAX; 1367 td->ts.lat_stat[i].min_val = ULONG_MAX; 1368 td->ts.bw_stat[i].min_val = ULONG_MAX; 1369 } de98bd30 1370 td->ddir_seq_nr = o->ddir_seq_nr; ebac4655 1371 de98bd30 1372 if ((o->stonewall || o->new_group) && prev_group_jobs) { 3c5df6fa 1373 prev_group_jobs = 0; ebac4655 1374 groupid++; d883efbd JA 1375 if (groupid == INT_MAX) { 1376 log_err("fio: too many groups defined\n"); 1377 goto err; 1378 } 3c5df6fa 1379 } ebac4655 JA 1380 1381 td->groupid = groupid; 3c5df6fa 1382 prev_group_jobs++; ebac4655 1383 9dbc7bfe 1384 if (setup_random_seeds(td)) { 93bcfd20 1385 td_verror(td, errno, "init_random_state"); 9c60ce64 1386 goto err; 93bcfd20 1387 } 9c60ce64 1388 ebac4655 JA 1389 if (setup_rate(td)) 1390 goto err; 1391 cb7e0ace 1392 if (o->lat_log_file) { aee2ab67 JA 1393 struct log_params p = { 1394 .td = td, 1395 .avg_msec = o->log_avg_msec, 1e613c9c KC 1396 .hist_msec = o->log_hist_msec, 1397 .hist_coarseness = o->log_hist_coarseness, aee2ab67 JA 1398 .log_type = IO_LOG_TYPE_LAT, 1399 .log_offset = o->log_offset, 1400 .log_gz = o->log_gz, b26317c9 1401 .log_gz_store = o->log_gz_store, aee2ab67 1402 }; b26317c9 1403 const char *suf; aee2ab67 1404 b26317c9 JA 1405 if (p.log_gz_store) 1406 suf = "log.fz"; 1407 else 1408 suf = "log"; 1409 3a5db920 JA 1410 gen_log_name(logname, sizeof(logname), "lat", o->lat_log_file, 1411 td->thread_number, suf, o->per_job_logs); aee2ab67 1412 setup_log(&td->lat_log, &p, logname); 3a5db920 JA 1413 1414 gen_log_name(logname, sizeof(logname), "slat", o->lat_log_file, 1415 td->thread_number, suf, o->per_job_logs); aee2ab67 1416 setup_log(&td->slat_log, &p, logname); 3a5db920 JA 1417 1418 gen_log_name(logname, sizeof(logname), "clat", o->lat_log_file, 1419 td->thread_number, suf, o->per_job_logs); aee2ab67 1420 setup_log(&td->clat_log, &p, logname); ebac4655 1421 } 1e613c9c KC 1422 1423 if (o->hist_log_file) { 1424 struct log_params p = { 1425 .td = td, 1426 .avg_msec = o->log_avg_msec, 1427 .hist_msec = o->log_hist_msec, 1428 .hist_coarseness = o->log_hist_coarseness, 1429 .log_type = IO_LOG_TYPE_HIST, 1430 .log_offset = o->log_offset, 1431 .log_gz = o->log_gz, 1432 .log_gz_store = o->log_gz_store, 1433 }; 1434 const char *suf; 1435 66da8a60 JA 1436#ifndef CONFIG_ZLIB 1437 if (td->client_type) { 1438 log_err("fio: --write_hist_log requires zlib in client/server mode\n"); 1439 goto err; 1440 } 1441#endif 1442 1e613c9c KC 1443 if (p.log_gz_store) 1444 suf = "log.fz"; 1445 else 1446 suf = "log"; 1447 1448 gen_log_name(logname, sizeof(logname), "clat_hist", o->hist_log_file, 1449 td->thread_number, suf, o->per_job_logs); 1450 setup_log(&td->clat_hist_log, &p, logname); 1451 } 1452 cb7e0ace 1453 if (o->bw_log_file) { aee2ab67 JA 1454 struct log_params p = { 1455 .td = td, 1456 .avg_msec = o->log_avg_msec, 1e613c9c KC 1457 .hist_msec = o->log_hist_msec, 1458 .hist_coarseness = o->log_hist_coarseness, aee2ab67 JA 1459 .log_type = IO_LOG_TYPE_BW, 1460 .log_offset = o->log_offset, 1461 .log_gz = o->log_gz, b26317c9 1462 .log_gz_store = o->log_gz_store, aee2ab67 1463 }; b26317c9 JA 1464 const char *suf; 1465 a47591e4 JA 1466 if (fio_option_is_set(o, bw_avg_time)) 1467 p.avg_msec = min(o->log_avg_msec, o->bw_avg_time); 1468 else 1469 o->bw_avg_time = p.avg_msec; 1e613c9c KC 1470 1471 p.hist_msec = o->log_hist_msec; 1472 p.hist_coarseness = o->log_hist_coarseness; a47591e4 1473 b26317c9 JA 1474 if (p.log_gz_store) 1475 suf = "log.fz"; 1476 else 1477 suf = "log"; aee2ab67 1478 3a5db920 JA 1479 gen_log_name(logname, sizeof(logname), "bw", o->bw_log_file, 1480 td->thread_number, suf, o->per_job_logs); aee2ab67 1481 setup_log(&td->bw_log, &p, logname); cb7e0ace JA 1482 } 1483 if (o->iops_log_file) { aee2ab67 JA 1484 struct log_params p = { 1485 .td = td, 1486 .avg_msec = o->log_avg_msec, 1e613c9c KC 1487 .hist_msec = o->log_hist_msec, 1488 .hist_coarseness = o->log_hist_coarseness, aee2ab67 JA 1489 .log_type = IO_LOG_TYPE_IOPS, 1490 .log_offset = o->log_offset, 1491 .log_gz = o->log_gz, b26317c9 1492 .log_gz_store = o->log_gz_store, aee2ab67 1493 }; b26317c9 JA 1494 const char *suf; 1495 a47591e4 JA 1496 if (fio_option_is_set(o, iops_avg_time)) 1497 p.avg_msec = min(o->log_avg_msec, o->iops_avg_time); 1498 else 1499 o->iops_avg_time = p.avg_msec; 1e613c9c KC 1500 1501 p.hist_msec = o->log_hist_msec; 1502 p.hist_coarseness = o->log_hist_coarseness; a47591e4 1503 b26317c9 JA 1504 if (p.log_gz_store) 1505 suf = "log.fz"; 1506 else 1507 suf = "log"; aee2ab67 1508 3a5db920 JA 1509 gen_log_name(logname, sizeof(logname), "iops", o->iops_log_file, 1510 td->thread_number, suf, o->per_job_logs); aee2ab67 1511 setup_log(&td->iops_log, &p, logname); cb7e0ace 1512 } ebac4655 1513 de98bd30 JA 1514 if (!o->name) 1515 o->name = strdup(jobname); 01452055 1516 129fb2d4 1517 if (output_format & FIO_OUTPUT_NORMAL) { b990b5c0 1518 if (!job_add_num) { b7cfb2e1 1519 if (is_backend && !recursed) 2f122b13 1520 fio_server_send_add_job(td); 807f9971 1521 9b87f09b 1522 if (!td_ioengine_flagged(td, FIO_NOIO)) { d21e2535 JA 1523 char *c1, *c2, *c3, *c4; 1524 char *c5 = NULL, *c6 = NULL; f8977ee6 1525 22f80458 JA 1526 c1 = fio_uint_to_kmg(o->min_bs[DDIR_READ]); 1527 c2 = fio_uint_to_kmg(o->max_bs[DDIR_READ]); 1528 c3 = fio_uint_to_kmg(o->min_bs[DDIR_WRITE]); 1529 c4 = fio_uint_to_kmg(o->max_bs[DDIR_WRITE]); d21e2535 JA 1530 1531 if (!o->bs_is_seq_rand) { 1532 c5 = fio_uint_to_kmg(o->min_bs[DDIR_TRIM]); 1533 c6 = fio_uint_to_kmg(o->max_bs[DDIR_TRIM]); 1534 } 1535 1536 log_info("%s: (g=%d): rw=%s, ", td->o.name, 1537 td->groupid, 1538 ddir_str(o->td_ddir)); 1539 1540 if (o->bs_is_seq_rand) 1541 log_info("bs(seq/rand)=%s-%s/%s-%s, ", 1542 c1, c2, c3, c4); 1543 else 1544 log_info("bs=%s-%s/%s-%s/%s-%s, ", 1545 c1, c2, c3, c4, c5, c6); 1546 1547 log_info("ioengine=%s, iodepth=%u\n", de98bd30 1548 td->io_ops->name, o->iodepth); f8977ee6 JA 1549 1550 free(c1); 1551 free(c2); 1552 free(c3); 1553 free(c4); 6eaf09d6 SL 1554 free(c5); 1555 free(c6); f8977ee6 1556 } b990b5c0 1557 } else if (job_add_num == 1) 6d86144d 1558 log_info("...\n"); c6ae0a5b 1559 } ebac4655 JA 1560 1561 /* 1562 * recurse add identical jobs, clear numjobs and stonewall options 1563 * as they don't apply to sub-jobs 1564 */ de98bd30 1565 numjobs = o->numjobs; ebac4655 1566 while (--numjobs) { ef3d8e53 1567 struct thread_data *td_new = get_new_job(0, td, 1, jobname); ebac4655 JA 1568 1569 if (!td_new) 1570 goto err; 1571 2dc1bbeb JA 1572 td_new->o.numjobs = 1; 1573 td_new->o.stonewall = 0; 92c1d41f 1574 td_new->o.new_group = 0; 69bdd6ba 1575 td_new->subjob_number = numjobs; e132cbae JA 1576 1577 if (file_alloced) { 455a50fa CE 1578 if (td_new->files) { 1579 struct fio_file *f; 1580 for_each_file(td_new, f, i) { 1581 if (f->file_name) ece6d647 AK 1582 sfree(f->file_name); 1583 sfree(f); 455a50fa 1584 } ece6d647 1585 free(td_new->files); 455a50fa CE 1586 td_new->files = NULL; 1587 } ece6d647 AK 1588 td_new->files_index = 0; 1589 td_new->files_size = 0; 455a50fa CE 1590 if (td_new->o.filename) { 1591 free(td_new->o.filename); 1592 td_new->o.filename = NULL; 1593 } e132cbae JA 1594 } 1595 bcbfeefa 1596 if (add_job(td_new, jobname, numjobs, 1, client_type)) ebac4655 JA 1597 goto err; 1598 } 3c5df6fa 1599 ebac4655 JA 1600 return 0; 1601err: 1602 put_job(td); 1603 return -1; 1604} 1605 79d16311 JA 1606/* 1607 * Parse as if 'o' was a command line 1608 */ 46bcd498 1609void add_job_opts(const char **o, int client_type) 79d16311 JA 1610{ 1611 struct thread_data *td, *td_parent; 1612 int i, in_global = 1; 1613 char jobname[32]; 1614 1615 i = 0; 1616 td_parent = td = NULL; 1617 while (o[i]) { 1618 if (!strncmp(o[i], "name", 4)) { 1619 in_global = 0; 1620 if (td) 46bcd498 1621 add_job(td, jobname, 0, 0, client_type); 79d16311 JA 1622 td = NULL; 1623 sprintf(jobname, "%s", o[i] + 5); 1624 } 1625 if (in_global && !td_parent) ef3d8e53 1626 td_parent = get_new_job(1, &def_thread, 0, jobname); 79d16311 JA 1627 else if (!in_global && !td) { 1628 if (!td_parent) 1629 td_parent = &def_thread; ef3d8e53 1630 td = get_new_job(0, td_parent, 0, jobname); 79d16311 JA 1631 } 1632 if (in_global) c2292325 1633 fio_options_parse(td_parent, (char **) &o[i], 1); 79d16311 1634 else c2292325 1635 fio_options_parse(td, (char **) &o[i], 1); 79d16311 JA 1636 i++; 1637 } 1638 1639 if (td) 46bcd498 1640 add_job(td, jobname, 0, 0, client_type); 79d16311 JA 1641} 1642 01f06b63 JA 1643static int skip_this_section(const char *name) 1644{ ad0a2735 JA 1645 int i; 1646 1647 if (!nr_job_sections) 01f06b63 JA 1648 return 0; 1649 if (!strncmp(name, "global", 6)) 1650 return 0; 1651 ad0a2735 JA 1652 for (i = 0; i < nr_job_sections; i++) 1653 if (!strcmp(job_sections[i], name)) 1654 return 0; 1655 1656 return 1; 01f06b63 JA 1657} 1658 ebac4655 JA 1659static int is_empty_or_comment(char *line) 1660{ 1661 unsigned int i; 1662 1663 for (i = 0; i < strlen(line); i++) { 1664 if (line[i] == ';') 1665 return 1; 5cc2da30 IM 1666 if (line[i] == '#') 1667 return 1; 76cd9378 1668 if (!isspace((int) line[i]) && !iscntrl((int) line[i])) ebac4655 JA 1669 return 0; 1670 } 1671 1672 return 1; 1673} 1674 07261983 JA 1675/* 1676 * This is our [ini] type file parser. 1677 */ b3270921 AK 1678int __parse_jobs_ini(struct thread_data *td, 1679 char *file, int is_buf, int stonewall_flag, int type, 1680 int nested, char *name, char ***popts, int *aopts, int *nopts) ebac4655 1681{ b3270921 AK 1682 unsigned int global = 0; 1683 char *string; ebac4655 JA 1684 FILE *f; 1685 char *p; 0c7e37a0 1686 int ret = 0, stonewall; 097b2991 1687 int first_sect = 1; ccf8f127 1688 int skip_fgets = 0; 01f06b63 1689 int inside_skip = 0; 3b8b7135 JA 1690 char **opts; 1691 int i, alloc_opts, num_opts; ebac4655 1692 b3270921 AK 1693 dprint(FD_PARSE, "Parsing ini file %s\n", file); 1694 assert(td || !nested); 1695 50d16976 JA 1696 if (is_buf) 1697 f = NULL; 1698 else { 1699 if (!strcmp(file, "-")) 1700 f = stdin; 1701 else 1702 f = fopen(file, "r"); 5a729cbe 1703 50d16976 1704 if (!f) { 323255cc JA 1705 int __err = errno; 1706 1707 log_err("fio: unable to open '%s' job file\n", file); 02957cd4 JA 1708 if (td) 1709 td_verror(td, __err, "job file open"); 50d16976 JA 1710 return 1; 1711 } ebac4655 JA 1712 } 1713 1714 string = malloc(4096); 7f7e6e59 JA 1715 1716 /* 1717 * it's really 256 + small bit, 280 should suffice 1718 */ b3270921 AK 1719 if (!nested) { 1720 name = malloc(280); 1721 memset(name, 0, 280); 1722 } ebac4655 1723 b3270921 AK 1724 opts = NULL; 1725 if (nested && popts) { 1726 opts = *popts; 1727 alloc_opts = *aopts; 1728 num_opts = *nopts; 1729 } 1730 1731 if (!opts) { 1732 alloc_opts = 8; 1733 opts = malloc(sizeof(char *) * alloc_opts); 1734 num_opts = 0; 1735 } 3b8b7135 1736 0c7e37a0 1737 stonewall = stonewall_flag; 7c124ac1 1738 do { ccf8f127 JA 1739 /* 1740 * if skip_fgets is set, we already have loaded a line we 1741 * haven't handled. 1742 */ 1743 if (!skip_fgets) { 50d16976 JA 1744 if (is_buf) 1745 p = strsep(&file, "\n"); 1746 else 43f74792 1747 p = fgets(string, 4096, f); ccf8f127 JA 1748 if (!p) 1749 break; 1750 } cdc7f193 1751 ccf8f127 1752 skip_fgets = 0; cdc7f193 1753 strip_blank_front(&p); 6c7c7da1 1754 strip_blank_end(p); cdc7f193 1755 b3270921 1756 dprint(FD_PARSE, "%s\n", p); ebac4655 JA 1757 if (is_empty_or_comment(p)) 1758 continue; b3270921 AK 1759 1760 if (!nested) { 1761 if (sscanf(p, "[%255[^\n]]", name) != 1) { 1762 if (inside_skip) 1763 continue; 1764 1765 log_err("fio: option <%s> outside of " 1766 "[] job section\n", p); 0e9c21a2 1767 ret = 1; b3270921 AK 1768 break; 1769 } 1770 1771 name[strlen(name) - 1] = '\0'; 1772 1773 if (skip_this_section(name)) { 1774 inside_skip = 1; 01f06b63 1775 continue; b3270921 AK 1776 } else 1777 inside_skip = 0; ebac4655 1778 b3270921 1779 dprint(FD_PARSE, "Parsing section [%s]\n", name); 7a4b80a1 1780 b3270921 1781 global = !strncmp(name, "global", 6); 01f06b63 1782 b3270921 AK 1783 if (dump_cmdline) { 1784 if (first_sect) 1785 log_info("fio "); 1786 if (!global) 1787 log_info("--name=%s ", name); 1788 first_sect = 0; 1789 } ebac4655 1790 b3270921 AK 1791 td = get_new_job(global, &def_thread, 0, name); 1792 if (!td) { 1793 ret = 1; 1794 break; 1795 } cca73aa7 1796 b3270921 AK 1797 /* 1798 * Separate multiple job files by a stonewall 1799 */ 1800 if (!global && stonewall) { 1801 td->o.stonewall = stonewall; 1802 stonewall = 0; 1803 } ebac4655 1804 b3270921 AK 1805 num_opts = 0; 1806 memset(opts, 0, alloc_opts * sizeof(char *)); 972cfd25 1807 } b3270921 AK 1808 else 1809 skip_fgets = 1; 3b8b7135 1810 50d16976 1811 while (1) { b3270921 AK 1812 if (!skip_fgets) { 1813 if (is_buf) 1814 p = strsep(&file, "\n"); 1815 else 1816 p = fgets(string, 4096, f); 1817 if (!p) 1818 break; 1819 dprint(FD_PARSE, "%s", p); 1820 } 50d16976 1821 else b3270921 1822 skip_fgets = 0; 50d16976 1823 ebac4655 JA 1824 if (is_empty_or_comment(p)) 1825 continue; e1f36503 1826 b6754f9d 1827 strip_blank_front(&p); 7c124ac1 1828 ccf8f127 JA 1829 /* 1830 * new section, break out and make sure we don't 1831 * fgets() a new line at the top. 1832 */ 1833 if (p[0] == '[') { b3270921 AK 1834 if (nested) { 1835 log_err("No new sections in included files\n"); 1836 return 1; 1837 } 1838 ccf8f127 1839 skip_fgets = 1; 7c124ac1 1840 break; ccf8f127 1841 } 7c124ac1 1842 4ae3f763 1843 strip_blank_end(p); aea47d44 1844 b3270921 1845 if (!strncmp(p, "include", strlen("include"))) { c0821a08 AK 1846 char *filename = p + strlen("include") + 1, 1847 *ts, *full_fn = NULL; 1848 1849 /* 1850 * Allow for the include filename 1851 * specification to be relative. 1852 */ 1853 if (access(filename, F_OK) && 1854 (ts = strrchr(file, '/'))) { 1855 int len = ts - file + 1856 strlen(filename) + 2; 1857 1858 if (!(full_fn = calloc(1, len))) { 1859 ret = ENOMEM; 1860 break; 1861 } 1862 1863 strncpy(full_fn, 1864 file, (ts - file) + 1); 1865 strncpy(full_fn + (ts - file) + 1, 1866 filename, strlen(filename)); 1867 full_fn[len - 1] = 0; 1868 filename = full_fn; 1869 } 1870 1871 ret = __parse_jobs_ini(td, filename, is_buf, 1872 stonewall_flag, type, 1, 1873 name, &opts, 1874 &alloc_opts, &num_opts); b3270921 1875 c0821a08 AK 1876 if (ret) { 1877 log_err("Error %d while parsing " 1878 "include file %s\n", b3270921 1879 ret, filename); b3270921 1880 } c0821a08 AK 1881 1882 if (full_fn) 1883 free(full_fn); 1884 1885 if (ret) 1886 break; 1887 b3270921 AK 1888 continue; 1889 } 1890 3b8b7135 JA 1891 if (num_opts == alloc_opts) { 1892 alloc_opts <<= 1; 1893 opts = realloc(opts, 1894 alloc_opts * sizeof(char *)); 1895 } 1896 1897 opts[num_opts] = strdup(p); 1898 num_opts++; ebac4655 1899 } ebac4655 1900 b3270921 AK 1901 if (nested) { 1902 *popts = opts; 1903 *aopts = alloc_opts; 1904 *nopts = num_opts; 1905 goto out; 1906 } 1907 c2292325 JA 1908 ret = fio_options_parse(td, opts, num_opts); 1909 if (!ret) { 1910 if (dump_cmdline) 1911 dump_opt_list(td); 1912 46bcd498 1913 ret = add_job(td, name, 0, 0, type); c2292325 1914 } else { b1508cf9 JA 1915 log_err("fio: job %s dropped\n", name); 1916 put_job(td); 45410acb 1917 } 3b8b7135 JA 1918 1919 for (i = 0; i < num_opts; i++) 1920 free(opts[i]); 1921 num_opts = 0; 7c124ac1 1922 } while (!ret); ebac4655 1923 cca73aa7 JA 1924 if (dump_cmdline) 1925 log_info("\n"); 1926 7e356b2d JA 1927 i = 0; 1928 while (i < nr_job_sections) { 1929 free(job_sections[i]); 1930 i++; 1931 } 1932 2577594d 1933 free(opts); b3270921 AK 1934out: 1935 free(string); 1936 if (!nested) 1937 free(name); 50d16976 1938 if (!is_buf && f != stdin) 5a729cbe 1939 fclose(f); 45410acb 1940 return ret; ebac4655 JA 1941} 1942 b3270921 AK 1943int parse_jobs_ini(char *file, int is_buf, int stonewall_flag, int type) 1944{ 1945 return __parse_jobs_ini(NULL, file, is_buf, stonewall_flag, type, 1946 0, NULL, NULL, NULL, NULL); 1947} 1948 ebac4655 JA 1949static int fill_def_thread(void) 1950{ 1951 memset(&def_thread, 0, sizeof(def_thread)); 66e19a38 1952 INIT_FLIST_HEAD(&def_thread.opt_list); ebac4655 1953 375b2695 1954 fio_getaffinity(getpid(), &def_thread.o.cpumask); 8b28bd41 1955 def_thread.o.error_dump = 1; 25bd16ce 1956 ebac4655 1957 /* ee738499 1958 * fill default options ebac4655 1959 */ 214e1eca 1960 fio_fill_default_options(&def_thread); ebac4655 JA 1961 return 0; 1962} 1963 2236df3d JA 1964static void show_debug_categories(void) 1965{ 44acdcf4 1966#ifdef FIO_INC_DEBUG 2236df3d JA 1967 struct debug_level *dl = &debug_levels[0]; 1968 int curlen, first = 1; 1969 1970 curlen = 0; 1971 while (dl->name) { 1972 int has_next = (dl + 1)->name != NULL; 1973 1974 if (first || curlen + strlen(dl->name) >= 80) { 1975 if (!first) { 1976 printf("\n"); 1977 curlen = 0; 1978 } 1979 curlen += printf("\t\t\t%s", dl->name); 1980 curlen += 3 * (8 - 1); 1981 if (has_next) 1982 curlen += printf(","); 1983 } else { 1984 curlen += printf("%s", dl->name); 1985 if (has_next) 1986 curlen += printf(","); 1987 } 1988 dl++; 1989 first = 0; 1990 } 1991 printf("\n"); 44acdcf4 1992#endif 2236df3d JA 1993} 1994 45378b35 1995static void usage(const char *name) 4785f995 1996{ 3d43382c 1997 printf("%s\n", fio_version_string); 45378b35 1998 printf("%s [options] [job options] <job file(s)>\n", name); 2236df3d JA 1999 printf(" --debug=options\tEnable debug logging. May be one/more of:\n"); 2000 show_debug_categories(); 111e032d 2001 printf(" --parse-only\t\tParse options only, don't start any IO\n"); 83881283 2002 printf(" --output\t\tWrite output to file\n"); b2cecdc2 2003 printf(" --runtime\t\tRuntime in seconds\n"); 83881283 JA 2004 printf(" --bandwidth-log\tGenerate per-job bandwidth logs\n"); 2005 printf(" --minimal\t\tMinimal (terse) output\n"); 513e37ee 2006 printf(" --output-format=x\tOutput format (terse,json,json+,normal)\n"); 75db6e2c 2007 printf(" --terse-version=x\tSet terse version output format to 'x'\n"); f3afa57e 2008 printf(" --version\t\tPrint version info and exit\n"); 83881283 2009 printf(" --help\t\tPrint this page\n"); 23893646 2010 printf(" --cpuclock-test\tPerform test/validation of CPU clock\n"); fec0f21c 2011 printf(" --crctest\t\tTest speed of checksum functions\n"); 83881283 2012 printf(" --cmdhelp=cmd\t\tPrint command help, \"all\" for all of" 5ec10eaa 2013 " them\n"); de890a1e SL 2014 printf(" --enghelp=engine\tPrint ioengine help, or list" 2015 " available ioengines\n"); 2016 printf(" --enghelp=engine,cmd\tPrint help for an ioengine" 2017 " cmd\n"); 83881283 JA 2018 printf(" --showcmd\t\tTurn a job file into command line options\n"); 2019 printf(" --eta=when\t\tWhen ETA estimate should be printed\n"); 2020 printf(" \t\tMay be \"always\", \"never\" or \"auto\"\n"); e382e661 JA 2021 printf(" --eta-newline=time\tForce a new line for every 'time'"); 2022 printf(" period passed\n"); 06464907 JA 2023 printf(" --status-interval=t\tForce full status dump every"); 2024 printf(" 't' period passed\n"); 83881283 2025 printf(" --readonly\t\tTurn on safety read-only checks, preventing" 5ec10eaa 2026 " writes\n"); 83881283 JA 2027 printf(" --section=name\tOnly run specified section in job file\n"); 2028 printf(" --alloc-size=kb\tSet smalloc pool to this size in kb" 2b386d25 2029 " (def 1024)\n"); 83881283 JA 2030 printf(" --warnings-fatal\tFio parser warnings are fatal\n"); 2031 printf(" --max-jobs=nr\t\tMaximum number of threads/processes to support\n"); 2032 printf(" --server=args\t\tStart a backend fio server\n"); 2033 printf(" --daemonize=pidfile\tBackground fio server, write pid to file\n"); 2034 printf(" --client=hostname\tTalk to remote backend fio server at hostname\n"); 323255cc 2035 printf(" --remote-config=file\tTell fio server to load this local job file\n"); f2a2ce0e HL 2036 printf(" --idle-prof=option\tReport cpu idleness on a system or percpu basis\n" 2037 "\t\t\t(option=system,percpu) or run unit work\n" 2038 "\t\t\tcalibration only (option=calibrate)\n"); b26317c9 JA 2039#ifdef CONFIG_ZLIB 2040 printf(" --inflate-log=log\tInflate and output compressed log\n"); 2041#endif b63efd30 2042 printf(" --trigger-file=file\tExecute trigger cmd when file exists\n"); ca09be4b 2043 printf(" --trigger-timeout=t\tExecute trigger af this time\n"); b63efd30 JA 2044 printf(" --trigger=cmd\t\tSet this command as local trigger\n"); 2045 printf(" --trigger-remote=cmd\tSet this command as remote trigger\n"); d264264a 2046 printf(" --aux-path=path\tUse this path for fio state generated files\n"); aa58d252 2047 printf("\nFio was written by Jens Axboe <[email protected]>"); bfca2c74 JA 2048 printf("\n Jens Axboe <[email protected]>"); 2049 printf("\n Jens Axboe <[email protected]>\n"); 4785f995 JA 2050} 2051 79e48f72 2052#ifdef FIO_INC_DEBUG ee56ad50 2053struct debug_level debug_levels[] = { 0b8d11ed JA 2054 { .name = "process", 2055 .help = "Process creation/exit logging", 2056 .shift = FD_PROCESS, 2057 }, 2058 { .name = "file", 2059 .help = "File related action logging", 2060 .shift = FD_FILE, 2061 }, 2062 { .name = "io", 2063 .help = "IO and IO engine action logging (offsets, queue, completions, etc)", 2064 .shift = FD_IO, 2065 }, 2066 { .name = "mem", 2067 .help = "Memory allocation/freeing logging", 2068 .shift = FD_MEM, 2069 }, 2070 { .name = "blktrace", 2071 .help = "blktrace action logging", 2072 .shift = FD_BLKTRACE, 2073 }, 2074 { .name = "verify", 2075 .help = "IO verification action logging", 2076 .shift = FD_VERIFY, 2077 }, 2078 { .name = "random", 2079 .help = "Random generation logging", 2080 .shift = FD_RANDOM, 2081 }, 2082 { .name = "parse", 2083 .help = "Parser logging", 2084 .shift = FD_PARSE, 2085 }, 2086 { .name = "diskutil", 2087 .help = "Disk utility logging actions", 2088 .shift = FD_DISKUTIL, 2089 }, 2090 { .name = "job", 2091 .help = "Logging related to creating/destroying jobs", 2092 .shift = FD_JOB, 2093 }, 2094 { .name = "mutex", 2095 .help = "Mutex logging", 2096 .shift = FD_MUTEX 2097 }, 2098 { .name = "profile", 2099 .help = "Logging related to profiles", 2100 .shift = FD_PROFILE, 2101 }, 2102 { .name = "time", 2103 .help = "Logging related to time keeping functions", 2104 .shift = FD_TIME, 2105 }, 2106 { .name = "net", 2107 .help = "Network logging", 2108 .shift = FD_NET, 2109 }, 3e260a46 JA 2110 { .name = "rate", 2111 .help = "Rate logging", 2112 .shift = FD_RATE, 2113 }, 0c56718d JA 2114 { .name = "compress", 2115 .help = "Log compression logging", 2116 .shift = FD_COMPRESS, 2117 }, 02444ad1 2118 { .name = NULL, }, ee56ad50 JA 2119}; 2120 c09823ab 2121static int set_debug(const char *string) ee56ad50 JA 2122{ 2123 struct debug_level *dl; 2124 char *p = (char *) string; 2125 char *opt; 2126 int i; 2127 2706fa00 JA 2128 if (!string) 2129 return 0; 2130 ee56ad50 2131 if (!strcmp(string, "?") || !strcmp(string, "help")) { ee56ad50 JA 2132 log_info("fio: dumping debug options:"); 2133 for (i = 0; debug_levels[i].name; i++) { 2134 dl = &debug_levels[i]; 2135 log_info("%s,", dl->name); 2136 } bd6f78b2 2137 log_info("all\n"); c09823ab 2138 return 1; ee56ad50 JA 2139 } 2140 2141 while ((opt = strsep(&p, ",")) != NULL) { 2142 int found = 0; 2143 5e1d306e JA 2144 if (!strncmp(opt, "all", 3)) { 2145 log_info("fio: set all debug options\n"); 2146 fio_debug = ~0UL; 2147 continue; 2148 } 2149 ee56ad50 JA 2150 for (i = 0; debug_levels[i].name; i++) { 2151 dl = &debug_levels[i]; 5e1d306e JA 2152 found = !strncmp(opt, dl->name, strlen(dl->name)); 2153 if (!found) 2154 continue; 2155 2156 if (dl->shift == FD_JOB) { 2157 opt = strchr(opt, ':'); 2158 if (!opt) { 2159 log_err("fio: missing job number\n"); 2160 break; 2161 } 2162 opt++; 2163 fio_debug_jobno = atoi(opt); 2164 log_info("fio: set debug jobno %d\n", 2165 fio_debug_jobno); 2166 } else { ee56ad50 2167 log_info("fio: set debug option %s\n", opt); bd6f78b2 2168 fio_debug |= (1UL << dl->shift); ee56ad50 2169 } 5e1d306e 2170 break; ee56ad50 JA 2171 } 2172 2173 if (!found) 2174 log_err("fio: debug mask %s not found\n", opt); 2175 } c09823ab 2176 return 0; ee56ad50 2177} 79e48f72 2178#else 69b98d4c 2179static int set_debug(const char *string) 79e48f72 JA 2180{ 2181 log_err("fio: debug tracing not included in build\n"); c09823ab 2182 return 1; 79e48f72 JA 2183} 2184#endif 9ac8a797 2185 4c6107ff JA 2186static void fio_options_fill_optstring(void) 2187{ 2188 char *ostr = cmd_optstr; 2189 int i, c; 2190 2191 c = i = 0; 2192 while (l_opts[i].name) { 2193 ostr[c++] = l_opts[i].val; 2194 if (l_opts[i].has_arg == required_argument) 2195 ostr[c++] = ':'; 2196 else if (l_opts[i].has_arg == optional_argument) { 2197 ostr[c++] = ':'; 2198 ostr[c++] = ':'; 2199 } 2200 i++; 2201 } 2202 ostr[c] = '\0'; 2203} 2204 c2c94585 JA 2205static int client_flag_set(char c) 2206{ 2207 int i; 2208 2209 i = 0; 2210 while (l_opts[i].name) { 2211 int val = l_opts[i].val; 2212 2213 if (c == (val & 0xff)) 2214 return (val & FIO_CLIENT_FLAG); 2215 2216 i++; 2217 } 2218 2219 return 0; 2220} 2221 10aa136b 2222static void parse_cmd_client(void *client, char *opt) 7a4b8240 2223{ fa2ea806 2224 fio_client_add_cmd_option(client, opt); 7a4b8240 JA 2225} 2226 a893c261 JA 2227static void show_closest_option(const char *name) 2228{ 2229 int best_option, best_distance; 2230 int i, distance; 2231 2232 while (*name == '-') 2233 name++; 2234 2235 best_option = -1; 2236 best_distance = INT_MAX; 2237 i = 0; 2238 while (l_opts[i].name) { 2239 distance = string_distance(name, l_opts[i].name); 2240 if (distance < best_distance) { 2241 best_distance = distance; 2242 best_option = i; 2243 } 2244 i++; 2245 } 2246 3701636d 2247 if (best_option != -1 && string_distance_ok(name, best_distance)) a893c261 JA 2248 log_err("Did you mean %s?\n", l_opts[best_option].name); 2249} 2250 a666cab8 JA 2251static int parse_output_format(const char *optarg) 2252{ 2253 char *p, *orig, *opt; 2254 int ret = 0; 2255 2256 p = orig = strdup(optarg); 2257 2258 output_format = 0; 2259 2260 while ((opt = strsep(&p, ",")) != NULL) { 2261 if (!strcmp(opt, "minimal") || 2262 !strcmp(opt, "terse") || 2263 !strcmp(opt, "csv")) 2264 output_format |= FIO_OUTPUT_TERSE; 2265 else if (!strcmp(opt, "json")) 2266 output_format |= FIO_OUTPUT_JSON; 513e37ee VF 2267 else if (!strcmp(opt, "json+")) 2268 output_format |= (FIO_OUTPUT_JSON | FIO_OUTPUT_JSON_PLUS); a666cab8 JA 2269 else if (!strcmp(opt, "normal")) 2270 output_format |= FIO_OUTPUT_NORMAL; 2271 else { 2272 log_err("fio: invalid output format %s\n", opt); 2273 ret = 1; 2274 break; 2275 } 2276 } 2277 2278 free(orig); 2279 return ret; 2280} 2281 46bcd498 2282int parse_cmd_line(int argc, char *argv[], int client_type) ebac4655 2283{ b4692828 2284 struct thread_data *td = NULL; c09823ab 2285 int c, ini_idx = 0, lidx, ret = 0, do_exit = 0, exit_val = 0; 4c6107ff 2286 char *ostr = cmd_optstr; 977b9596 2287 char *pid_file = NULL; 9bae26e8 2288 void *cur_client = NULL; 81179eec 2289 int backend = 0; ebac4655 2290 c2c94585 JA 2291 /* 2292 * Reset optind handling, since we may call this multiple times 2293 * for the backend. 2294 */ 2295 optind = 1; 7a4b8240 2296 c2c94585 JA 2297 while ((c = getopt_long_only(argc, argv, ostr, l_opts, &lidx)) != -1) { 2298 if ((c & FIO_CLIENT_FLAG) || client_flag_set(c)) { fa2ea806 2299 parse_cmd_client(cur_client, argv[optind - 1]); 7a4b8240 JA 2300 c &= ~FIO_CLIENT_FLAG; 2301 } 2302 ebac4655 2303 switch (c) { 2b386d25 JA 2304 case 'a': 2305 smalloc_pool_size = atoi(optarg); 52d892b2 2306 smalloc_pool_size <<= 10; 5f9454a2 2307 sinit(); 2b386d25 2308 break; b4692828 2309 case 't': 25bd16ce JA 2310 if (check_str_time(optarg, &def_timeout, 1)) { 2311 log_err("fio: failed parsing time %s\n", optarg); 2312 do_exit++; 2313 exit_val = 1; 2314 } b4692828 JA 2315 break; 2316 case 'l': cb7e0ace JA 2317 log_err("fio: --latency-log is deprecated. Use per-job latency log options.\n"); 2318 do_exit++; 2319 exit_val = 1; b4692828 2320 break; 3d73e5a9 2321 case 'b': b4692828 JA 2322 write_bw_log = 1; 2323 break; 2324 case 'o': 988da120 2325 if (f_out && f_out != stdout) 5e1d8745 JA 2326 fclose(f_out); 2327 b4692828 JA 2328 f_out = fopen(optarg, "w+"); 2329 if (!f_out) { 2330 perror("fopen output"); 2331 exit(1); 2332 } 2333 f_err = f_out; 2334 break; 2335 case 'm': f3afa57e 2336 output_format = FIO_OUTPUT_TERSE; b4692828 2337 break; f3afa57e 2338 case 'F': a666cab8 JA 2339 if (parse_output_format(optarg)) { 2340 log_err("fio: failed parsing output-format\n"); 2341 exit_val = 1; 2342 do_exit++; 2343 break; 2344 } 2b8c71b0 2345 break; f6a7df53 JA 2346 case 'f': 2347 output_format |= FIO_OUTPUT_TERSE; 2348 break; b4692828 2349 case 'h': 7d8ea970 2350 did_arg = 1; 7874f8b7 2351 if (!cur_client) { c2c94585 2352 usage(argv[0]); 7874f8b7 JA 2353 do_exit++; 2354 } c2c94585 2355 break; fd28ca49 2356 case 'c': 7d8ea970 2357 did_arg = 1; 7874f8b7 2358 if (!cur_client) { c2c94585 2359 fio_show_option_help(optarg); 7874f8b7 JA 2360 do_exit++; 2361 } c2c94585 2362 break; de890a1e 2363 case 'i': 7d8ea970 2364 did_arg = 1; de890a1e SL 2365 if (!cur_client) { 2366 fio_show_ioengine_help(optarg); 2367 do_exit++; 2368 } 2369 break; cca73aa7 2370 case 's': 7d8ea970 2371 did_arg = 1; cca73aa7 JA 2372 dump_cmdline = 1; 2373 break; 724e4435 JA 2374 case 'r': 2375 read_only = 1; 2376 break; b4692828 2377 case 'v': 7d8ea970 2378 did_arg = 1; 7874f8b7 2379 if (!cur_client) { 3d43382c 2380 log_info("%s\n", fio_version_string); 7874f8b7 JA 2381 do_exit++; 2382 } c2c94585 2383 break; f57a9c59 JA 2384 case 'V': 2385 terse_version = atoi(optarg); 09786f5f JA 2386 if (!(terse_version == 2 || terse_version == 3 || 2387 terse_version == 4)) { f57a9c59 JA 2388 log_err("fio: bad terse version format\n"); 2389 exit_val = 1; 2390 do_exit++; 2391 } 2392 break; e592a06b AC 2393 case 'e': 2394 if (!strcmp("always", optarg)) 2395 eta_print = FIO_ETA_ALWAYS; 2396 else if (!strcmp("never", optarg)) 2397 eta_print = FIO_ETA_NEVER; 2398 break; e382e661 JA 2399 case 'E': { 2400 long long t = 0; 2401 88038bc7 2402 if (check_str_time(optarg, &t, 1)) { e382e661 JA 2403 log_err("fio: failed parsing eta time %s\n", optarg); 2404 exit_val = 1; 2405 do_exit++; 2406 } 88038bc7 2407 eta_new_line = t / 1000; e382e661 JA 2408 break; 2409 } ee56ad50 2410 case 'd': c09823ab JA 2411 if (set_debug(optarg)) 2412 do_exit++; ee56ad50 2413 break; 111e032d 2414 case 'P': 7d8ea970 2415 did_arg = 1; 111e032d JA 2416 parse_only = 1; 2417 break; ad0a2735 JA 2418 case 'x': { 2419 size_t new_size; 2420 01f06b63 2421 if (!strcmp(optarg, "global")) { 5ec10eaa JA 2422 log_err("fio: can't use global as only " 2423 "section\n"); c09823ab JA 2424 do_exit++; 2425 exit_val = 1; 01f06b63 JA 2426 break; 2427 } ad0a2735 JA 2428 new_size = (nr_job_sections + 1) * sizeof(char *); 2429 job_sections = realloc(job_sections, new_size); 2430 job_sections[nr_job_sections] = strdup(optarg); 2431 nr_job_sections++; 01f06b63 2432 break; ad0a2735 2433 } b26317c9 JA 2434#ifdef CONFIG_ZLIB 2435 case 'X': 2436 exit_val = iolog_file_inflate(optarg); 2437 did_arg++; 2438 do_exit++; 2439 break; 2440#endif 9ac8a797 2441 case 'p': 7d8ea970 2442 did_arg = 1; 4af7c007 JA 2443 if (exec_profile) 2444 free(exec_profile); 07b3232d 2445 exec_profile = strdup(optarg); 9ac8a797 2446 break; b4692828 2447 case FIO_GETOPT_JOB: { 5ec10eaa 2448 const char *opt = l_opts[lidx].name; b4692828 JA 2449 char *val = optarg; 2450 c2b1e753 2451 if (!strncmp(opt, "name", 4) && td) { 46bcd498 2452 ret = add_job(td, td->o.name ?: "fio", 0, 0, client_type); 66251554 2453 if (ret) 4a4ac4e3 2454 goto out_free; c2b1e753 2455 td = NULL; 7d8ea970 2456 did_arg = 1; c2b1e753 2457 } b4692828 2458 if (!td) { 01f06b63 2459 int is_section = !strncmp(opt, "name", 4); 3106f220 JA 2460 int global = 0; 2461 01f06b63 2462 if (!is_section || !strncmp(val, "global", 6)) 3106f220 2463 global = 1; c2b1e753 2464 01f06b63 JA 2465 if (is_section && skip_this_section(val)) 2466 continue; 2467 ef3d8e53 2468 td = get_new_job(global, &def_thread, 1, NULL); f6ff24cb JA 2469 if (!td || ioengine_load(td)) { 2470 if (td) { 2471 put_job(td); 2472 td = NULL; 2473 } 2474 do_exit++; 205a719e 2475 exit_val = 1; f6ff24cb JA 2476 break; 2477 } de890a1e 2478 fio_options_set_ioengine_opts(l_opts, td); b4692828 2479 } 38d0adb0 2480 9d918187 JA 2481 if ((!val || !strlen(val)) && 2482 l_opts[lidx].has_arg == required_argument) { 2483 log_err("fio: option %s requires an argument\n", opt); 2484 ret = 1; 2485 } else 2486 ret = fio_cmd_option_parse(td, opt, val); 2487 bfb3ea22 JA 2488 if (ret) { 2489 if (td) { 2490 put_job(td); 2491 td = NULL; 2492 } a88c8c14 2493 do_exit++; 205a719e 2494 exit_val = 1; bfb3ea22 2495 } de890a1e SL 2496 2497 if (!ret && !strcmp(opt, "ioengine")) { 2498 free_ioengine(td); f6ff24cb 2499 if (ioengine_load(td)) { 342f570e JA 2500 put_job(td); 2501 td = NULL; f6ff24cb 2502 do_exit++; 205a719e 2503 exit_val = 1; f6ff24cb JA 2504 break; 2505 } de890a1e SL 2506 fio_options_set_ioengine_opts(l_opts, td); 2507 } 2508 break; 2509 } 2510 case FIO_GETOPT_IOENGINE: { 2511 const char *opt = l_opts[lidx].name; 2512 char *val = optarg; b1ed74e0 JA 2513 2514 if (!td) 2515 break; 2516 de890a1e 2517 ret = fio_cmd_ioengine_option_parse(td, opt, val); b4692828 JA 2518 break; 2519 } a9523c6f JA 2520 case 'w': 2521 warnings_fatal = 1; 2522 break; fca70358 JA 2523 case 'j': 2524 max_jobs = atoi(optarg); 2525 if (!max_jobs || max_jobs > REAL_MAX_JOBS) { 2526 log_err("fio: invalid max jobs: %d\n", max_jobs); 2527 do_exit++; 2528 exit_val = 1; 2529 } 2530 break; 50d16976 2531 case 'S': 7d8ea970 2532 did_arg = 1; c8931876 2533#ifndef CONFIG_NO_SHM a37f69b7 2534 if (nr_clients) { 132159a5 JA 2535 log_err("fio: can't be both client and server\n"); 2536 do_exit++; 2537 exit_val = 1; 2538 break; 2539 } 87aa8f19 2540 if (optarg) bebe6398 2541 fio_server_set_arg(optarg); 50d16976 2542 is_backend = 1; 81179eec 2543 backend = 1; c8931876 JA 2544#else 2545 log_err("fio: client/server requires SHM support\n"); 2546 do_exit++; 2547 exit_val = 1; 2548#endif 50d16976 2549 break; e46d8091 2550 case 'D': 6cfe9a8c JA 2551 if (pid_file) 2552 free(pid_file); 402668f3 2553 pid_file = strdup(optarg); e46d8091 2554 break; f2a2ce0e HL 2555 case 'I': 2556 if ((ret = fio_idle_prof_parse_opt(optarg))) { 2557 /* exit on error and calibration only */ 7d8ea970 2558 did_arg = 1; f2a2ce0e 2559 do_exit++; 7d8ea970 2560 if (ret == -1) f2a2ce0e HL 2561 exit_val = 1; 2562 } 2563 break; 132159a5 2564 case 'C': 7d8ea970 2565 did_arg = 1; 132159a5 JA 2566 if (is_backend) { 2567 log_err("fio: can't be both client and server\n"); 2568 do_exit++; 2569 exit_val = 1; 2570 break; 2571 } 08633c32 BE 2572 /* if --client parameter contains a pathname */ 2573 if (0 == access(optarg, R_OK)) { 2574 /* file contains a list of host addrs or names */ 2972708f 2575 char hostaddr[PATH_MAX] = {0}; 08633c32 BE 2576 char formatstr[8]; 2577 FILE * hostf = fopen(optarg, "r"); 2578 if (!hostf) { 2579 log_err("fio: could not open client list file %s for read\n", optarg); 2580 do_exit++; 2581 exit_val = 1; 2582 break; 2583 } 2972708f JA 2584 sprintf(formatstr, "%%%ds", PATH_MAX - 1); 2585 /* 2586 * read at most PATH_MAX-1 chars from each 2587 * record in this file 2588 */ 08633c32 BE 2589 while (fscanf(hostf, formatstr, hostaddr) == 1) { 2590 /* expect EVERY host in file to be valid */ 2591 if (fio_client_add(&fio_client_ops, hostaddr, &cur_client)) { 2592 log_err("fio: failed adding client %s from file %s\n", hostaddr, optarg); 2593 do_exit++; 2594 exit_val = 1; 2595 break; 2596 } 2597 } 2598 fclose(hostf); 2599 break; /* no possibility of job file for "this client only" */ 2600 } a5276616 2601 if (fio_client_add(&fio_client_ops, optarg, &cur_client)) { bebe6398 JA 2602 log_err("fio: failed adding client %s\n", optarg); 2603 do_exit++; 2604 exit_val = 1; 2605 break; 2606 } 14ea90ed JA 2607 /* 2608 * If the next argument exists and isn't an option, 2609 * assume it's a job file for this client only. 2610 */ 2611 while (optind < argc) { 2612 if (!strncmp(argv[optind], "--", 2) || 2613 !strncmp(argv[optind], "-", 1)) 2614 break; 2615 bb781643 2616 if (fio_client_add_ini_file(cur_client, argv[optind], false)) 323255cc 2617 break; 14ea90ed JA 2618 optind++; 2619 } 132159a5 2620 break; 323255cc JA 2621 case 'R': 2622 did_arg = 1; bb781643 2623 if (fio_client_add_ini_file(cur_client, optarg, true)) { 323255cc JA 2624 do_exit++; 2625 exit_val = 1; 2626 } 2627 break; 7d11f871 2628 case 'T': 7d8ea970 2629 did_arg = 1; 7d11f871 2630 do_exit++; aad918e4 2631 exit_val = fio_monotonic_clocktest(1); 7d11f871 2632 break; fec0f21c 2633 case 'G': 7d8ea970 2634 did_arg = 1; fec0f21c JA 2635 do_exit++; 2636 exit_val = fio_crctest(optarg); 2637 break; 06464907 JA 2638 case 'L': { 2639 long long val; 2640 88038bc7 2641 if (check_str_time(optarg, &val, 1)) { 06464907 JA 2642 log_err("fio: failed parsing time %s\n", optarg); 2643 do_exit++; 2644 exit_val = 1; 2645 break; 2646 } 88038bc7 2647 status_interval = val / 1000; 06464907 JA 2648 break; 2649 } b63efd30 2650 case 'W': f9cbae9e JA 2651 if (trigger_file) 2652 free(trigger_file); b63efd30 JA 2653 trigger_file = strdup(optarg); 2654 break; 2655 case 'H': f9cbae9e JA 2656 if (trigger_cmd) 2657 free(trigger_cmd); b63efd30 JA 2658 trigger_cmd = strdup(optarg); 2659 break; 2660 case 'J': f9cbae9e JA 2661 if (trigger_remote_cmd) 2662 free(trigger_remote_cmd); b63efd30 2663 trigger_remote_cmd = strdup(optarg); ca09be4b 2664 break; d264264a JA 2665 case 'K': 2666 if (aux_path) 2667 free(aux_path); 2668 aux_path = strdup(optarg); 2669 break; ca09be4b JA 2670 case 'B': 2671 if (check_str_time(optarg, &trigger_timeout, 1)) { 2672 log_err("fio: failed parsing time %s\n", optarg); 2673 do_exit++; 2674 exit_val = 1; 2675 } 2676 trigger_timeout /= 1000000; 2677 break; 798827c8 JA 2678 case '?': 2679 log_err("%s: unrecognized option '%s'\n", argv[0], 2680 argv[optind - 1]); a893c261 2681 show_closest_option(argv[optind - 1]); b4692828 2682 default: c09823ab JA 2683 do_exit++; 2684 exit_val = 1; b4692828 2685 break; ebac4655 2686 } c7d5c941 JA 2687 if (do_exit) 2688 break; ebac4655 2689 } c9fad893 2690 0b14f0a8 JA 2691 if (do_exit && !(is_backend || nr_clients)) 2692 exit(exit_val); 536582bf 2693 fb296043 JA 2694 if (nr_clients && fio_clients_connect()) 2695 exit(1); 132159a5 2696 81179eec 2697 if (is_backend && backend) 402668f3 2698 return fio_start_server(pid_file); b8ba87ac JA 2699 else if (pid_file) 2700 free(pid_file); 50d16976 2701 b4692828 2702 if (td) { 7d8ea970 2703 if (!ret) { 46bcd498 2704 ret = add_job(td, td->o.name ?: "fio", 0, 0, client_type); 7d8ea970 JA 2705 if (ret) 2706 did_arg = 1; 2707 } 972cfd25 2708 } 774a6177 2709 7874f8b7 2710 while (!ret && optind < argc) { b4692828 JA 2711 ini_idx++; 2712 ini_file = realloc(ini_file, ini_idx * sizeof(char *)); 2713 ini_file[ini_idx - 1] = strdup(argv[optind]); 2714 optind++; eb8bbf48 2715 } 972cfd25 2716 4a4ac4e3 JA 2717out_free: 2718 if (pid_file) 2719 free(pid_file); 2720 972cfd25 2721 return ini_idx; ebac4655 JA 2722} 2723 0420ba6a 2724int fio_init_options(void) ebac4655 2725{ b4692828 JA 2726 f_out = stdout; 2727 f_err = stderr; 2728 4c6107ff 2729 fio_options_fill_optstring(); 5ec10eaa 2730 fio_options_dup_and_init(l_opts); b4692828 2731 9d9eb2e7 JA 2732 atexit(free_shm); 2733 ebac4655 JA 2734 if (fill_def_thread()) 2735 return 1; 2736 0420ba6a JA 2737 return 0; 2738} 2739 51167799 JA 2740extern int fio_check_options(struct thread_options *); 2741 0420ba6a JA 2742int parse_options(int argc, char *argv[]) 2743{ 46bcd498 2744 const int type = FIO_CLIENT_TYPE_CLI; 0420ba6a JA 2745 int job_files, i; 2746 2747 if (fio_init_options()) 2748 return 1; 51167799 JA 2749 if (fio_test_cconv(&def_thread.o)) 2750 log_err("fio: failed internal cconv test\n"); 0420ba6a 2751 46bcd498 2752 job_files = parse_cmd_line(argc, argv, type); ebac4655 2753 cdf54d85 JA 2754 if (job_files > 0) { 2755 for (i = 0; i < job_files; i++) { bc4f5ef6 2756 if (i && fill_def_thread()) 132159a5 2757 return 1; cdf54d85 JA 2758 if (nr_clients) { 2759 if (fio_clients_send_ini(ini_file[i])) 2760 return 1; 2761 free(ini_file[i]); 2762 } else if (!is_backend) { 46bcd498 2763 if (parse_jobs_ini(ini_file[i], 0, i, type)) cdf54d85 JA 2764 return 1; 2765 free(ini_file[i]); 2766 } 132159a5 2767 } 14ea90ed JA 2768 } else if (nr_clients) { 2769 if (fill_def_thread()) 2770 return 1; 2771 if (fio_clients_send_ini(NULL)) 2772 return 1; 972cfd25 2773 } ebac4655 2774 88c6ed80 2775 free(ini_file); 7e356b2d 2776 fio_options_free(&def_thread); bcbfeefa 2777 filesetup_mem_free(); b4692828 JA 2778 2779 if (!thread_number) { f9633d72 2780 if (parse_dryrun()) cca73aa7 2781 return 0; 07b3232d JA 2782 if (exec_profile) 2783 return 0; a37f69b7 2784 if (is_backend || nr_clients) 5c341e9a 2785 return 0; 085399db JA 2786 if (did_arg) 2787 return 0; cca73aa7 2788 d65e11c6 2789 log_err("No jobs(s) defined\n\n"); 085399db JA 2790 2791 if (!did_arg) { 2792 usage(argv[0]); 2793 return 1; 2794 } 2795 2796 return 0; b4692828 JA 2797 } 2798 129fb2d4 2799 if (output_format & FIO_OUTPUT_NORMAL) 3d43382c 2800 log_info("%s\n", fio_version_string); f6dea4d3 2801 ebac4655 JA 2802 return 0; 2803} 588b7f09 JA 2804 2805void options_default_fill(struct thread_options *o) 2806{ 2807 memcpy(o, &def_thread.o, sizeof(*o)); 2808} c2292325 2809 66e19a38 2810struct thread_data *get_global_options(void) c2292325 2811{ 66e19a38 2812 return &def_thread; c2292325 2813}
__label__pos
0.895068
Home > Articles > Operating Systems, Server Journaling Filesystems for Linux • Print • + Share This Minimizing system restart time is the primary advantage of using a journaling filesystem, but there are many others. As "newer" filesystems, journaling filesystems can take advantage of newer techniques for enhancing filesystem performance. The previous article in this series provided background information on how data storage is organized and allocated on Linux and Unix systems, highlighting some of the more modern approaches used to improve performance, deal with larger files, and so on. One constant among all classic Linux and Unix filesystems is the general approach to the way the disk is updated when writing to a disk. Writing to a disk drive or other long-term storage is one of the slowest operations performed by computers, simply because it requires physical rather than electronic motion. For this reason, writing to a filesystem is usually done asynchronously, so that other processes on the system can continue to execute while data is being written to disk. Many filesystems cache data in memory until sufficient processor time is available, or a specific amount of data needs to be written to disk. The problem with standard caching and asynchronous disk updates is that if the system goes down in the middle of an update, the filesystem is usually left in an inconsistent state. File information may not have been updated to reflect blocks that have been added to or deallocated those files, and directories may not have been correctly updated to reflect files that have been created or deleted. Similarly, the free list or filesystem bitmap may not have been correctly updated to reflect blocks that have been allocated or deallocated from files and directories. To verify the consistency of a filesystem before attempting to mount and use it, Linux systems run a program called fsck, which stands for "file system check." If a filesystem isn't marked as being clean (by a bit in the filesystem superblock), the filesystem must be exhaustively checked for consistency before it can be mounted. Among other things, the fsck program for the ext2 filesystem verifies the consistency of all of the inodes, files, and directories in the filesystem, checks that all blocks marked as allocated are actually owned by some file or directory, and verifies that all blocks owned by files and directories are marked as allocated in the filesystem bitmap. As you can imagine, this can take quite a while to do on huge filesystems, and could therefore substantially delay making your system available to users. Journaling filesystems keep a journal (or log) of the changes that are to be made to the filesystem, and then asynchronously apply those changes to the filesystem. Sets of related changes in the log are marked as being completed when they have been successfully written to the filesystem, and are then deleted from the log. If a computer crashes during the middle of these updates, the operating system need only replay the pending transactions in the log to restore the filesystem to a consistent state, rather than having to check the entire filesystem. Journaling filesystems therefore minimize system downtime due to filesystem corruption—by replacing the need to check the consistency of an entire filesystem with the requirement of replaying a fairly small log of changes, systems that use journaling filesystems can be made available to the user much more quickly after a system crash or any other type of downtime. Minimizing system restart time is the primary advantage of using a journaling filesystem, but there are many others. As "newer" filesystems, journaling filesystems can take advantage of newer techniques for enhancing filesystem performance. Many journaling filesystems create and allocate inodes as they are needed, rather than preallocating a specific number of inodes when the filesystem is created. This removes limitations on the number of files and directories that can be created on that partition, increases performance, and reduces the overhead involved if you subsequently want to change the size of a journaling filesystem. Journaling filesystems also typically incorporate enhanced algorithms for storing and locating file and directory data, such as B-Trees, B+Trees, or B*Trees. Nowadays, the terms "logging" and "journaling" are usually used interchangeably when referring to filesystems that record changes to filesystem structures and data to minimize restart time and maximize consistency. Classically, log-based filesystems are actually a distinct type of filesystem that uses a log-oriented representation for the filesystem itself, and also usually require a garbage collection process to reclaim space internally. Journaling filesystems use a log, which is simply a distinct portion of a filesystem or disk. Where and how logs are stored and used differs with each type of journaling filesystem. I tend to use the term "journaling filesystem" so as not to anger any of my old Computer Science professors who may still be living. More Filesystems than You Can Shake a Memory Stick at One of the biggest features of Linux as an Open Source endeavor is that the availability of the source code for the operating system makes it easy to understand and extend the operating system itself. All operating systems provide APIs for integrating low-level services, but having the source code is like the difference between reading the blueprints for a house and being allowed inside it with a toolbelt. Having the source code also eliminates the chance of undocumented APIs, which you might only be familiar with if your mailing address is in Redmond. The availability of kernel source code and decent APIs for integrating low-level operating system services has resulted in some excellent extensions to the core capabilities of Linux, especially including support for new and existing filesystems. The best-known journaling filesystem for Linux, the Reiser File System, is an excellent example of this. The ReiserFS was born on Linux, and was the first journaling filesystem whose source code was integrated into the standard Linux kernel development tree. More recently (later versions of the 2.4 kernel family), the source code for the ext3 and JFS journaling filesystems have been integrated into the core Linux kernel source tree. As you see later in this article, the ext3 filesystem is a truly impressive effort—a logical follow-on to the ext2 filesystem that is completely compatible with existing ext2 filesystems and data structures. However, Linux has also benefited from some excellent journaling filesystems (such as JFS) with surprising roots—proprietary Unix vendors. To a large extent, Linux is ringing the death knell for proprietary versions of Unix. Why spend a zillion dollars for hardware and a proprietary version of Unix when Linux is freely available and will run on everything from a sexy SMP machine to the PDA in your pocket? Most of the standard Unix vendors have seen the light to some extent, and understand the importance of embracing (or at least playing nicely) with Linux. To this end, existing Unix vendors—such as IBM and Silicon Graphics—have contributed the source code for some of their most exciting research efforts, the journaling filesystems that these proprietary vendors use on some or all of their hardware. IBM released the source code for its Journal File System, JFS, as Open Source in 2000. Similarly, Silicon Graphics released the source code for its XFS (eXtended File System) as Open Source at the same time. Regardless of the PR value inherent in releasing projects on which they've spent millions of research dollars, the bottom line of these contributions is the tremendous benefit that the capability to understand and use these filesystems brings to Linux systems. The next few sections highlight the most popular journaling filesystems that are available for Linux and discuss some of the things that make each of them unique. As you'd expect, there are plenty of other journaling filesystems that are under development for Linux, as both research and open source projects. This article focuses on the ones that are actively used on Linux systems today, and which you may therefore actually encounter in the near future. • + Share This • 🔖 Save To Your Account Related Resources There are currently no related titles. Please check back later.
__label__pos
0.790896
How to filter krumbleate-ads.info referral traffic in Google Analytics krumbleate-ads.info referrer spam Krumbleate-ads.info Learn what krumbleate-ads.info is, why they’re spamming you, and how to filter krumbleate-ads.info referral traffic in Google Analytics. Links on this page What is Krumbleate-ads.info? Krumbleate-ads.info is a domain name utilized by a questionable Amazon advertisement automation service called Krumble. Krumble uses the krumbleate-ads.info domain name as a tool to spam your Google Analytics data with fake referral traffic. The unethical tactic of projecting fake visitors in your Google Analytics data is recognized as referrer spam indexing. Referrer spam indexing allows Krumble to target your Google Analytics data and show you fake referral traffic in order to obtain your attention and persuade you to visit krumbleate-ads.info. krumbleate-ads.info referral When you visit krumbleate-ads.info you will be forwarded to https://www.krumble.net/#ck. The new website says “Automation of Amazon Affiliate Program for Publishers” and “Krumble is a simple and reliable technology that automates the management of the Amazon Affiliate Program for Publishers.” krumbleate-ads.info website The reason why Krumble has spammed your Google Analytics data with fake referral traffic is to get you to visit this site and read it. They consider it something you might be interested in since you are a website owner or someone who monitors data. However, in my personal opinion, businesses that spam your data with fake referral traffic are not trustworthy. They disregard your data in order to benefit themselves. They show that they do not care about ruining your website’s data as long as you become aware of who they are. Referrer spam may sound harmless to some people but it can actually ruin your website’s analytical data and make it difficult to understand your website’s real traffic information. Fake referral traffic can affect most of the data in your reports such as your bounce rate. For example, the fake referrals will appear to land on a single page on your site and leave from the same page. This will create a 100% bounce rate. To add to this, the spammers usually hit your data with multiple fake visits which can inflate the bounce rate even more. Campaign Source Filter A campaign source filter can be used to block all krumbleate-ads.info referral traffic in Google Analytics. 1. Open your Google Analytics account and go to the Admin tab > Click Filters on the right side in the VIEW section. 2. Click the + ADD FILTER button to create a new exclude filter. 3. Add krumbleate-ads.info or something you can easily remember as the Filter Name. 4. Select the Custom Filter Type. 5. In Filter Field, find and select Campaign Source in the list. In the Filter Pattern text box, add krumbleate-ads.info and click the blue Save button on the bottom of the web page. To add multiple URLs to the same filter you can make a Filter Pattern similar to this with a | between each URL: Example.com | Example\.com | krumbleate-ads.info Campaign Referral Path Filter A campaign referral path filter can be used to block single web pages. 1. Open your Google Analytics account and go to the Admin tab > Click Filters on the right side in the VIEW section. 2. Click the + ADD FILTER button to create a new exclude filter. 3. Add krumbleate-ads.info or something you can easily remember as the Filter Name. 4. Select the Custom Filter Type. 5. In Filter Field, find and select Campaign Referral Path in the list. In the Filter Pattern text box, add a permalink from the referred URL and click the blue Save button on the bottom of the web page Language Settings Filter Some spam may appear in your language settings as keywords, phrases, and searched terms. A language settings filter can be used to block language spam in Google Analytics. 1. Log in to your Google Analytics account and go to the Admin tab 2. In the “View” column select Filters and then click + Add Filter 3. Add a Filter Name: Language Spam (or something you can easily remember) 4. Go to: Filter Type > Custom > Exclude 5. Select Filter FieldLanguage settings 6. Add a Filter Pattern\s[^s]*\s|.{15,}|\.|, 7. Click on the blue text that says Verify this filter to see a preview table of how this filter will work in your account. You should only see language spam on the left side of the table: filter-verification-language-spam 8. After you verify the filter click the Save button on the bottom of the page Sean Doyle Sean Doyle is an engineer from Los Angeles, California. Sean's primary focuses include Cyber Security, Web Spam, and Online Marketing.
__label__pos
0.814137
File:  [gforth] / gforth / configure.in Revision 1.216: download - view: text, annotated - select for diffs Sat Feb 23 13:03:55 2008 UTC (13 years, 9 months ago) by pazsan Branches: MAIN CVS tags: HEAD Fixed build problem Some changes for NXT 1: dnl Process this file with autoconf to produce a configure script. 2: 3: #Copyright (C) 1995,1996,1997,1998,2000,2003,2004,2005,2006,2007 Free Software Foundation, Inc. 4: 5: #This file is part of Gforth. 6: 7: #Gforth is free software; you can redistribute it and/or 8: #modify it under the terms of the GNU General Public License 9: #as published by the Free Software Foundation, either version 3 10: #of the License, or (at your option) any later version. 11: 12: #This program is distributed in the hope that it will be useful, 13: #but WITHOUT ANY WARRANTY; without even the implied warranty of 14: #MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.#See the 15: #GNU General Public License for more details. 16: 17: #You should have received a copy of the GNU General Public License 18: #along with this program. If not, see http://www.gnu.org/licenses/. 19: 20: 21: dnl We use some automake macros here, 22: dnl but don't use automake for creating Makefile.in 23: AC_INIT([gforth],[0.6.9-20080223],[https://savannah.gnu.org/bugs/?func=addbug&group=gforth]) 24: AC_PREREQ(2.54) 25: #snapshots have numbers major.minor.release-YYYYMMDD 26: #note that lexicographic ordering must be heeded. 27: #I.e., 0.4.1-YYYYMMDD must not exist before 0.4.1! 28: UPDATED="February 23, 2008" 29: AC_SUBST(UPDATED) 30: AC_CONFIG_HEADERS(engine/config.h) 31: 32: #suppress the "-g -O2" default 33: test "$CFLAGS" || CFLAGS=-O2 34: 35: AC_ARG_ENABLE(force-cdiv, 36: AC_HELP_STRING([--enable-force-cdiv], 37: [ Use the native C division - symmetric - instead of 38: floored division (default disabled).]), 39: ,enable_force_cdiv=no) 40: test "$enable_force_cdiv" = "no"|| 41: AC_DEFINE(FORCE_CDIV,,[Define if you want to use explicit symmetric division for better performance]) 42: 43: AC_SUBST(PROFEXES) 44: AC_SUBST(PROFOBJS) 45: AC_ARG_ENABLE(prof, 46: AC_HELP_STRING([--enable-prof], 47: [ Build gforth-prof, which outputs frequently occuring 48: sequences of primitives.]), 49: ,enable_prof=no) 50: if test "$enable_prof" != "no"; then 51: PROFEXES='gforth-prof$(OPT)$(EXE)'; PROFOBJS='engine-prof$(OPT).o main-prof$(OPT).o profile$(OPT).o' 52: fi 53: 54: AC_ARG_WITH(debug, 55: [ --with-debug specifies option -g to compile with debug info 56: --without-debug omits the -g switch and creates smaller images on 57: machines where "strip" has problems with gcc style 58: debugging informations.], 59: if test "$withval" = "yes"; then DEBUGFLAG=-g; fi) 60: 61: GCC_LD="\$(GCC)" 62: EC_MODE="false" 63: EC="" 64: engine2='engine2$(OPT).o' 65: engine_fast2='engine-fast2$(OPT).o' 66: no_dynamic="" 67: image_i="" 68: signals_o="io.o signals.o" 69: 70: AC_ARG_WITH(ec, 71: AC_HELP_STRING([--with-ec=<arch>], 72: [ Build gforth for systems without OS.]), 73: [if test "$withval" = "no"; then 74: echo "defining hosted system" 75: else 76: echo "defining standalone system (${withval})" 77: AC_DEFINE(STANDALONE,,[Define if you want a Gforth without OS]) 78: EC_MODE="true" 79: EC="-ec" 80: engine2="" 81: engine_fast2="" 82: no_dynamic="-DNO_DYNAMIC" 83: image_i="image.i" 84: if test "$withval" != "yes"; then 85: signals_o="io-${withval}.o" 86: else 87: signals_o="io.o" 88: fi 89: GCC_PATH=$(which $CC) 90: LIB_PATH=${GCC_PATH%/*/*} 91: GCC_LD="\$(LD)" 92: platform=${withval} 93: fi]) 94: 95: #variables mentioned in INSTALL 96: AC_ARG_VAR(CC, [The C compiler (must support GNU C 2.x).]) 97: AC_ARG_VAR(FORTHSIZES, [Gforth command line options for the default stack and dictionary sizes (see INSTALL).]) 98: AC_ARG_VAR(STACK_CACHE_REGS, [number of registers in the maximum stack cache state for gforth-fast and gforth-native (default platform-dependent).]) 99: AC_ARG_VAR(STACK_CACHE_DEFAULT_FAST, [number of registers in the default stack cache state for gforth-fast and gforth-native (default 1).]) 100: AC_ARG_VAR(GCC_PR15242_WORKAROUND, [Force the enabling (1) or disabling (0) of a workaround for a gcc-3.x performance bug (default unset: use workaround for gcc-3.x)]) 101: 102: AC_ARG_VAR(ac_cv_sizeof_char_p, [sizeof(char *)]) 103: AC_ARG_VAR(ac_cv_sizeof_void_p, [sizeof(void *)]) 104: AC_ARG_VAR(ac_cv_sizeof_char, [sizeof(char)]) 105: AC_ARG_VAR(ac_cv_sizeof_short, [sizeof(short)]) 106: AC_ARG_VAR(ac_cv_sizeof_int, [sizeof(int)]) 107: AC_ARG_VAR(ac_cv_sizeof_long, [sizeof(long)]) 108: AC_ARG_VAR(ac_cv_sizeof_long_long, [sizeof(long long)]) 109: AC_ARG_VAR(ac_cv_sizeof_intptr_t, [sizeof(intptr_t)]) 110: AC_ARG_VAR(ac_cv_c_bigendian, [Is the target big-endian ("yes" or "no")?]) 111: AC_ARG_VAR(no_dynamic_default, [run gforth with --dynamic (0) or --no-dynamic (1) by default]) 112: AC_ARG_VAR(condbranch_opt, [enable (1) or disable (0) using two dispatches for conditional branches]) 113: AC_ARG_VAR(skipcode, [assembly code for skipping 16 bytes of code]) 114: AC_ARG_VAR(asmcomment, [assembler comment start string]) 115: AC_ARG_VAR(arm_cacheflush, [file containing ARM cacheflush function (without .c)]) 116: 117: #set up feature test macros, so the tests get them right: 118: # turn on all POSIX, SUSv3, and GNU features if available 119: AC_GNU_SOURCE 120: dnl AC_DEFINE_UNQUOTED([_GNU_SOURCE],1,[feature test macro]) 121: 122: dnl Don't define _POSIX_SOURCE etc. because some OSs (in particular 123: dnl MacOSX) disable some features then (MacOSX checks for _POSIX_SOURCE, 124: dnl but not for _XOPEN_SOURCE) 125: dnl AC_DEFINE_UNQUOTED([_POSIX_SOURCE],1,[feature test macro]) 126: dnl AC_DEFINE_UNQUOTED([_POSIX_C_SOURCE],199506L,[feature test macro]) 127: dnl AC_DEFINE_UNQUOTED([_XOPEN_SOURCE],600,[feature test macro]) 128: # turn on large file support with 64-bit off_t where available 129: AC_SYS_LARGEFILE 130: dnl AC_DEFINE_UNQUOTED([_LARGEFILE_SOURCE],1,[feature test macro]) 131: dnl AC_DEFINE_UNQUOTED([_FILE_OFFSET_BITS],64,[feature test macro]) 132: 133: #currently we force direct threading this way. Eventually we should 134: #setup in the arch and engine files right 135: 136: AC_PROG_CC 137: 138: test "$GCC" = "yes" || AC_MSG_ERROR(Gforth uses GNU C extensions and requires GCC 2.0 or higher) 139: 140: AC_MSG_CHECKING([whether to use two dispatches per conditional branch]) 141: test x$condbranch_opt = x && 142: if ($CC -v 2>&1 |grep -q 'gcc version 3'); then 143: condbranch_opt=0 144: else 145: condbranch_opt=1 146: fi 147: AC_MSG_RESULT($condbranch_opt) 148: AC_SUBST(condbranch_opt) 149: 150: AC_SUBST(CC) 151: AC_SUBST(GCC_LD) 152: AC_SUBST(DEBUGFLAG) 153: AC_SUBST(EC) 154: AC_SUBST(EC_MODE) 155: AC_SUBST(engine2) 156: AC_SUBST(engine_fast2) 157: AC_SUBST(no_dynamic) 158: AC_SUBST(image_i) 159: AC_SUBST(signals_o) 160: 161: #this is used to disable some (not generally essential) part of the 162: #Makefile that some makes don't grok. It would be better to test for 163: #this specific Makefile feature than the make version. 164: AC_MSG_CHECKING(make type) 165: make_type=`make -n -v 2>&1|grep 'ake'|sed 's/ake .*/ake/'` 166: GNUMAKE='#' 167: test "$make_type" = "GNU Make" && GNUMAKE='' 168: AC_MSG_RESULT($make_type) 169: AC_SUBST(GNUMAKE) 170: 171: AC_MSG_CHECKING([whether the linker accepts -export-dynamic]) 172: OLDLDFLAGS=$LDFLAGS 173: LDFLAGS="$LDFLAGS -export-dynamic" 174: dnl AC_TRY_LINK gives false positive on rs6000-ibm-aix4.2.1.0 175: dnl AC_TRY_LINK(,,ac_export_dynamic=yes,ac_export_dynamic=no) 176: AC_TRY_RUN(main(){exit(0);},ac_export_dynamic=yes,ac_export_dynamic=no,ac_export_dynamic=no) 177: test $ac_export_dynamic = yes|| LDFLAGS=$OLDLDFLAGS 178: AC_MSG_RESULT($ac_export_dynamic) 179: 180: #terminology is a bit unusual here: The host is the system on which 181: #gforth will run; the system on which configure will run is the `build' 182: AC_CANONICAL_HOST 183: case "$host_cpu" in 184: arm*) 185: machine=arm 186: CFLAGS="$CFLAGS -fomit-frame-pointer" 187: if test x$platform = xnxt; then 188: CFLAGS="$CFLAGS -mthumb -mthumb-interwork" 189: fi 190: if test -z $arm_cacheflush; then 191: case "$host_os" in 192: *linux*) 193: arm_cacheflush=arch/arm/cacheflush-linux 194: ;; 195: *) 196: no_dynamic_default=1 197: arm_cacheflush=arch/arm/cacheflush0 198: AC_MSG_WARN([No I-cache flush code known, disabling dynamic native code generation]) 199: ;; 200: esac 201: fi 202: AC_LIBSOURCES([../arch/arm/cacheflush0, dnl 203: ../arch/arm/cacheflush-linux]) 204: AC_LIBOBJ(../$arm_cacheflush) 205: #longer skipcodes lead to problems on ARM, and it uses 206: #only 4-byte alignment anyway 207: test "$skipcode" || skipcode="nop" 208: ;; 209: hppa*) 210: machine=hppa 211: $srcdir/mkinstalldirs arch/hppa 212: AC_LIBOBJ(../arch/hppa/cache) 213: #-N needed for --dynamic <[email protected]> 214: LDFLAGS="$LDFLAGS -Xlinker -N" 215: LIBS="$LIBS -L/lib/pa1.1/" 216: ;; 217: sparc*) 218: machine=sparc 219: ;; 220: i386) 221: machine=386 222: CFLAGS="$CFLAGS -fomit-frame-pointer -fforce-addr" 223: ;; 224: i486) 225: machine=386 226: CFLAGS="$CFLAGS -fomit-frame-pointer -fforce-addr -m486" 227: ;; 228: i*86) 229: machine=386 230: CFLAGS="$CFLAGS -fomit-frame-pointer -fforce-addr" 231: CFLAGS_1="$CFLAGS" 232: CFLAGS="$CFLAGS -march=pentium" 233: AC_TRY_COMPILE(,,,CFLAGS="$CFLAGS_1 -m486") 234: ;; 235: x86_64) 236: case $CC 237: in 238: *-m32*) 239: machine=386 240: CFLAGS="$CFLAGS -fomit-frame-pointer -fforce-addr" 241: CFLAGS_1="$CFLAGS" 242: CFLAGS="$CFLAGS -march=athlon64" 243: ;; 244: *) 245: machine=amd64 246: ;; 247: esac 248: ;; 249: ia64*) 250: machine=ia64 251: AC_LIBOBJ(../arch/ia64/flush_icache_block) 252: test "$skipcode" || skipcode="nop.i 0" 253: #".skip 16" passes the test below, 254: # but gives an assembler error in engine 255: ;; 256: m68k) 257: machine=m68k 258: CFLAGS="$CFLAGS -fomit-frame-pointer" 259: if test "$host_os" = "nextstep3" 260: then 261: AC_LIBOBJ(termios) 262: fi 263: ;; 264: mips*) 265: machine=mips 266: #dynamic native code has the following problems on MIPS: 267: # 268: #1) J/JAL seems relocatable, but is are only 269: #relocatable within a 256MB-segment. While we try to 270: #get the linker to arrange this, there is no guarantee 271: #that this will succeed (and if the user uses a lot of 272: #memory, it is likely to fail). 273: # 274: #2) The way we generate dynamic native code may 275: #violate MIPS architectural restrictions (in 276: #particular, the delay slots of LW, MFLO, etc.) 277: # 278: #Therefore we disable dynamic native code by default: 279: if test -z $no_dynamic_default; then 280: no_dynamic_default=1 281: AC_MSG_WARN([Disabling default dynamic native code generation (relocation and delay slot issues)]) 282: fi 283: ;; 284: alpha*) 285: machine=alpha 286: #full IEEE FP support for more uniformity across platforms: 287: CFLAGS="$CFLAGS -mieee" 288: ;; 289: power*|rs6000) 290: machine=power 291: $srcdir/mkinstalldirs arch/power 292: AC_CHECK_FUNC(_sync_cache_range,[true],[AC_LIBOBJ(../arch/power/_sync_cache_range)]) 293: #long long is broken on (at least) gcc-2.95.* for PPC 294: test x$ac_cv_sizeof_long_long = x && 295: ($CC -v 2>&1 |grep -q 'gcc version 2.95') && 296: ac_cv_sizeof_long_long=0 297: #The only architecture with enough callee-saved registers 298: test x$STACK_CACHE_REGS = x && STACK_CACHE_REGS=3 299: #or use 2, hardly slower at run-time and starts up faster 300: ;; 301: *) 302: AC_MSG_WARN([Using a generic machine description]) 303: AC_MSG_WARN([Assuming C floats and doubles are IEEE floats and doubles (for SF@ DF@ SF! DF!)]) 304: AC_MSG_WARN([FLUSH-ICACHE will do nothing, so END-CODE may not work properly!]) 305: machine=generic 306: #I-cache flushing would be needed for dynamic code generation 307: if test -z $no_dynamic_default; then 308: no_dynamic_default=1 309: AC_MSG_WARN([No I-cache flush code known, disabling dynamic native code generation]) 310: fi 311: esac 312: AC_SUBST(host) 313: 314: MAKEINC="" 315: 316: echo "Check for arch/$machine/$platform/gforth.ld ($EC_MODE)" 317: if test x$EC_MODE = xtrue 318: then 319: echo "Check for arch/$machine/$platform/gforth.ld" 320: if test -f arch/$machine/$platform/gforth.ld 321: then 322: LDFLAGS="-T ../arch/$machine/$platform/gforth.ld -Map \[email protected] -cref --gc-sections $LDFLAGS" 323: if test x$platform = xnxt; then 324: LIBS="$LIB_PATH/lib/gcc/arm-elf/$($CC --version | grep GCC | cut -d' ' -f3)/interwork/libgcc.a $LIB_PATH/arm-elf/lib/interwork/libc.a $LIBS" 325: fi 326: fi 327: if test -f arch/$machine/$platform/make.inc 328: then 329: MAKEINC="include ../arch/$machine/$platform/make.inc" 330: fi 331: fi 332: AC_SUBST(MAKEINC) 333: 334: AC_ARG_VAR(STACK_CACHE_REGS, [number of registers in the maximum stack cache state for gforth-fast and gforth-native (default platform-dependent).]) 335: 336: test x$STACK_CACHE_REGS = x && STACK_CACHE_REGS=1 337: AC_DEFINE_UNQUOTED(STACK_CACHE_REGS, $STACK_CACHE_REGS, 338: [number of registers in the maximum stack cache state for gforth-fast and gforth-native]) 339: test x$STACK_CACHE_DEFAULT_FAST = x && STACK_CACHE_DEFAULT_FAST=1 340: AC_DEFINE_UNQUOTED(STACK_CACHE_DEFAULT_FAST, $STACK_CACHE_DEFAULT_FAST, 341: [number of registers in the default stack cache state for gforth-fast and gforth-native]) 342: 343: test x$GCC_PR15242_WORKAROUND = x || 344: AC_DEFINE_UNQUOTED(GCC_PR15242_WORKAROUND, $GCC_PR15242_WORKAROUND, 345: [force (1) or forbid (0) use of a workaround for a gcc performance bug]) 346: 347: dnl the following macro produces a warning with autoconf-2.1 348: AC_CHECK_SIZEOF(char *) 349: case "$ac_cv_sizeof_char_p" in 350: 2) 351: wordsize=16 352: ;; 353: 4) 354: wordsize=32 355: ;; 356: 8) 357: wordsize=64 358: ;; 359: esac 360: 361: AC_CHECK_SIZEOF(void *) 362: case "$ac_cv_sizeof_void_p" in 363: 2) 364: vwordsize=16 365: ;; 366: 4) 367: vwordsize=32 368: ;; 369: 8) 370: vwordsize=64 371: ;; 372: esac 373: 374: AC_CHECK_SIZEOF(char) 375: AC_CHECK_SIZEOF(short) 376: AC_CHECK_SIZEOF(int) 377: AC_CHECK_SIZEOF(long) 378: AC_CHECK_SIZEOF(long long) 379: AC_CHECK_SIZEOF(intptr_t) 380: AC_CHECK_SIZEOF(int128_t) 381: AC_CHECK_SIZEOF(uint128_t) 382: 383: AC_MSG_CHECKING([for a C type for cells]) 384: ac_cv_int_type_cell=none 385: case "$ac_cv_sizeof_char_p" in 386: $ac_cv_sizeof_int) 387: ac_cv_int_type_cell=int 388: ;; 389: $ac_cv_sizeof_short) 390: ac_cv_int_type_cell=short 391: ;; 392: $ac_cv_sizeof_char) 393: ac_cv_int_type_cell=char 394: ;; 395: $ac_cv_sizeof_long) 396: ac_cv_int_type_cell=long 397: ;; 398: $ac_cv_sizeof_long_long) 399: ac_cv_int_type_cell="long long" 400: ;; 401: $ac_cv_sizeof_intptr_t) 402: ac_cv_int_type_cell="intptr_t" 403: ;; 404: esac 405: AC_MSG_RESULT($ac_cv_int_type_cell) 406: AC_DEFINE_UNQUOTED(CELL_TYPE,$ac_cv_int_type_cell,[an integer type that is as long as a pointer]) 407: 408: AC_MSG_CHECKING([for a C type for wydes]) 409: ac_cv_wyde_type_cell=none 410: case 2 in 411: $ac_cv_sizeof_int) 412: ac_cv_wyde_type_cell=int 413: ;; 414: $ac_cv_sizeof_short) 415: ac_cv_wyde_type_cell=short 416: ;; 417: $ac_cv_sizeof_char) 418: ac_cv_wyde_type_cell=char 419: ;; 420: $ac_cv_sizeof_long) 421: ac_cv_wyde_type_cell=long 422: ;; 423: $ac_cv_sizeof_long_long) 424: ac_cv_wyde_type_cell="long long" 425: ;; 426: $ac_cv_sizeof_intptr_t) 427: ac_cv_wyde_type_cell="intptr_t" 428: ;; 429: esac 430: AC_MSG_RESULT($ac_cv_wyde_type_cell) 431: AC_DEFINE_UNQUOTED(WYDE_TYPE,$ac_cv_wyde_type_cell,[an integer type that is 2 bytes long]) 432: 433: AC_MSG_CHECKING([for a C type for tetrabytes]) 434: ac_cv_tetrabyte_type_cell=none 435: case 4 in 436: $ac_cv_sizeof_int) 437: ac_cv_tetrabyte_type_cell=int 438: ;; 439: $ac_cv_sizeof_short) 440: ac_cv_tetrabyte_type_cell=short 441: ;; 442: $ac_cv_sizeof_char) 443: ac_cv_tetrabyte_type_cell=char 444: ;; 445: $ac_cv_sizeof_long) 446: ac_cv_tetrabyte_type_cell=long 447: ;; 448: $ac_cv_sizeof_long_long) 449: ac_cv_tetrabyte_type_cell="long long" 450: ;; 451: $ac_cv_sizeof_intptr_t) 452: ac_cv_tetrabyte_type_cell="intptr_t" 453: ;; 454: esac 455: AC_MSG_RESULT($ac_cv_tetrabyte_type_cell) 456: AC_DEFINE_UNQUOTED(TETRABYTE_TYPE,$ac_cv_tetrabyte_type_cell,[an integer type that is 4 bytes long]) 457: 458: AC_MSG_CHECKING([for a C type for double-cells]) 459: ac_cv_int_type_double_cell=none 460: case `expr 2 '*' "$ac_cv_sizeof_char_p"` in 461: $ac_cv_sizeof_short) 462: ac_cv_int_type_double_cell=short 463: ;; 464: $ac_cv_sizeof_int) 465: ac_cv_int_type_double_cell=int 466: ;; 467: $ac_cv_sizeof_long) 468: ac_cv_int_type_double_cell=long 469: ;; 470: $ac_cv_sizeof_long_long) 471: ac_cv_int_type_double_cell="long long" 472: ;; 473: $ac_cv_sizeof_intptr_t) 474: ac_cv_int_type_double_cell="intptr_t" 475: ;; 476: $ac_cv_sizeof_int128_t) 477: ac_cv_int_type_double_cell="int128_t" 478: ;; 479: esac 480: AC_MSG_RESULT($ac_cv_int_type_double_cell) 481: 482: AC_MSG_CHECKING([for a C type for unsigned double-cells]) 483: ac_cv_int_type_double_ucell=none 484: case `expr 2 '*' "$ac_cv_sizeof_char_p"` in 485: $ac_cv_sizeof_short) 486: ac_cv_int_type_double_ucell="unsigned short" 487: ;; 488: $ac_cv_sizeof_int) 489: ac_cv_int_type_double_ucell="unsigned int" 490: ;; 491: $ac_cv_sizeof_long) 492: ac_cv_int_type_double_ucell="unsigned long" 493: ;; 494: $ac_cv_sizeof_long_long) 495: ac_cv_int_type_double_ucell="unsigned long long" 496: ;; 497: $ac_cv_sizeof_intptr_t) 498: ac_cv_int_type_double_ucell="unsigned intptr_t" 499: ;; 500: $ac_cv_sizeof_uint128_t) 501: ac_cv_int_type_double_ucell="uint128_t" 502: ;; 503: esac 504: AC_MSG_RESULT($ac_cv_int_type_double_ucell) 505: 506: if test "$ac_cv_int_type_double_cell" != none && \ 507: test "$ac_cv_int_type_double_ucell" != none 508: then 509: AC_DEFINE_UNQUOTED(DOUBLE_CELL_TYPE,$ac_cv_int_type_double_cell,[an integer type that is twice as long as a pointer]) 510: AC_DEFINE_UNQUOTED(DOUBLE_UCELL_TYPE,$ac_cv_int_type_double_ucell,[an unsigned integer type that is twice as long as a pointer]) 511: OPTS=-ll 512: else 513: if test "$ac_cv_sizeof_char_p" == 8; then 514: OPTS="-ll -noll" 515: else 516: OPTS=-noll 517: fi 518: fi 519: 520: if grep -q FORCE_REG arch/$machine/machine.h; then 521: OPTS=`for i in $OPTS; do echo -n "$i-reg "; done`$OPTS 522: fi 523: AC_SUBST(OPTS) 524: 525: AC_TYPE_OFF_T 526: AC_CHECK_SIZEOF(off_t) 527: test $ac_cv_sizeof_off_t -gt $ac_cv_sizeof_char_p 528: ac_small_off_t=$? 529: AC_DEFINE_UNQUOTED(SMALL_OFF_T,$ac_small_off_t,[1 if off_t fits in a Cell]) 530: 531: ENGINE_FLAGS= 532: AC_SUBST(ENGINE_FLAGS) 533: 534: # Try if GCC understands -fno-gcse 535: 536: AC_MSG_CHECKING([if $CC understands -fno-gcse]) 537: CFLAGS_1="$CFLAGS" 538: CFLAGS="$CFLAGS -fno-gcse" 539: AC_TRY_COMPILE(,,ac_nogcse=yes;ENGINE_FLAGS="$ENGINE_FLAGS -fno-gcse",ac_nogcse=no) 540: CFLAGS="$CFLAGS_1" 541: AC_MSG_RESULT($ac_nogcse) 542: 543: # Try if GCC understands -fno-strict-aliasing 544: AC_MSG_CHECKING([if $CC understands -fno-strict-aliasing]) 545: CFLAGS_1="$CFLAGS" 546: CFLAGS="$CFLAGS -fno-strict-aliasing" 547: AC_TRY_COMPILE(,,ac_nostrictaliasing=yes;ENGINE_FLAGS="$ENGINE_FLAGS -fno-strict-aliasing",ac_nostrictaliasing=no) 548: CFLAGS="$CFLAGS_1" 549: AC_MSG_RESULT($ac_nostrictaliasing) 550: 551: # Try if GCC understands -fno-crossjumping 552: AC_MSG_CHECKING([if $CC understands -fno-crossjumping]) 553: CFLAGS_1="$CFLAGS" 554: CFLAGS="$CFLAGS -fno-crossjumping" 555: AC_TRY_COMPILE(,,ac_nocrossjumping=yes;ENGINE_FLAGS="$ENGINE_FLAGS -fno-crossjumping",ac_nocrossjumping=no) 556: CFLAGS="$CFLAGS_1" 557: AC_MSG_RESULT($ac_nocrossjumping) 558: 559: # Try if GCC understands -fno-reorder-blocks 560: AC_MSG_CHECKING([if $CC understands -fno-reorder-blocks]) 561: CFLAGS_1="$CFLAGS" 562: CFLAGS="$CFLAGS -fno-reorder-blocks" 563: AC_TRY_COMPILE(,,ac_noreorder_blocks=yes;ENGINE_FLAGS="$ENGINE_FLAGS -fno-reorder-blocks",ac_noreorder_blocks=no) 564: CFLAGS="$CFLAGS_1" 565: AC_MSG_RESULT($ac_noreorder_blocks) 566: 567: # Try if GCC understands -falign-labels=1 568: AC_MSG_CHECKING([if $CC understands -falign-labels=1]) 569: CFLAGS_1="$CFLAGS" 570: CFLAGS="$CFLAGS -falign-labels=1" 571: AC_TRY_COMPILE(,,ac_align_labels=yes;ENGINE_FLAGS="$ENGINE_FLAGS -falign-labels=1",ac_align_labels=no) 572: CFLAGS="$CFLAGS_1" 573: AC_MSG_RESULT($ac_align_labels) 574: 575: # Try if GCC understands -falign-loops=1 576: AC_MSG_CHECKING([if $CC understands -falign-loops=1]) 577: CFLAGS_1="$CFLAGS" 578: CFLAGS="$CFLAGS -falign-loops=1" 579: AC_TRY_COMPILE(,,ac_align_loops=yes;ENGINE_FLAGS="$ENGINE_FLAGS -falign-loops=1",ac_align_loops=no) 580: CFLAGS="$CFLAGS_1" 581: AC_MSG_RESULT($ac_align_loops) 582: 583: # Try if GCC understands -falign-jumps=1 584: AC_MSG_CHECKING([if $CC understands -falign-jumps=1]) 585: CFLAGS_1="$CFLAGS" 586: CFLAGS="$CFLAGS -falign-jumps=1" 587: AC_TRY_COMPILE(,,ac_align_jumps=yes;ENGINE_FLAGS="$ENGINE_FLAGS -falign-jumps=1",ac_align_jumps=no) 588: CFLAGS="$CFLAGS_1" 589: AC_MSG_RESULT($ac_align_jumps) 590: 591: # Try if GCC understands __attribute__((unused)) 592: AC_MSG_CHECKING([how to suppress 'unused variable' warnings]) 593: AC_TRY_COMPILE(,[int __attribute__((unused)) foo;], MAYBE_UNUSED='__attribute__((unused))',) 594: AC_DEFINE_UNQUOTED(MAYBE_UNUSED,$MAYBE_UNUSED,[attribute for possibly unused variables]) 595: AC_MSG_RESULT($MAYBE_UNUSED) 596: 597: #try if m4 understands -s 598: AC_MSG_CHECKING([how to invoke m4]) 599: if m4 -s /dev/null >/dev/null 2>&1; then 600: M4="m4 -s" 601: else 602: M4=m4 603: fi 604: AC_SUBST(M4) 605: AC_DEFINE_UNQUOTED(M4,"$M4",[How to invoke m4]) 606: AC_MSG_RESULT($M4) 607: 608: # Find installed Gforth 609: AC_MSG_CHECKING([for gforth]) 610: GFORTH="`cd / && which gforth 2>/dev/null`" 611: if test -z "$GFORTH"; then 612: PREFORTH='echo "You need to configure with a gforth in \$PATH to build this part" && false' 613: else 614: PREFORTH="$GFORTH -i `cd / && gforth --debug -e bye 2>&1 |grep "Opened image file: "|sed 's/Opened image file: //'`" ; 615: fi 616: AC_SUBST(PREFORTH) 617: AC_DEFINE_UNQUOTED(PREFORTH,"$PREFORTH",[How to invoke the pre-installed gforth]) 618: AC_MSG_RESULT($PREFORTH) 619: 620: #echo "machine='$machine'" 621: 622: dnl AC_CHECK_PROG(asm_fs,asm.fs,arch/$machine/asm.fs,,$srcdir/arch/$machine) 623: AC_CHECK_FILE($srcdir/arch/$machine/asm.fs,[asm_fs=arch/$machine/asm.fs],) 624: AC_SUBST(asm_fs) 625: 626: dnl AC_CHECK_PROG(disasm_fs,disasm.fs,arch/$machine/disasm.fs,,$srcdir/arch/$machine) 627: AC_CHECK_FILE($srcdir/arch/$machine/disasm.fs,[disasm_fs=arch/$machine/disasm.fs],) 628: AC_SUBST(disasm_fs) 629: 630: AC_PATH_PROG(INSTALL_INFO,install-info,[echo '>>>>Please make info dir entry:'],$PATH:/sbin:/usr/sbin:/usr/local/sbin) 631: 632: case "$host_os" in 633: *win32*) 634: # !!!FIXME!!! problems with cygwin and ';' as path separator 635: DIRSEP="\\\\" 636: PATHSEP=";" 637: #we want the builtins of command.com/cmd.exe and its 638: # handling of .com files. 639: #$COMSPEC contains the name of the Windows shell; 640: # the ./ is there, because the bash does not recognize 641: # absolute DOS filenames 642: DEFAULTSYSTEMPREFIX="./$COMSPEC /c " 643: ;; 644: *darwin*) 645: #Darwin uses some funny preprocessor by default; eliminate it: 646: AC_MSG_NOTICE([using -no-cpp-precomp on Darwin]) 647: CFLAGS="$CFLAGS -no-cpp-precomp" 648: DIRSEP="/" 649: PATHSEP=":" 650: DEFAULTSYSTEMPREFIX="" 651: ;; 652: *) 653: DIRSEP="/" 654: PATHSEP=":" 655: DEFAULTSYSTEMPREFIX="" 656: ;; 657: esac 658: AC_SUBST(DIRSEP) 659: AC_DEFINE_UNQUOTED(DIRSEP,'$DIRSEP',[a directory separator character]) 660: AC_SUBST(PATHSEP) 661: AC_DEFINE_UNQUOTED(PATHSEP,'$PATHSEP',[a path separator character]) 662: AC_SUBST(DEFAULTSYSTEMPREFIX) 663: AC_DEFINE_UNQUOTED(DEFAULTSYSTEMPREFIX,"$DEFAULTSYSTEMPREFIX",[default for environment variable GFORTHSYSTEMPREFIX]) 664: 665: #work around SELinux brain damage (from Andrew Haley <[email protected]>) 666: #This magic incantation seems to be completely undocumented. 667: AC_CHECK_PROG([MASSAGE_EXE],[chcon],[chcon -t unconfined_execmem_exec_t],[true]) 668: 669: dnl Now a little support for DOS/DJGCC 670: AC_SUBST(GFORTH_EXE) 671: GFORTH_EXE="true" 672: AC_SUBST(GFORTHFAST_EXE) 673: GFORTHFAST_EXE="true" 674: AC_SUBST(GFORTHITC_EXE) 675: GFORTHITC_EXE="true" 676: AC_SUBST(GFORTHDITC_EXE) 677: GFORTHDITC_EXE="true" 678: 679: AC_SUBST(FORTHSIZES) 680: 681: dnl if test "$PEEPHOLE" = "yes" 682: dnl then 683: dnl PEEPHOLEFLAG="true" 684: dnl AC_DEFINE(HAS_PEEPHOLE,,[Define if you want to use peephole optimization]) 685: dnl else 686: dnl PEEPHOLEFLAG="false" 687: dnl fi 688: PEEPHOLEFLAG="true" 689: AC_SUBST(PEEPHOLEFLAG) 690: 691: dnl copy commands for systems that don't have links 692: AC_SUBST(LINK_KERNL) 693: LINK_KERNL="" 694: 695: #if test $host_os=dos 696: #then 697: # echo Configuring for DOS!!! 698: # MAKE_EXE="coff2exe gforth" 699: # LINK_KERNL='$(CP) kernl32l.fi kernel.fi' 700: #fi 701: 702: dnl the following macro produces a warning with autoconf-2.1 703: AC_C_BIGENDIAN 704: AC_SUBST(KERNEL) 705: dnl ac_cv_c_bigendian is an undocumented variable of autoconf-2.1 706: if test $ac_cv_c_bigendian = yes; then 707: bytesex=b 708: KERNEL="kernl16b.fi kernl16l.fi kernl32b.fi kernl32l.fi kernl64b.fi kernl64l.fi" 709: else 710: bytesex=l 711: KERNEL="kernl16l.fi kernl16b.fi kernl32l.fi kernl32b.fi kernl64l.fi kernl64b.fi" 712: fi 713: 714: #check how to do asm(".skip 16") 715: #echo "CFLAGS=$CFLAGS" 716: #echo "ac_link=$ac_link" 717: AC_MSG_CHECKING([if and how we can waste code space]) 718: if test -z "$skipcode"; then 719: skipcode=no 720: CFLAGS_1="$CFLAGS" 721: CFLAGS="$CFLAGS $ENGINE_FLAGS" 722: for i in ".skip 16" ".block 16" ".org .+16" ".=.+16" ".space 16" 723: do 724: AC_TRY_RUN( 725: [int foo(int,int,int); 726: main() 727: { 728: exit(foo(0,0,0)!=16); 729: } 730: int foo(int x, int y, int z) 731: { 732: static void *labels[]={&&label1, &&label2}; 733: if (x) { 734: y++; /* workaround for http://gcc.gnu.org/bugzilla/show_bug.cgi?id=12108 */ 735: label1: 736: asm("$i"); /* or ".space 16" or somesuch */ 737: label2: ; 738: } 739: { 740: if (y) goto *labels[z]; /* workaround for gcc PR12108 */ 741: return labels[1]-labels[0]; 742: } 743: }] 744: ,skipcode=$i; break 745: ,,) 746: done 747: CFLAGS=$CFLAGS_1 748: fi 749: AC_MSG_RESULT($skipcode) 750: if test "$skipcode" = no 751: then 752: if test -z $no_dynamic_default; then 753: no_dynamic_default=1 754: AC_MSG_WARN(Disabling default dynamic native code generation) 755: fi 756: AC_DEFINE_UNQUOTED(SKIP16,((void)0),statement for skipping 16 bytes) 757: else 758: AC_DEFINE_UNQUOTED(SKIP16,asm("$skipcode"),statement for skipping 16 bytes) 759: fi 760: 761: AC_MSG_CHECKING([if and how we can do comments in asm statements]) 762: #the point here is to get asm statements that look different to 763: #gcc's "optimizer" 764: if test -z "$asmcomment"; then 765: asmcomment=no 766: CFLAGS_1="$CFLAGS" 767: CFLAGS="$CFLAGS $ENGINE_FLAGS" 768: for i in '"# "' '"! "' '"; "'; do 769: AC_TRY_COMPILE(,[asm($i"fluffystunk");],asmcomment=$i; break,) 770: done 771: CFLAGS=$CFLAGS_1 772: fi 773: AC_MSG_RESULT($asmcomment) 774: if test "$asmcomment" != no 775: then 776: AC_DEFINE_UNQUOTED(ASMCOMMENT,$asmcomment,[assembler comment start string]) 777: fi 778: 779: test "$no_dynamic_default" || no_dynamic_default=0 780: AC_DEFINE_UNQUOTED(NO_DYNAMIC_DEFAULT,$no_dynamic_default,default value for no_dynamic) 781: 782: dnl Checks for programs. 783: AC_PROG_LN_S 784: AC_PROG_INSTALL 785: AC_CHECK_PROGS(TEXI2DVI,texi2dvi4a2ps texi2dvi) 786: 787: dnl MacOS X has a libtool that does something else 788: AC_CHECK_PROGS(GNU_LIBTOOL,glibtool libtool) 789: 790: dnl Checks for library functions 791: dnl This check is just for making later checks link with libm. 792: dnl using sin here is no good idea since it is built-into gcc and typechecked 793: AC_CHECK_LIB(m,asin) 794: AC_CHECK_LIB(ltdl,lt_dlinit) 795: AC_CHECK_LIB(dl,dlopen) 796: dnl check for libffi 2.x 797: AC_CHECK_LIB(ffi,ffi_call) 798: if test $ac_cv_lib_ffi_ffi_call = yes 799: then 800: LIBFFIFLAG="true" 801: FFCALLFLAG="false" 802: OLDCALLFLAG="false" 803: AC_DEFINE(HAS_LIBFFI,,[define this if you want to use the ffcall interface with libffi 2.0]) 804: else 805: dnl check for ffcall libraries 806: dnl unfortunately, these four calls are separated out into a library each. 807: AC_CHECK_LIB(avcall,__builtin_avcall) 808: AC_CHECK_LIB(callback,__vacall_r) 809: AC_CHECK_LIB(vacall,vacall) 810: AC_CHECK_LIB(trampoline,alloc_trampoline) 811: LIBFFIFLAG="false" 812: FFCALLFLAG="false" 813: OLDCALLFLAG="true" 814: test $ac_cv_lib_avcall___builtin_avcall = yes && FFCALLFLAG="true" && OLDCALLFLAG="false" && AC_DEFINE(HAS_FFCALL,,[define this if you want to use the ffcall libraries]) 815: test $ac_cv_lib_avcall___builtin_avcall = no && AC_DEFINE(HAS_OLDCALL,,[define this if you want to use the old call libraries]) 816: fi 817: AC_SUBST(LIBFFIFLAG) 818: AC_SUBST(FFCALLFLAG) 819: AC_SUBST(OLDCALLFLAG) 820: if test "$host_os" != "nextstep3" 821: then 822: AC_FUNC_MEMCMP 823: fi 824: AC_REPLACE_FUNCS(memmove strtoul pow10 strerror strsignal atanh) 825: AC_FUNC_FSEEKO 826: AC_CHECK_FUNCS(ftello dlopen sys_siglist getrusage nanosleep) 827: AC_CHECK_TYPES(stack_t,,,[#include <signal.h>]) 828: AC_DECL_SYS_SIGLIST 829: AC_CHECK_FUNC(getopt_long,[true],[AC_LIBOBJ(getopt) AC_LIBOBJ(getopt1)]) 830: AC_CHECK_FUNCS(expm1 log1p) 831: AC_REPLACE_FUNCS(rint ecvt) 832: dnl No check for select, because our replacement is no good under 833: dnl anything but DOS 834: AC_CHECK_HEADERS(sys/mman.h fnmatch.h) 835: AC_FUNC_FNMATCH 836: test $ac_cv_func_fnmatch_works = yes || AC_LIBOBJ(fnmatch) 837: AC_CHECK_FUNCS(mmap sysconf getpagesize) 838: AM_PATH_LISPDIR 839: 840: kernel_fi=kernl${vwordsize}${bytesex}.fi 841: include_fi=kernl${wordsize}${bytesex}${EC}.fi 842: AC_SUBST(kernel_fi) 843: AC_SUBST(include_fi) 844: 845: #this breaks bindists 846: #dnl replace srource directory by absolute value 847: #if test $srcdir = "."; then srcdir=`pwd` 848: #fi 849: 850: AC_SUBST(machine) 851: AC_CONFIG_FILES([ 852: Makefile 853: Makedist 854: gforthmi 855: vmgen 856: machpc.fs 857: envos.fs 858: engine/Makefile 859: engine/libcc.h 860: doc/version.texi 861: build-ec ]) 862: AC_CONFIG_COMMANDS([stamp-h],[[echo timestamp > stamp-h 863: chmod +x gforthmi 864: chmod +x vmgen 865: chmod +x build-ec 866: test -d kernel||mkdir kernel 867: $srcdir/mkinstalldirs include/gforth/$PACKAGE_VERSION 868: ln -sf ../../../engine/config.h ../../../engine/libcc.h include/gforth/$PACKAGE_VERSION]],[[PACKAGE_VERSION=$PACKAGE_VERSION]]) 869: AC_OUTPUT 870: FreeBSD-CVSweb <[email protected]>
__label__pos
0.960003
支付宝红包 京东盲盒抽奖 幸运转盘 秒杀 自营热卖 支付宝红包 2000行代码用go语言实现的比特币基本的相关模型功能 雨后、云初霁 1年前   阅读数 281 0 |版权声明:本文为博主原创文章,未经博主允许不得转载。博客地址:https://blog.csdn.net/sgsgy5 前言:闲暇时期,参考了一些资料,用go简单的实现了比特币中的一些相关功能,实现完全大概2000行代码左右,现在刚利用闲暇时间写了一点小功能,大概500多行代码左右,只实现了基本的区块链接,存储是选择了github上面的一个开源库,bolt轻量级数据库,还有很多需要迭代完善,这个月内打算完善好,可供参考 • 区块的结构,和区块链的结构定义,一些相关功能没有加入,后期会加入,比如交易,UTXO等 • pow工作量证明,现在不够完善,后期会完善 • 钱包结点的相关功能,后期会完善 • 加密的相关功能,后期会完善 type Block struct { //区块的一个结构 Version uint64 //版本号 PrevBlockHash []byte //前区块哈希值 MerkelRoot []byte //这是一个哈希值,后面完善 TimeStamp uint64 //时间戳,从1970.1.1到现在的秒数 Difficulty uint64 //通过这个数字,算出一个哈希值:0x00010000000xxx,暂时写死难度,代码里面大概5个前导0以上本地普通电脑就跑不动了 Nonce uint64 // 这是我们要找的随机数,挖矿就找证书 Hash []byte //当前区块哈希值, 正常的区块不存在,我们为了方便放进来 Data []byte //数据本身,区块体,先用字符串表示,v4版本的时候会引用真正的交易结构 } //定义一个区块链结构,使用bolt数据库进行保存 type BlockChain struct { //存储在数据库中,会生成一个文件 //数据库的句柄 Db *bolt.DB //最后一个区块的哈希值 lastHash []byte //在内存中的临时值,只保存最后一个区块哈希 } const difficulty = 16 //难度值写死暂时,后期完善 //1. 定义工作量证明, block, 难度值 type ProofOfWork struct { //数据来源 block Block //难度值 target *big.Int //一个能够处理大数的内置的类型,有比较方法 } 上面是一些基本的结构, 这里比较主要的就是工作量证明函数 我们来看一下实现方式 func (pow *ProofOfWork) Run() ([]byte, uint64) { //1. 拿到区块数据 //block := pow.block //区块的哈希值 var currentHash [32]byte //挖矿的随机值 var nonce uint64 for { info := pow.prepareData(nonce) //2. 对数据做哈希运算 currentHash = sha256.Sum256(info) //3. 比较 //引用big.int,将获取的[]byte类型的哈希值转成big.int var currentHashInt big.Int currentHashInt.SetBytes(currentHash[:]) // -1 if x < y // 0 if x == y // +1 if x > y // //func (x *Int) Cmp(y *Int) (r int) { if currentHashInt.Cmp(pow.target) == -1 { //a. 比目标小,成功,返回哈希和nonce break } else { //b. 比目标大,继续nonce++ nonce++ } } return currentHash[:], nonce } 这里的哈希运算就是,加上随机数来运算哈希,一直找找到一个随机数使得哈希符合难度值 比目标哈希是00001xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx,那么就需要找比这个哈希小的哈希,就是前面一定要有5个0的哈希才符合要求,这样打包区块才是有效的, 我们这里就简单实现了,后期迭代会加入UTXO,交易机制,还有钱包结点相关功能,欢迎提意见, 如果想详细参考的见github地址:https://github.com/wumansgy/btcmodel 阅读更多 (function(){ function setArticleH(btnReadmore,posi){ var winH = $(window).height(); var articleBox = $("div.article_content"); var artH = articleBox.height(); if(artH > winH*posi){ articleBox.css({ 'height':winH*posi+'px', 'overflow':'hidden' }) btnReadmore.click(function(){ if(typeof window.localStorage === "object" && typeof window.csdn.anonymousUserLimit === "object"){ if(!window.csdn.anonymousUserLimit.judgment()){ window.csdn.anonymousUserLimit.Jumplogin(); return false; }else if(!currentUserName){ window.csdn.anonymousUserLimit.updata(); } } articleBox.removeAttr("style"); $(this).parent().remove(); }) }else{ btnReadmore.parent().remove(); } } var btnReadmore = $("#btn-readmore"); if(btnReadmore.length>0){ if(currentUserName){ setArticleH(btnReadmore,3); }else{ setArticleH(btnReadmore,1.2); } } })() 注意:本文归作者所有,未经作者允许,不得转载 全部评论: 0 我有话说:
__label__pos
0.834261
If Not True Then False Install Apache/PHP 7.1.2 on Fedora 25/24, CentOS/RHEL 7.3/6.8 - Comment Page: 4 This guide shows howto install Apache HTTP Server (httpd) with PHP 7.1.2 and following modules on Fedora 25/24/23, CentOS 7.3/6.8 and Red Hat (RHEL) 7.3/6.8 systems. OPcache (php-opcache) – The Zend OPcache provides faster PHP execution through opcode caching and optimization. APCu (php-pecl-apcu) – APCu userland caching CLI (php-cli) – Command-line interface for PHP PEAR... Categories: 259 Comments Leave a Comment Your email address will not be published. Required fields are marked * Input your comment. help You may use these HTML tags and attributes: <a href="" title=""> <abbr title=""> <acronym title=""> <b> <blockquote cite=""> <cite> <code> <del datetime=""> <em> <i> <q cite=""> <s> <strike> <strong> Input your name. Novi help !!!! i have this message when i run ‘systemctl status httpd.service’ in terminal httpd.service – The Apache HTTP Server (prefork MPM) Loaded: loaded (/usr/lib/systemd/system/httpd.service; disabled) Active: failed (Result: exit-code) since Thu, 07 Jun 2012 13:30:47 +0200; 3min 13s ago Process: 4165 ExecStart=/usr/sbin/httpd $OPTIONS -k start (code=exited, status=1/FAILURE) CGroup: name=systemd:/system/httpd.service Jun 07 13:30:47 localhost.localdomain httpd[4165]: [Thu Jun 07 13:30:47 2012] [crit] (2)No such …’. Jun 07 13:30:47 localhost.localdomain httpd[4165]: httpd: Could not reliably determine the serve…me Jun 07 13:30:47 localhost.localdomain httpd[4165]: (98)Address already in use: make_sock: could …43 Jun 07 13:30:47 localhost.localdomain httpd[4165]: no listening sockets available, shutting down Jun 07 13:30:47 localhost.localdomain httpd[4165]: Unable to open logs reply Reply JR Hi Ramy, Not any significant news yet, I found some Fedora 18 builds here. Of course it is possible compile all packages manually… reply Reply Venkata Krishnan Kindly help me solve this problem # /etc/init.d/httpd start Starting httpd: Syntax error on line 1010 of /etc/httpd/conf/httpd.conf: Invalid command ‘localhost’, perhaps misspelled or defined by a module not included in the server configuration [FAILED] # service httpd start Starting httpd: Syntax error on line 1010 of /etc/httpd/conf/httpd.conf: Invalid command ‘localhost’, perhaps misspelled or defined by a module not included in the server configuration [FAILED] reply Reply Venkata Krishnan Hello JR, I have posted the contents of the file httpd.conf. Please find the same here too: # # This is the main Apache server configuration file. It contains the # configuration directives that give the server its instructions. # See for detailed information. # In particular, see # # for a discussion of each configuration directive. # # # Do NOT simply read the instructions in here without understanding # what they do. They're here only as hints or reminders. If you are unsure # consult the online docs. You have been warned. # # The configuration directives are grouped into three basic sections: # 1. Directives that control the operation of the Apache server process as a # whole (the 'global environment'). # 2. Directives that define the parameters of the 'main' or 'default' server, # which responds to requests that aren't handled by a virtual host. # These directives also provide default values for the settings # of all virtual hosts. # 3. Settings for virtual hosts, which allow Web requests to be sent to # different IP addresses or hostnames and have them handled by the # same Apache server process. # # Configuration and logfile names: If the filenames you specify for many # of the server's control files begin with "/" (or "drive:/" for Win32), the # server will use that explicit path. If the filenames do *not* begin # with "/", the value of ServerRoot is prepended -- so "logs/foo.log" # with ServerRoot set to "/etc/httpd" will be interpreted by the # server as "/etc/httpd/logs/foo.log". # ### Section 1: Global Environment # # The directives in this section affect the overall operation of Apache, # such as the number of concurrent requests it can handle or where it # can find its configuration files. # # # Don't give away too much information about all the subcomponents # we are running. Comment out this line if you don't mind remote sites # finding out what major optional modules you are running ServerTokens OS # # ServerRoot: The top of the directory tree under which the server's # configuration, error, and log files are kept. # # NOTE! If you intend to place this on an NFS (or otherwise network) # mounted filesystem then please read the LockFile documentation # (available at ); # you will save yourself a lot of trouble. # # Do NOT add a slash at the end of the directory path. # ServerRoot "/etc/httpd" # # PidFile: The file in which the server should record its process # identification number when it starts. Note the PIDFILE variable in # /etc/sysconfig/httpd must be set appropriately if this location is # changed. # PidFile run/httpd.pid # # Timeout: The number of seconds before receives and sends time out. # Timeout 60 # # KeepAlive: Whether or not to allow persistent connections (more than # one request per connection). Set to "Off" to deactivate. # KeepAlive Off # # MaxKeepAliveRequests: The maximum number of requests to allow # during a persistent connection. Set to 0 to allow an unlimited amount. # We recommend you leave this number high, for maximum performance. # MaxKeepAliveRequests 100 # # KeepAliveTimeout: Number of seconds to wait for the next request from the # same client on the same connection. # KeepAliveTimeout 15 ## ## Server-Pool Size Regulation (MPM specific) ## # prefork MPM # StartServers: number of server processes to start # MinSpareServers: minimum number of server processes which are kept spare # MaxSpareServers: maximum number of server processes which are kept spare # ServerLimit: maximum value for MaxClients for the lifetime of the server # MaxClients: maximum number of server processes allowed to start # MaxRequestsPerChild: maximum number of requests a server process serves StartServers 8 MinSpareServers 5 MaxSpareServers 20 ServerLimit 256 MaxClients 256 MaxRequestsPerChild 4000 # worker MPM # StartServers: initial number of server processes to start # MaxClients: maximum number of simultaneous client connections # MinSpareThreads: minimum number of worker threads which are kept spare # MaxSpareThreads: maximum number of worker threads which are kept spare # ThreadsPerChild: constant number of worker threads in each server process # MaxRequestsPerChild: maximum number of requests a server process serves StartServers 4 MaxClients 300 MinSpareThreads 25 MaxSpareThreads 75 ThreadsPerChild 25 MaxRequestsPerChild 0 # # Listen: Allows you to bind Apache to specific IP addresses and/or # ports, in addition to the default. See also the # directive. # # Change this to Listen on specific IP addresses as shown below to # prevent Apache from glomming onto all bound IP addresses (0.0.0.0) # #Listen 12.34.56.78:80 Listen 80 # # Dynamic Shared Object (DSO) Support # # To be able to use the functionality of a module which was built as a DSO you # have to place corresponding `LoadModule' lines at this location so the # directives contained in it are actually available _before_ they are used. # Statically compiled modules (those listed by `httpd -l') do not need # to be loaded here. # # Example: # LoadModule foo_module modules/mod_foo.so # LoadModule auth_basic_module modules/mod_auth_basic.so LoadModule auth_digest_module modules/mod_auth_digest.so LoadModule authn_file_module modules/mod_authn_file.so LoadModule authn_alias_module modules/mod_authn_alias.so LoadModule authn_anon_module modules/mod_authn_anon.so LoadModule authn_dbm_module modules/mod_authn_dbm.so LoadModule authn_default_module modules/mod_authn_default.so LoadModule authz_host_module modules/mod_authz_host.so LoadModule authz_user_module modules/mod_authz_user.so LoadModule authz_owner_module modules/mod_authz_owner.so LoadModule authz_groupfile_module modules/mod_authz_groupfile.so LoadModule authz_dbm_module modules/mod_authz_dbm.so LoadModule authz_default_module modules/mod_authz_default.so LoadModule ldap_module modules/mod_ldap.so LoadModule authnz_ldap_module modules/mod_authnz_ldap.so LoadModule include_module modules/mod_include.so LoadModule log_config_module modules/mod_log_config.so LoadModule logio_module modules/mod_logio.so LoadModule env_module modules/mod_env.so LoadModule ext_filter_module modules/mod_ext_filter.so LoadModule mime_magic_module modules/mod_mime_magic.so LoadModule expires_module modules/mod_expires.so LoadModule deflate_module modules/mod_deflate.so LoadModule headers_module modules/mod_headers.so LoadModule usertrack_module modules/mod_usertrack.so LoadModule setenvif_module modules/mod_setenvif.so LoadModule mime_module modules/mod_mime.so LoadModule dav_module modules/mod_dav.so LoadModule status_module modules/mod_status.so LoadModule autoindex_module modules/mod_autoindex.so LoadModule info_module modules/mod_info.so LoadModule dav_fs_module modules/mod_dav_fs.so LoadModule vhost_alias_module modules/mod_vhost_alias.so LoadModule negotiation_module modules/mod_negotiation.so LoadModule dir_module modules/mod_dir.so LoadModule actions_module modules/mod_actions.so LoadModule speling_module modules/mod_speling.so LoadModule userdir_module modules/mod_userdir.so LoadModule alias_module modules/mod_alias.so LoadModule substitute_module modules/mod_substitute.so LoadModule rewrite_module modules/mod_rewrite.so LoadModule proxy_module modules/mod_proxy.so LoadModule proxy_balancer_module modules/mod_proxy_balancer.so LoadModule proxy_ftp_module modules/mod_proxy_ftp.so LoadModule proxy_http_module modules/mod_proxy_http.so LoadModule proxy_ajp_module modules/mod_proxy_ajp.so LoadModule proxy_connect_module modules/mod_proxy_connect.so LoadModule cache_module modules/mod_cache.so LoadModule suexec_module modules/mod_suexec.so LoadModule disk_cache_module modules/mod_disk_cache.so LoadModule cgi_module modules/mod_cgi.so LoadModule version_module modules/mod_version.so # # The following modules are not loaded by default: # #LoadModule asis_module modules/mod_asis.so #LoadModule authn_dbd_module modules/mod_authn_dbd.so #LoadModule cern_meta_module modules/mod_cern_meta.so #LoadModule cgid_module modules/mod_cgid.so #LoadModule dbd_module modules/mod_dbd.so #LoadModule dumpio_module modules/mod_dumpio.so #LoadModule filter_module modules/mod_filter.so #LoadModule ident_module modules/mod_ident.so #LoadModule log_forensic_module modules/mod_log_forensic.so #LoadModule unique_id_module modules/mod_unique_id.so # # # Load config files from the config directory "/etc/httpd/conf.d". # Include conf.d/*.conf # # ExtendedStatus controls whether Apache will generate "full" status # information (ExtendedStatus On) or just basic information (ExtendedStatus # Off) when the "server-status" handler is called. The default is Off. # #ExtendedStatus On # # If you wish httpd to run as a different user or group, you must run # httpd as root initially and it will switch. # # User/Group: The name (or #number) of the user/group to run httpd as. # . On SCO (ODT 3) use "User nouser" and "Group nogroup". # . On HPUX you may not be able to use shared memory as nobody, and the # suggested workaround is to create a user www and use that user. # NOTE that some kernels refuse to setgid(Group) or semctl(IPC_SET) # when the value of (unsigned)Group is above 60000; # don't use Group #-1 on these systems! # User apache Group apache ### Section 2: 'Main' server configuration # # The directives in this section set up the values used by the 'main' # server, which responds to any requests that aren't handled by a # definition. These values also provide defaults for # any containers you may define later in the file. # # All of these directives may appear inside containers, # in which case these default settings will be overridden for the # virtual host being defined. # # # ServerAdmin: Your address, where problems with the server should be # e-mailed. This address appears on some server-generated pages, such # as error documents. e.g. [email protected] # ServerAdmin root@localhost # # ServerName gives the name and port that the server uses to identify itself. # This can often be determined automatically, but we recommend you specify # it explicitly to prevent problems during startup. # # If this is not set to valid DNS name for your host, server-generated # redirections will not work. See also the UseCanonicalName directive. # # If your host doesn't have a registered DNS name, enter its IP address here. # You will have to access it by its address anyway, and this will make # redirections work in a sensible way. # #ServerName www.example.com:80 # # UseCanonicalName: Determines how Apache constructs self-referencing # URLs and the SERVER_NAME and SERVER_PORT variables. # When set "Off", Apache will use the Hostname and Port supplied # by the client. When set "On", Apache will use the value of the # ServerName directive. # UseCanonicalName Off # # DocumentRoot: The directory out of which you will serve your # documents. By default, all requests are taken from this directory, but # symbolic links and aliases may be used to point to other locations. # DocumentRoot "/var/www/html" # # Each directory to which Apache has access can be configured with respect # to which services and features are allowed and/or disabled in that # directory (and its subdirectories). # # First, we configure the "default" to be a very restrictive set of # features. # Options FollowSymLinks AllowOverride None # # Note that from this point forward you must specifically allow # particular features to be enabled - so if something's not working as # you might expect, make sure that you have specifically enabled it # below. # # # This should be changed to whatever you set DocumentRoot to. # # # Possible values for the Options directive are "None", "All", # or any combination of: # Indexes Includes FollowSymLinks SymLinksifOwnerMatch ExecCGI MultiViews # # Note that "MultiViews" must be named *explicitly* --- "Options All" # doesn't give it to you. # # The Options directive is both complicated and important. Please see # http://httpd.apache.org/docs/2.2/mod/core.html#options # for more information. # Options Indexes FollowSymLinks # # AllowOverride controls what directives may be placed in .htaccess files. # It can be "All", "None", or any combination of the keywords: # Options FileInfo AuthConfig Limit # AllowOverride None # # Controls who can get stuff from this server. # Order allow,deny Allow from all # # UserDir: The name of the directory that is appended onto a user's home # directory if a ~user request is received. # # The path to the end user account 'public_html' directory must be # accessible to the webserver userid. This usually means that ~userid # must have permissions of 711, ~userid/public_html must have permissions # of 755, and documents contained therein must be world-readable. # Otherwise, the client will only receive a "403 Forbidden" message. # # See also: http://httpd.apache.org/docs/misc/FAQ.html#forbidden # # # UserDir is disabled by default since it can confirm the presence # of a username on the system (depending on home directory # permissions). # UserDir disabled # # To enable requests to /~user/ to serve the user's public_html # directory, remove the "UserDir disabled" line above, and uncomment # the following line instead: # #UserDir public_html # # Control access to UserDir directories. The following is an example # for a site where these directories are restricted to read-only. # # # AllowOverride FileInfo AuthConfig Limit # Options MultiViews Indexes SymLinksIfOwnerMatch IncludesNoExec # # Order allow,deny # Allow from all # # # Order deny,allow # Deny from all # # # # DirectoryIndex: sets the file that Apache will serve if a directory # is requested. # # The index.html.var file (a type-map) is used to deliver content- # negotiated documents. The MultiViews Option can be used for the # same purpose, but it is much slower. # DirectoryIndex index.html index.html.var # # AccessFileName: The name of the file to look for in each directory # for additional configuration directives. See also the AllowOverride # directive. # AccessFileName .htaccess # # The following lines prevent .htaccess and .htpasswd files from being # viewed by Web clients. # Order allow,deny Deny from all Satisfy All # # TypesConfig describes where the mime.types file (or equivalent) is # to be found. # TypesConfig /etc/mime.types # # DefaultType is the default MIME type the server will use for a document # if it cannot otherwise determine one, such as from filename extensions. # If your server contains mostly text or HTML documents, "text/plain" is # a good value. If most of your content is binary, such as applications # or images, you may want to use "application/octet-stream" instead to # keep browsers from trying to display binary files as though they are # text. # DefaultType text/plain # # The mod_mime_magic module allows the server to use various hints from the # contents of the file itself to determine its type. The MIMEMagicFile # directive tells the module where the hint definitions are located. # # MIMEMagicFile /usr/share/magic.mime MIMEMagicFile conf/magic # # HostnameLookups: Log the names of clients or just their IP addresses # e.g., www.apache.org (on) or 204.62.129.132 (off). # The default is off because it'd be overall better for the net if people # had to knowingly turn this feature on, since enabling it means that # each client request will result in AT LEAST one lookup request to the # nameserver. # HostnameLookups Off # # EnableMMAP: Control whether memory-mapping is used to deliver # files (assuming that the underlying OS supports it). # The default is on; turn this off if you serve from NFS-mounted # filesystems. On some systems, turning it off (regardless of # filesystem) can improve performance; for details, please see # http://httpd.apache.org/docs/2.2/mod/core.html#enablemmap # #EnableMMAP off # # EnableSendfile: Control whether the sendfile kernel support is # used to deliver files (assuming that the OS supports it). # The default is on; turn this off if you serve from NFS-mounted # filesystems. Please see # http://httpd.apache.org/docs/2.2/mod/core.html#enablesendfile # #EnableSendfile off # # ErrorLog: The location of the error log file. # If you do not specify an ErrorLog directive within a # container, error messages relating to that virtual host will be # logged here. If you *do* define an error logfile for a # container, that host's errors will be logged there and not here. # ErrorLog logs/error_log # # LogLevel: Control the number of messages logged to the error_log. # Possible values include: debug, info, notice, warn, error, crit, # alert, emerg. # LogLevel warn # # The following directives define some format nicknames for use with # a CustomLog directive (see below). # LogFormat "%h %l %u %t \"%r\" %>s %b \"%{Referer}i\" \"%{User-Agent}i\"" combined LogFormat "%h %l %u %t \"%r\" %>s %b" common LogFormat "%{Referer}i -> %U" referer LogFormat "%{User-agent}i" agent # "combinedio" includes actual counts of actual bytes received (%I) and sent (%O); this # requires the mod_logio module to be loaded. #LogFormat "%h %l %u %t \"%r\" %>s %b \"%{Referer}i\" \"%{User-Agent}i\" %I %O" combinedio # # The location and format of the access logfile (Common Logfile Format). # If you do not define any access logfiles within a # container, they will be logged here. Contrariwise, if you *do* # define per- access logfiles, transactions will be # logged therein and *not* in this file. # #CustomLog logs/access_log common # # If you would like to have separate agent and referer logfiles, uncomment # the following directives. # #CustomLog logs/referer_log referer #CustomLog logs/agent_log agent # # For a single logfile with access, agent, and referer information # (Combined Logfile Format), use the following directive: # CustomLog logs/access_log combined # # Optionally add a line containing the server version and virtual host # name to server-generated pages (internal error documents, FTP directory # listings, mod_status and mod_info output etc., but not CGI generated # documents or custom error documents). # Set to "EMail" to also include a mailto: link to the ServerAdmin. # Set to one of: On | Off | EMail # ServerSignature On # # Aliases: Add here as many aliases as you need (with no limit). The format is # Alias fakename realname # # Note that if you include a trailing / on fakename then the server will # require it to be present in the URL. So "/icons" isn't aliased in this # example, only "/icons/". If the fakename is slash-terminated, then the # realname must also be slash terminated, and if the fakename omits the # trailing slash, the realname must also omit it. # # We include the /icons/ alias for FancyIndexed directory listings. If you # do not use FancyIndexing, you may comment this out. # Alias /icons/ "/var/www/icons/" Options Indexes MultiViews FollowSymLinks AllowOverride None Order allow,deny Allow from all # # WebDAV module configuration section. # # Location of the WebDAV lock database. DAVLockDB /var/lib/dav/lockdb # # ScriptAlias: This controls which directories contain server scripts. # ScriptAliases are essentially the same as Aliases, except that # documents in the realname directory are treated as applications and # run by the server when requested rather than as documents sent to the client. # The same rules about trailing "/" apply to ScriptAlias directives as to # Alias. # ScriptAlias /cgi-bin/ "/var/www/cgi-bin/" # # "/var/www/cgi-bin" should be changed to whatever your ScriptAliased # CGI directory exists, if you have that configured. # AllowOverride None Options None Order allow,deny Allow from all # # Redirect allows you to tell clients about documents which used to exist in # your server's namespace, but do not anymore. This allows you to tell the # clients where to look for the relocated document. # Example: # Redirect permanent /foo http://www.example.com/bar # # Directives controlling the display of server-generated directory listings. # # # IndexOptions: Controls the appearance of server-generated directory # listings. # IndexOptions FancyIndexing VersionSort NameWidth=* HTMLTable Charset=UTF-8 # # AddIcon* directives tell the server which icon to show for different # files or filename extensions. These are only displayed for # FancyIndexed directories. # AddIconByEncoding (CMP,/icons/compressed.gif) x-compress x-gzip AddIconByType (TXT,/icons/text.gif) text/* AddIconByType (IMG,/icons/image2.gif) image/* AddIconByType (SND,/icons/sound2.gif) audio/* AddIconByType (VID,/icons/movie.gif) video/* AddIcon /icons/binary.gif .bin .exe AddIcon /icons/binhex.gif .hqx AddIcon /icons/tar.gif .tar AddIcon /icons/world2.gif .wrl .wrl.gz .vrml .vrm .iv AddIcon /icons/compressed.gif .Z .z .tgz .gz .zip AddIcon /icons/a.gif .ps .ai .eps AddIcon /icons/layout.gif .html .shtml .htm .pdf AddIcon /icons/text.gif .txt AddIcon /icons/c.gif .c AddIcon /icons/p.gif .pl .py AddIcon /icons/f.gif .for AddIcon /icons/dvi.gif .dvi AddIcon /icons/uuencoded.gif .uu AddIcon /icons/script.gif .conf .sh .shar .csh .ksh .tcl AddIcon /icons/tex.gif .tex AddIcon /icons/bomb.gif core AddIcon /icons/back.gif .. AddIcon /icons/hand.right.gif README AddIcon /icons/folder.gif ^^DIRECTORY^^ AddIcon /icons/blank.gif ^^BLANKICON^^ # # DefaultIcon is which icon to show for files which do not have an icon # explicitly set. # DefaultIcon /icons/unknown.gif # # AddDescription allows you to place a short description after a file in # server-generated indexes. These are only displayed for FancyIndexed # directories. # Format: AddDescription "description" filename # #AddDescription "GZIP compressed document" .gz #AddDescription "tar archive" .tar #AddDescription "GZIP compressed tar archive" .tgz # # ReadmeName is the name of the README file the server will look for by # default, and append to directory listings. # # HeaderName is the name of a file which should be prepended to # directory indexes. ReadmeName README.html HeaderName HEADER.html # # IndexIgnore is a set of filenames which directory indexing should ignore # and not include in the listing. Shell-style wildcarding is permitted. # IndexIgnore .??* *~ *# HEADER* README* RCS CVS *,v *,t # # DefaultLanguage and AddLanguage allows you to specify the language of # a document. You can then use content negotiation to give a browser a # file in a language the user can understand. # # Specify a default language. This means that all data # going out without a specific language tag (see below) will # be marked with this one. You probably do NOT want to set # this unless you are sure it is correct for all cases. # # * It is generally better to not mark a page as # * being a certain language than marking it with the wrong # * language! # # DefaultLanguage nl # # Note 1: The suffix does not have to be the same as the language # keyword --- those with documents in Polish (whose net-standard # language code is pl) may wish to use "AddLanguage pl .po" to # avoid the ambiguity with the common suffix for perl scripts. # # Note 2: The example entries below illustrate that in some cases # the two character 'Language' abbreviation is not identical to # the two character 'Country' code for its country, # E.g. 'Danmark/dk' versus 'Danish/da'. # # Note 3: In the case of 'ltz' we violate the RFC by using a three char # specifier. There is 'work in progress' to fix this and get # the reference data for rfc1766 cleaned up. # # Catalan (ca) - Croatian (hr) - Czech (cs) - Danish (da) - Dutch (nl) # English (en) - Esperanto (eo) - Estonian (et) - French (fr) - German (de) # Greek-Modern (el) - Hebrew (he) - Italian (it) - Japanese (ja) # Korean (ko) - Luxembourgeois* (ltz) - Norwegian Nynorsk (nn) # Norwegian (no) - Polish (pl) - Portugese (pt) # Brazilian Portuguese (pt-BR) - Russian (ru) - Swedish (sv) # Simplified Chinese (zh-CN) - Spanish (es) - Traditional Chinese (zh-TW) # AddLanguage ca .ca AddLanguage cs .cz .cs AddLanguage da .dk AddLanguage de .de AddLanguage el .el AddLanguage en .en AddLanguage eo .eo AddLanguage es .es AddLanguage et .et AddLanguage fr .fr AddLanguage he .he AddLanguage hr .hr AddLanguage it .it AddLanguage ja .ja AddLanguage ko .ko AddLanguage ltz .ltz AddLanguage nl .nl AddLanguage nn .nn AddLanguage no .no AddLanguage pl .po AddLanguage pt .pt AddLanguage pt-BR .pt-br AddLanguage ru .ru AddLanguage sv .sv AddLanguage zh-CN .zh-cn AddLanguage zh-TW .zh-tw # # LanguagePriority allows you to give precedence to some languages # in case of a tie during content negotiation. # # Just list the languages in decreasing order of preference. We have # more or less alphabetized them here. You probably want to change this. # LanguagePriority en ca cs da de el eo es et fr he hr it ja ko ltz nl nn no pl pt pt-BR ru sv zh-CN zh-TW # # ForceLanguagePriority allows you to serve a result page rather than # MULTIPLE CHOICES (Prefer) [in case of a tie] or NOT ACCEPTABLE (Fallback) # [in case no accepted languages matched the available variants] # ForceLanguagePriority Prefer Fallback # # Specify a default charset for all content served; this enables # interpretation of all content as UTF-8 by default. To use the # default browser choice (ISO-8859-1), or to allow the META tags # in HTML content to override this choice, comment out this # directive: # AddDefaultCharset UTF-8 # # AddType allows you to add to or override the MIME configuration # file mime.types for specific file types. # #AddType application/x-tar .tgz # # AddEncoding allows you to have certain browsers uncompress # information on the fly. Note: Not all browsers support this. # Despite the name similarity, the following Add* directives have nothing # to do with the FancyIndexing customization directives above. # #AddEncoding x-compress .Z #AddEncoding x-gzip .gz .tgz # If the AddEncoding directives above are commented-out, then you # probably should define those extensions to indicate media types: # AddType application/x-compress .Z AddType application/x-gzip .gz .tgz # # MIME-types for downloading Certificates and CRLs # AddType application/x-x509-ca-cert .crt AddType application/x-pkcs7-crl .crl # # AddHandler allows you to map certain file extensions to "handlers": # actions unrelated to filetype. These can be either built into the server # or added with the Action directive (see below) # # To use CGI scripts outside of ScriptAliased directories: # (You will also need to add "ExecCGI" to the "Options" directive.) # #AddHandler cgi-script .cgi # # For files that include their own HTTP headers: # #AddHandler send-as-is asis # # For type maps (negotiated resources): # (This is enabled by default to allow the Apache "It Worked" page # to be distributed in multiple languages.) # AddHandler type-map var # # Filters allow you to process content before it is sent to the client. # # To parse .shtml files for server-side includes (SSI): # (You will also need to add "Includes" to the "Options" directive.) # AddType text/html .shtml AddOutputFilter INCLUDES .shtml # # Action lets you define media types that will execute a script whenever # a matching file is called. This eliminates the need for repeated URL # pathnames for oft-used CGI file processors. # Format: Action media/type /cgi-script/location # Format: Action handler-name /cgi-script/location # # # Customizable error responses come in three flavors: # 1) plain text 2) local redirects 3) external redirects # # Some examples: #ErrorDocument 500 "The server made a boo boo." #ErrorDocument 404 /missing.html #ErrorDocument 404 "/cgi-bin/missing_handler.pl" #ErrorDocument 402 http://www.example.com/subscription_info.html # # # Putting this all together, we can internationalize error responses. # # We use Alias to redirect any /error/HTTP_.html.var response to # our collection of by-error message multi-language collections. We use # includes to substitute the appropriate text. # # You can modify the messages' appearance without changing any of the # default HTTP_.html.var files by adding the line: # # Alias /error/include/ "/your/include/path/" # # which allows you to create your own set of files by starting with the # /var/www/error/include/ files and # copying them to /your/include/path/, even on a per-VirtualHost basis. # Alias /error/ "/var/www/error/" AllowOverride None Options IncludesNoExec AddOutputFilter Includes html AddHandler type-map var Order allow,deny Allow from all LanguagePriority en es de fr ForceLanguagePriority Prefer Fallback # ErrorDocument 400 /error/HTTP_BAD_REQUEST.html.var # ErrorDocument 401 /error/HTTP_UNAUTHORIZED.html.var # ErrorDocument 403 /error/HTTP_FORBIDDEN.html.var # ErrorDocument 404 /error/HTTP_NOT_FOUND.html.var # ErrorDocument 405 /error/HTTP_METHOD_NOT_ALLOWED.html.var # ErrorDocument 408 /error/HTTP_REQUEST_TIME_OUT.html.var # ErrorDocument 410 /error/HTTP_GONE.html.var # ErrorDocument 411 /error/HTTP_LENGTH_REQUIRED.html.var # ErrorDocument 412 /error/HTTP_PRECONDITION_FAILED.html.var # ErrorDocument 413 /error/HTTP_REQUEST_ENTITY_TOO_LARGE.html.var # ErrorDocument 414 /error/HTTP_REQUEST_URI_TOO_LARGE.html.var # ErrorDocument 415 /error/HTTP_UNSUPPORTED_MEDIA_TYPE.html.var # ErrorDocument 500 /error/HTTP_INTERNAL_SERVER_ERROR.html.var # ErrorDocument 501 /error/HTTP_NOT_IMPLEMENTED.html.var # ErrorDocument 502 /error/HTTP_BAD_GATEWAY.html.var # ErrorDocument 503 /error/HTTP_SERVICE_UNAVAILABLE.html.var # ErrorDocument 506 /error/HTTP_VARIANT_ALSO_VARIES.html.var # # The following directives modify normal HTTP response behavior to # handle known problems with browser implementations. # BrowserMatch "Mozilla/2" nokeepalive BrowserMatch "MSIE 4\.0b2;" nokeepalive downgrade-1.0 force-response-1.0 BrowserMatch "RealPlayer 4\.0" force-response-1.0 BrowserMatch "Java/1\.0" force-response-1.0 BrowserMatch "JDK/1\.0" force-response-1.0 # # The following directive disables redirects on non-GET requests for # a directory that does not include the trailing slash. This fixes a # problem with Microsoft WebFolders which does not appropriately handle # redirects for folders with DAV methods. # Same deal with Apple's DAV filesystem and Gnome VFS support for DAV. # BrowserMatch "Microsoft Data Access Internet Publishing Provider" redirect-carefully BrowserMatch "MS FrontPage" redirect-carefully BrowserMatch "^WebDrive" redirect-carefully BrowserMatch "^WebDAVFS/1.[0123]" redirect-carefully BrowserMatch "^gnome-vfs/1.0" redirect-carefully BrowserMatch "^XML Spy" redirect-carefully BrowserMatch "^Dreamweaver-WebDAV-SCM1" redirect-carefully # # Allow server status reports generated by mod_status, # with the URL of http://servername/server-status # Change the ".example.com" to match your domain to enable. # # # SetHandler server-status # Order deny,allow # Deny from all # Allow from .example.com # # # Allow remote server configuration reports, with the URL of # http://servername/server-info (requires that mod_info.c be loaded). # Change the ".example.com" to match your domain to enable. # # # SetHandler server-info # Order deny,allow # Deny from all # Allow from .example.com # # # Proxy Server directives. Uncomment the following lines to # enable the proxy server: # # #ProxyRequests On # # # Order deny,allow # Deny from all # Allow from .example.com # # # Enable/disable the handling of HTTP/1.1 "Via:" headers. # ("Full" adds the server version; "Block" removes all outgoing Via: headers) # Set to one of: Off | On | Full | Block # #ProxyVia On # # To enable a cache of proxied content, uncomment the following lines. # See http://httpd.apache.org/docs/2.2/mod/mod_cache.html for more details. # # # CacheEnable disk / # CacheRoot "/var/cache/mod_proxy" # # # # End of proxy directives. ### Section 3: Virtual Hosts # # VirtualHost: If you want to maintain multiple domains/hostnames on your # machine you can setup VirtualHost containers for them. Most configurations # use only name-based virtual hosts so the server doesn't need to worry about # IP addresses. This is indicated by the asterisks in the directives below. # # Please see the documentation at # # for further details before you try to setup virtual hosts. # # You may use the command line option '-S' to verify your virtual host # configuration. # # Use name-based virtual hosting. # #NameVirtualHost *:80 # # NOTE: NameVirtualHost cannot be used without a port specifier # (e.g. :80) if mod_ssl is being used, due to the nature of the # SSL protocol. # # # VirtualHost example: # Almost any Apache directive may go into a VirtualHost container. # The first VirtualHost section is used for requests without a known # server name. # # # ServerAdmin [email protected] # DocumentRoot /www/docs/dummy-host.example.com # ServerName dummy-host.example.com # ErrorLog logs/dummy-host.example.com-error_log # CustomLog logs/dummy-host.example.com-access_log common # localhost 127.0.0.1 reply Reply JR Do you have some reason why you have localhost 127.0.0.1 alone in last line? If not then remove it or comment it out and try start/restart apache again. reply Reply Venkata Krishnan Thank you JR. Its working perectly well now after removing localhost 127.0.0.1 from the file. Do I need to be superuser to add/edit files in /var/www/html/ directory to use with Apache. Kinldy let me know the procedure to access the contents of the folder as a user. reply Reply Venkata Krishnan Thank you very much JR. All set to use Apache now. Your guidance was understandable, easy and effective. My thanks again for the prompt help too. reply Reply JR You are very welcome! Nice to hear that everything is working! And thank you for compliment! :) reply Reply JR Hi Boubakr, yum groupinstall -y web-server Doesn’t work on RHEL 6/5, CentOS 6/5 and earlier Fedora versions. And this installs only Apache not PHP with modules. reply Reply UZ [root@phi umar]# /etc/init.d/httpd restart Restarting httpd (via systemctl): Job failed. See system logs and ‘systemctl status’ for details. [FAILED] Help Please! reply Reply UZ Hello JR, [root@phi umar]# service httpd restart Restarting httpd (via systemctl): Job failed. See system logs and ‘systemctl status’ for details. [FAILED] reply Reply UZ [root@phi umar]# service httpd restart Restarting httpd (via systemctl): Job failed. See system logs and ‘systemctl status’ for details. [FAILED] reply Reply JR Hi again UZ, Could you post output of following commands: systemctl status httpd.service cat /var/log/messages | grep httpd httpd -t tail -n 150 /var/log/httpd/error_log reply Reply ReynierPM Hi, I’ve use the command as follow: yum --enablerepo=remi install httpd php php-common so PHP 5.3.17 was installed instead of 5.4.7. It’s possible to upgrade from 5.3.17 to 5.4.7 without mess the system? How? Thanks reply Reply JR Hi ReynierPM, Following should work: yum --enablerepo=remi,remi-test update httpd php php-common service httpd restart You should also make sure that you update also all PHP modules and all your PHP scripts should work with PHP 5.4. Best and safest method is backup your system before update and test this change on test server or some virtual machine. reply Reply Danny If it is not working as you expect, just do yum downgrade php* and all will be set to working stock versions. reply Reply JR Hi tarmo, You have conflict between your current MySQL 5.0.95 installation and remi repo php-mysql (mysql-libs) 5.5.28 packages. reply Reply Faisal Hi, When i try to update the php, this error show:- Error: Package: php-gd-5.4.9-1.el6.remi.x86_64 (remi-test) Requires: libt1.so.5()(64bit) reply Reply JR Hi Faisal, Could you post output of following commands: yum list installed t1lib ls -la /usr/lib64/libt1.* uname -a lsb_release -a reply Reply joe i get this error : Error unpacking rpm package php-common-5.4.10-1.el5.remi.i386 warning: /etc/php.ini created as /etc/php.ini.rpmnew error: unpacking of archive failed on file /usr/lib/php: cpio: chown reply Reply JR Hi joe, Could you post output of following commands: df free reply Reply Jerry Everything always posts when they have problems.. Everything worked fine for me on CentOS 6.3 x86_64. This saved me some time, thanks! reply Reply Jeff Thanks for this, worked great. I was having problems previously with permissions and the GD library so I wiped and used this as a clean install. My only issue now is that mod_rewrite doesn’t seem to work properly. An existing wordpress install I had gives me 404’s when I follow any of the “pretty permalinks”. When I change them back to query strings, they work just fine. I checked httpd.conf and the module is imported. I refreshed the permalink settings in WordPress and made sure the .htaccess file was writeable by the server. The rewrite rules are in the .htaccess file properly. Any ideas? Thanks! reply Reply Jeff Fixed! Had to change AllowOverride from None to All under . Thanks again for this great guide. I’ll be coming back here often. reply Reply
__label__pos
0.762516
How do you insert lines in pages? How do you insert lines in pages? Add borders and rules (lines) in Pages on MacClick the line or paragraph (or select multiple paragraphs) where you want to add the border or rule.In the Format sidebar, click the Layout button near the top.Click the pop-up menu next to Borders & Rules, then choose a line type (solid, dashed, or dotted). How can I remove a line in Word? Click the line, connector, or shape that you want to delete, and then press Delete. Tip: If you want to delete multiple lines or connectors, select the first line, press and hold Ctrl while you select the other lines, and then press Delete. How do I remove a page break line in Word? Remove a manual page breakGo to Home and select Show all nonprinting characters . This displays page breaks while you’re working on your document.Click or tap just after the paragraph mark in the page break, and then press Delete. How do I get rid of a black line in Word 2013? RemovalIn Office 2013 go to the Design tab and look to the far right for the Page Borders button.Within the borders settings, click on the leftmost tab titled Borders (not Page Border)select the top left option of None. How do I make a thick black line in Word? WordOn the Home tab, under Insert, click Shape, point to Lines and Connectors, and then click the line style that you want.In your document, hold down the mouse button and draw the line where you want. Tip: To draw a line at a pre-set angle, such as vertical or horizontal, hold down SHIFT as you draw the line. Why is there a line on the right side of my Word document? When Word inserts a vertical red line into your margins when you create a new paragraph, it means someone has enabled change tracking in the document. Change tracking is often used with shared documents so that each user’s changes can be tracked and even undone if necessary. Can you open a PDF in Word? Go to File > Open. Find the PDF, and open it (you might have to select Browse and find the PDF in a folder). Word tells you that it’s going to make a copy of the PDF and convert its contents into a format that Word can display. The original PDF won’t be changed at all. How do I remove margins in Word 2016? Page marginsSelect the Layout tab, then click the Margins command.A drop-down menu will appear. Click the predefined margin size you want.The margins of the document will be changed. Which are paper size options in Word? How to Choose Paper Size in Microsoft WordOn the Layout tab, in the Page Setup group, click Size .Select More Paper Sizes .In the Page Setup dialog box, choose a paper size and for Apply to , select Whole document .Click OK . How do you remove custom margins in Word? select marginated text> page layout>margins>custom margins>layout>Border> None I hope it works. How do I make 1 inch margins? To be sure you have the margins set to 1-inch:Click on the Page Layout tab.Click on Margins to see a drop-down menu.Make sure Normal is selected. Where is margin in Word? To change margins, click on the Margins button, found on the Page Layout tab. Word lists a number of pre-formatted options, but you can also make your own margins by selecting “Custom Margins,” found at the bottom of the Margins list. You can change each of the four margins in the dialog box that appears. How do you set margins? Click Margins, click Custom Margins, and then in the Top, Bottom, Left, and Right boxes, enter new values for the margins.To change the default margins, click Margins after you select a new margin, and then click Custom Margins. To restore the original margin settings, click Margins and then click Custom Margins.
__label__pos
0.999953
Game Theory Is Used To Create Perfect Strategies And The Ultimate Scenario For A Game Of Stone-Paper-Scissors A game is usually a structured form of playful, often recreational, play, and at times used as an academic tool F95ZONE. Games are quite different from chore, which normally is carried out for monetary reward, and from literature, which is usually an expression of literary or aesthetic elements. Games have an important social role in our society. The game has become a vital part of many people’s everyday life. The origin of the word ‘game’ is uncertain; however, most scholars agree that it came from the Greek word kerastes, which meant challenge. In the context of academic game theory, a game is defined as a set of interacting agents, where each player can affect the state and outcome of any state other than their own. A typical game would involve a team of two players who compete to accumulate a certain number of points. Every point they collect is divided equally between them. If a team member collects less points than his or her teammate, they lose that team’s point and that player’s teammate must then take that player’s remaining points from the opponent’s pool. However, we do not always play games such as these for pure recreation. We engage in this activity because it provides us with a unique and satisfying way to internalize and communicate certain values, such as sharing and cooperation. As humans, we typically have a limited understanding of the world around us. We are exposed to a very narrow view of the world, surrounded by masses of propaganda and mass-produced entertainment. A prime example of a game that provides a solid and realistic view of the world would be, for instance, the popular game called Solitaire. Solitaire is one of the most common experiences that human beings have, giving us the ability to concentrate and solve problems by applying our cognitive powers alone. In contrast to this common experience, the objective of the game Mastermind, is to solve a problem without any reliance on any other human players, using pure strategy and logic. A prime example of a game with this goal in mind is chess. In this game, the objective is to form the best possible five-man combination from the available squares by choosing pieces that fit together well and controlling the board using pieces that are on the same row, column or diagonal. One of the most important factors that separate a game of pure strategy and that of pure chance is the level of skill that the players have when they play a game of master mind. The concept of the perfect information set is an extremely important part of the development of the game Mastermind concept. The term “perfect information” refers to the set of physical facts that can be accessed at will by any participant in the game. For example, if we had a hundred people standing on a street, all of those individuals could observe all of the physical details of the hundred individuals’ bodies at the same time. However, it would still be very difficult for these hundred individuals to form any sort of pure strategies that would allow them to win the game; all of the physical facts that could be accessed would be considered “perfect information” for that particular moment. This problem is solved for the game Mastermind when each player chooses individual tiles and chooses the corresponding groupings of tiles that they will place into their game tray. After all of the players have placed all of their tiles into their game trays, they then randomly select the tiles that do not fit into their groupings, and then the groupings of tiles that do fit are drawn from the tile tray, one tile at a time, until all of the tiles are in place. The goal of the game is then to select the optimal combination of tiles to form the perfect information set. Although there is no way to guarantee that a player will actually draw the optimal combinations, through the use of dice, a random number generator, or by playing a few games with varying the size of the numbers that are being drawn, the game Mastermind is able to at least control the chances of drawing optimal combinations. Author: Leave a Reply Your email address will not be published. Required fields are marked *
__label__pos
0.91334
Sunday April 26, 2015 Homework Help: calculus Posted by Jillian on Saturday, April 9, 2011 at 12:29pm. An open box is to be constructed so that the length of the base is 4 times larger than the width of the base. If the cost to construct the base is 5 dollars per square foot and the cost to construct the four sides is 3 dollars per square foot, determine the dimensions for a box to have volume = 25 cubic feet which would minimize the cost of construction. height? dimensions of the base? Answer this Question First Name: School Subject: Answer: Related Questions Members
__label__pos
0.528611
在线英汉词典 Search 共找到1项关于identity-element意思的翻译解释和用法说明   相关词 identity的意思 element的意思 相关短语 identity crisis identity element identity matrix identity sign rare-earth element trace element transition element 单词意思查询Top5 outlet的意思 logo的意思 menu的意思 combo的意思 premium的意思 单词翻译查询Top5 profile的翻译 account的翻译 compliance的翻译 service的翻译 portfolio的翻译 单词解释查询Top5 audio的解释 review的解释 bean的解释 award的解释 eagle的解释 单词用法查询Top5 let的用法 compare的用法 would的用法 return的用法 see的用法 缩写词Top10 CEO NBA BBC DIY SW CV WTO MS DJ FLTA 词组、短语、俚语及习惯用语 • element antenna 振子天线。 • element breakdown [工管]基本动作分解。 • element cell [电]单元室。 • element cylinder 柱母线。 • element error 元件误差(同批元件产品的互差)。 • element mesh 单元网。 • element number (天线)振子数。 • element of best approximation 最佳逼近元素。 • element of coal seam occurrence 煤层要素。 • element of composite symmetry 复合对称元。 identity element identity-element的词性: n.Mathematics (名词)【数学】 1. The element of a set of numbers that when combined with another number in an operation leaves that number unchanged. For example, 0 is the identity element under addition for the real numbers, since if a is any real number, a + 0 = 0 + a = a. Similarly, 1 is the identity element under multiplication for the real numbers, since a × 1 = 1 × a = a. Also called unity 单位元素,幺元:与其它数字进行运算后而此数字不被改变的数字元素。例如,0在加法运算中是任何实数的单位元素,因为假设a 是任意实数 a 加上0等于0加上 a 还是等于 a。 同样,1在乘法运算中是任意实数的单位元素,因为 a ×1=1× aa 也作 unity
__label__pos
0.918559
Hammond_Robotics_ 1. I think it's because You've downloaded it. Downloading mixes also cost lot of storage cause it changes songs daily so you end up downloading tracks you don't want automatically. 2. No, no. It does that for me too even without downloading it. I think it's a problem on their end. 3. That's why the supermix isn't updating. 4. But it should be an automated process made by an AI, well unless they need to manually hit a button every week idk 5. you can link your watch under a samsung account, check the advanced features of your settings 6. For the left-handed thing it's okay, you can wear it on your right arm, but you need to tell your Watch you're wearing it on your right arm. Go to Settings -> General -> Orientation. 7. Not sure if that's still the case, but maybe you can create two playlists for that? 8. Probably just stick with spotify then, I like to just have the entire playlist play at random and occasionally hear an old song. Would suck to randomly pick between two. 9. Yeah I can totally understand that. I've never created my own playlist tho because I find YouTube Music's recommendations and automated playlists to be very good and always on point in recommending me what I want to listen to, but I understand why you want your playlist, and why you'd want to stick with it. 10. No lmao what? You can’t disable the “feature” that google just harvests your data. 11. Don't worry. All the apps you've installed on your phone are also "stealing" your data. What do you think Meta is doing for example? Don't believe you're safe because you have an iPhone. That's what Apple wants you to believe. 12. It's so stupid that Samsung removed the default amdroid feature to block internet access for apps and I have to use a VPN like workaround for that.. 13. Amazon Music Unlimited. The audio quality for a track (SD/HD/UltraHD/etc) displays above the track title. Click the quality indicator and you’ll see the screens I posted. 14. Oh okay thanks. Will I be able to see this screen for my Buds Pro (1st gen) if I'm not subbed to Amazon Music? I'm already subbed to YouTube Music but I want to see this screen 15. That screen is a feature of the Amazon Music app. I don’t see an equivalent in the YouTube Music app. 16. Ni I was asking if I would be able to see this screen if I downloaded Amazon Music even if I'm not subbed to premium? 17. Does it have a smart timer that wakes you up when you're in light sleep stage ? 18. It used to exist on older Galaxy Watches that don't have WearOS but TizenOS 19. J'ai ReVanced depuis quelques mois maintenant, le manager fonctionne tout à fait bien et je n'ai eu aucun soucis d'installation. En tout cas, je suis content que les devs aient décidé de ne pas abandonner le projet. 20. Xbox one running apex is tough! Glad the homies upgraded him! 21. Upgraded this year from an Xbox One to Series S, and holy moly I didn't realize I was only getting 30 fps in the dropship. Getting a constant 60 on the Series is game changer really. 22. R5R servers are self-hosted. This will be hosted by EA so makes sense they'd put limits on it since hosting them still costs money. 23. halo has been self-hosting? I think you are referring to Peer to peer, and no halo is not peer to peer anymore. I believe they canned that back near halo 3/halo 4 24. When you start a custom game, you can choose to host it on the Xbox Live servers, or locally. So actually it's Microsoft who's hosting the servers but it's working fine anyways. 25. They are certified IP67 so water resistant up to a certain level, and I go run with them even on rainy days and they are fine. 26. Not what I’m asking. I’m saying if I were to remove that account from the app, would they still take up memory 27. Youtube Music isn't bad IMO, although it isn't the best service for that at least you can use it for free 28. It works for PS4, I was soooo disappointed when I got my Series S. Seems like the most basic feature. 29. There's no Quick Resume or such feature on PS4, not even PS5. What feature are you talking about on PlayStation? 30. There is nothing to fix. If you leave any multiplayer game running, it will kick you. 31. As I said, they could let us have the option to refresh the servers and try to connect to a new one or let us go back to the main menu, at least something. 32. I don't even know what this tastes like it doesn't exist in my country. Leave a Reply Your email address will not be published. Required fields are marked * Author: admin
__label__pos
0.905595
id summary reporter owner description type status priority component severity resolution keywords cc release 5456 Wiki formatting is not properly displayed in referenced ticket rjollos rjollos "I added the following comment to one of my tickets: --- 1. Complete the following tickets. 1. Ticket #39 1. Ticket #54 1. Ticket #55 1. Ticket #56 1. Merge all development work from version 0.1 into the trunk and destroy the development branches. --- The actual wiki markup is: {{{ 1. Complete the following tickets. 1. Ticket #39 1. Ticket #54 1. Ticket #55 1. Ticket #56 1. Merge all development work from version 0.1 into the trunk and destroy the development branches. }}} When I look at ticket `#55` I see exactly: {{{ This ticket has been referenced in ticket #66: ... Complete the following tickets. 1. Ticket #39 2. Ticket #54 3. Ticket #55 4. Ticket #56 1. Merge all development work from version 0.1 into the trun... }}} The issues here are: 1. The indentation of the comment in `ticket #55` not correct. 1. A linebreak has been added, which probably results in issue `#3`. 1. The list numbering is not correct. " defect new normal TracBacksPlugin normal 0.11
__label__pos
0.931966
upgrade to 2.2 required libc5 upgrade? upgrade to 2.2 required libc5 upgrade? Post by Andrew Robertso » Thu, 08 Apr 1999 04:00:00 <snip> Quote:> Except....However, the 2.2 Changes file indicates that part of the minimal > requirements is to have libc5 5.4.46 installed. > I only have libc5 5.3.12 that came default with rh 5.2 with 2.0.36-7 kernel, > and the libc5 upgrade is not indicated on the redhat 2.2 upgrade HOWTO. Redhat 5.2 is based on Glibc2 (libc 6), and the libc5 libraries included are only there for backwards compatibility for pre-compiled binaries. This means that you have more than the minimum requirements for upgrading. Andy       1. Kernel 2.2, to upgrade or not to upgrade??? I've installed RH 5.1 and 5.2 half a dozen times and slackware a couple.  I currently run a 5.1 machine as a firewall running IP Masq. and NAT.  Samba's installed but not configured (although I intend to). I use two NE2000 compatible adapters. Should I consider installing the new 2.2 kernel? Should I consider installing the new(er) 2.0 Samba? I know that I probably want XFREE86 but are there any other pieces that I need? I'd like to use an older MCA 486/33 (which 2.2 supports natively) any comments? 2. 2.5.69 still can't write DVD... 3. How much free space required in / for Solaris 2.2 upgrade? 4. resuming ftp server. 5. What are required upgrades for 2.1/2.2? 6. PATCH: clean up es968, fix build 7. Upgraded from 2.2-GAMMA to 2.2-RELEASE: problems... 8. Incomplete lines received by a socket 9. Problem upgrading to RedHat kernal 2.2-17 from 2.2-16 10. Upgrading kernel 2.2.x to 2.4.x and GLibc 2.1.3 to 2.2.x 11. FBSD 2.2.1R -> 2.2.2R cdrom upgrade problem 12. rvplayer doesn't work after kernel 2.2 upgrade 13. Is it still possible to boot from 2.0.x once upgraded to 2.2.x?
__label__pos
0.999284
Search Ask a question 0 Express “ln va “ as a product Express “ln √a “ as a product 2 Answers by Expert Tutors Tutors, sign in to answer this question. Imtiazur S. | Tutor with master's degree and experience teaching CalculusTutor with master's degree and experienc... 4.9 4.9 (302 lesson ratings) (302) 0 Logarithm of a number raised to a power can be expressed as the power * logarithm of the number   Expressed Algebraically, log(xa) = a * log(x)   ln is natural logarithm which is a logarithm to base e (natural number)   The given expression ln(√a) can be expressed as ln(a0.5)   This can be expressed as a product as 0.5 * ln(a)  Luis A. | Physics / Advance Math / Spanish TutorPhysics / Advance Math / Spanish Tutor 5.0 5.0 (3 lesson ratings) (3) 0 Express “ln va “ as a product Solution: ================================================================== from x^a/b = r(b) [x^a], where r = root, and r(b) = root of index b; a = exponent ================================================================== from our example: va = r(2) [a] = a^1/2 and using logarims law: Ln [a^b/c] = (b/c) * ln [a], then ln va = ln ( a^1/2) = (1/2)*ln [a] = ln[a] / 2  
__label__pos
0.979156
Handy SOQL/SOSL Queries for Knowledge  Happy Monday!! In my last post, I have explained the Data Model around Salesforce Lightning Knowledge. If you haven’t gone through that post, I highly recommend you going through the post “Understanding Salesforce Lightning Knowledge Data Model” as it is going to help you to understand the queries in this post.Let’s first create the article – “Hello World” and below are the different versions with different publishing status. Below article properties which I will be using the query -Article Number: 000001000Knowledge Article Id: kA03t00000063w2CAAFetching Latest Knowledge Article VersionSELECT Id,PublishStatus,Title FROM Knowledge__kav WHERE KnowledgeArticleId = ‘kA03t00000063w2CAA’ AND IsLatestVersion = TrueFetching Published Knowledge Article VersionSELECT Id,PublishStatus,Title FROM Knowledge__kav WHERE KnowledgeArticleId = ‘kA03t00000063w2CAA’ AND PublishStatus = ‘Online’Fetching Draft Knowledge Article VersionSELECT Id,PublishStatus,Title FROM Knowledge__kav WHERE KnowledgeArticleId = ‘kA03t00000063w2CAA’ AND PublishStatus = ‘Draft’Fetching Archived Knowledge Article VersionsSELECT Id,PublishStatus,Title FROM Knowledge__kav WHERE KnowledgeArticleId = ‘kA03t00000063w2CAA’ AND PublishStatus = ‘archived’ AND IsLatestVersion = FalseThis one is tricky, as to query Archived article versions, KnowledgeArticleId and IsLatestVersion = False and PublishStatus = archived should be mentioned.Note -Using bind variables with Apex SOQL Statements with KnowledgeArticleVersion is not allowed. So you need to use dynamic SOQL like below -final String ONLINE_ARTICLE = ‘Online’;final String myQuery = ‘SELECT Id FROM Knowledge__kav WHERE PublishStatus = :ONLINE_ARTICLE’;List<Knowledge__kav> allArtciles = Database.query(myQuery);Working with DATA CATEGORYWITH DATA CATEGORY is an optional clause in SOQL and it helps to identify articles linked with one or more... Read More
__label__pos
0.942997
0 Tengo en un programa dos métodos en clases separadas de la main, en uno valido un nombre de usuario introducido en el main, y en el otro una contraseña. Mi problema es que cuando introduzco estas en un mapa, la contraseña aparece siempre como "null", así que sospecho que viene a ser un error de utilización de expresiones regulares. Main: public class MenuTreeMap { public static String menu(){ System.out.println("Elija lo que quiera hacer a continuación"); System.out.println("----------------------------------------"); System.out.println("1.-Alta de usuario."); System.out.println("----------------------------------------"); return null; } static String mUsu; static String mPass; public static void main(String[] args) { boolean flag = false; boolean mFlag = false; menu(); while (mFlag==false){ TreeMap <String, String> t = new TreeMap<String, String>(); Scanner input = new Scanner(System.in); Scanner menu = new Scanner(System.in); String mENU=menu.nextLine(); switch(mENU) { case "1": //Validar e introducir Usuario //-------------------------------------- System.out.println("Introduzca un nombre de usuario, de 8 a 20 caracteres"); while(!flag){ mUsu=input.nextLine(); System.out.println(""); ValidarUsuario valUsu= new ValidarUsuario(); if( valUsu.valUsu()==true ){ System.out.println("Usuario valido, nombre de usuario " + mUsu); flag=true; }else{ System.out.println("El usuario no era válido"); } } //-------------------------------------- //Validar e introducir contraseña System.out.println("Introduzca una contraseña, de 8 a 20. Debe incluir un símbolo especial."); while(!flag){ mUsu=input.nextLine(); System.out.println(""); ValidarContrasena valPass= new ValidarContrasena(); if( valPass.valPass()==true ){ System.out.println("Contraseña válida"); flag=true; }else{ System.out.println("El Contraseña no válida"); } } //--------------------------------------- t.put("mUsu", "mPass"); break; } } } Clase ValidarUsuario: import java.util.regex.Matcher; import java.util.regex.Pattern; /** * * @author julia */ public class ValidarUsuario { //Validar un usuario public static boolean valUsu(){ Scanner input = new Scanner(System.in); Pattern p=Pattern.compile("[\\w]+"); boolean comp=true; System.out.println("Validando... ... ... ..."); System.out.println(""); Matcher m=p.matcher(Tarea051GestionarCredenciales.mUsu); try{ if( m.matches() && Tarea051GestionarCredenciales.mUsu.length()>=8 && Tarea051GestionarCredenciales.mUsu.length()<=20){ comp=true; }else{ comp=false; } }catch(Exception e){ System.out.println("Introduzca un usuario válido"); } return comp; } } Clase ValidarContraseña import java.util.regex.Matcher; import java.util.regex.Pattern; /** * * @author julia */ public class ValidarContrasena { public static boolean valPass(){ Scanner input = new Scanner(System.in); Pattern p=Pattern.compile("[\\w!@#$]{8,20}"); boolean comp=true; System.out.println("Validando... ... ... ..."); System.out.println(""); Matcher m=p.matcher(Tarea051GestionarCredenciales.mPass); try{ if( m.matches() && Tarea051GestionarCredenciales.mPass.length()>=8 && Tarea051GestionarCredenciales.mPass.length()<=20){ comp=true; }else{ comp=false; } }catch(Exception e){ System.out.println("Introduzca un usuario válido"); } return comp; } } Con esa expresión regular en la contraseña, lo que quiero es que la contraseña sea de entre 8 y 20 caracteres, y tenga que incluir un caracter especial. Como podéis observar, tanto ValidarUsuario como ValidarContrasena están construidas de la misma manera. Entre las opciones que he intentado, ha sido decirle en la expresión regular que longitud debe tener. También añadir que los mensajes de la clase principal para comprobar que se ha introducido una contraseña adecuada no se imprimen, imagino que porque la contraseña no es adecuada, pero tampoco salta el mensaje de que no lo sea. Gracias por vuestro tiempo. 3 • Bienvenido a SOe. No entiendo dónde tienes el problema, hablas de que hay un null pero tus métodos devuelve boolean. Te aconsejo que reduzcas el ámbito de la pregunta; primero que detectes realmente cuál es el problema (¿expresión regular?¿introduces los datos de forma incorrecta?) y luego hagas un mini-programa que solo compruebe eso (por ejemplo, que compruebe si la expresión regular funciona usando como prueba una cadena que le pasas por código. – SJuan76 el 16 may. 20 a las 18:46 • Y, en general, es buena práctica separar el código en métodos o clases con funciones específicas; en este caso solo el validar la contraseña podría ser su propia clase (o método) y así lo puedes probar de forma independente al código de leer e imprimir datos. – SJuan76 el 16 may. 20 a las 18:48 • Cuidado con las variables estáticas. Las variables estáticas no pertenecen a los objetos creados por una clase, sino que son comunes a la clase. Si creas una variable estática "nombre" en una clase persona, cuando tengas 3 objetos persona, los 3 tendrán el mismo nombre. Y si cambias el nombre a 1, lo cambias a los 3. (y para cambiarlo no lo haces por objeto.propiedad sino por clase.propiedad.) Así que lo dicho, cuidado con las variables estáticas. – Jesús el 16 may. 20 a las 19:55 0 El problema es de control de lfujo... y si... hay detalles en la expresión regular pero eso no impide correrlo. Hay un principio de diseño agrupado en SOLID llamado Single Responsability, que quiere decir que no intentes hacer encargarte de hacer varias cosas a la vez, sino que te enfoques, haciéndote responsable solo de una de ellas. En este caso el problema es que tu variable flag la estás usando para controlar tanto la validación de tu usuario como de tu password, y pasando la validación de tu usuario se queda en verdadero, así que nunca entra a la validadción de tu password, y como tu bandera para el menu nunca se apaga se queda esperando infinitamente. Una corrección rápida sería: Menu tree map package com.stackoverflow.es.question356293; import java.util.Scanner; import java.util.TreeMap; public class MenuTreeMap { public static void menu() { System.out.println("Elija lo que quiera hacer a continuación"); System.out.println("----------------------------------------"); System.out.println("1.-Alta de usuario."); System.out.println("----------------------------------------"); } static String mUsu; static String mPass; public static void main(String[] args) { boolean userFlag = false; boolean passwordFlag = false; boolean mFlag = false; menu(); while (!mFlag) { TreeMap<String, String> t = new TreeMap<String, String>(); Scanner input = new Scanner(System.in); Scanner menu = new Scanner(System.in); String mENU = menu.nextLine(); switch (mENU) { case "1": //Validar e introducir Usuario //-------------------------------------- System.out.println("Introduzca un nombre de usuario, de 8 a 20 caracteres"); while (!userFlag) { mUsu = input.nextLine(); System.out.println(""); if (ValidarUsuario.valUsu()) { System.out.println("Usuario valido, nombre de usuario " + mUsu); userFlag = true; } else { System.out.println("El usuario no era válido"); } } //-------------------------------------- //Validar e introducir contraseña System.out.println("Introduzca una contraseña, de 8 a 20. Debe incluir un símbolo especial."); passwordFlag = false; while (!passwordFlag) { mPass = input.nextLine(); System.out.println(""); if (ValidarContrasena.valPass()) { System.out.println("Contraseña válida"); passwordFlag = true; } else { System.out.println("El Contraseña no válida"); } } //--------------------------------------- t.put("mUsu", "mPass"); //if (passwordFlag && userFlag) { mFlag = true; // hacemos que salga //} break; } } } } ValidarUsuario package com.stackoverflow.es.question356293; import java.util.regex.Matcher; import java.util.regex.Pattern; /** * @author julia */ public class ValidarUsuario { //Validar un usuario public static boolean valUsu() { Pattern p = Pattern.compile("[\\w]+"); boolean comp = true; System.out.println("Validando... ... ... ..."); System.out.println(""); Matcher m = p.matcher(MenuTreeMap.mUsu); try { comp = m.matches() && MenuTreeMap.mUsu.length() >= 8 && MenuTreeMap.mUsu.length() <= 20; } catch (Exception e) { System.out.println("Introduzca un usuario válido"); } return comp; } } ValidarContrasena package com.stackoverflow.es.question356293; import java.util.regex.Matcher; import java.util.regex.Pattern; /** * @author julia */ public class ValidarContrasena { public static boolean valPass() { Pattern p = Pattern.compile("[\\w!@#$]{8,20}"); boolean comp = true; System.out.println("Validando... ... ... ..."); System.out.println(""); Matcher m = p.matcher(MenuTreeMap.mPass); try { comp = m.matches() && MenuTreeMap.mPass.length() >= 8 && MenuTreeMap.mPass.length() <= 20; } catch (Exception e) { System.out.println("Introduzca una contraseña válida"); } return comp; } } Como verás para desatorarlo hice una variable para manejar aparte la validación del usuario de la validadción de la contraseña. Una mejora adicional es que no necesitas instanciar tus clases si los métodos que piensas llamar son estáticos, a este tipo de clases se les ocnoce como clases de utilidades. De hecho en tu caso es muy buen diseño, ya que sin darte cuenta hiciste de tus validadores interfaces funcionales. Aplicando el primer principio de Solid podemos modificar el código de MenuTreeMap a algo más fácil de modificar como esto: package com.stackoverflow.es.question356293; import java.util.Scanner; import java.util.TreeMap; public class MenuTreeMap { public static void menu() { System.out.println("Elija lo que quiera hacer a continuación"); System.out.println("----------------------------------------"); System.out.println("1.-Alta de usuario."); System.out.println("----------------------------------------"); } static String mUsu; static String mPass; public static void main(String[] args) { boolean mFlag = false; menu(); while (!mFlag) { TreeMap<String, String> t = new TreeMap<String, String>(); Scanner input = new Scanner(System.in); String mENU = input.nextLine(); switch (mENU) { case "1": leerUsuarioValido(input); //-------------------------------------- leerPasswordValido(input); //--------------------------------------- t.put("mUsu", "mPass"); //if (passwordFlag && userFlag) { mFlag = true; // hacemos que salga //} break; } } } /** * Validar e introducir contraseña */ private static boolean leerPasswordValido(Scanner input) { boolean passwordFlag = false; while (!passwordFlag) { System.out.println("Introduzca una contraseña, de 8 a 20. Debe incluir un símbolo especial."); mPass = input.nextLine(); System.out.println(""); if (ValidarContrasena.valPass()) { System.out.println("Contraseña válida"); passwordFlag = true; } else { System.out.println("El Contraseña no válida"); } } return true; } /** * Validar e introducir Usuario */ private static boolean leerUsuarioValido(Scanner input) { boolean userFlag = false; while (!userFlag) { System.out.println("Introduzca un nombre de usuario, de 8 a 20 caracteres"); mUsu = input.nextLine(); System.out.println(""); if (ValidarUsuario.valUsu()) { System.out.println("Usuario valido, nombre de usuario " + mUsu); userFlag = true; } else { System.out.println("El usuario no era válido"); } } return true; } } Tu Respuesta Al pulsar en “Publica tu respuesta”, muestras tu consentimiento a nuestros términos de servicio, política de privacidad y política de cookies ¿No es la respuesta que buscas? Examina otras preguntas con la etiqueta o formula tu propia pregunta.
__label__pos
0.978952
Introduction This post uses some LaTeX. You may want to read it on the original site. In my last post I showed how SymPy can benefit from Theano. In particular Theano provided a mature platform for code generation that outperformed SymPy’s attempt at the same problem. I argued that projects should stick to one specialty and depend on others for secondary concerns. Interfaces are better than add-ons. In this post I’ll show how Theano can benefit from SymPy. In particular I’ll demonstrate the practicality of SymPy’s impressive scalar simplification routines for generating efficient programs. After re-reading over this post I realize that it’s somewhat long. I’ve decided to put the results first in hopes that it’ll motivate you to keep reading. Project operation count SymPy 27 Theano 24 SymPy+Theano 17 Now, lets find out what those numbers mean. Example problem We use a larger version of our problem from last time; a radial wavefunction corresponding to n = 6 and l = 2 for Carbon (Z = 6) from sympy.physics.hydrogen import R_nl from sympy.abc import x n, l, Z = 6, 2, 6 expr = R_nl(n, l, x, Z) print latex(expr) \[\frac{1}{210} \sqrt{70} x^{2} \left(- \frac{4}{3} x^{3} + 16 x^{2} - 56 x + 56\right) e^{- x}\] We want to generate code to compute both this expression and its derivative. Both SymPy and Theano can compute and simplify derivatives. In this post we’ll measure the complexity of a computation that simultaneously computes both the above expression and its derivative. We’ll arrive at this computation through a couple of different routes that use overlapping parts of SymPy and Theano. This will supply a couple of direct comparisons. Disclaimer: I’ve chosen a larger expression here to exaggerate results. Simpler expressions yield less impressive results. Simplification We show the expression, it’s derivative, and SymPy’s simplification of that derivative. In each case we quantify the complexity of the expression by the number of algebraic operations The target expression: print latex(expr) \[\frac{1}{210} \sqrt{70} x^{2} \left(- \frac{4}{3} x^{3} + 16 x^{2} - 56 x + 56\right) e^{- x}\] print "Operations: ", count_ops(expr) Operations: 17 It’s derivative print latex(expr.diff(x)) \[\frac{1}{210} \sqrt{70} x^{2} \left(- 4 x^{2} + 32 x - 56\right) e^{- x} - \frac{1}{210} \sqrt{70} x^{2} \left(- \frac{4}{3} x^{3} + 16 x^{2} - 56 x + 56\right) e^{- x} + \frac{1}{105} \sqrt{70} x \left(- \frac{4}{3} x^{3} + 16 x^{2} - 56 x + 56\right) e^{- x}\] print "Operations: ", count_ops(expr.diff(x)) Operations: 48 The result of simplify on the derivative. Note the significant cancellation of the above expression. print latex(simplify(expr.diff(x))) \[\frac{2}{315} \sqrt{70} x \left(x^{4} - 17 x^{3} + 90 x^{2} - 168 x + 84\right) e^{- x}\] print "Operations: ", count_ops(simplify(expr.diff(x))) Operations: 18 An unevaluated derivative object. We’ll end up passing this to Theano so that it computes the derivative on its own. print latex(Derivative(expr, x)) \[\frac{\partial}{\partial x}\left(\frac{1}{210} \sqrt{70} x^{2} \left(- \frac{4}{3} x^{3} + 16 x^{2} - 56 x + 56\right) e^{- x}\right)\] Bounds on the cost of Differentiation Scalar differentiation is actually a very simple transformation. You need to know how to transform all of the elementary functions (exp, log, sin, cos, polynomials, etc...), the chain rule, and that’s it. Theorems behind automatic differentiation state that the cost of a derivative will be at most five times the cost of the original. In this case we’re guaranteed to have at most 17*5 == 85 operations in the derivative computation; this holds in our case because 48 < 85 However derivatives are often far simpler than this upper bound. We see that after simplification the operation count of the derivative is 18, only one more than the original. This is common in practice. Theano Simplification Like SymPy, Theano transforms graphs to mathematically equivalent but computationally more efficient representations. It provides standard compiler optimizations like constant folding, and common sub-expressions as well as array specific optimizations like the element-wise operation fusion. Because users regularly handle mathematical terms Theano also provides a set of optimizations to simplify some common scalar expressions. For example Theano will convert expressions like x*y/x to y. In this sense it overlaps with SymPy’s simplify functions. This post is largely a demonstration that SymPy’s scalar simplifications are far more powerful than Theano’s and that their use can result in significant improvements. This shouldn’t be surprising. Sympians are devoted to scalar simplification to a degree that far exceeds the Theano community’s devotion to this topic. Experiment We’ll compute the derivative of our radial wavefunction and then simplify the result. We’ll do this using both SymPy’s derivative and simplify routines and using Theano’s derivative and simplify routines. We’ll then compare the two results by counting the number of required operations. Here is some setup code that you can safely ignore: def fgraph_of(*exprs): """ Transform SymPy expressions into Theano Computation """ outs = map(theano_code, exprs) ins = theano.gof.graph.inputs(outs) ins, outs = theano.gof.graph.clone(ins, outs) return theano.gof.FunctionGraph(ins, outs) def theano_simplify(fgraph): """ Simplify a Theano Computation """ mode = theano.compile.get_default_mode().excluding("fusion") fgraph = fgraph.clone() mode.optimizer.optimize(fgraph) return fgraph def theano_count_ops(fgraph): """ Count the number of Scalar operations in a Theano Computation """ return len(filter(lambda n: isinstance(n.op, theano.tensor.Elemwise), fgraph.apply_nodes)) In SymPy we create both an unevaluated derivative and a fully evaluated and sympy-simplified version. We translate each to Theano, simplify within Theano, and then count the number of operations both before and after simplification. In this way we can see the value added by both SymPy’s and Theano’s optimizations. exprs = [Derivative(expr, x), # derivative computed in Theano simplify(expr.diff(x))] # derivative computed in SymPy, also sympy-simplified for expr in exprs: fgraph = fgraph_of(expr) simp_fgraph = theano_simplify(fgraph) print latex(expr) print "Operations: ", theano_count_ops(fgraph) print "Operations after Theano Simplification: ", theano_count_ops(simp_fgraph) Theano Only \[\frac{\partial}{\partial x}\left(\frac{1}{210} \sqrt{70} x^{2} \left(- \frac{4}{3} x^{3} + 16 x^{2} - 56 x + 56\right) e^{- x}\right)\] Operations: 40 Operations after Theano Simplification: 21 SymPy + Theano \[\frac{2}{315} \sqrt{70} x \left(x^{4} - 17 x^{3} + 90 x^{2} - 168 x + 84\right) e^{- x}\] Operations: 13 Operations after Theano Simplification: 10 Analysis On its own Theano produces a derivative expression that is about as complex as the unsimplified SymPy version. Theano simplification then does a surprisingly good job, roughly halving the amount of work needed (40 -> 21) to compute the result. If you dig deeper however you find that this isn’t because it was able to algebraically simplify the computation (it wasn’t) but rather because the computation contained several common sub-expressions. The Theano version looks a lot like the unsimplified SymPy version. Note the common sub-expressions like 56*x below. \[\frac{1}{210} \sqrt{70} x^{2} \left(- 4 x^{2} + 32 x - 56\right) e^{- x} - \frac{1}{210} \sqrt{70} x^{2} \left(- \frac{4}{3} x^{3} + 16 x^{2} - 56 x + 56\right) e^{- x} + \frac{1}{105} \sqrt{70} x \left(- \frac{4}{3} x^{3} + 16 x^{2} - 56 x + 56\right) e^{- x}\] The pure-SymPy simplified result is again substantially more efficient (13 operations). Interestingly Theano is still able to improve on this, again not because of additional algebraic simplification but rather due to constant folding. The two projects simplify in orthogonal ways. Simultaneous Computation When we compute both the expression and its derivative simultaneously we find substantial benefits from using the two projects together. orig_expr = R_nl(n, l, x, Z) for expr in exprs: fgraph = fgraph_of(expr, orig_expr) simp_fgraph = theano_simplify(fgraph) print latex((expr, orig_expr)) print "Operations: ", len(fgraph.apply_nodes) print "Operations after Theano Simplification: ", len(simp_fgraph.apply_nodes) \[\begin{pmatrix}\frac{\partial}{\partial x}\left(\frac{1}{210} \sqrt{70} x^{2} \left(- \frac{4}{3} x^{3} + 16 x^{2} - 56 x + 56\right) e^{- x}\right), & \frac{1}{210} \sqrt{70} x^{2} \left(- \frac{4}{3} x^{3} + 16 x^{2} - 56 x + 56\right) e^{- x}\end{pmatrix}\] Operations: 57 Operations after Theano Simplification: 24 \[\begin{pmatrix}\frac{2}{315} \sqrt{70} x \left(x^{4} - 17 x^{3} + 90 x^{2} - 168 x + 84\right) e^{- x}, & \frac{1}{210} \sqrt{70} x^{2} \left(- \frac{4}{3} x^{3} + 16 x^{2} - 56 x + 56\right) e^{- x}\end{pmatrix}\] Operations: 27 Operations after Theano Simplification: 17 The combination of SymPy’s scalar simplification and Theano’s common sub-expression optimization yields a significantly simpler computation than either project could do independently. To summarize Project operation count SymPy 27 Theano 24 SymPy+Theano 17 References blog comments powered by Disqus
__label__pos
0.997305
Checked content Emmy Noether Related subjects: Mathematicians Background to the schools Wikipedia SOS Children offer a complete download of this selection for schools for use on schools intranets. Sponsoring children helps children in the developing world to learn too. Emmy Noether Noether.jpg Born Amalie Emmy Noether (1882-03-23)23 March 1882 Erlangen, Bavaria, Germany Died 14 April 1935(1935-04-14) (aged 53) Bryn Mawr, Pennsylvania, USA Nationality German Fields Mathematics and physics Institutions University of Göttingen Bryn Mawr College Alma mater University of Erlangen Doctoral advisor Paul Gordan Doctoral students Max Deuring Hans Fitting Grete Hermann Zeng Jiongzhi Jacob Levitzki Otto Schilling Ernst Witt Known for Abstract algebra Theoretical physics Emmy Noether (German: [ˈnøːtɐ]; official name Amalie Emmy Noether, 23 March 1882 – 14 April 1935), was an influential German mathematician known for her groundbreaking contributions to abstract algebra and theoretical physics. Described by Pavel Alexandrov, Albert Einstein, Jean Dieudonné, Hermann Weyl, Norbert Wiener and others as the most important woman in the history of mathematics, she revolutionized the theories of rings, fields, and algebras. In physics, Noether's theorem explains the fundamental connection between symmetry and conservation laws. She was born to a Jewish family in the Bavarian town of Erlangen; her father was mathematician Max Noether. Emmy originally planned to teach French and English after passing the required examinations, but instead studied mathematics at the University of Erlangen, where her father lectured. After completing her dissertation in 1907 under the supervision of Paul Gordan, she worked at the Mathematical Institute of Erlangen without pay for seven years (at the time women were largely excluded from academic positions). In 1915, she was invited by David Hilbert and Felix Klein to join the mathematics department at the University of Göttingen, a world-renowned centre of mathematical research. The philosophical faculty objected, however, and she spent four years lecturing under Hilbert's name. Her habilitation was approved in 1919, allowing her to obtain the rank of Privatdozent. Noether remained a leading member of the Göttingen mathematics department until 1933; her students were sometimes called the "Noether boys". In 1924, Dutch mathematician B. L. van der Waerden joined her circle and soon became the leading expositor of Noether's ideas: her work was the foundation for the second volume of his influential 1931 textbook, Moderne Algebra. By the time of her plenary address at the 1932 International Congress of Mathematicians in Zürich, her algebraic acumen was recognized around the world. The following year, Germany's Nazi government dismissed Jews from university positions, and Noether moved to the United States to take up a position at Bryn Mawr College in Pennsylvania. In 1935 she underwent surgery for an ovarian cyst and, despite signs of a recovery, died four days later at the age of 53. Noether's mathematical work has been divided into three "epochs". In the first (1908–19), she made significant contributions to the theories of algebraic invariants and number fields. Her work on differential invariants in the calculus of variations, Noether's theorem, has been called "one of the most important mathematical theorems ever proved in guiding the development of modern physics". In the second epoch (1920–26), she began work that "changed the face of [abstract] algebra". In her classic paper Idealtheorie in Ringbereichen (Theory of Ideals in Ring Domains, 1921) Noether developed the theory of ideals in commutative rings into a powerful tool with wide-ranging applications. She made elegant use of the ascending chain condition, and objects satisfying it are named Noetherian in her honour. In the third epoch (1927–35), she published major works on noncommutative algebras and hypercomplex numbers and united the representation theory of groups with the theory of modules and ideals. In addition to her own publications, Noether was generous with her ideas and is credited with several lines of research published by other mathematicians, even in fields far removed from her main work, such as algebraic topology. Biography Noether grew up in the Bavarian city of Erlangen, depicted here in a 1916 postcard Emmy's father, Max Noether, was descended from a family of wholesale traders in Germany. He had been paralyzed by poliomyelitis at the age of fourteen. He regained mobility, but one leg remained affected. Largely self-taught, he was awarded a doctorate from the University of Heidelberg in 1868. After teaching there for seven years, he took a position in the Bavarian city of Erlangen, where he met and married Ida Amalia Kaufmann, the daughter of a prosperous merchant. Max Noether's mathematical contributions were to algebraic geometry mainly, following in the footsteps of Alfred Clebsch. His best known results are the Brill–Noether theorem and the residue, or AF+BG theorem; several other theorems are associated with him, including Max Noether's theorem. Emmy Noether was born on 23 March 1882, the first of four children. Her first name was "Amalie", after her mother and paternal grandmother, but she began using her middle name at a young age. As a girl, she was well liked. She did not stand out academically although she was known for being clever and friendly. Emmy was near-sighted and talked with a minor lisp during childhood. A family friend recounted a story years later about young Emmy quickly solving a brain teaser at a children's party, showing logical acumen at that early age. Emmy was taught to cook and clean, as were most girls of the time, and she took piano lessons. She pursued none of these activities with passion, although she loved to dance. She had three younger brothers. The eldest, Alfred, was born in 1883, was awarded a doctorate in chemistry from Erlangen in 1909, but died nine years later. Fritz Noether, born in 1884, is remembered for his academic accomplishments: after studying in Munich he made a reputation for himself in applied mathematics. The youngest, Gustav Robert, was born in 1889. Very little is known about his life; he suffered from chronic illness and died in 1928. University of Erlangen Paul Gordan supervised Noether's doctoral dissertation on invariants of biquadratic forms Emmy Noether showed early proficiency in French and English. In the spring of 1900 she took the examination for teachers of these languages and received an overall score of sehr gut (very good). Her performance qualified her to teach languages at schools reserved for girls, but she chose instead to continue her studies at the University of Erlangen. This was an unconventional decision; two years earlier, the Academic Senate of the university had declared that allowing mixed-sex education would "overthrow all academic order". One of only two women students in a university of 986, Noether was only allowed to audit classes rather than participate fully, and required the permission of individual professors whose lectures she wished to attend. Despite the obstacles, on 14 July 1903 she passed the graduation exam at a Realgymnasium in Nuremberg. During the 1903–04 winter semester, she studied at the University of Göttingen, attending lectures given by astronomer Karl Schwarzschild and mathematicians Hermann Minkowski, Otto Blumenthal, Felix Klein, and David Hilbert. Soon thereafter, restrictions on women's participation in that university were rescinded. Noether returned to Erlangen. She officially reentered the university on 24 October 1904, and declared her intention to focus solely on mathematics. Under the supervision of Paul Gordan she wrote her dissertation, Über die Bildung des Formensystems der ternären biquadratischen Form (On Complete Systems of Invariants for Ternary Biquadratic Forms, 1907). Although it had been well received, Noether later described her thesis as "crap". For the next seven years (1908–15) she taught at the University of Erlangen's Mathematical Institute without pay, occasionally substituting for her father when he was too ill to lecture. In 1910 and 1911 she published an extension of her thesis work from three variables to n variables. Noether sometimes used postcards to discuss abstract algebra with her colleague, Ernst Fischer; this card is postmarked 10 April 1915 Gordan retired in the spring of 1910, but continued to teach occasionally with his successor, Erhard Schmidt, who left shortly afterward for a position in Breslau. Gordan retired from teaching altogether in 1911 with the arrival of Schmidt's successor Ernst Fischer, and died in December 1912. According to Hermann Weyl, Fischer was an important influence on Noether, in particular by introducing her to the work of David Hilbert. From 1913 to 1916 Noether published several papers extending and applying Hilbert's methods to mathematical objects such as fields of rational functions and the invariants of finite groups. This phase marks the beginning of her engagement with abstract algebra, the field of mathematics to which she would make groundbreaking contributions. Noether and Fischer shared lively enjoyment of mathematics and would often discuss lectures long after they were over; Noether is known to have sent postcards to Fischer continuing her train of mathematical thoughts. University of Göttingen In the spring of 1915, Noether was invited to return to the University of Göttingen by David Hilbert and Felix Klein. Their effort to recruit her, however, was blocked by the philologists and historians among the philosophical faculty: women, they insisted, should not become privatdozent. One faculty member protested: "What will our soldiers think when they return to the university and find that they are required to learn at the feet of a woman?" Hilbert responded with indignation, stating, "I do not see that the sex of the candidate is an argument against her admission as privatdozent. After all, we are a university, not a bath house." In 1915 David Hilbert invited Noether to join the Göttingen mathematics department, challenging the views of some of his colleagues that a woman should not be allowed to teach at a university Noether left for Göttingen in late April; two weeks later her mother died suddenly in Erlangen. She had previously received medical care for an eye condition, but its nature and impact on her death is unknown. At about the same time Noether's father retired and her brother joined the German Army to serve in World War I. She returned to Erlangen for several weeks, mostly to care for her aging father. During her first years teaching at Göttingen she did not have an official position and was not paid; her family paid for her room and board and supported her academic work. Her lectures often were advertised under Hilbert's name, and Noether would provide "assistance". Soon after arriving at Göttingen, however, she demonstrated her capabilities by proving the theorem now known as Noether's theorem, which shows that a conservation law is associated with any differentiable symmetry of a physical system. American physicists Leon M. Lederman and Christopher T. Hill argue in their book Symmetry and the Beautiful Universe that Noether's theorem is "certainly one of the most important mathematical theorems ever proved in guiding the development of modern physics, possibly on a par with the Pythagorean theorem". The mathematics department at the University of Göttingen allowed Noether's habilitation in 1919, four years after she had begun lecturing at the school When World War I ended, the German Revolution of 1918–19 brought a significant change in social attitudes, including more rights for women. In 1919 the University of Göttingen allowed Noether to proceed with her habilitation (eligibility for tenure). Her oral examination was held in late May, and she successfully delivered her habilitation lecture in June. Three years later she received a letter from the Prussian Minister for Science, Art, and Public Education, in which he conferred on her the title of nicht beamteter ausserordentlicher Professor (an untenured professor with limited internal administrative rights and functions). This was an unpaid "extraordinary" professorship, not the higher "ordinary" professorship, which was a civil-service position. Although it recognized the importance of her work, the position still provided no salary. Noether was not paid for her lectures until she was appointed to the special position of Lehrbeauftragte für Algebra a year later. Seminal work in abstract algebra Although Noether's theorem had a profound effect upon physics, among mathematicians she is best remembered for her seminal contributions to abstract algebra. As Nathan Jacobson says in his Introduction to Noether's Collected Papers, The development of abstract algebra, which is one of the most distinctive innovations of twentieth century mathematics, is largely due to her – in published papers, in lectures, and in personal influence on her contemporaries. Noether's groundbreaking work in algebra began in 1920. In collaboration with W. Schmeidler, she then published a paper about the theory of ideals in which they defined left and right ideals in a ring. The following year she published a landmark paper called Idealtheorie in Ringbereichen, analyzing ascending chain conditions with regard to (mathematical) ideals. Noted algebraist Irving Kaplansky called this work "revolutionary"; the publication gave rise to the term " Noetherian ring", and several other mathematical objects being called Noetherian. In 1924 a young Dutch mathematician, B. L. van der Waerden, arrived at the University of Göttingen. He immediately began working with Noether, who provided invaluable methods of abstract conceptualization. van der Waerden later said that her originality was "absolute beyond comparison". In 1931 he published Moderne Algebra, a central text in the field; its second volume borrowed heavily from Noether's work. Although Emmy Noether did not seek recognition, he included as a note in the seventh edition "based in part on lectures by E. Artin and E. Noether". She sometimes allowed her colleagues and students to receive credit for her ideas, helping them develop their careers at the expense of her own. van der Waerden's visit was part of a convergence of mathematicians from all over the world to Göttingen, which became a major hub of mathematical and physical research. From 1926 to 1930 Russian topologist Pavel Alexandrov lectured at the university, and he and Noether quickly became good friends. He began referring to her as der Noether, using the masculine German article as a term of endearment to show his respect. She tried to arrange for him to obtain a position at Göttingen as a regular professor, but was only able to help him secure a scholarship from the Rockefeller Foundation. They met regularly and enjoyed discussions about the intersections of algebra and topology. In his 1935 memorial address, Alexandrov named Emmy Noether "the greatest woman mathematician of all time". Lecturing and students In Göttingen, Noether supervised more than a dozen doctoral students; her first was Grete Hermann, who defended her dissertation in February 1925. She later spoke reverently of her "dissertation-mother".. Noether also supervised Max Deuring, who distinguished himself as an undergraduate and went on to contribute significantly to the field of arithmetic geometry; Hans Fitting, remembered for Fitting's theorem and the Fitting lemma; and Zeng Jiongzhi (also rendered "Chiungtze C. Tsen" in English), who proved Tsen's theorem. She also worked closely with Wolfgang Krull, who greatly advanced commutative algebra with his Hauptidealsatz and his dimension theory for commutative rings. In addition to her mathematical insight, Noether was respected for her consideration of others. Although she sometimes acted rudely toward those who disagreed with her, she nevertheless gained a reputation for constant helpfulness and patient guidance of new students. Her loyalty to mathematical precision caused one colleague to name her "a severe critic", but she combined this demand for accuracy with a nurturing attitude. A colleague later described her this way: "Completely unegotistical and free of vanity, she never claimed anything for herself, but promoted the works of her students above all." Her frugal lifestyle at first was due to being denied pay for her work; however, even after the university began paying her a small salary in 1923, she continued to live a simple and modest life. She was paid more generously later in her life, but saved half of her salary to bequeath to her nephew, Gottfried E. Noether. Mostly unconcerned about appearance and manners, she focused on her studies to the exclusion of romance and fashion. A distinguished algebraist Olga Taussky-Todd described a luncheon, during which Noether, wholly engrossed in a discussion of mathematics, "gesticulated wildly" as she ate and "spilled her food constantly and wiped it off from her dress, completely unperturbed". Appearance-conscious students cringed as she retrieved the handkerchief from her blouse and ignored the increasing disarray of her hair during a lecture. Two female students once approached her during a break in a two-hour class to express their concern, but they were unable to break through the energetic mathematics discussion she was having with other students. According to van der Waerden's obituary of Emmy Noether, she did not follow a lesson plan for her lectures, which frustrated some students. Instead, she used her lectures as a spontaneous discussion time with her students, to think through and clarify important cutting-edge problems in mathematics. Some of her most important results were developed in these lectures, and the lecture notes of her students formed the basis for several important textbooks, such as those of van der Waerden and Deuring. Several of her colleagues attended her lectures, and she allowed some of her ideas, such as the crossed product (verschränktes Produkt in German) of associative algebras, to be published by others. Noether was recorded as having given at least five semester-long courses at Göttingen: • Winter 1924/25: Gruppentheorie und hyperkomplexe Zahlen (Group Theory and Hypercomplex Numbers) • Winter 1927/28: Hyperkomplexe Grössen und Darstellungstheorie (Hypercomplex Quantities and Representation Theory) • Summer 1928: Nichtkommutative Algebra (Noncommutative Algebra) • Summer 1929: Nichtkommutative Arithmetik (Noncommutative Arithmetic) • Winter 1929/30: Algebra der hyperkomplexen Grössen (Algebra of Hypercomplex Quantities). These courses often preceded major publications in these areas. Noether spoke quickly—reflecting the speed of her thoughts, many said—and demanded great concentration from her students. Students who disliked her style often felt alienated. Some pupils felt that she relied too much on spontaneous discussions. Her most dedicated students, however, relished the enthusiasm with which she approached mathematics, especially since her lectures often built on earlier work they had done together. She developed a close circle of colleagues and students who thought along similar lines and tended to exclude those who did not. "Outsiders" who occasionally visited Noether's lectures usually spent only 30 minutes in the room before leaving in frustration or confusion. A regular student said of one such instance: "The enemy has been defeated; he has cleared out." Noether showed a devotion to her subject and her students that extended beyond the academic day. Once, when the building was closed for a state holiday, she gathered the class on the steps outside, led them through the woods, and lectured at a local coffee house. Later, after she had been dismissed by the Third Reich, she invited students into her home to discuss their future plans and mathematical concepts. Moscow Noether taught at the Moscow State University during the winter of 1928–29 In the winter of 1928–29 Noether accepted an invitation to Moscow State University, where she continued working with P. S. Alexandrov. In addition to carrying on with her research, she taught classes in abstract algebra and algebraic geometry. She worked with the topologists, Lev Pontryagin and Nikolai Chebotaryov, who later praised her contributions to the development of Galois theory. Although politics was not central to her life, Noether took a keen interest in political matters and, according to Alexandrov, showed considerable support for the Russian Revolution (1917). She was especially happy to see Soviet advancements in the fields of science and mathematics, which she considered indicative of new opportunities made possible by the Bolshevik project. This attitude caused her problems in Germany, culminating in her eviction from a pension lodging building, after student leaders complained of living with "a Marxist-leaning Jewess". Pavel Alexandrov Noether planned to return to Moscow, an effort for which she received support from Alexandrov. After she left Germany in 1933 he tried to help her gain a chair at Moscow State University through the Soviet Education Ministry. Although this effort proved unsuccessful, they corresponded frequently during the 1930s, and in 1935 she made plans for a return to the Soviet Union. Meanwhile her brother, Fritz accepted a position at the Research Institute for Mathematics and Mechanics in Tomsk, in the Siberian Federal District of Russia, after losing his job in Germany. Recognition In 1932 Emmy Noether and Emil Artin received the Ackermann–Teubner Memorial Award for their contributions to mathematics. The prize carried a monetary reward of 500  Reichsmarks and was seen as a long-overdue official recognition of her considerable work in the field. Nevertheless, her colleagues expressed frustration at the fact that she was not elected to the Göttingen Gesellschaft der Wissenschaften (academy of sciences) and was never promoted to the position of Ordentlicher Professor (full professor). Noether visited Zürich in 1932 to deliver a plenary address at the International Congress of Mathematicians Noether's colleagues celebrated her fiftieth birthday in 1932, in typical mathematicians' style. Helmut Hasse dedicated an article to her in the Mathematische Annalen, wherein he confirmed her suspicion that some aspects of noncommutative algebra are simpler than those of commutative algebra, by proving a noncommutative reciprocity law. This pleased her immensely. He also sent her a mathematical riddle, the "mμν-riddle of syllables", which she solved immediately; the riddle has been lost. In November of the same year, Noether delivered a plenary address (großer Vortrag) on "Hyper-complex systems in their relations to commutative algebra and to number theory" at the International Congress of Mathematicians in Zürich. The congress was attended by 800 people, including Noether's colleagues Hermann Weyl, Edmund Landau, and Wolfgang Krull. There were 420 official participants and twenty-one plenary addresses presented. Apparently, Noether's prominent speaking position was a recognition of the importance of her contributions to mathematics. The 1932 congress is sometimes described as the high point of her career. Expulsion from Göttingen When Adolf Hitler became the German Reichskanzler in January 1933, Nazi activity around the country increased dramatically. At the University of Göttingen the German Student Association led the attack on the "un-German spirit" attributed to Jews and was aided by a privatdozent named Werner Weber, a former student of Emmy Noether. Antisemitic attitudes created a climate hostile to Jewish professors. One young protester reportedly demanded: "Aryan students want Aryan mathematics and not Jewish mathematics." One of the first actions of Hitler's administration was the Law for the Restoration of the Professional Civil Service which removed Jews and politically suspect government employees (including university professors) from their jobs unless they had "demonstrated their loyalty to Germany" by serving in World War I. In April 1933 Noether received a notice from the Prussian Ministry for Sciences, Art, and Public Education which read: "On the basis of paragraph 3 of the Civil Service Code of 7 April 1933, I hereby withdraw from you the right to teach at the University of Göttingen." Several of Noether's colleagues, including Max Born and Richard Courant, also had their positions revoked. Noether accepted the decision calmly, providing support for others during this difficult time. Hermann Weyl later wrote that "Emmy Noether—her courage, her frankness, her unconcern about her own fate, her conciliatory spirit—was in the midst of all the hatred and meanness, despair and sorrow surrounding us, a moral solace." Typically, Noether remained focused on mathematics, gathering students in her apartment to discuss class field theory. When one of her students appeared in the uniform of the Nazi paramilitary organization Sturmabteilung (SA), she showed no sign of agitation and, reportedly, even laughed about it later. Bryn Mawr Bryn Mawr College provided a welcoming home for Noether during the last two years of her life As dozens of newly unemployed professors began searching for positions outside of Germany, their colleagues in the United States sought to provide assistance and job opportunities for them. Albert Einstein and Hermann Weyl were appointed by the Institute for Advanced Study in Princeton, while others worked to find a sponsor required for legal immigration. Noether was contacted by representatives of two educational institutions, Bryn Mawr College in the United States and Somerville College at the University of Oxford in England. After a series of negotiations with the Rockefeller Foundation, a grant to Bryn Mawr was approved for Noether and she took a position there, starting in late 1933. At Bryn Mawr, Noether met and befriended Anna Wheeler, who had studied at Göttingen just before Noether arrived there. Another source of support at the college was the Bryn Mawr president, Marion Edwards Park, who enthusiastically invited mathematicians in the area to "see Dr. Noether in action!" Noether and a small team of students worked quickly through van der Waerden's 1930 book Moderne Algebra I and parts of Erich Hecke's Theorie der algebraischen Zahlen (Theory of algebraic numbers, 1908). In 1934, Noether began lecturing at the Institute for Advanced Study in Princeton upon the invitation of Abraham Flexner and Oswald Veblen. She also worked with and supervised Abraham Albert and Harry Vandiver. However, she remarked about Princeton University that she was not welcome at the "men's university, where nothing female is admitted". Her time in the United States was pleasant, surrounded as she was by supportive colleagues and absorbed in her favorite subjects. In the summer of 1934 she briefly returned to Germany to see Emil Artin and her brother Fritz before he left for Tomsk. Although many of her former colleagues had been forced out of the universities, she was able to use the library as a "foreign scholar". Death Noether's remains were placed under the walkway surrounding the cloisters of Bryn Mawr's M. Carey Thomas Library In April 1935 doctors discovered a tumor in Noether's pelvis. Worried about complications from surgery, they ordered two days of bed rest first. During the operation they discovered an ovarian cyst "the size of a large cantaloupe". Two smaller tumors in her uterus appeared to be benign and were not removed, to avoid prolonging surgery. For three days she appeared to convalesce normally, and she recovered quickly from a circulatory collapse on the fourth. On 14 April she fell unconscious, her temperature soared to 109 °F (42.8 °C), and she died. "[I]t is not easy to say what had occurred in Dr. Noether", one of the physicians wrote. "It is possible that there was some form of unusual and virulent infection, which struck the base of the brain where the heat centers are supposed to be located." A few days after Noether's death her friends and associates at Bryn Mawr held a small memorial service at College President Park's house. Hermann Weyl and Richard Brauer traveled from Princeton and spoke with Wheeler and Taussky about their departed colleague. In the months which followed, written tributes began to appear around the globe: Albert Einstein joined van der Waerden, Weyl, and Pavel Alexandrov in paying their respects. Her body was cremated and the ashes interred under the walkway around the cloisters of the M. Carey Thomas Library at Bryn Mawr. Contributions to mathematics and physics First and foremost Noether is remembered by mathematicians as an algebraist and for her work in topology. Physicists appreciate her best for her famous theorem because of its far-ranging consequences for theoretical physics and dynamic systems. She showed an acute propensity for abstract thought, which allowed her to approach problems of mathematics in fresh and original ways. Her friend and colleague Hermann Weyl described her scholarly output in three epochs: Emmy Noether's scientific production fell into three clearly distinct epochs: (1) the period of relative dependence, 1907–1919; (2) the investigations grouped around the general theory of ideals 1920–1926; (3) the study of the non-commutative algebras, their representations by linear transformations, and their application to the study of commutative number fields and their arithmetics. —Weyl 1935 In the first epoch (1907–19), Noether dealt primarily with differential and algebraic invariants, beginning with her dissertation under Paul Gordan. Her mathematical horizons broadened, and her work became more general and abstract, as she became acquainted with the work of David Hilbert, through close interactions with a successor to Gordan, Ernst Sigismund Fischer. After moving to Göttingen in 1915, she produced her seminal work for physics, the two Noether's theorems. In the second epoch (1920–26), Noether devoted herself to developing the theory of mathematical rings. In the third epoch (1927–35), Noether focused on noncommutative algebra, linear transformations, and commutative number fields. Historical context In the century from 1832 to Noether's death in 1935, the field of mathematics—specifically algebra—underwent a profound revolution, whose reverberations are still being felt. Mathematicians of previous centuries had worked on practical methods for solving specific types of equations, e.g., cubic, quartic, and quintic equations, as well as on the related problem of constructing regular polygons using compass and straightedge. Beginning with Carl Friedrich Gauss's 1832 proof that prime numbers such as five can be factored in Gaussian integers, Évariste Galois's introduction of permutation groups in 1832 (although, because of his death, his papers were only published in 1846 by Liouville), William Rowan Hamilton's discovery of quaternions in 1843, and Arthur Cayley's more modern definition of groups in 1854, research turned to determining the properties of ever-more-abstract systems defined by ever-more-universal rules. Noether's most important contributions to mathematics were to the development of this new field, abstract algebra. Abstract algebra and begriffliche Mathematik (conceptual mathematics) Two of the most basic objects in abstract algebra are groups and rings. A group consists of a set of elements and a single operation which combines a first and a second element and returns a third. The operation must satisfy certain constraints for it to determine a group: It must be closed (when applied to any pair of elements of the associated set, the generated element must also be a member of that set), it must be associative, there must be an identity element (an element which, when combined with another element using the operation, results in the original element, such as adding zero to a number or multiplying it by one), and for every element there must be an inverse element. A ring likewise, has a set of elements, but now has two operations. The first operation must make the set a group, and the second operation is associative and distributive with respect to the first operation. It may or may not be commutative; this means that the result of applying the operation to a first and a second element is the same as to the second and first—the order of the elements does not matter. If every non-zero element has a multiplicative inverse (an element x such that ax = xa = 1), the ring is called a division ring. A field is defined as a commutative division ring. Groups are frequently studied through group representations. In their most general form, these consist of a choice of group, a set, and an action of the group on the set, that is, an operation which takes an element of the group and an element of the set and returns an element of the set. Most often, the set is a vector space, and the group represents symmetries of the vector space. For example, there is a group which represents the rigid rotations of space. This is a type of symmetry of space, because space itself does not change when it is rotated even though the positions of objects in it do. Noether used these sorts of symmetries in her work on invariants in physics. A powerful way of studying rings is through their modules. A module consists of a choice of ring, another set, usually distinct from the underlying set of the ring and called the underlying set of the module, an operation on pairs of elements of the underlying set of the module, and an operation which takes an element of the ring and an element of the module and returns an element of the module. The underlying set of the module and its operation must form a group. A module is a ring-theoretic version of a group representation: Ignoring the second ring operation and the operation on pairs of module elements determines a group representation. The real utility of modules is that the kinds of modules that exist and their interactions, reveal the structure of the ring in ways that are not apparent from the ring itself. An important special case of this is an algebra. (The word algebra means both a subject within mathematics as well as an object studied in the subject of algebra.) An algebra consists of a choice of two rings and an operation which takes an element from each ring and returns an element of the second ring. This operation makes the second ring into a module over the first. Often the first ring is a field. Words such as "element" and "combining operation" are very general, and can be applied to many real-world and abstract situations. Any set of things that obeys all the rules for one (or two) operation(s) is, by definition, a group (or ring), and obeys all theorems about groups (or rings). Integer numbers, and the operations of addition and multiplication, are just one example. For example, the elements might be computer data words, where the first combining operation is exclusive or and the second is logical conjunction. Theorems of abstract algebra are powerful because they are general; they govern many systems. It might be imagined that little could be concluded about objects defined with so few properties, but precisely therein lay Noether's gift: to discover the maximum that could be concluded from a given set of properties, or conversely, to identify the minimum set, the essential properties responsible for a particular observation. Unlike most mathematicians, she did not make abstractions by generalizing from known examples; rather, she worked directly with the abstractions. As van der Waerden recalled in his obituary of her, The maxim by which Emmy Noether was guided throughout her work might be formulated as follows: "Any relationships between numbers, functions, and operations become transparent, generally applicable, and fully productive only after they have been isolated from their particular objects and been formulated as universally valid concepts. This is the begriffliche Mathematik (purely conceptual mathematics) that was characteristic of Noether. This style of mathematics was adopted by other mathematicians and, after her death, flowered into new forms, such as category theory. Integers as an example of a ring The integers form a commutative ring whose elements are the integers, and the combining operations are addition and multiplication. Any pair of integers can be added or multiplied, always resulting in another integer, and the first operation, addition, is commutative, i.e., for any elements a and b in the ring, a + b = b + a. The second operation, multiplication, also is commutative, but that need not be true for other rings, meaning that a combined with b might be different from b combined with a. Examples of noncommutative rings include matrices and quaternions. The integers do not form a division ring, because the second operation cannot always be inverted; there is no integer a such that 3 × a = 1. The integers have additional properties which do not generalize to all commutative rings. An important example is the fundamental theorem of arithmetic, which says that every positive integer can be factored uniquely into prime numbers. Unique factorizations do not always exist in other rings, but Noether found a unique factorization theorem, now called the Lasker–Noether theorem, for the ideals of many rings. Much of Noether's work lay in determining what properties do hold for all rings, in devising novel analogs of the old integer theorems, and in determining the minimal set of assumptions required to yield certain properties of rings. First epoch (1908–19) Algebraic invariant theory Table 2 from Noether's dissertation on invariant theory. This table collects 202 of the 331 invariants of ternary biquadratic forms. These forms are graded in two variables x and u. The horizontal direction of the table lists the invariants with increasing grades in x, while the vertical direction lists them with increasing grades in u. Much of Noether's work in the first epoch of her career was associated with invariant theory, principally algebraic invariant theory. Invariant theory is concerned with expressions that remain constant (invariant) under a group of transformations. As an everyday example, if a rigid yardstick is rotated, the coordinates (x, y, z) of its endpoints change, but its length L given by the formula L2 = Δx2 + Δy2 + Δz2 remains the same. Invariant theory was an active area of research in the later nineteenth century, prompted in part by Felix Klein's Erlangen program, according to which different types of geometry should be characterized by their invariants under transformations, e.g., the cross-ratio of projective geometry. The archetypal example of an invariant is the discriminant B2 − 4AC of a binary quadratic form Ax2 + Bxy + Cy2. This is called an invariant because it is unchanged by linear substitutions xax + by, ycx + dy with determinant adbc = 1. These substitutions form the special linear group SL2. (There are no invariants under the general linear group of all invertible linear transformations because these transformations can be multiplication by a scaling factor. To remedy this, classical invariant theory also considered relative invariants, which were forms invariant up to a scale factor.) One can ask for all polynomials in A, B, and C that are unchanged by the action of SL2; these are called the invariants of binary quadratic forms, and turn out to be the polynomials in the discriminant. More generally, one can ask for the invariants of homogeneous polynomials A0xry0 + ... + Arx0yr of higher degree, which will be certain polynomials in the coefficients A0, ..., Ar, and more generally still, one can ask the similar question for homogeneous polynomials in more than two variables. One of the main goals of invariant theory was to solve the "finite basis problem". The sum or product of any two invariants is invariant, and the finite basis problem asked whether it was possible to get all the invariants by starting with a finite list of invariants, called generators, and then, adding or multiplying the generators together. For example, the discriminant gives a finite basis (with one element) for the invariants of binary quadratic forms. Noether's advisor, Paul Gordan, was known as the "king of invariant theory", and his chief contribution to mathematics was his 1870 solution of the finite basis problem for invariants of homogeneous polynomials in two variables. He proved this by giving a constructive method for finding all of the invariants and their generators, but was not able to carry out this constructive approach for invariants in three or more variables. In 1890, David Hilbert proved a similar statement for the invariants of homogeneous polynomials in any number of variables. Furthermore, his method worked, not only for the special linear group, but also for some of its subgroups such as the special orthogonal group. His first proof caused some controversy because it did not give a method for constructing the generators, although in later work he made his method constructive. For her thesis, Noether extended Gordan's computational proof to homogeneous polynomials in three variables. Noether's constructive approach made it possible to study the relationships among the invariants. Later, after she had turned to more abstract methods, Noether called her thesis Mist (crap) and Formelngestrüpp (a jungle of equations). Galois theory Galois theory concerns transformations of number fields that permute the roots of an equation. Consider a polynomial equation of a variable x of degree n, in which the coefficients are drawn from some ground field, which might be, for example, the field of real numbers, rational numbers, or the integers modulo 7. There may or may not be choices of x, which make this polynomial evaluate to zero. Such choices, if they exist, are called roots. If the polynomial is x2 + 1 and the field is the real numbers, then the polynomial has no roots, because any choice of x makes the polynomial greater than or equal to one. If the field is extended, however, then the polynomial may gain roots, and if it is extended enough, then it always has a number of roots equal to its degree. Continuing the previous example, if the field is enlarged to the complex numbers, then the polynomial gains two roots, i and −i, where i is the imaginary unit, that is, i 2 = −1. More generally, the extension field in which a polynomial can be factored into its roots is known as the splitting field of the polynomial. The Galois group of a polynomial is the set of all ways of transforming the splitting field, while preserving the ground field and the roots of the polynomial. (In mathematical jargon, these transformations are called automorphisms.) The Galois group of x2 + 1 consists of two elements: The identity transformation, which sends every complex number to itself, and complex conjugation, which sends i to −i. Since the Galois group does not change the ground field, it leaves the coefficients of the polynomial unchanged, so it must leave the set of all roots unchanged. Each root can move to another root, however, so transformation determines a permutation of the n roots among themselves. The significance of the Galois group derives from the fundamental theorem of Galois theory, which proves that the fields lying between the ground field and the splitting field are in one-to-one correspondence with the subgroups of the Galois group. In 1918, Noether published a seminal paper on the inverse Galois problem. Instead of determining the Galois group of transformations of a given field and its extension, Noether asked whether, given a field and a group, it always is possible to find an extension of the field that has the given group as its Galois group. She reduced this to " Noether's problem", which asks whether the fixed field of a subgroup G of the permutation group Sn acting on the field k(x1, ... , xn) always is a pure transcendental extension of the field k. (She first mentioned this problem in a 1913 paper, where she attributed the problem to her colleague Fischer.) She showed this was true for n = 2, 3, or 4. In 1969, R. G. Swan found a counter-example to Noether's problem, with n = 47 and G a cyclic group of order 47 (although this group can be realized as a Galois group over the rationals in other ways). The inverse Galois problem remains unsolved. Physics Noether was brought to Göttingen in 1915 by David Hilbert and Felix Klein, who wanted her expertise in invariant theory to help them in understanding general relativity, a geometrical theory of gravitation developed mainly by Albert Einstein. Hilbert had observed that the conservation of energy seemed to be violated in general relativity, due to the fact that gravitational energy could itself gravitate. Noether provided the resolution of this paradox, and a fundamental tool of modern theoretical physics, with Noether's first theorem, which she proved in 1915, but did not publish until 1918. She solved the problem not only for general relativity, but determined the conserved quantities for every system of physical laws that possesses some continuous symmetry. Upon receiving her work, Einstein wrote to Hilbert: "Yesterday I received from Miss Noether a very interesting paper on invariants. I'm impressed that such things can be understood in such a general way. The old guard at Göttingen should take some lessons from Miss Noether! She seems to know her stuff." For illustration, if a physical system behaves the same, regardless of how it is oriented in space, the physical laws that govern it are rotationally symmetric; from this symmetry, Noether's theorem shows the angular momentum of the system must be conserved. The physical system itself need not be symmetric; a jagged asteroid tumbling in space conserves angular momentum despite its asymmetry. Rather, the symmetry of the physical laws governing the system is responsible for the conservation law. As another example, if a physical experiment has the same outcome at any place and at any time, then its laws are symmetric under continuous translations in space and time; by Noether's theorem, these symmetries account for the conservation laws of linear momentum and energy within this system, respectively. Noether's theorem has become a fundamental tool of modern theoretical physics, both because of the insight it gives into conservation laws, and also, as a practical calculation tool. Her theorem allows researchers to determine the conserved quantities from the observed symmetries of a physical system. Conversely, it facilitates the description of a physical system based on classes of hypothetical physical laws. For illustration, suppose that a new physical phenomenon is discovered. Noether's theorem provides a test for theoretical models of the phenomenon: if the theory has a continuous symmetry, then Noether's theorem guarantees that the theory has a conserved quantity, and for the theory to be correct, this conservation must be observable in experiments. Second epoch (1920–26) Although the results of Noether's first epoch were impressive and useful, her fame as a mathematician rests more on the groundbreaking work she did in her second and third epochs, as noted by Hermann Weyl and B. L. van der Waerden in their obituaries of her. In these epochs, she was not merely applying ideas and methods of earlier mathematicians; rather, she was crafting new systems of mathematical definitions that would be used by future mathematicians. In particular, she developed a completely new theory of ideals in rings, generalizing earlier work of Richard Dedekind. She is also renowned for developing ascending chain conditions, a simple finiteness condition that yielded powerful results in her hands. Such conditions and the theory of ideals enabled Noether to generalize many older results and to treat old problems from a new perspective, such as elimination theory and the algebraic varieties that had been studied by her father. Ascending and descending chain conditions In this epoch, Noether became famous for her deft use of ascending (Teilerkettensatz) or descending (Vielfachenkettensatz) chain conditions. A sequence of non-empty subsets A1, A2, A3, etc. of a set S is usually said to be ascending, if each is a subset of the next A_{1} \subset A_{2} \subset A_{3} \subset \cdots. Conversely, a sequence of subsets of S is called descending if each contains the next subset: A_{1} \supset A_{2} \supset A_{3} \supset \cdots. A chain becomes constant after a finite number of steps if there is an n such that A_n =A_m for all m ≥ n. A collection of subsets of a given set satisfies the ascending chain condition if any ascending sequence becomes constant after a finite number of steps. It satisfies the descending chain condition if any descending sequence becomes constant after a finite number of steps. Ascending and descending chain conditions are general, meaning that they can be applied to many types of mathematical objects—and, on the surface, they might not seem very powerful. Noether showed how to exploit such conditions, however, to maximum advantage: for example, how to use them to show that every set of sub-objects has a maximal/minimal element or that a complex object can be generated by a smaller number of elements. These conclusions often are crucial steps in a proof. Many types of objects in abstract algebra can satisfy chain conditions, and usually if they satisfy an ascending chain condition, they are called Noetherian in her honour. By definition, a Noetherian ring satisfies an ascending chain condition on its left and right ideals, whereas a Noetherian group is defined as a group in which every strictly ascending chain of subgroups is finite. A Noetherian module is a module in which every strictly ascending chain of submodules breaks off after a finite number. A Noetherian space is a topological space in which every strictly increasing chain of open subspaces breaks off after a finite number of terms; this definition is made so that the spectrum of a Noetherian ring is a Noetherian topological space. The chain condition often is "inherited" by sub-objects. For example, all subspaces of a Noetherian space, are Noetherian themselves; all subgroups and quotient groups of a Noetherian group are likewise, Noetherian; and, mutatis mutandis, the same holds for submodules and quotient modules of a Noetherian module. All quotient rings of a Noetherian ring are Noetherian, but that does not necessarily hold for its subrings. The chain condition also may be inherited by combinations or extensions of a Noetherian object. For example, finite direct sums of Noetherian rings are Noetherian, as is the ring of formal power series over a Noetherian ring. Another application of such chain conditions is in Noetherian induction—also known as well-founded induction—which is a generalization of mathematical induction. It frequently is used to reduce general statements about collections of objects to statements about specific objects in that collection. Suppose that S is a partially ordered set. One way of proving a statement about the objects of S is to assume the existence of a counterexample and deduce a contradiction, thereby proving the contrapositive of the original statement. The basic premise of Noetherian induction is that every non-empty subset of S contains a minimal element. In particular, the set of all counterexamples contains a minimal element, the minimal counterexample. In order to prove the original statement, therefore, it suffices to prove something seemingly much weaker: For any counterexample, there is a smaller counterexample. Commutative rings, ideals, and modules Noether's paper, Idealtheorie in Ringbereichen (Theory of Ideals in Ring Domains, 1921), is the foundation of general commutative ring theory, and gives one of the first general definitions of a commutative ring. Before her paper, most results in commutative algebra were restricted to special examples of commutative rings, such as polynomial rings over fields or rings of algebraic integers. Noether proved that in a ring which satisfies the ascending chain condition on ideals, every ideal is finitely generated. In 1943, French mathematician Claude Chevalley coined the term, Noetherian ring, to describe this property. A major result in Noether's 1921 paper is the Lasker–Noether theorem, which extends Lasker's theorem on the primary decomposition of ideals of polynomial rings to all Noetherian rings. The Lasker–Noether theorem can be viewed as a generalization of the fundamental theorem of arithmetic which states that any positive integer can be expressed as a product of prime numbers, and that this decomposition is unique. Noether's work Abstrakter Aufbau der Idealtheorie in algebraischen Zahl- und Funktionenkörpern (Abstract Structure of the Theory of Ideals in Algebraic Number and Function Fields, 1927) characterized the rings in which the ideals have unique factorization into prime ideals as the Dedekind domains: integral domains that are Noetherian, 0 or 1- dimensional, and integrally closed in their quotient fields. This paper also contains what now are called the isomorphism theorems, which describe some fundamental natural isomorphisms, and some other basic results on Noetherian and Artinian modules. Elimination theory In 1923–24, Noether applied her ideal theory to elimination theory—in a formulation that she attributed to her student, Kurt Hentzelt—showing that fundamental theorems about the factorization of polynomials could be carried over directly. Traditionally, elimination theory is concerned with eliminating one or more variables from a system of polynomial equations, usually by the method of resultants. For illustration, the system of equations often can be written in the form of a matrix M (missing the variable x) times a vector v (having only different powers of x) equaling the zero vector, M•v = 0. Hence, the determinant of the matrix M must be zero, providing a new equation in which the variable x has been eliminated. Invariant theory of finite groups Techniques such as Hilbert's original non-constructive solution to the finite basis problem could not be used to get quantitative information about the invariants of a group action, and furthermore, they did not apply to all group actions. In her 1915 paper, Noether found a solution to the finite basis problem for a finite group of transformations G acting on a finite dimensional vector space over a field of characteristic zero. Her solution shows that the ring of invariants is generated by homogenous invariants whose degree is less than, or equal to, the order of the finite group; this is called, Noether's bound. Her paper gave two proofs of Noether's bound, both of which also work when the characteristic of the field is coprime to |G|!, the factorial of the order |G| of the group G. The number of generators need not satisfy Noether's bound when the characteristic of the field divides the |G|, but Noether was not able to determine whether the bound was correct when the characteristic of the field divides |G|! but not |G|. For many years, determining the truth or falsity of the bound in this case was an open problem called "Noether's gap". It finally was resolved independently by Fleischmann in 2000 and Fogarty in 2001, who both showed that the bound remains true. In her 1926 paper, Noether extended Hilbert's theorem to representations of a finite group over any field; the new case that did not follow from Hilbert's work, is when the characteristic of the field divides the order of the group. Noether's result was later extended by William Haboush to all reductive groups by his proof of the Mumford conjecture. In this paper Noether also introduced the Noether normalization lemma, showing that a finitely generated domain A over a field k has a set x1, ... , xn of algebraically independent elements such that A is integral over k[x1, ... , xn]. Contributions to topology A continuous deformation (homotopy) of a coffee cup into a doughnut (torus) and back As noted by Pavel Alexandrov and Hermann Weyl in their obituaries, Noether's contributions to topology illustrate her generosity with ideas and how her insights could transform entire fields of mathematics. In topology, mathematicians study the properties of objects that remain invariant even under deformation, properties such as their connectedness. A common joke is that a topologist cannot distinguish a donut from a coffee mug, since they can be continuously deformed into one another. Noether is credited with the fundamental ideas that led to the development of algebraic topology from the earlier combinatorial topology, specifically, the idea of homology groups. According to the account of Alexandrov, Noether attended lectures given by Heinz Hopf and him in the summers of 1926 and 1927, where "she continually made observations, which were often deep and subtle" and he continues that, When... she first became acquainted with a systematic construction of combinatorial topology, she immediately observed that it would be worthwhile to study directly the groups of algebraic complexes and cycles of a given polyhedron and the subgroup of the cycle group consisting of cycles homologous to zero; instead of the usual definition of Betti numbers, she suggested immediately defining the Betti group as the complementary (quotient) group of the group of all cycles by the subgroup of cycles homologous to zero. This observation now seems self-evident. But in those years (1925–28) this was a completely new point of view. Noether's suggestion that topology be studied algebraically, was adopted immediately by Hopf, Alexandrov, and others, and it became a frequent topic of discussion among the mathematicians of Göttingen. Noether observed that her idea of a Betti group makes the Euler–Poincaré formula simpler to understand, and Hopf's own work on this subject "bears the imprint of these remarks of Emmy Noether". Noether mentions her own topology ideas only as an aside in one 1926 publication, where she cites it as an application of group theory. The algebraic approach to topology was developed independently in Austria. In a 1926–27 course given in Vienna, Leopold Vietoris defined a homology group, which was developed by Walther Mayer, into an axiomatic definition in 1928. Helmut Hasse worked with Noether and others to found the theory of central simple algebras Third epoch (1927–35) Hypercomplex numbers and representation theory Much work on hypercomplex numbers and group representations was carried out in the nineteenth and early twentieth centuries, but remained disparate. Noether united the results and gave the first general representation theory of groups and algebras. Briefly, Noether subsumed the structure theory of associative algebras and the representation theory of groups into a single arithmetic theory of modules and ideals in rings satisfying ascending chain conditions. This single work by Noether was of fundamental importance for the development of modern algebra. Noncommutative algebra Noether also was responsible for a number of other advancements in the field of algebra. With Emil Artin, Richard Brauer, and Helmut Hasse, she founded the theory of central simple algebras. A seminal paper by Noether, Helmut Hasse, and Richard Brauer pertains to division algebras, which are algebraic systems in which division is possible. They proved two important theorems: a local-global theorem stating that if a finite dimensional central division algebra over a number field splits locally everywhere then it splits globally (so is trivial), and from this, deduced their Hauptsatz ("main theorem"): every finite dimensional central division algebra over an algebraic number field F splits over a cyclic cyclotomic extension. These theorems allow one to classify all finite dimensional central division algebras over a given number field. A subsequent paper by Noether showed, as a special case of a more general theorem, that all maximal subfields of a division algebra D are splitting fields. This paper also contains the Skolem–Noether theorem which states that any two embeddings of an extension of a field k into a finite dimensional central simple algebra over k, are conjugate. The Brauer–Noether theorem gives a characterization of the splitting fields of a central division algebra over a field. Assessment, recognition, and memorials The Emmy Noether Campus at the University of Siegen is home to its mathematics and physics departments Noether's work continues to be relevant for the development of theoretical physics and mathematics and she is consistently ranked as one of the greatest mathematicians of the twentieth century. In his obituary, fellow algebraist BL van der Waerden says that her mathematical originality was "absolute beyond comparison", and Hermann Weyl said that Noether "changed the face of algebra by her work". During her lifetime and even until today, Noether has been characterized as the greatest woman mathematician in recorded history by mathematicians such as Pavel Alexandrov, Hermann Weyl, and Jean Dieudonné. In a letter to The New York Times, Albert Einstein wrote: In the judgment of the most competent living mathematicians, Fräulein Noether was the most significant creative mathematical genius thus far produced since the higher education of women began. In the realm of algebra, in which the most gifted mathematicians have been busy for centuries, she discovered methods which have proved of enormous importance in the development of the present-day younger generation of mathematicians. On 2 January 1935, a few months before her death, mathematician Norbert Wiener wrote that Miss Noether is... the greatest woman mathematician who has ever lived; and the greatest woman scientist of any sort now living, and a scholar at least on the plane of Madame Curie. At an exhibition at the 1964 World's Fair devoted to Modern Mathematicians, Noether was the only woman represented among the notable mathematicians of the modern world. Noether has been honored in several memorials, • The Association for Women in Mathematics holds a Noether Lecture to honour women in mathematics every year; in its 2005 pamphlet for the event, the Association characterizes Noether as "one of the great mathematicians of her time, someone who worked and struggled for what she loved and believed in. Her life and work remain a tremendous inspiration". • Consistent with her dedication to her students, the University of Siegen houses its mathematics and physics departments in buildings on the Emmy Noether Campus. • The German Research Foundation ( Deutsche Forschungsgemeinschaft) operates the Emmy Noether Programme, a scholarship providing funding to promising young post-doctorate scholars in their further research and teaching activities. • A street in her hometown, Erlangen, has been named after Emmy Noether and her father, Max Noether. • The successor to the secondary school she attended in Erlangen has been renamed as the Emmy Noether School. In fiction, Emmy Nutter, the physics professor in "The God Patent" by Ransom Stephens, is based on Emmy Noether Farther from home, • The crater Nöther on the far side of the Moon is named after her. • The 7001 Noether asteroid also is named for Emmy Noether. List of doctoral students Date Student name Dissertation title and English translation University Publication 1911.12.16 Falckenberg, Hans Verzweigungen von Lösungen nichtlinearer Differentialgleichungen Ramifications of Solutions of Nonlinear Differential Equations§ Erlangen Leipzig 1912 1916.03.04 Seidelmann, Fritz Die Gesamtheit der kubischen und biquadratischen Gleichungen mit Affekt bei beliebigem Rationalitätsbereich Complete Set of Cubic and Biquadratic Equations with Affect in an Arbitrary Rationality Domain§ Erlangen Erlangen 1916 1925.02.25 Hermann, Grete Die Frage der endlich vielen Schritte in der Theorie der Polynomideale unter Benutzung nachgelassener Sätze von Kurt Hentzelt The Question of the Finite Number of Steps in the Theory of Ideals of Polynomials using Theorems of the Late Kurt Hentzelt§ Göttingen Berlin 1926 1926.07.14 Grell, Heinrich Beziehungen zwischen den Idealen verschiedener Ringe Relationships between the Ideals of Various Rings§ Göttingen Berlin 1927 1927 Doräte, Wilhelm Über einem verallgemeinerten Gruppenbegriff On a Generalized Conceptions of Groups§ Göttingen Berlin 1927 died before defense Hölzer, Rudolf Zur Theorie der primären Ringe On the Theory of Primary Rings§ Göttingen Berlin 1927 1929.06.12 Weber, Werner Idealtheoretische Deutung der Darstellbarkeit beliebiger natürlicher Zahlen durch quadratische Formen Ideal-theoretic Interpretation of the Representability of Arbitrary Natural Numbers by Quadratic Forms§ Göttingen Berlin 1930 1929.06.26 Levitski, Jakob Über vollständig reduzible Ringe und Unterringe On Completely Reducible Rings and Subrings§ Göttingen Berlin 1931 1930.06.18 Deuring, Max Zur arithmetischen Theorie der algebraischen Funktionen On the Arithmetic Theory of Algebraic Functions§ Göttingen Berlin 1932 1931.07.29 Fitting, Hans Zur Theorie der Automorphismenringe Abelscher Gruppen und ihr Analogon bei nichtkommutativen Gruppen On the Theory of Automorphism-Rings of Abelian Groups and Their Analogs in Noncommutative Groups§ Göttingen Berlin 1933 1933.07.27 Witt, Ernst Riemann-Rochscher Satz und Zeta-Funktion im Hyperkomplexen The Riemann-Roch Theorem and Zeta Function in Hypercomplex Numbers§ Göttingen Berlin 1934 1933.12.06 Tsen, Chiungtze Algebren über Funktionenkörpern Algebras over Function Fields§ Göttingen Göttingen 1934 1934 Schilling, Otto Über gewisse Beziehungen zwischen der Arithmetik hyperkomplexer Zahlsysteme und algebraischer Zahlkörper On Certain Relationships between the Arithmetic of Hypercomplex Number Systems and Algebraic Number Fields§ Marburg Braunschweig 1935 1935 Stauffer, Ruth The construction of a normal basis in a separable extension field Bryn Mawr Baltimore 1936 1935 Vorbeck, Werner Nichtgaloissche Zerfällungskörper einfacher Systeme Non-Galois Splitting Fields of Simple Systems§ Göttingen 1936 Wichmann, Wolfgang Anwendungen der p-adischen Theorie im Nichtkommutativen Applications of the p-adic Theory in Noncommutative Algebras§ Göttingen Monatshefte für Mathematik und Physik (1936) 44, 203–24. Eponymous mathematical topics • Noetherian • Noetherian group • Noetherian ring • Noetherian module • Noetherian space • Noetherian induction • Noetherian scheme • Noether normalization lemma • Noether problem • Noether's theorem • Noether's second theorem • Lasker–Noether theorem • Skolem–Noether theorem • Albert–Brauer–Hasse–Noether theorem Retrieved from " http://en.wikipedia.org/w/index.php?title=Emmy_Noether&oldid=548871169"
__label__pos
0.896006
~kennylevinsen/wlsunset wlsunset/color_math.c -rw-r--r-- 6.0 KiB f32f6963Kenny Levinsen readme: Update discuss section 26 days ago 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 #define _POSIX_C_SOURCE 200809L #include <math.h> #include <errno.h> #include <time.h> #include "color_math.h" static double SOLAR_START_TWILIGHT = RADIANS(90.833 + 6.0); static double SOLAR_END_TWILIGHT = RADIANS(90.833 - 3.0); static int days_in_year(int year) { int leap = (year % 4 == 0 && year % 100 != 0) || year % 400 == 0; return leap ? 366 : 365; } static double date_orbit_angle(struct tm *tm) { return 2 * M_PI / (double)days_in_year(tm->tm_year + 1900) * tm->tm_yday; } static double equation_of_time(double orbit_angle) { // https://www.esrl.noaa.gov/gmd/grad/solcalc/solareqns.PDF return 4 * (0.000075 + 0.001868 * cos(orbit_angle) - 0.032077 * sin(orbit_angle) - 0.014615 * cos(2*orbit_angle) - 0.040849 * sin(2*orbit_angle)); } static double sun_declination(double orbit_angle) { // https://www.esrl.noaa.gov/gmd/grad/solcalc/solareqns.PDF return 0.006918 - 0.399912 * cos(orbit_angle) + 0.070257 * sin(orbit_angle) - 0.006758 * cos(2*orbit_angle) + 0.000907 * sin(2*orbit_angle) - 0.002697 * cos(3*orbit_angle) + 0.00148 * sin(3*orbit_angle); } static double sun_hour_angle(double latitude, double declination, double target_sun) { // https://www.esrl.noaa.gov/gmd/grad/solcalc/solareqns.PDF return acos(cos(target_sun) / cos(latitude) * cos(declination) - tan(latitude) * tan(declination)); } static time_t hour_angle_to_time(double hour_angle, double eqtime) { // https://www.esrl.noaa.gov/gmd/grad/solcalc/solareqns.PDF return DEGREES((4.0 * M_PI - 4 * hour_angle - eqtime) * 60); } static enum sun_condition condition(double latitude_rad, double sun_declination) { int sign_lat = signbit(latitude_rad) == 0; int sign_decl = signbit(sun_declination) == 0; return sign_lat == sign_decl ? MIDNIGHT_SUN : POLAR_NIGHT; } enum sun_condition calc_sun(struct tm *tm, double latitude, struct sun *sun) { double orbit_angle = date_orbit_angle(tm); double decl = sun_declination(orbit_angle); double eqtime = equation_of_time(orbit_angle); double ha_twilight = sun_hour_angle(latitude, decl, SOLAR_START_TWILIGHT); double ha_daylight = sun_hour_angle(latitude, decl, SOLAR_END_TWILIGHT); sun->dawn = hour_angle_to_time(fabs(ha_twilight), eqtime); sun->dusk = hour_angle_to_time(-fabs(ha_twilight), eqtime); sun->sunrise = hour_angle_to_time(fabs(ha_daylight), eqtime); sun->sunset = hour_angle_to_time(-fabs(ha_daylight), eqtime); return isnan(ha_twilight) || isnan(ha_daylight) ? condition(latitude, decl) : NORMAL; } /* * Illuminant D, or daylight locus, is is a "standard illuminant" used to * describe natural daylight. It is on this locus that D65, the whitepoint used * by most monitors and assumed by wlsunset, is defined. * * This approximation is strictly speaking only well-defined between 4000K and * 25000K, but we stretch it a bit further down for transition purposes. */ static int illuminant_d(int temp, double *x, double *y) { // https://en.wikipedia.org/wiki/Standard_illuminant#Illuminant_series_D if (temp >= 2500 && temp <= 7000) { *x = 0.244063 + 0.09911e3 / temp + 2.9678e6 / pow(temp, 2) - 4.6070e9 / pow(temp, 3); } else if (temp > 7000 && temp <= 25000) { *x = 0.237040 + 0.24748e3 / temp + 1.9018e6 / pow(temp, 2) - 2.0064e9 / pow(temp, 3); } else { errno = EINVAL; return -1; } *y = (-3 * pow(*x, 2)) + (2.870 * (*x)) - 0.275; return 0; } /* * Planckian locus, or black body locus, describes the color of a black body at * a certain temperatures. This is not entirely equivalent to daylight due to * atmospheric effects. * * This approximation is only valid from 1667K to 25000K. */ static int planckian_locus(int temp, double *x, double *y) { // https://en.wikipedia.org/wiki/Planckian_locus#Approximation if (temp >= 1667 && temp <= 4000) { *x = -0.2661239e9 / pow(temp, 3) - 0.2343589e6 / pow(temp, 2) + 0.8776956e3 / temp + 0.179910; if (temp <= 2222) { *y = -1.1064814 * pow(*x, 3) - 1.34811020 * pow(*x, 2) + 2.18555832 * (*x) - 0.20219683; } else { *y = -0.9549476 * pow(*x, 3) - 1.37418593 * pow(*x, 2) + 2.09137015 * (*x) - 0.16748867; } } else if (temp > 4000 && temp < 25000) { *x = -3.0258469e9 / pow(temp, 3) + 2.1070379e6 / pow(temp, 2) + 0.2226347e3 / temp + 0.240390; *y = 3.0817580 * pow(*x, 3) - 5.87338670 * pow(*x, 2) + 3.75112997 * (*x) - 0.37001483; } else { errno = EINVAL; return -1; } return 0; } static double srgb_gamma(double value, double gamma) { // https://en.wikipedia.org/wiki/SRGB if (value <= 0.0031308) { return 12.92 * value; } else { return pow(1.055 * value, 1.0/gamma) - 0.055; } } static double clamp(double value) { if (value > 1.0) { return 1.0; } else if (value < 0.0) { return 0.0; } else { return value; } } static void xyz_to_srgb(double x, double y, double z, double *r, double *g, double *b) { // http://www.brucelindbloom.com/index.html?Eqn_RGB_XYZ_Matrix.html *r = srgb_gamma(clamp(3.2404542 * x - 1.5371385 * y - 0.4985314 * z), 2.2); *g = srgb_gamma(clamp(-0.9692660 * x + 1.8760108 * y + 0.0415560 * z), 2.2); *b = srgb_gamma(clamp(0.0556434 * x - 0.2040259 * y + 1.0572252 * z), 2.2); } static void srgb_normalize(double *r, double *g, double *b) { double maxw = fmaxl(*r, fmaxl(*g, *b)); *r /= maxw; *g /= maxw; *b /= maxw; } void calc_whitepoint(int temp, double *rw, double *gw, double *bw) { if (temp == 6500) { *rw = *gw = *bw = 1.0; return; } double x = 1.0, y = 1.0; if (temp >= 25000) { illuminant_d(25000, &x, &y); } else if (temp >= 4000) { illuminant_d(temp, &x, &y); } else if (temp >= 2500) { double x1, y1, x2, y2; illuminant_d(temp, &x1, &y1); planckian_locus(temp, &x2, &y2); double factor = (4000 - temp) / 1500; double sinefactor = (cos(M_PI*factor) + 1.0) / 2.0; x = x1 * sinefactor + x2 * (1.0 - sinefactor); y = y1 * sinefactor + y2 * (1.0 - sinefactor); } else if (temp >= 1667) { planckian_locus(temp, &x, &y); } else { planckian_locus(1667, &x, &y); } double z = 1.0 - x - y; xyz_to_srgb(x, y, z, rw, gw, bw); srgb_normalize(rw, gw, bw); }
__label__pos
0.980852
MongoDB vs. ClustrixDB vs. Hadoop Get help choosing one of these Get news updates about these tools MongoDB ClustrixDB Hadoop Favorites 388 Favorites 0 Favorites 43 Hacker News, Reddit, Stack Overflow Stats • 623 • 535 • 95.1K • 224 • 26 • 0 • - • 29 • 37.1K GitHub Stats No public GitHub repository stats available Description What is MongoDB? MongoDB stores data in JSON-like documents that can vary in structure, offering a dynamic, flexible schema. MongoDB was also designed for high availability and scalability, with built-in replication and auto-sharding. What is ClustrixDB? ClustrixDB is a scale-out SQL database built from the ground up with a distributed shared nothing architecture, automatic data redistribution (so you never need to shard), with built in fault tolerance, all accessible by a simple SQL interface and support for business critical MySQL features – replication, triggers, stored routines, etc. What is Hadoop? The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Pros about this tool Why do you like MongoDB? Why do you like ClustrixDB? Why do you like Hadoop? Cons about this tool Pricing ClustrixDB Pricing Customers Integrations Latest News This Week in Data with Colin Charles 28: Percona Liv... MongoDB Drops ACID MongoDB’s Drive to Multi-Document Transactions Introducing Spark Structured Streaming Support in ES... Spring for Apache Hadoop 2.5.0 GA released Elasticsearch for Apache Hadoop 5.5.0 Interest Over Time Get help choosing one of these
__label__pos
0.955346
close Вход Забыли? вход по аккаунту ? ...Обработка результатов эксперимента. Метод наименьших квадратов код для вставкиСкачать Алексеев Е.Р., Чеснокова О.В. Введение в Octave 1 11. Обработка результатов наименьших квадратов эксперимента . Метод Данная глава посвящена решению часто встречающихся на практике задач по обработке реальных количественных экспериментальных данных, полученных в результате всевозможных научных опытов, технических испытаний методом наименьших квадратов. В первых четырех параграфах читатель познакомится с математическими основами метода наименьших квадратов. Последний пятый параграф посвящён решению задач обработки экспериментальных данных методом наименьших квадратов с использованием пакета Octave. 11.1 Пос т ановка задачи Метод наименьших квадратов (МНК) позволяет по экспериментальным данным подобрать такую аналитическую функцию, которая проходит настолько близко к экспериментальным точкам, насколько это возможно. В общем случае задачу можно сформулировать следующим образом. Пусть в результате эксперимента были получены некая экспериментальная зависимость y ( x ) , представленная в таблице 11.1. Таблица 11.1: x x1 x2 x3 … x n−1 xn y y1 y2 y3 ... y n−1 yn Необходимо построить аналитическую зависимость f ( x , a1 , a 2 ,… a k ) , наиболее точно описывающую результаты эксперимента. Для построения параметров функции f ( x , a1 , a 2 ,… a k ) будем использовать метод наименьших квадратов. Идея метода наименьших квадратов заключается в том, что функцию f (x , a1 , a 2 ,… a k ) необходимо подобрать таким образом, чтобы сумма квадратов отклонений измеренных значений yi от расчётных Y i = f (x i , a1 , a2 ,... , a k ) была бы наименьшей (см. рис. 11.1): n 2 n 2 S ( a1, a 2, ... , a k )=∑ [ y i−Y i ] =∑ [ y i− f ( x i , a 1, a2, ... , ak ) ] → min i=1 (11.1) i=1 Задача состоит из двух этапов: 1. По результатам эксперимента определить внешний вид подбираемой зависимости. 2. Подобрать коэффициенты зависимости Y = f (x , a1 , a 2 , ... , ak ) . Математически задача подбора коэффициентов зависимости сводится к определению коэффициентов a i из условия (11.1). В Octave её можно решать несколькими способами: 1. Решать как задачу поиска минимума функции многих переменных без ограничений с использованием функции sqp. 2. Использовать специализированную функцию polyfit(x,y,n). 3. Используя аппарат высшей математики составить и решить систему алгебраических уравнений для определения коэффициентов a i . Алексеев Е.Р., Чеснокова О.В. Введение в Octave 2 Рисунок 11.1 Геометрическая интерпретация метода наименьших квадра тов 11.2 Подбор параме т ров эксперимен т альной зависимос т и ме т одом наименьших квадра т о в Вспомним некоторые сведения из высшей математики, необходимые для решения задачи подбора зависимости методом наименьших квадратов. Достаточным условием минимума функции S ( a1 , a 2 , ... , ak ) (11.1) является равенство нулю всех её частных производных. Поэтому задача поиска минимума функции (11.1) эквивалентна решению системы алгебраических уравнений: { ∂S =0 ∂ a1 ∂S =0 ∂ a2 . ... ∂S =0 ∂ ak Если параметры a i входят в зависимость Y = f ( x , a1 , a 2 , ... , ak ) получим систему (11.3) из k линейных уравнений с k неизвестным. (11.2) линейно, то Алексеев Е.Р., Чеснокова О.В. Введение в Octave { n ∂f ∑ 2 [ y i− f ( x i , a 1, a2, ... , a k ) ]∂ a =0 ∑ 2 [ y i− f ( x i , a 1, a2, ... , a k ) ]∂∂ af =0 i=1 n 1 i=1 2 3 . (11.3) ... n ∑ 2 [ y i− f ( x i , a 1, a2, ... , a k ) ]∂∂ af i=1 =0 k Составим систему (11.3) для наиболее часто используемых функций. 11.2.1 Подбор коэффициентов линейной зависимости Y =a1 +a 2 x , составим функцию (11.1) Для подбора параметров линейной функции для линейной зависимости: n 2 S (a1 , a 2)=∑ [ y i−a1−a2 x i ] → min . (11.4) i=1 S по a 1 и a 2 , получим систему уравнений: Продифференцировав функцию { { n 2 ∑ [ y i−a1−a2 x i ](−1)=0 i=1 n n n a1 n+a2 ∑ x i =∑ y i ⇒ 2 ∑ [ y i−a1−a2 x i ](−x i )=0 i=1 i=1 n n i =1 i=1 , n (11.5) a1 ∑ x i+a2 ∑ x =∑ y i xi i=1 2 i i =1 решив которую, определим коэффициенты функции Y =a1 +a 2 x : { n ∑ a1 = a 2= 11.2.2 n ∑ xi yi i=1 n n −a2 i=1 n n n i=1 i =1 2 n ∑ y i x i −∑ y i ∑ x i . i=1 n n (11.6) (∑ ) n ∑ xi2− i=1 i=1 xi Подбор коэффициентов полинома k–й степени Для определения параметров зависимости S ( a1 , a 2 , a 3) (11.1): n Y =a1 +a 2 x +a3 x 2 2 S (a1 , a 2 , a 3)=∑ [ y i−a1−a2 x i−a3 x 2i ] → min . i=1 После дифференцирования алгебраических уравнений: S по a1 , a2 составим функцию (11.7) и a3 получим систему линейных Алексеев Е.Р., Чеснокова О.В. Введение в Octave { n n 4 n 2 i a1 n+a 2 ∑ x i+a3 ∑ x =∑ y i i =1 i =1 i =1 n n n n i=1 n i=1 n i =1 n i=1 n i=1 i=1 i=1 i=1 a1 ∑ x i +a 2 ∑ xi2+a 3 ∑ x 3i =∑ y i x i . (11.8) a1 ∑ x 2i +a 2 ∑ xi3+a 3 ∑ x i4=∑ y i x 2i Решив систему (11.8), найдём значения параметров a 1 , a 2 Аналогично определим параметры многочлена 2 3 Y =a1 +a 2 x +a 3 x +a 4 x . Составим функцию S ( a1 , a 2 , a 3 , a 4) : n и a3 . третьей 2 S (a1 , a 2 , a 3 , a 4)=∑ [ y i−a1−a2 x i −a3 x 2i −a 4 x 3i ] → min . степени: (11.9) i=1 После дифференцирования S по a 1 , a 2 , a 3 и a 4 система линейных алгебраических уравнений для вычисления параметров a 1 , a 2 , a 3 , a 4 примет вид: { n n n n a1 n+a 2 ∑ x i+a3 ∑ x +a4 ∑ x =∑ yi 2 i i =1 3 i i =1 n n i=1 n i=1 n i=1 n i =1 n n a1 ∑ x i+a 2 ∑ x +a 3 ∑ x +a 4 ∑ x =∑ y i x i 2 i i=1 n 3 i 4 i i =1 n i =1 n . a1 ∑ x +a 2 ∑ x +a3 ∑ x +a 4 ∑ x =∑ y i x 2 i i=1 n 3 i i=1 n i=1 n 4 i 5 i i =1 n i=1 n (11.10) 2 i a1 ∑ x 3i +a2 ∑ x 4i +a 3 ∑ x 5i +a 4 ∑ x6i =∑ y i x 3i i=1 i=1 i=1 i=1 i=1 Решив систему (11.10), найдём коэффициенты a 1 , a 2 , a 3 и a 4 . В общем случае система уравнений для вычисления параметров ai многочлена k-й k+1 i−1 степени Y =∑ ai x имеет вид: i=1 { n n n 2 i n k i a1 n+a 2 ∑ x i+a3 ∑ x +…+a k+1 ∑ x =∑ y i i =1 i =1 i =1 i=1 n n n n n i=1 i=1 i=1 i=1 i=1 a1 ∑ x i+a 2 ∑ xi2+a 3 ∑ x3i +…+a k +1 ∑ x ik+1=∑ y i x i . (11.11) ... n n a1 ∑ x k +1 i i=1 +a 2 ∑ x i =1 n k+2 i +a 3 ∑ x n k+2 i i=1 n 2k i k …+a k+1 ∑ x =∑ y i x i i=1 i=1 В матричном виде систему (11.11) можно записать Ca=g , Элементы матрицы C и вектора g рассчитываются по формулам (11.12) n C i , j =∑ x ik+ j−2 , i=1,... , k +1, j=1, ... , k +1 , (11.13) k=1 n g i =∑ y k x i−1 k , i=1, ... , k +1 (11.14) k=1 Решив систему (11.12), 2 k Y =a1 +a 2 x +a3 x +...+a k+1 x . определим параметры зависимости Алексеев Е.Р., Чеснокова О.В. Введение в Octave 11.2.3 5 Подбор коэффициентов функции Y =ax b e cx Параметры b и c входят в зависимость Y =ax b e cx нелинейным образом. Чтобы избавиться от нелинейности предварительно прологарифмируем 1 выражение Y =ax b e cx . ln Y =ln a+b ln x+cx Сделаем замену Y1=ln Y , A=ln a : Y1= A+b ln x+cx . Составим функцию S(A, b, c) по формуле (11.1): n 2 S ( A , b , c)=∑ [ Y1i− A−b ln x i −c x i ] → min (11.15) i =1 После дифференцирования получим систему трёх линейных алгебраических уравнений для определения коэффициентов A, b и c. { n n n n A+b ∑ ln xi +c ∑ x i=∑ Y1i i=1 i=1 i =1 n n n n i=1 n i=1 i =1 i =1 n n n i=1 i=1 i =1 i=1 A ∑ ln x i+b ∑ (ln x i ) +b ∑ x i ln x i=∑ Y1i ln x i . 2 (11.16) A ∑ x i+b ∑ x i ln x i+c ∑ x 2i =∑ Y1i x i После решения системы (11.6) необходимо вычислить значение коэффициента а по формуле a=е A . 11.2.4 Функции , приводим ые к линейной Для вычисления параметров функции Y =ax b необходимо предварительно ее прологарифмировать ln Y =ln ax b =ln a+b ln x . После чего замена Z =ln Y , X =ln x , A=ln a приводит заданную функцию к линейному виду Z =bX + A , где коэффициенты A и b вычисляются по формулам (11.6) и, соответственно, a=e A . Аналогично можно подобрать параметры функции вида Y =ae bx . Прологарифмируем ln y=ln a+bx ln e , ln y=ln a+bx . Проведём замену Y =ln y , заданную функцию A=ln a и получим линейную зависимость Y =bx+ A . По формулам (11.6) найдем A и b, а затем вычислим a=e A . Рассмотрим ещё ряд зависимостей, которые сводятся к линейной. 1 1 Для подбора параметров функции Y = сделаем замену Z = . В результате ax+b Y x Y= Z =ax+b . Функция получим линейную зависимость заменами ax+b 1 1 Z= , X = Z =a+bX . Для определения коэффициентов сводится к линейной Y x 1 функциональной зависимости Y = −x необходимо сделать следующие замены ae +b 1 Z = , X =e−x . В результате также получим линейную функцию Z =aX +b . Y Аналогичными приемами (логарифмированием, заменами и т. п.) можно многие подбираемые зависимости преобразовать к такому виду, что получаемая при решении задачи оптимизации система (11.2) была системой линейных алгебраических уравнений. При использовании Octave можно напрямую решать задачу подбора параметров, как задачу 1 Можно и не проводить предварительное логарифмирование выражения Y =ax b e cx , однако в этом случаем получаемая система уравнений будет нелинейной, которую решать сложнее. Алексеев Е.Р., Чеснокова О.В. Введение в Octave 6 оптимизации (11.1) с использованием функции sqp. f ( x , a1 , a 2 ,… a k ) возникает вопрос После нахождения параметров зависимости насколько адекватно описывает подобранная зависимость экспериментальные данные. Чем ближе величина n 2 S=∑ [ y i− f (x i , a1, a2, ... , a k )] , (11.17) i =1 называемая суммарной квадратичной ошибкой, к нулю, тем точнее подобранная кривая описывает экспериментальные данные. 11.3 Уравнение регрессии и коэффициен т корреляции Линия, описываемая уравнением вида y=a1+a2 x , называется линией регрессии y на x, параметры a 1 и a 2 называются коэффициентами регрессии и определяются формулами (11.6). n 2 Чем меньше величина S=∑ [ y i−a 1−a 2 x i ] , тем более обоснованно предположение, i =1 что экспериментальные данные описываются линейной функцией. Существует показатель, характеризующий тесноту линейной связи между x и y, который называется коэффициентом корреляции и рассчитывается по формуле: n ∑ ( x i− M x )( y i−M y ) r= i =1 √∑ n i =1 2 n n ∑ xi , M x = i=1 n 2 ( x i− M x ) ∑ ( y i − M y ) n ∑ yi , M y = i=1 n (11.18) i=1 Значение коэффициента корреляции удовлетворяет соотношению – 1≤r ≤1 . Чем меньше отличается абсолютная величина r от единицы, тем ближе к линии ∣r∣=1 , то все регрессии располагаются экспериментальные точки. Если экспериментальные точки находятся на линии регрессии. Если коэффициент корреляции близок к нулю, то это означает, что между x и y не существует линейной связи, но между ними может существовать зависимость, отличная от линейной. Для того, чтобы проверить, значимо ли отличается от нуля коэффициент корреляции, можно использовать критерий Стьюдента. Вычисленное значение критерия определяется по формуле: n−2 t=r (11.19) 1−r 2 Рассчитанное по формуле (11.19) значение t сравнивается со значением, взятым из таблицы распределения Стьюдента (см. табл. 11.2) в соответствии с уровнем значимости p (стандартное значение p=0.95 ) и числом степеней свободы k =n – 2 . Если полученная по формуле (3.10) величина t больше табличного значения, то коэффициент корреляции значимо отличен от нуля. √ Таблица 11.2: p 0,99 0,98 0,95 0,90 0,80 0,70 0,60 1 63,657 31,821 12,706 6,314 3,078 1,963 1,376 2 9,925 6,965 4,303 2,920 1,886 1,386 1,061 3 5,841 4,541 3,182 2,353 1,638 1,250 0,978 4 4,604 3,747 2,776 2,132 1,533 1,190 0,941 k Алексеев Е.Р., Чеснокова О.В. Введение в Octave 7 5 4,032 3,365 2,571 2,05 1,476 1,156 0,920 6 3,707 3,141 2,447 1,943 1,440 1,134 0,906 7 3,499 2,998 2,365 1,895 1,415 1,119 0,896 8 3,355 2,896 2,306 1,860 1,387 1,108 0,889 9 3,250 2,821 2,261 1,833 1,383 1,100 0,883 10 3,169 2,764 2,228 1,812 1,372 1,093 0,879 11 3,106 2,718 2,201 1,796 1,363 1,088 0,876 12 3,055 2,681 2,179 1,782 1,356 1,083 0,873 13 3,012 2,650 2,160 1,771 1,350 1,079 0,870 14 2,977 2,624 2,145 1,761 1,345 1,076 0,868 15 2,947 2,602 2,131 1,753 1,341 1,074 0,866 16 2,921 2,583 2,120 1,746 1,337 1,071 0,865 17 2,898 2,567 2,110 1,740 1,333 1,069 0,863 18 2,878 2,552 2,101 1,734 1,330 1,067 0,862 19 2,861 2,539 2,093 1,729 1,328 1,066 0,861 20 2,845 2,528 2,086 1,725 1,325 1,064 0,860 21 2,831 2,518 2,080 1,721 1,323 1,063 0,859 22 2,819 2,508 2,074 1,717 1,321 1,061 0,858 23 2,807 2,500 2,069 1,714 1,319 1,060 0,858 24 2,797 2,492 2,064 1,711 1,318 1,059 0,857 25 2,779 2,485 2,060 1,708 1,316 1,058 0,856 26 2,771 2,479 2,056 1,706 1,315 1,058 0,856 27 2,763 2,473 2,052 1,703 1,314 1,057 0,855 28 2,756 2,467 2,048 1,701 1,313 1,056 0,855 29 2,750 2,462 2,045 1,699 1,311 1,055 0,854 30 2,704 2,457 2,042 1,697 1,310 1,055 0,854 40 2,660 2,423 2,021 1,684 1,303 1,050 0,851 60 2,612 2,390 2,000 1,671 1,296 1,046 0,848 120 2,617 2,358 1,980 1,658 1,289 1,041 0,845 ∞ 2,576 2,326 1,960 1,645 1,282 1,036 0,842 11.4 Нелинейная корреляция Коэффициент корреляции r применяется только в тех случаях, когда между данными существует прямолинейная связь. Если же связь нелиненая, то для выявления тесноты связи между переменными y и x в случае нелинейной зависимости пользуются индексом корреляции. Он показывает тесноту связи между фактором x и зависимой переменной y и рассчитывается по формуле: Алексеев Е.Р., Чеснокова О.В. Введение в Octave √ 8 n ∑ ( y i−Y i )2 R= 1− i =1 n , (11.20) 2 ∑ ( y i−M y ) i =1 где y – экспериментальные значения, Y – теоретические значения (рассчитанные по подобранной методом наименьших квадратов формуле), M y – среднее значение y . Индекс корреляции лежит в пределах от 0 до 1. При наличии функциональной зависимости индекс корреляции близок к 1. При отсутствии связи R практически равен нулю. Если коэффициент корреляции r является мерой тесноты связи только для линейной формы связи, то индекс корреляции R – как для линейной, так и для нелинейной. При прямолинейной связи коэффициент корреляции по своей абсолютной величине равен индексу корреляции: ∣r∣= R . 11.5 Использование Octave ме т одом наименьших квадра т о в 11.5.1 Функции зависимости МНК Octave, для подбора используемые зависимос т ей для подбора Для решения задач подбора аналитических зависимостей по экспериментальным данным можно использовать следующие функции Octave: polyfit(x,y,k) – функция подбора коэффициентов полинома k-й степени методом наименьших квадратов (x – массив абсцисс экспериментальных точек, y – массив ординат экспериментальных точек, k – степень полинома), функция возвращает массив коэффициентов полинома; sqp(x0,phi,g,h,lb,ub,maxiter,tolerance) – функция поиска минимума (функция подробно описана в десятой главе); cor(x,y) – функция вычисления коэффициента корреляции (x – массив абсцисс экспериментальных точек, y – массив ординат экспериментальных точек); mean(x) – функция вычисления среднего арифметического. 11.5.2 Примеры решения задач ПРИМЕР 11.1. В «Основах химии» Д.И. Менделеева приводятся данные о растворимости азотнокислого натрия NaNO3 в зависимости от температуры воды. В 100 частях воды (табл. 11.3) растворяется следующее число условных частей NaNO3 при соответствующих температурах. Требуется определить растворимость азотнокислого натрия при температуре t=32°С в случае линейной зависимости и найти коэффициент корреляции. Таблица 11.3: t P 0° 4° 10° 15° 21° 29° 36° 51° 66,7 71,0 76,3 80,6 85,7 92,9 99,4 113,6 Решение задачи 11.1 с комментариями в приведено на листинге 11.1. %Ввод экспериментальных данных 68° 125,1 Алексеев Е.Р., Чеснокова О.В. Введение в Octave 9 X=[0 4 10 15 21 29 36 51 68]; Y=[66.7 71.0 76.3 80.6 85.7 92.9 99.4 113.6 125.1]; %Вычисление вектора коэффициентов полинома y=a1*x+a2 [a]=polyfit(X,Y,1) %Вычисление значения полинома y=a1*x+a2 в точке t=32 t=32; yt=a(1)*t+a(2) %Построение графика полинома y=a1*x+a2, %экспериментальных точек и значения в заданной точке %в одной графической области x=0:68; y=a(1)*x+a(2); plot(X,Y,'ok',x,y,'-k',t,yt,'xk') grid % Вычисление коэффициента корреляции k =cor(x,y) Листинг 11.1 Результаты работы программы приведены ниже >>>a = 0.87064 67.50779 >>>yt = 95.368 >>>k = 1 На рис. 11.2 приведено графическое решение этой задачи, изображены экспериментальные точки, линия регрессии y=a1 x+a 2 , на котором отмечена точка t=32. ПРИМЕР 11.2. В результате эксперимента получена табличная зависимость y(x) (см. табл. 11.4). Подобрать аналитическую зависимость Y =ax b e cx методом наименьших квадратов. Вычислить ожидаемое значение в точках 2, 3, 4. Вычислить индекс корреляции. Таблица 11.4: x -2 -1,3 -0,6 0,1 0,8 1,5 2,2 2,9 3,6 4,3 5 5,7 6,4 y -10 -5 0 0,7 0,8 2 3 5 8 30 60 100 238 b cx Решение задачи подбора параметров функции f (x )=ax e в Octave возможно двумя способами: 1. Решение задачи путём поиска минимума функции (11.15)2. После чего надо пересчитать значение коэффициента а по формуле a=е A . 2. Формирование системы линейных алгебраических уравнений (11.16)3 и её решение. 2 Может не получится решать задачу подбора зависимости «в лоб» путём оптимизации функции n 2 S ( a ,b , c )=∑ [ y i −ax bi e cx ] , это связано с тем, что при решении задачи оптимизации с помощью i i=1 sqp итерационными методами может возникнуть проблема возникнуть проблема возведения отрицательного числа в дробную степень (см. главу 2). Да и с точки зрения математики, если есть возможность решать линейную задачу вместо нелинейной, то лучше решать линейную. 3 Следует помнить, что при отрицательных значениях y необходимо будет решать проблему замены Y =ln y . Алексеев Е.Р., Чеснокова О.В. Введение в Octave 10 Рисунок 11.2 Рассмотрим последовательно оба варианта решения задачи. Способ 1. Функция (11.15) реализована в Octave с помощью функции f_mnk. Полный текст программы решения задачи способом 1 с комментариями приведён на листинге 11.2. Вместо A , b , c коэффициентов из формул (11.15)-(11.16) в программе на Octave используется массив c. function s=f_mnk(c) %Переменные x,y являются глобальными, % используются в нескольких функциях global x; global y; s=0; for i=1:length(x) s=s+(log(y(i))-c(1)-c(2)*log(x(i)) -c(3)*x(i))^2; end end %------------------------------------------------global x; global y; %Задание начального значения вектора c, при неправильном его % определении, экстремум может быть найден неправильно. c=[2;1;3]; Алексеев Е.Р., Чеснокова О.В. Введение в Octave 11 %Определение координат экспериментальных точек x=[1 1.4 1.8 2.2 2.6 3 3.4 3.8 4.2 4.6 5 5.4 5.8]; y=[0.7 0.75 0.67 0.62 0.51 0.45 0.4 0.32 0.28 0.25 0.22 0.16 0.1]; Решение задачи оптимизации функции 11.15 с помощью sqp. c=sqp(c,@f_mnk) % Вычисление суммарной квадратичной ошибки для подобранной %зависимости и вывод её на экран. sum1=f_mnk(c) %Формирование точек для построения графика подобранной кривой. x1=1:0.1:6; y1=exp(c(1)).*x1.^c(2).*exp(c(3).*x1); %Вычисление значений на подобранной кривой в заданных точках. yr=exp(c(1)).*x.^c(2).*exp(c(3).*x); %Вычисление ожидаемого значения подобранной функции в точках %x=[2,3,4]; x2=[2 3 4] y2=exp(c(1)).*x2.^c(2).*exp(c(3).*x2) %Построение графика: подобранная кривая, f(x2) % и экспериментальные точки. plot(x1,y1,'-r',x,y,'*b',x2,y2,'pk'); %Вычисление индекса корреляции. R=sqrt(1-sum((y-yr).^2)/sum((y-mean(y)).^2)) Листинг 11.2 Результаты программы приведены ниже. >>>c = 0.33503 0.90183 -0.69337 >>>sum1=0.090533 >>>x2= 2 3 4 >>>y2= 0.65272 0.47033 >>>R=0.99533 0.30475 0.90183 −0.9337x Таким образом подобрана зависимость . Вычислено Y =0.33503x e ожидаемое значение в точках 2, 3, 4: Y (2)=0.65272 ,Y (3)=0.47033 ,Y ( 4)=0.30475 . График подобранной зависимости вместе с экспериментальными точками и расчётными значениями изображён на рис. 11.3. Индекс корреляции равен 0.99533. Способ 2. Теперь рассмотрим решение задачи 11.2 путём решения системы 11.16. Решение с комментариями приведено на листинге 11.3. Результаты и графики при решении обоими способами полностью совпадают. function s=f_mnk(c) %Переменные x,y являются глобальными, % используются в нескольких функциях global x; global y; s=0; for i=1:length(x) s=s+(log(y(i))-c(1)-c(2)*log(x(i)) -c(3)*x(i))^2; end end Алексеев Е.Р., Чеснокова О.В. Введение в Octave 12 Рисунок 11.3 График к примеру 11.2: экспериментальные точки и подобранная методом наименьших квадра тов зависимос ть global x; global y; %Определение координат экспериментальных точек x=[1 1.4 1.8 2.2 2.6 3 3.4 3.8 4.2 4.6 5 5.4 5.8]; y=[0.7 0.75 0.67 0.62 0.51 0.45 0.4 0.32 0.28 0.25 0.22 0.16 0.1]; %Формирование СЛАУ (11.16) G=[length(x) sum(log(x)) sum(x);... sum(log(x)) sum(log(x).*log(x)) sum(x.*log(x));... sum(x) sum(x.*log(x)) sum(x.*x)]; H=[sum(log(y)); sum(log(y).*log(x)); sum(log(y).*x)]; %Решение СЛАУ методом Гаусса с помощью функции rref. C=rref([G H]); n=size(C); c=C(:,n(2)) % Вычисление суммарной квадратичной ошибки для подобранной %зависимости и вывод её на экран. sum1=f_mnk(c) %Формирование точек для построения графика подобранной кривой. x1=1:0.1:6; y1=exp(c(1)).*x1.^c(2).*exp(c(3).*x1); %Вычисление значений на подобранной кривой в заданных точках. Алексеев Е.Р., Чеснокова О.В. Введение в Octave 13 yr=exp(c(1)).*x.^c(2).*exp(c(3).*x); %Вычисление ожидаемого значения подобранной функции в точках %x=[2,3,4]; x2=[2 3 4] y2=exp(c(1)).*x2.^c(2).*exp(c(3).*x2) %Построение графика: подобранная кривая, f(x2) % и экспериментальные точки. plot(x1,y1,'-r',x,y,'*b',x2,y2,'pk'); %Вычисление индекса корреляции. R=sqrt(1-sum((y-yr).^2)/sum((y-mean(y)).^2)) Листинг 11.3 ПРИМЕР 11.3. В результате эксперимента получена табличная зависимость y(x) (см. табл. 11.5). 2 3 4 5 f (x )=b 1+b 2 x+b 3 x +b 4 x +b 5 x +b 6 x , Подобрать аналитические зависимости g ( x)=a1+a2 x+a3 x 2+a4 x 3+a5 x 5 и φ ( x)=c1 +c 2 x+c 4 x 3+c5 x 5 методом наименьших квадратов. Пользуясь значением индекса корреляции выбрать наилучшую из них, с помощью которой вычислить ожидаемое значение в точках 1, 2.5, 4.8. Построить графики экспериментальных точек, подобранных зависимостей. На графиках отобразить рассчитанные значения в точках 1, 2.5, 4.8. Таблица 11.5: x -2 -1,3 -0,6 0,1 0,8 1,5 2,2 2,9 3,6 4,3 5 5,7 6,4 y -10 -5 0 0,7 0,8 2 3 5 8 30 60 100 238 Как рассматривалось ранее решать задачу подбора параметров полинома методом наименьших квадратов, в Octave можно тремя способами. 1. Сформировать и решить систему уравнений (11.3). k+1 2. Решить задачу оптимизации (11.1). В случае полинома f (x )=∑ ai x i −1 i =1 подбираемые коэффициенты a i будут входить в функцию (11.1) линейным образом и не должно возникнуть проблем при решении задачи оптимизации с помощью функции sqp. 3. Использовать функцию polyfit. Чтобы продемонстрировать использование всех трех методов для подбора 6 f ( x )=∑ bi x i−1 воспользуемся функцией polyfit, для формирования коэффициентов i=1 функции g ( x) сформируем и решим систему уравнений (11.3), а функцию φ ( x) будем искать с помощью функции sqp. Для формирования подбора коэффициентов функции 2 3 5 сформируем систему уравнений. Составим функцию g ( x)=a1+a2 x+a3 x +a4 x +a5 x n 2 S (a1 , a 2 , a 3 , a 4 , a 5)=∑ [ y i−a1−a2 x i−a3 x 2i −a4 x 3i −a5 x 5i ] . i=1 После дифференцирования S по a 1 , a 2 , a 3 , a 4 и a 5 система линейных алгебраических уравнений для вычисления параметров a 1 , a 2 , a 3 , a 4 , a 5 примет вид: Алексеев Е.Р., Чеснокова О.В. Введение в Octave { n n n 14 n n a1 n+a 2 ∑ x i+a3 ∑ x +a4 ∑ x +a5 ∑ x =∑ y i i =1 n i =1 n 2 i i=1 n 3 i i=1 n 5 i i=1 n n a1 ∑ x i+a 2 ∑ x +a 3 ∑ x +a 4 ∑ x +a5 ∑ x =∑ y i x i 2 i i=1 n 3 i i=1 n i=1 n 4 i i =1 n 6 i i =1 n i=1 n a1 ∑ x 2i +a 2 ∑ x3i +a3 ∑ x 4i +a 4 ∑ x 5i +a 5 ∑ x7i =∑ y i x 2i . i=1 n i=1 n i=1 n i =1 n i=1 n i=1 n i=1 n i=1 n i=1 n i=1 n i=1 n i =1 n i=1 i=1 i=1 i=1 i=1 (11.21) a1 ∑ x 3i +a2 ∑ x 4i +a 3 ∑ x 5i +a 4 ∑ x6i +a5 ∑ x 8i =∑ yi x 3i 5 a1 ∑ x 5i +a2 ∑ x 6i +a3 ∑ x 7i +a4 ∑ x 8i +a5 ∑ x 10 i =∑ y i x i i=1 Решив систему (11.21), найдём коэффициенты a 1 , a 2 2 3 , a 3 , a 4 и a 5 функции 5 g ( x)=a1+a2 x+a3 x +a4 x +a5 x . Для поиска функциональной зависимости вида φ ( x)=c1 +c 2 x+c 4 x 3+c5 x 5 необходимо будет найти такие значения c 1 , c 2 , c3 , c 4 , при которых функция n 2 S (c1 , c2 , c 3 , c 4)=∑ [ y i−c 1−c 2 x i−c 3 x 3i −c 4 x 5i ] (11.22) i=1 принимала наименьшее значение. После вывода необходимых формул приступим к реализации в Octave. Текст программы в Octave очень подробными комментариями приведён на листинге 11.4. % Функция для подбора зависимости fi(x) методом наименьших % квадратов. function s=f_mnk(c) %Переменные x,y являются глобальными, используются в % функции f_mnk и главной функции. global x; global y; %Формирование суммы квадратов отклонений 11.22. s=0; for i=1:length(x) s=s+(y(i)-c(1)-c(2)*x(i) -c(3)*x(i)^3 - c(4)*x(i)^5 )^2; end end %-Главная функция------------------------------------%Переменные x,y являются глобальными, используются в % функции f_mnk и главной функции. global x; global y; %Определение координат экспериментальных точек x=[-2 -1.3 -0.6 0.1 0.8 1.5 2.2 2.9 3.6 4.3 5 5.7 6.4]; y=[-10 -5 0 0.7 0.8 2 3 5 8 30 60 100 238]; z=[1 2.5 4.8] %Подбор коэффициентов зависимости f(x) (полинома пятой %степени) методом наименьших квадратов, используя функцию %polyfit. Коэффициенты полинома будут хранится в переменной B. B=polyfit(x,y,5) %Формирование точек для построения графиков подобранных Алексеев Е.Р., Чеснокова О.В. Введение в Octave 15 % функций. X1=-2:0.1:6.5; %Вычисление ординат точек графика первой функции f(x). Y1=polyval(B,X1); %Формирование системы (11.21) для подбора функции g(x). %Здесь GGL — матрица коэффициентов, H – вектор правых частей % системы (11.21). матрица G — первые 4 строки и 4 столбца %матрицы коэффициентов, G1 – пятый столбец матрицы %коэффициентов, G2 — пятая строка матрицы коэффициентов. for i = 1:4 for j=1:4 G(i,j)=sum(x.^(i+j-2)); endfor endfor for i = 1:4 G1(i)=sum(x.^(i+5)); H(i)=sum(y.*x.^(i-1)); endfor for i=1:4 G2(i)=sum(x.^(i+4)); endfor G2(5)=sum(x.^10); %Формирование матрицы коэффициентов системы (11.21) из матриц % G, G1 и G2. GGL=[G G1'; G2] H(5)=sum(y.*x.^5); %Решение системы (11.21) методом обратной матрицы и % формирование коэффициентов А функции g(x). A=inv(GGL)*H' %Подбор коэффициентов зависимости fi(x) методом наименьших %квадратов, используя функцию %sqp. Коэффициенты функции будут %хранится в переменной C. %Задание начального значения вектора С, при неправильном его %определении, экстремум функции может быть найден неправильно. C=[2;1;3;1]; %Поиск вектора С, при котором функция (11.22) достигает своего %минимального значения с помощью функции sqp, %вектор С — коэффиценты функции fi. C=sqp(C,@f_mnk) %Вычисление ординат точек графика второй функции g(x). Y2=A(1)+A(2)*X1+A(3)*X1.^2+A(4)*X1.^3+A(5)*X1.^5; %Вычисление ординат точек графика третьей функции fi(x). Y3=C(1)+C(2)*X1+C(3)*X1.^3 + C(4)*X1.^5; %Вычисление значений на подобранной на первой функции f(x) %в заданных точках. yr1=polyval(B,x); %Вычисление значений на подобранной на второй функции g(x) %в заданных точках. yr2=A(1)+A(2)*x+A(3)*x.^2+A(4)*x.^3+A(5)*x.^5; %Вычисление значений на подобранной на второй функции fi(x) %в заданных точках. Алексеев Е.Р., Чеснокова О.В. Введение в Octave 16 yr3=C(1)+C(2)*x+C(3)*x.^3 + C(4)*x.^5; %Вычисление индекса корреляции для первой функции f(x). R1=sqrt(1-sum((y-yr1).^2)/sum((y-mean(y)).^2)) %Вычисление индекса корреляции для второй функции g(x). R2=sqrt(1-sum((y-yr2).^2)/sum((y-mean(y)).^2)) %Вычисление индекса корреляции для третьей функции fi(x). R3=sqrt(1-sum((y-yr3).^2)/sum((y-mean(y)).^2)) %Сравнивая значения трёх индексов корреляции, выбираем %наилучшую функцию и с её помощью вычисляем ожидаемое значение %в точнах 1, 2.5, 4.8. if R1>R2 & R1>R3 yz=polyval(B,z) "R1=" R1 endif if R2>R1 & R2>R3 yz=C2(1)+C2(2)*z+C2(3)*z.^2+C2(4)*z.^3+C2(5)*z.^5 "R2=" R2 endif if R3>R1 & R3>R2 yz=C(1)+C(2)*z+C(3)*z.^3 + C(4)*z.^5 "R3=" R3 endif %Построение графика. plot(x,y,"*r;esperiment;",X1,Y1,'-b;f(x);',... X1,Y2,'dr;g(x);',X1,Y3,'ok;fi(x);',z,yz,'sb;f(z);'); grid(); Листинг 11.4 Ниже приведены результаты работы программы. >>>z = 1.0000 2.5000 >>>B = 0.083039 -0.567892 0.906779 GGL = 1.3000e+01 2.8600e+01 1.5210e+02 2.8600e+01 1.5210e+02 7.2701e+02 1.5210e+02 7.2701e+02 3.9868e+03 7.2701e+02 3.9868e+03 2.2183e+04 2.2183e+04 1.2793e+05 7.5030e+05 >>>A = 9.4262e+00 -3.6516e+00 -5.7767e+00 1.7888e+00 -5.8179e-05 >>>C = -1.030345 5.080391 -0.609721 0.033534 >>>R1=0.99690 >>>R2 =0.98136 >>>R3 =0.99573 4.8000 1.609432 -1.115925 7.2701e+02 3.9868e+03 2.2183e+04 1.2793e+05 4.4706e+06 1.2793e+05 7.5030e+05 4.4706e+06 2.6938e+07 1.6383e+08 -1.355075 Алексеев Е.Р., Чеснокова О.В. Введение в Octave >>>yz = -0.43964 ans=R1= R1=0.99690 6.00854 17 40.77972 На рисунке 11.4 представлено графическое решение задачи. Рисунок 11.4 Рассмотренная задача демонстрирует основные приёмы подбора зависимости методом наименьших квадратов в Octave. Авторы рекомендует внимательно рассмотреть её для понимания методов решения подобных задач в Octave. В заключении авторы позволят несколько советов по решению задачи аппроксимации. 1. Подбор каждой зависимости по экспериментальным данным – довольно сложная математическая задача, поэтому следует аккуратно выбирать вид зависимости, наиболее точно описывающей экспериментальные точки. 2. Необходимо сформировать реальную систему уравнений исходя из соотношений (11.1) – (11.3). Следует помнить, что проще и точнее решать систему линейных алгебраических уравнений, чем систему нелинейных уравнений. Поэтому, может быть, следует преобразовать исходную функцию (прологарифмировать, сделать замену и т. д.) и только после этого составлять систему уравнений. 3. При том, что функция sqp – довольно мощная, лучше использовать методы и функции решения систем линейных алгебраических уравнений, функцию polyfit, чем функцию sqp. Этот совет связан с тем, что функция sqp – приближённые итерационные алгоритмы, поэтому получаемый результат иногда может быть менее точен, чем при точных методах решения систем линейных алгебраических уравнений. Но, иногда, именно функция sqp – единственный Алексеев Е.Р., Чеснокова О.В. Введение в Octave 18 метод решения задачи. 4. Для оценки корректности подобранной зависимости следует использовать коэффициент корреляции, критерий Стьюдента (для линейной зависимости) и индекс корреляции и суммарную квадратичную ошибку (для нелинейных зависимостей). Документ Категория Физико-математические науки Просмотров 331 Размер файла 284 Кб Теги 1/--страниц Пожаловаться на содержимое документа
__label__pos
0.662173
   Access 2K splash screen  Author Message  Access 2K splash screen I have made my own splash screen, and access uses it, but i flashes by so quickly that I can't really see it. How can I tell Access to slow down and display it for a few seconds? Sun, 15 Sep 2002 03:00:00 GMT    Access 2K splash screen Quote: > I have made my own splash screen, and access uses it, but i flashes by so > quickly that I can't really see it. How can I tell Access to slow down and > display it for a few seconds? I use a form as a splash screen with an On Timer event which closes the form and launches the switchboard.  Setting the timer interval to 5000 (5 seconds) seems about right. Specifying this form in the startup options will effectively use it like a splash screen. Incidentally, please try not to cross-post to unrelated NGs as many people get very annoyed when they have to trawl through messages which are not related to the NG they are in. Regards, -- Roger E K Stout Programmer Integration & Support EASAMS Ltd Tel 01785 782339 Fax 01785 244397 Mon, 16 Sep 2002 03:00:00 GMT    Access 2K splash screen I am just taking a stab at this, but is the splash part of the web page?  If so, couldn't you add more frames to the animation? I am no great VB programmer, but you could create a Timer/Counter event that looped, with every loop the splash screen would be displayed.  Once the conditions of the loop are met, screen disappears.  I am approaching this from a newbie perspective and my programming experience is from high school . Quote: > I have made my own splash screen, and access uses it, but i flashes by so > quickly that I can't really see it. How can I tell Access to slow down and > display it for a few seconds? Tue, 19 Nov 2002 03:00:00 GMT      [ 3 post ]   Relevant Pages  1. Supress Access Splash Screen in Access 2000 2. Splash Screen Delay] Note on splashes 3. Splash screens in Access '97 4. Cannot access disposed object named (Splash Screen related problem) 5. Access 2K with Win 2K 6. Save and close excel 2k from access 2k 7. Merge access 2k with word 2k 8. Splash screen hints? 9. Splash screens 10. Product registration info in splash screen 11. Suppress default splash screen 12. Splash screen A97     Powered by phpBB® Forum Software
__label__pos
0.987821
qemu/block/block-copy.c << >> Prefs 1/* 2 * block_copy API 3 * 4 * Copyright (C) 2013 Proxmox Server Solutions 5 * Copyright (c) 2019 Virtuozzo International GmbH. 6 * 7 * Authors: 8 * Dietmar Maurer ([email protected]) 9 * Vladimir Sementsov-Ogievskiy <[email protected]> 10 * 11 * This work is licensed under the terms of the GNU GPL, version 2 or later. 12 * See the COPYING file in the top-level directory. 13 */ 14 15#include "qemu/osdep.h" 16 17#include "trace.h" 18#include "qapi/error.h" 19#include "block/block-copy.h" 20#include "sysemu/block-backend.h" 21#include "qemu/units.h" 22#include "qemu/coroutine.h" 23#include "block/aio_task.h" 24#include "qemu/error-report.h" 25 26#define BLOCK_COPY_MAX_COPY_RANGE (16 * MiB) 27#define BLOCK_COPY_MAX_BUFFER (1 * MiB) 28#define BLOCK_COPY_MAX_MEM (128 * MiB) 29#define BLOCK_COPY_MAX_WORKERS 64 30#define BLOCK_COPY_SLICE_TIME 100000000ULL /* ns */ 31#define BLOCK_COPY_CLUSTER_SIZE_DEFAULT (1 << 16) 32 33typedef enum { 34 COPY_READ_WRITE_CLUSTER, 35 COPY_READ_WRITE, 36 COPY_WRITE_ZEROES, 37 COPY_RANGE_SMALL, 38 COPY_RANGE_FULL 39} BlockCopyMethod; 40 41static coroutine_fn int block_copy_task_entry(AioTask *task); 42 43typedef struct BlockCopyCallState { 44 /* Fields initialized in block_copy_async() and never changed. */ 45 BlockCopyState *s; 46 int64_t offset; 47 int64_t bytes; 48 int max_workers; 49 int64_t max_chunk; 50 bool ignore_ratelimit; 51 BlockCopyAsyncCallbackFunc cb; 52 void *cb_opaque; 53 /* Coroutine where async block-copy is running */ 54 Coroutine *co; 55 56 /* Fields whose state changes throughout the execution */ 57 bool finished; /* atomic */ 58 QemuCoSleep sleep; /* TODO: protect API with a lock */ 59 bool cancelled; /* atomic */ 60 /* To reference all call states from BlockCopyState */ 61 QLIST_ENTRY(BlockCopyCallState) list; 62 63 /* 64 * Fields that report information about return values and erros. 65 * Protected by lock in BlockCopyState. 66 */ 67 bool error_is_read; 68 /* 69 * @ret is set concurrently by tasks under mutex. Only set once by first 70 * failed task (and untouched if no task failed). 71 * After finishing (call_state->finished is true), it is not modified 72 * anymore and may be safely read without mutex. 73 */ 74 int ret; 75} BlockCopyCallState; 76 77typedef struct BlockCopyTask { 78 AioTask task; 79 80 /* 81 * Fields initialized in block_copy_task_create() 82 * and never changed. 83 */ 84 BlockCopyState *s; 85 BlockCopyCallState *call_state; 86 int64_t offset; 87 /* 88 * @method can also be set again in the while loop of 89 * block_copy_dirty_clusters(), but it is never accessed concurrently 90 * because the only other function that reads it is 91 * block_copy_task_entry() and it is invoked afterwards in the same 92 * iteration. 93 */ 94 BlockCopyMethod method; 95 96 /* 97 * Fields whose state changes throughout the execution 98 * Protected by lock in BlockCopyState. 99 */ 100 CoQueue wait_queue; /* coroutines blocked on this task */ 101 /* 102 * Only protect the case of parallel read while updating @bytes 103 * value in block_copy_task_shrink(). 104 */ 105 int64_t bytes; 106 QLIST_ENTRY(BlockCopyTask) list; 107} BlockCopyTask; 108 109static int64_t task_end(BlockCopyTask *task) 110{ 111 return task->offset + task->bytes; 112} 113 114typedef struct BlockCopyState { 115 /* 116 * BdrvChild objects are not owned or managed by block-copy. They are 117 * provided by block-copy user and user is responsible for appropriate 118 * permissions on these children. 119 */ 120 BdrvChild *source; 121 BdrvChild *target; 122 123 /* 124 * Fields initialized in block_copy_state_new() 125 * and never changed. 126 */ 127 int64_t cluster_size; 128 int64_t max_transfer; 129 uint64_t len; 130 BdrvRequestFlags write_flags; 131 132 /* 133 * Fields whose state changes throughout the execution 134 * Protected by lock. 135 */ 136 CoMutex lock; 137 int64_t in_flight_bytes; 138 BlockCopyMethod method; 139 QLIST_HEAD(, BlockCopyTask) tasks; /* All tasks from all block-copy calls */ 140 QLIST_HEAD(, BlockCopyCallState) calls; 141 /* 142 * skip_unallocated: 143 * 144 * Used by sync=top jobs, which first scan the source node for unallocated 145 * areas and clear them in the copy_bitmap. During this process, the bitmap 146 * is thus not fully initialized: It may still have bits set for areas that 147 * are unallocated and should actually not be copied. 148 * 149 * This is indicated by skip_unallocated. 150 * 151 * In this case, block_copy() will query the source’s allocation status, 152 * skip unallocated regions, clear them in the copy_bitmap, and invoke 153 * block_copy_reset_unallocated() every time it does. 154 */ 155 bool skip_unallocated; /* atomic */ 156 /* State fields that use a thread-safe API */ 157 BdrvDirtyBitmap *copy_bitmap; 158 ProgressMeter *progress; 159 SharedResource *mem; 160 RateLimit rate_limit; 161} BlockCopyState; 162 163/* Called with lock held */ 164static BlockCopyTask *find_conflicting_task(BlockCopyState *s, 165 int64_t offset, int64_t bytes) 166{ 167 BlockCopyTask *t; 168 169 QLIST_FOREACH(t, &s->tasks, list) { 170 if (offset + bytes > t->offset && offset < t->offset + t->bytes) { 171 return t; 172 } 173 } 174 175 return NULL; 176} 177 178/* 179 * If there are no intersecting tasks return false. Otherwise, wait for the 180 * first found intersecting tasks to finish and return true. 181 * 182 * Called with lock held. May temporary release the lock. 183 * Return value of 0 proves that lock was NOT released. 184 */ 185static bool coroutine_fn block_copy_wait_one(BlockCopyState *s, int64_t offset, 186 int64_t bytes) 187{ 188 BlockCopyTask *task = find_conflicting_task(s, offset, bytes); 189 190 if (!task) { 191 return false; 192 } 193 194 qemu_co_queue_wait(&task->wait_queue, &s->lock); 195 196 return true; 197} 198 199/* Called with lock held */ 200static int64_t block_copy_chunk_size(BlockCopyState *s) 201{ 202 switch (s->method) { 203 case COPY_READ_WRITE_CLUSTER: 204 return s->cluster_size; 205 case COPY_READ_WRITE: 206 case COPY_RANGE_SMALL: 207 return MIN(MAX(s->cluster_size, BLOCK_COPY_MAX_BUFFER), 208 s->max_transfer); 209 case COPY_RANGE_FULL: 210 return MIN(MAX(s->cluster_size, BLOCK_COPY_MAX_COPY_RANGE), 211 s->max_transfer); 212 default: 213 /* Cannot have COPY_WRITE_ZEROES here. */ 214 abort(); 215 } 216} 217 218/* 219 * Search for the first dirty area in offset/bytes range and create task at 220 * the beginning of it. 221 */ 222static coroutine_fn BlockCopyTask * 223block_copy_task_create(BlockCopyState *s, BlockCopyCallState *call_state, 224 int64_t offset, int64_t bytes) 225{ 226 BlockCopyTask *task; 227 int64_t max_chunk; 228 229 QEMU_LOCK_GUARD(&s->lock); 230 max_chunk = MIN_NON_ZERO(block_copy_chunk_size(s), call_state->max_chunk); 231 if (!bdrv_dirty_bitmap_next_dirty_area(s->copy_bitmap, 232 offset, offset + bytes, 233 max_chunk, &offset, &bytes)) 234 { 235 return NULL; 236 } 237 238 assert(QEMU_IS_ALIGNED(offset, s->cluster_size)); 239 bytes = QEMU_ALIGN_UP(bytes, s->cluster_size); 240 241 /* region is dirty, so no existent tasks possible in it */ 242 assert(!find_conflicting_task(s, offset, bytes)); 243 244 bdrv_reset_dirty_bitmap(s->copy_bitmap, offset, bytes); 245 s->in_flight_bytes += bytes; 246 247 task = g_new(BlockCopyTask, 1); 248 *task = (BlockCopyTask) { 249 .task.func = block_copy_task_entry, 250 .s = s, 251 .call_state = call_state, 252 .offset = offset, 253 .bytes = bytes, 254 .method = s->method, 255 }; 256 qemu_co_queue_init(&task->wait_queue); 257 QLIST_INSERT_HEAD(&s->tasks, task, list); 258 259 return task; 260} 261 262/* 263 * block_copy_task_shrink 264 * 265 * Drop the tail of the task to be handled later. Set dirty bits back and 266 * wake up all tasks waiting for us (may be some of them are not intersecting 267 * with shrunk task) 268 */ 269static void coroutine_fn block_copy_task_shrink(BlockCopyTask *task, 270 int64_t new_bytes) 271{ 272 QEMU_LOCK_GUARD(&task->s->lock); 273 if (new_bytes == task->bytes) { 274 return; 275 } 276 277 assert(new_bytes > 0 && new_bytes < task->bytes); 278 279 task->s->in_flight_bytes -= task->bytes - new_bytes; 280 bdrv_set_dirty_bitmap(task->s->copy_bitmap, 281 task->offset + new_bytes, task->bytes - new_bytes); 282 283 task->bytes = new_bytes; 284 qemu_co_queue_restart_all(&task->wait_queue); 285} 286 287static void coroutine_fn block_copy_task_end(BlockCopyTask *task, int ret) 288{ 289 QEMU_LOCK_GUARD(&task->s->lock); 290 task->s->in_flight_bytes -= task->bytes; 291 if (ret < 0) { 292 bdrv_set_dirty_bitmap(task->s->copy_bitmap, task->offset, task->bytes); 293 } 294 QLIST_REMOVE(task, list); 295 if (task->s->progress) { 296 progress_set_remaining(task->s->progress, 297 bdrv_get_dirty_count(task->s->copy_bitmap) + 298 task->s->in_flight_bytes); 299 } 300 qemu_co_queue_restart_all(&task->wait_queue); 301} 302 303void block_copy_state_free(BlockCopyState *s) 304{ 305 if (!s) { 306 return; 307 } 308 309 ratelimit_destroy(&s->rate_limit); 310 bdrv_release_dirty_bitmap(s->copy_bitmap); 311 shres_destroy(s->mem); 312 g_free(s); 313} 314 315static uint32_t block_copy_max_transfer(BdrvChild *source, BdrvChild *target) 316{ 317 return MIN_NON_ZERO(INT_MAX, 318 MIN_NON_ZERO(source->bs->bl.max_transfer, 319 target->bs->bl.max_transfer)); 320} 321 322void block_copy_set_copy_opts(BlockCopyState *s, bool use_copy_range, 323 bool compress) 324{ 325 /* Keep BDRV_REQ_SERIALISING set (or not set) in block_copy_state_new() */ 326 s->write_flags = (s->write_flags & BDRV_REQ_SERIALISING) | 327 (compress ? BDRV_REQ_WRITE_COMPRESSED : 0); 328 329 if (s->max_transfer < s->cluster_size) { 330 /* 331 * copy_range does not respect max_transfer. We don't want to bother 332 * with requests smaller than block-copy cluster size, so fallback to 333 * buffered copying (read and write respect max_transfer on their 334 * behalf). 335 */ 336 s->method = COPY_READ_WRITE_CLUSTER; 337 } else if (compress) { 338 /* Compression supports only cluster-size writes and no copy-range. */ 339 s->method = COPY_READ_WRITE_CLUSTER; 340 } else { 341 /* 342 * If copy range enabled, start with COPY_RANGE_SMALL, until first 343 * successful copy_range (look at block_copy_do_copy). 344 */ 345 s->method = use_copy_range ? COPY_RANGE_SMALL : COPY_READ_WRITE; 346 } 347} 348 349static int64_t block_copy_calculate_cluster_size(BlockDriverState *target, 350 Error **errp) 351{ 352 int ret; 353 BlockDriverInfo bdi; 354 bool target_does_cow = bdrv_backing_chain_next(target); 355 356 /* 357 * If there is no backing file on the target, we cannot rely on COW if our 358 * backup cluster size is smaller than the target cluster size. Even for 359 * targets with a backing file, try to avoid COW if possible. 360 */ 361 ret = bdrv_get_info(target, &bdi); 362 if (ret == -ENOTSUP && !target_does_cow) { 363 /* Cluster size is not defined */ 364 warn_report("The target block device doesn't provide " 365 "information about the block size and it doesn't have a " 366 "backing file. The default block size of %u bytes is " 367 "used. If the actual block size of the target exceeds " 368 "this default, the backup may be unusable", 369 BLOCK_COPY_CLUSTER_SIZE_DEFAULT); 370 return BLOCK_COPY_CLUSTER_SIZE_DEFAULT; 371 } else if (ret < 0 && !target_does_cow) { 372 error_setg_errno(errp, -ret, 373 "Couldn't determine the cluster size of the target image, " 374 "which has no backing file"); 375 error_append_hint(errp, 376 "Aborting, since this may create an unusable destination image\n"); 377 return ret; 378 } else if (ret < 0 && target_does_cow) { 379 /* Not fatal; just trudge on ahead. */ 380 return BLOCK_COPY_CLUSTER_SIZE_DEFAULT; 381 } 382 383 return MAX(BLOCK_COPY_CLUSTER_SIZE_DEFAULT, bdi.cluster_size); 384} 385 386BlockCopyState *block_copy_state_new(BdrvChild *source, BdrvChild *target, 387 Error **errp) 388{ 389 BlockCopyState *s; 390 int64_t cluster_size; 391 BdrvDirtyBitmap *copy_bitmap; 392 bool is_fleecing; 393 394 cluster_size = block_copy_calculate_cluster_size(target->bs, errp); 395 if (cluster_size < 0) { 396 return NULL; 397 } 398 399 copy_bitmap = bdrv_create_dirty_bitmap(source->bs, cluster_size, NULL, 400 errp); 401 if (!copy_bitmap) { 402 return NULL; 403 } 404 bdrv_disable_dirty_bitmap(copy_bitmap); 405 406 /* 407 * If source is in backing chain of target assume that target is going to be 408 * used for "image fleecing", i.e. it should represent a kind of snapshot of 409 * source at backup-start point in time. And target is going to be read by 410 * somebody (for example, used as NBD export) during backup job. 411 * 412 * In this case, we need to add BDRV_REQ_SERIALISING write flag to avoid 413 * intersection of backup writes and third party reads from target, 414 * otherwise reading from target we may occasionally read already updated by 415 * guest data. 416 * 417 * For more information see commit f8d59dfb40bb and test 418 * tests/qemu-iotests/222 419 */ 420 is_fleecing = bdrv_chain_contains(target->bs, source->bs); 421 422 s = g_new(BlockCopyState, 1); 423 *s = (BlockCopyState) { 424 .source = source, 425 .target = target, 426 .copy_bitmap = copy_bitmap, 427 .cluster_size = cluster_size, 428 .len = bdrv_dirty_bitmap_size(copy_bitmap), 429 .write_flags = (is_fleecing ? BDRV_REQ_SERIALISING : 0), 430 .mem = shres_create(BLOCK_COPY_MAX_MEM), 431 .max_transfer = QEMU_ALIGN_DOWN( 432 block_copy_max_transfer(source, target), 433 cluster_size), 434 }; 435 436 block_copy_set_copy_opts(s, false, false); 437 438 ratelimit_init(&s->rate_limit); 439 qemu_co_mutex_init(&s->lock); 440 QLIST_INIT(&s->tasks); 441 QLIST_INIT(&s->calls); 442 443 return s; 444} 445 446/* Only set before running the job, no need for locking. */ 447void block_copy_set_progress_meter(BlockCopyState *s, ProgressMeter *pm) 448{ 449 s->progress = pm; 450} 451 452/* 453 * Takes ownership of @task 454 * 455 * If pool is NULL directly run the task, otherwise schedule it into the pool. 456 * 457 * Returns: task.func return code if pool is NULL 458 * otherwise -ECANCELED if pool status is bad 459 * otherwise 0 (successfully scheduled) 460 */ 461static coroutine_fn int block_copy_task_run(AioTaskPool *pool, 462 BlockCopyTask *task) 463{ 464 if (!pool) { 465 int ret = task->task.func(&task->task); 466 467 g_free(task); 468 return ret; 469 } 470 471 aio_task_pool_wait_slot(pool); 472 if (aio_task_pool_status(pool) < 0) { 473 co_put_to_shres(task->s->mem, task->bytes); 474 block_copy_task_end(task, -ECANCELED); 475 g_free(task); 476 return -ECANCELED; 477 } 478 479 aio_task_pool_start_task(pool, &task->task); 480 481 return 0; 482} 483 484/* 485 * block_copy_do_copy 486 * 487 * Do copy of cluster-aligned chunk. Requested region is allowed to exceed 488 * s->len only to cover last cluster when s->len is not aligned to clusters. 489 * 490 * No sync here: nor bitmap neighter intersecting requests handling, only copy. 491 * 492 * @method is an in-out argument, so that copy_range can be either extended to 493 * a full-size buffer or disabled if the copy_range attempt fails. The output 494 * value of @method should be used for subsequent tasks. 495 * Returns 0 on success. 496 */ 497static int coroutine_fn block_copy_do_copy(BlockCopyState *s, 498 int64_t offset, int64_t bytes, 499 BlockCopyMethod *method, 500 bool *error_is_read) 501{ 502 int ret; 503 int64_t nbytes = MIN(offset + bytes, s->len) - offset; 504 void *bounce_buffer = NULL; 505 506 assert(offset >= 0 && bytes > 0 && INT64_MAX - offset >= bytes); 507 assert(QEMU_IS_ALIGNED(offset, s->cluster_size)); 508 assert(QEMU_IS_ALIGNED(bytes, s->cluster_size)); 509 assert(offset < s->len); 510 assert(offset + bytes <= s->len || 511 offset + bytes == QEMU_ALIGN_UP(s->len, s->cluster_size)); 512 assert(nbytes < INT_MAX); 513 514 switch (*method) { 515 case COPY_WRITE_ZEROES: 516 ret = bdrv_co_pwrite_zeroes(s->target, offset, nbytes, s->write_flags & 517 ~BDRV_REQ_WRITE_COMPRESSED); 518 if (ret < 0) { 519 trace_block_copy_write_zeroes_fail(s, offset, ret); 520 *error_is_read = false; 521 } 522 return ret; 523 524 case COPY_RANGE_SMALL: 525 case COPY_RANGE_FULL: 526 ret = bdrv_co_copy_range(s->source, offset, s->target, offset, nbytes, 527 0, s->write_flags); 528 if (ret >= 0) { 529 /* Successful copy-range, increase chunk size. */ 530 *method = COPY_RANGE_FULL; 531 return 0; 532 } 533 534 trace_block_copy_copy_range_fail(s, offset, ret); 535 *method = COPY_READ_WRITE; 536 /* Fall through to read+write with allocated buffer */ 537 538 case COPY_READ_WRITE_CLUSTER: 539 case COPY_READ_WRITE: 540 /* 541 * In case of failed copy_range request above, we may proceed with 542 * buffered request larger than BLOCK_COPY_MAX_BUFFER. 543 * Still, further requests will be properly limited, so don't care too 544 * much. Moreover the most likely case (copy_range is unsupported for 545 * the configuration, so the very first copy_range request fails) 546 * is handled by setting large copy_size only after first successful 547 * copy_range. 548 */ 549 550 bounce_buffer = qemu_blockalign(s->source->bs, nbytes); 551 552 ret = bdrv_co_pread(s->source, offset, nbytes, bounce_buffer, 0); 553 if (ret < 0) { 554 trace_block_copy_read_fail(s, offset, ret); 555 *error_is_read = true; 556 goto out; 557 } 558 559 ret = bdrv_co_pwrite(s->target, offset, nbytes, bounce_buffer, 560 s->write_flags); 561 if (ret < 0) { 562 trace_block_copy_write_fail(s, offset, ret); 563 *error_is_read = false; 564 goto out; 565 } 566 567 out: 568 qemu_vfree(bounce_buffer); 569 break; 570 571 default: 572 abort(); 573 } 574 575 return ret; 576} 577 578static coroutine_fn int block_copy_task_entry(AioTask *task) 579{ 580 BlockCopyTask *t = container_of(task, BlockCopyTask, task); 581 BlockCopyState *s = t->s; 582 bool error_is_read = false; 583 BlockCopyMethod method = t->method; 584 int ret; 585 586 ret = block_copy_do_copy(s, t->offset, t->bytes, &method, &error_is_read); 587 588 WITH_QEMU_LOCK_GUARD(&s->lock) { 589 if (s->method == t->method) { 590 s->method = method; 591 } 592 593 if (ret < 0) { 594 if (!t->call_state->ret) { 595 t->call_state->ret = ret; 596 t->call_state->error_is_read = error_is_read; 597 } 598 } else if (s->progress) { 599 progress_work_done(s->progress, t->bytes); 600 } 601 } 602 co_put_to_shres(s->mem, t->bytes); 603 block_copy_task_end(t, ret); 604 605 return ret; 606} 607 608static int block_copy_block_status(BlockCopyState *s, int64_t offset, 609 int64_t bytes, int64_t *pnum) 610{ 611 int64_t num; 612 BlockDriverState *base; 613 int ret; 614 615 if (qatomic_read(&s->skip_unallocated)) { 616 base = bdrv_backing_chain_next(s->source->bs); 617 } else { 618 base = NULL; 619 } 620 621 ret = bdrv_block_status_above(s->source->bs, base, offset, bytes, &num, 622 NULL, NULL); 623 if (ret < 0 || num < s->cluster_size) { 624 /* 625 * On error or if failed to obtain large enough chunk just fallback to 626 * copy one cluster. 627 */ 628 num = s->cluster_size; 629 ret = BDRV_BLOCK_ALLOCATED | BDRV_BLOCK_DATA; 630 } else if (offset + num == s->len) { 631 num = QEMU_ALIGN_UP(num, s->cluster_size); 632 } else { 633 num = QEMU_ALIGN_DOWN(num, s->cluster_size); 634 } 635 636 *pnum = num; 637 return ret; 638} 639 640/* 641 * Check if the cluster starting at offset is allocated or not. 642 * return via pnum the number of contiguous clusters sharing this allocation. 643 */ 644static int block_copy_is_cluster_allocated(BlockCopyState *s, int64_t offset, 645 int64_t *pnum) 646{ 647 BlockDriverState *bs = s->source->bs; 648 int64_t count, total_count = 0; 649 int64_t bytes = s->len - offset; 650 int ret; 651 652 assert(QEMU_IS_ALIGNED(offset, s->cluster_size)); 653 654 while (true) { 655 ret = bdrv_is_allocated(bs, offset, bytes, &count); 656 if (ret < 0) { 657 return ret; 658 } 659 660 total_count += count; 661 662 if (ret || count == 0) { 663 /* 664 * ret: partial segment(s) are considered allocated. 665 * otherwise: unallocated tail is treated as an entire segment. 666 */ 667 *pnum = DIV_ROUND_UP(total_count, s->cluster_size); 668 return ret; 669 } 670 671 /* Unallocated segment(s) with uncertain following segment(s) */ 672 if (total_count >= s->cluster_size) { 673 *pnum = total_count / s->cluster_size; 674 return 0; 675 } 676 677 offset += count; 678 bytes -= count; 679 } 680} 681 682/* 683 * Reset bits in copy_bitmap starting at offset if they represent unallocated 684 * data in the image. May reset subsequent contiguous bits. 685 * @return 0 when the cluster at @offset was unallocated, 686 * 1 otherwise, and -ret on error. 687 */ 688int64_t block_copy_reset_unallocated(BlockCopyState *s, 689 int64_t offset, int64_t *count) 690{ 691 int ret; 692 int64_t clusters, bytes; 693 694 ret = block_copy_is_cluster_allocated(s, offset, &clusters); 695 if (ret < 0) { 696 return ret; 697 } 698 699 bytes = clusters * s->cluster_size; 700 701 if (!ret) { 702 qemu_co_mutex_lock(&s->lock); 703 bdrv_reset_dirty_bitmap(s->copy_bitmap, offset, bytes); 704 if (s->progress) { 705 progress_set_remaining(s->progress, 706 bdrv_get_dirty_count(s->copy_bitmap) + 707 s->in_flight_bytes); 708 } 709 qemu_co_mutex_unlock(&s->lock); 710 } 711 712 *count = bytes; 713 return ret; 714} 715 716/* 717 * block_copy_dirty_clusters 718 * 719 * Copy dirty clusters in @offset/@bytes range. 720 * Returns 1 if dirty clusters found and successfully copied, 0 if no dirty 721 * clusters found and -errno on failure. 722 */ 723static int coroutine_fn 724block_copy_dirty_clusters(BlockCopyCallState *call_state) 725{ 726 BlockCopyState *s = call_state->s; 727 int64_t offset = call_state->offset; 728 int64_t bytes = call_state->bytes; 729 730 int ret = 0; 731 bool found_dirty = false; 732 int64_t end = offset + bytes; 733 AioTaskPool *aio = NULL; 734 735 /* 736 * block_copy() user is responsible for keeping source and target in same 737 * aio context 738 */ 739 assert(bdrv_get_aio_context(s->source->bs) == 740 bdrv_get_aio_context(s->target->bs)); 741 742 assert(QEMU_IS_ALIGNED(offset, s->cluster_size)); 743 assert(QEMU_IS_ALIGNED(bytes, s->cluster_size)); 744 745 while (bytes && aio_task_pool_status(aio) == 0 && 746 !qatomic_read(&call_state->cancelled)) { 747 BlockCopyTask *task; 748 int64_t status_bytes; 749 750 task = block_copy_task_create(s, call_state, offset, bytes); 751 if (!task) { 752 /* No more dirty bits in the bitmap */ 753 trace_block_copy_skip_range(s, offset, bytes); 754 break; 755 } 756 if (task->offset > offset) { 757 trace_block_copy_skip_range(s, offset, task->offset - offset); 758 } 759 760 found_dirty = true; 761 762 ret = block_copy_block_status(s, task->offset, task->bytes, 763 &status_bytes); 764 assert(ret >= 0); /* never fail */ 765 if (status_bytes < task->bytes) { 766 block_copy_task_shrink(task, status_bytes); 767 } 768 if (qatomic_read(&s->skip_unallocated) && 769 !(ret & BDRV_BLOCK_ALLOCATED)) { 770 block_copy_task_end(task, 0); 771 trace_block_copy_skip_range(s, task->offset, task->bytes); 772 offset = task_end(task); 773 bytes = end - offset; 774 g_free(task); 775 continue; 776 } 777 if (ret & BDRV_BLOCK_ZERO) { 778 task->method = COPY_WRITE_ZEROES; 779 } 780 781 if (!call_state->ignore_ratelimit) { 782 uint64_t ns = ratelimit_calculate_delay(&s->rate_limit, 0); 783 if (ns > 0) { 784 block_copy_task_end(task, -EAGAIN); 785 g_free(task); 786 qemu_co_sleep_ns_wakeable(&call_state->sleep, 787 QEMU_CLOCK_REALTIME, ns); 788 continue; 789 } 790 } 791 792 ratelimit_calculate_delay(&s->rate_limit, task->bytes); 793 794 trace_block_copy_process(s, task->offset); 795 796 co_get_from_shres(s->mem, task->bytes); 797 798 offset = task_end(task); 799 bytes = end - offset; 800 801 if (!aio && bytes) { 802 aio = aio_task_pool_new(call_state->max_workers); 803 } 804 805 ret = block_copy_task_run(aio, task); 806 if (ret < 0) { 807 goto out; 808 } 809 } 810 811out: 812 if (aio) { 813 aio_task_pool_wait_all(aio); 814 815 /* 816 * We are not really interested in -ECANCELED returned from 817 * block_copy_task_run. If it fails, it means some task already failed 818 * for real reason, let's return first failure. 819 * Still, assert that we don't rewrite failure by success. 820 * 821 * Note: ret may be positive here because of block-status result. 822 */ 823 assert(ret >= 0 || aio_task_pool_status(aio) < 0); 824 ret = aio_task_pool_status(aio); 825 826 aio_task_pool_free(aio); 827 } 828 829 return ret < 0 ? ret : found_dirty; 830} 831 832void block_copy_kick(BlockCopyCallState *call_state) 833{ 834 qemu_co_sleep_wake(&call_state->sleep); 835} 836 837/* 838 * block_copy_common 839 * 840 * Copy requested region, accordingly to dirty bitmap. 841 * Collaborate with parallel block_copy requests: if they succeed it will help 842 * us. If they fail, we will retry not-copied regions. So, if we return error, 843 * it means that some I/O operation failed in context of _this_ block_copy call, 844 * not some parallel operation. 845 */ 846static int coroutine_fn block_copy_common(BlockCopyCallState *call_state) 847{ 848 int ret; 849 BlockCopyState *s = call_state->s; 850 851 qemu_co_mutex_lock(&s->lock); 852 QLIST_INSERT_HEAD(&s->calls, call_state, list); 853 qemu_co_mutex_unlock(&s->lock); 854 855 do { 856 ret = block_copy_dirty_clusters(call_state); 857 858 if (ret == 0 && !qatomic_read(&call_state->cancelled)) { 859 WITH_QEMU_LOCK_GUARD(&s->lock) { 860 /* 861 * Check that there is no task we still need to 862 * wait to complete 863 */ 864 ret = block_copy_wait_one(s, call_state->offset, 865 call_state->bytes); 866 if (ret == 0) { 867 /* 868 * No pending tasks, but check again the bitmap in this 869 * same critical section, since a task might have failed 870 * between this and the critical section in 871 * block_copy_dirty_clusters(). 872 * 873 * block_copy_wait_one return value 0 also means that it 874 * didn't release the lock. So, we are still in the same 875 * critical section, not interrupted by any concurrent 876 * access to state. 877 */ 878 ret = bdrv_dirty_bitmap_next_dirty(s->copy_bitmap, 879 call_state->offset, 880 call_state->bytes) >= 0; 881 } 882 } 883 } 884 885 /* 886 * We retry in two cases: 887 * 1. Some progress done 888 * Something was copied, which means that there were yield points 889 * and some new dirty bits may have appeared (due to failed parallel 890 * block-copy requests). 891 * 2. We have waited for some intersecting block-copy request 892 * It may have failed and produced new dirty bits. 893 */ 894 } while (ret > 0 && !qatomic_read(&call_state->cancelled)); 895 896 qatomic_store_release(&call_state->finished, true); 897 898 if (call_state->cb) { 899 call_state->cb(call_state->cb_opaque); 900 } 901 902 qemu_co_mutex_lock(&s->lock); 903 QLIST_REMOVE(call_state, list); 904 qemu_co_mutex_unlock(&s->lock); 905 906 return ret; 907} 908 909int coroutine_fn block_copy(BlockCopyState *s, int64_t start, int64_t bytes, 910 bool ignore_ratelimit) 911{ 912 BlockCopyCallState call_state = { 913 .s = s, 914 .offset = start, 915 .bytes = bytes, 916 .ignore_ratelimit = ignore_ratelimit, 917 .max_workers = BLOCK_COPY_MAX_WORKERS, 918 }; 919 920 return block_copy_common(&call_state); 921} 922 923static void coroutine_fn block_copy_async_co_entry(void *opaque) 924{ 925 block_copy_common(opaque); 926} 927 928BlockCopyCallState *block_copy_async(BlockCopyState *s, 929 int64_t offset, int64_t bytes, 930 int max_workers, int64_t max_chunk, 931 BlockCopyAsyncCallbackFunc cb, 932 void *cb_opaque) 933{ 934 BlockCopyCallState *call_state = g_new(BlockCopyCallState, 1); 935 936 *call_state = (BlockCopyCallState) { 937 .s = s, 938 .offset = offset, 939 .bytes = bytes, 940 .max_workers = max_workers, 941 .max_chunk = max_chunk, 942 .cb = cb, 943 .cb_opaque = cb_opaque, 944 945 .co = qemu_coroutine_create(block_copy_async_co_entry, call_state), 946 }; 947 948 qemu_coroutine_enter(call_state->co); 949 950 return call_state; 951} 952 953void block_copy_call_free(BlockCopyCallState *call_state) 954{ 955 if (!call_state) { 956 return; 957 } 958 959 assert(qatomic_read(&call_state->finished)); 960 g_free(call_state); 961} 962 963bool block_copy_call_finished(BlockCopyCallState *call_state) 964{ 965 return qatomic_read(&call_state->finished); 966} 967 968bool block_copy_call_succeeded(BlockCopyCallState *call_state) 969{ 970 return qatomic_load_acquire(&call_state->finished) && 971 !qatomic_read(&call_state->cancelled) && 972 call_state->ret == 0; 973} 974 975bool block_copy_call_failed(BlockCopyCallState *call_state) 976{ 977 return qatomic_load_acquire(&call_state->finished) && 978 !qatomic_read(&call_state->cancelled) && 979 call_state->ret < 0; 980} 981 982bool block_copy_call_cancelled(BlockCopyCallState *call_state) 983{ 984 return qatomic_read(&call_state->cancelled); 985} 986 987int block_copy_call_status(BlockCopyCallState *call_state, bool *error_is_read) 988{ 989 assert(qatomic_load_acquire(&call_state->finished)); 990 if (error_is_read) { 991 *error_is_read = call_state->error_is_read; 992 } 993 return call_state->ret; 994} 995 996/* 997 * Note that cancelling and finishing are racy. 998 * User can cancel a block-copy that is already finished. 999 */ 1000void block_copy_call_cancel(BlockCopyCallState *call_state) 1001{ 1002 qatomic_set(&call_state->cancelled, true); 1003 block_copy_kick(call_state); 1004} 1005 1006BdrvDirtyBitmap *block_copy_dirty_bitmap(BlockCopyState *s) 1007{ 1008 return s->copy_bitmap; 1009} 1010 1011int64_t block_copy_cluster_size(BlockCopyState *s) 1012{ 1013 return s->cluster_size; 1014} 1015 1016void block_copy_set_skip_unallocated(BlockCopyState *s, bool skip) 1017{ 1018 qatomic_set(&s->skip_unallocated, skip); 1019} 1020 1021void block_copy_set_speed(BlockCopyState *s, uint64_t speed) 1022{ 1023 ratelimit_set_speed(&s->rate_limit, speed, BLOCK_COPY_SLICE_TIME); 1024 1025 /* 1026 * Note: it's good to kick all call states from here, but it should be done 1027 * only from a coroutine, to not crash if s->calls list changed while 1028 * entering one call. So for now, the only user of this function kicks its 1029 * only one call_state by hand. 1030 */ 1031} 1032
__label__pos
0.999156
Help us improve your experience. Let us know what you think. Do you have time for a two-minute survey?   Managing Off-site Hosts   An off-site host is a JSA appliance that can't be accessed through the JSA console in your current deployment. You can configure an off-site host to transfer data to or to receive data from your JSA deployment. Configuring an Off-site Source To forward event and flow data to an Event Collector in another deployment, configure the target deployment to include an off-site source so that it knows which computer is sending the data. To prevent connection errors, when you configure off-site source and target components, deploy the JSA Console with the off-site source first. Then, deploy the JSA console with the off-site target. 1. On the navigation menu (), click Admin to open the admin tab. 2. In the System Configuration section, click System and License Management. 3. In the Display list, select Systems. 4. On the Deployment Actions menu, click Manage Off-site Sources. 5. Click Add and configure the parameters. The name can be up to 20 characters in length and can include underscores or hyphens. 6. Click Save. 7. Click Manage Connections to specify which JSA hosts you want to receive the data. The host must have an Event Collector to receive the data. 8. Repeat the steps to configure all off-site sources that you want to configure. 9. Deploy the changes and restart the event collection service. Configuring an Off-site Target To forward event and flow data to an Event Collector in another deployment, configure the source deployment to include an off-site target so that it knows which computer to send the data to. You must know the listening ports for the off-site target appliance. By default, the listening port for events is 32004, and 32000 for flows. To find the listening port on the target appliance, follow these steps: 1. In the target deployment, click the System and License Management icon. 2. Select the host and click Deployment Actions >Edit Host. 3. Click the Component Management settings icon (), and find the ports in the Event Forwarding Listening Port and Flow Forwarding Listening Port fields. To prevent connection errors, when you configure off-site source and target components, deploy the JSA Console with the off-site source first. Then, deploy the JSA console with the off-site target. 1. On the navigation menu (), click Admin to open the admin tab. 2. In the System Configuration section, click System and License Management. 3. In the Display list, select Systems. 4. On the Deployment Actions menu, click Manage Off-site Targets. 5. Click Add and configure the parameters. The name can be up to 20 characters in length and can include underscores or hyphens. The default port to listen for events is 32004, and 32000 for flows. 6. Click Save. 7. Click Manage Connections to specify which JSA hosts you want to receive the data. Only hosts that have an Event Collector are shown in the list. 8. Repeat the steps to configure all off-site targets that you want to configure. 9. On the Admin tab, click Deploy changes. Generating Public Keys for JSA Products To forward normalized events in JSA, you must copy the public key file, /root/.ssh/id_rsa.pub, from the off-site source to the off-site target. If the off-site source and off-site target are on separate systems, the public key is automatically generated. If the off-site source and target are both on an all-in-one system, the public key is not automatically generated. You must manually generate the public key. To manually generate the public key, follow these steps: 1. Use SSH to log in to your system as the root user. 2. To generate the public key, type the following command: opt/qradar/bin/ssh-key-generating 3. Press Enter. The public and private key pair is generated and saved in the /root/.ssh/id_rsa folder. Forwarding Filtered Flows You can set up forwarding of filtered flows. You can use filtered flows to split flow forwarding across multiple boxes, and to forward specific flows for specific investigations. 1. On the target system, set up the source system as an off-site source. 1. On the navigation menu (), click Admin to open the admin tab. 2. Click System and License Management > Deployment Actions > Manage Off-Site Sources. 3. Add the source system IP address, and select Receive Events and/or Receive Flows. 4. Select Manage Connections and select which host is expecting to receive the off-site connection. 5. Click Save. 6. Select Deploy Full Configuration from the Advanced menu for the changes to take effect. 2. On the source system, set up the forwarding destination, IP address, and port number. 1. Click Main menu > Admin. 2. Click Forwarding Destinations > Add. 3. Set the IP address of the target system and the destination port. 4. Enter 32000 for the port number on the source system. Port 32000 is used for flow forwarding. 5. Select Normalized from the Event Format list. 3. Set up routing rules. 1. Click Main menu > Admin. 2. Click Routing Rules > Add. 3. Select the rules that you want to add. Note Rules forward flows that are based on offenses, or based on CRE information when Offline Forwarding is selected on the Routing Rules page. The flows that are filtered on the Routing Rules screen are forwarded. Example: Forwarding Normalized Events and Flows To forward normalized events and flows, configure the target deployment to include an off-site source so that it knows which computer is sending the data. Configure the source deployment to include an off-site target so that it knows which computer to send the data to. The following diagram shows forwarding event and flow data between deployments. Figure 1: Forwarding Data Between Deployments by Using SSH Forwarding Data Between Deployments by Using SSH If the off-site source or target is an all-in-one system, the public key is not automatically generated; therefore, you must manually generate the public key. For more information, see Generating public keys for JSA products. To forward normalized events and flows from Deployment A to Deployment B: 1. Configure an off-site target in Deployment A. The off-site target configuration includes the IP address of the Event Collector in Deployment B that receives the data. 2. Configure an off-site source in Deployment B. The off-site source configuration includes the IP address and the port number of the Event Collector in Deployment A that is sending the data. 3. To transfer encrypted data, you must enable encryption on both the off-site source and the off-site target. To ensure appropriate access, the SSH public key for the source system (Deployment A) must be available to the target system (Deployment B). For example, to enable encryption between Deployment A and Deployment B, follow these steps: 4. Create ssh keys by using the ssh-keygen -1 -t rsa command, and press enter when prompted about the directory and passphrase. By default, the id_rsa.pub file is stored in the /root/.ssh directory. 5. Copy the id_rsa.pub file to the /root/.ssh directory on the Event Collector and on the JSA console in the source system (Deployment A). 6. Rename the file to authorized_keys. Ensure that the source system is configured with the appropriate permissions to send event and flow data to the target system. 7. If you didn't use the chmod 600 authorized_keys command to assign rw owner privileges to the file and the parent directory, use the ssh-copy-id command with the -i parameter to specify that the identity file /root/.ssh/id_rsa.pub be used. For example, type the following command to append entries or create a new authorized_keys file on the target console with the right privileges. This command does not check for duplicate entries. ssh-copy-id -i [email protected] 8. Configure the source system to ensure that forwarding of events and flows is not interrupted by other configuration activities, such as adding a managed host to one of the consoles. For example, if a managed host is added to a console that is forwarding events, then an authorized_keys file must exist in the /root/.ssh directory on the managed host. If not, adding a managed host fails. This file is required regardless of whether encryption is used between the managed host and the console. 9. On the JSA console in the source system (Deployment A), create a ssh_keys_created file under /opt/qradar/conf. 10. Change the owner and group to nobody and the permission to 775 to make sure that the file can be backed up and restored properly. chown nobody:nobody /opt/qradar/conf/ssh_keys_created chmod 775 /opt/qradar/conf/ssh_keys_created 11. To prevent connection errors, deploy the changes in the target system (Deployment B) before you deploy the changes in the source system (Deployment A). If you update the Event Collector configuration or the monitoring ports, you must manually update the configuration for the off-site source and off-site target to maintain the connection between the two deployments. If you want to disconnect the source system (Deployment A), you must remove the connections from both deployments. Remove the off-site target from the source system (Deployment A), and then remove the off-site source from the target system (Deployment B).    
__label__pos
0.540705
Survo-ristikoita: www.survo.fi/ristikot --------------------------------------- Tehtävät 9.4.2007 Survo-ristikko 51/2007 (35) A B C D E 1 5 2 * * 8 35 2 * * * 11 * 40 3 * * 10 * * 45 28 23 24 21 24 Survo-ristikko 52/2007 (530) A B C D E 1 * * 17 3 * 49 2 * * * * 11 60 3 5 18 * * * 46 4 * 7 12 * * 55 42 41 41 42 44 Survo-ristikko 53/2007 (1100 ilman vihjettä) A B C D 1 * * * * 27 2 * * * * 13 3 * * * * 53 4 * * * * 43 47 39 29 21 Vihje: Yhtälön 456*X^2+1416*X-131=0 positiivisen juuren ketjumurtolukukehitelmä 1 X = ----------------- 1 A1 + ----------------- 1 B1 + ------------ 1 C1 + ------- D1 + X antaa ristikon ensimmäisen rivin. Survossa luvut saa lasketuksi esim. seuraavilla rekursiivisilla, editoriaalisen laskennan kaavoilla Y(N,X):=if(N=0)then(X)else(1/(Y(N-1,X)-A(N-1,X))) A(N,X):=if(N=0)then(int(X))else(int(Y(N,X))) Copyright (c) Survo Systems 2006-2007
__label__pos
0.996527
user-select La propiedad CSS user-select controla si el usuario puede seleccionar el texto. Esto no tiene ningún efecto en el contenido cargado bajo chrome, excepto en cuadros de texto. /* Valores de palabras clave */ user-select: none; user-select: auto; user-select: text; user-select: contain; user-select: all; /* Varoles globales */ user-select: inherit; user-select: initial; user-select: unset; /* Valores Mozilla-specific */ -moz-user-select: none; -moz-user-select: text; -moz-user-select: all; /* Valores WebKit-specific */ -webkit-user-select: none; -webkit-user-select: text; -webkit-user-select: all; /* No funciona el Safari; solo usa "none" or "text", o si no hará permitir escribir en el contenedor <html> */ /* Valores Microsoft-specific */ -ms-user-select: none; -ms-user-select: text; -ms-user-select: element; Valor inicialauto Applies toall elements Heredableno Valor calculadocomo se especifica Animation typediscrete Syntaxis none El texto y sus sub elementos no son seleccionables. Tenga en cuenta que el objeto Selection puede contener estos elementos. auto El valor calculado auto se determina de la siguiente manera: En los pseudo elementos ::before y ::after, el valor calculado es none • Si el elemento es un elemento editable, el valor calculado es contain • De lo contrario, si el valor calculador de user-select en la matriz de este elemento es all, el valor calculado es all • De lo contrario, si el valor calulado de user-select  en la matriz de este elemento es  none, el valor calculado es none • De lo contrario, el valor calculado es text text El texto puede ser seleccionado por el usuario. all En el editor HTML, si  se realiza doble-click o click-contextual en el subelemento, se seleccionará el antecesor más alto de el valor. contain Permite que la selección comience dentro del elemento; sin embargo, la selección estará contenida por los límites de ese elemento. element (IE-specific alias) Igual que contain. Solo lo soportado en Internet Explorer. Formal syntax auto | (en-US) text | (en-US) none | (en-US) contain | (en-US) all Ejemplos HTML <p>Debería poder seleccionar este texto.</p> <p class="unselectable">No puedes seleccionar este texto</p> <p class="all">Al hacer clic una vez se seleccionará todo este texto.</p> CSS .unselectable { -moz-user-select: none; -webkit-user-select: none; -ms-user-select: none; user-select: none; } .all { -moz-user-select: all; -webkit-user-select: all; -ms-user-select: all; user-select: all; } Resultado Especificaciones Especificación Estado Comentario CSS Basic User Interface Module Level 4 La definición de 'user-select' en esta especificación. Working Draft Initial definition. Also defines -webkit-user-select as a deprecated alias of user-select. Compatibilidad con navegadores BCD tables only load in the browser Véase también
__label__pos
0.736949
Depth-first search From PEGWiki (Redirected from DFS) Jump to: navigation, search Animated example of DFS. Depth-first search (DFS) is one of the most-basic and well-known types of algorithm in graph theory. The basic idea of DFS is deceptively simple, but it can be extended to yield asymptotically optimal solutions to many important problems in graph theory. It is a type of graph search (what it means to search a graph is explained in that article). Principles[edit] DFS is distinguished from other graph search techniques in that, after visiting a vertex and adding its neighbours to the fringe, the next vertex to be visited is always the newest vertex on the fringe, so that as long as the visited vertex has unvisited neighbours, each neighbour is itself visited immediately. For example, suppose a graph consists of the vertices {1,2,3,4,5} and the edges {(1,2),(1,3),(2,4),(3,5)}. If we start by visiting 1, then we add vertices 2 and 3 to the fringe. Suppose that we visit 2 next. (2 and 3 could have been added in any order, so it doesn't really matter.) Then, 4 is placed on the fringe, which now consists of 3 and 4. Since 4 was necessarily added after 3, we visit 4 next. Finding that it has no neighbours, we visit 3, and finally 5. It is called depth-first search because it goes as far down a path (deep) as possible before visiting other vertices. Stack-based implementation[edit] The basic template given at Graph search can be easily adapted for DFS: input G for all u ∈ V(G) let visited[u] = false for each u ∈ V(G) if visited[u] = false let S be empty push(S,u) while not empty(S) v = pop(S) if visited[v] = false visit v visited[v] = true for each w ∈ V(G) such that (v,w) ∈ E(G) and visited[w] = false push(S,w) F has been replaced by a stack, S, because a stack satisfies exactly the necessary property for DFS: the most recently added item is the first to be removed. Recursive implementation[edit] There is a much cleaner implementation of DFS using recursion. Notice that no matter which vertex we happen to be examining, we execute the same code. That suggests that we might be able to wrap this code in a recursive function, in which the argument is the vertex to be visited next. Also, after a particular vertex is visited, we know that we want to immediately visit any of its unvisited neighbours. We can do so by recursing on that vertex: function DFS(u) if not visited[u] visit u visited[u] = true for each v ∈ V(G) such that (u,v) ∈ E(G) DFS(v) input G for each u ∈ V(G) let visited[u]=false for each u ∈ V(G) DFS(u) Note that we do not need to check if a vertex has been visited before recursing on it, because it is checked immediately after recursing. The recursive implementation will be preferred and assumed whenever DFS is discussed on this wiki. Performance characteristics[edit] Time[edit] Most DFS algorithms feature a constant time "visit" operation. Suppose, for the sake of simplicity, that the entire graph consists of one connected component. Each vertex will be visited exactly once, with each visit being a constant-time operation, so \mathcal{O}(V) time is spent visiting vertices. Whenever a vertex is visited, every edge radiating from that vertex is considered, with constant time spent on each edge. It follows that every edge in the graph will be considered exactly twice -- once for each vertex upon which it is incident. Thus, the time spent considering edges is again a constant factor times the number of edges, \mathcal{O}(E). This makes DFS \mathcal{O}(E+V) overall, linear time, which, like BFS, is asymptotically optimal for a graph search. Space[edit] Recursive implementation[edit] For every additional level of recursive depth, a constant amount of memory is required. Notice that on the recursive stack, no vertex may be repeated (except the one at the top, in the current instance, just before the if not visited[u] check), because a vertex must be marked visited before any vertices can be pushed onto the stack above it, and recursion stops when a vertex is encountered which has already been marked visited. Therefore, DFS requires \mathcal{O}(V) additional memory, and can be called "linear space". In practice this is often less than the memory used to store the graph in the first place, if the graph is explicit (\mathcal{O}(E+V) space with an adjacency list). Non-recursive implementation[edit] The less-preferred non-recursive implementation cannot be guaranteed to take only \mathcal{O}(V) additional memory, since vertices are not visited immediately after being pushed on the stack (and therefore a vertex may be on the stack several times at any given moment). However, it is also clear that there can be no more than \mathcal{O}(E) vertices on the stack at any given time, because every addition of a vertex to the stack is due to the traversal of an edge (except for that of the initial vertex) and no edge can be traversed more than twice. The stack space is therefore bounded by \mathcal{O}(E), and it is easy to see that a complete graph does indeed meet this upper bound. \mathcal{O}(V) space is also required to store the list of visited vertices, so \mathcal{O}(E+V) space is required overall. The DFS tree[edit] Running DFS on a graph starting from a vertex from which all other vertices are reachable produces a depth-first spanning tree or simply DFS tree, because every vertex is visited and no vertex is visited on more than one path from the root (starting vertex). Depth-first spanning trees tend to be tall and slender, because DFS tends to always keep going along one path until it cannot go any further, and, hence, produce relatively long paths. The DFS tree has a number of useful properties that allow efficient solutions to important problems (see the Applications section). Undirected graphs[edit] When we run DFS on an undirected graph, we generally consider it as a directed graph in which every edge is bidirectional. (This is akin to adding an arrowhead to each end of each edge in a diagram of the graph using circles and line segments.) Running DFS on this graph selects some of the edges (|V|-1 of them, to be specific) to form the spanning tree. Note that these edges point generally away from the source. They are known as tree edges. When we traverse a tree link, we make a recursive call to a previously unvisited vertex. The directed edges running in the reverse direction - from a node to its parent in the DFS tree - are known as parent edges, for obvious reasons. They correspond to a recursive call reaching the end of the function and returning to the caller. The remaining edges are all classified as either down edges or back edges. If a edge points from a node to one of its descendants in the spanning tree rooted at the source vertex, but it is not a tree edge, then it is a down edge; otherwise, it points to an ancestor, and if it is not a parent edge then it is a back edge. In the recursive implementation given, encountering either a down edge or a back edge results in an immediate return as the vertex has already been visited. Evidently, the reverse of a down edge is always a back edge and vice versa. Every (directed) edge in the graph must be classifiable into one of the four categories: tree, parent, down, or back. To see this, suppose that the edge from node A to node B in some graph does fit into any of these categories. Then, in the spanning tree, A is neither an ancestor nor a descendant of B. This means that the lowest common ancestor of A and B in the spanning tree (which we will denote as C) is neither A nor B. Evidently, C is visited before either A or B, and A and B lie in different subtrees rooted at immediate children of C. So at some point C has been visited but none of its descendants have been visited. Now, because DFS always visits an entire subtree of the current vertex before backtracking and entering another subtree, the entire subtree containing either A or B (whichever one is visited first) must be explored before visiting the subtree containing the other. Suppose the subtree containing A is visited first. Then, when A is visited, B has not yet been visited, and thus the edge between them must be a tree edge, which is a contradiction. Notice that given some graph, it is not possible to assign each edge a fixed classification. For example, in the complete graph of two vertices (K_2), two of the (undirected) edges will be tree/parent, and the other one will be down/back, but which are which depends on the starting vertex and the order in which vertices are visited. Thus, when we say that a particular edge in the graph is a tree edge (or a parent, down, or back link), it is assumed that we are referring to the classification induced by a specific invocation of DFS on that graph. The beauty of DFS-based algorithms is that any valid edge classification (that is, edge classification produced by any correct invocation of DFS) is useful. Directed graphs[edit] In directed graphs, edges are also classified into four categories. Tree edges, back edges, and down edges are classified just as they are in undirected graphs. There are no longer parent edges. If an tree edge exists from A to B, this does not guarantee the existence of an edge from B to A; if this edge does exist, it is considered a back edge, since it points from a node to its ancestor in the spanning tree. However, it is not true in a directed graph that every edge must be either tree, back, or down. In the sketch proof given for the corresponding fact in an undirected graph, we use the fact that it doesn't matter if A or B is visited first, since edges are bidirectional. But in the same circumstances in a directed graph, if there is an edge from A to B but no edge from B to A, and the subtree containing B is visited first, then the edge A-B is not a tree edge, since B has already been visited at the time that A is visited. Thus, when performing DFS on a directed graph, it is possible for an edge to lead from a vertex to another vertex which is neither a descendant nor an ancestor. Such an edge is called a cross edge. Again, these labels become meaningful only after DFS has been run. Traversal of the spanning tree[edit] In order to classify edges as tree, parent, down, back, or cross, we need to know which nodes are descendants of which. In effect, we would like to be able to answer the question "what is the relationship between nodes X and Y?", preferably in constant time per query (giving an optimal algorithm). The simplest way to do this, which is useful for other algorithms as well, is by preorder and postorder numbering of the nodes. In these numberings, nodes are assigned consecutive increasing integers (usually starting from 0 or 1). In preorder, a node is numbered as soon as it is visited: that is, the numbering of a node occurs pre-recursion, before any other nodes are visited. In postorder, a node is numbered after all the recursive calls arising from tree edges out of it have terminated - it occurs post-recursion. Here is code that assigns both preordering and postordering numbers at once. (Notice that we must visit vertices in the same order when assigning preordering and postordering numbers, otherwise the classification scheme for edges in a directed graph will fail. The best way to ensure that they are visited in the same order in both DFSs is by combining them into a single DFS.) function DFS(u) if pre[u] = ∞ pre[u] = precnt precnt = precnt + 1 for each v ∈ V(G) such that (u,v) ∈ E(G) DFS(v) post[u] = postcnt postcnt = postcnt + 1 input G for each u ∈ V(G) let pre[u] = ∞ let post[u] = ∞ let precnt=0 let postcnt=0 for each u ∈ V(G) DFS(u) Before a node has had a preorder or postorder number assigned, its number is ∞. At the termination of the algorithm, precnt=postcnt=|V|. Classification of edges in undirected graphs[edit] In an undirected graph, preorder numbers may be used to distinguish between down edges and back edges. (It becomes obvious when an edge is a tree edge or a parent edge, while running DFS, of course.) If node A has a higher preorder number than node B, then B is an ancestor and the edge is a back edge. Otherwise, the edge is a down edge. Classification of edges in directed graphs[edit] The situation is more complicated in directed graphs. Again it is obvious when an edge is a tree edge. Otherwise, consider the other possible classifications for an edge from A to B. If A is a descendant of B (back edge), then A has a higher preorder number but lower postorder number than B. If A is an ancestor of B (down edge), then A has a lower preorder number but higher postorder number than B. Finally, if A is neither an ancestor nor a descendant of B (cross edge), then an edge doesn't exist from B to A and B was visited before A, so A has a higher preorder number and a higher postorder number. Note that we do not actually know the postorder number for a vertex until after we have examined all the edges leaving it. This is not a problem for the sake of classification, since we know that any number that hasn't been assigned yet is higher than any number that has been so far assigned, and that using ∞ as a placeholder therefore doesn't change the results of comparisons. You could also use -1 as a placeholder, but you would have to modify the code somewhat. Topological sort[edit] A topological ordering of a DAG can be obtained by reversing the postordering. A proof of this fact is given in the Topological sorting article. Applications[edit] DFS, by virtue of being a subcategory of graph search, is able to find connected components, paths (including augmenting paths), spanning trees, and topological orderings in linear time. The special properties of DFS trees also make it uniquely suitable for finding biconnected components (or bridges and articulation points) and strongly connected components. There are three well-known algorithms for the latter, those of Tarjan, Kosaraju, and Gabow. The classic algorithm for the former appears to have no unique name in the literature. The following example shows how to find connected components and a DFS forest (a collection of DFS trees, one for each connected component). function DFS(u,v) if id[v] = -1 id[v] = cnt let pred[v] = u for each w ∈ V(G) such that (v,w) ∈ E(G) DFS(v,w) input G cnt = 0 for each u ∈ V(G) let id[u] = -1 for each u ∈ V(G) if id[u] = -1 DFS(u,u) cnt = cnt + 1 After the termination of this algorithm, cnt will contain the total number of connected components, id[u] will contain an ID number indicating the connected component to which vertex u belongs (an integer in the range [0,cnt)) and pred[u] will contain the parent vertex of u in some spanning tree of the connected component to which u belongs, unless u is the first vertex visited in its component, in which case pred[u] will be u itself. Flood fill[edit] The flood fill problem is often used as an example application of DFS. It avoids some of the implementational complexity associated with graph-theoretic algorithms by using an implicitly defined graph; there is no need to construct an adjacency matrix or adjacency list or any other such structure. Instead, we can just identify each vertex—a square in the grid—by its row and column indices; and it is very easy to find the adjacent squares. This is what a typical recursive implementation looks like (C++): // (dx[i], dy[i]) represents one of the cardinal directions const int dx[4] = {0, 1, 0, -1}; const int dy[4] = {1, 0, -1, 0}; void do_flood_fill(vector<vector<int> >& grid, int r, int c, int oldval, int newval) { // invalid coordinates if (r < 0 || c < 0 || r >= grid.size() || c >= grid[0].size()) return; // outside the blob if (grid[r][c] != oldval) return; grid[r][c] = newval; for (int i = 0; i < 4; i++) do_flood_fill(grid, r + dx[i], c + dy[i], oldval, newval); } void flood_fill(vector<vector<int> >& grid, int r, int c, int newval) { if (grid[r][c] != newval) do_flood_fill(grid, r, c, grid[r][c], newval); } References[edit]
__label__pos
0.980097
Properties Label 4-3920e2-1.1-c1e2-0-1 Degree $4$ Conductor $15366400$ Sign $1$ Analytic cond. $979.774$ Root an. cond. $5.59476$ Motivic weight $1$ Arithmetic yes Rational yes Primitive no Self-dual yes Analytic rank $0$ Origins Origins of factors Downloads Learn more Normalization:   Dirichlet series L(s)  = 1  + 2·3-s − 2·5-s − 9-s + 6·11-s − 6·13-s − 4·15-s + 2·17-s + 12·19-s − 12·23-s + 3·25-s − 6·27-s − 6·29-s + 12·31-s + 12·33-s − 4·37-s − 12·39-s + 4·41-s − 4·43-s + 2·45-s − 6·47-s + 4·51-s − 12·55-s + 24·57-s + 4·59-s + 12·65-s + 8·67-s − 24·69-s + ⋯ L(s)  = 1  + 1.15·3-s − 0.894·5-s − 1/3·9-s + 1.80·11-s − 1.66·13-s − 1.03·15-s + 0.485·17-s + 2.75·19-s − 2.50·23-s + 3/5·25-s − 1.15·27-s − 1.11·29-s + 2.15·31-s + 2.08·33-s − 0.657·37-s − 1.92·39-s + 0.624·41-s − 0.609·43-s + 0.298·45-s − 0.875·47-s + 0.560·51-s − 1.61·55-s + 3.17·57-s + 0.520·59-s + 1.48·65-s + 0.977·67-s − 2.88·69-s + ⋯ Functional equation \[\begin{aligned}\Lambda(s)=\mathstrut & 15366400 ^{s/2} \, \Gamma_{\C}(s)^{2} \, L(s)\cr =\mathstrut & \, \Lambda(2-s) \end{aligned}\] \[\begin{aligned}\Lambda(s)=\mathstrut & 15366400 ^{s/2} \, \Gamma_{\C}(s+1/2)^{2} \, L(s)\cr =\mathstrut & \, \Lambda(1-s) \end{aligned}\] Invariants Degree: \(4\) Conductor: \(15366400\)    =    \(2^{8} \cdot 5^{2} \cdot 7^{4}\) Sign: $1$ Analytic conductor: \(979.774\) Root analytic conductor: \(5.59476\) Motivic weight: \(1\) Rational: yes Arithmetic: yes Character: induced by $\chi_{3920} (1, \cdot )$ Primitive: no Self-dual: yes Analytic rank: \(0\) Selberg data: \((4,\ 15366400,\ (\ :1/2, 1/2),\ 1)\) Particular Values \(L(1)\) \(\approx\) \(3.192341266\) \(L(\frac12)\) \(\approx\) \(3.192341266\) \(L(\frac{3}{2})\) not available \(L(1)\) not available Euler product    \(L(s) = \displaystyle \prod_{p} F_p(p^{-s})^{-1} \) $p$$\Gal(F_p)$$F_p(T)$ bad2 \( 1 \) 5$C_1$ \( ( 1 + T )^{2} \) 7 \( 1 \) good3$D_{4}$ \( 1 - 2 T + 5 T^{2} - 2 p T^{3} + p^{2} T^{4} \) 11$C_2^2$ \( 1 - 6 T + 23 T^{2} - 6 p T^{3} + p^{2} T^{4} \) 13$D_{4}$ \( 1 + 6 T + 33 T^{2} + 6 p T^{3} + p^{2} T^{4} \) 17$D_{4}$ \( 1 - 2 T + p T^{2} - 2 p T^{3} + p^{2} T^{4} \) 19$C_2$ \( ( 1 - 6 T + p T^{2} )^{2} \) 23$D_{4}$ \( 1 + 12 T + 80 T^{2} + 12 p T^{3} + p^{2} T^{4} \) 29$D_{4}$ \( 1 + 6 T + 35 T^{2} + 6 p T^{3} + p^{2} T^{4} \) 31$D_{4}$ \( 1 - 12 T + 80 T^{2} - 12 p T^{3} + p^{2} T^{4} \) 37$D_{4}$ \( 1 + 4 T + 60 T^{2} + 4 p T^{3} + p^{2} T^{4} \) 41$D_{4}$ \( 1 - 4 T + 68 T^{2} - 4 p T^{3} + p^{2} T^{4} \) 43$C_2$ \( ( 1 + 2 T + p T^{2} )^{2} \) 47$D_{4}$ \( 1 + 6 T + 85 T^{2} + 6 p T^{3} + p^{2} T^{4} \) 53$C_2^2$ \( 1 + 88 T^{2} + p^{2} T^{4} \) 59$D_{4}$ \( 1 - 4 T + 104 T^{2} - 4 p T^{3} + p^{2} T^{4} \) 61$C_2^2$ \( 1 + 114 T^{2} + p^{2} T^{4} \) 67$D_{4}$ \( 1 - 8 T + 132 T^{2} - 8 p T^{3} + p^{2} T^{4} \) 71$C_4$ \( 1 - 12 T + 170 T^{2} - 12 p T^{3} + p^{2} T^{4} \) 73$C_2^2$ \( 1 + 74 T^{2} + p^{2} T^{4} \) 79$D_{4}$ \( 1 - 14 T + 135 T^{2} - 14 p T^{3} + p^{2} T^{4} \) 83$C_2$ \( ( 1 + p T^{2} )^{2} \) 89$C_2$ \( ( 1 - 8 T + p T^{2} )^{2} \) 97$D_{4}$ \( 1 + 18 T + 257 T^{2} + 18 p T^{3} + p^{2} T^{4} \) show more show less    \(L(s) = \displaystyle\prod_p \ \prod_{j=1}^{4} (1 - \alpha_{j,p}\, p^{-s})^{-1}\) Imaginary part of the first few zeros on the critical line −8.441442790812856523883432212045, −8.256325891091565577939245090735, −7.85065505292972220561530559851, −7.82713484789237901041498766749, −7.24035620634240993771593864480, −7.03679416669897953252115658571, −6.44770162826730772498000048947, −6.22515159726651672373737081452, −5.45995213551298798063887261836, −5.43603910847013014003027763175, −4.76043098823771432097158270741, −4.42309441543727187005896293854, −3.88483953880235867248184423100, −3.51597730171840436274601175836, −3.28239995162409282042144765754, −2.92303880235447586509322072964, −2.11145328900084220562022357581, −2.03756948013824219197765158176, −1.07790614661032914947477398655, −0.53499513789367496019032163026, 0.53499513789367496019032163026, 1.07790614661032914947477398655, 2.03756948013824219197765158176, 2.11145328900084220562022357581, 2.92303880235447586509322072964, 3.28239995162409282042144765754, 3.51597730171840436274601175836, 3.88483953880235867248184423100, 4.42309441543727187005896293854, 4.76043098823771432097158270741, 5.43603910847013014003027763175, 5.45995213551298798063887261836, 6.22515159726651672373737081452, 6.44770162826730772498000048947, 7.03679416669897953252115658571, 7.24035620634240993771593864480, 7.82713484789237901041498766749, 7.85065505292972220561530559851, 8.256325891091565577939245090735, 8.441442790812856523883432212045 Graph of the $Z$-function along the critical line
__label__pos
0.737377
A 3D model may be composed of a few different parts. We want to find all the independent parts. In the following code snippet, the interface GetConnectedCellIds can help us to find all the parts, DrawCells will mark a part which is represented by some cell ids red color. #include <iostream> #include <vtkSmartPointer.h> #include <vtkSphereSource.h> #include <vtkActor.h> #include <vtkConeSource.h> #include <vtkRenderer.h> #include <vtkRenderWindow.h> #include <vtkPolyDataMapper.h> #include <vtkRenderWindowInteractor.h> #include <vtkCellData.h> #include <vtkNamedColors.h> #include <vtkColorTransferFunction.h> #include <vtkTriangleFilter.h> #include <vtkXMLPolyDataReader.h> #include <vtkCharArray.h> #define vtkSPtr vtkSmartPointer #define vtkSPtrNew(Var, Type) vtkSPtr<Type> Var = vtkSPtr<Type>::New(); using namespace std; struct Edge { vtkIdType cellId; vtkIdType edgePt1; vtkIdType edgePt2; }; void DrawCells( vtkSPtr<vtkPolyData> pd, vtkSPtr<vtkPolyDataMapper> mapper, vtkSPtr<vtkIdList> visitedCellIds ) { vtkNew<vtkCharArray> cellTypes; cellTypes->SetNumberOfComponents( 1 ); for( int i = 0; i < pd->GetNumberOfCells(); ++i ) { cellTypes->InsertNextValue( 0 ); } for( int i = 0; i < visitedCellIds->GetNumberOfIds(); ++i ) { auto cellId = visitedCellIds->GetId( i ); cellTypes->InsertValue( cellId, 1 ); } pd->GetCellData()->SetScalars( cellTypes ); pd->GetCellData()->Modified(); mapper->SetScalarModeToUseCellData(); mapper->SetColorModeToMapScalars(); vtkNew<vtkColorTransferFunction> lut; lut->AddRGBPoint( 0, 1, 1, 1 ); lut->AddRGBPoint( 1, 0.8, 0, 0 ); mapper->SetLookupTable( lut ); } vtkSPtr<vtkIdList> GetConnectedCellIds( vtkPolyData *pd, vtkIdType seedCellId ) { vtkIdType nPts; vtkIdType *pts; pd->GetCellPoints( seedCellId, nPts, pts ); std::vector<Edge> currrentEdges; for( int i = 0; i < 3; ++i ) { Edge edge; edge.cellId = seedCellId; edge.edgePt1 = pts[i]; edge.edgePt2 = pts[ (i+1) % 3 ]; currrentEdges.push_back( edge ); } vtkNew<vtkIdList> visitedCellIds; visitedCellIds->InsertNextId( seedCellId ); std::vector<Edge> nextEdges; while ( currrentEdges.size() > 0 ) { for( int i = 0; i < currrentEdges.size(); ++i ) { Edge edge = currrentEdges[i]; vtkNew<vtkIdList> neighborCellIds; pd->GetCellEdgeNeighbors( edge.cellId, edge.edgePt1, edge.edgePt2, neighborCellIds ); for( int j = 0; j < neighborCellIds->GetNumberOfIds(); ++j ) { auto neiCellId = neighborCellIds->GetId( j ); if( -1 != visitedCellIds->IsId( neiCellId ) ) { continue; } pd->GetCellPoints( neiCellId, nPts, pts ); vtkIdType thirdPt = -1; for( int k = 0; k < 3; ++k ) { if( pts[k] != edge.edgePt1 && pts[k] != edge.edgePt2 ) { thirdPt = pts[k]; break; } } if( -1 == thirdPt ) { continue; } Edge edge1; edge1.cellId = neiCellId; edge1.edgePt1 = edge.edgePt1; edge1.edgePt2 = thirdPt; Edge edge2; edge2.cellId = neiCellId; edge2.edgePt1 = edge.edgePt2; edge2.edgePt2 = thirdPt; nextEdges.push_back( edge1 ); nextEdges.push_back( edge2 ); visitedCellIds->InsertNextId( neiCellId ); } } currrentEdges.swap( nextEdges ); nextEdges.clear(); } return visitedCellIds; } int main() { vtkSPtrNew( reader, vtkXMLPolyDataReader ); reader->SetFileName( "/Users/weiyang/Desktop/test.vtp" ); reader->Update(); auto *polyData = reader->GetOutput(); vtkSPtrNew( triangleFilter, vtkTriangleFilter ); triangleFilter->SetInputData( polyData ); triangleFilter->PassLinesOff(); triangleFilter->PassVertsOff(); triangleFilter->Update(); polyData = triangleFilter->GetOutput(); vtkSPtrNew( mapper, vtkPolyDataMapper ); mapper->SetInputData( polyData ); vtkSPtrNew( actor, vtkActor ); actor->SetMapper( mapper ); // ============== start to split ================= polyData->BuildCells(); polyData->BuildLinks(); vtkIdType pdCellCount = polyData->GetNumberOfCells(); vtkIdType visCellCount = 0; std::vector<bool> markCellIds; std::vector< vtkSPtr<vtkIdList> > visCellIdsList; for( int i = 0; i < pdCellCount; ++i ) { markCellIds.push_back( false ); } while ( visCellCount != pdCellCount ) { vtkIdType seedCellId = -1; for( int i = 0; i < pdCellCount; ++i ) { if( !markCellIds[i] ) { seedCellId = i; } } if( -1 == seedCellId ) { break; } vtkSPtr<vtkIdList> visitedCells = GetConnectedCellIds( polyData, seedCellId ); visCellCount += visitedCells->GetNumberOfIds(); visCellIdsList.push_back( visitedCells ); for( int i = 0; i < visitedCells->GetNumberOfIds(); ++i ) { auto id = visitedCells->GetId( i ); markCellIds[id] = true; } } cout << visCellIdsList.size() << endl; DrawCells( polyData, mapper, visCellIdsList[1] ); // ============== finish: split ================= vtkSPtrNew( renderer, vtkRenderer ); renderer->AddActor(actor); renderer->SetBackground( 0, 0, 0 ); vtkSPtrNew( renderWindow, vtkRenderWindow ); renderWindow->AddRenderer( renderer ); vtkSPtrNew( renderWindowInteractor, vtkRenderWindowInteractor ); renderWindowInteractor->SetRenderWindow( renderWindow ); renderer->ResetCamera(); renderWindow->Render(); renderWindowInteractor->Start(); return 0; } Categories: VTK 0 0 votes Article Rating Subscribe Notify of guest 0 Comments Inline Feedbacks View all comments XOR Strings : Input your strings, the tool can encrypt and decrypt them. X 0 Would love your thoughts, please comment.x () x
__label__pos
0.973291
Documentation This is machine translation Translated by Microsoft Mouseover text to see original. Click the button below to return to the English version of the page. Note: This page has been translated by MathWorks. Please click here To view all translated materials including this page, select Japan from the country navigator on the bottom of this page. ismembertol Members of set within tolerance Syntax LIA = ismembertol(A,B,tol) LIA = ismembertol(A,B) [LIA,LocB] = ismembertol(___) [___] = ismembertol(___,Name,Value) Description example LIA = ismembertol(A,B,tol) returns an array containing logical 1 (true) where the elements of A are within tolerance of the elements in B. Otherwise, the array contains logical 0 (false). Two values, u and v, are within tolerance if abs(u-v) <= tol*max(abs([A(:);B(:)])) That is, ismembertol scales the tol input based on the magnitude of the data. ismembertol is similar to ismember. Whereas ismember performs exact comparisons, ismembertol performs comparisons using a tolerance. example LIA = ismembertol(A,B) uses a default tolerance of 1e-6 for single-precision inputs and 1e-12 for double-precision inputs. example [LIA,LocB] = ismembertol(___) also returns an array, LocB, that contains the index location in B for each element in A that is a member of B. You can use any of the input arguments in previous syntaxes. example [___] = ismembertol(___,Name,Value) uses additional options specified by one or more Name-Value pair arguments using any of the input or output argument combinations in previous syntaxes. For example, ismembertol(A,B,'ByRows',true) compares the rows of A and B and returns a logical column vector. Examples collapse all Create a vector, x. Obtain a second vector, y, by transforming and untransforming x. This transformation introduces round-off differences to y. x = (1:6)'*pi; y = 10.^log10(x); Verify that x and y are not identical by taking the difference. x-y ans = 1.0e-14 * 0.0444 0 0 0 0 -0.3553 Use ismember to find the elements of x that are in y. The ismember function performs exact comparisons and determines that some of the matrix elements in x are not members of y. lia = ismember(x,y) lia = 6x1 logical array 0 1 1 1 1 0 Use ismembertol to perform the comparison using a small tolerance. ismembertol treats elements that are within tolerance as equal and determines that all of the elements in x are members of y. LIA = ismembertol(x,y) LIA = 6x1 logical array 1 1 1 1 1 1 By default, ismembertol looks for elements that are within tolerance, but it also can find rows of a matrix that are within tolerance. Create a numeric matrix, A. Obtain a second matrix, B, by transforming and untransforming A. This transformation introduces round-off differences to B. A = [0.05 0.11 0.18; 0.18 0.21 0.29; 0.34 0.36 0.41; ... 0.46 0.52 0.76; 0.82 0.91 1.00]; B = log10(10.^A); Use ismember to find the rows of A that are in B. ismember performs exact comparisons and thus determines that most of the rows in A are not members of B, even though some of the rows differ by only a small amount. lia = ismember(A,B,'rows') lia = 5x1 logical array 0 0 0 0 1 Use ismembertol to perform the row comparison using a small tolerance. ismembertol treats rows that are within tolerance as equal and thus determines that all of the rows in A are members of B. LIA = ismembertol(A,B,'ByRows',true) LIA = 5x1 logical array 1 1 1 1 1 Create two vectors of random numbers and determine which values in A are also members of B, using a tolerance. Specify OutputAllIndices as true to return all of the indices for the elements in B that are within tolerance of the corresponding elements in A. rng(5) A = rand(1,15); B = rand(1,5); [LIA,LocAllB] = ismembertol(A,B,0.2,'OutputAllIndices',true) LIA = 1x15 logical array 1 0 1 0 1 1 1 1 1 1 0 1 1 1 0 LocAllB = 1x15 cell array Columns 1 through 5 {2x1 double} {[0]} {2x1 double} {[0]} {3x1 double} Columns 6 through 10 {2x1 double} {[4]} {3x1 double} {3x1 double} {2x1 double} Columns 11 through 15 {[0]} {2x1 double} {4x1 double} {2x1 double} {[0]} Find the average value of the elements in B that are within tolerance of the value A(13). The cell LocAllB{13} contains all the indices for elements in B that are within tolerance of A(13). A(13) ans = 0.4413 allB = B(LocAllB{13}) allB = 0.2741 0.4142 0.2961 0.5798 aveB = mean(allB) aveB = 0.3911 By default, ismembertol uses a tolerance test of the form abs(u-v) <= tol*DS, where DS automatically scales based on the magnitude of the input data. You can specify a different DS value to use with the DataScale option. However, absolute tolerances (where DS is a scalar) do not scale based on the magnitude of the input data. First, compare two small values that are a distance eps apart. Specify tol and DS to make the within tolerance equation abs(u-v) <= 10^-6. x = 0.1; ismembertol(x, exp(log(x)), 10^-6, 'DataScale', 1) ans = logical 1 Next, increase the magnitude of the values. The round-off error in the calculation exp(log(x)) is proportional to the magnitude of the values, specifically to eps(x). Even though the two large values are a distance eps from one another, eps(x) is now much larger. Therefore, 10^-6 is no longer a suitable tolerance. x = 10^10; ismembertol(x, exp(log(x)), 10^-6, 'DataScale', 1) ans = logical 0 Correct this issue by using the default (scaled) value of DS. Y = [0.1 10^10]; ismembertol(Y, exp(log(Y))) ans = 1x2 logical array 1 1 Create a set of random 2-D points, and then use ismembertol to group the points into vertical bands that have a similar (within-tolerance) x-coordinate to a small set of query points, B. Use these options with ismembertol: • Specify ByRows as true, since the point coordinates are in the rows of A and B. • Specify OutputAllIndices as true to return the indices for all points in A that have an x-coordinate within tolerance of the query points in B. • Specify DataScale as [1 Inf] to use an absolute tolerance for the x-coordinate, while ignoring the y-coordinate. A = rand(1000,2); B = [(0:.2:1)',0.5*ones(6,1)]; [LIA,LocAllB] = ismembertol(B, A, 0.1, 'ByRows', true, ... 'OutputAllIndices', true, 'DataScale', [1,Inf]) LIA = 6x1 logical array 1 1 1 1 1 1 LocAllB = 6x1 cell array { 94x1 double} {223x1 double} {195x1 double} {212x1 double} {187x1 double} { 89x1 double} Plot the points in A that are within tolerance of each query point in B. hold on plot(B(:,1),B(:,2),'x') for k = 1:length(LocAllB) plot(A(LocAllB{k},1), A(LocAllB{k},2),'.') end Input Arguments collapse all Query array, specified as a scalar, vector, matrix, or multidimensional array. Inputs A and B must be full. If you specify the ByRows option, then A and B must have the same number of columns. Data Types: single | double Query array, specified as a scalar, vector, matrix, or multidimensional array. Inputs A and B must be full. If you specify the ByRows option, then A and B must have the same number of columns. Data Types: single | double Comparison tolerance, specified as a positive real scalar. ismembertol scales the tol input using the maximum absolute values in the input arrays A and B. Then ismembertol uses the resulting scaled comparison tolerance to determine which elements in A are also a member of B. If two elements are within tolerance of each other, then ismembertol considers them to be equal. Two values, u and v, are within tolerance if abs(u-v) <= tol*max(abs([A(:);B(:)])). To specify an absolute tolerance, specify both tol and the 'DataScale' Name-Value pair. Example: tol = 0.05 Example: tol = 1e-8 Example: tol = eps Data Types: single | double Name-Value Pair Arguments Specify optional comma-separated pairs of Name,Value arguments. Name is the argument name and Value is the corresponding value. Name must appear inside single quotes (' '). You can specify several name and value pair arguments in any order as Name1,Value1,...,NameN,ValueN. Example: LIA = ismembertol(A,B,'ByRows',true) collapse all Output index type, specified as the comma-separated pair consisting of 'OutputAllIndices' and either false (default), true, 0, or 1. ismembertol interprets numeric 0 as false and numeric 1 as true. When OutputAllIndices is true, the ismembertol function returns the second output, LocB, as a cell array. The cell array contains the indices for all elements in B that are within tolerance of the corresponding value in A. That is, each cell in LocB corresponds to a value in A, and the values in each cell correspond to locations in B. Example: [LIA,LocAllB] = ismembertol(A,B,tol,'OutputAllIndices',true) Row comparison toggle, specified as the comma-separated pair consisting of 'ByRows' and either false (default), true, 0, or 1. ismembertol interprets numeric 0 as false and numeric 1 as true. Use this option to find rows in A and B that are within tolerance. When ByRows is true: • ismembertol compares the rows of A and B by considering each column separately. Thus, A and B must be 2-D arrays with the same number of columns. • If the corresponding row in A is within tolerance of a row in B, then LIA contains logical 1 (true). Otherwise, it contains logical 0 (false). Two rows, u and v, are within tolerance if all(abs(u-v) <= tol*max(abs([A;B]))). Example: LIA = ismembertol(A,B,tol,'ByRows',true) Scale of data, specified as the comma-separated pair consisting of 'DataScale' and either a scalar or vector. Specify DataScale as a numeric scalar, DS, to change the tolerance test to be, abs(u-v) <= tol*DS. When used together with the ByRows option, the DataScale value also can be a vector. In this case, each element of the vector specifies DS for a corresponding column in A. If a value in the DataScale vector is Inf, then ismembertol ignores the corresponding column in A. Example: LIA = ismembertol(A,B,'DataScale',1) Example: [LIA,LocB] = ismembertol(A,B,'ByRows',true,'DataScale',[eps(1) eps(10) eps(100)]) Data Types: single | double Output Arguments collapse all Logical index to A, returned as a vector or matrix containing logical 1 (true) wherever the elements (or rows) in A are members of B (within tolerance). Elsewhere, LIA contains logical 0 (false). LIA is the same size as A, unless you specify the ByRows option. In that case, LIA is a column vector with the same number of rows as A. Locations in B, returned as a vector, matrix, or cell array. LocB contains the indices to the elements (or rows) in B that are found in A (within tolerance). LocB contains 0 wherever an element in A is not a member of B. If OutputAllIndices is true, then ismembertol returns LocB as a cell array. The cell array contains the indices for all elements in B that are within tolerance of the corresponding value in A. That is, each cell in LocB corresponds to a value in A, and the values in each cell correspond to locations in B. LocB is the same size as A, unless you specify the ByRows option. In that case, LocB is a column vector with the same number of rows as A. Introduced in R2015a Was this topic helpful?
__label__pos
0.897867
WP-ShortStat是数据库空间杀手   很奇怪数据库空间为什么一直在快速增加,今天备份数据库的时候发现数据库已经到35M了,把备份文件看了下,发现wp_ss_stats表几乎占据了整个数据库的使用量,删除这个表的数据后,数据库空间下降到只有1M多一点,呵呵,果断的停用了这个插件,还是使用外部的访问统计好了。我想WP-ShortStat可能只适用与每天只有几个点击量的Blog?我的这个Blog的访问量也不是很大,每天的点击数最多300(Google Analytics统计数据),三个多月的数据量就有35M,要是一天的点击量上千,那还不是早暴了? 这是找到的鸟文资料,o(︶︿︶)o 唉 In this article, I present a simple method for dramatically decreasing the size of your WordPress database by partially emptying old data from the WP-ShortStat table via the following SQL command: DELETE FROM `wp_ss_stats` ORDER BY `id` ASC LIMIT n That is the point of this entire article, which dips into just about everything one might need to know before employing such strategy. If you are familiar with SQL and understand the purpose and functionality of this command, feel free to grab, gulp and go. Otherwise, read on for the full story.. A little context, please.. Many WordPress users enjoy the convenient statistics provided by one of theexcellent ShortStat plugins. WP-ShortStat keeps track of many essential types of data: recent referrers, search queries, referring domains, keywords, locations, browsers, and many more. Over time, the copious amount of statistical data collected by WP-ShortStat increasingly inflates the size of your WordPress database.   For example, before installing WP-ShortStat, my WP database was around 8 megabytes in size. After installing ShortStat and using it for several months, the size of my database ballooned to well over 30 megabytes! WP-ShortStat uses these data sets to calculate cumulative totals, such as total hits, daily visits, referral counts, and various others. Fortunately, I supplement the functionality of ShortStat with Mint, which provides all the statistical data I will ever need — without bloating the size of my WordPress database (Note: the Mint database is maintained independently of the WP database). Thus, my statistical strategy involves using Mint to record permanent, long-term data, while simultaneously using WP-ShortStat to track daily hits, referrals, and visitors 1. At the end of the day, the ShortStat data is completely expendable and fails to justify its behemoth table size. Nevertheless, even without running a secondary statistical package to supplement WP-ShortStat, many users would gladly trade their running tally of cumulative statistical data for a more lightweight and agile WP database. If this sounds like you, or if you really don’t care about all this long-winded nonsense, here is a simple way to reduce the overall size of your WordPress database by cleaning up your WP-ShortStat table. Getting on with it, then.. Thanks to an informative discussion with Mark from c77studios.com, a single SQLcommand was forged to quickly and efficiently remove old data from the WP-ShortStat database table: DELETE FROM `wp_ss_stats` ORDER BY `id` ASC LIMIT n As discussed in the original conversation, this command is a generalized SQL query that is independent of table data such as ID. Execution of this command will effectively delete the oldest n number of records from the WP-ShortStat table,wp_ss_stats. Further, the generalized syntax of this command enables us to automate the procedure at specified intervals via cron job, PHP, etc. I use this snippet of SQL to shrink my WordPress database every month or so (kind of like shaving). Rather than try to explain the overall effect of this command, here are a few screenshots showing my WP-ShortStat panel and its associated wp_ss_statstable at various points throughout the process 1: I began with this hellishly sized table, which was over 180,000 rows thick and over 45,000 Kilobytes in size: [ Screenshot: showing the relatively immense size of the WP-ShortStat table ] Two chop the table size in half, I divided the total number of records by two and entered the following query: [ Screenshot: showing the SQL query discussed in this article with a value of 97257 ] After executing the query, the oldest 97,257 records were removed from the table, effectively reducing its size by roughly half: [ Screenshot: showing the halving effect of the previously discussed SQL query ] Within the WordPress Admin area, here is a portion of the WP-ShortStat panel beforedeleting any data: [ Screenshot: showing a portion of the WP-ShortStat panel before data removal ] After executing the query, the cumulative totals have been decreased by roughly half (as expected), while the most recent data remains intact: [ Screenshot: showing a portion of the WP-ShortStat panel after data removal ] Conclusion As you can see, this simple query serves as an effective tool at reducing the the overall size of your WordPress database. Especially if you are fronting from a limited hosting plan, this trick may enable you to run one of the WP-ShortStat plugins without exceeding any limits. Keep in mind that, depending on how many records you decide to dump (based on the value of n), the amount of data that remains for various daily totals may also be affected. If you are unsure as to which value to use for n, determine the total number of entries via the cardinality value and then try using a relatively small n value; run the query, examine the results, and then try a slightly larger number. If you remember to backup your wp_ss_stats table before beginning, it is totally safe to experiment with different values for n until you determine the optimum value. If you happen to delete too much data from your table, simply restore the backup and try again. Either way, I wouldn’t stress too much — the table will eventually be refilled with more data 😉 欢迎大佬支持本博客的发展 -- Donate -- 本文链接:WP-ShortStat是数据库空间杀手 转载声明:本站文章若无特别说明,皆为原创,转载请注明来源:三十岁,谢谢!^^ 分享到:           1. 没有评论 1. 没有通告
__label__pos
0.864902
Contenu | Rechercher | Menus Annonce L'équipe des administrateurs et modérateurs du forum vous invite à prendre connaissance des nouvelles règles. En cas de besoin, vous pouvez intervenir dans cette discussion. N'oubliez pas de cocher la case « Ajustement pour l'heure d'été » dans votre profil. Ubuntu 16.04 LTS Commandez vos DVD et clés USB Ubuntu-fr ! Pour en savoir un peu plus sur l'équipe du forum. Si vous avez des soucis pour rester connecté, déconnectez-vous puis reconnectez-vous depuis ce lien en cochant la case Me connecter automatiquement lors de mes prochaines visites. #1 Le 08/06/2008, à 13:10 karch Gmail notifier : ajout de la langue "francais" J'ai rajouté le francais dans les langues du Gmail Notifier Le fichier langs.xml est a remplacer dans : /usr/share/apps/gmail-notify/langs.xml A télécharger ici : http://karch.fr.tc/gmail_notifier/langs.xml Ou alors ajouter ceci apres le premier <langs> du fichier  /usr/share/apps/gmail-notify/langs.xml     <lang name="Francais">         <string id="1" >Configuration</string>         <string id="2" >Utilisateur:</string>         <string id="3" >Navigateur:</string>         <string id="4" >Langue:</string>         <string id="5" >Voulez vous vraiment quitter Gmail Notifier?</string>         <string id="6" >Information Quota</string>         <string id="24">%(u)s utilisé, %(t)s total (%(p)s)</string>         <string id="9" >_Vérifier maintenant</string>         <string id="10">_Quota info...</string>         <string id="11">_Configuration...</string>         <string id="12">_Quitter</string>         <string id="13">Connexion...</string>         <string id="14">Connecté</string>         <string id="15">Connexion échouée</string>         <string id="16">Connexion à la messagerie échouée, ré-essai</string>         <string id="17">Nouveau message de </string>         <string id="18">Pas de messages non-lus</string>         <string id="19">%(u)d Messages non-lu%(s)s</string>         <string id="20">Connexion...</string>         <string id="21">Gmail Notifier</string>         <string id="22">Mot de passe:</string>         <string id="23">Aller à la messagerie...</string>         <string id="25">Vérification des messages échoué, ré-essai</string>         <string id="26">Avancé >></string>         <string id="27">Décalage horizontal :</string>         <string id="28">Décalage vertical :</string>         <string id="29">Vitesse d'animation:</string>         <string id="30">Délai des Popups:</string>         <string id="31">Intervalle de vérification:</string>         <string id="32">Vitesse de connexion</string>         <string id="33">Des valeurs manques</string>         <string id="34">Sauver Utilisateur/Mot de passe</string>         <string id="35">Proxy</string>     </lang> Voili voilou smile Dernière modification par karch (Le 08/06/2008, à 13:29) Hors ligne #2 Le 28/05/2009, à 13:06 rwaan Re : Gmail notifier : ajout de la langue "francais" bonjour j'essaie d'inclure la partie français au fichier mais au moment d'enregistrer, j'ai le droit à : "Impossible d'enregistrer le fichier /usr/share/apps/gmail-notify/langs.xml. Vous n'avez pas les permissions nécessaires pour enregistrer ce fichier. Vérifiez que vous avez saisi l'emplacement correctement et réessayez." comme c'est mon 2e jour sur ubuntu, je sollicite votre aide roll Hors ligne #3 Le 28/05/2009, à 20:47 omega13 Re : Gmail notifier : ajout de la langue "francais" Essaye en ouvrantle fichier en mode root (gksudo gedit /usr/share/apps/gmail-notify/langs.xml) ca devrai solutionner ton probleme. N'oublie jamais que par defaut, pour une question de sécurité tu n'as pas le droit de modification sur les fichiers systemes Gnome Ubuntu 16.04 Hors ligne #4 Le 29/03/2011, à 13:35 snipe2004 Re : Gmail notifier : ajout de la langue "francais" Bon bin cette méthode ne marche plus pour la version 0.10.1. Petite mise à jour par bibi : remplacer les fichiers .xml situés dans /usr/share/gnome-gmail-notifier par ceux-ci. • ggn-prefs.xml <?xml version="1.0"?> <!--*- mode: xml -*--> <interface> <object class="GtkAdjustment" id="adjustment1"> <property name="upper">60</property> <property name="lower">1</property> <property name="page_increment">0</property> <property name="step_increment">1</property> <property name="page_size">0</property> <property name="value">10</property> </object> <object class="GtkWindow" id="GgnPrefsWindow"> <property name="border_width">4</property> <property name="title" translatable="yes">Préférences de Gmail-Notifier</property> <property name="resizable">False</property> <property name="window_position">GTK_WIN_POS_CENTER</property> <property name="icon">ggn-normal-lg.svg</property> <signal handler="ggn_prefs_window_deleted" name="delete_event"/> <signal handler="ggn_prefs_window_key_pressed" name="key_press_event"/> <child> <object class="GtkVBox" id="vboxMain"> <property name="visible">True</property> <property name="spacing">6</property> <child> <object class="GtkFrame" id="fraAccounts"> <property name="visible">True</property> <property name="label_xalign">0</property> <property name="shadow_type">GTK_SHADOW_NONE</property> <child> <object class="GtkAlignment" id="alnFraAccounts"> <property name="visible">True</property> <property name="left_padding">12</property> <child> <object class="GtkVBox" id="vboxAccounts"> <property name="visible">True</property> <property name="border_width">4</property> <child> <object class="GtkHBox" id="hboxAccounts"> <property name="visible">True</property> <property name="spacing">2</property> <child> <object class="GtkScrolledWindow" id="scrWinAccounts"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="hscrollbar_policy">GTK_POLICY_NEVER</property> <property name="vscrollbar_policy">GTK_POLICY_AUTOMATIC</property> <property name="shadow_type">GTK_SHADOW_IN</property> <child> <object class="GtkTreeView" id="treeAccounts"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="headers_visible">False</property> <property name="enable_search">False</property> <signal handler="ggn_prefs_window_account_activated" name="row_activated"/> </object> </child> </object> </child> <child> <object class="GtkVBox" id="vboxAccountButtons"> <property name="visible">True</property> <property name="border_width">2</property> <property name="spacing">2</property> <child> <object class="GtkButton" id="btnAccountAdd"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="label">Ajouter</property> <property name="use_stock">True</property> <signal handler="ggn_prefs_window_account_add" name="clicked"/> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> </packing> </child> <child> <object class="GtkButton" id="btnAccountDel"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="label">Effacer</property> <property name="use_stock">True</property> <signal handler="ggn_prefs_window_account_del" name="clicked"/> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> <property name="position">1</property> </packing> </child> <child> <object class="GtkButton" id="btnAccountEdit"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="label">Éditer</property> <property name="use_stock">True</property> <signal handler="ggn_prefs_window_account_edit" name="clicked"/> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> <property name="position">2</property> </packing> </child> </object> <packing> <property name="expand">False</property> <property name="position">1</property> </packing> </child> </object> </child> </object> </child> </object> </child> <child type="label"> <object class="GtkLabel" id="lblFraAccounts"> <property name="visible">True</property> <property name="label" translatable="yes">&lt;b&gt;Comptes Gmail&lt;/b&gt;</property> <property name="use_markup">True</property> </object> </child> </object> </child> <child> <object class="GtkFrame" id="fraUpdates"> <property name="visible">True</property> <property name="label_xalign">0</property> <property name="shadow_type">GTK_SHADOW_NONE</property> <child> <object class="GtkAlignment" id="alnFraUpdates"> <property name="visible">True</property> <property name="left_padding">12</property> <child> <object class="GtkVBox" id="vboxUpdates"> <property name="visible">True</property> <property name="border_width">4</property> <child> <object class="GtkLabel" id="lblUpdates"> <property name="visible">True</property> <property name="xalign">0</property> <property name="label" translatable="yes">Vérifier l'arrivée d'e-mails toutes les :</property> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> </packing> </child> <child> <object class="GtkHScale" id="slideUpdates"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="adjustment">adjustment1</property> <property name="digits">0</property> <property name="value_pos">GTK_POS_BOTTOM</property> <signal handler="ggn_prefs_window_rate_changed" name="value_changed"/> <signal handler="ggn_prefs_window_rate_format" name="format_value"/> </object> <packing> <property name="position">1</property> </packing> </child> </object> </child> </object> </child> <child type="label"> <object class="GtkLabel" id="lblFraUpdates"> <property name="visible">True</property> <property name="label" translatable="yes">&lt;b&gt;Actualisation&lt;/b&gt;</property> <property name="use_markup">True</property> </object> </child> </object> <packing> <property name="position">1</property> </packing> </child> <child> <object class="GtkFrame" id="fraNotes"> <property name="visible">True</property> <property name="label_xalign">0</property> <property name="shadow_type">GTK_SHADOW_NONE</property> <child> <object class="GtkAlignment" id="alnFraNotes"> <property name="visible">True</property> <property name="left_padding">12</property> <child> <object class="GtkVBox" id="vboxNotes"> <property name="visible">True</property> <property name="border_width">4</property> <child> <object class="GtkCheckButton" id="chkNotesMsgs"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="label" translatable="yes">Prévenir de l'arrivée de nouveaux e-mails</property> <property name="use_underline">True</property> <property name="active">True</property> <property name="draw_indicator">True</property> <signal handler="ggn_prefs_window_notify_msgs_toggled" name="toggled"/> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> </packing> </child> <child> <object class="GtkCheckButton" id="chkNotesErrs"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="label" translatable="yes">Prévenir en cas d'erreur</property> <property name="use_underline">True</property> <property name="draw_indicator">True</property> <signal handler="ggn_prefs_window_notify_errs_toggled" name="toggled"/> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> <property name="position">1</property> </packing> </child> </object> </child> </object> </child> <child type="label"> <object class="GtkLabel" id="lblFraNotes"> <property name="visible">True</property> <property name="label" translatable="yes">&lt;b&gt;Notifications visuelles&lt;/b&gt;</property> <property name="use_markup">True</property> </object> </child> </object> <packing> <property name="position">2</property> </packing> </child> <child> <object class="GtkFrame" id="fraSounds"> <property name="visible">True</property> <property name="label_xalign">0</property> <property name="shadow_type">GTK_SHADOW_NONE</property> <child> <object class="GtkAlignment" id="alnFraSounds"> <property name="visible">True</property> <property name="left_padding">12</property> <child> <object class="GtkVBox" id="vboxSounds"> <property name="visible">True</property> <property name="border_width">4</property> <child> <object class="GtkCheckButton" id="chkSounds"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="label" translatable="yes">Jouer un son à l'arrivée d'un nouvel e-mail</property> <property name="use_underline">True</property> <property name="active">True</property> <property name="draw_indicator">True</property> <signal handler="ggn_prefs_window_sound_enab_toggled" name="toggled"/> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> </packing> </child> <child> <object class="GtkHBox" id="hboxSounds"> <property name="visible">True</property> <property name="spacing">4</property> <child> <object class="GtkFileChooserButton" id="btnSoundsChoose"> <property name="visible">True</property> <property name="events">GDK_POINTER_MOTION_MASK | GDK_POINTER_MOTION_HINT_MASK | GDK_BUTTON_PRESS_MASK | GDK_BUTTON_RELEASE_MASK</property> <property name="title" translatable="yes">Choisir un fichier son</property> </object> </child> <child> <object class="GtkButton" id="btnSoundsPlay"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="label">Jouer</property> <property name="use_stock">True</property> <signal handler="ggn_prefs_window_test_sound" name="clicked"/> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> <property name="position">1</property> </packing> </child> </object> <packing> <property name="position">1</property> </packing> </child> </object> </child> </object> </child> <child type="label"> <object class="GtkLabel" id="lblFraSounds"> <property name="visible">True</property> <property name="label" translatable="yes">&lt;b&gt;Notifications audios&lt;/b&gt;</property> <property name="use_markup">True</property> </object> </child> </object> <packing> <property name="position">3</property> </packing> </child> <child> <object class="GtkHBox" id="hboxClose"> <property name="visible">True</property> <property name="border_width">2</property> <child> <placeholder/> </child> <child> <object class="GtkButton" id="btnClose"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="label">Fermer</property> <property name="use_stock">True</property> <signal handler="ggn_prefs_window_closed" name="clicked"/> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> <property name="pack_type">GTK_PACK_END</property> <property name="position">1</property> </packing> </child> </object> <packing> <property name="expand">False</property> <property name="position">4</property> </packing> </child> </object> </child> </object> </interface> • ggn-menu.xml <?xml version="1.0"?> <!--*- mode: xml -*--> <interface> <object class="GtkUIManager" id="uimanager1"> <child> <object class="GtkActionGroup" id="actiongroup1"> <child> <object class="GtkAction" id="itemCheck"> <property name="stock_id">gtk-connect</property> <property name="name">itemCheck</property> <property name="label" translatable="yes">_Vérifier les e-mails</property> <signal handler="ggn_icon_check_selected" name="activate"/> </object> </child> <child> <object class="GtkAction" id="itemPrefs"> <property name="stock_id">gtk-preferences</property> <property name="name">itemPrefs</property> <property name="label" translatable="yes">_Préférences</property> <signal handler="ggn_icon_prefs_selected" name="activate"/> </object> </child> <child> <object class="GtkAction" id="itemAbout"> <property name="stock_id">gtk-about</property> <property name="name">itemAbout</property> <property name="label" translatable="yes">_A propos</property> <signal handler="ggn_icon_about_selected" name="activate"/> </object> </child> <child> <object class="GtkAction" id="itemQuit"> <property name="stock_id">gtk-quit</property> <property name="name">itemQuit</property> <property name="label" translatable="yes">_Quitter</property> <signal handler="ggn_icon_quit_selected" name="activate"/> </object> </child> </object> </child> <ui> <popup name="GgnContextMenu"> <menuitem action="itemCheck"/> <separator/> <menuitem action="itemPrefs"/> <menuitem action="itemAbout"/> <separator/> <menuitem action="itemQuit"/> </popup> </ui> </object> <object class="GtkMenu" constructor="uimanager1" id="GgnContextMenu"> </object> </interface> • ggn-edit.xml <?xml version="1.0"?> <!--*- mode: xml -*--> <interface> <object class="GtkWindow" id="GgnEditWindow"> <property name="title" translatable="yes">Éditer le compte</property> <property name="resizable">False</property> <property name="window_position">GTK_WIN_POS_CENTER_ON_PARENT</property> <property name="default_width">460</property> <property name="icon">ggn-normal-lg.svg</property> <signal handler="ggn_edit_window_deleted" name="delete_event"/> <signal handler="ggn_edit_window_key_pressed" name="key_press_event"/> <child> <object class="GtkVBox" id="vboxMain"> <property name="visible">True</property> <property name="border_width">4</property> <child> <object class="GtkFrame" id="fraDesc"> <property name="visible">True</property> <property name="border_width">2</property> <property name="label_xalign">0</property> <property name="shadow_type">GTK_SHADOW_NONE</property> <child> <object class="GtkAlignment" id="alnDesc"> <property name="visible">True</property> <property name="border_width">4</property> <property name="left_padding">12</property> <child> <object class="GtkHBox" id="hboxDesc"> <property name="visible">True</property> <property name="spacing">4</property> <child> <object class="GtkLabel" id="lblDesc"> <property name="visible">True</property> <property name="label" translatable="yes">Nom:</property> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> </packing> </child> <child> <object class="GtkEntry" id="txtName"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="invisible_char">&#x2022;</property> </object> <packing> <property name="position">1</property> </packing> </child> </object> </child> </object> </child> <child type="label"> <object class="GtkLabel" id="lblFraDesc"> <property name="visible">True</property> <property name="label" translatable="yes">&lt;b&gt;Description&lt;/b&gt;</property> <property name="use_markup">True</property> </object> </child> </object> </child> <child> <object class="GtkFrame" id="fraCreds"> <property name="visible">True</property> <property name="border_width">2</property> <property name="label_xalign">0</property> <property name="shadow_type">GTK_SHADOW_NONE</property> <child> <object class="GtkAlignment" id="alnFraCreds"> <property name="visible">True</property> <property name="border_width">4</property> <property name="left_padding">12</property> <child> <object class="GtkTable" id="tblCreds"> <property name="visible">True</property> <property name="border_width">4</property> <property name="n_rows">3</property> <property name="n_columns">2</property> <property name="column_spacing">4</property> <property name="row_spacing">4</property> <child> <object class="GtkEntry" id="txtPass"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="visibility">False</property> <property name="invisible_char">&#x2022;</property> </object> <packing> <property name="left_attach">1</property> <property name="right_attach">2</property> <property name="top_attach">2</property> <property name="bottom_attach">3</property> </packing> </child> <child> <object class="GtkEntry" id="txtDomain"> <property name="visible">True</property> <property name="can_focus">True</property> </object> <packing> <property name="left_attach">1</property> <property name="right_attach">2</property> <property name="top_attach">1</property> <property name="bottom_attach">2</property> </packing> </child> <child> <object class="GtkEntry" id="txtUser"> <property name="visible">True</property> <property name="can_focus">True</property> </object> <packing> <property name="left_attach">1</property> <property name="right_attach">2</property> </packing> </child> <child> <object class="GtkLabel" id="lblPass"> <property name="visible">True</property> <property name="label" translatable="yes">Mot de passe:</property> </object> <packing> <property name="top_attach">2</property> <property name="bottom_attach">3</property> <property name="x_options"/> </packing> </child> <child> <object class="GtkLabel" id="lblDomain"> <property name="visible">True</property> <property name="label" translatable="yes">Domaine:</property> </object> <packing> <property name="top_attach">1</property> <property name="bottom_attach">2</property> <property name="x_options"/> </packing> </child> <child> <object class="GtkLabel" id="lblUser"> <property name="visible">True</property> <property name="label" translatable="yes">Nom d'utilisateur:</property> </object> <packing> <property name="x_options"/> </packing> </child> </object> </child> </object> </child> <child type="label"> <object class="GtkLabel" id="lblFraCreds"> <property name="visible">True</property> <property name="label" translatable="yes">&lt;b&gt;Nom d'utilisateur Gmail&lt;/b&gt;</property> <property name="use_markup">True</property> </object> </child> </object> <packing> <property name="position">1</property> </packing> </child> <child> <object class="GtkFrame" id="fraSend"> <property name="visible">True</property> <property name="border_width">2</property> <property name="label_xalign">0</property> <property name="shadow_type">GTK_SHADOW_NONE</property> <child> <object class="GtkAlignment" id="alnFraSend"> <property name="visible">True</property> <property name="border_width">4</property> <property name="left_padding">12</property> <child> <object class="GtkCheckButton" id="chkDefault"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="label" translatable="yes">Utilise ce compte pour écrire de nouveaux e-mails</property> <property name="draw_indicator">True</property> </object> </child> </object> </child> <child type="label"> <object class="GtkLabel" id="lblFraSend"> <property name="visible">True</property> <property name="label" translatable="yes">&lt;b&gt;Composition&lt;/b&gt;</property> <property name="use_markup">True</property> </object> </child> </object> <packing> <property name="position">2</property> </packing> </child> <child> <object class="GtkHBox" id="hboxButtons"> <property name="visible">True</property> <property name="border_width">4</property> <property name="spacing">4</property> <child> <placeholder/> </child> <child> <object class="GtkButton" id="btnCancel"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="label">gtk-cancel</property> <property name="use_stock">True</property> <signal handler="ggn_edit_window_cancelled" name="clicked"/> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> <property name="position">1</property> </packing> </child> <child> <object class="GtkButton" id="btnOK"> <property name="visible">True</property> <property name="can_focus">True</property> <property name="label">gtk-ok</property> <property name="use_stock">True</property> <signal handler="ggn_edit_window_confirmed" name="clicked"/> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> <property name="pack_type">GTK_PACK_END</property> <property name="position">2</property> </packing> </child> </object> <packing> <property name="expand">False</property> <property name="fill">False</property> <property name="position">3</property> </packing> </child> </object> </child> </object> </interface> Ca marche (h) Dernière modification par snipe2004 (Le 29/03/2011, à 13:37) MEDION AKOYA P7624 (MD 98921) Processeur Intel Core™ i5-2430M 2,40 GHz (Turbo Boost 2,0 jusqu‘à 3,0 GHz, 3 Mo Intel Smart Cache, Intel Hyper-Threading) NVIDIA GeForce GT630M DirectX 11 (1024 Mo de mémoire, sortie numérique audio/vidéo HDMI et Optimus) + Intel Centrino Advanced-N 1030 (Bluetooth 3.0 intégré) Ubuntu 13.10 Hors ligne #5 Le 29/03/2011, à 13:47 HP Re : Gmail notifier : ajout de la langue "francais" Sinon, il y a un fork ici : Gmail Notifier En Français (parce que le développeur, moi, est francophone) mais aussi ar, de, eo, es, it, jp, nl, pl, sv, tr Par contre pas du tout recommandé pour un environnement Gnome « standard » * wink * surtout à cause de gnome-panel et, dans une moindre mesure, de notify-osd Dernière modification par HP (Le 29/03/2011, à 13:57) cat /dev/urandom >/dev/null 2>&1 #621141 - Pi Hors ligne
__label__pos
0.920799
How To Tag Someone In Google Sheets? 9 how to tag someone in google sheets Have you ever wondered how to tag someone in Google Sheets? It’s a great way to keep track of their data and make sure you don’t miss any important information. Here’s how to do it: 1. Open Google Sheets and click on the “File” button in the top left corner. 2. Click on “Make a copy” and select the desired sheet (or sheets) from your Google Drive. 3. In the “Sheet name” column, type in a name for your tag and hit enter. 4. In the “People you’re tagging” column, type in the names of the people you want to tag. 5. To add a new tag, type in a keyword or phrase in the “Tag” column and hit enter. 6. To remove someone from your tag list, simply click on their name and hit delete. 7. You’re ready to start tagging! Now you can easily find all the data related to your tag in one place. Table of Contents What is a Tag? A tag is an identifier you add to someone’s data in a spreadsheet. It allows you to quickly identify that person and their data later on. How to Tag Someone in Google Sheets? To tag someone in Google Sheets, follow these steps: 1. Open the spreadsheet you want to tag someone in. 2. Go to the cell where you want to add the tag. 3. Choose the ‘tag’ icon from the top toolbar or menu bar. 4. Enter the name of the person you want to tag and click ‘OK’. 5. Repeat steps 3 and 4 for each person you want to tag. 6. Click ‘Save’ when you’re done. 7. You’re done! Now everyone can easily find them with just a few clicks. What Are the Benefits of Tagging Someone in Google Sheets? Tagging someone in Google Sheets can be beneficial for many reasons, such as tracking their progress, keeping track of their data, and making sure they don’t miss anything important. It’s also a great way to stay organized and keep track of your work. What Are The Steps To Tag Someone In Google Sheets? Here are the steps to tag someone in Google Sheets: 1. Open the spreadsheet you want to tag someone in. 2. Go to the cell where you want to add the tag. 3. Choose the ‘tag’ icon from the top toolbar or menu bar. 4. Enter the name of the person you want to tag and click ‘OK’. 5. Repeat steps 3 and 4 for each person you want to tag. 6. Click ‘Save’ when you’re done. 7. You’re done! Now everyone can easily find them with just a few clicks. 8. If necessary, adjust the tags later on by clicking ‘Edit Tags’ on the top toolbar or menu bar of your spreadsheet. 9. Delete tags by clicking on them and then clicking ‘Delete’ from the top toolbar or menu bar of your spreadsheet. Why can’t I tag people on Google Sheets? Have you ever wondered why you can’t tag people on Google Sheets? It sounds like a simple feature, but it’s not available for everyone. This article will explain why you can’t tag people on Google Sheets, and how to get around it. What is Google Sheets? Google Sheets is a web-based spreadsheet tool that allows you to create, organize and manipulate data. It also features such as formulae, conditional formatting, and advanced data manipulation. Why Can’t I Tag People on Google Sheets? The reason you can’t tag people on Google Sheets has to do with the way that it works. There are two main reasons for this: 1. The way that Google Sheets is designed doesn’t allow you to tag people directly. 2. If you try to tag someone, it will show up as a comment, rather than an actual tag. What Other Ways Can I Tag People on Google Sheets? There are other ways to tag people on Google Sheets, such as using the advanced tagging feature or using the data labels feature. How Do I Get Around the Issue of Not Being Able to Tag People on Google Sheets? There are a few ways that you can get around this issue. The first is to use the advanced tagging feature or the data labels feature. The second option is to use a third-party add-on such as PowerSheet or SpreadsheetGems. We hope this article has helped you understand why you can’t tag people on Google Sheets and how to get around it. With the right add-on or feature, you can easily tag people on Google Sheets without having to worry about it breaking your workflow. How do you mention someone in a spreadsheet? Have you ever tried to add someone to a spreadsheet? It can be tricky, especially when you don’t want to mention their name. Here are some tips on how to mention someone in a spreadsheet. Start with a Name The first step is to start with a name. Make sure you include their full name, title, and any other information that may help you identify them. If you’re not sure what to include, make a note of the person’s name, title or position, and any other relevant details. Add Their Title Next, add their title or position. This can be anything from “Vice President of Sales” to “Accountant”. If you have more than one person in the same role, make sure to list each person’s title or position individually. Add Their Email Address Next, add their email address. This is important because it will allow you to easily contact them if needed. Make sure to include the person’s full name and title too. Add Their Phone Number Next, add their phone number. This is an important piece of information because it will allow you to quickly reach them if needed. Make sure to include the person’s full name and title too. Add Their Link Finally, add their link or website address. This is an important piece of information as it allows you to easily contact them without having to go through email or phone calls. Make sure to include the person’s full name and title too. How do I add a tag to a Google sheet? Have you ever wanted to add a tag to a Google Sheets spreadsheet? If so, you’ve come to the right place. In this article, we’ll discuss how to add a tag to a Google Sheets spreadsheet. To add a tag to a Google Sheets spreadsheet, first, open the spreadsheet you want to add the tag to. Next, click on the Labels tab located at the top of the spreadsheet. From here, you can select the Add a new label button. Once the Add a new label button is clicked, a new window will open. In this window, you will need to provide a name for your label. After providing a name for your label, click on the OK button. Now that you have created your new label, you need to specify what it will be used for. To do this, click on the Selector Box next to your newly created label and select one of the following options: -Category: This option can be used to group different sheets together based on a specific topic or subject. For example, if you have multiple sheets containing financial data, you could use Category as your Label selector option and divide your sheets into categories such as Income (income), Expenses (expenses), and so on. -Subject: This option can be used to group different sheets together based on their content or purpose. For example, if you have multiple sheets containing customer data, you could use Subject as your Label selector option and group them by company (AOL), product category (cell phone plans), or other relevant information. -Date added: This option can be used to group different sheets together based on when they were added to the Google Sheets spreadsheet. For example, if you have multiple sheets containing sales data from different dates, you could use Date added as your Label selector option and group them by month (January 2017), week (July 2017), or day (October 10th). Once you have selected your Label selector option, click on the OK button. Now that you have added a new label, you will need to specify what data it will be used for. To do this, click on the Data tab located at the top of the spreadsheet. From here, you can select the data you want to use for your label. After selecting the data you want to use, click on the OK button. Finally, you need to add your new label to your sheet. To do this, click on the Labels tab located at the top of the spreadsheet and select your newly created label from the list of labels. Click on the OK button when finished. That is how you add a tag to a Google Sheets spreadsheet. What is a Tag? A tag is an attribute associated with a spreadsheet cell or range of cells. It can be used for organizing and filtering data, as well as for tracking data over time. How to Add a Tag to a Google Sheets Spreadsheet? To add a tag to a Google Sheets spreadsheet, follow these steps: 1. Open the spreadsheet in question. 2. Click on the ‘Data’ tab at the top of the page. 3. Select ‘Add Column’ and choose ‘Tag’ from the list of options. 4. Type the name of the tag you want to add and select the appropriate options. 5. Click ‘OK’ when finished. What Are the Benefits of Adding a Tag to a Google Sheets Spreadsheet? Adding tags can help organize and filter data, as well as track data over time. It can also be used for tracking customer data, project data, and more. How Do I Add a Tag to a Google Sheets Spreadsheet? To add a tag to a Google Sheets spreadsheet, follow these steps: 1. Open the spreadsheet in question. 2. Click on the ‘Data’ tab at the top of the page. 3. Select ‘Add Column’ and choose ‘Tag’ from the list of options. 4. Type the name of the tag you want to add and select the appropriate options. 5. Click ‘OK’ when finished. 6. Save your changes and close the spreadsheet if necessary. 7. Enjoy! How do I tag someone in a shared Google sheet? Are you looking for a way to tag someone in a shared Google Sheet? You’re not alone. Many people need help remembering who is in a shared spreadsheet and how to tag them. In this article, we’ll teach you how to tag someone in a Google Sheet. To tag someone in a shared Google Sheet, first, open the spreadsheet and find the person you want to tag. Next, click on their name or picture. (If they’re not in the current sheet, you can search for them.) Now that you’ve found them, click on the “Tag” button next to their name or picture. (It’s on the right side of the screen.) A drop-down menu will appear. (If you don’t see the “Tag” button, hover your mouse over it and it will become visible.) Select the option that best describes who you’re tagging. You can also use the dropdown menu to filter the list of tags by type (e.g., people, location, date). Once you’ve selected your tag, click on the “OK” button. Your tag will now appear next to the person’s name or picture in all future copies of the sheet that you make. What Are the Benefits of Tagging Someone in a Shared Google Sheet? Tags help you quickly identify and manage who is in a spreadsheet, making it easier to collaborate and access data from multiple sources. Additionally, you can use tags to keep track of who has been added or removed from a spreadsheet. How Do I Tag Someone in a Shared Google Sheet? To tag someone in a shared Google Sheet, you need to first create a new column and enter the person’s name or alias into the column. Then, simply add the tag associated with that person. What Is the Best Way to Tag Multiple People? The best way to tag multiple people is to use an Excel formula to do so. This will allow you to quickly add multiple tags without having to manually enter each one. What If I Need to Remove Someone from a Shared Google Sheet? If you need to remove someone from a shared spreadsheet, simply remove the tag they have been given and re-tag them using the same process as above. What Are Some Tips for Using Tags Effectively? Here are some tips for using tags effectively: – Make sure you use the right tag for the right person or item. – Use tags to group items together, so you can more easily find and use them. – Use different tags for different types of data (e.g., people, locations, dates). – Remove tags when you no longer need them. Why can’t I tag a person? Are you looking for an answer to the question: Why can’t I tag a person? You’re not alone. Many people have been trying to figure out this question for years. There are a few reasons why you might not be able to tag a person in a post. Maybe the person hasn’t been tagged in the post yet, or maybe they’ve been tagged by someone else and you don’t have permission to tag them. If you’re trying to tag someone who’s already been tagged, there’s a good chance that their name is in the post already. If it’s not, try looking for an identification number or an annotation that explains who the person is. And if you still can’t find the information you’re looking for, feel free to ask us in the comments below. We’re happy to help. What is a Tag? A tag is a way to identify content in a social media post or profile. It’s usually used to show who is contributing to the post or who is the original creator of the content. Why Can’t I Tag a Person? Generally speaking, you can’t tag someone unless they have given permission to be tagged. This can be done through their social media profile, their email address, or by contacting them directly. What Should I Do If I Can’t Tag a Person? If you can’t tag a person, there are a few things you can do. You can leave a comment on their social media post or profile, or you can contact them directly to ask for permission. What Are Some Other Ways I Can Help Out? If you want to help out, there are other ways you can do so. You can share their content, like and comment on their posts, and even follow them on social media. Is There Anything Else I Should Know About Tagging People? There are some other things you should know when it comes to tagging people. You should always give permission before tagging someone, and you should make sure that the person you’re tagging is okay with being tagged. Why can’t I tag someone as a collaborator? Have you ever been unable to tag someone as a collaborator in a project? It’s a common problem, but don’t worry – there’s a good explanation for why it can be difficult. In this essay, we’ll explore the reasons why you might not be able to tag someone as a collaborator. Hopefully, by the end of it, you’ll be able to find a workaround or solve the issue on your own. When you’re working on a project with someone, it can be helpful to tag them as a collaborator. This will allow you to easily find their contributions and track the progress of the project. Unfortunately, not everyone understands how tagging works. Sometimes, people won’t tag themselves as collaborators, or they might forget to tag someone. There are several reasons why tagging someone as a collaborator can be difficult. First, sometimes people don’t want to be tagged as collaborators. They might feel that their contributions are minor or that they don’t deserve to be included in the project. Second, sometimes people don’t have access to the tagging feature on the platform where the project is being worked on. In these cases, it can be difficult to tag the person correctly. If you’re having trouble tagging someone as a collaborator, there are several solutions that you can try. You can ask them if they would like to be tagged as a collaborator and give them the opportunity to do so. Alternatively, you can create a separate project to which only the collaborators will have access and use tags to identify each person’s contribution separately. Finally, you can use other methods of tracking progress such as Trello or Asana if tagging is not an option. How Does Tag Collaboration Work? Tag collaboration is a feature on many platforms and services that allows you to tag people in your project as collaborators. This makes it easy for others to see who is involved in your project and who can help out with it. When you tag someone as a collaborator, they are automatically added to the list of people who can access the project and make contributions. When you tag someone as a collaborator, they are automatically added to the list of people who can access the project and make contributions. Why Do People Get Flagged as Non-Collaborators? There are several reasons why someone might get flagged as a non-collaborator. These include not having permission to access the platform, not having the necessary permissions, or just not being involved in the project itself. What Are Some Solutions to This Problem? There are several solutions to this problem, including creating an account on the platform, getting permissions, and making sure you are actually part of the project. What Else Should I Know About Tag Collaboration? Aside from understanding why you can’t tag someone as a collaborator, you should also know how to fix this problem. You can check out tips for using tag collaboration effectively, and even use tools like Slack or Trello to help manage projects more effectively.     LEAVE A REPLY Please enter your comment! Please enter your name here
__label__pos
0.984856
The best way to Block Discord on a Cellphone, PC, Router, or in Chrome There’s no denying that Discord is an efficient streaming software program program! Nonetheless, like many alternative web pages and apps, it isn’t the proper place for youths – Discord might keep delicate data or simply become addictive. Within the occasion you’re concerned about your toddler using Discord, be taught our data. How to Block Discord on a Phone, PC, Router, or in Chrome On this text, we’re going to make clear one of the simplest ways to dam Discord on Chromebook, Mac, Residence home windows, cell devices, and routers. We’ll even take a look at one of the simplest ways to dam Discord audio on Obs. Be taught on to discover ways to deal with app entry in your system. The best way to Block Discord on a Chromebook? You probably can block Discord on a Chromebook an identical to each different app with the help of parental administration. To do that, observe the steps underneath: 1. Create a separate account in your toddler. First, sign out of your account. 2. On the bottom of the sign-in internet web page, click on on ‘’Add particular person’’. 3. Form in your toddler’s Google account e mail and password, click on on ‘’Subsequent’’ and observe the on-screen instructions. 4. As quickly because the model new account is about up, limit entry to your Chromebook. Register to the admin account. 5. Navigate to the ‘’Setting menu’’. 6. Click on on ‘’Deal with totally different people’’ beneath the People half. 7. Choose your toddler’s account beneath the “Restrict sign-in to the following prospects” half. 8. To restrict entry to Discord, go to the Family Hyperlink app. 9. Go to your toddler’s profile, then to ‘’Settings’’. 10. Click on on ‘’Apps put in’’, then ‘’Further’’. 11. Select Discord and shift the toggle button to Off” to dam the entry. 12. To dam Discord inside the browser, navigate once more to the child’s account settings, then click on on ‘’Filters’’ on Google Chrome. 13. Click on on ‘’Deal with web sites’’, then ‘’Blocked’’. 14. Click on on the plus icon on the bottom of your show and paste the Discord URL to the textual content material enter area, then shut the window. The best way to Block Discord on a Mac? To dam Discord on Mac using Show Time, observe the instructions underneath: 1. Organize Show Time in your toddler. To do that, log in to your toddler’s Mac account. 2. Navigate to Apple Menu, then to ‘’System Preferences,’’ and select ‘’Show Time.’’ 3. Select ‘’Selections’’ from the menu on the left. 4. Select ‘’Activate’’ inside the upper-right nook of your show. 5. Select the ‘’Use Show Time Passcode’’ chance. 6. Return to ‘’Show Time settings’’ and click on on ‘’Content material materials & Privateness,’’ Click on on the ‘’Flip On’’ button. 7. Click on on ‘’Apps,’’ uncover the Discord app, and prohibit entry to it. You may should enter your passcode. 8. To dam Discord inside the browser, return to ‘’Content material materials & Privateness settings’’ and select ‘’Content material materials,’’ then paste Discord URL and prohibit it. The best way to Block Discord on a Residence home windows PC? Within the occasion you’re a Residence home windows client, you might limit your toddler’s entry to Discord by following the instructions underneath: 1. Create a family group on the Microsoft website. Create a separate account in your toddler. 2. Register to your toddler’s account in your system, set it up following on-screen instructions, then sign out. 3. Register to your Microsoft account. 4. Navigate to the Start menu, then to ‘’Settings.’’ 5. Click on on ‘’Accounts,’’ then select ‘’Family & Completely different Prospects’’ from the left sidebar. 6. Uncover your toddler’s account and click on on ‘’Allow’’ beneath their account determine. 7. Return to your family members group on the Microsoft website online. 8. Select your toddler’s account and navigate to the ‘’App and sport limits’’ tab. 9. Scroll down until you uncover the Discord app, then click on on ‘’Block app.’’ The best way to Block Discord on an iPhone? Proscribing app entry on an iPhone isn’t quite a bit fully totally different from doing it on a Mac – it is essential to make use of Show Time. To do that, observe the instructions underneath: 1. Open the Settings app and navigate to the ‘’Show Time’’ settings. 2. Select “That’s my system” or “That’s my toddler’s system.” 3. Within the occasion you choose the second chance, you’ll be requested to create a model new passcode. 4. Faucet ‘’Content material materials & Privateness Restrictions’’ and enter your passcode. 5. Shift the toggle button subsequent to ‘’Content material materials & Privateness’’ to “On.” 6. Faucet ‘’Allowed apps.’’ . 7. Scroll down until you uncover the Discord app, then shift the toggle button subsequent to it to the “Off” place. The best way to Block Discord on an Android Gadget? You probably can stop your toddler from downloading Discord on Android by way of the Play Retailer app. To do that, observe the steps underneath: 1. Open the Play Retailer app. 2. Faucet the three-line icon inside the upper-left nook of your show. 3. Faucet ‘’Settings’’, then select ‘’Parental controls.’’ 4. Shift the toggle button subsequent to “Parental controls are off” to point out them on. 5. Organize a passcode, then affirm. 6. Select “Rated for 12+” or youthful to cease your toddler from downloading Discord – it’s rated 13+ inside the Play Retailer. The best way to Block Discord on a Netgear Router? You probably can limit entry to the Discord website online by organising Smart Wizard in your Netgear router. To do that, observe the instructions underneath: 1. Open browser on a laptop linked to your Netgear router. 2. Register to routerlogin.net. Within the occasion you didn’t have any login credentials organize, use “admin” as a result of the login and “password” as a password. 3. Navigate to ‘’Content material materials Filtering’’, then to ‘’Blocked Web sites.’’ 4. Select the ‘’Always’’ chance to fully block Discord. To dam Discord solely at specified events, select the ‘’Per Schedule’’ chance. 5. Paste Discord URL to the “Form key phrase or space determine proper right here” area. 6. Affirm by clicking ‘’Add key phrase’’, then ‘’Apply.’’ Optionally, you might organize parental controls in your Netgear router. To do that, observe the steps underneath: 1. Acquire and open the Orbi app in your phone and tap ‘’Parental Controls.’’ 2. Select a profile, then faucet ‘’Historic previous.’’ 3. Uncover the Discord website online and swipe from the left to the appropriate to dam it. 4. Select the ‘’Set as Filtered’’ option to completely block Discord. The best way to Block Discord on a Xfinity Router? Xfinity router permits blocking web pages with the help of parental controls. Adjust to the instructions underneath to limit entry to Discord: 1. Register to the Xfinity website. 2. From the left sidebar, select ‘’Parental Administration.’’ 3. Select ‘’Managed Web sites’’ from the dropdown menu. 4. Click on on ‘’Enable’’, then click on on ‘’Add.’’ 5. Paste Discord URL to the textual content material enter area and ensure. 6. Optionally, click on on ‘’Managed Devices’’ to limit entry to Discord only for specified devices. 7. Click on on ‘’Enable,’’ then click on on ‘’Add’’ and select a instrument. The best way to Block Discord on an Asus Router? To dam Discord on an Asus router, do the following: 1. Register to the Asus router website. 2. From the left sidebar, select ‘’Firewall.’’ 3. Navigate to the ‘’URL Filter’’ tab. 4. Paste Discord URL to the textual content material enter area on the bottom of your show. 5. Click on on ‘’Apply.’’ The best way to Block Discord on Chrome? To restrict entry to Discord in Google Chrome, observe the steps underneath: 1. Make sure that your toddler has a separate Google Account. 2. Launch the Family Hyperlink app. 3. Click on on in your toddler’s profile. 4. Open the ‘’Settings’’ tab. Click on on ‘’Deal with Settings,’’ then ‘’Filters’’ on Google Chrome. 5. Click on on ‘’Deal with Web sites’’, then ‘’Blocked.’’ 6. Click on on the plus icon on the bottom correct nook of your show. 7. Paste Discord URL to the cope with enter area, then shut the window. Observe: Family Hyperlink website online restrictions gained’t work on iPhone or iPad. It’s essential to block Discord by way of Show Time. The best way to Block Discord on Obs? You probably can block audio from Discord on Obs by following the steps underneath: 1. Launch Obs. 2. Navigate to the ‘’Sources Panel.’’ 3. Select “Audio Output Seize.” 4. Uncover the Gadget tab and select the system you utilize to stream on Discord. 5. Click on on ‘’Delete.’’ Deal with Discord Entry Hopefully, with the help of our data, now you may block Discord irrespective of your system. Parental administration is a good instrument that helps to deal with your toddler’s content material materials. You don’t primarily should restrict entry to Discord completely, though – have in mind setting a time limit as an alternative. This way, your toddler will nonetheless be succesful to make use of the required app with out spending all of their time on it. What’s your opinion on children using Discord? Share your concepts inside the suggestions half underneath. Leave a Reply Your email address will not be published. Required fields are marked * Press ESC to close
__label__pos
0.996144
0 Hello, I am currently trying to figure out some event stuff with Javascript. I have the page capturing all onkeypress events by calling my function keyPress, and passing the event as an argument. The key presses get passed to a text field that I have on the page (even if it's out of focus at the time of the key press) by using textfield.focus() within my keyPress function. This seems to pass the event along to the text field so that it registers the stroke and handles it as if the field was in focus at the time of the key press. My problem lies in that I need to then grab the new value of the text field for use with another part of the script. It seems though that with the way I'm setting focus, it'll execute the rest of my keyPress function (with the outdated text field value) before the text field handles the event. Is there a way to yield the event to this text field first? Sorry, this was a long post, but I guess here's a short recap: If I handle key presses via the body of the page, so that regardless of the text field's current state of focus it updates the text field accordingly, is there a way to have that happen first before the rest of my function that needs to use the new value of the text field? Thanks in advance! 2 Contributors 3 Replies 4 Views 7 Years Discussion Span Last Post by codejoust 0 I don't think you can do it that way, initiating the keypress runs the code, and the keypress IS what would change the value of a textarea. What I can recommend, though, is capturing the keypress key and appending that key's value to the text input. 0 What I can recommend, though, is capturing the keypress key and appending that key's value to the text input. Thanks, that would be nice if it were that simple :P I'd have to handle the proper action if the key was not a regular, printable character, such as if the user pressed left arrow, backspace, delete, etc. On top of that, there goes support for different keyboards, because of the placement of different keys in different places on international keyboards. Because the keycode would be run through my script instead of the native handling on the client's browser/OS. I think I'll just find some alternate solution for this... 0 window.onload=(function(){ window.onkeypress = fillTextarea; }); function fillTextarea(e) { var code; if (!e) var e = window.event; if (e.keyCode) code = e.keyCode; else if (e.which) code = e.which; var character = String.fromCharCode(code).replace(/[^A-Za-z 0-9]/,''); var textarea = document.getElementById('text'); if (textarea.focused){ } else { if (character != ''){ textarea.value += charater; } } } window.onload = function() { elem = document.getElementById('text'); elem.focused = false; elem.hasFocus = function() { return this.focused; }; elem.onfocus=function() { this.focused=true; }; elem.onblur=function() { this.focused=false; }; } I can rewrite this for jquery, It'll be much cleaner. Edited by codejoust: n/a This topic has been dead for over six months. Start a new discussion instead. Have something to contribute to this discussion? Please be thoughtful, detailed and courteous, and be sure to adhere to our posting rules.
__label__pos
0.988577
Misframe Jun 15, 2020 Dark theme I added a dark theme to this blog over the weekend. Here’s how you can do something similar with just a few CSS updates. The first thing I did was move all color codes to variables. This is useful in general so you can define colors once and reuse them instead of copying and pasting color codes everywhere. :root { --main-bg-color: white; --main-fg-color: black; --date-fg-color: #c0c0c0; --border-color: #eee; } Those are the only 4 colors I use. Next, I updated styles to reference those variables, like this: .mf-header-nav-links a { color: var(--main-fg-color); } I also had to add rules for things that assumed certain defaults, like the body text color and background: body { color: var(--main-fg-color); background-color: var(--main-bg-color); } Finally, I added a media selector for dark color scheme preferences. All it does is update the variables I defined earlier. @media (prefers-color-scheme: dark) { :root { --main-bg-color: black; --main-fg-color: white; --date-fg-color: #888; --border-color: #333; } } And that’s it! It only took a few minutes to do. All of the websites I have created recently use CSS variables and the prefers-color-scheme media selector because they’re so useful. Next read these:
__label__pos
0.834878
O'Reilly logo Stay ahead with the world's most comprehensive technology and business learning platform. With Safari, you learn the way you learn best. Get unlimited access to videos, live online training, learning paths, books, tutorials, and more. Start Free Trial No credit card required WordPress 3 for Business Bloggers Book Description Promote and grow your WordPress blog with advanced marketing techniques, plugins, advertising, and SEO • Use WordPress to create a winning blog for your business • Develop and transform your blog with strategic goals • Market and measure the success of your blog In Detail WordPress makes the business of blogging easy. But there’s more to a successful business blog than just churning out posts. You need to understand the advanced marketing and promotion techniques to make your blog stand out from the crowd, attract visitors, benefit your brand, and deliver a worthwhile return on your investment. WordPress 3 for Business Bloggers shows you how to use WordPress to run your business blog. It covers everything you need to develop a custom look for your blog, use analytics to understand your visitors, market your blog online, and foster connections with other bloggers to increase your traffic and the value of your blog. You begin by identifying your blog’s strategic goals before going step-by-step through the advanced techniques that will grow your blog to its full business potential. You will learn how to build a custom theme for your blog and incorporate multimedia content like images and video. Advanced promotion techniques like SEO and social media marketing are covered in detail before you learn how to monetize your blog and manage its growth. WordPress 3 for Business Bloggers will help you to create a blog that brings real benefits to your business. " Table of Contents 1. WordPress 3 for Business Bloggers 1. WordPress 3 for Business Bloggers 2. Credits 3. About the Author 4. About the Reviewers 5. www.PacktPub.com 1. Support files, eBooks, discount offers and more 1. Why Subscribe? 2. Free Access for Packt account holders 6. Preface 1. What this book covers 2. What you need for this book 3. Who this book is for 4. Conventions 5. Reader feedback 6. Customer support 1. Downloading the example code 2. Errata 3. Piracy 4. Questions 7. 1. A Blog Less Ordinary—What Makes a Great Blog? 1. You can stand out from the crowd 2. Where do you fit in? 3. Not all business blogs are the same 1. Increasing sales 2. Adding value 3. A dialog with your customers 4. Raising awareness 5. Showing expertise 6. Providing customer service 7. Public relations 8. Driving traffic 9. Add some personality 4. Categorizing business blogs 1. Product blogs 2. Corporate or company blogs 3. News blogs 4. Expert blogs 5. The WordPress arsenal 1. Good design 2. Maximizing usability 3. Promoting your blog 4. Analyzing the statistics 5. Managing content 6. Monetizing your blog 7. Measuring success 1. Google PageRank 2. Alexa ranking 6. Summary 8. 2. Introducing our Case Study—WPBizGuru 1. WPBizGuru—the man behind the blog 2. Before and after 3. Goals and planning 1. Business situation 2. Strategic goals 3. The blog plan 1. Tactical goals 4. Implementation 4. An overview of the WPBizGuru makeover 1. Design 2. Content 3. Promotion and analysis 4. Generating revenue 5. Enabling growth 5. Summary 9. 3. Designing your Blog 1. Blog design principles 1. Layout 2. Color 1. Web color theory 3. Typography 1. Font replacement 4. Usability and accessibility 2. Implementing your blog design 3. A brief introduction to CSS 1. The early days of the web 2. Content and style 3. Looking at the code 4. The stylesheet 5. Applying the stylesheet 6. Tweaking the styles 4. Setting up a local development environment 1. Installing XAMPP 2. Setting the 'root' password for MySQL 3. Installing WordPress locally 5. Case study—WPBizGuru design 1. Setting up a child theme 1. Dummy content 2. Installing a new text editor 3. Creating your child theme 4. A closer look at style.css 2. The page layout 3. The default stylesheet 4. The header 5. The menu 6. Colors and fonts 7. The main content area 8. The sidebars 9. The footer 10. The finished theme 6. Summary 10. 4. Images and Videos 1. Image theory basics 1. Optimization 2. Images in WordPress posts 1. Thumbnail creation 2. Thumbnail size 3. Attachment size 4. Styling images 3. Setting up an image gallery 1. NextGEN Gallery 1. Creating an image gallery page 4. Using video 1. Embedding a YouTube video 5. Adding a favicon 6. Summary 11. 5. Content is King 1. Blog writing tips 1. Killer headlines 2. Length of posts 3. Post frequency 4. Links to other blogs 5. Establishing your tone and voice 6. The structure of a post 7. Ending with a question 8. A quick checklist 2. Categories and tags 1. The difference between categories and tags 2. Using categories 3. Using tags 4. Applying tags and categories to WPBizGuru 3. The About page 1. About you 2. About your blog 3. Anything to declare 4. The WPBizGuru About page 4. Other static content 5. Backing up 1. Backing up wp-content 2. Backing up the database using phpMyAdmin 3. Restoring the database from a backup file 6. Summary 12. 6. Search Engine Optimization 1. The principles of SEO 1. How search engines find stuff 2. Keywords 1. Choosing your keywords 2. Using your keywords 3. Permalinks 4. Title tags 5. Sitemaps 1. Adding a Google Sitemap 6. Inbound links 7. Robots.txt optimization 8. Using excerpts on the home page 9. Search engine submissions 1. The big four 2. DMOZ.org 3. Minor search engines and directories 10. SEO software and tools 1. Web CEO 2. Google webmaster tools 3. Firefox SEO extensions 11. Seeing results 12. Summary 13. 7. Supercharged Promotion 1. Syndication 1. WordPress feeds 1. Excerpts or full posts? 2. FeedBurner 1. Setting up FeedBurner 2. Using FeedBurner 2. Blog indexes and search engines 1. Ping-O-Matic 2. FeedBurner's Pingshot 3. Technorati 4. Minor blog indexes 3. Using social networks 1. Facebook 2. LinkedIn 4. Using Twitter 1. Setting up Twitter in WordPress 5. Social bookmarking 1. Adding links 2. Bookmarking tips 6. Summary 14. 8. Connecting with the Blogosphere 1. Defining the blogosphere 2. Why it's so important to be connected 3. How to engage with the blogosphere 4. The blogroll 1. Managing your blogroll 1. Adding categories and links 5. Feeding off the blogosphere 6. The importance of comments 1. Fishing for comments 2. Managing the conversation 3. Moderation 4. Dealing with negative comments 5. Trackbacks 6. Comment and trackback spam 7. Installing a contact form 1. Using the Contact Form 7 plugin 2. Preventing contact form spam 8. Summary 15. 9. Analyzing your Blog Stats 1. Key performance indicators 1. Traffic 1. Hits 2. Unique visitors 3. Visits 4. Page views 2. Subscribers 1. RSS subscriptions 2. E-mail subscriptions 3. Comments and feedback 4. Search engine results 5. Inbound links 2. Web analytics tools 1. WordPress.com Stats 2. Google Analytics 3. Using Google Analytics 1. Getting started 2. Visitors 3. Traffic sources 4. Google AdWords 5. Content 4. Not an exact science 5. FeedBurner Stats 1. Subscribers 2. Item use 3. Uncommon uses 3. Alexa rankings 4. Summary 16. 10. Monetizing your Blog 1. Google AdSense 1. Getting started with AdSense 2. Creating AdSense ad units 3. Using the AdSense code in WordPress 2. Affiliate programs 1. Amazon Associates 1. Creating an Amazon Associates widget 2. Using your Amazon widget in WordPress 2. Affiliate networks 3. Direct ad sales 1. Banner sizes 2. Where to place banner ads 3. How much to charge 4. Your media pack and rate card 5. Rotating banner ads 4. Paid reviews 5. Case study review 6. Summary 17. 11. Managing Growth 1. Keeping up with the workload 2. Going mobile 3. Managing increased traffic 1. Installing WP Super Cache 4. Outgrowing your web host 1. Virtual Private Servers and Cloud Servers 2. Moving WordPress to a new server 5. Bringing in other writers 1. How to find guest writers 6. Introducing WordPress Multisite 1. Getting started with WordPress Multisite 2. Installing a network 3. Managing your network 4. Developing a blog network 7. Summary
__label__pos
1
Was this page helpful? Your feedback about this content is important. Let us know what you think. Additional feedback? 1500 characters remaining Export (0) Print Expand All Collapse the table of content Expand the table of content Expand Minimize Get-DnsServerDiagnostics Get-DnsServerDiagnostics Retrieves DNS event logging details. Syntax Parameter Set: Get0 Get-DnsServerDiagnostics [-AsJob] [-CimSession <CimSession[]> ] [-ComputerName <String> ] [-ThrottleLimit <Int32> ] [ <CommonParameters>] Detailed Description The Get-DnsServerDiagnostics cmdlet retrieves Domain Name System (DNS) server diagnostic and logging parameters. Parameters -AsJob Aliases none Required? false Position? named Default Value none Accept Pipeline Input? false Accept Wildcard Characters? false -CimSession<CimSession[]> Runs the cmdlet in a remote session or on a remote computer. Enter a computer name or a session object, such as the output of a New-CimSession or Get-CimSession cmdlet. The default is the current session on the local computer. Aliases none Required? false Position? named Default Value none Accept Pipeline Input? false Accept Wildcard Characters? false -ComputerName<String> Specifies a DNS server. The acceptable values for this parameter are: an IP V4 address; an IP V6 address; any other value that resolves to an IP address, such as a fully qualified domain name (FQDN), host name, or NETBIOS name. Aliases none Required? false Position? named Default Value none Accept Pipeline Input? false Accept Wildcard Characters? false -ThrottleLimit<Int32> Specifies the maximum number of concurrent operations that can be established to run the cmdlet. If this parameter is omitted or a value of 0 is entered, then Windows PowerShell® calculates an optimum throttle limit for the cmdlet based on the number of CIM cmdlets that are running on the computer. The throttle limit applies only to the current cmdlet, not to the session or to the computer. Aliases none Required? false Position? named Default Value none Accept Pipeline Input? false Accept Wildcard Characters? false <CommonParameters> This cmdlet supports the common parameters: -Verbose, -Debug, -ErrorAction, -ErrorVariable, -OutBuffer, and -OutVariable. For more information, see    about_CommonParameters (http://go.microsoft.com/fwlink/p/?LinkID=113216). Inputs The input type is the type of the objects that you can pipe to the cmdlet. Outputs The output type is the type of the objects that the cmdlet emits. • Microsoft.Management.Infrastructure.CimInstance#DnsServerDiagnostics Examples Example 1: Get DNS event logging details This command gets DNS event logging details for the local DNS server. PS C:\> Get-DnsServerDiagnostics Related topics Was this page helpful? (1500 characters remaining) Thank you for your feedback Community Additions ADD Show: © 2015 Microsoft
__label__pos
0.99995
Disclaimer : All the postings on this site are my own and don’t necessarily represent IBM’s positions, strategies or opinions.   This post is part of the “Learn debugging using AIX DBX” series. It provides tips for customizing the DBX debugging environment using .dbxinit file. If you looking for some other aspect of debugging with AIX DBX go to the parent topic or other sub-topics here : http://www.sangeek.com/day2dayunix/2013/08/learn-debugging-using-aix-dbx/ .dbxinit file gives an opportunity for the debuggers to provide initial set of commands that should be executed at the beginning of the dbx session. Using this we can provide “Useful customization for the dbx debugging environment”. For the .dbxinit file to be automatically picked up by the dbx command. it has to be placed in the $HOME directory of the user or the user’s current directory from where DBX is being executed. Here’s the list of subcommands that I find very useful : set $repeat set $deferevents set $expandunions set $catchbp Here’s why each is for : 1. “set $repeat” saves you the pain of typing the same commands again and again. It re-executes the last command on pressing the enter key. e.g. (dbx) r [1] stopped in main at line 18 18           int val = 0, a, b, c, d; (dbx) next stopped in main at line 20 20           a = 12; (dbx) <———————————————————- Pressing “enter” key execute the “next” subcommand stopped in main at line 21 21           b = 5; (dbx)<———————————————————- Pressing “enter” key execute the “next” subcommand stopped in main at line 22 22           c = 3; (dbx) 2. “set $deferevents” turns on the defer events feature allowing the user to set breakpoints in functions which are not yet loaded e.g.  I want to stop in function process_output(), which is defined/declared in a separate library which is not yet loaded by the binary being debugged : (dbx) st in process_output “process_output” is not loaded. Creating deferred event: <————- creates a deferred breakpoint <3> stop in process_output (dbx) r Processing deferred event# 3 <————- Identifies a library load and checks if the symbol of deferred event is loaded [3] stopped in new_lib.process_output [/usr/lib/libtest.a] at line 234 <————- Stops at the deferred breakpoint 234           int a = 20; This subcommand helps the user by allowing us to set breakpoints in functions of libraries which are not yet loaded. User does not have to wait till the library is loaded for setting a breakpoint. 3. “set $expandunions” allows the user print and see values of all the individual members of a union. By default, dbx allows us to print the values of each union member separately. e.g. In your program you have a union defined like this : union un { int ival; float fval; }; Without “set $expandunions” subcommand, you would have to print each member of the union : (dbx) print un1 [union] (dbx) print un1.ival 1 (dbx) print un1.fval 1.40129846e-45 With “set $expandunions” enabled, you can just do this : (dbx) print un1 union:(ival = 1, fval = 1.40129846e-45) <————- prints all the members of union in one shot 4. By default dbx skips the breakpoints e.g. the ones set in “called functions” while the user is stepping through the program using “next”. With “set $catchbp” set we can ensure that we don’t miss on those breakpoints.         What do you think ? Set your Twitter account name in your settings to use the TwitterBar Section. %d bloggers like this:
__label__pos
0.741486
1 Why the form validation does not work on Safari ? I tried not to load the JHTML::_(behavior.formvalidator) I load the fields from the XML file <field name="title" type="text" label="COM_MYCOMP_LBL_TITLE" required="true" class="form-control" /> With this result <input id="jform_title" class="form-control required" type="text" aria-required="true" required="required" value="" name="jform[title]"> This result contains the attributes of validation aria-required="true" required="required" that I would not use How do I turn off html5fallback.js and use a different library of validation ? 1 Answer 1 4 The required="required" is standard HTML5 validation, - the html5fallback.js is just a polyfill for those browsers that do not support html5 validations. To answer your question add novalidate to your <form> markup: <form novalidate> .... </form> This turns off html5 validation. Your Answer By clicking “Post Your Answer”, you agree to our terms of service and acknowledge you have read our privacy policy. Not the answer you're looking for? Browse other questions tagged or ask your own question.
__label__pos
0.967358
How can I get soap4r to work with digest authentication? Discussion in 'Ruby' started by [email protected], Dec 21, 2005. 1. Guest Hello, I used wsdl2rb.rb to create a SOAP client and got it to work with basic authentication by adding one single line of code to the defaultDriver.rb file that was built by wsdl2rb.rb. (I also had to install http-access2 over my Ruby installation to avoid a runtime error.) options["protocol.http.basic_auth"]<< "https://www.example.com/somewhere/, "user", "passwd"] So far so good. Now: How do I add support for digest authentication to my client? I'm using Ruby 1.8.1 for Windows. I've installed soap4r 1.5.5. I've installed http-access2 2.0.6. Thanks, Yonatan   , Dec 21, 2005 #1 1. Advertisements 2. Guest I was able to get digest authentication working by incorporating the patch I found here into http-access2: http://dev.ctor.org/http-access2/ticket/27 For what it's worth, here is my version with the patch. I really hardcoded things quite a bit. Concept wise, it's totally not correct, because it relies on the basic_auth property. If the basic_auth property was set, it attempts to add both basic and digest. I did verify, though, that it works (on my setup) against both a server that requires basic and a server that requires digest: # HTTPAccess2 - HTTP accessing library. # Copyright (C) 2000-2005 NAKAMURA, Hiroshi <>. # This program is copyrighted free software by NAKAMURA, Hiroshi. You can # redistribute it and/or modify it under the same terms of Ruby's license; # either the dual license version in 2003, or any later version. # http-access2.rb is based on http-access.rb in http-access/0.0.4. Some part # of code in http-access.rb was recycled in http-access2.rb. Those part is # copyrighted by Maehashi-san. # Ruby standard library require 'timeout' require 'uri' require 'socket' require 'thread' require 'digest/md5' # Extra library require 'http-access2/http' require 'http-access2/cookie' module HTTPAccess2 VERSION = '2.0.6' RUBY_VERSION_STRING = "ruby #{RUBY_VERSION} (#{RUBY_RELEASE_DATE}) [#{RUBY_PLATFORM}]" s = %w$Id: http-access2.rb 114 2005-09-13 03:20:38Z nahi $ RCS_FILE, RCS_REVISION = s[1][/.*(?=,v$)/], s[2] SSLEnabled = begin require 'openssl' true rescue LoadError false end DEBUG_SSL = true # DESCRIPTION # HTTPAccess2::Client -- Client to retrieve web resources via HTTP. # # How to create your client. # 1. Create simple client. # clnt = HTTPAccess2::Client.new # # 2. Accessing resources through HTTP proxy. # clnt = HTTPAccess2::Client.new("http://myproxy:8080") # # 3. Set User-Agent and From in HTTP request header.(nil means "No proxy") # clnt = HTTPAccess2::Client.new(nil, "MyAgent", "") # # How to retrieve web resources. # 1. Get content of specified URL. # puts clnt.get_content("http://www.ruby-lang.org/en/") # # 2. Do HEAD request. # res = clnt.head(uri) # # 3. Do GET request with query. # res = clnt.get(uri) # # 4. Do POST request. # res = clnt.post(uri) # res = clnt.get|post|head(uri, proxy) # class Client attr_reader :agent_name attr_reader :from attr_reader :ssl_config attr_accessor :cookie_manager attr_reader :test_loopback_response class << self %w(get_content head get post put delete options trace).each do |name| eval <<-EOD def #{name}(*arg) new.#{name}(*arg) end EOD end end # SYNOPSIS # Client.new(proxy = nil, agent_name = nil, from = nil) # # ARGS # proxy A String of HTTP proxy URL. ex. "http://proxy:8080". # agent_name A String for "User-Agent" HTTP request header. # from A String for "From" HTTP request header. # # DESCRIPTION # Create an instance. # SSLConfig cannot be re-initialized. Create new client. # def initialize(proxy = nil, agent_name = nil, from = nil) @proxy = nil # assigned later. @no_proxy = nil @agent_name = agent_name @from = from @basic_auth = BasicAuth.new(self) @digest_auth = DigestAuth.new(self) @debug_dev = nil @ssl_config = SSLConfig.new(self) @redirect_uri_callback = method:)default_redirect_uri_callback) @test_loopback_response = [] @session_manager = SessionManager.new @session_manager.agent_name = @agent_name @session_manager.from = @from @session_manager.ssl_config = @ssl_config @cookie_manager = WebAgent::CookieManager.new self.proxy = proxy end def debug_dev @debug_dev end def debug_dev=(dev) @debug_dev = dev reset_all @session_manager.debug_dev = dev end def protocol_version @session_manager.protocol_version end def protocol_version=(protocol_version) reset_all @session_manager.protocol_version = protocol_version end def connect_timeout @session_manager.connect_timeout end def connect_timeout=(connect_timeout) reset_all @session_manager.connect_timeout = connect_timeout end def send_timeout @session_manager.send_timeout end def send_timeout=(send_timeout) reset_all @session_manager.send_timeout = send_timeout end def receive_timeout @session_manager.receive_timeout end def receive_timeout=(receive_timeout) reset_all @session_manager.receive_timeout = receive_timeout end def proxy @proxy end def proxy=(proxy) if proxy.nil? @proxy = nil else if proxy.is_a?(URI) @proxy = proxy else @proxy = URI.parse(proxy) end if @proxy.scheme == nil or @proxy.scheme.downcase != 'http' or @proxy.host == nil or @proxy.port == nil raise ArgumentError.new("unsupported proxy `#{proxy}'") end end reset_all @proxy end def no_proxy @no_proxy end def no_proxy=(no_proxy) @no_proxy = no_proxy reset_all end # if your ruby is older than 2005-09-06, do not set socket_sync = false to # avoid an SSL socket blocking bug in openssl/buffering.rb. def socket_sync=(socket_sync) @session_manager.socket_sync = socket_sync end def set_basic_auth(uri, user_id, passwd) unless uri.is_a?(URI) uri = URI.parse(uri) end @basic_auth.set(uri, user_id, passwd) @digest_auth.set(uri, user_id, passwd) end =begin def set_digest_auth(uri, user_id, passwd) unless uri.is_a?(URI) uri = URI.parse(uri) end @digest_auth.set(uri, user_id, passwd) end =end def set_cookie_store(filename) if @cookie_manager.cookies_file raise RuntimeError.new("overriding cookie file location") end @cookie_manager.cookies_file = filename @cookie_manager.load_cookies if filename end def save_cookie_store @cookie_manager.save_cookies end def redirect_uri_callback=(redirect_uri_callback) @redirect_uri_callback = redirect_uri_callback end # SYNOPSIS # Client#get_content(uri, query = nil, extheader = {}, &block = nil) # # ARGS # uri an_URI or a_string of uri to connect. # query a_hash or an_array of query part. e.g. { "a" => "b" }. # Give an array to pass multiple value like # [["a" => "b"], ["a" => "c"]]. # extheader # a_hash of extra headers like { "SOAPAction" => "urn:foo" }. # &block Give a block to get chunked message-body of response like # get_content(uri) { |chunked_body| ... } # Size of each chunk may not be the same. # # DESCRIPTION # Get a_sring of message-body of response. # def get_content(uri, query = nil, extheader = {}, &block) retry_connect(uri, query) do |uri, query| get(uri, query, extheader, &block) end end def post_content(uri, body = nil, extheader = {}, &block) retry_connect(uri, nil) do |uri, query| post(uri, body, extheader, &block) end end def default_redirect_uri_callback(res) uri = res.header['location'][0] puts "Redirect to: #{uri}" if $DEBUG uri end def head(uri, query = nil, extheader = {}) request('HEAD', uri, query, nil, extheader) end def get(uri, query = nil, extheader = {}, &block) request('GET', uri, query, nil, extheader, &block) end def post(uri, body = nil, extheader = {}, &block) request('POST', uri, nil, body, extheader, &block) end def put(uri, body = nil, extheader = {}, &block) request('PUT', uri, nil, body, extheader, &block) end def delete(uri, extheader = {}, &block) request('DELETE', uri, nil, nil, extheader, &block) end def options(uri, extheader = {}, &block) request('OPTIONS', uri, nil, nil, extheader, &block) end def trace(uri, query = nil, body = nil, extheader = {}, &block) request('TRACE', uri, query, body, extheader, &block) end def request(method, uri, query = nil, body = nil, extheader = {}, &block) conn = Connection.new conn_request(conn, method, uri, query, body, extheader, &block) conn.pop end # Async interface. def head_async(uri, query = nil, extheader = {}) request_async('HEAD', uri, query, nil, extheader) end def get_async(uri, query = nil, extheader = {}) request_async('GET', uri, query, nil, extheader) end def post_async(uri, body = nil, extheader = {}) request_async('POST', uri, nil, body, extheader) end def put_async(uri, body = nil, extheader = {}) request_async('PUT', uri, nil, body, extheader) end def delete_async(uri, extheader = {}) request_async('DELETE', uri, nil, nil, extheader) end def options_async(uri, extheader = {}) request_async('OPTIONS', uri, nil, nil, extheader) end def trace_async(uri, query = nil, body = nil, extheader = {}) request_async('TRACE', uri, query, body, extheader) end def request_async(method, uri, query = nil, body = nil, extheader = {}) conn = Connection.new t = Thread.new(conn) { |tconn| conn_request(tconn, method, uri, query, body, extheader) } conn.async_thread = t conn end ## # Multiple call interface. # ??? ## # Management interface. def reset(uri) @session_manager.reset(uri) end def reset_all @session_manager.reset_all end private def retry_connect(uri, query = nil) retry_number = 0 while retry_number < 10 res = yield(uri, query) if res.status == HTTP::Status::OK return res.content elsif HTTP::Status.redirect?(res.status) uri = @redirect_uri_callback.call(res) query = nil retry_number += 1 else raise RuntimeError.new("Unexpected response: #{res.header.inspect}") end end raise RuntimeError.new("Retry count exceeded.") end def conn_request(conn, method, uri, query, body, extheader, &block) unless uri.is_a?(URI) uri = URI.parse(uri) end proxy = no_proxy?(uri) ? nil : @proxy begin req = create_request(method, uri, query, body, extheader, !proxy.nil?) do_get_block(req, proxy, conn, &block) rescue Session::KeepAliveDisconnected req = create_request(method, uri, query, body, extheader, !proxy.nil?) do_get_block(req, proxy, conn, &block) end end def create_request(method, uri, query, body, extheader, proxy) if extheader.is_a?(Hash) extheader = extheader.to_a end if cred = @digest_auth.get(uri) then extheader << ['Authorization', "Digest " << cred] end if cred = @basic_auth.get(uri) then extheader << ['Authorization', "Basic " << cred] end =begin if cred = @digest_auth.get(uri) then extheader << ['Authorization', "Digest" << cred] elsif cred = @basic_auth.get(uri) extheader << ['Authorization', "Basic" << cred] end =end if cookies = @cookie_manager.find(uri) extheader << ['Cookie', cookies] end boundary = nil content_type = extheader.find { |key, value| key.downcase == 'content-type' } if content_type && content_type[1] =~ /boundary=(.+)\z/ boundary = $1 end req = HTTP::Message.new_request(method, uri, query, body, proxy, boundary) extheader.each do |key, value| req.header.set(key, value) end if content_type.nil? and !body.nil? req.header.set('content-type', 'application/x-www-form-urlencoded') end req end NO_PROXY_HOSTS = ['localhost'] def no_proxy?(uri) if !@proxy or NO_PROXY_HOSTS.include?(uri.host) return true end unless @no_proxy return false end @no_proxy.scan(/([^:,]+)(?::(\d+))?/) do |host, port| if /(\A|\.)#{Regexp.quote(host)}\z/i =~ uri.host && (!port || uri.port == port.to_i) return true end end false end # !! CAUTION !! # Method 'do_get*' runs under MT conditon. Be careful to change. def do_get_block(req, proxy, conn, &block) if str = @test_loopback_response.shift dump_dummy_request_response(req.body.dump, str) if @debug_dev conn.push(HTTP::Message.new_response(str)) return end content = '' res = HTTP::Message.new_response(content) @debug_dev << "= Request\n\n" if @debug_dev sess = @session_manager.query(req, proxy) @debug_dev << "\n\n= Response\n\n" if @debug_dev do_get_header(req, res, sess) conn.push(res) sess.get_data() do |str| block.call(str) if block content << str end @session_manager.keep(sess) unless sess.closed? end def do_get_stream(req, proxy, conn) if str = @test_loopback_response.shift dump_dummy_request_response(req.body.dump, str) if @debug_dev conn.push(HTTP::Message.new_response(str)) return end piper, pipew = IO.pipe res = HTTP::Message.new_response(piper) @debug_dev << "= Request\n\n" if @debug_dev sess = @session_manager.query(req, proxy) @debug_dev << "\n\n= Response\n\n" if @debug_dev do_get_header(req, res, sess) conn.push(res) sess.get_data() do |str| pipew.syswrite(str) end pipew.close @session_manager.keep(sess) unless sess.closed? end def do_get_header(req, res, sess) res.version, res.status, res.reason = sess.get_status sess.get_header().each do |line| unless /^([^:]+)\s*:\s*(.*)$/ =~ line raise RuntimeError.new("Unparsable header: '#{line}'.") if $DEBUG end res.header.set($1, $2) end if res.header['set-cookie'] res.header['set-cookie'].each do |cookie| @cookie_manager.parse(cookie, req.header.request_uri) end end end def dump_dummy_request_response(req, res) @debug_dev << "= Dummy Request\n\n" @debug_dev << req @debug_dev << "\n\n= Dummy Response\n\n" @debug_dev << res end end # HTTPAccess2::SSLConfig -- SSL configuration of a client. # class SSLConfig # :nodoc: attr_reader :client_cert attr_reader :client_key attr_reader :client_ca attr_reader :verify_mode attr_reader :verify_depth attr_reader :verify_callback attr_reader :timeout attr_reader :eek:ptions attr_reader :ciphers attr_reader :cert_store # don't use if you don't know what it is. def initialize(client) return unless SSLEnabled @client = client @cert_store = OpenSSL::X509::Store.new @client_cert = @client_key = @client_ca = nil @verify_mode = OpenSSL::SSL::VERIFY_PEER | OpenSSL::SSL::VERIFY_FAIL_IF_NO_PEER_CERT @verify_depth = nil @verify_callback = nil @dest = nil @timeout = nil @options = defined?(OpenSSL::SSL::OP_ALL) ? OpenSSL::SSL::OP_ALL | OpenSSL::SSL::OP_NO_SSLv2 : nil @ciphers = "ALL:!ADH:!LOW:!EXP:!MD5:mad:STRENGTH" end def set_client_cert_file(cert_file, key_file) @client_cert = OpenSSL::X509::Certificate.new(File.open(cert_file).read) @client_key = OpenSSL::pKey::RSA.new(File.open(key_file).read) change_notify end def set_trust_ca(trust_ca_file_or_hashed_dir) if FileTest.directory?(trust_ca_file_or_hashed_dir) @cert_store.add_path(trust_ca_file_or_hashed_dir) else @cert_store.add_file(trust_ca_file_or_hashed_dir) end change_notify end def set_crl(crl_file) crl = OpenSSL::X509::CRL.new(File.open(crl_file).read) @cert_store.add_crl(crl) @cert_store.flags = OpenSSL::X509::V_FLAG_CRL_CHECK | OpenSSL::X509::V_FLAG_CRL_CHECK_ALL change_notify end def client_cert=(client_cert) @client_cert = client_cert change_notify end def client_key=(client_key) @client_key = client_key change_notify end def client_ca=(client_ca) @client_ca = client_ca change_notify end def verify_mode=(verify_mode) @verify_mode = verify_mode change_notify end def verify_depth=(verify_depth) @verify_depth = verify_depth change_notify end def verify_callback=(verify_callback) @verify_callback = verify_callback change_notify end def timeout=(timeout) @timeout = timeout change_notify end def options=(options) @options = options change_notify end def ciphers=(ciphers) @ciphers = ciphers change_notify end # don't use if you don't know what it is. def cert_store=(cert_store) @cert_store = cert_store change_notify end # interfaces for SSLSocketWrap. def set_context(ctx) # Verification: Use Store#verify_callback instead of SSLContext#verify*? ctx.cert_store = @cert_store ctx.verify_mode = @verify_mode ctx.verify_depth = @verify_depth if @verify_depth ctx.verify_callback = @verify_callback || method:)default_verify_callback) # SSL config ctx.cert = @client_cert ctx.key = @client_key ctx.client_ca = @client_ca ctx.timeout = @timeout ctx.options = @options ctx.ciphers = @ciphers end # this definition must match with the one in ext/openssl/lib/openssl/ssl.rb def post_connection_check(peer_cert, hostname) check_common_name = true cert = peer_cert cert.extensions.each{|ext| next if ext.oid != "subjectAltName" ext.value.split(/,\s+/).each{|general_name| if /\ADNS:(.*)/ =~ general_name check_common_name = false reg = Regexp.escape($1).gsub(/\\\*/, "[^.]+") return true if /\A#{reg}\z/i =~ hostname elsif /\AIP Address:(.*)/ =~ general_name check_common_name = false return true if $1 == hostname end } } if check_common_name cert.subject.to_a.each{|oid, value| if oid == "CN" && value.casecmp(hostname) == 0 return true end } end raise OpenSSL::SSL::SSLError, "hostname not match" end # Default callback for verification: only dumps error. def default_verify_callback(is_ok, ctx) if $DEBUG puts "#{ is_ok ? 'ok' : 'ng' }: #{ctx.current_cert.subject}" end if !is_ok depth = ctx.error_depth code = ctx.error msg = ctx.error_string STDERR.puts "at depth #{depth} - #{code}: #{msg}" end is_ok end # Sample callback method: CAUTION: does not check CRL/ARL. def sample_verify_callback(is_ok, ctx) unless is_ok depth = ctx.error_depth code = ctx.error msg = ctx.error_string STDERR.puts "at depth #{depth} - #{code}: #{msg}" if $DEBUG return false end cert = ctx.current_cert self_signed = false ca = false pathlen = nil server_auth = true self_signed = (cert.subject.cmp(cert.issuer) == 0) # Check extensions whatever its criticality is. (sample) cert.extensions.each do |ex| case ex.oid when 'basicConstraints' /CA:(TRUE|FALSE), pathlen:(\d+)/ =~ ex.value ca = ($1 == 'TRUE') pathlen = $2.to_i when 'keyUsage' usage = ex.value.split(/\s*,\s*/) ca = usage.include?('Certificate Sign') server_auth = usage.include?('Key Encipherment') when 'extendedKeyUsage' usage = ex.value.split(/\s*,\s*/) server_auth = usage.include?('Netscape Server Gated Crypto') when 'nsCertType' usage = ex.value.split(/\s*,\s*/) ca = usage.include?('SSL CA') server_auth = usage.include?('SSL Server') end end if self_signed STDERR.puts 'self signing CA' if $DEBUG return true elsif ca STDERR.puts 'middle level CA' if $DEBUG return true elsif server_auth STDERR.puts 'for server authentication' if $DEBUG return true end return false end private def change_notify @client.reset_all end end # HTTPAccess2::BasicAuth -- BasicAuth repository. # class BasicAuth # :nodoc: def initialize(client) @client = client @auth = {} end def set(uri, user_id, passwd) uri = uri.clone uri.path = uri.path.sub(/\/[^\/]*$/, '/') @auth[uri] = ["#{user_id}:#{passwd}"].pack('m').strip @client.reset_all end def get(uri) @auth.each do |realm_uri, cred| if ((realm_uri.host == uri.host) and (realm_uri.scheme == uri.scheme) and (realm_uri.port == uri.port) and uri.path.upcase.index(realm_uri.path.upcase) == 0) return cred end end nil end end # HTTPAccess2::BasicAuth -- BasicAuth repository. # class DigestAuth # :nodoc: def initialize(client) @client = client @auth = {} @params = {} @nonce_count = 0 end def set(uri, user_id, passwd) uri = uri.clone @client.head(uri).header['WWW-Authenticate'].to_s.gsub(/(\w+)="(.*?)"/) { @params[$1] = $2 } @a_1 = "#{user_id}:#{@params['realm']}:#{passwd}" # XXX: need to obtain proper method (e.g. GET/POST/HEAD) #@a_2 = "GET:#{uri.path}" @a_2 = "POST:#{uri.path}" @cnonce = Digest::MD5.new(Time.now.to_s).hexdigest @nonce_count += 1 @message_digest = '' @message_digest << Digest::MD5.new(@a_1).hexdigest return if nil == @params['nonce'] @message_digest << ':' << @params['nonce'] @message_digest << ':' << ('%08x' % @nonce_count) @message_digest << ':' << @cnonce @message_digest << ':' << @params['qop'] @message_digest << ':' << Digest::MD5.new(@a_2).hexdigest @header = '' @header << "username=\"#{user_id}\", " @header << "realm=\"#{@params['realm']}\", " @header << "qop=\"#{@params['qop']}\", " @header << "algorithm=\"MD5\", " @header << "uri=\"#{uri.path}\", " @header << "nonce=\"#{@params['nonce']}\", " @header << "nc=#{'%08x' % @nonce_count}, " @header << "cnonce=\"#{@cnonce}\", " @header << "response=\"#{Digest::MD5.new(@message_digest).hexdigest}\", " @header << "opaque=\"#{@params['opaque']}\" " puts "header->#{@header}" puts "message_digest->#{@message_digest}" uri.path = uri.path.sub(/\/[^\/]*$/, '/') # @auth[uri] = ["#{user_id}:#{passwd}"].pack('m').strip @auth[uri] = @header @client.reset_all end def get(uri) @auth.each do |realm_uri, cred| if ((realm_uri.host == uri.host) and (realm_uri.scheme == uri.scheme) and (realm_uri.port == uri.port) and uri.path.upcase.index(realm_uri.path.upcase) == 0) return cred end end nil end end # HTTPAccess2::Site -- manage a site(host and port) # class Site # :nodoc: attr_accessor :scheme attr_accessor :host attr_reader :port def initialize(uri = nil) if uri @scheme = uri.scheme @host = uri.host @port = uri.port.to_i else @scheme = 'tcp' @host = '0.0.0.0' @port = 0 end end def addr "#{@scheme}://#{@host}:#{@port.to_s}" end def port=(port) @port = port.to_i end def ==(rhs) if rhs.is_a?(Site) ((@scheme == rhs.scheme) and (@host == rhs.host) and (@port == rhs.port)) else false end end def to_s addr end def inspect sprintf("#<%s:0x%x %s>", self.class.name, __id__, addr) end end # HTTPAccess2::Connection -- magage a connection(one request and response to it). # class Connection # :nodoc: attr_accessor :async_thread def initialize(header_queue = [], body_queue = []) @headers = header_queue @body = body_queue @async_thread = nil @queue = Queue.new end def finished? if !@async_thread # Not in async mode. true elsif @async_thread.alive? # Working... false else # Async thread have been finished. @async_thread.join true end end def pop @queue.pop end def push(result) @queue.push(result) end def join unless @async_thread false else @async_thread.join end end end # HTTPAccess2::SessionManager -- manage several sessions. # class SessionManager # :nodoc: attr_accessor :agent_name # Name of this client. attr_accessor :from # Owner of this client. attr_accessor :protocol_version # Requested protocol version attr_accessor :chunk_size # Chunk size for chunked request attr_accessor :debug_dev # Device for dumping log for debugging attr_accessor :socket_sync # Boolean value for Socket#sync # These parameters are not used now... attr_accessor :connect_timeout attr_accessor :connect_retry # Maximum retry count. 0 for infinite. attr_accessor :send_timeout attr_accessor :receive_timeout attr_accessor :read_block_size attr_accessor :ssl_config def initialize @proxy = nil @agent_name = nil @from = nil @protocol_version = nil @debug_dev = nil @socket_sync = true @chunk_size = 4096 @connect_timeout = 60 @connect_retry = 1 @send_timeout = 120 @receive_timeout = 60 # For each read_block_size bytes @read_block_size = 8192 @ssl_config = nil @sess_pool = [] @sess_pool_mutex = Mutex.new end def proxy=(proxy) if proxy.nil? @proxy = nil else @proxy = Site.new(proxy) end end def query(req, proxy) req.body.chunk_size = @chunk_size dest_site = Site.new(req.header.request_uri) proxy_site = if proxy Site.new(proxy) else @proxy end sess = open(dest_site, proxy_site) begin sess.query(req) rescue sess.close raise end sess end def reset(uri) unless uri.is_a?(URI) uri = URI.parse(uri.to_s) end site = Site.new(uri) close(site) end def reset_all close_all end def keep(sess) add_cached_session(sess) end private def open(dest, proxy = nil) sess = nil if cached = get_cached_session(dest) sess = cached else sess = Session.new(dest, @agent_name, @from) sess.proxy = proxy sess.socket_sync = @socket_sync sess.requested_version = @protocol_version if @protocol_version sess.connect_timeout = @connect_timeout sess.connect_retry = @connect_retry sess.send_timeout = @send_timeout sess.receive_timeout = @receive_timeout sess.read_block_size = @read_block_size sess.ssl_config = @ssl_config sess.debug_dev = @debug_dev end sess end def close_all each_sess do |sess| sess.close end @sess_pool.clear end def close(dest) if cached = get_cached_session(dest) cached.close true else false end end def get_cached_session(dest) cached = nil @sess_pool_mutex.synchronize do new_pool = [] @sess_pool.each do |s| if s.dest == dest cached = s else new_pool << s end end @sess_pool = new_pool end cached end def add_cached_session(sess) @sess_pool_mutex.synchronize do @sess_pool << sess end end def each_sess @sess_pool_mutex.synchronize do @sess_pool.each do |sess| yield(sess) end end end end # HTTPAccess2::SSLSocketWrap # class SSLSocketWrap def initialize(socket, context, debug_dev = nil) unless SSLEnabled raise RuntimeError.new( "Ruby/OpenSSL module is required for https access.") end @context = context @socket = socket @ssl_socket = create_ssl_socket(@socket) @debug_dev = debug_dev end def ssl_connect @ssl_socket.connect end def post_connection_check(host) verify_mode = @context.verify_mode || OpenSSL::SSL::VERIFY_NONE if verify_mode == OpenSSL::SSL::VERIFY_NONE return elsif @ssl_socket.peer_cert.nil? and check_mask(verify_mode, OpenSSL::SSL::VERIFY_FAIL_IF_NO_PEER_CERT) raise OpenSSL::SSL::SSLError, "no peer cert" end hostname = host.host if @ssl_socket.respond_to?:)post_connection_check) @ssl_socket.post_connection_check(hostname) end @context.post_connection_check(@ssl_socket.peer_cert, hostname) end def peer_cert @ssl_socket.peer_cert end def addr @socket.addr end def close @ssl_socket.close @socket.close end def closed? @socket.closed? end def eof? @ssl_socket.eof? end def gets(*args) str = @ssl_socket.gets(*args) @debug_dev << str if @debug_dev str end def read(*args) str = @ssl_socket.read(*args) @debug_dev << str if @debug_dev str end def <<(str) rv = @ssl_socket.write(str) @debug_dev << str if @debug_dev rv end def flush @ssl_socket.flush end def sync @ssl_socket.sync end def sync=(sync) @ssl_socket.sync = sync end private def check_mask(value, mask) value & mask == mask end def create_ssl_socket(socket) ssl_socket = nil if OpenSSL::SSL.const_defined?("SSLContext") ctx = OpenSSL::SSL::SSLContext.new @context.set_context(ctx) ssl_socket = OpenSSL::SSL::SSLSocket.new(socket, ctx) else ssl_socket = OpenSSL::SSL::SSLSocket.new(socket) @context.set_context(ssl_socket) end ssl_socket end end # HTTPAccess2::DebugSocket -- debugging support # class DebugSocket < TCPSocket attr_accessor :debug_dev # Device for logging. class << self def create_socket(host, port, debug_dev) debug_dev << "! CONNECT TO #{host}:#{port}\n" socket = new(host, port) socket.debug_dev = debug_dev socket.log_connect socket end private :new end def initialize(*args) super @debug_dev = nil end def log_connect @debug_dev << '! CONNECTION ESTABLISHED' << "\n" end def close super @debug_dev << '! CONNECTION CLOSED' << "\n" end def gets(*args) str = super @debug_dev << str if str str end def read(*args) str = super @debug_dev << str if str str end def <<(str) super @debug_dev << str end end # HTTPAccess2::Session -- manage http session with one site. # One or more TCP sessions with the site may be created. # Only 1 TCP session is live at the same time. # class Session # :nodoc: class Error < StandardError # :nodoc: end class InvalidState < Error # :nodoc: end class BadResponse < Error # :nodoc: end class KeepAliveDisconnected < Error # :nodoc: end attr_reader :dest # Destination site attr_reader :src # Source site attr_accessor :proxy # Proxy site attr_accessor :socket_sync # Boolean value for Socket#sync attr_accessor :requested_version # Requested protocol version attr_accessor :debug_dev # Device for dumping log for debugging # These session parameters are not used now... attr_accessor :connect_timeout attr_accessor :connect_retry attr_accessor :send_timeout attr_accessor :receive_timeout attr_accessor :read_block_size attr_accessor :ssl_config def initialize(dest, user_agent, from) @dest = dest @src = Site.new @proxy = nil @socket_sync = true @requested_version = nil @debug_dev = nil @connect_timeout = nil @connect_retry = 1 @send_timeout = nil @receive_timeout = nil @read_block_size = nil @ssl_config = nil @user_agent = user_agent @from = from @state = :INIT @requests = [] @status = nil @reason = nil @headers = [] @socket = nil end # Send a request to the server def query(req) connect() if @state == :INIT begin timeout(@send_timeout) do set_header(req) req.dump(@socket) # flush the IO stream as IO::sync mode is false @socket.flush unless @socket_sync end rescue Errno::ECONNABORTED close raise KeepAliveDisconnected.new rescue if SSLEnabled and $!.is_a?(OpenSSL::SSL::SSLError) raise KeepAliveDisconnected.new elsif $!.is_a?(TimeoutError) close raise else raise end end @state = :META if @state == :WAIT @next_connection = nil @requests.push(req) end def close unless @socket.nil? @socket.flush @socket.close unless @socket.closed? end @state = :INIT end def closed? @state == :INIT end def get_status version = status = reason = nil begin if @state != :META raise RuntimeError.new("get_status must be called at the beginning of a session.") end version, status, reason = read_header() rescue close raise end return version, status, reason end def get_header(&block) begin read_header() if @state == :META rescue close raise end if block @headers.each do |line| block.call(line) end else @headers end end def eof? if @content_length == 0 true elsif @readbuf.length > 0 false else @socket.closed? or @socket.eof? end end def get_data(&block) begin read_header() if @state == :META return nil if @state != :DATA unless @state == :DATA raise InvalidState.new('state != DATA') end data = nil if block until eof? begin timeout(@receive_timeout) do data = read_body() end rescue TimeoutError raise end block.call(data) if data end data = nil # Calling with block returns nil. else begin timeout(@receive_timeout) do data = read_body() end rescue TimeoutError raise end end rescue close raise end if eof? if @next_connection @state = :WAIT else close end end data end private LibNames = "(#{RCS_FILE}/#{RCS_REVISION}, #{RUBY_VERSION_STRING})" def set_header(req) req.version = @requested_version if @requested_version if @user_agent req.header.set('User-Agent', "#{@user_agent} #{LibNames}") end if @from req.header.set('From', @from) end req.header.set('Date', Time.now) end # Connect to the server def connect site = @proxy || @dest begin retry_number = 0 timeout(@connect_timeout) do @socket = create_socket(site) begin @src.host = @socket.addr[3] @src.port = @socket.addr[1] rescue SocketError # to avoid IPSocket#addr problem on Mac OS X 10.3 + ruby-1.8.1. # cf. [ruby-talk:84909], [ruby-talk:95827] end if @dest.scheme == 'https' @socket = create_ssl_socket(@socket) connect_ssl_proxy(@socket) if @proxy @socket.ssl_connect @socket.post_connection_check(@dest) end # Use Ruby internal buffering instead of passing data immediatly # to the underlying layer # => we need to to call explicitely flush on the socket @socket.sync = @socket_sync end rescue TimeoutError if @connect_retry == 0 retry else retry_number += 1 retry if retry_number < @connect_retry end close raise end @state = :WAIT @readbuf = '' end def create_socket(site) begin if @debug_dev DebugSocket.create_socket(site.host, site.port, @debug_dev) else TCPSocket.new(site.host, site.port) end rescue SystemCallError => e e.message << " (#{site.host}, ##{site.port})" raise end end # wrap socket with OpenSSL. def create_ssl_socket(raw_socket) SSLSocketWrap.new(raw_socket, @ssl_config, (DEBUG_SSL ? @debug_dev : nil)) end def connect_ssl_proxy(socket) socket << sprintf("CONNECT %s:%s HTTP/1.1\r\n\r\n", @dest.host, @dest.port) parse_header(socket) unless @status == 200 raise BadResponse.new( "connect to ssl proxy failed with status #{@status} #{@reason}") end end # Read status block. def read_header if @state == :DATA get_data {} check_state() end unless @state == :META raise InvalidState, 'state != :META' end parse_header(@socket) @content_length = nil @chunked = false @headers.each do |line| case line when /^Content-Length:\s+(\d+)/i @content_length = $1.to_i when /^Transfer-Encoding:\s+chunked/i @chunked = true @content_length = true # how? @chunk_length = 0 when /^Connection:\s+([\-\w]+)/i, /^Proxy-Connection:\s+([\-\w]+)/i case $1 when /^Keep-Alive$/i @next_connection = true when /^close$/i @next_connection = false end else # Nothing to parse. end end # Head of the request has been parsed. @state = :DATA req = @requests.shift if req.header.request_method == 'HEAD' @content_length = 0 if @next_connection @state = :WAIT else close end end @next_connection = false unless @content_length return [@version, @status, @reason] end StatusParseRegexp = %r(\AHTTP/(\d+\.\d+)\s+(\d+)(?:\s+(.*))?\r?\n\z) def parse_header(socket) begin timeout(@receive_timeout) do begin initial_line = socket.gets("\n") if initial_line.nil? raise KeepAliveDisconnected.new end if StatusParseRegexp =~ initial_line @version, @status, @reason = $1, $2.to_i, $3 @next_connection = HTTP.keep_alive_enabled?(@version) else @version = '0.9' @status = nil @reason = nil @next_connection = false @readbuf = initial_line break end @headers = [] while true line = socket.gets("\n") unless line raise BadResponse.new('Unexpected EOF.') end line.sub!(/\r?\n\z/, '') break if line.empty? if line.sub!(/^\t/, '') @headers[-1] << line else @headers.push(line) end end end while (@version == '1.1' && @status == 100) end rescue TimeoutError raise end end def read_body if @chunked return read_body_chunked() elsif @content_length == 0 return nil elsif @content_length return read_body_length() else if @readbuf.length > 0 data = @readbuf @readbuf = '' return data else data = @socket.read(@read_block_size) data = nil if data.empty? # Absorbing interface mismatch. return data end end end def read_body_length maxbytes = @read_block_size if @readbuf.length > 0 data = @readbuf[0, @content_length] @readbuf[0, @content_length] = '' @content_length -= data.length return data end maxbytes = @content_length if maxbytes > @content_length data = @socket.read(maxbytes) if data @content_length -= data.length else @content_length = 0 end return data end RS = "\r\n" ChunkDelimiter = "0#{RS}" ChunkTrailer = "0#{RS}#{RS}" def read_body_chunked if @chunk_length == 0 until (i = @readbuf.index(RS)) @readbuf << @socket.gets(RS) end i += 2 if @readbuf[0, i] == ChunkDelimiter @content_length = 0 unless @readbuf[0, 5] == ChunkTrailer @readbuf << @socket.gets(RS) end @readbuf[0, 5] = '' return nil end @chunk_length = @readbuf[0, i].hex @readbuf[0, i] = '' end while @readbuf.length < @chunk_length + 2 @readbuf << @socket.read(@chunk_length + 2 - @readbuf.length) end data = @readbuf[0, @chunk_length] @readbuf[0, @chunk_length + 2] = '' @chunk_length = 0 return data end def check_state if @state == :DATA if eof? if @next_connection if @requests.empty? @state = :WAIT else @state = :META end end end end end end end HTTPClient = HTTPAccess2::Client   , Dec 22, 2005 #2 1. Advertisements 3. Guest Security alert: Don't use my code as-is! I checked with a sniffer and it looks like I'm attempting to send the password in the clear before sending the digest response. It looks like the flow is wrong . . . (Sorry for not testing this earlier!)   , Dec 25, 2005 #3 4. Guest I fixed the problem of always sending the password in the clear by checking the HTTP and making sure that the server is indeed requesting basic realm. In class BasicAuth I added lines with +: def set(uri, user_id, passwd) uri = uri.clone + # Make sure that the server is really requesting Basic Authentication! + serverRealm = (@client.head(uri).header['WWW-Authenticate']).join + return nil if ("Basic realm".downcase != serverRealm[0,11].downcase) uri.path = uri.path.sub(/\/[^\/]*$/, '/') @auth[uri] = ["#{user_id}:#{passwd}"].pack('m').strip @client.reset_all end I forgot to mention that the digest authentication in the code I posted earlier is hardcoded to build the digest response with "POST" which always works for me, as I'm sending SOAP. I'm not sure how to get the correct HTTP method. (http-access2 is about 1700 lines of code and I haven't had a chance to understand the flow . . . )   , Dec 26, 2005 #4 5. -----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Hi, Sorry I couldn't reply sooner. wrote: > I fixed the problem of always sending the password in the clear by > checking the HTTP and making sure that the server is indeed requesting > basic realm. In class BasicAuth I added lines with +: > > def set(uri, user_id, passwd) > uri = uri.clone > > + # Make sure that the server is really requesting Basic > Authentication! > + serverRealm = (@client.head(uri).header['WWW-Authenticate']).join > + return nil if ("Basic realm".downcase != > serverRealm[0,11].downcase) > > uri.path = uri.path.sub(/\/[^\/]*$/, '/') > @auth[uri] = ["#{user_id}:#{passwd}"].pack('m').strip > @client.reset_all > end I think I understood the problem but the problem is in BasicAuth#get, not in BasicAuth#set, right? http-access2 now sends password to a defined realm even if WWW-Authenticate is missing. I'll fix this. Thanks. Regards, // NaHi -----BEGIN PGP SIGNATURE----- Version: GnuPG v1.4.1 (Cygwin) iD8DBQFD5curf6b33ts2dPkRAhqxAJ9D5uOyVOpDuLSj3h2csm4n+RKXWQCgjUsE ynbn9wvGnjpZ4+jVC9GucmY= =k6RG -----END PGP SIGNATURE-----   NAKAMURA, Hiroshi, Feb 5, 2006 #5 1. Advertisements Want to reply to this thread or ask your own question? It takes just 2 minutes to sign up (and it's free!). Just click the sign up button to choose a username and then you can ask your own questions on the forum. Similar Threads 1. Wayne Carlaw Replies: 0 Views: 476 Wayne Carlaw Mar 29, 2005 2. Tammy Mc Replies: 3 Views: 314 Tammy Mc Oct 1, 2006 3. Ryan Q. Replies: 8 Views: 177 Ryan Q. Sep 7, 2008 4. myalo Replies: 4 Views: 1,650 A. Sinan Unur Nov 28, 2007 5. Replies: 2 Views: 467 Julian Cromarty Jun 26, 2013 Loading... Share This Page
__label__pos
0.52739
Epifanio Epifanio - 1 year ago 54 AngularJS Question Error access the properties of a query - AngularJS + Firebase I have problems with basic query firebase, it is with the rest of the application have no problem to bring data and work with them. And even with authentication firebase. However, with these queries, when I try to access its properties returns an undefined me. Although it is clearly defined. Does anyone see the stupid error? Thanks Versions with whom I work are: "dependencies": { "angular": "^1.5.8", "bootstrap": "^3.3.7", "angular-bootstrap": "^2.1.0", "angularfire": "^2.0.2", "firebase": "^3.3.2", "angular-animate": "^1.5.8", "angular-route": "^1.5.8", "textAngular": "^1.5.11", "angular-socialshare": "angularjs-socialshare#^2.3.1", "angular-timeago": "^0.4.3" } This is the view: <div ng-controller="blogpostCtrl"> {{post}} {{post.Title}} </div> This is the content of firebase: { "blog" : { "posts" : { "-KT9vxYoCkly93GqTSY7" : { "CreationDate" : 1475504757165, "Post" : "<p>esto es una prueba de post</p>", "Title" : "esto es una prueba", "Url" : "esto_es_una_prueba", "isPublished" : true }, "-KTA-f-BDAzC-4T4a50K" : { "CreationDate" : 1475505991860, "Post" : "<p>dsf sdf sdfs fd</p>", "Title" : "o1i21o3i", "Url" : "o1i21o3i", "isPublished" : true }, "-KTA-ggQSQLnDr4eQtfi" : { "CreationDate" : 1475505998783, "Post" : "<p>ds sdf sdf</p>", "Title" : "4 45654456 2546345562535", "Url" : "4_45654456_2546345562535", "isPublished" : true } } } } This is the controller: app.controller("blogpostCtrl", function($scope, $location, $routeParams, $firebaseArray) { var ref = firebase.database().ref("blog/posts"); $scope.post = $firebaseArray(ref.orderByChild("Url").equalTo($routeParams.uri)); console.log($scope.post) console.log($scope.post.Title) //<------------ undefined firebase.database().ref("blog/posts").orderByChild("Url").equalTo($routeParams.uri).once('value').then(function(snapshot) { console.log(snapshot.val()) console.log(snapshot.val().Title) //<------------ undefined }); firebase.database().ref("blog/posts").on('value', function(snapshot) { console.log(snapshot.val()) console.log(snapshot.val().Title) //<------------ undefined }); }); And this is the console output: []0: Object$$added: ()$$error: ()$$getKey: ()$$moved: ()$$notify: ()$$process: ()$$removed: ()$$updated: ()$add: ()$destroy: ()$getRecord: ()$indexFor: ()$keyAt: ()$loaded: ()$ref: ()$remove: ()$save: ()$watch: ()length: 1__proto__: Array[0] ctrl.js:163 undefined ctrl.js:170 Object {-KT9vxYoCkly93GqTSY7: Object, -KTA-f-BDAzC-4T4a50K: Object, -KTA-ggQSQLnDr4eQtfi: Object}-KT9vxYoCkly93GqTSY7: Object-KTA-f-BDAzC-4T4a50K: Object-KTA-ggQSQLnDr4eQtfi: Object__proto__: Object ctrl.js:171 undefined ctrl.js:166 Object {-KTA-ggQSQLnDr4eQtfi: Object}-KTA-ggQSQLnDr4eQtfi: ObjectCreationDate: 1475505998783Post: "<p>ds sdf sdf</p>"Title: "4 45654456 2546345562535"Url: "4_45654456_2546345562535"isPublished: true__proto__: Object__proto__: Object ctrl.js:167 undefined Answer Source .Title is a property of the child object so this is why you are getting undefined. If you need that Title you must include the object key as well, for example: console.log($scope.post.blog.posts['-KT9vxYoCkly93GqTSY7'].Title) Check this fiddle based on your json response http://jsfiddle.net/b2pfLp26/ About your $filterArray you have to wait before getting the console printed so you have to make use of .then. var array = $firebaseArray(ref.orderByChild("Url").equalTo($routeParams.uri)); array.$loaded().then(function(response){ console.log(response); });
__label__pos
0.793179
编译且移植FFTW3到Android手机上(2) 本文主要对如何将FFTW3编译且移植到Android App上进行介绍,同时对各FFTW提供的一些快速傅里叶变换的方法在手机进行性能测试,总结出使用FFTW3进行小规模傅里叶变换的最佳方式。 文章重点内容有:FFTW configure;编译so库;ARM NEON优化;float加速;多线程 第2部分为详细说明版,如想查看快速使用,请查看第1部分 : http://he-kai.com/?p=11 内容 详细代码:https://github.com/hekai/fftw_android 准备工作: 确保已完成 快速入门 部分的准备工作 编译FFTW: 打开fftw_android目录下的build.sh文件,默认的脚本中已启用neon优化和float支持 #!/bin/sh # Compiles fftw3 for Android # Make sure you have NDK_DIR defined in .bashrc or .bash_profile #NDK Version r9c http://dl.google.com/android/ndk/android-ndk-r9c-linux-x86_64.tar.bz2 NDK_DIR="/home/hekai/software/android-ndk-r9c" INSTALL_DIR="`pwd`/jni/fftw3" SRC_DIR="`pwd`/../fftw-3.3.3" cd $SRC_DIR export PATH="$NDK_DIR/toolchains/arm-linux-androideabi-4.8/prebuilt/linux-x86_64/bin/:$PATH" export SYS_ROOT="$NDK_DIR/platforms/android-8/arch-arm/" export CC="arm-linux-androideabi-gcc --sysroot=$SYS_ROOT -march=armv7-a -mfloat-abi=softfp" export LD="arm-linux-androideabi-ld" export AR="arm-linux-androideabi-ar" export RANLIB="arm-linux-androideabi-ranlib" export STRIP="arm-linux-androideabi-strip" #export CFLAGS="-mfpu=neon -mfloat-abi=softfp" mkdir -p $INSTALL_DIR ./configure --host=arm-eabi --prefix=$INSTALL_DIR LIBS="-lc -lgcc" --enable-float --enable-threads # --with-combined-threads --enable-neon make make install exit 0 将NDK_DIR的路径修改为本机实际的NDK路径,如果版本不是r9c,请查看$NDK_DIR/toolchains/arm-linux-androideabi-4.8/prebuilt/linux-x86_64/bin/该路径是否存在,如不存在,请从NDK根目录开始找,寻找类似命名的文件夹,并将最终路径替换到build.sh中的export PATH=中 为验证编译效果,可将fftw_android/jni/fftw3中的文件夹全部删除,Android.mk留下! 打开终端,进入fftw_android所在目录,输入./build.sh,回车即可,终端上会显示编译过程,先会检查各编译器是否OK,再生成相应的config.h,最后再生成相关.a库文件,并自动复制到fftw_android/jni/fftw3目录下 通过默认的build.sh生成的只是支持float,neon和threads库文件(libfftwf3.a和libfftwf3_threads.a),由于代码中还用到了double做对比测试,请将–enable-float,–enable-threads,–enable-neon对应的行用#号注释,将export CC中的-march=armv7-a -mfloat-abi=softfp直接删除,再运行./build.sh,从而生成支持double的库文件(libfftw3.a) 编写代码: .a库生成好后,通过已写好的Android.mk在NDK或者Eclipse中即可生成完整so文件,下面描述如何使用这些so库进行编程 下面的过程就是常见的JNI编程过程,基本代码搭建示例很多,这里不做过多阐述,最简单的示例程序在NDK路径下的sample文件中,有hello-jni项目可供参考。大概过程是:Java代码中定义native的方法,并loadLibrary将后续要生成的so库载入内存;在jni中编写相应c/cpp文件实现native方法;Android.mk中定义好so库名,以及代码文件和依赖的库。 现主要对如何使用FFTW进行描述: 为方便测试,预先定义了几个常用方法如下: #define SIZE 160 //SIZE x SIZE , default: 160 x 160 int init_in_fftw_complex(fftw_complex* in){ int i,j,index; for (i = 0; i < SIZE; i++) { for (j = 0; j < SIZE; j++) { index = j + i * SIZE; in[index][0] = index + 1; in[index][1] = 0; } } return 0; } int init_in_fftwf_float(float* in){ int i,j,index; for (i = 0; i < SIZE; i++) { for (j = 0; j < SIZE; j++) { index = j + i * SIZE; in[index] = (float)(index + 1); } } return 0; } static double now_ms(void) { struct timeval tv; gettimeofday(&tv, NULL); return tv.tv_sec*1000. + tv.tv_usec/1000.; } 上述代码主要用于初始化二维矩阵,以及获取当前时间用来判断消耗。 基本 先写一个最基本的二维矩阵进行快速傅里叶变换,代码如下: JNIEXPORT jstring JNICALL Java_com_hekai_fftw_1android_Utils_fftw_1dft_12d( JNIEnv * env, jobject thiz) { double t_start, t_end, t_span; t_start = now_ms(); fftw_complex *in, *out; fftw_plan p; in = (fftw_complex*) fftw_malloc(sizeof(fftw_complex) * SIZE * SIZE); out = (fftw_complex*) fftw_malloc(sizeof(fftw_complex) * SIZE * SIZE); init_in_fftw_complex(in); p = fftw_plan_dft_2d(SIZE, SIZE, in, out, FFTW_FORWARD, FFTW_ESTIMATE); fftw_execute(p); fftw_destroy_plan(p); fftw_free(in); fftw_free(out); t_end = now_ms(); t_span = t_end - t_start; LOGD("fftw_dft_2d() costs time %f ms", t_span); return (*env)->NewStringUTF(env, "fftw_dft_2d"); } 上述代码使用的是fftw_complex结构体,这会使用double类型做运算,需要依赖的库为libfftw3.a。通过fftw_plan_dft_2d 创建plan,并调用fftw_execute执行plan(可多次调用)。执行后,如果想查看结果,可以遍历out来查看,需要注意的是out是二维的,out[index][0]和out[index][1]中分别为结果的实数和虚数部分。运行结果可与Matlab或Octave的结果进行对比,以确定是否运算正确。 纯实数 针对纯实数矩阵,fftw提供有专门的函数加快运算过程,具体为携带有r2c标识的,如下所示: JNIEXPORT jstring JNICALL Java_com_hekai_fftw_1android_Utils_fftw_1dft_1r2c_12d( JNIEnv * env, jobject thiz) { double t_start, t_end, t_span; t_start = now_ms(); fftw_complex *in, *out; fftw_plan p; int NTmp = floor(SIZE/2 +1); in = (fftw_complex*) fftw_malloc(sizeof(fftw_complex) * SIZE * SIZE); out = (fftw_complex*) fftw_malloc(sizeof(fftw_complex) * SIZE * NTmp); init_in_fftw_complex(in); p = fftw_plan_dft_r2c_2d(SIZE, SIZE, in, out, FFTW_ESTIMATE); fftw_execute(p); fftw_destroy_plan(p); fftw_free(in); fftw_free(out); t_end = now_ms(); t_span = t_end - t_start; LOGD("fftw_dft_r2c_2d() costs time %f ms", t_span); return (*env)->NewStringUTF(env, "fftw_dft_r2c_2d"); } 需要注意的是,纯实数矩阵做傅里叶变换后的结果具有对称性,因此out里没有N x N的结果,而是只有N x floor(N/2 + 1)的结果,遍历out时需注意。 Float 当不要求double的运算精度时,FFTW支持float运算,从而节省相应时间,代码如下: JNIEXPORT jstring JNICALL Java_com_hekai_fftw_1android_Utils_fftwf_1dft_1r2c_12d( JNIEnv * env, jobject thiz) { double t_start, t_end, t_span; t_start = now_ms(); float *in; fftwf_complex *out; fftwf_plan p; int NTmp = floor(SIZE / 2 + 1); in = (float*) fftw_malloc(sizeof(float) * SIZE * SIZE); out = (fftwf_complex*) fftwf_malloc(sizeof(fftwf_complex) * SIZE * NTmp); init_in_fftwf_float(in); p = fftwf_plan_dft_r2c_2d(SIZE, SIZE, in, out, FFTW_ESTIMATE); fftwf_execute(p); fftwf_destroy_plan(p); fftwf_free(in); fftwf_free(out); t_end = now_ms(); t_span = t_end - t_start; LOGD("fftwf_dft_r2c_2d() costs time %f ms", t_span); return (*env)->NewStringUTF(env, "fftwf_dft_r2c_2d"); } 使用float运算时需注意,输入变为float*了,且相关的方法或数据结构都需要变为fftwf_***,即增加一个f,其他过程一样。 多线程 FFTW还支持多线程,但在小规模时反而会拖慢时间,具体代码如下: JNIEXPORT jstring JNICALL Java_com_hekai_fftw_1android_Utils_fftwf_1dft_1r2c_12d_1thread( JNIEnv * env, jobject thiz) { double t_start, t_end, t_span; t_start = now_ms(); int thread_ok = 0; int n_threads = 4; float *in; fftwf_complex *out; fftwf_plan p; int NTmp = floor(SIZE / 2 + 1); in = (float*) fftw_malloc(sizeof(float) * SIZE * SIZE); out = (fftwf_complex*) fftwf_malloc(sizeof(fftwf_complex) * SIZE * NTmp); init_in_fftwf_float(in); thread_ok = fftwf_init_threads(); if(thread_ok) fftwf_plan_with_nthreads(n_threads); p = fftwf_plan_dft_r2c_2d(SIZE, SIZE, in, out, FFTW_ESTIMATE); fftwf_execute(p); fftwf_destroy_plan(p); if(thread_ok) fftwf_cleanup_threads(); fftwf_free(in); fftwf_free(out); t_end = now_ms(); t_span = t_end - t_start; LOGD("fftwf_dft_r2c_2d_thread() costs time %f ms. thread_ok = %d", t_span, thread_ok); return (*env)->NewStringUTF(env, "fftwf_dft_r2c_2d_thread"); } 使用多线程需要在对FFTW做configure时,开启–enable-threads,代码中通过thread_ok = fftwf_init_threads();获取当前是否支持多线程,1为支持。后续再通过fftwf_plan_with_nthreads(n_threads);指定线程数,以及fftwf_cleanup_threads();回收资源。 NEON优化 对于Android手机,如果CPU是ARM-v7架构的,可编译neon优化的so库加快运算速度,具体如下: 在build.sh中开启–enable-float和–enable-neon,并在CC中加入-march=armv7-a -mfloat-abi=softfp 。要开启NEON优化,必须开启float,即double运算不能进行NEON优化。最根本的原因主要是因为ARM-v7架构的问题,寄存器不够NEON优化来做double运算,而新的64位架构的ARM-v8a则解决了这个问题,不够需要重新porting相关NEON优化代码。ARM架构更多的信息请查阅ARM官网相关文档,这里不再做展开。 开启NEON优化后,c/cpp代码无需更改,直接可以运行查看运行效果。 在手机上,经过NEON优化后的so库运行效率有明显提高,当然不同的CPU可能有不同的改善效果… 总结: • 在Android上使用so库,能支持NEON优化的,尽量使用以提高效率。 • FFTW的多线程在小规模时可以不用开启了,具体阈值多大需要具体去测试才行。 • float和r2c的方法请根据实际情况选用。 参考文献: 1. Compiling open source libraries with Android NDK: Part 2 2. 一个FFTW NDK的例子 https://github.com/jimjh/fftw-ndk-example 3. FFTW官方文档 http://fftw.org/fftw3_doc/   27 thoughts to “编译且移植FFTW3到Android手机上(2)” 1. Hi , 在搜尋了網上大部分的文章,您是我在網上唯一看到有編譯出NDK的FFTW的強者; 在使用上有些問題想請教您 我直接將您fftw-ndk-example project jni下的fftw拿出作為prebuilt-static-libary, include ; 如此可以編譯成功並且執行,但程式會卡在convfftm()之類的function沒有反應,但也不會報錯,請問是因為您文章提到的”通过默认的build.sh生成的只是支持float,neon和threads库文件(libfftwf3.a和libfftwf3_threads.a)” 所以convfftm()之類的function並沒有支援的原因嗎 若是如此該如何解決呢 謝謝 2. 抱歉 搞錯了 卡住的function 是 fftw_plan_dft_1d,是否要呼叫同功能的function是要呼叫您命名的新function 感謝 1. 如果你使用的是float的so库,请使用FFTW中的fftwf_plan_dft_1d的函数,即所有的fftw后再加上f,代表调用的是相对于的float的函数,需要注意的是,有些函数内可能还需要传入些参数,可能也要是fftwf_xx类型。具体你可以对比文章中对于Double和Float的代码区别。 1. 对于Android来说,用到的是so库,而fftw编译生成的是静态库,即这里看到的.a库,无法直接在Android里loadLibrary来使用,因此我们通过配置Android.mk,编写jni文件去封装,达到调用.a库中的方法的目的 3. 楼主,我现在在寻求一种代码优化的解决方案,我用来测试的手机是小米4,我查了一下,其处理器是骁龙801,支持“ARMv7, NEON”,而且是一枚32位处理器。我的代码中需要进行double运算,这样的话,优化方案只能设定为“r2c”这一个,“NEON”这种优化就用不了了是吧。 4. 楼主,有时候我会显得很啰嗦,希望你不要介意。我还有一个问题,就是关于FFTW_MEASURE和FFTW_ESTIMATE的使用,不知道你有没有类似使用体会,我在网上查了一些资料,说是FFTW_MEASURE可以使用在“一次初始化,加快多次使用计算速度”,我看到一篇文档里面这样说到:FFTW_MEASURE:告诉FFTW通过几种算法计算输入阵列实际所用的时间来确定最优算法。根据你的机器硬件配置情况,可能需要点时间(通常为几秒钟),该参数是默认选项。我的问题中计算的是一维数组,长度为8192*2 = 16384,循环计算的次数为100~500次不等。不知道楼主是否有类似的处理经验可以分享一下。 1. 这一块我后来没有进行更细的调研了,不过你可以看下https://github.com/hekai/fftw_android上面贴的图表,在使用ffwtf_plan__dft_r2c_2d(neon)时有跟你的描述类似的效果,第一次时间较长,后面的运算速度明显提升。我不确定这里面是否默认使用了你所提到的fftw_measure方法,具体可能得你多测试对比下,并看下fftw的源码实现 5. 楼主,我又来啦,嘻嘻,我有个困惑,就是我的问题中需要用到复数乘法,也就是fftw_complex的乘法,我今天试了一下直接相乘的语句,也就是complex1*complex2,发现竟然没有报错,不知道是什么原因,难道是因为fftw3.h中已经重载了fftw_complex类型的乘法运算吗。 1. 这种情况我也不清楚,你看下结果正常不,如果用float的话最好还是用fftwf_complex吧,省得以后出现结果不对,诡异难查的问题 6. 楼主,您好,我还是上次问你问题的那位同学。我这次来,是有个不情之请,不知道你是否还保存有博文中说的支持double的库文件(libfftw3.a),其实我原本是想自己来做的,后来由于自己重装了系统,如果自己生成还需要配置虚拟机等复杂的步骤,而我由于小论文的压力,需要尽快对自己的代码进行测试。如果楼主还保留double的库文件(libfftw3.a)的库文件的话,希望能发送给我一份,我衷心的表示十分感谢,麻烦您了。这是我的邮箱:[email protected] 1. 我这也没有现成的支持double的库,建议您还是自行编译个,后续你也可以自行更改编译参数进行对比测试,具体方法文章中有讲到,即修改build.sh里的配置,你全新安装环境并编译出a文件应该在1天内可以搞定,如果对Linux比较熟悉,那么时间可以更短。 7. 楼主您好,显示了这样的信息:checking for a BSD-compatible install… /usr/bin/install -cchecking whether build environment is sane… yeschecking for arm-eabi-strip… arm-linux-androideabi-stripchecking for a thread-safe mkdir -p… /bin/mkdir -pchecking for gawk… nochecking for mawk… mawkchecking whether make sets $(MAKE)… yeschecking whether make supports nested variables… yeschecking whether to enable maintainer-specific portions of Makefiles… nochecking build system type… x86_64-unknown-linux-gnuchecking host system type… arm-unknown-eabichecking for arm-eabi-gcc… arm-linux-androideabi-gcc –sysroot=/Home/Program_Files/Path/android-ndk-r11b/platforms/android-19/arch-arm/ checking whether the C compiler works… noconfigure: error: in `/home/songyuc/Documents/Code/FFTW/fftw-3.3.4′:configure: error: C compiler cannot create executablesSee `config.log’ for more detailsmake: *** 没有指明目标并且找不到 makefile。 停止。make: *** No rule to make target ‘install’。 停止。 1. 你可以自己查看下’config.log’里的log,看下具体,从目前的log来看,checking whether the C compiler works… no,猜测是这里的问题,少什么装什么。简单粗暴的方法是去 https://source.android.com/source/initializing.html 上面根据自己的Ubuntu版本直接把编译整个Android源码的依赖软件都安装上,也可以通过百度之类搜索,看别人安装了哪些依赖。我一般都是配置好可以编译整个源码的环境 8. 楼主,在Ubuntu18.04版本上,编译遇到问题: …… config.status: executing depfiles commands config.status: executing libtool commands ./build.sh: 29: ./build(ndk r9c).sh: –enable-neon: not found ./build.sh: 31: ./build(ndk r9c).sh: make: not found ./build.sh: 32: ./build(ndk r9c).sh: make: not found 已经试了一天了,除了系统不一样其他都是一样的,就是编译不成功。试过各种版本ndk了,就r9c这个版本是最接近成功的。。。想请教一下,该怎么解决。 发表评论 电子邮件地址不会被公开。 必填项已用*标注
__label__pos
0.537811
ParaView Visualization The section ParaView Support describes how to install Ascent with ParaView support and how to run the example integrations using insitu ParaView pipelines. In this section we describe in detail the ParaView visualization pipeline for cloverleaf3d, one of the example integrations, and we provide implementation details for the Ascent ParaView integration. The ParaView pipeline for cloverleaf3d First we need to tell Ascent that we are using a Python script to visualize data using ascent-actions.json. [ { "action": "add_extracts", "extracts": { "e1": { "type": "python", "params": { "file": "paraview-vis.py" } } } } ] The ParaView pipeline for the cloverleaf3d sample integration is in paraview-vis.py. We use a variable count to be able to distinguish timestep 0, when we setup the visualization pipeline. For all timesteps including timestep 0, we execute the visualization pipeline we setup at timestep 0. try: count = count + 1 except NameError: count = 0 For timestep 0, we initialize ParaView, if count == 0: import paraview paraview.options.batch = True paraview.options.symmetric = True then we load the AscentSource plugin and we create the object that presents the simulation data as a VTK dataset. We also create a view of the same size as the image we want to save. # LoadPlugin("@PARAVIEW_ASCENT_SOURCE@", remote=True, ns=globals()) ascentSource = AscentSource() view = CreateRenderView() view.ViewSize = [1024, 1024] From the VTK dataset, we select only the cells that are not ghosts and show them colored by the energy scalar. Note that for a ParaView filter that has no input specified, the output data from the previous command in the program is used. So SelectCells uses the output data from ascentSource. # sel = SelectCells("vtkGhostType < 1") e = ExtractSelection(Selection=sel) rep = Show() ColorBy(rep, ("CELLS", "energy")) We rescale the transfer function, show a scalar bar, and change the viewing direction # transferFunction = GetColorTransferFunction('energy') transferFunction.RescaleTransferFunction(1, 5.5) renderView1 = GetActiveView() scalarBar = GetScalarBar(transferFunction, renderView1) scalarBar.Title = 'energy' scalarBar.ComponentTitle = '' scalarBar.Visibility = 1 rep.SetScalarBarVisibility(renderView1, True) cam = GetActiveCamera() cam.Elevation(30) cam.Azimuth(-30) For all timesteps, UpdateAscentData sets the new Ascent data and marks the VTK source as modified. This insures that a new VTK dataset will be computed when we need to Render. We also call UpdatePropertyInformation which insures that property values are available to the script. There are two properties setup on AscentSource: Time (this represents the simulation time and is the same as state/time in Conduit Blueprint Mesh specification) and Cycle (this represents the simulation time step when the visualization pipeline is called and is the same as state/cycle in Conduit Blueprint Mesh specification). After that, we ResetCamera so that the image fits the screen properly, we render and save the image to a file. ascentSource.UpdateAscentData() ascentSource.UpdatePropertyInformation() cycle = GetProperty(ascentSource, "Cycle").GetElement(0) imageName = "image_{0:04d}.png".format(int(cycle)) ResetCamera() Render() SaveScreenshot(imageName, ImageResolution=(1024, 1024)) This script saves an image for each cycle with the image for cycle 200 shown next. ../_images/paraview_clover_example.png Fig. 44 CloverLeaf3D visualized with a ParaView pipeline Implementation details The Ascent ParaView integration is implemented in the src/examples/paraview-vis directory in the Ascent distribution. AscentSource class, found in paraview_ascent_source.py, derives from VTKPythonAlgorithmBase and produces one of the following datasets: vtkImageData, vtkRectilinearGrid, vtkStructuredGrid or vtkUnstructuredGrid. AscentSource receives from an instrumented simulation a tree structure (json like) that describes the simulation data using the Conduit Blueprint Mesh specification. This data is converted to a VTK format using zero copy for data arrays. Global extents are not passed for the existing example integrations so they are computed (using MPI communication) for uniform and rectilinear topologies but they are not computed for a structured topology (lulesh integration example). This means that for lulesh and datasets that have a structured topology we cannot save a correct parallel file that represents the whole dataset, unless the global extents are passed from the simulation. A ParaView pipeline for each sample simulation is specified in a paraview-vis-XXX.py file where XXX is the name of the simulation. In this file, we load the ParaView plugin and setup the pipeline for timestep 0 and update the pipeline and save a screenshot for each timestep of the simulation.
__label__pos
0.98761
Binding Methodologies in wireless system Question: What are the different binding methodologies that we can have in a generic wireless system?   Answer:          Question:  Repsonse: Binding is the mechanism by which a network is created or by which new devices are added into an existing network. Binding can be achieved in many different ways. Examples of binding options include: ■ Binding in the manufacturing process, such as through use of special test software/hardware ■ Binding that occurs as a result of a user action, such as pressing buttons ■ Binding that occurs upon the power up sequence of a device ■ Binding that can occur in a dynamic or ad-hoc basis, such as whenever a device comes in proximity of a network ■ Completely manual binding, such as selection of the channel through a switch, or entering parameters via a software interface.
__label__pos
0.999451
Arrays Arrays Now you know how to create variables and functions, and you know how to use for loops to repeat a block of code. So far, the variables you’ve seen have held a single value. This tutorial introduces arrays, which hold multiple values. Multiple Variables Let’s start with an example sketch: This sketch uses a circleY variable to show a circle falling down the screen. Incrementing the circleY each frame causes it to fall. The if statement detects when the circle reaches the bottom, and resets the circle back to the top of the screen. falling circle The Bad Way What if you want to add another circle? You might be tempted to use another variable: This code uses two variables: circleYOne and circleYTwo to show two circles that fall from the top of the screen. two falling circles Creating an Array What if you wanted to add a third circle? Or ten more circles? You could keep adding variables, but that’s going to make your program very long and hard to work with. Instead, you can use an array. An array is a single variable that holds multiple values. Remember that to create a variable you use the let keyword to give it a name and a value. Creating an array is similar: • Use the let keyword. You can also use the const keyword, but to keep it simple I’m going to stick with let. • Give it a name. • Give it an array value. An array value is multiple values inside square brackets [] and separated by commas. For example, this line of code creates an array named circleY that holds two values, 10 and 20: let circleY = [10, 20]; Accessing an Array An array is a variable that holds multiple values. To use an individual value inside an array, you can use the array access operator. The array access operator is a number value inside square brackets []. That number provides the index of the array value that you want to use. For example, this line of code accesses the first and second values from the array to draw two circles: circle(100, circleY[0], 25); circle(200, circleY[1], 25); This line of code does the same thing as before, but now it’s getting the values from an array instead of two separate variables. Start at Zero You might notice that the code uses 0 instead of 1 to get the first value from the array. That’s because array indexes start at zero! The second value from the array has 1 as an index: // draw the second circle circle(200, circleY[1], 25); This can be pretty confusing, but remember that array indexes start at zero. So if you have an array with ten values, the last index is 9. Setting an Array Index Just like you can modify the value a variable holds, you can modify the value an array index holds. This line of code reassigns the first index of the array to a new value: circleY[0] = 42; And this line of code adds 5 to the first array index: circleY[0] = circleY[0] + 5; Which can be shortened to: circleY[0] += 5; The Bad Way with Arrays Putting it all together, you could rewrite the sketch to use arrays instead of single-value variables: I’m using this example to show how arrays work, but you wouldn’t actually write code like this. If you added a third value to the circleY array, you’d still need to add the code that uses that new value. That’s going to get very annoying! Instead of copying the same line of code over and over again, you can use for loops to make your life easier. For Loops Let’s say the circleY array holds five values. You can write code that draws five circles: circle(50, circleY[0], 25); circle(100, circleY[1], 25); circle(150, circleY[2], 25); circle(200, circleY[3], 25); circle(250, circleY[4], 25); (I’m leaving out the code for moving and resetting the circles, but imagine how long that code would be!) This will work, but notice that this code contains a pattern: it uses an index that starts at 0, increases by 1, and stops at 4. That means you can rewrite this code to use a for loop instead! for (let i = 0; i < 5; i++) { circle(50 * (i+1), circleY[i], 25); } This code uses a for loop with a loop variable i that goes from 0 to 4. When the i variable reaches 5, then i < 5 evaluates to false and the loop exits. Inside the body of the loop, the code uses that loop variable to access every index of the array. It also uses that loop variable to calculate the x value of each circle. You can rewrite the code to use a for loop: five falling circles And that’s the cool thing about arrays, especially when you use for loops with them: you now have 5 falling circles, without any extra code! You only have to write the code that draws, moves, and resets a circle once, and then you can apply that code to every circle in the array. Array Length When you’re using a for loop with an array, you have to know how many values are in the array, so you know which index to stop at. When there are two values, the for loop looks like this: for (let i = 0; i < 2; i++) { And when there are ten values, the for loop looks like this: for (let i = 0; i < 10; i++) { In other words, you always want to stop the loop when its loop variable equals the number of elements in the array, which is also called the length of the array. If you try to access an index that’s larger than the length, you’ll get an error! So if you add a variable to the array initialization (the values in the square brackets []), you’ll have to change the check in the for loop. Wouldn’t it be nice if the computer could keep track of that for you? You guessed it: the computer does keep track of the length of an array! To use the length value, you type .length after the name of an array: let numberOfValues = circleY.length; You can use this length variable exactly like you can any other variable, including in a for loop check: for (let i = 0; i < circleY.length; i++) { Now if you add values to the array, you no longer have to modify the for loop check yourself. The length variable will always contain the length of the array, so the for loop will work no matter how many elements the array contains. Delayed Initialization Remember that declaring a variable means creating in by giving it a name, and initializing a variable means giving it a starting value. Reassigning a variable means changing its value. So far, the code above has initialized arrays as soon as it declares them, using values inside square brackets []: let circleY = [50, 100, 150, 200, 250]; But what if you don’t know what the values should be yet? In this case, you can delay the initialization of the array. To create an array without initializing its values, you can use empty square brackets []. This line of code creates an empty array: let circleY = []; Now you can set the value of each of the indexes individually: circleY[0] = 50; circleY[1] = 100; circleY[2] = 150; circleY[3] = 200; circleY[4] = 250; Or better yet, you can use a for loop: for (let i = 0; i < circleY.length; i++) { circleY[i] = (i + 1) * 50; } The Payoff Putting all of this together, here’s an example that shows 25 falling circles: 25 falling circles Imagine how much code this would take if it wasn’t using arrays and for loops! Challenge: Change this code to show 100 falling circles, all falling at different speeds! Summary Arrays are variables that hold multiple values. By combining them with for loops, you can write programs that handle a lot of data in just a few lines of code. Homework • Create a sketch that shows rain drops or snow flakes falling. • Create a sketch that shows a trail of circles that follow the mouse. Hint:. store the previous 25 positions of the mouse in an array and draw those to the screen! Arrays Examples Comments Happy Coding is a community of folks just like you learning about coding. Do you have a comment or question? Post it here! Comments are powered by the Happy Coding forum. This page has a corresponding forum post, and replies to that post show up as comments here. Click the button above to go to the forum to post a comment!
__label__pos
0.919426
How to Clean an Xbox One S Fan Without Opening? Are you an Xbox One S owner? If so, then you might be wondering how to clean the fan without opening up the console. The Xbox One S is a great console, but like all electronics, it can get dirty over time. The fan is one of the most important parts of the console, and if it gets too dirty, it can start to cause problems. In this article, we’ll show you how to clean the fan on your Xbox One S without opening up the console. We’ll also give you some tips on how to prevent the fan from getting dirty in the first place. How to Clean an Xbox One S Fan Without Opening? What You’ll Need To clean the fan on your Xbox One S, you’ll need the following supplies: • Isopropyl alcohol • Cotton swabs • A soft, clean cloth Step 1: Turn Off Your Xbox One S Before you start cleaning, you’ll need to turn off your Xbox One S. To do this, press and hold the power button on the front of the console for 10 seconds. See also  How Do I Delete a Team in Yahoo Fantasy Football? Step 2: Remove the Xbox One S Power Supply Next, you’ll need to remove the power supply from the Xbox One S. To do this, locate the power supply on the back of the console. There are two screws holding the power supply in place. Remove these screws and then carefully pull the power supply out of the console. Step 3: Clean the Fan Blade Once you have the power supply removed, you’ll be able to see the fan blade. Take a cotton swab and dip it in some isopropyl alcohol. Gently wipe the cotton swab around the fan blade to remove any dust or dirt. Be careful not to touch the fan with your fingers as this could damage it. Step 4: Clean the Fan Housing The next step is to clean the fan housing. This is the part of the console where the fan blade is located. Take a cotton swab and dip it in some isopropyl alcohol. Gently wipe the cotton swab around the inside of the fan housing. See also  How Strong Is Gorr the God Butcher? Be careful not to touch the fan blade with the cotton swab as this could damage it. Step 5: Clean the Outside of the Console Once you’ve cleaned the fan, you’ll need to clean the outside of the console. Use a soft, clean cloth to wipe down the outside of the Xbox One S. Make sure to get all of the nooks and crannies, including the vents on the side of the console. Step 6: Put Everything Back Together Once you’ve cleaned everything, it’s time to put the Xbox One S back together. Start by reconnecting the power supply to the back of the console. Then, screw the power supply back into place. Finally, turn on your Xbox One S by pressing the power button on the front of the console. How to Prevent the Fan from Getting Dirty? Now that you know how to clean the fan on your Xbox One S, you might be wondering how you can prevent it from getting dirty in the first place. See also  What Is Evil Dead Game? Here are some tips: 1. Don’t keep your Xbox One S in an enclosed space. This will allow dust and dirt to build up on the console. 2. Keep your Xbox One S away from pets. Pets can shed hair and dander that can end up in the console. 3. Don’t smoke near your Xbox One S. Smoke can settle on the console and cause the fan to get dirty. Conclusion Cleaning the fan on your Xbox One S is important to keep your console running smoothly. Luckily, it’s not difficult to do and only takes a few minutes. Follow the steps in this article and you’ll have your Xbox One S fan cleaned in no time.
__label__pos
0.919244
CPM Homework Banner 12-58. Using the triangle at right, determine: 1.   Remember that is opposite over hypotenuse. 1.   Which trigonometric ratio is  related to? How is it related? 1.   1.   How is related to ? Right triangle: horizontal side, A, C, labeled, b, vertical side, C, b, labeled, a, hypotenuse side, A, b, labeled, c.
__label__pos
1
Force connection to wifi network while at work even if phone tries to connect to other networks Learn how you can make your phone always connect to the same networks if it wants otherwise 1. R4V3N_2010 R4V3N_2010 5/5, Exactly what i need because my phone drops the wifi connection all the time. Thank you
__label__pos
0.989507
  Beyond.ca Registration is free! Car Forums - Member Rides Car Forums - Find other members Car Forums - Calendar Car Forums - Frequently Asked Questions Forum Rules and Regulations Car Forums - Search Logout What is vB Code? vB code is a set of tags based on the HTML language that you may already be familiar with. They allow you to add formatting to your messages in the same way as HTML does, but have a simpler syntax and will never break the layout of the pages you are viewing. The ability to use vB Code is set on a forum-by-forum basis by the administrator, so you should check the forum rules when you post a new message. URL Hyperlinking If vB Code is enabled in a forum, you can simply type in the full address of the page you are linking to, and the hyperlink will be created automatically. Here are some example links: • http://www.vbulletin.com/forum/ • www.vbulletin.com Notice that if the address begins with www. you do not need to add the http:// part of the address. If the address does not begin with www. you will need to add the http:// section. You may also use https:// and ftp:// links, and these will be converted into links. If you want to include the vB Code, you may simply surround the address with [url] tags as shown below. (The vB Code tags are shown in red). [url]www.vbulletin.com/forum/[/url] You can also include true hyperlinks using the [url] tag. Just use the following format: [url=http://www.vbulletin.com/forum/]Click here to visit the vBulletin forums[/url] This will produce a hyperlink like this: Click here to visit the vBulletin forums. Note that once again, you need not include the http:// if the address begins www. Email Links To add a link to an email address, you can simply include the email address in your message like this: [email protected] Note that there must be a blank space, such as a space or a carriage return before the beginning of the address. You can also use vB Code tags to specify an email address, like this: [email][email protected][/email] You can also add a true email hyperlink using the following format: [[email protected]]Click here to email me[/email] This will produce a hyperlink like this: Click here to email me. Bold, Underlined and Italic Text You can make text bold, underlined or italicized by simply surrounding your text with tags as shown below: • [b]some text[/b] produces some text • [u]some text[/u] produces some text • [i]some text[/i] produces some text Using Different Colors, Sizes and Fonts You can alter the size, color and font of text using the following tags: • [color=blue]some text[/color] produces some text (colored blue) • [size=4]some text[/size] produces some text (size 4 text) • [font=courier]some text[/font] produces some text (using courier font) You can also combine all the various text formatting tags. This example uses bold, underlined, purple text: [color=purple][u][b]Wow there's lots of formatting here![/b][/u][/color] This example produces this: Wow there's lots of formatting here! Bullets and Lists You can create bulleted or ordered lists in the following way: Unordered, bulleted list: [list] [*]first bulleted item [*]second bulleted item [/list] This produces: • first bulleted item • second bulleted item Note that you must remember to close the list with the [/list] tag. If you would like to create a list ordered numerically or alphabetically, this is just as easy. You simply need to add a little extra code to your [list] and [/list] tags. The extra code looks like =1 (for a numbered list) or =A (for a list from A to Z). Here are some examples: [list=1] [*]this is the first numbered item [*]this is the second numbered item [/list=1] This produces: 1. this is the first numbered item 2. this is the second numbered item [list=A] [*]this is the first alphabetically ordered item [*]this is the second alphabetically ordered item [/list=A] This produces: 1. this is the first alphabetically ordered item 2. this is the second alphabetically ordered item Adding Images To include a picture or graphic within the body of your message, you can simply surround the address of the image as shown here: [img]http://forums.beyond.ca//beyond-b.png[/img] Note that the http:// part of the image URL is required for the [img] code. You can even create a thumbnail-type hyperlink by surrounding your [img] code with a [url] code like this: [url=http://forums.beyond.ca//beyond-b.png][img]http://forums.beyond.ca//images/vb_bullet.gif[/img][/url] This produces a link like this: 0. Quoting Other Messages To quote something that has already been posted, simply cut-and-paste the text you want to quote, and enclose it as follows: [quote]No. Try not. Do or do not, there is no try.[/quote] The [quote] tags will automatically indent the enclosed text. The Code and PHP Tags If you want to post some programming source code, or perhaps some ASCII art, which would require a non-proportional font, you can use the [code] tag to achieve this. For example: [code] <script language="Javascript"> <!-- alert("Hello world!"); //--> </script> [/code] In the example above, the text enclosed in the [code] tags would be automatically indented, and the spacing would be preserved like this: <script language="Javascript"> <!-- alert("Hello world!"); //--> </script> A special case is for code written in the PHP language. If you are posting PHP code, you can enclose the source code in [php] tags, and the script will automatically have syntax highlighting applied: [php] $myvar = "Hello World!"; for ($i=0; $i<10; $i++) { echo $myvar."\n"; } [/php] This would produce: $myvar "Hello World!"; for ($i=0$i<10$i++) { echo $myvar."\n"; } Incorrect vB Code Usage: • [url] www.vbulletin.com [/url] - don't put spaces between the bracketed code and the text you are applying the code to. • [email][email protected][email] - the end brackets must include a forward slash ([/email]) Google   Web beyond.ca Terms of Use - Contact Us - Advertising Info - Archives - Car Blog Powered by: vBulletin Version 2.3.9 Copyright ©2009 Jelsoft Enterprises Limited. Bringing Car Enthusiasts together in discussion on our car forums Page Statistics : Page generated in 0.03853703 seconds (87.02% PHP - 12.98% MySQL) with 10 queries.
__label__pos
0.621687
Java Tutorial       Basic Programming Use of Comparison operator (==) for String Bitwise Operator (^) XOR (Exclusive OR) Change char in StringBuffer obj- reverse content Use of While Loop Use of Switch Statement Example of Multiple Levels of Inheritance Standard Libraray Class Methods Example of extracting subString from a String Use of Nested If-Else Statements Example of using a recursive method in class Declare, initialize, print an array of integers Bitwise Operator (&) AND StringBuffer,String obj,Basic Types to strbuf obj Switch Statement for different case labels Bitwise Operators Calling Base Class Method from Derived Class Example of arrays of String Example of searching String for character Use of If Statement Example of overloading constructors in a class Use of Break Statement Classes Boolean variable on Standard Output Device Show some properties of StringBuffer object Logical or Boolean Negation i.e. NOT (!) Example of Driving and using a Class Declare, initialize, print 2D array of integers Compare Strings by successive corresponding char Operator (>>) Shift Right - sign bit from left Initialization block - initialize data of class Use of the Continue Statement String Use of some Math Methods Increment and Decrement Operators Example of creating string obj from char Array Conditional OR Operator (||) in If Statement Example of using Non-Static Nested Class Search element in array using Linear Search Compare strings for equality Bitwise Operator (~) Complement Example creating String obj from StringBuffer obj Use of Do-While Loop Continue/Break Statement Example using Abstract Class and Abstract Method Print string on Standard Output Device Example of modifying string objects Conditional AND Operator (&&) in If Statement Example of using user-defined package Copy element of array in another in reverse order Example of concatination of Strings Bitwise Operator (|) OR Example extraction of char from StringBuffer obj Use of For Loop Conditional Statements - Logical Operators Example of Polymorphism Example of searching string for subString Use of If-Else Statement Example of using multiple classes Use of Labeled Break Statement Extending Classes and Inheritance Print on same line using multiple print statement Append StringBuffer by StringBuffer, String obj Conditional (Ternary) Operator - ? : Input - Output - Variables Example of Overriding a Base Class Method Declare, initialize and print String object Example of getting at characters in a String Operator (>>>) Shift Right - with zeros from left Constructor in class to initialize data member Use of Labeled Continue Statement StringBuffer Use of some Character Methods Read input from Standard Input Device Declare, initialize, print StringBuffer object Logical OR Operator (|) in If Statement Static Nested Class outside Top-Level Class Declare, initialize, print array of characters Check start and end of a String Operator (<<) Shift Left - with zero from right Example of creating class and using its object Use of Nested Loop Arrays Example of Copying i.e. Clonning Objects Integer variable on Standard Output Device Example of creating char Array from String obj Logical AND Operator (&) in If Statement Example of using a Static Nested Class Sort content of an array using Bubble Sort     public class JAVA_037 { public static void main(String[] args) { int[] array={10,-1,28,13,44,5,36,97,-18,11}; System.out.println(\" The contents of the Array in original order are :\"); for(int i=0;i<array.length;i++) System.out.println(\"\\t\\t\\t\\t\\t Array[\" + i + \"] = \" + array[i]); for(int j=0;j<(array.length-1);j++) { for(int k=0;k<(array.length-1);k++) { if(array[k]>array[(k+1)]) { int temp=array[k]; array[k]=array[(k+1)]; array[(k+1)]=temp; } } } System.out.println(\"\\n The contents of the Array after sorting are :\"); for(int l=0;l<array.length;l++) System.out.println(\"\\t\\t\\t\\t\\t Array[\" + l + \"] = \" + array[l]); } } Related Post: 1. Program to generate Fibonacci Series of a given number 2. Program which creates an Array of character. Make one function with one argument as a character and that function throw a user defined exception 3. Develop a game application in CORBA for which the player will guess a number between 1 to 100, which will be compared to the random number generated b 4. Program of linear search 5. An applet program to display sum of two number of textField 6. An applet program that display blinking rectangle 7. Game application in CORBA for which player will guess a number between 1 to 100, which will be compared to the random number generated by the class 8. Write a class whose objects holds a current value and have a method to add that value, printing the new value 9. Program of storing and retrieving integers using data streams on a single file 10. Program to show the use of Increment and Decrement Operators 11. Program to extract a portion of a character string and print the extracted string 12. Biological Sequence Allignment Using Dynamic Programing 13. An applet program to give demo of getDocumentBase() and getCodeBase() methods 14. CORBA program for displaying the date and time of the server machine, client machine and the difference between these two date and time 15. Program to print triangle of numbers 16. Develop a CORBA application that takes a string from client and returns whether it is palindrome 17. Program to show an example of Driving and using a Class 18. RMI based application, which reads a file with list of marks of student from a client, send it to server and find how many students having distinct 19. Even no or Prime no????? 20. Program to show an example of using Abstract Class and Abstract Method     Didn't find what you were looking for? Find more on Program to sort the contents of an array using Bubble Sort
__label__pos
0.999685
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.     107 lines 2.6 KiB <?php namespace p3k\XRay\Formats; use HTMLPurifier, HTMLPurifier_Config; use DOMDocument, DOMXPath; use p3k\XRay\Formats; class JSONFeed extends Format { public static function matches_host($url) { return true; } public static function matches($url) { return true; } public static function parse($feed, $url) { $result = [ 'data' => [ 'type' => 'unknown', ], 'url' => $url, 'source-format' => 'feed+json', ]; if($feed) { $result['data']['type'] = 'feed'; foreach($feed['items'] as $item) { $result['data']['items'][] = self::_hEntryFromFeedItem($item, $feed); } } return $result; } private static function _hEntryFromFeedItem($item, $feed) { $entry = [ 'type' => 'entry', 'author' => [ 'name' => null, 'url' => null, 'photo' => null ] ]; if(isset($item['author']['name'])) { $entry['author']['name'] = $item['author']['name']; } if(isset($item['author']['url'])) { $entry['author']['url'] = $item['author']['url']; } elseif(isset($feed['home_page_url'])) { $entry['author']['url'] = $feed['home_page_url']; } if(isset($item['author']['avatar'])) { $entry['author']['photo'] = $item['author']['avatar']; } if(isset($item['url'])) { $entry['url'] = $item['url']; } if(isset($item['id'])) { $entry['uid'] = $item['id']; } if(isset($item['title']) && trim($item['title'])) { $entry['name'] = trim($item['title']); } if(isset($item['content_html']) && isset($item['content_text'])) { $entry['content'] = [ 'html' => self::sanitizeHTML($item['content_html']), 'text' => trim($item['content_text']) ]; } elseif(isset($item['content_html'])) { $entry['content'] = [ 'html' => self::sanitizeHTML($item['content_html']), 'text' => self::stripHTML($item['content_html']) ]; } elseif(isset($item['content_text'])) { $entry['content'] = [ 'text' => trim($item['content_text']) ]; } if(isset($item['summary'])) { $entry['summary'] = $item['summary']; } if(isset($item['date_published'])) { $entry['published'] = $item['date_published']; } if(isset($item['date_modified'])) { $entry['updated'] = $item['date_modified']; } if(isset($item['image'])) { $entry['photo'] = $item['image']; } if(isset($item['tags'])) { $entry['category'] = $item['tags']; } $entry['post-type'] = \p3k\XRay\PostType::discover($entry); return $entry; } }
__label__pos
0.999996
Advanced Topic with Python Channel Access This chapter contains a variety of “usage notes” and implementation details that may help in getting the best performance from the pyepics module. The wait and timeout options for get(), ca.get_complete() The get functions, epics.caget(), pv.get() and epics.ca.get() all ask for data to be transferred over the network. For large data arrays or slow networks, this can can take a noticeable amount of time. For PVs that have been disconnected, the get call will fail to return a value at all. For this reason, these functions all take a timeout keyword option. The lowest level epics.ca.get() also has a wait option, and a companion function epics.ca.get_complete(). This section describes the details of these. If you’re using epics.caget() or pv.get() you can supply a timeout value. If the value returned is None, then either the PV has truly disconnected or the timeout passed before receiving the value. If the get is incomplete, in that the PV is connected but the data has simply not been received yet, a subsequent epics.caget() or pv.get() will eventually complete and receive the value. That is, if a PV for a large waveform record reports that it is connected, but a pv.get() returns None, simply trying again later will probably work: >>> p = epics.PV('LargeWaveform') >>> val = p.get() >>> val >>> time.sleep(10) >>> val = p.get() At the lowest level (which pv.get() and epics.caget() use), epics.ca.get() issues a get-request with an internal callback function. That is, it calls the CA library function libca.ca_array_get_callback() with a pre-defined callback function. With wait=True (the default), epics.ca.get() then waits up to the timeout or until the CA library calls the specified callback function. If the callback has been called, the value can then be converted and returned. If the callback is not called in time or if wait=False is used but the PV is connected, the callback will be called eventually, and simply waiting (or using epics.ca.pend_event() if epics.ca.PREEMPTIVE_CALLBACK is False) may be sufficient for the data to arrive. Under this condition, you can call epics.ca.get_complete(), which will NOT issue a new request for data to be sent, but wait (for up to a timeout time) for the previous get request to complete. epics.ca.get_complete() will return None if the timeout is exceeded or if there is not an “incomplete get” that it can wait to complete. Thus, you should use the return value from epics.ca.get_complete() with care. Note that pv.get() (and so epics.caget()) will normally rely on the PV value to be filled in automatically by monitor callbacks. If monitor callbacks are disabled (as is done for large arrays and can be turned off) or if the monitor hasn’t been called yet, pv.get() will check whether it should can epics.ca.get() or epics.ca.get_complete(). If not specified, the timeout for epics.ca.get_complete() (and all other get functions) will be set to: timeout = 0.5 + log10(count) Again, that’s the maximum time that will be waited, and if the data is received faster than that, the get will return as soon as it can. Strategies for connecting to a large number of PVs Occasionally, you may find that you need to quickly connect to a large number of PVs, say to write values to disk. The most straightforward way to do this, say: import epics pvnamelist = read_list_pvs() pv_vals = {} for name in pvnamelist: pv = epics.PV(name) pv_vals[name] = pv.get() or even just: values = [epics.caget(name) for name in pvnamelist] does incur some performance penalty. To minimize the penalty, we need to understand its cause. Creating a PV object (using any of pv.PV, or pv.get_pv(), or epics.caget()) will automatically use connection and event callbacks in an attempt to keep the PV alive and up-to-date during the seesion. Normally, this is an advantage, as you don’t need to explicitly deal with many aspects of Channel Access. But creating a PV does request some network traffic, and the PV will not be “fully connected” and ready to do a PV.get() until all the connection and event callbacks are established. In fact, PV.get() will not run until those connections are all established. This takes very close to 30 milliseconds for each PV. That is, for 1000 PVs, the above approach will take about 30 seconds. The simplest remedy is to allow all those connections to happen in parallel and in the background by first creating all the PVs and then getting their values. That would look like: # improve time to get multiple PVs: Method 1 import epics pvnamelist = read_list_pvs() pvs = [epics.PV(name) for name in pvnamelist] values = [p.get() for p in pvs] Though it doesn’t look that different, this improves performance by a factor of 100, so that getting 1000 PV values will take around 0.4 seconds. Can it be improved further? The answer is Yes, but at a price. For the discussion here, we’ll can the original version “Method 0” and the method of creating all the PVs then getting their values “Method 1”. With both of these approaches, the script has fully connected PV objects for all PVs named, so that subsequent use of these PVs will be very efficient. But this can be made even faster by turning off any connection or event callbacks, avoiding PV objects altogether, and using the epics.ca interface. This has been encapsulated into epics.caget_many() which can be used as: # get multiple PVs as fast as possible: Method 2 import epics pvnamelist = read_list_pvs() values = epics.caget_many(pvlist) In tests using 1000 PVs that were all really connected, Method 2 will take about 0.25 seconds, compared to 0.4 seconds for Method 1 and 30 seconds for Method 0. To understand what epics.caget_many() is doing, a more complete version of this looks like this: # epics.caget_many made explicit: Method 3 from epics import ca pvnamelist = read_list_pvs() pvdata = {} pvchids = [] # create, don't connect or create callbacks for name in pvnamelist: chid = ca.create_channel(name, connect=False, auto_cb=False) # note 1 pvchids.append(chid) # connect for chid in pvchids: ca.connect_channel(chid) # request get, but do not wait for result ca.poll() for chid in pvchids: ca.get(chid, wait=False) # note 2 # now wait for get() to complete ca.poll() for chid in pvchids: val = ca.get_complete(data[0]) pvdata[ca.name(chid)] = val The code here probably needs detailed explanation. As mentioned above, it uses the ca level, not PV objects. Second, the call to epics.ca.create_channel() (Note 1) uses connect=False and auto_cb=False which mean to not wait for a connection before returning, and to not automatically assign a connection callback. Normally, these are not what you want, as you want a connected channel and to be informed if the connection state changes, but we’re aiming for maximum speed here. We then use epics.ca.connect_channel() to connect all the channels. Next (Note 2), we tell the CA library to request the data for the channel without waiting around to receive it. The main point of not having epics.ca.get() wait for the data for each channel as we go is that each data transfer takes time. Instead we request data to be sent in a separate thread for all channels without waiting. Then we do wait by calling epics.ca.poll() once and only once, (not len(pvnamelist) times!). Finally, we use the epics.ca.get_complete() method to convert the data that has now been received by the companion thread to a python value. Method 2 and 3 have essentially the same runtime, which is somewhat faster than Method 1, and much faster than Method 0. Which method you should use depends on use case. In fact, the test shown here only gets the PV values once. If you’re writing a script to get 1000 PVs, write them to disk, and exit, then Method 2 (epics.caget_many()) may be exactly what you want. But if your script will get 1000 PVs and stay alive doing other work, or even if it runs a loop to get 1000 PVs and write them to disk once a minute, then Method 1 will actually be faster. That is doing epics.caget_many() in a loop, as with: # caget_many() 10 times import epics import time pvnamelist = read_list_pvs() for i in range(10): values = epics.caget_many(pvlist) time.sleep(0.01) will take around considerably longer than creating the PVs once and getting their values in a loop with: # pv.get() 10 times import epics import time pvnamelist = read_list_pvs() pvs = [epics.PV(name) for name in pvnamelist] for i in range(10): values = [p.get() for p in pvs] time.sleep(0.01) In tests with 1000 PVs, looping with epics.caget_many() took about 1.5 seconds, while the version looping over PV.get() took about 0.5 seconds. To be clear, it is connecting to Epics PVs that is expensive, not the retreiving of data from connected PVs. You can lower the connection expense by not retaining the connection or creating monitors on the PVs, but if you are going to re-use the PVs, that savings will be lost quickly. In short, use Method 1 over epics.caget_many() unless you’ve benchmarked your use-case and have demonstrated that epics.caget_many() is better for your needs. time.sleep() or epics.poll()? In order for a program to communicate with Epics devices, it needs to allow some time for this communication to happen. With epics.ca.PREEMPTIVE_CALLBACK set to True, this communication will be handled in a thread separate from the main Python thread. This means that CA events can happen at any time, and epics.ca.pend_event() does not need to be called to explicitly allow for event processing. Still, some time must be released from the main Python thread on occasion in order for events to be processed. The simplest way to do this is with time.sleep(), so that an event loop can simply be: >>> while True: >>> time.sleep(0.001) Unfortunately, the time.sleep() method is not a very high-resolution clock, with typical resolutions of 1 to 10 ms, depending on the system. Thus, even though events will be asynchronously generated and epics with pre-emptive callbacks does not require epics.ca.pend_event() or epics.ca.poll() to be run, better performance may be achieved with an event loop of: >>> while True: >>> epics.poll(evt=1.e-5, iot=0.1) as the loop will be run more often than using time.sleep(). Using Python Threads An important feature of the PyEpics package is that it can be used with Python threads, as Epics 3.14 supports threads for client code. Even in the best of cases, working with threads can be somewhat tricky and lead to unexpected behavior, and the Channel Access library adds a small level of complication for using CA with Python threads. The result is that some precautions may be in order when using PyEpics and threads. This section discusses the strategies for using threads with PyEpics. First, to use threads with Channel Access, you must have epics.ca.PREEMPTIVE_CALLBACK = True. This is the default value, but if epics.ca.PREEMPTIVE_CALLBACK has been set to False, threading will not work. Second, if you are using PV objects and not making heavy use of the epics.ca module (that is, not making and passing around chids), then the complications below are mostly hidden from you. If you’re writing threaded code, it’s probably a good idea to read this just to understand what the issues are. Channel Access Contexts The Channel Access library uses a concept of contexts for its own thread model, with contexts holding sets of threads as well as Channels and Process Variables. For non-threaded work, a process will use a single context that is initialized prior doing any real CA work (done in epics.ca.initialize_libca()). In a threaded application, each new thread begins with a new, uninitialized context that must be initialized or replaced. Thus each new python thread that will interact with CA must either explicitly create its own context with epics.ca.create_context() (and then, being a good citizen, destroy this context as the thread ends with epics.ca.destroy_context()) or attach to an existing context. The generally recommended approach is to use a single CA context throughout an entire process and have each thread attach to the first context created (probably from the main thread). This avoids many potential pitfalls (and crashes), and can be done fairly simply. It is the default mode when using PV objects. The most explicit use of contexts is to put epics.ca.create_context() at the start of each function call as a thread target, and epics.ca.destroy_context() at the end of each thread. This will cause all the activity in that thread to be done in its own context. This works, but means more care is needed, and so is not the recommended. The best way to attach to the initially created context is to call epics.ca.use_initial_context() before any other CA calls in each function that will be called by Thread.run(). Equivalently, you can add a withInitialContext() decorator to the function. Creating a PV object will implicitly do this for you, as long as it is your first CA action in the function. Each time you do a PV.get() or PV.put() (or a few other methods), it will also check that the initial context is being used. Of course, this approach requires CA to be initialized already. Doing that in the main thread is highly recommended. If it happens in a child thread, that thread must exist for all CA work, so either the life of the process or with great care for processes that do only some CA calls. If you are writing a threaded application in which the first real CA calls are inside a child thread, it is recommended that you initialize CA in the main thread, As a convenience, the CAThread in the epics.ca module is is a very thin wrapper around the standard threading.Thread which adding a call of epics.ca.use_initial_context() just before your threaded function is run. This allows your target functions to not explicitly set the context, but still ensures that the initial context is used in all functions. How to work with CA and Threads Summarizing the discussion above, to use threads you must use run in PREEMPTIVE_CALLBACK mode. Furthermore, it is recommended that you use a single context, and that you initialize CA in the main program thread so that your single CA context belongs to the main thread. Using PV objects exclusively makes this easy, but it can also be accomplished relatively easily using the lower-level ca interface. The options for using threads (in approximate order of reliability) are then: 1. use PV objects for threading work. This ensures you’re working in a single CA context. 2. use CAThread instead of Thread for threads that will use CA calls. 3. put epics.ca.use_initial_context() at the top of all functions that might be a Thread target function, or decorate them with withInitialContext() decorator, @withInitialContext. 4. use epics.ca.create_context() at the top of all functions that are inside a new thread, and be sure to put epics.ca.destroy_context() at the end of the function. 5. ignore this advise and hope for the best. If you’re not creating new PVs and only reading values of PVs created in the main thread inside a child thread, you may not see a problems, at least not until you try to do something fancier. Thread Examples This is a simplified version of test code using Python threads. It is based on code originally from Friedrich Schotte, NIH, and included as thread_test.py in the tests directory of the source distribution. In this example, we define a run_test procedure which will create PVs from a supplied list, and monitor these PVs, printing out the values when they change. Two threads are created and run concurrently, with overlapping PV lists, though one thread is run for a shorter time than the other. """This script tests using EPICS CA and Python threads together Based on code from Friedrich Schotte, NIH, modified by Matt Newville 19-Apr-2010 """ import time from sys import stdout from threading import Thread import epics from epics.ca import CAThread from pvnames import updating_pvlist pvlist_a = updating_pvlist[:-1] pvlist_b = updating_pvlist[1:] def run_test(runtime=1, pvnames=None, run_name='thread c'): msg = '-> thread "%s" will run for %.3f sec, monitoring %s\n' stdout.write(msg % (run_name, runtime, pvnames)) def onChanges(pvname=None, value=None, char_value=None, **kw): stdout.write(' %s = %s (%s)\n' % (pvname, char_value, run_name)) stdout.flush() # epics.ca.use_initial_context() # epics.ca.create_context() start_time = time.time() pvs = [epics.PV(pvn, callback=onChanges) for pvn in pvnames] while time.time()-start_time < runtime: time.sleep(0.1) [p.clear_callbacks() for p in pvs] stdout.write( 'Completed Thread %s\n' % ( run_name)) stdout.write( "First, create a PV in the main thread:\n") p = epics.PV(updating_pvlist[0]) stdout.write("Run 2 Background Threads simultaneously:\n") th1 = CAThread(target=run_test,args=(3, pvlist_a, 'A')) th1.start() th2 = CAThread(target=run_test,args=(6, pvlist_b, 'B')) th2.start() th2.join() th1.join() stdout.write('Done\n') In light of the long discussion above, a few remarks are in order: This code uses the standard Thread library and explicitly calls epics.ca.use_initial_context() prior to any CA calls in the target function. Also note that the run_test() function is first called from the main thread, so that the initial CA context does belong to the main thread. Finally, the epics.ca.use_initial_context() call in run_test() above could be replaced with epics.ca.create_context(), and run OK. The output from this will look like: First, create a PV in the main thread: Run 2 Background Threads simultaneously: -> thread "A" will run for 3.000 sec, monitoring ['Py:ao1', 'Py:ai1', 'Py:long1'] -> thread "B" will run for 6.000 sec, monitoring ['Py:ai1', 'Py:long1', 'Py:ao2'] Py:ao1 = 8.3948 (A) Py:ai1 = 3.14 (B) Py:ai1 = 3.14 (A) Py:ao1 = 0.7404 (A) Py:ai1 = 4.07 (B) Py:ai1 = 4.07 (A) Py:long1 = 3 (B) Py:long1 = 3 (A) Py:ao1 = 13.0861 (A) Py:ai1 = 8.49 (B) Py:ai1 = 8.49 (A) Py:ao2 = 30 (B) Completed Thread A Py:ai1 = 9.42 (B) Py:ao2 = 30 (B) Py:long1 = 4 (B) Py:ai1 = 3.35 (B) Py:ao2 = 31 (B) Py:ai1 = 4.27 (B) Py:ao2 = 31 (B) Py:long1 = 5 (B) Py:ai1 = 8.20 (B) Py:ao2 = 31 (B) Completed Thread B Done Note that while both threads A and B are running, a callback for the PV Py:ai1 is generated in each thread. Note also that the callbacks for the PVs created in each thread are explicitly cleared with: [p.clear_callbacks() for p in pvs] Without this, the callbacks for thread A will persist even after the thread has completed! Using Multiprocessing with PyEpics An alternative to Python threads that has some very interesting and important features is to use multiple processes, as with the standard Python multiprocessing module. While using multiple processes has some advantages over threads, it also has important implications for use with PyEpics. The basic issue is that multiple processes need to be fully separate, and do not share global state. For epics Channel Access, this means that all those things like established communication channels, callbacks, and Channel Access context cannot easily be share between processes. The solution is to use a CAProcess, which acts just like multiprocessing.Process, but knows how to separate contexts between processes. This means that you will have to create PV objects for each process (even if they point to the same PV). class CAProcess(group=None, target=None, name=None, args=(), kwargs={}) a subclass of multiprocessing.Process that clears the global Channel Access context before running you target function in its own process. class CAPool(processes=None, initializer=None, initargs=(), maxtasksperchild=None) a subclass of multiprocessing.pool.Pool, creating a Pool of CAProcess instances. A simple example of using multiprocessing successfully is given: from __future__ import print_function import epics import time import multiprocessing as mp import threading import pvnames PVN1 = pvnames.double_pv # 'Py:ao2' PVN2 = pvnames.double_pv2 # 'Py:ao3' def subprocess(*args): print('==subprocess==', args) mypvs = [epics.get_pv(pvname) for pvname in args] for i in range(10): time.sleep(0.750) out = [(p.pvname, p.get(as_string=True)) for p in mypvs] out = ', '.join(["%s=%s" % o for o in out]) print('==sub (%d): %s' % (i, out)) def main_process(): def monitor(pvname=None, char_value=None, **kwargs): print('--main:monitor %s=%s' % (pvname, char_value)) print('--main:') pv1 = epics.get_pv(PVN1) print('--main:init %s=%s' % (PVN1, pv1.get())) pv1.add_callback(callback=monitor) try: proc1 = epics.CAProcess(target=subprocess, args=(PVN1, PVN2)) proc1.start() proc1.join() except KeyboardInterrupt: print('--main: killing subprocess') proc1.terminate() print('--main: subprocess complete') time.sleep(0.9) print('--main:final %s=%s' % (PVN1, pv1.get())) if __name__ == '__main__': main_process() here, the main process and the subprocess can each interact with the same PV, though they need to create a separate connection (here, using PV) in each process. Note that different CAProcess instances can communicate via standard multiprocessing.Queue. At this writing, no testing has been done on using multiprocessing Managers.
__label__pos
0.924038
Answers Solutions by everydaycalculation.com Answers.everydaycalculation.com » Subtract fractions Subtract 15/30 from 56/5 1st number: 11 1/5, 2nd number: 15/30 56/5 - 15/30 is 107/10. Steps for subtracting fractions 1. Find the least common denominator or LCM of the two denominators: LCM of 5 and 30 is 30 Next, find the equivalent fraction of both fractional numbers with denominator 30 2. For the 1st fraction, since 5 × 6 = 30, 56/5 = 56 × 6/5 × 6 = 336/30 3. Likewise, for the 2nd fraction, since 30 × 1 = 30, 15/30 = 15 × 1/30 × 1 = 15/30 4. Subtract the two like fractions: 336/30 - 15/30 = 336 - 15/30 = 321/30 5. After reducing the fraction, the answer is 107/10 6. In mixed form: 107/10 MathStep (Works offline) Download our mobile app and learn to work with fractions in your own time: Android and iPhone/ iPad Related: © everydaycalculation.com
__label__pos
0.99569
MANA Decentraland and BCBA Bolsa de Comercio de Buenos Aires Understanding the Connection Decentraland s virtual reality platform, powered by MANA cryptocurrency, has made significant strides in recent years. Its innovative approach to creating a decentralized metaverse has captured the attention of investors and technology enthusiasts alike. However, one may wonder what connection does MANA have with the BCBA Bolsa de Comercio de Buenos Aires, Argentina s largest stock exchange? The connection lies in the fact that Decentraland s MANA token is listed on the BCBA stock exchange, giving investors the opportunity to trade and invest in this cryptocurrency. This listing signifies the growing recognition and acceptance of digital currencies in traditional financial markets. It also highlights the potential for blockchain technology to transform various industries, including finance and gaming. Introduction: Decentraland (MANA) is a virtual reality platform built on the Ethereum blockchain that allows users to create, experience, and monetize virtual reality content. It is a decentralized platform where users can own virtual land, build and create 3D content, and interact with other users. On the other hand, the BCBA (Bolsa de Comercio de Buenos Aires) is the largest stock exchange in Argentina and one of the oldest in Latin America. Understanding the connection between MANA and BCBA is important as it sheds light on the potential impact of blockchain technology on the traditional financial industry. The partnership between Decentraland and BCBA demonstrates how blockchain technology can be leveraged to transform the way assets are traded and owned. Decentraland is a virtual reality platform built on the Ethereum blockchain. Decentraland is a unique and innovative virtual reality platform that is built on the Ethereum blockchain. It aims to provide users with a decentralized and immersive virtual world experience. By utilizing blockchain technology, Decentraland ensures transparency, security, and ownership of virtual assets. With Decentraland, users have the freedom to create, explore, and monetize virtual worlds. They can purchase virtual land, known as LAND, and use it to build various assets such as buildings, structures, and landscapes. These virtual assets can then be sold, rented, or used for various purposes within the Decentraland ecosystem. One of the key features of Decentraland is its native cryptocurrency called MANA. MANA, short for Decentraland, is an ERC-20 token that is used for various transactions within the Decentraland platform. It serves as the primary currency for purchasing virtual land, virtual assets, and other goods and services within the virtual world. The supply of MANA is limited to a maximum of 2.64 billion tokens, which adds to its scarcity and value. Users can acquire MANA through various means, including buying it on cryptocurrency exchanges or earning it through participating in the Decentraland ecosystem. MANA has several use cases within the Decentraland platform. It can be used to purchase virtual land, invest in virtual assets, participate in virtual events, and trade with other users. It also enables creators to monetize their content by selling virtual goods or services to other users. Furthermore, MANA plays a crucial role in providing governance and decision-making powers within the Decentraland ecosystem. Token holders can participate in voting on proposals and shaping the future development of the platform. Overall, Decentraland and its native cryptocurrency MANA offer a unique and exciting virtual reality experience. By leveraging the power of blockchain technology, Decentraland aims to revolutionize the way we interact and engage with virtual worlds. 1. What is MANA Decentraland? MANA Decentraland is the native cryptocurrency of the Decentraland virtual reality platform. Decentraland is a decentralized, blockchain-based virtual world where users can explore, create, and trade virtual assets. The platform is built on the Ethereum blockchain, allowing users to buy and sell virtual land, known as LAND, and other virtual items using the MANA token. MANA can be used for a variety of purposes within the Decentraland platform. Users can use MANA to buy and sell virtual land and other assets, participate in Decentraland s virtual economy, and contribute to the development and governance of the platform. It also serves as the in-game currency, allowing users to purchase virtual goods and services from other players. MANA is an ERC-20 token that serves as the primary currency within the Decentraland platform. Decentraland is a virtual reality platform built on the Ethereum blockchain that allows users to create, explore, and trade virtual assets. The platform is divided into parcels of virtual land that users can purchase and develop using the platform s native cryptocurrency, MANA. MANA, short for Decentraland token, is an ERC-20 token that operates on the Ethereum blockchain. It is used as a medium of exchange within the Decentraland ecosystem, enabling users to buy, sell, and trade virtual goods, such as land, wearables, and art. The value of MANA is determined by market demand and supply, with users able to convert MANA to other cryptocurrencies or fiat currencies on various cryptocurrency exchanges. As the primary currency within Decentraland, MANA plays a crucial role in the platform s economy. Users can earn MANA through various activities within the virtual world, such as participating in events, completing quests, or selling virtual goods. They can then use MANA to purchase additional land, art, wearables, or even trade it for other digital assets. The connection between MANA and BCBA (Bolsa de Comercio de Buenos Aires) lies in the partnership between the two entities. BCBA, the main stock exchange in Argentina, joined forces with Decentraland to offer its users the opportunity to trade virtual assets using traditional financial instruments. This partnership aims to bridge the gap between the traditional financial world and the emerging virtual reality market, allowing BCBA users to invest in and trade virtual land and goods using their existing trading accounts. Through this connection, Decentraland and BCBA hope to attract additional users to the Decentraland platform and provide a more diverse and liquid market for virtual assets. This collaboration also demonstrates the growing recognition of blockchain technology and virtual reality as a legitimate and valuable sector within the global economy. It is used by users to purchase virtual land, known as LAND, and various in-world assets like avatars, wearables, and virtual goods. MANA, the native cryptocurrency of the Decentraland virtual reality platform, is an integral part of the ecosystem. It serves as the medium of exchange, allowing users to buy and sell virtual assets within the Decentraland metaverse. MANA is an ERC-20 token built on the Ethereum blockchain, ensuring security, transparency, and immutability. Decentraland offers a unique virtual world where users can explore, interact, and build their own experiences. The virtual land, or LAND, is the foundation of this virtual reality universe. It is purchased and owned by users using MANA, giving them full control and ownership over their virtual properties. LAND can be used to create and monetize various experiences, such as games, art galleries, virtual stores, and much more. Additionally, MANA is used to acquire various in-world assets like avatars, wearables, and virtual goods. Avatars are digital representations of users in the virtual world, allowing them to customize and express their unique identity. Wearables are virtual accessories and clothing items that users can equip their avatars with, enhancing their appearance and personal style. Virtual goods, on the other hand, are a wide range of digital assets that can be bought and sold within Decentraland. These include virtual art, collectibles, landmarks, and even virtual real estate. With MANA, users can participate in the thriving virtual economy of Decentraland, buying and selling these assets to enhance their virtual experiences. Win the Game: Get Expert Answers to Questions about Cryptocurrency Mining What is Decentraland? Decentraland is a virtual reality platform powered by blockchain technology, where users can create and monetize their own virtual experiences. What is BCBA? BCBA stands for Bolsa de Comercio de Buenos Aires, which is the stock exchange of Buenos Aires, Argentina. How are MANA and BCBA connected? MANA, the virtual currency of Decentraland, has partnered with BCBA to create a virtual representation of the stock exchange, allowing users to explore and interact with the trading floor in a virtual reality environment. What benefits does the partnership between MANA and BCBA bring? The partnership between MANA and BCBA brings several benefits. It brings more visibility to both platforms and enhances the user experience for traders and investors. It also opens up new opportunities for virtual reality experiences in the financial sector and showcases the potential of blockchain technology in traditional industries.
__label__pos
0.934096
Definition of arc tangent in English: arc tangent (also arctan) noun • A mathematical function that is the inverse of the tangent function. • ‘In the 14th century, Madhava, isolated in South India, developed a power series for the arc tangent function, apparently without the use of calculus, allowing the calculation of pi to any number of decimal places.’ • ‘To see how this description of the series fits with Gregory's series for arctan see the biography of Madhava.’ • ‘When we find an arctan of a reciprocal of an even-indexed Fibonacci number, we can use to replace it by a sum of two terms, one an odd-indexed Fibonacci number and another even-indexed Fibonacci number.’ • ‘We can put this in words as ‘The final arctan is just 1 more than the product of the other two (whose denominators differ by one).’’ • ‘Using a bit more trigonometry, we can determine the angle between two subsequent samples by multiplying one by the complex conjugate of the other and then taking the arc tangent of the product.’
__label__pos
0.979538
2 yellow name yellow_name 于 2017.09.08 22:14 提问 sqlalchemy 同步数据库问题python from sqlalchemy import create_engine, MetaData from sqlalchemy.orm import sessionmaker from sqlalchemy.ext.declarative import declarative_base engine = create_engine('host':'mysql+pymysql://root:123456@localhost:3306/spider') Base = declarative_base() class BaseModel(Base): __abstract__ = True __metadata__ = MetaData(bind=engine) __table_arhs__ = { 'mysql_engine': 'InnoDb', 'mysql_charset': 'utf8mb4' } session = _Session(autocommit=False) if __name__ == '__main__': BaseModel.__metadata__.create_all() 用 BaseModel.__metadata__.create_all() 不能同步数据库表 其中一个表类如下: from sqlalchemy import Column from sqlalchemy.dialects.mysql import INTEGER, VARCHAR from common.db import BaseModel class CityModel(BaseModel): __tablename__ = 'city' id = Column(INTEGER, primary_key=True) name = Column(VARCHAR(64)) 这样怎么同步mysql 自动生成表? 大佬们 1个回答 caozhy caozhy   Ds   Rxr 2017.09.09 23:53 Csdn user default icon 上传中... 上传图片 插入图片 准确详细的回答,更有利于被提问者采纳,从而获得C币。复制、灌水、广告等回答会被删除,是时候展现真正的技术了!
__label__pos
0.574102