url
stringlengths
14
1.76k
text
stringlengths
100
1.02M
metadata
stringlengths
1.06k
1.1k
http://accountingexplained.com/financial/revenue/percentage-of-completion-method
# Revenue Recognition: Percentage of Completion Method Percentage of completion method is a basis for revenue recognition in long-term construction contracts which span over more than one accounting periods. In case of long-term contracts, accountants need a basis to apportion the total contract revenue between the multiple accounting periods. Percentage of completion method provides one of those bases, other being full-contract method. It determines revenue recognized in one accounting period in accordance with the following formula: Revenue Recognized = Percentage of Work Completed in the Period Total Contract Value Total contract value is the total revenue from the long-term contract. Percentage of work completed is the proportion of work completed in a period to total work for the contract. It is usually estimated using the following formula: Percentage of Work Completed = Expenditures Incurred from Inception to Date Total Estimated Costs for the Contract Expenditures incurred from inception to date represent costs incurred from the start of the project to the date of estimation. Total estimated expenditures for the contract represent the total budgeted cost for the project. It includes costs that have been incurred to date and costs that are expected to be incurred in future periods. There is another method for estimating percentage of completion called survey method which is based on the physical progress of the contract. Under this method engineers and other experts observe the activities and determine their judgment of the percentage of work completed. ## Example Construction Inc. is engaged in constructing a massive bridge in Wonderland. The contract is worth $200 million and the company is expected to complete it in 3 years. In Year 1 the company has incurred an amount of$50 million on the contract and the engineers estimate that in the next 2 years the company is expected to expend $110 million more. Based on the physical progress of the project the engineers also estimate that 40% of the work has been carried out. Solution: Under the survey method the engineers have provided their judgment of the percentage of work completed and it is 40%. Based on costs incurred to date and total costs the percentage of completion comes out to be: Percentage of work completed =$50 million ÷ ($50 million +$110 million) = 31.25%. Total costs include costs incurred to date and costs expected to be incurred over the remaining period. Based on the percentage of completion calculated using cost date we determine than revenue of $62.5 million has been earned (31.25% multiplied by$200 million total contract value). On the other hand based on the engineer's survey the revenue recognized should be $80 million (40% multiplied by$200 million). Written by Obaidullah Jan, ACA, CFA <--- Hire me on Upwork
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3207091689109802, "perplexity": 2707.948308895792}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-22/segments/1464049288709.26/warc/CC-MAIN-20160524002128-00063-ip-10-185-217-139.ec2.internal.warc.gz"}
https://worldwidescience.org/topicpages/h/high+quality+subset.html
#### Sample records for high quality subset 1. Criteria to Extract High-Quality Protein Data Bank Subsets for Structure Users. Science.gov (United States) Carugo, Oliviero; Djinović-Carugo, Kristina 2016-01-01 It is often necessary to build subsets of the Protein Data Bank to extract structural trends and average values. For this purpose it is mandatory that the subsets are non-redundant and of high quality. The first problem can be solved relatively easily at the sequence level or at the structural level. The second, on the contrary, needs special attention. It is not sufficient, in fact, to consider the crystallographic resolution and other feature must be taken into account: the absence of strings of residues from the electron density maps and from the files deposited in the Protein Data Bank; the B-factor values; the appropriate validation of the structural models; the quality of the electron density maps, which is not uniform; and the temperature of the diffraction experiments. More stringent criteria produce smaller subsets, which can be enlarged with more tolerant selection criteria. The incessant growth of the Protein Data Bank and especially of the number of high-resolution structures is allowing the use of more stringent selection criteria, with a consequent improvement of the quality of the subsets of the Protein Data Bank. 2. Bayesian Subset Modeling for High-Dimensional Generalized Linear Models KAUST Repository Liang, Faming 2013-06-01 This article presents a new prior setting for high-dimensional generalized linear models, which leads to a Bayesian subset regression (BSR) with the maximum a posteriori model approximately equivalent to the minimum extended Bayesian information criterion model. The consistency of the resulting posterior is established under mild conditions. Further, a variable screening procedure is proposed based on the marginal inclusion probability, which shares the same properties of sure screening and consistency with the existing sure independence screening (SIS) and iterative sure independence screening (ISIS) procedures. However, since the proposed procedure makes use of joint information from all predictors, it generally outperforms SIS and ISIS in real applications. This article also makes extensive comparisons of BSR with the popular penalized likelihood methods, including Lasso, elastic net, SIS, and ISIS. The numerical results indicate that BSR can generally outperform the penalized likelihood methods. The models selected by BSR tend to be sparser and, more importantly, of higher prediction ability. In addition, the performance of the penalized likelihood methods tends to deteriorate as the number of predictors increases, while this is not significant for BSR. Supplementary materials for this article are available online. © 2013 American Statistical Association. 3. A Feature Subset Selection Method Based On High-Dimensional Mutual Information Directory of Open Access Journals (Sweden) Chee Keong Kwoh 2011-04-01 Full Text Available Feature selection is an important step in building accurate classifiers and provides better understanding of the data sets. In this paper, we propose a feature subset selection method based on high-dimensional mutual information. We also propose to use the entropy of the class attribute as a criterion to determine the appropriate subset of features when building classifiers. We prove that if the mutual information between a feature set X and the class attribute Y equals to the entropy of Y , then X is a Markov Blanket of Y . We show that in some cases, it is infeasible to approximate the high-dimensional mutual information with algebraic combinations of pairwise mutual information in any forms. In addition, the exhaustive searches of all combinations of features are prerequisite for finding the optimal feature subsets for classifying these kinds of data sets. We show that our approach outperforms existing filter feature subset selection methods for most of the 24 selected benchmark data sets. 4. High Aldehyde Dehydrogenase Activity Identifies a Subset of Human Mesenchymal Stromal Cells with Vascular Regenerative Potential. Science.gov (United States) Sherman, Stephen E; Kuljanin, Miljan; Cooper, Tyler T; Putman, David M; Lajoie, Gilles A; Hess, David A 2017-06-01 During culture expansion, multipotent mesenchymal stromal cells (MSCs) differentially express aldehyde dehydrogenase (ALDH), an intracellular detoxification enzyme that protects long-lived cells against oxidative stress. Thus, MSC selection based on ALDH-activity may be used to reduce heterogeneity and distinguish MSC subsets with improved regenerative potency. After expansion of human bone marrow-derived MSCs, cell progeny was purified based on low versus high ALDH-activity (ALDH hi ) by fluorescence-activated cell sorting, and each subset was compared for multipotent stromal and provascular regenerative functions. Both ALDH l ° and ALDH hi MSC subsets demonstrated similar expression of stromal cell (>95% CD73 + , CD90 + , CD105 + ) and pericyte (>95% CD146 + ) surface markers and showed multipotent differentiation into bone, cartilage, and adipose cells in vitro. Conditioned media (CDM) generated by ALDH hi MSCs demonstrated a potent proliferative and prosurvival effect on human microvascular endothelial cells (HMVECs) under serum-free conditions and augmented HMVEC tube-forming capacity in growth factor-reduced matrices. After subcutaneous transplantation within directed in vivo angiogenesis assay implants into immunodeficient mice, ALDH hi MSC or CDM produced by ALDH hi MSC significantly augmented murine vascular cell recruitment and perfused vessel infiltration compared with ALDH l ° MSC. Although both subsets demonstrated strikingly similar mRNA expression patterns, quantitative proteomic analyses performed on subset-specific CDM revealed the ALDH hi MSC subset uniquely secreted multiple proangiogenic cytokines (vascular endothelial growth factor beta, platelet derived growth factor alpha, and angiogenin) and actively produced multiple factors with chemoattractant (transforming growth factor-β, C-X-C motif chemokine ligand 1, 2, and 3 (GRO), C-C motif chemokine ligand 5 (RANTES), monocyte chemotactic protein 1 (MCP-1), interleukin [IL]-6, IL-8) and matrix 5. Long term impact of high titer Edmonston-Zagreb measles vaccine on T lymphocyte subsets DEFF Research Database (Denmark) Lisse, I M; Aaby, P; Knudsen, K 1994-01-01 Several trials of high titer measles vaccine (> 10(4.7) plaque-forming unit) have found female recipients of Edmonston-Zagreb (EZ) vaccine to have lower survival than female recipients of standard measles vaccine. Two trials with medium and high titer EZ vaccine from the age of 4 months were...... unlikely to explain the reduced survival which has been associated with high titer EZ measles vaccination. In the 2 years after the investigation of T cell subsets, there was no increased mortality for recipients of EZ vaccine. Hence it is unlikely that high titer vaccine has an persistent adverse effect... 6. Tired telomeres: Poor global sleep quality, perceived stress, and telomere length in immune cell subsets in obese men and women. Science.gov (United States) Prather, Aric A; Gurfein, Blake; Moran, Patricia; Daubenmier, Jennifer; Acree, Michael; Bacchetti, Peter; Sinclair, Elizabeth; Lin, Jue; Blackburn, Elizabeth; Hecht, Frederick M; Epel, Elissa S 2015-07-01 Poor sleep quality and short sleep duration are associated with increased incidence and progression of a number of chronic health conditions observed at greater frequency among the obese and those experiencing high levels of stress. Accelerated cellular aging, as indexed by telomere attrition in immune cells, is a plausible pathway linking sleep and disease risk. Prior studies linking sleep and telomere length are mixed. One factor may be reliance on leukocytes, which are composed of varied immune cell types, as the sole measure of telomere length. To better clarify these associations, we investigated the relationships of global sleep quality, measured by the Pittsburgh Sleep Quality Index (PSQI), and diary-reported sleep duration with telomere length in different immune cell subsets, including granulocytes, peripheral blood mononuclear cells (PBMCs), CD8+ and CD4+ T lymphocytes, and B lymphocytes in a sample of 87 obese men and women (BMI mean=35.4, SD=3.6; 81.6% women; 62.8% Caucasian). Multiple linear regression analyses were performed adjusting for age, gender, race, education, BMI, sleep apnea risk, and perceived stress. Poorer PSQI global sleep quality was associated with statistically significantly shorter telomere length in lymphocytes but not granulocytes and in particular CD8+ T cells (b=-56.8 base pairs per one point increase in PSQI, SE=20.4, p=0.007) and CD4+ T cells (b=-37.2, SE=15.9, p=0.022). Among separate aspects of global sleep quality, low perceived sleep quality and decrements in daytime function were most related to shorter telomeres. In addition, perceived stress moderated the sleep-CD8+ telomere association. Poorer global sleep quality predicted shorter telomere length in CD8+ T cells among those with high perceived stress but not in low stress participants. These findings provide preliminary evidence that poorer global sleep quality is related to telomere length in several immune cell types, which may serve as a pathway linking sleep and 7. A Hybrid Feature Subset Selection Algorithm for Analysis of High Correlation Proteomic Data Science.gov (United States) 2012-01-01 Pathological changes within an organ can be reflected as proteomic patterns in biological fluids such as plasma, serum, and urine. The surface-enhanced laser desorption and ionization time-of-flight mass spectrometry (SELDI-TOF MS) has been used to generate proteomic profiles from biological fluids. Mass spectrometry yields redundant noisy data that the most data points are irrelevant features for differentiating between cancer and normal cases. In this paper, we have proposed a hybrid feature subset selection algorithm based on maximum-discrimination and minimum-correlation coupled with peak scoring criteria. Our algorithm has been applied to two independent SELDI-TOF MS datasets of ovarian cancer obtained from the NCI-FDA clinical proteomics databank. The proposed algorithm has used to extract a set of proteins as potential biomarkers in each dataset. We applied the linear discriminate analysis to identify the important biomarkers. The selected biomarkers have been able to successfully diagnose the ovarian cancer patients from the noncancer control group with an accuracy of 100%, a sensitivity of 100%, and a specificity of 100% in the two datasets. The hybrid algorithm has the advantage that increases reproducibility of selected biomarkers and able to find a small set of proteins with high discrimination power. PMID:23717808 8. The CD44(high tumorigenic subsets in lung cancer biospecimens are enriched for low miR-34a expression. Directory of Open Access Journals (Sweden) Saroj K Basak Full Text Available Cellular heterogeneity is an integral part of cancer development and progression. Progression can be associated with emergence of cells that exhibit high phenotypic plasticity (including "de-differentiation" to primitive developmental states, and aggressive behavioral properties (including high tumorigenic potentials. We observed that many biomarkers that are used to identify Cancer Stem Cells (CSC can label cell subsets in an advanced clinical stage of lung cancer (malignant pleural effusions, or MPE. Thus, CSC-biomarkers may be useful for live sorting functionally distinct cell subsets from individual tumors, which may enable investigators to hone in on the molecular basis for functional heterogeneity. We demonstrate that the CD44(hi (CD44-high cancer cell subsets display higher clonal, colony forming potential than CD44(lo cells (n=3 and are also tumorigenic (n=2/2 when transplanted in mouse xenograft model. The CD44(hi subsets express different levels of embryonal (de-differentiation markers or chromatin regulators. In archived lung cancer tissues, ALDH markers co-localize more with CD44 in squamous cell carcinoma (n=5/7 than Adeno Carcinoma (n=1/12. MPE cancer cells and a lung cancer cell line (NCI-H-2122 exhibit chromosomal abnormalities and 1p36 deletion (n=3/3. Since miR-34a maps to the 1p36 deletion site, low miR-34a expression levels were detected in these cells. The colony forming efficiency of CD44(hi cells, characteristic property of CSC, can be inhibited by mir-34a replacement in these samples. In addition the highly tumorigenic CD44(hi cells are enriched for cells in the G2 phase of cell cycle. 9. Long term impact of high titer Edmonston-Zagreb measles vaccine on T lymphocyte subsets DEFF Research Database (Denmark) Lisse, I M; Aaby, P; Knudsen, K 1994-01-01 Several trials of high titer measles vaccine (> 10(4.7) plaque-forming unit) have found female recipients of Edmonston-Zagreb (EZ) vaccine to have lower survival than female recipients of standard measles vaccine. Two trials with medium and high titer EZ vaccine from the age of 4 months were... 10. High incidence of interchromosomal transpositions in the evolutionary history of a subset of or genes in Drosophila. Science.gov (United States) 2008-04-01 In insects, the odorant receptor (Or) multigene family is an intermediate-sized family with genes present in all chromosomes, indicating that duplication followed by interchromosomal transposition played an important role in the early stages of the family evolution. Here, we have explored the occurrence of interchromosomal transpositions in more recent stages through the comparative analysis of a subset of Or genes in Drosophila, where the gene content of chromosomal arms is highly conserved. The studied subset consisted of 11 Or genes located on the left arm of chromosome 3 (Muller's D element) in D. melanogaster. Our study focused on the number and chromosomal arm location of these members of the family across the 12 Drosophila species with complete genome sequences. In contrast to previous results from in situ hybridization comparative mapping that were mainly based on single-copy genes, our study, based on members of a multigene family of moderate size, revealed repeated interchromosomal transposition events and a complex history of some of the studied genes. 11. Neonatal posterior sagittal anorectoplasty for a subset of males with high anorectal malformations Directory of Open Access Journals (Sweden) Nilesh G Nagdeve 2011-01-01 12. A Critical Subset Model Provides a Conceptual Basis for the High Antiviral Activity of Major HIV Drugs** Science.gov (United States) Shen, Lin; Rabi, S. Alireza; Sedaghat, Ahmad R.; Shan, Liang; Lai, Jun; Xing, Sifei; Siliciano, Robert F. 2012-01-01 Control of HIV-1 replication was first achieved with regimens that included a nonnucleoside reverse transcriptase inhibitor (NNRTI) or a protease inhibitor (PI); however, an explanation for the high antiviral activity of these drugs has been lacking. Indeed, conventional pharmacodynamic measures like IC50 (drug concentration causing 50% inhibition) do not differentiate NNRTIs and PIs from less active nucleoside reverse transcriptase inhibitors (NRTIs). Drug inhibitory potential depends on the slope of the dose-response curve (m), which represents how inhibition increases as a function of increasing drug concentration and is related to the Hill coefficient, a measure of intramolecular cooperativity in ligand binding to a multivalent receptor. Although NNRTIs and PIs bind univalent targets, they unexpectedly exhibit cooperative dose-response curves (m > 1). We show that this cooperative inhibition can be explained by a model in which infectivity requires participation of multiple copies of a drug target in an individual life cycle stage. A critical subset of these target molecules must be in the unbound state. Consistent with experimental observations, this model predicts m > 1 for NNRTIs and PIs and m = 1 in situations where a single drug target/virus mediates a step in the life cycle, as is the case with NRTIs and integrase strand transfer inhibitors. This model was tested experimentally by modulating the number of functional drug targets per virus, and dose-response curves for modulated virus populations fit model predictions. This model explains the high antiviral activity of two drug classes important for successful HIV-1 treatment and defines a characteristic of good targets for antiviral drugs in general, namely, intermolecular cooperativity. PMID:21753122 13. Different tumor microenvironments contain functionally distinct subsets of macrophages derived from Ly6C(high) monocytes NARCIS (Netherlands) Movahedi, Kiavash; Laoui, Damya; Gysemans, Conny; Baeten, Martijn; Stangé, Geert; van den Bossche, Jan; Mack, Matthias; Pipeleers, Daniel; In't Veld, Peter; de Baetselier, Patrick; van Ginderachter, Jo A. 2010-01-01 Tumor-associated macrophages (TAM) form a major component of the tumor stroma. However, important concepts such as TAM heterogeneity and the nature of the monocytic TAM precursors remain speculative. Here, we show for the first time that mouse mammary tumors contained functionally distinct subsets 14. T-lymphocyte subsets in HIV-infected and high-risk HIV-uninfected adolescents - Retention of naive T lymphocytes in HIV-infected adolescents NARCIS (Netherlands) Douglas, SD; Rudy, B; Muenz, L; Starr, SE; Campbell, DE; Wilson, C; Holland, C; Crowley-Nowick, P; Vermund, SH Background: The capacity of the immune system of adolescents to generate and repopulate naive and memory cell populations under conditions of normal homeostasis and human immunodeficiency virus (HIV) infection is largely unknown. Objective: To assess lymphocyte subsets in HIV-infected and high-risk 15. Meta-analysis of the risk of gastric stump cancer: detection of high risk patient subsets for stomach cancer after remote partial gastrectomy for benign conditions NARCIS (Netherlands) Tersmette, A. C.; Offerhaus, G. J.; Tersmette, K. W.; Giardiello, F. M.; Moore, G. W.; Tytgat, G. N.; Vandenbroucke, J. P. 1990-01-01 Controversy about gastric cancer risk after partial gastrectomy exists, especially in the United States. Therefore, we performed a meta-analysis to determine overall relative risk and weighted mean relative risk for subsets of postgastrectomy patients, define possible high risk patients suitable for 16. High seroprevalence of Toxoplasma gondii infection in a subset of Mexican patients with work accidents and low socioeconomic status. Science.gov (United States) 2012-01-11 Toxoplasma gondii has been associated with reflex impairment and traffic accidents. It is unknown whether Toxoplasma infection might be associated with work accidents. Therefore, using a case-control seroprevalence study design, 133 patients with a recent work accident and 266 control subjects of the general population from the same region were examined with enzyme-linked immunoassays for the presence and levels of anti-Toxoplasma IgG antibodies and anti-Toxoplasma IgM antibodies. Socio-demographic, work, clinical and behavioral characteristics from each worker were obtained. Eleven (8.3%) of 133 patients, and 14 (5.3%) of 266 controls had anti-T. gondii IgG antibodies. Anti-T. gondii IgG levels were higher than 150 IU/ml in 8 (6%) patients and 10 (3.8%) controls. Anti-T. gondii IgM antibodies were found in one (0.8%) of the workers, and in 6 (2.3%) of the controls. No statistically significant differences in the IgG seroprevalences, frequencies of high IgG levels, and IgM seroprevalences among patients and controls were found. In contrast, a low socio-economic level in patients with work accidents was associated with Toxoplasma seropositivity (P = 0.01). Patients with work accidents and low socioeconomic status showed a significantly (OR = 3.38; 95% CI: 0.84-16.06; P = 0.04) higher seroprevalence of T. gondii infection than controls of the same socioeconomic status (15.1% vs. 5%, respectively). Multivariate analysis showed a positive association of T. gondii infection with boar meat consumption (OR = 3.04; 95% CI: 1.03-8.94; P = 0.04). In contrast, a negative association between T. gondii infection and national trips (OR = 0.40; 95% CI: 0.17-0.96; P = 0.04), sausage consumption (OR = 0.20; 95% CI: 0.05-0.68; P = 0.01), and ham consumption (OR = 0.16; 95% CI: 0.05-0.51; P = 0.002) was found. In the study described here seropositivity to T. gondii was associated to work accidents in a subset of patients with low socioeconomic status. This is the first report of an 17. In Vitro Measles Virus Infection of Human Lymphocyte Subsets Demonstrates High Susceptibility and Permissiveness of both Naive and Memory B Cells. Science.gov (United States) Laksono, Brigitta M; Grosserichter-Wagener, Christina; de Vries, Rory D; Langeveld, Simone A G; Brem, Maarten D; van Dongen, Jacques J M; Katsikis, Peter D; Koopmans, Marion P G; van Zelm, Menno C; de Swart, Rik L 2018-04-15 Measles is characterized by a transient immune suppression, leading to an increased risk of opportunistic infections. Measles virus (MV) infection of immune cells is mediated by the cellular receptor CD150, expressed by subsets of lymphocytes, dendritic cells, macrophages, and thymocytes. Previous studies showed that human and nonhuman primate memory T cells express higher levels of CD150 than naive cells and are more susceptible to MV infection. However, limited information is available about the CD150 expression and relative susceptibility to MV infection of B-cell subsets. In this study, we assessed the susceptibility and permissiveness of naive and memory T- and B-cell subsets from human peripheral blood or tonsils to in vitro MV infection. Our study demonstrates that naive and memory B cells express CD150, but at lower frequencies than memory T cells. Nevertheless, both naive and memory B cells proved to be highly permissive to MV infection. Furthermore, we assessed the susceptibility and permissiveness of various functionally distinct T and B cells, such as helper T (T H ) cell subsets and IgG- and IgA-positive memory B cells, in peripheral blood and tonsils. We demonstrated that T H 1T H 17 cells and plasma and germinal center B cells were the subsets most susceptible and permissive to MV infection. Our study suggests that both naive and memory B cells, along with several other antigen-experienced lymphocytes, are important target cells of MV infection. Depletion of these cells potentially contributes to the pathogenesis of measles immune suppression. IMPORTANCE Measles is associated with immune suppression and is often complicated by bacterial pneumonia, otitis media, or gastroenteritis. Measles virus infects antigen-presenting cells and T and B cells, and depletion of these cells may contribute to lymphopenia and immune suppression. Measles has been associated with follicular exhaustion in lymphoid tissues in humans and nonhuman primates, emphasizing the 18. Unsupervised Feature Subset Selection DEFF Research Database (Denmark) Søndberg-Madsen, Nicolaj; Thomsen, C.; Pena, Jose 2003-01-01 This paper studies filter and hybrid filter-wrapper feature subset selection for unsupervised learning (data clustering). We constrain the search for the best feature subset by scoring the dependence of every feature on the rest of the features, conjecturing that these scores discriminate some ir...... irrelevant features. We report experimental results on artificial and real data for unsupervised learning of naive Bayes models. Both the filter and hybrid approaches perform satisfactorily.... 19. Sleep quality and its relationship with quality of life among high-risk pregnant women (gestational diabetes and hypertension). Science.gov (United States) Saadati, Fatemeh; Sehhatiei Shafaei, Fahimeh; Mirghafourvand, Mozhgan 2018-01-01 Sleep is one of the most basic human requirements. This research aims at determining the status of sleep quality and its relationship with quality of life among high-risk pregnant women in Tabriz, Iran, in 2015. This research was a sectional study done on 364 qualified women in 28-36 weeks of pregnancy suffering from mild preeclampsia and gestational diabetes. The sampling was done as convenience. Personal-social-midwifery questionnaire, Pittsburg sleep quality, and quality of life in pregnancy (QOL-ORAV) were used for gathering data. Multivariate linear regression model was used for determining the relationship between sleep quality and its subsets with quality of life and controlling confounders. In the current study, the prevalence of sleep disturbance was 96.4%. Mean (SD) of the total score of sleep quality was 10.1 (4.1) and the total score of quality of life was 61.7 (17.3). According to Pearson's correlation test, there was statistically significant relationship between quality of life and sleep quality and all its subsets except sleep duration and use of sleep medication (p quality of life. The findings of current research show that sleep quality is low among high-risk pregnant women and quality of life is medium. So, it is necessary that required training is given by health cares for improving sleep quality and quality of life to mothers. 20. Subset selection in regression CERN Document Server Miller, Alan 2002-01-01 Originally published in 1990, the first edition of Subset Selection in Regression filled a significant gap in the literature, and its critical and popular success has continued for more than a decade. Thoroughly revised to reflect progress in theory, methods, and computing power, the second edition promises to continue that tradition. The author has thoroughly updated each chapter, incorporated new material on recent developments, and included more examples and references. New in the Second Edition:A separate chapter on Bayesian methodsComplete revision of the chapter on estimationA major example from the field of near infrared spectroscopyMore emphasis on cross-validationGreater focus on bootstrappingStochastic algorithms for finding good subsets from large numbers of predictors when an exhaustive search is not feasible Software available on the Internet for implementing many of the algorithms presentedMore examplesSubset Selection in Regression, Second Edition remains dedicated to the techniques for fitting... 1. Fast High-Quality Noise DEFF Research Database (Denmark) 2007-01-01 At the moment the noise functions available in a graphics programmer's toolbox are either slow to compute or they involve grid-line artifacts making them of lower quality. In this paper we present a real-time noise computation with no grid-line artifacts or other regularity problems. In other words......, we put a new tool in the box that computes fast high-quality noise. In addition to being free of artifacts, the noise we present does not rely on tabulated data (everything is computed on the fly) and it is easy to adjust quality vs. quantity for the noise. The noise is based on point rendering (like...... spot noise), but it extends to more than two dimensions. The fact that it is based on point rendering makes art direction of the noise much easier.... 2. The Subset Sum game. Science.gov (United States) Darmann, Andreas; Nicosia, Gaia; Pferschy, Ulrich; Schauer, Joachim 2014-03-16 In this work we address a game theoretic variant of the Subset Sum problem, in which two decision makers (agents/players) compete for the usage of a common resource represented by a knapsack capacity. Each agent owns a set of integer weighted items and wants to maximize the total weight of its own items included in the knapsack. The solution is built as follows: Each agent, in turn, selects one of its items (not previously selected) and includes it in the knapsack if there is enough capacity. The process ends when the remaining capacity is too small for including any item left. We look at the problem from a single agent point of view and show that finding an optimal sequence of items to select is an [Formula: see text]-hard problem. Therefore we propose two natural heuristic strategies and analyze their worst-case performance when (1) the opponent is able to play optimally and (2) the opponent adopts a greedy strategy. From a centralized perspective we observe that some known results on the approximation of the classical Subset Sum can be effectively adapted to the multi-agent version of the problem. 3. A subset of osteoblasts expressing high endogenous levels of PPARgamma switches fate to adipocytes in the rat calvaria cell culture model. Directory of Open Access Journals (Sweden) Yuji Yoshiko 2010-07-01 4. Subsets of Women With Close or Positive Margins After Breast-Conserving Surgery With High Local Recurrence Risk Despite Breast Plus Boost Radiotherapy International Nuclear Information System (INIS) Lupe, Krystine; Truong, Pauline T.; Alexander, Cheryl; Lesperance, Mary; Speers, Caroline; Tyldesley, Scott 2011-01-01 Purpose: (1) To examine the effect of surgical margin status on local recurrence (LR) and survival following breast-conserving therapy; (2) To identify subsets with close or positive margins with high LR risk despite whole breast radiotherapy (RT) plus boost. Methods and Materials: Subjects were 2,264 women with pT1–3, any N, M0 invasive breast cancer, treated with breast-conserving surgery and whole breast ± boost RT. Five-year Kaplan-Meier (KM) LR, breast cancer–specific and overall survival (BCSS and OS) were compared between cohorts with negative (n = 1,980), close (n = 222), and positive (n = 62) margins. LR rates were analyzed according to clinicopathologic characteristics. Multivariable Cox regression modeling and matched analysis of close/positive margin cases and negative margin controls were performed. Results: Median follow-up was 5.2 years. Boost RT was used in 92% of patients with close or positive margins. Five-year KM LR rates in the negative, close and positive margin cohorts were 1.3%, 4.0%, and 5.2%, respectively (p = 0.001). BCSS and OS were similar in the three margin subgroups. In the close/positive margin cohort, LR rates were 10.2% with age 10% despite whole breast plus boost RT. These patients should be considered for more definitive surgery. 5. Prediction based on mean subset DEFF Research Database (Denmark) Øjelund, Henrik; Brown, P. J.; Madsen, Henrik 2002-01-01 , it is found that the proposed mean subset method has superior prediction performance than prediction based on the best subset method, and in some settings also better than the ridge regression and lasso methods. The conclusions drawn from the Monte Carlo study is corroborated in an example in which prediction......Shrinkage methods have traditionally been applied in prediction problems. In this article we develop a shrinkage method (mean subset) that forms an average of regression coefficients from individual subsets of the explanatory variables. A Bayesian approach is taken to derive an expression of how...... the coefficient vectors from each subset should be weighted. It is not computationally feasible to calculate the mean subset coefficient vector for larger problems, and thus we suggest an algorithm to find an approximation to the mean subset coefficient vector. In a comprehensive Monte Carlo simulation study... 6. In vitro measles virus infection of human lymphocyte subsets demonstrates high susceptibility and permissiveness of both naive and memory B cells NARCIS (Netherlands) B.M. Laksono (Brigitta); C. Grosserichter-Wagener (Christina); R.D. de Vries (Rory); Langeveld, S.A.G. (Simone A.G.); M.D. Brem (Maarten); J.J.M. van Dongen (Jacques); Katsikis, P.D. (Peter D.); M.P.G. Koopmans D.V.M. (Marion); M.C. van Zelm (Menno); R.L. de Swart (Rik) 2018-01-01 textabstractMeasles is characterized by a transient immune suppression, leading to an increased risk of opportunistic infections. Measles virus (MV) infection of immune cells is mediated by the cellular receptor CD150, expressed by subsets of lymphocytes, dendritic cells, macrophages, and 7. Core Hunter 3: flexible core subset selection. Science.gov (United States) De Beukelaer, Herman; Davenport, Guy F; Fack, Veerle 2018-05-31 Core collections provide genebank curators and plant breeders a way to reduce size of their collections and populations, while minimizing impact on genetic diversity and allele frequency. Many methods have been proposed to generate core collections, often using distance metrics to quantify the similarity of two accessions, based on genetic marker data or phenotypic traits. Core Hunter is a multi-purpose core subset selection tool that uses local search algorithms to generate subsets relying on one or more metrics, including several distance metrics and allelic richness. In version 3 of Core Hunter (CH3) we have incorporated two new, improved methods for summarizing distances to quantify diversity or representativeness of the core collection. A comparison of CH3 and Core Hunter 2 (CH2) showed that these new metrics can be effectively optimized with less complex algorithms, as compared to those used in CH2. CH3 is more effective at maximizing the improved diversity metric than CH2, still ensures a high average and minimum distance, and is faster for large datasets. Using CH3, a simple stochastic hill-climber is able to find highly diverse core collections, and the more advanced parallel tempering algorithm further increases the quality of the core and further reduces variability across independent samples. We also evaluate the ability of CH3 to simultaneously maximize diversity, and either representativeness or allelic richness, and compare the results with those of the GDOpt and SimEli methods. CH3 can sample equally representative cores as GDOpt, which was specifically designed for this purpose, and is able to construct cores that are simultaneously more diverse, and either are more representative or have higher allelic richness, than those obtained by SimEli. In version 3, Core Hunter has been updated to include two new core subset selection metrics that construct cores for representativeness or diversity, with improved performance. It combines and outperforms the 8. Availability of high quality weather data measurements DEFF Research Database (Denmark) Andersen, Elsa; Johansen, Jakob Berg; Furbo, Simon In the period 2016-2017 the project “Availability of high quality weather data measurements” is carried out at Department of Civil Engineering at the Technical University of Denmark. The aim of the project is to establish measured high quality weather data which will be easily available...... for the building energy branch and the solar energy branch in their efforts to achieve energy savings and for researchers and students carrying out projects where measured high quality weather data are needed.... 9. Zerodur polishing process for high surface quality and high efficiency International Nuclear Information System (INIS) Tesar, A.; Fuchs, B. 1992-08-01 Zerodur is a glass-ceramic composite importance in applications where temperature instabilities influence optical and mechanical performance, such as in earthbound and spaceborne telescope mirror substrates. Polished Zerodur surfaces of high quality have been required for laser gyro mirrors. Polished surface quality of substrates affects performance of high reflection coatings. Thus, the interest in improving Zerodur polished surface quality has become more general. Beyond eliminating subsurface damage, high quality surfaces are produced by reducing the amount of hydrated material redeposited on the surface during polishing. With the proper control of polishing parameters, such surfaces exhibit roughnesses of < l Angstrom rms. Zerodur polishing was studied to recommend a high surface quality polishing process which could be easily adapted to standard planetary continuous polishing machines and spindles. This summary contains information on a polishing process developed at LLNL which reproducibly provides high quality polished Zerodur surfaces at very high polishing efficiencies 10. Engineering high quality medical software CERN Document Server Coronato, Antonio 2018-01-01 This book focuses on high-confidence medical software in the growing field of e-health, telecare services and health technology. It covers the development of methodologies and engineering tasks together with standards and regulations for medical software. 11. Quality management manual for production of high quality cassava flour DEFF Research Database (Denmark) Dziedzoave, Nanam Tay; Abass, Adebayo Busura; Amoa-Awua, Wisdom K. The high quality cassava flour (HQCF) industry has just started to evolve in Africa and elsewhere. The sustainability of the growing industry, the profitability of small- and medium-scale enterprises (SMEs) that are active in the industry and good-health of consumers can best be guaranteed through...... the adoption of proper quality and food safety procedures. Cassava processing enterprises involved in the productionof HQCF must therefore be commited to the quality and food safety of the HQCF. They must have the right technology, appropriate processing machhinery, standard testing instruments...... and the necessary technical expertise. This quality manual was therefore developed to guide small- to medium-scale cassava in the design and implematation of Hazard Analysis Critical Control Point (HACCP) system and Good manufacturing Practices (GMP) plans for HQCF production. It describes the HQCF production... 12. High quality-factor optical resonators International Nuclear Information System (INIS) Henriet, Rémi; Salzenstein, Patrice; Coillet, Aurélien; Saleh, Khaldoun; Chembo, Yanne K; Ristic, Davor; Ferrari, Maurizio; Mortier, Michel; Rasoloniaina, Alphonse; Dumeige, Yannick; Féron, Patrice; Cibiel, Gilles; Llopis, Olivier 2014-01-01 Various resonators are investigated for microwave photonic applications. Micro-sphere, disk and fiber ring resonators were designed, realized and characterized. Obtained quality factors are as high as Q = 10 10 . (paper) 13. High-quality compressive ghost imaging Science.gov (United States) Huang, Heyan; Zhou, Cheng; Tian, Tian; Liu, Dongqi; Song, Lijun 2018-04-01 We propose a high-quality compressive ghost imaging method based on projected Landweber regularization and guided filter, which effectively reduce the undersampling noise and improve the resolution. In our scheme, the original object is reconstructed by decomposing of regularization and denoising steps instead of solving a minimization problem in compressive reconstruction process. The simulation and experimental results show that our method can obtain high ghost imaging quality in terms of PSNR and visual observation. 14. Fabrication of high-quality brazed joints International Nuclear Information System (INIS) Orlov, A.V. 1980-01-01 Problem of ensuring of joint high-quality when brazing different parts in power engineering is considered. To obtain high-quality joints it is necessary to correctly design brazed joint and to choose a gap width, overlap length and fillet radius; to clean up carefully the surfaces to be brazed and fix them properly one relative to another; to apply a solder so as to provide its flowing into the gap and sticking in it; to exactly regulate thermal conditions of brazing. High quality and reliability of brazed joints are ensured by the application of solders based on noble metals, and cheap solders based on nickel, manganese and copper. Joints brazed with nickel base solders may operate at temperatures as high as 888 deg C 15. TOTAL QUALITY MANAGEMENT IN HIGH EDUCATION Directory of Open Access Journals (Sweden) Hasan SERİN, Alper AYTEKİN 2009-01-01 Full Text Available The approach of Total Quality Management (TQM has been even more common and most recently its use in high education has been discussed. Likewise the enterprises producing various products, universities have also inputs, processes, and outputs. Due to conditions of competition, universities have to improve the qualities of these inputs, processes, and outputs, according to satisfaction, demands, and expectations of internal and external customers. If the TQM has been implemented in the universities with a manner that aims for customer satisfaction (students, lecturers, public and private establishments, and families, supports constant development, ensures participatory approach, and encourages working in groups, it will provide universities with effectiveness, efficiency, dynamics, and economics. In this study, common problems of universities, definitions of quality and TQM in high education, customer concept at universities, and factors affecting the quality of education have been explained. Besides, in order TQM approach to be successfully implemented in the universities, various suggestions have been presented. 16. Regulatory effects of resveratrol on glucose metabolism and T-lymphocyte subsets in the development of high-fat diet-induced obesity in C57BL/6 mice. Science.gov (United States) Wang, Bin; Sun, Jin; Li, Longnan; Zheng, Jing; Shi, Yonghui; Le, Guowei 2014-07-25 High-fat diet (HFD)-induced obesity is often associated with immune dysfunction. Resveratrol (trans-3,5,4'-trihydroxystilbene), which has well-founded immunity-related beneficial properties, was used to elucidate the regulatory effect on glucose metabolism and T-lymphocyte subsets in the development of HFD-induced obesity. Resveratrol, being associated with decreases of plasma leptin and plasma lipids and the release of oxidative stress, significantly decreased the body weight and fat masses in HF mice after 26 weeks of feeding. Furthermore, resveratrol decreased the fasting blood glucose and fasting plasma insulin and increased the CD3(+)CD4(+)/CD3(+)CD8(+) subsets percentages and the regulatory T cells (Tregs) production after 13 and 26 weeks of feeding. The results indicate that resveratrol, as an effective supplement for HFD, maintained glucose homeostasis by activating the PI3K and SIRT1 signaling pathways. Moreover, resveratrol activated the Nrf2 signaling pathway-mediated antioxidant enzyme expression to alleviate inflammation by protecting against oxidative damage and T-lymphocyte subset-related chronic inflammatory response in the development of HFD-induced obesity. 17. Efficient Secure Multiparty Subset Computation Directory of Open Access Journals (Sweden) Sufang Zhou 2017-01-01 Full Text Available Secure subset problem is important in secure multiparty computation, which is a vital field in cryptography. Most of the existing protocols for this problem can only keep the elements of one set private, while leaking the elements of the other set. In other words, they cannot solve the secure subset problem perfectly. While a few studies have addressed actual secure subsets, these protocols were mainly based on the oblivious polynomial evaluations with inefficient computation. In this study, we first design an efficient secure subset protocol for sets whose elements are drawn from a known set based on a new encoding method and homomorphic encryption scheme. If the elements of the sets are taken from a large domain, the existing protocol is inefficient. Using the Bloom filter and homomorphic encryption scheme, we further present an efficient protocol with linear computational complexity in the cardinality of the large set, and this is considered to be practical for inputs consisting of a large number of data. However, the second protocol that we design may yield a false positive. This probability can be rapidly decreased by reexecuting the protocol with different hash functions. Furthermore, we present the experimental performance analyses of these protocols. 18. Effect of forage quality in faeces from different ruminant species fed high and low quality forage DEFF Research Database (Denmark) Jalali, A R; Nørgaard, P; Nielsen, M O 2010-01-01 Effect of forage quality in faeces from different ruminant species fed high and low quality forage......Effect of forage quality in faeces from different ruminant species fed high and low quality forage... 19. Assuring quality in high-consequence engineering Energy Technology Data Exchange (ETDEWEB) Hoover, Marcey L.; Kolb, Rachel R. 2014-03-01 In high-consequence engineering organizations, such as Sandia, quality assurance may be heavily dependent on staff competency. Competency-dependent quality assurance models are at risk when the environment changes, as it has with increasing attrition rates, budget and schedule cuts, and competing program priorities. Risks in Sandia's competency-dependent culture can be mitigated through changes to hiring, training, and customer engagement approaches to manage people, partners, and products. Sandia's technical quality engineering organization has been able to mitigate corporate-level risks by driving changes that benefit all departments, and in doing so has assured Sandia's commitment to excellence in high-consequence engineering and national service. 20. Producing high-quality slash pine seeds Science.gov (United States) James Barnett; Sue Varela 2003-01-01 Slash pine is a desirable species. It serves many purposes and is well adapted to poorly drained flatwoods and seasonally flooded areas along the lower Coastal Plain of the Southeastern US. The use of high-quality seeds has been shown to produce uniform seedlings for outplanting, which is key to silvicultural success along the Coastal Plain and elsewhere. We present... 1. Endorectal high dose rate brachytherapy quality assurance International Nuclear Information System (INIS) Devic, S.; Vuong, T.; Evans, M.; Podgorsak, E. 2008-01-01 We describe our quality assurance method for preoperative high dose rate (HDR) brachytherapy of endorectal tumours. Reproduction of the treatment planning dose distribution on a daily basis is crucial for treatment success. Due to the cylindrical symmetry, two types of adjustments are necessary: applicator rotation and dose distribution shift along the applicator axis. (author) 2. High Quality Virtual Reality for Architectural Exhibitions DEFF Research Database (Denmark) Kreutzberg, Anette 2016-01-01 This paper will summarise the findings from creating and implementing a visually high quality Virtual Reality (VR) experiment as part of an international architecture exhibition. It was the aim to represent the architectural spatial qualities as well as the atmosphere created from combining natural...... and artificial lighting in a prominent not yet built project. The outcome is twofold: Findings concerning the integration of VR in an exhibition space and findings concerning the experience of the virtual space itself. In the exhibition, an important aspect was the unmanned exhibition space, requiring the VR...... experience to be self-explanatory. Observations of different visitor reactions to the unmanned VR experience compared with visitor reactions at guided tours with personal instructions are evaluated. Data on perception of realism, spatial quality and light in the VR model were collected with qualitative... 3. High quality transportation fuels from renewable feedstock Energy Technology Data Exchange (ETDEWEB) Lindfors, Lars Peter 2010-09-15 Hydrotreating of vegetable oils is novel process for producing high quality renewable diesel. Hydrotreated vegetable oils (HVO) are paraffinic hydrocarbons. They are free of aromatics, have high cetane numbers and reduce emissions. HVO can be used as component or as such. HVO processes can also be modified to produce jet fuel. GHG savings by HVO use are significant compared to fossil fuels. HVO is already in commercial production. Neste Oil is producing its NExBTL diesel in two plants. Production of renewable fuels will be limited by availability of sustainable feedstock. Therefore R and D efforts are made to expand feedstock base further. 4. Boiling curve in high quality flow boiling International Nuclear Information System (INIS) Shiralkar, B.S.; Hein, R.A.; Yadigaroglu, G. 1980-01-01 The post dry-out heat transfer regime of the flow boiling curve was investigated experimentally for high pressure water at high qualities. The test section was a short round tube located downstream of a hot patch created by a temperature controlled segment of tubing. Results from the experiment showed that the distance from the dryout point has a significant effect on the downstream temperatures and there was no unique boiling curve. The heat transfer coefficients measured sufficiently downstream of the dryout point could be correlated using the Heineman correlation for superheated steam, indicating that the droplet deposition effects could be neglected in this region 5. Breeding and maintaining high-quality insects DEFF Research Database (Denmark) Jensen, Kim; Kristensen, Torsten Nygård; Heckmann, Lars-Henrik 2017-01-01 Insects have a large potential for sustainably enhancing global food and feed production, and commercial insect production is a rising industry of high economic value. Insects suitable for production typically have fast growth, short generation time, efficient nutrient utilization, high...... reproductive potential, and thrive at high density. Insects may cost-efficiently convert agricultural and industrial food by-products into valuable protein once the technology is finetuned. However, since insect mass production is a new industry, the technology needed to efficiently farm these animals is still...... in a starting phase. Here, we discuss the challenges and precautions that need to be considered when breeding and maintaining high-quality insect populations for food and feed. This involves techniques typically used in domestic animal breeding programs including maintaining genetically healthy populations... 6. Overexpression of the A Disintegrin and Metalloproteinase ADAM15 is linked to a Small but Highly Aggressive Subset of Prostate Cancers Directory of Open Access Journals (Sweden) Christoph Burdelski 2017-04-01 7. Plasma Cell Alloantigen 1 and IL-10 Secretion Define Two Distinct Peritoneal B1a B Cell Subsets With Opposite Functions, PC1high Cells Being Protective and PC1low Cells Harmful for the Growing Fetus Directory of Open Access Journals (Sweden) Anne Schumacher 2018-05-01 Full Text Available B cells possess various immuno regulatory functions. However, research about their participation in tolerance induction toward the fetus is just emerging. Accumulating evidence supports the idea that B cells can play seemingly conflicting roles during pregnancy, either protecting or harming the fetus. Previous findings indicated the presence of two different peritoneal B cell subsets, defined by the expression of the plasma cell alloantigen 1 (PC1 and with distinct immune modulatory functions. Here, we aimed to study the participation of these two B cell subsets, on pregnancy outcome in a murine model of disturbed fetal tolerance. The frequencies and cell numbers of peritoneal and splenic CD19+IL-10+ and CD19+CD5+IL-10+PC1+ cells were assessed in virgin as well as normal pregnant (NP and abortion-prone (AP females during the course of gestation. Peritoneal PC1low or PC1high B1a B cells were sorted, analyzed for their ability to secrete IL-10 and adoptively transferred into NP or AP females. On gestation day (gd 12, the abortion rate as well as the frequencies and cell numbers of regulatory T cells, TH1 and TH17 cells were determined in spleens and decidua. In addition, mRNA expression of IL-10, TGF-β, IFN-γ, and TNF-α was analyzed in decidual tissue. Peritoneal CD19+IL-10+ and CD19+CD5+IL-10+PC1+ frequencies fluctuated during the progression of normal pregnancies while no significant changes were observed in spleen. AP females showed significantly reduced frequencies of both B cell populations and exhibited an altered peritoneal PC1high/PC1low ratio at gd10. Adoptive transfers of PC1low B1a B cells into NP females increased the abortion rate in association with a reduced splenic regulatory T/TH17 ratio. By contrast, the transfer of PC1high B1a B cells into AP females significantly diminished the fetal rejection rate and significantly reduced the numbers of splenic TH17 cells. Our results suggest that the peritoneum harbors two distinct B1a B 8. MODIS/Aqua Atmosphere Aeronet Subsetting Product Data.gov (United States) National Aeronautics and Space Administration — The MODIS/Aqua Atmosphere Aeronet Subsetting Product (MYDARNSS) consists of MODIS Atmosphere and Ancillary Products subsets that are generated over a number of... 9. On subset selection from Logistic populations NARCIS (Netherlands) Laan, van der P. 1990-01-01 Some distributional results are derived for subset selection from Logistic populations, differing only in their location parameter. The probability of correct selection is determined. Exact and numerical results concerning the expected subset size are presented. 10. Method for synthesis of high quality graphene Science.gov (United States) Lanzara, Alessandra [Piedmont, CA; Schmid, Andreas K [Berkeley, CA; Yu, Xiaozhu [Berkeley, CA; Hwang, Choonkyu [Albany, CA; Kohl, Annemarie [Beneditkbeuern, DE; Jozwiak, Chris M [Oakland, CA 2012-03-27 A method is described herein for the providing of high quality graphene layers on silicon carbide wafers in a thermal process. With two wafers facing each other in close proximity, in a first vacuum heating stage, while maintained at a vacuum of around 10.sup.-6 Torr, the wafer temperature is raised to about 1500.degree. C., whereby silicon evaporates from the wafer leaving a carbon rich surface, the evaporated silicon trapped in the gap between the wafers, such that the higher vapor pressure of silicon above each of the wafers suppresses further silicon evaporation. As the temperature of the wafers is raised to about 1530.degree. C. or more, the carbon atoms self assemble themselves into graphene. 11. High quality data: An evaluation of AIM data quality and data quality procedures Science.gov (United States) The goal of every monitoring program is to collect high-quality data which can then be used to provide information to decision makers. The Bureau of Land Management (BLM) Assessment, Inventory, and Monitoring (AIM) program is one such data set which provides rangeland status, condition, and trend in... 12. High-picture quality industrial CT scanner International Nuclear Information System (INIS) Shoji, Takao; Nishide, Akihiko; Fujii, Masashi. 1989-01-01 Industrial X-ray-CT-scanners, which provide cross-sectional images of a tested sample without destroying it, are attracting attention as a new nondestructive inspection device. In 1982, Toshiba commenced the development of industrial CT scanners, and introduced the 'TOSCANER' -3000 and-4000 series. Now, the state of the art 'TOSCANER'-20000 series of CT systems has been developed incorporating the latest computer tomography and image processing technology, such as the T9506 image processor. One of the advantages of this system is its applicability to a wide range of X-ray energy . The 'TOSCANER'-20000 series can be utilized for inspecting castings and other materials with relatively low-transparency to X-rays, as well as ceramics, composite materials and other materials with high X-ray transparency. A further feature of the new system is its high-picture quality, with a high-spatial resolution resulting from a pixel size of 0.2x0.2(mm). (author) 13. [Quality management is associated with high quality services in health care]. Science.gov (United States) Nielsen, Tenna Hassert; Riis, Allan; Mainz, Jan; Jensen, Anne-Louise Degn 2013-12-09 In these years, quality management has been the focus in order to meet high quality services for the patients in Danish health care. This article provides information on quality management and quality improvement and it evaluates its effectiveness in achieving better organizational structures, processes and results in Danish health-care organizations. Our findings generally support that quality management is associated with high quality services in health care. 14. Optical studies of high quality synthetic diamond International Nuclear Information System (INIS) Sharp, S.J. 1999-01-01 This thesis is concerned with the study of fundamental and defect induced optical properties of synthetic diamond grown using high pressure, high temperature (HPHT) synthesis or chemical vapour deposition (CVD). The primary technique used for investigation is cathodoluminescence (including imaging and decay-time measurements) in addition to other forms of optical spectroscopy. This thesis is timely in that the crystallinity and purity of synthetic diamond has increased ten fold over the last few years. The diamond exciton emission, which is easily quenched by the presence of defects, is studied in high quality samples in detail. In addition the ability now exists to engineer the isotopic content of synthetic diamond to a high degree of accuracy. The experimental chapters are divided as follows: Chapter 2: High resolution, low temperature spectra reveal a splitting of the free-exciton phonon recombination emission peaks and the bound-exciton zero phonon line. Included are measurements of the variation in intensity and decay-time as a function of temperature. Chapter 3: The shift in energy of the phonon-assisted free-exciton phonon replicas with isotopic content has been measured. The shift is in agreement with the results of interatomic force model for phonon scattering due to isotope disorder. Chapter 4: A study of the shift in energy with isotopic content of the diamond of the GR1 band due to the neutral vacancy has allowed a verification of the theoretical predictions due to the Jahn Teller effect. Chapter 5: The spatial distribution of the free-exciton luminescence is studied in HPHT synthetic and CVD diamond. A variation in intensity with distance from the surface is interpreted as a significant non-radiative loss of excitons to the surface. Chapter 6: The decay-times of all known self-interstitial related centres have been measured in order to calculate the concentration of these centres present in electron irradiated diamond. (author) 15. High Quality Data for Grid Integration Studies Energy Technology Data Exchange (ETDEWEB) Clifton, Andrew; Draxl, Caroline; Sengupta, Manajit; Hodge, Bri-Mathias 2017-01-22 As variable renewable power penetration levels increase in power systems worldwide, renewable integration studies are crucial to ensure continued economic and reliable operation of the power grid. The existing electric grid infrastructure in the US in particular poses significant limitations on wind power expansion. In this presentation we will shed light on requirements for grid integration studies as far as wind and solar energy are concerned. Because wind and solar plants are strongly impacted by weather, high-resolution and high-quality weather data are required to drive power system simulations. Future data sets will have to push limits of numerical weather prediction to yield these high-resolution data sets, and wind data will have to be time-synchronized with solar data. Current wind and solar integration data sets are presented. The Wind Integration National Dataset (WIND) Toolkit is the largest and most complete grid integration data set publicly available to date. A meteorological data set, wind power production time series, and simulated forecasts created using the Weather Research and Forecasting Model run on a 2-km grid over the continental United States at a 5-min resolution is now publicly available for more than 126,000 land-based and offshore wind power production sites. The National Solar Radiation Database (NSRDB) is a similar high temporal- and spatial resolution database of 18 years of solar resource data for North America and India. The need for high-resolution weather data pushes modeling towards finer scales and closer synchronization. We also present how we anticipate such datasets developing in the future, their benefits, and the challenges with using and disseminating such large amounts of data. 16. Subset simulation for structural reliability sensitivity analysis International Nuclear Information System (INIS) Song Shufang; Lu Zhenzhou; Qiao Hongwei 2009-01-01 Based on two procedures for efficiently generating conditional samples, i.e. Markov chain Monte Carlo (MCMC) simulation and importance sampling (IS), two reliability sensitivity (RS) algorithms are presented. On the basis of reliability analysis of Subset simulation (Subsim), the RS of the failure probability with respect to the distribution parameter of the basic variable is transformed as a set of RS of conditional failure probabilities with respect to the distribution parameter of the basic variable. By use of the conditional samples generated by MCMC simulation and IS, procedures are established to estimate the RS of the conditional failure probabilities. The formulae of the RS estimator, its variance and its coefficient of variation are derived in detail. The results of the illustrations show high efficiency and high precision of the presented algorithms, and it is suitable for highly nonlinear limit state equation and structural system with single and multiple failure modes 17. Subset Selection by Local Convex Approximation DEFF Research Database (Denmark) 1999-01-01 This paper concerns selection of the optimal subset of variables in a lenear regression setting. The posed problem is combinatiorial and the globally best subset can only be found in exponential time. We define a cost function for the subset selection problem by adding the penalty term to the usual...... of the subset selection problem so as to guarantee positive definiteness of the Hessian term, hence avoiding numerical instability. The backward Elemination type algorithm attempts to improve the results upon termination of the modified Newton-Raphson search by sing the current solution as an initial guess... 18. Ensuring High-Quality Learning for All Science.gov (United States) Núñez, Elsa M. 2018-01-01 The Association of American Colleges and Universities (AAC&U) has embarked on a sustained program to enhance the quality of student learning on campuses, while also supporting AAC&U members' efforts to bring liberal education to all sectors of society. This commitment to quality and equity in service to democracy forms the basis for… 19. High Efficiency, Illumination Quality OLEDs for Lighting Energy Technology Data Exchange (ETDEWEB) Joseph Shiang; James Cella; Kelly Chichak; Anil Duggal; Kevin Janora; Chris Heller; Gautam Parthasarathy; Jeffery Youmans; Joseph Shiang 2008-03-31 The goal of the program was to demonstrate a 45 lumen per watt white light device based upon the use of multiple emission colors through the use of solution processing. This performance level is a dramatic extension of the team's previous 15 LPW large area illumination device. The fundamental material system was based upon commercial polymer materials. The team was largely able to achieve these goals, and was able to deliver to DOE a 90 lumen illumination source that had an average performance of 34 LPW a 1000 cd/m{sup 2} with peak performances near 40LPW. The average color temperature is 3200K and the calculated CRI 85. The device operated at a brightness of approximately 1000cd/m{sup 2}. The use of multiple emission colors particularly red and blue, provided additional degrees of design flexibility in achieving white light, but also required the use of a multilayered structure to separate the different recombination zones and prevent interconversion of blue emission to red emission. The use of commercial materials had the advantage that improvements by the chemical manufacturers in charge transport efficiency, operating life and material purity could be rapidly incorporated without the expenditure of additional effort. The program was designed to take maximum advantage of the known characteristics of these material and proceeded in seven steps. (1) Identify the most promising materials, (2) assemble them into multi-layer structures to control excitation and transport within the OLED, (3) identify materials development needs that would optimize performance within multilayer structures, (4) build a prototype that demonstrates the potential entitlement of the novel multilayer OLED architecture (5) integrate all of the developments to find the single best materials set to implement the novel multilayer architecture, (6) further optimize the best materials set, (7) make a large area high illumination quality white OLED. A photo of the final deliverable is shown 20. Analysis of high-quality modes in open chaotic microcavities International Nuclear Information System (INIS) Fang, W.; Yamilov, A.; Cao, H. 2005-01-01 We present a numerical study of the high-quality modes in two-dimensional dielectric stadium microcavities. Although the classical ray mechanics is fully chaotic in a stadium billiard, all of the high-quality modes show a 'strong scar' around unstable periodic orbits. When the deformation (ratio of the length of the straight segments over the diameter of the half circles) is small, the high-quality modes correspond to whispering-gallery-type trajectories and their quality factors decrease monotonically with increasing deformation. At large deformation, each high-quality mode is associated with multiple unstable periodic orbits. Its quality factor changes nonmonotonically with the deformation, and there exists an optimal deformation for each mode at which its quality factor reaches a local maximum. This unusual behavior is attributed to the interference of waves propagating along different constituent orbits that could minimize light leakage out of the cavity 1. High quality steel casting for energy technics International Nuclear Information System (INIS) Schuster, F.; Koefler, G. 1982-01-01 The casting of several chromium-molybdenum steels for steam and hydraulic turbines is discussed. Non-destructive testing of the castings is performed demonstrating the safety for use in nuclear technology. The effect of metallurgical parameters on steel casting quality, the heat treatment, and the effect of construction design on costs for fettling and repair weldings are considered. (Auth.) 2. Two-stage atlas subset selection in multi-atlas based image segmentation. Science.gov (United States) Zhao, Tingting; Ruan, Dan 2015-06-01 Fast growing access to large databases and cloud stored data presents a unique opportunity for multi-atlas based image segmentation and also presents challenges in heterogeneous atlas quality and computation burden. This work aims to develop a novel two-stage method tailored to the special needs in the face of large atlas collection with varied quality, so that high-accuracy segmentation can be achieved with low computational cost. An atlas subset selection scheme is proposed to substitute a significant portion of the computationally expensive full-fledged registration in the conventional scheme with a low-cost alternative. More specifically, the authors introduce a two-stage atlas subset selection method. In the first stage, an augmented subset is obtained based on a low-cost registration configuration and a preliminary relevance metric; in the second stage, the subset is further narrowed down to a fusion set of desired size, based on full-fledged registration and a refined relevance metric. An inference model is developed to characterize the relationship between the preliminary and refined relevance metrics, and a proper augmented subset size is derived to ensure that the desired atlases survive the preliminary selection with high probability. The performance of the proposed scheme has been assessed with cross validation based on two clinical datasets consisting of manually segmented prostate and brain magnetic resonance images, respectively. The proposed scheme demonstrates comparable end-to-end segmentation performance as the conventional single-stage selection method, but with significant computation reduction. Compared with the alternative computation reduction method, their scheme improves the mean and medium Dice similarity coefficient value from (0.74, 0.78) to (0.83, 0.85) and from (0.82, 0.84) to (0.95, 0.95) for prostate and corpus callosum segmentation, respectively, with statistical significance. The authors have developed a novel two-stage atlas 3. Two-stage atlas subset selection in multi-atlas based image segmentation Energy Technology Data Exchange (ETDEWEB) Zhao, Tingting, E-mail: [email protected]; Ruan, Dan, E-mail: [email protected] [The Department of Radiation Oncology, University of California, Los Angeles, California 90095 (United States) 2015-06-15 Purpose: Fast growing access to large databases and cloud stored data presents a unique opportunity for multi-atlas based image segmentation and also presents challenges in heterogeneous atlas quality and computation burden. This work aims to develop a novel two-stage method tailored to the special needs in the face of large atlas collection with varied quality, so that high-accuracy segmentation can be achieved with low computational cost. Methods: An atlas subset selection scheme is proposed to substitute a significant portion of the computationally expensive full-fledged registration in the conventional scheme with a low-cost alternative. More specifically, the authors introduce a two-stage atlas subset selection method. In the first stage, an augmented subset is obtained based on a low-cost registration configuration and a preliminary relevance metric; in the second stage, the subset is further narrowed down to a fusion set of desired size, based on full-fledged registration and a refined relevance metric. An inference model is developed to characterize the relationship between the preliminary and refined relevance metrics, and a proper augmented subset size is derived to ensure that the desired atlases survive the preliminary selection with high probability. Results: The performance of the proposed scheme has been assessed with cross validation based on two clinical datasets consisting of manually segmented prostate and brain magnetic resonance images, respectively. The proposed scheme demonstrates comparable end-to-end segmentation performance as the conventional single-stage selection method, but with significant computation reduction. Compared with the alternative computation reduction method, their scheme improves the mean and medium Dice similarity coefficient value from (0.74, 0.78) to (0.83, 0.85) and from (0.82, 0.84) to (0.95, 0.95) for prostate and corpus callosum segmentation, respectively, with statistical significance. Conclusions: The authors 4. Two-stage atlas subset selection in multi-atlas based image segmentation International Nuclear Information System (INIS) Zhao, Tingting; Ruan, Dan 2015-01-01 Purpose: Fast growing access to large databases and cloud stored data presents a unique opportunity for multi-atlas based image segmentation and also presents challenges in heterogeneous atlas quality and computation burden. This work aims to develop a novel two-stage method tailored to the special needs in the face of large atlas collection with varied quality, so that high-accuracy segmentation can be achieved with low computational cost. Methods: An atlas subset selection scheme is proposed to substitute a significant portion of the computationally expensive full-fledged registration in the conventional scheme with a low-cost alternative. More specifically, the authors introduce a two-stage atlas subset selection method. In the first stage, an augmented subset is obtained based on a low-cost registration configuration and a preliminary relevance metric; in the second stage, the subset is further narrowed down to a fusion set of desired size, based on full-fledged registration and a refined relevance metric. An inference model is developed to characterize the relationship between the preliminary and refined relevance metrics, and a proper augmented subset size is derived to ensure that the desired atlases survive the preliminary selection with high probability. Results: The performance of the proposed scheme has been assessed with cross validation based on two clinical datasets consisting of manually segmented prostate and brain magnetic resonance images, respectively. The proposed scheme demonstrates comparable end-to-end segmentation performance as the conventional single-stage selection method, but with significant computation reduction. Compared with the alternative computation reduction method, their scheme improves the mean and medium Dice similarity coefficient value from (0.74, 0.78) to (0.83, 0.85) and from (0.82, 0.84) to (0.95, 0.95) for prostate and corpus callosum segmentation, respectively, with statistical significance. Conclusions: The authors 5. Variable and subset selection in PLS regression DEFF Research Database (Denmark) Høskuldsson, Agnar 2001-01-01 The purpose of this paper is to present some useful methods for introductory analysis of variables and subsets in relation to PLS regression. We present here methods that are efficient in finding the appropriate variables or subset to use in the PLS regression. The general conclusion...... is that variable selection is important for successful analysis of chemometric data. An important aspect of the results presented is that lack of variable selection can spoil the PLS regression, and that cross-validation measures using a test set can show larger variation, when we use different subsets of X, than... 6. Indoor Air Quality in High Performance Schools Science.gov (United States) High performance schools are facilities that improve the learning environment while saving energy, resources, and money. The key is understanding the lifetime value of high performance schools and effectively managing priorities, time, and budget. 7. Identification of a novel stereotypic IGHV4-59/IGHJ5-encoded B-cell receptor subset expressed by various B-cell lymphomas with high affinity rheumatoid factor activity NARCIS (Netherlands) Bende, Richard J.; Janssen, Jerry; Wormhoudt, Thera A. M.; Wagner, Koen; Guikema, Jeroen E. J.; van Noesel, Carel J. M. 2016-01-01 Subsets of mucosa-associated lymphoid tissue (MALT) lymphoma, splenic marginal zone B-cell lymphoma (MZBCL) and chronic lymphocytic leukemia (CLL) have been identified that express near-identical B-cell receptors (BCRs), strongly suggesting selection by restricted antigenic epitopes. We here report 8. Development of a synchronous subset of AADL DEFF Research Database (Denmark) Filali, Mamoun; Lawall, Julia 2010-01-01 We study the definition and the mapping of an AADL subset: the so called synchronous subset. We show that the data port protocol used for delayed and immediate connections between periodic threads can be interpreted in a  synchronous way. In this paper, we formalize this interpretation and study ...... the development of its mapping such that the original synchronous semantics is preserved. For that purpose, we use refinements through the Event B method.... 9. Quality Assurance Strategy for Existing Homes: Final Quality Management Primer for High Performing Homes Energy Technology Data Exchange (ETDEWEB) Del Bianco, M.; Taggart, J.; Sikora, J.; Wood, A. 2012-12-01 This guide is designed to help Building America (BA) Teams understand quality management and its role in transitioning from conventional to high performance home building and remodeling. It explains what quality means, the value of quality management systems, the unique need for QMS when building high performing homes, and the first steps to a implementing a comprehensive QMS. This document provides a framework and context for BA teams when they encounter builders and remodelers. 10. Quality Assurance Strategy for Existing Homes. Final Quality Management Primer for High Performing Homes Energy Technology Data Exchange (ETDEWEB) Del Bianco, M. [NAHB Research Center Industry Partnership, Upper Marlboro, MD (United States); Taggart, J. [NAHB Research Center Industry Partnership, Upper Marlboro, MD (United States); Sikora, J. [NAHB Research Center Industry Partnership, Upper Marlboro, MD (United States); Wood, A. [NAHB Research Center Industry Partnership, Upper Marlboro, MD (United States) 2012-12-01 This guide is designed to help Building America (BA) teams understand quality management and its role in transitioning from conventional to high performance home building and remodeling. It explains what quality means, the value of quality management systems, the unique need for QMS when building high performing homes, and the first steps to a implementing a comprehensive QMS. This document provides a framework and context for BA teams when they encounter builders and remodelers. 11. Inequality in Preschool Quality? Community-Level Disparities in Access to High-Quality Learning Environments Science.gov (United States) Bassok, Daphna; Galdo, Eva 2016-01-01 In recent years, unequal access to high-quality preschool has emerged as a growing public policy concern. Because of data limitations, it is notoriously difficult to measure disparities in access to early learning opportunities across communities and particularly challenging to quantify gaps in access to "high-quality" programs. Research… 12. Band Subset Selection for Hyperspectral Image Classification Directory of Open Access Journals (Sweden) Chunyan Yu 2018-01-01 Full Text Available This paper develops a new approach to band subset selection (BSS for hyperspectral image classification (HSIC which selects multiple bands simultaneously as a band subset, referred to as simultaneous multiple band selection (SMMBS, rather than one band at a time sequentially, referred to as sequential multiple band selection (SQMBS, as most traditional band selection methods do. In doing so, a criterion is particularly developed for BSS that can be used for HSIC. It is a linearly constrained minimum variance (LCMV derived from adaptive beamforming in array signal processing which can be used to model misclassification errors as the minimum variance. To avoid an exhaustive search for all possible band subsets, two numerical algorithms, referred to as sequential (SQ and successive (SC algorithms are also developed for LCMV-based SMMBS, called SQ LCMV-BSS and SC LCMV-BSS. Experimental results demonstrate that LCMV-based BSS has advantages over SQMBS. 13. Prediction of bread-making quality using size exclusion high ... African Journals Online (AJOL) Variation in the distribution of protein molecular weight in wheat (Triticum aestivum), influences breadmaking quality of wheat cultivars, resulting in either poor or good bread. The objective of this study was to predict breadmaking quality of wheat cultivars using size exclusion high performance liquid chromatography. 14. Adoption and impact of high quality bambara flour (HQBF ... African Journals Online (AJOL) Adoption and impact of high quality bambara flour (HQBF) technology in the ... consumer acceptability/quality of products, credit, availability of raw materials, and ... as a result of 12.5 per cent increase in demand for bambara-based products. 15. Quack: A quality assurance tool for high throughput sequence data. Science.gov (United States) Thrash, Adam; Arick, Mark; Peterson, Daniel G 2018-05-01 The quality of data generated by high-throughput DNA sequencing tools must be rapidly assessed in order to determine how useful the data may be in making biological discoveries; higher quality data leads to more confident results and conclusions. Due to the ever-increasing size of data sets and the importance of rapid quality assessment, tools that analyze sequencing data should quickly produce easily interpretable graphics. Quack addresses these issues by generating information-dense visualizations from FASTQ files at a speed far surpassing other publicly available quality assurance tools in a manner independent of sequencing technology. Copyright © 2018 The Authors. Published by Elsevier Inc. All rights reserved. 16. Innovative and high quality education through Open Education and OER OpenAIRE Stracke, Christian M. 2017-01-01 Online presentation and webinar by Stracke, C. M. (2017, 18 December) on "Innovative and high quality education through Open Education and OER" for the Belt and Road Open Education Learning Week by the Beijing Normal University, China. 17. Improving high quality, equitable maternal health services in Malawi ... International Development Research Centre (IDRC) Digital Library (Canada) Improving high quality, equitable maternal health services in Malawi (IMCHA) ... In response, the Ministry of Health implemented a Standards-Based Management and Recognition for Reproductive Health initiative to improve ... Total funding. 18. High Quality Education and Learning for All through Open Education NARCIS (Netherlands) Stracke, Christian M. 2016-01-01 Keynote at the International Lensky Education Forum 2016, Yakutsk, Republic of Sakha, Russian Federation, by Stracke, C. M. (2016, 16 August): "High Quality Education and Learning for All through Open Education" 19. Emblems of Quality in Higher Education. Developing and Sustaining High-Quality Programs. Science.gov (United States) Haworth, Jennifer Grant; Conrad, Clifton F. This book proposes an "engagement" theory of program quality to evaluate and improve higher education programs at all degree levels. Based on interviews with 781 participants in a national study of Masters degree programs, it focuses on the interactive roles of students, faculty, and administrators in developing high-quality programs… 20. Distributional and efficiency results for subset selection NARCIS (Netherlands) Laan, van der P. 1996-01-01 Assume k (??k \\geq 2) populations are given. The associated independent random variables have continuous distribution functions with an unknown location parameter. The statistical selec??tion goal is to select a non??empty subset which contains the best population,?? that is the pop??ulation with 1. Invariant subsets under compact quantum group actions OpenAIRE Huang, Huichi 2012-01-01 We investigate compact quantum group actions on unital $C^*$-algebras by analyzing invariant subsets and invariant states. In particular, we come up with the concept of compact quantum group orbits and use it to show that countable compact metrizable spaces with infinitely many points are not quantum homogeneous spaces. 2. Subset specification of central serotonergic neurons Directory of Open Access Journals (Sweden) Marten P Smidt 2013-10-01 Full Text Available The last decade the serotonin (5-hydroxytryptamine; 5-HT system has received enormous attention due to its role in regulation of behavior, exemplified by the discovery that increased 5-HT tone in the central nervous system is able to alleviate affective disorders. Here, we review the developmental processes, with a special emphasis on subset specification, leading to the formation of the 5-HT system in the brain. Molecular classification of 5-HT neuronal groups leads to the definition of two independent rostral groups positioned in rhombomere 1 and 2/3 and a caudal group in rhombomere 5-8. In addition, more disperse refinement of these subsets is present as shown by the selective expression of the 5-HT1A autoreceptor, indicating functional diversity between 5-HT subsets. The functional significance of the molecular coding differences is not well known and the molecular basis of described specific connectivity patterns remain to be elucidated. Recent developments in genetic lineage tracing models will provide these data and form a major step-up towards the full understanding of the importance of developmental programming and function of 5-HT neuronal subsets. 3. High quality, high efficiency welding technology for nuclear power plants International Nuclear Information System (INIS) Aoki, Shigeyuki; Nagura, Yasumi 1996-01-01 For nuclear power plants, it is required to ensure the safety under the high reliability and to attain the high rate of operation. In the manufacture and installation of the machinery and equipment, the welding techniques which become the basis exert large influence to them. For the purpose of improving joint performance and excluding human errors, welding heat input and the number of passes have been reduced, the automation of welding has been advanced, and at present, narrow gap arc welding and high energy density welding such as electron beam welding and laser welding have been put to practical use. Also in the welding of pipings, automatic gas metal arc welding is employed. As for the welding of main machinery and equipment, there are the welding of the joints that constitute pressure boundaries, the build-up welding on the internal surfaces of pressure vessels for separating primary water from them, and the sealing welding of heating tubes and tube plates in steam generators. These weldings are explained. The welding of pipings and the state of development and application of new welding methods are reported. (K.I.) 4. Concrete Waste Recycling Process for High Quality Aggregate International Nuclear Information System (INIS) Ishikura, Takeshi; Fujii, Shin-ichi 2008-01-01 Large amount of concrete waste generates during nuclear power plant (NPP) dismantling. Non-contaminated concrete waste is assumed to be disposed in a landfill site, but that will not be the solution especially in the future, because of decreasing tendency of the site availability and natural resources. Concerning concrete recycling, demand for roadbeds and backfill tends to be less than the amount of dismantled concrete generated in a single rural site, and conventional recycled aggregate is limited of its use to non-structural concrete, because of its inferior quality to ordinary natural aggregate. Therefore, it is vital to develop high quality recycled aggregate for general uses of dismantled concrete. If recycled aggregate is available for high structural concrete, the dismantling concrete is recyclable as aggregate for industry including nuclear field. Authors developed techniques on high quality aggregate reclamation for large amount of concrete generated during NPP decommissioning. Concrete of NPP buildings has good features for recycling aggregate; large quantity of high quality aggregate from same origin, record keeping of the aggregate origin, and little impurities in dismantled concrete such as wood and plastics. The target of recycled aggregate in this development is to meet the quality criteria for NPP concrete as prescribed in JASS 5N 'Specification for Nuclear Power Facility Reinforced Concrete' and JASS 5 'Specification for Reinforced Concrete Work'. The target of recycled aggregate concrete is to be comparable performance with ordinary aggregate concrete. The high quality recycled aggregate production techniques are assumed to apply for recycling for large amount of non-contaminated concrete. These techniques can also be applied for slightly contaminated concrete dismantled from radiological control area (RCA), together with free release survey. In conclusion: a technology on dismantled concrete recycling for high quality aggregate was developed 5. A Genetic Algorithm for Selection of Fixed-Size Subsets with Application to Design Problems Directory of Open Access Journals (Sweden) Mark A. Wolters 2015-11-01 Full Text Available The R function kofnGA conducts a genetic algorithm search for the best subset of k items from a set of n alternatives, given an objective function that measures the quality of a subset. The function fills a gap in the presently available subset selection software, which typically searches over a range of subset sizes, restricts the types of objective functions considered, or does not include freely available code. The new function is demonstrated on two types of problem where a fixed-size subset search is desirable: design of environmental monitoring networks, and D-optimal design of experiments. Additionally, the performance is evaluated on a class of constructed test problems with a novel design that is interesting in its own right. 6. High-quality planar high-Tc Josephson junctions International Nuclear Information System (INIS) Bergeal, N.; Grison, X.; Lesueur, J.; Faini, G.; Aprili, M.; Contour, J.P. 2005-01-01 Reproducible high-T c Josephson junctions have been made in a rather simple two-step process using ion irradiation. A microbridge (1 to 5 μm wide) is firstly designed by ion irradiating a c-axis-oriented YBa 2 Cu 3 O 7-δ film through a gold mask such as the nonprotected part becomes insulating. A lower T c part is then defined within the bridge by irradiating with a much lower fluence through a narrow slit (20 nm) opened in a standard electronic photoresist. These planar junctions, whose settings can be finely tuned, exhibit reproducible and nearly ideal Josephson characteristics. This process can be used to produce complex Josephson circuits 7. Key factors for a high-quality VR experience Science.gov (United States) Champel, Mary-Luc; Doré, Renaud; Mollet, Nicolas 2017-09-01 For many years, Virtual Reality has been presented as a promising technology that could deliver a truly new experience to users. The media and entertainment industry is now investigating the possibility to offer a video-based VR 360 experience. Nevertheless, there is a substantial risk that VR 360 could have the same fate as 3DTV if it cannot offer more than just being the next fad. The present paper aims at presenting the various quality factors required for a high-quality VR experience. More specifically, this paper will focus on the main three VR quality pillars: visual, audio and immersion. 8. Process to Continuously Melt, Refine and Cast High Quality Steel Energy Technology Data Exchange (ETDEWEB) None 2005-09-01 The purpose of this project is to conduct research and development targeted at designing a revolutionary steelmaking process. This process will deliver high quality steel from scrap to the casting mold in one continuous process and will be safer, more productive, and less capital intensive to build and operate than conventional steelmaking. The new process will produce higher quality steel faster than traditional batch processes while consuming less energy and other resources. 9. High-quality cardiopulmonary resuscitation: current and future directions. Science.gov (United States) Abella, Benjamin S 2016-06-01 Cardiopulmonary resuscitation (CPR) represents the cornerstone of cardiac arrest resuscitation care. Prompt delivery of high-quality CPR can dramatically improve survival outcomes; however, the definitions of optimal CPR have evolved over several decades. The present review will discuss the metrics of CPR delivery, and the evidence supporting the importance of CPR quality to improve clinical outcomes. The introduction of new technologies to quantify metrics of CPR delivery has yielded important insights into CPR quality. Investigations using CPR recording devices have allowed the assessment of specific CPR performance parameters and their relative importance regarding return of spontaneous circulation and survival to hospital discharge. Additional work has suggested new opportunities to measure physiologic markers during CPR and potentially tailor CPR delivery to patient requirements. Through recent laboratory and clinical investigations, a more evidence-based definition of high-quality CPR continues to emerge. Exciting opportunities now exist to study quantitative metrics of CPR and potentially guide resuscitation care in a goal-directed fashion. Concepts of high-quality CPR have also informed new approaches to training and quality improvement efforts for cardiac arrest care. 10. Selecting Optimal Subset of Security Controls OpenAIRE Yevseyeva, I.; Basto-Fernandes, V.; Michael, Emmerich, T. M.; Moorsel, van, A. 2015-01-01 Open Access journal Choosing an optimal investment in information security is an issue most companies face these days. Which security controls to buy to protect the IT system of a company in the best way? Selecting a subset of security controls among many available ones can be seen as a resource allocation problem that should take into account conflicting objectives and constraints of the problem. In particular, the security of the system should be improved without hindering productivity, ... 11. Next Generation High Quality Videoconferencing Service for the LHC CERN Multimedia CERN. Geneva 2012-01-01 In recent times, we have witnessed an explosion of video initiatives in the industry worldwide. Several advancements in video technology are currently improving the way we interact and collaborate. These advancements are forcing tendencies and overall experiences: any device in any network can be used to collaborate, in most cases with an overall high quality. To cope with this technology progresses, CERN IT Department has taken the leading role to establish strategies and directions to improve the user experience in remote dispersed meetings and remote collaboration at large in the worldwide LHC communities. Due to the high rate of dispersion in the LHC user communities, these are critically dependent of videoconferencing technology, with a need of robustness and high quality for the best possible user experience. We will present an analysis of the factors that influenced the technical and strategic choices to improve the reliability, efficiency and overall quality of the LHC remote sessions. In particular, ... 12. IMPACT OF RESILIENCE, ICT SUPPORT AND QUALITY OF STUDENT'S LIFE ON QUALITY OF HIGH EDUCATION PROCESS Directory of Open Access Journals (Sweden) Zorica Lazic 2016-12-01 Full Text Available Answers to the questions of how and in what way the quality of life of students, resilience and ICT support affects the quality of high education we will get through this work where main objective is to define a network of processes and process management ensuring more quality and more innovative managing and service provision, therefore satisfying the needs of service users - in this case the students of the university. To collect the relevant data in the thematic analysis of this paper, the method of interviewing by questionnaires was applied. The sample survey was conducted among undergraduate students, teachers and staff of the Teacher Training Faculty in Uzice. 13. Implementing Modifed Burg Algorithms in Multivariate Subset Autoregressive Modeling Directory of Open Access Journals (Sweden) 2003-02-01 Full Text Available The large number of parameters in subset vector autoregressive models often leads one to procure fast, simple, and efficient alternatives or precursors to maximum likelihood estimation. We present the solution of the multivariate subset Yule-Walker equations as one such alternative. In recent work, Brockwell, Dahlhaus, and Trindade (2002, show that the Yule-Walker estimators can actually be obtained as a special case of a general recursive Burg-type algorithm. We illustrate the structure of this Algorithm, and discuss its implementation in a high-level programming language. Applications of the Algorithm in univariate and bivariate modeling are showcased in examples. Univariate and bivariate versions of the Algorithm written in Fortran 90 are included in the appendix, and their use illustrated. 14. Design of a verifiable subset for HAL/S Science.gov (United States) Browne, J. C.; Good, D. I.; Tripathi, A. R.; Young, W. D. 1979-01-01 An attempt to evaluate the applicability of program verification techniques to the existing programming language, HAL/S is discussed. HAL/S is a general purpose high level language designed to accommodate the software needs of the NASA Space Shuttle project. A diversity of features for scientific computing, concurrent and real-time programming, and error handling are discussed. The criteria by which features were evaluated for inclusion into the verifiable subset are described. Individual features of HAL/S with respect to these criteria are examined and justification for the omission of various features from the subset is provided. Conclusions drawn from the research are presented along with recommendations made for the use of HAL/S with respect to the area of program verification. 15. Learning Disabilities and Achieving High-Quality Education Standards Science.gov (United States) Gartland, Debi; Strosnider, Roberta 2017-01-01 This is an official document of the National Joint Committee on Learning Disabilities (NJCLD), of which Council for Learning Disabilities is a long-standing, active member. With this position paper, NJCLD advocates for the implementation of high-quality education standards (HQES) for students with learning disabilities (LD) and outlines the… 16. extraction of high quality dna from polysaccharides-secreting ... African Journals Online (AJOL) cistvr A DNA extraction method using CTAB was used for the isolation of genomic DNA from ten. Xanthomonas campestris pathovars, ten isolates of Xanthomonas albilineans and one isolate of. Pseudomonas rubrisubalbicans. High quality DNA was obtained that was ideal for molecular analy- ses. Extracellular polysaccharides ... 17. Negative Binomial charts for monitoring high-quality processes NARCIS (Netherlands) Albers, Willem/Wim Good control charts for high quality processes are often based on the number of successes between failures. Geometric charts are simplest in this respect, but slow in recognizing moderately increased failure rates p. Improvement can be achieved by waiting until r > 1 failures have occurred, i.e. by 18. Synthesis and spectroscopic study of high quality alloy Cdx S ... Wintec In the present study, we report the synthesis of high quality CdxZn1–xS nanocrystals alloy at. 150°C with .... (XRD) using a Siemens model D 500, powder X-ray ... decays were analysed using IBH DAS6 software. 3. ... This alloying process is. 19. Methods and systems for fabricating high quality superconducting tapes Science.gov (United States) Majkic, Goran; Selvamanickam, Venkat 2018-02-13 An MOCVD system fabricates high quality superconductor tapes with variable thicknesses. The MOCVD system can include a gas flow chamber between two parallel channels in a housing. A substrate tape is heated and then passed through the MOCVD housing such that the gas flow is perpendicular to the tape's surface. Precursors are injected into the gas flow for deposition on the substrate tape. In this way, superconductor tapes can be fabricated with variable thicknesses, uniform precursor deposition, and high critical current densities. 20. Managing quality inside a high-technology project organization OpenAIRE Jokinen, T. (Tauno) 2004-01-01 Abstract This action research addresses the deployment of Total Quality Management (TQM) principles in a high-technology new product development organisation. During the period of study, the organisation grew fast. High-technology product development and hypergrowth provided a unique combination of extreme conditions for the study. The existing concepts of TQM are presented as an organised map enabling strategic analysis for an implementation plan. The history of TQM dates back to the ... 1. [Quality of sleep and academic performance in high school students]. Science.gov (United States) Bugueño, Maithe; Curihual, Carolina; Olivares, Paulina; Wallace, Josefa; López-AlegrÍa, Fanny; Rivera-López, Gonzalo; Oyanedel, Juan Carlos 2017-09-01 Sleeping and studying are the day-to-day activities of a teenager attending school. To determine the quality of sleep and its relationship to the academic performance among students attending morning and afternoon shifts in a public high school. Students of the first and second year of high school answered an interview about socio-demographic background, academic performance, student activities and subjective sleep quality; they were evaluated using the Pittsburgh Sleep Quality Index (PSQI). The interview was answered by 322 first year students aged 15 ± 5 years attending the morning shift and 364 second year students, aged 16 ± 0.5 years, attending the afternoon shift. The components: sleep latency, habitual sleep efficiency, sleep disturbance, drug use and daytime dysfunction were similar and classified as good in both school shifts. The components subjective sleep quality and duration of sleep had higher scores among students of the morning shift. The mean grades during the first semester of the students attending morning and afternoon shifts were 5.9 and 5.8, respectively (of a scale from 1 to 7). Among students of both shifts, the PSQI scale was associated inversely and significantly with academic performance. A bad sleep quality influences academic performance in these students. 2. Highly qualified does not equal high quality: A study of urban stakeholders' perceptions of quality in science teaching Science.gov (United States) Miranda, Rommel Joseph By employing qualitative methods, this study sought to determine the perceptions that urban stakeholders hold about what characteristics should distinguish a high school science teacher whom they would consider to demonstrate high quality in science teaching. A maximum variation sample of six science teachers, three school administrators, six parents and six students from a large urban public school district were interviewed using semi-structured, in-depth interview techniques. From these data, a list of observable characteristics which urban stakeholders hold as evidence of high quality in science teaching was generated. Observational techniques were utilized to determine the extent to which six urban high school science teachers, who meet the NCLB Act criteria for being "highly qualified", actually possessed the characteristics which these stakeholders hold as evidence of high quality in science teaching. Constant comparative analysis was used to analyze the data set. The findings suggest that urban stakeholders perceive that a high school science teacher who demonstrates high quality in science teaching should be knowledgeable about their subject matter, their student population, and should be resourceful; should possess an academic background in science and professional experience in science teaching; should exhibit professionalism, a passion for science and teaching, and a dedication to teaching and student learning; should be skillful in planning and preparing science lessons and in organizing the classroom, in presenting the subject matter to students, in conducting a variety of hands-on activities, and in managing a classroom; and should assess whether students complete class goals and objectives, and provide feedback about grades for students promptly. The findings further reveal that some of the urban high school science teachers who were deemed to be "highly qualified", as defined by the NCLB Act, engaged in practices that threatened quality in science 3. High quality digital holographic reconstruction on analog film Science.gov (United States) Nelsen, B.; Hartmann, P. 2017-05-01 High quality real-time digital holographic reconstruction, i.e. at 30 Hz frame rates, has been at the forefront of research and has been hailed as the holy grail of display systems. While these efforts have produced a fascinating array of computer algorithms and technology, many applications of reconstructing high quality digital holograms do not require such high frame rates. In fact, applications such as 3D holographic lithography even require a stationary mask. Typical devices used for digital hologram reconstruction are based on spatial-light-modulator technology and this technology is great for reconstructing arbitrary holograms on the fly; however, it lacks the high spatial resolution achievable by its analog counterpart, holographic film. Analog holographic film is therefore the method of choice for reconstructing highquality static holograms. The challenge lies in taking a static, high-quality digitally calculated hologram and effectively writing it to holographic film. We have developed a theoretical system based on a tunable phase plate, an intensity adjustable high-coherence laser and a slip-stick based piezo rotation stage to effectively produce a digitally calculated hologram on analog film. The configuration reproduces the individual components, both the amplitude and phase, of the hologram in the Fourier domain. These Fourier components are then individually written on the holographic film after interfering with a reference beam. The system is analogous to writing angularly multiplexed plane waves with individual component phase control. 4. Long quantum channels for high-quality entanglement transfer International Nuclear Information System (INIS) Banchi, L; Apollaro, T J G; Cuccoli, A; Verrucchi, P; Vaia, R 2011-01-01 High-quality quantum-state and entanglement transfer can be achieved in an unmodulated spin bus operating in the ballistic regime, which occurs when the endpoint qubits A and B are nonperturbatively coupled to the chain by a suitable exchange interaction j 0 . Indeed, the transition amplitude characterizing the transfer quality exhibits a maximum for a finite optimal value j opt 0 (N), where N is the channel length. We show that j opt 0 (N) scales as N -1/6 for large N and that it ensures a high-quality entanglement transfer even in the limit of arbitrarily long channels, almost independently of the channel initialization. For instance, for any chain length the average quantum-state transmission fidelity exceeds 90% and decreases very little in a broad neighbourhood of j opt 0 (N). We emphasize that, taking the reverse point of view, should j 0 be experimentally constrained, high-quality transfer can still be obtained by adjusting the channel length to its optimal value. (paper) 5. Integration study of high quality teaching resources in universities Directory of Open Access Journals (Sweden) Honglu Liu 2012-12-01 Full Text Available Purpose: The development level and quality of education depend on the merits and efficiency in the use of teaching resources, especially in the case of obvious contradiction between the demand and supply of teaching resources. So to integrate teaching resources, improve the efficiency in the use of high quality teaching resources, and take the road of content development to enhance the competitiveness of education has become very important and urgent.Design/methodology/approach: On the basis of analysis on the teaching resources of universities and the problems they faced, this paper introduced the basic concepts of cloud storage, and built the integration architecture of high quality teaching resources in universities based on the cloud storage.Findings and Originality/value: The HDFS-based cloud storage proposed in this paper is a dynamically adjustable and Internet-based storage solution, and the users can access storage targets using the network through a common and easy-to-use protocol and application programming interfaces. This new technology is useful for end users benefits. With the continuous development and improvement of cloud storage, it will necessarily result in more and more applications in the institutions of higher learning and education network.Originality/value: This paper introduced the cloud storage into the integration of high quality teaching resources in universities first and as a new form of service, it can be a good solution. 6. Examination of Operation Quality for High-frequent Railway Operation DEFF Research Database (Denmark) Landex, Alex; Kaas, Anders H. 2009-01-01 take the first train in their direction. The article examines four different approaches to examine operation quality for high-frequent operation that are based on the experiences of the passengers. These approaches are the service frequency of the operation, travel time extension, a combination......The examination of operation quality for high-frequent operation requires other approaches than the typical evaluation of punctuality (trains on time) and reliability (operated trains). This is because passengers in high-frequent railway systems do not necessarily notice train delays as they just...... of the service frequency and travel time approaches, and passenger delays. The service frequency and travel time approaches are simple measurements with low complexity and complement each other. Therefore, the article recommends combining the service frequency and travel time approaches to get a more accurate... 7. Quality and efficiency in high dimensional Nearest neighbor search KAUST Repository Tao, Yufei; Yi, Ke; Sheng, Cheng; Kalnis, Panos 2009-01-01 Nearest neighbor (NN) search in high dimensional space is an important problem in many applications. Ideally, a practical solution (i) should be implementable in a relational database, and (ii) its query cost should grow sub-linearly with the dataset size, regardless of the data and query distributions. Despite the bulk of NN literature, no solution fulfills both requirements, except locality sensitive hashing (LSH). The existing LSH implementations are either rigorous or adhoc. Rigorous-LSH ensures good quality of query results, but requires expensive space and query cost. Although adhoc-LSH is more efficient, it abandons quality control, i.e., the neighbor it outputs can be arbitrarily bad. As a result, currently no method is able to ensure both quality and efficiency simultaneously in practice. Motivated by this, we propose a new access method called the locality sensitive B-tree (LSB-tree) that enables fast highdimensional NN search with excellent quality. The combination of several LSB-trees leads to a structure called the LSB-forest that ensures the same result quality as rigorous-LSH, but reduces its space and query cost dramatically. The LSB-forest also outperforms adhoc-LSH, even though the latter has no quality guarantee. Besides its appealing theoretical properties, the LSB-tree itself also serves as an effective index that consumes linear space, and supports efficient updates. Our extensive experiments confirm that the LSB-tree is faster than (i) the state of the art of exact NN search by two orders of magnitude, and (ii) the best (linear-space) method of approximate retrieval by an order of magnitude, and at the same time, returns neighbors with much better quality. © 2009 ACM. 8. High quality flux control system for electron gun evaporation International Nuclear Information System (INIS) Appelbloom, A.M.; Hadley, P.; van der Marel, D.; Mooij, J.E. 1991-01-01 This paper reports on a high quality flux control system for electron gun evaporation developed and tested for the MBE growth of high temperature superconductors. The system can be applied to any electron gun without altering the electron gun itself. Essential elements of the system are a high bandwidth mass spectrometer, control electronics and a high voltage modulator to sweep the electron beam over the melt at high frequencies. the sweep amplitude of the electron beam is used to control the evaporation flux at high frequencies. The feedback loop of the system has a bandwidth of over 100 Hz, which makes it possible to grow superlattices and layered structures in a fast and precisely controlled manner 9. Safety and quality management at the high flux reactor Petten International Nuclear Information System (INIS) Zurita, A.; Ahlf, J. 1995-01-01 The High Flux Reactor (HFR) is one high power multi-purpose materials testing research reactor of the tank-in-pool type, cooled and moderated by light-water. It is operated at 45 MW at a prescribed schedule of 11 cycles per year, each comprising 25 operation days and three shut-down days. Since the licence for the operation of HFR was granted in 1962, a total of 14 amendments to the original licence have been made following different modifications in the installations. In the meantime, international nuclear standards were developed, especially in the framework of the NUSS programme of the IAEA, which were adopted by the Dutch Licensing Authorities. In order to implement the new standards, the situation at the HFR was comprehensively reviewed in the course of an audit performed by the Dutch Licensing Authorities in 1988. This also resulted in formulating the task of setting-up an 'HFR - Integral Quality Assurance Handbook' (HFR-IQAD) involving both organizations JRCIAM and ECN, which had the unique framework and basic guideline to assure the safe and efficient operation and exploitation of the HFR and to promote safety and quality in all aspects of HFR related activities. The assurance of safe and efficient operation and exploitation of the HFR is condensed together under the concepts of safety and quality of services and is achieved through the safety and quality management. (orig.) 10. Wellbeing Understanding in High Quality Healthcare Informatics and Telepractice. Science.gov (United States) Fiorini, Rodolfo A; De Giacomo, Piero; L'Abate, Luciano 2016-01-01 The proper use of healthcare informatics technology and multidimensional conceptual clarity are fundamental to create and boost outstanding clinical and telepractice results. Avoiding even terminology ambiguities is mandatory for high quality of care service. For instance, well-being or wellbeing is a different way to write the same concept only, or there is a good deal of ambiguity around the meanings of these terms the way they are written. In personal health, healthcare and healthcare informatics, this kind of ambiguity and lack of conceptual clarity has been called out repeatedly over the past 50 years. It is time to get the right, terse scenario. We present a brief review to develop and achieve ultimate wellbeing understanding for practical high quality healthcare informatics and telepractice application. This article presents an innovative point of view on deeper wellbeing understanding towards its increased clinical effective application. 11. Lymphocytic subsets and low-dose exposure International Nuclear Information System (INIS) Tuschl, H.; Kovac, R.; Eybl, E. 1993-03-01 The present investigations proved the differential radiosensitivity of lymphocytic subpopulations: From in vivo and in vitro irradiations it may be followed that the most sensitive subset are CD8 positive suppressor T cells. CD4/CD8 ratios are increased both in peripheral blood and after mitogen stimulation of lymphocytes of exposed persons. The decrease in B cells is pronounced only at higher radiation doses. Though the rate of DNA synthesis after mitogen stimulation was reduced in some exposed persons, that was no general phenomenon. Especially after tritium exposure, the observed lymphopenia correlated with an increased stimulation by PHA and an increased rate of DNA synthesis in some probands. Thus the present investigations indicate that - despite an inhibition of some immune parameters by radioexposure - the body is able to maintain its immunological homoeostasis. (authors) 12. TCR tuning of T cell subsets. Science.gov (United States) Cho, Jae-Ho; Sprent, Jonathan 2018-05-01 After selection in the thymus, the post-thymic T cell compartments comprise heterogenous subsets of naive and memory T cells that make continuous T cell receptor (TCR) contact with self-ligands bound to major histocompatibility complex (MHC) molecules. T cell recognition of self-MHC ligands elicits covert TCR signaling and is particularly important for controlling survival of naive T cells. Such tonic TCR signaling is tightly controlled and maintains the cells in a quiescent state to avoid autoimmunity. Here, we review how naive and memory T cells are differentially tuned and wired for TCR sensitivity to self and foreign ligands. © 2018 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd. 13. Maximum parsimony on subsets of taxa. Science.gov (United States) Fischer, Mareike; Thatte, Bhalchandra D 2009-09-21 In this paper we investigate mathematical questions concerning the reliability (reconstruction accuracy) of Fitch's maximum parsimony algorithm for reconstructing the ancestral state given a phylogenetic tree and a character. In particular, we consider the question whether the maximum parsimony method applied to a subset of taxa can reconstruct the ancestral state of the root more accurately than when applied to all taxa, and we give an example showing that this indeed is possible. A surprising feature of our example is that ignoring a taxon closer to the root improves the reliability of the method. On the other hand, in the case of the two-state symmetric substitution model, we answer affirmatively a conjecture of Li, Steel and Zhang which states that under a molecular clock the probability that the state at a single taxon is a correct guess of the ancestral state is a lower bound on the reconstruction accuracy of Fitch's method applied to all taxa. 14. Quality assurance system for sitting high risk facilities International Nuclear Information System (INIS) Rodriguez, Aymee; Peralta, Jose L.; Fernandez, Manuel 1999-01-01 The paper shows how we have conceived and designed the quality assurance system for the site selection process of an area for sitting the facility of high risk in correspondence with the approved methodology. The results obtained in the implementation of the system have permitted the satisfactory performance of each one the expected stage, defining the most favorable sectors in order to continue the studies of the repository site for the disposal of low and intermedium. (author) 15. High-quality uniform dry transfer of graphene to polymers. Science.gov (United States) Lock, Evgeniya H; Baraket, Mira; Laskoski, Matthew; Mulvaney, Shawn P; Lee, Woo K; Sheehan, Paul E; Hines, Daniel R; Robinson, Jeremy T; Tosado, Jacob; Fuhrer, Michael S; Hernández, Sandra C; Walton, Scott G 2012-01-11 In this paper we demonstrate high-quality, uniform dry transfer of graphene grown by chemical vapor deposition on copper foil to polystyrene. The dry transfer exploits an azide linker molecule to establish a covalent bond to graphene and to generate greater graphene-polymer adhesion compared to that of the graphene-metal foil. Thus, this transfer approach provides a novel alternative route for graphene transfer, which allows for the metal foils to be reused. © 2011 American Chemical Society 16. Coupling of high-quality-factor optical resonators International Nuclear Information System (INIS) Salzenstein, Patrice; Henriet, Rémi; Coillet, Aurélien; Chembo, Yanne K; Mortier, Michel; Sérier-Brault, Hélène; Rasoloniaina, Alphonse; Dumeige, Yannick; Féron, Patrice 2013-01-01 We improve theoretically and experimentally the problem of the coupling between a high Q-factor resonator and its external coupler. We have observed oscillations of ringing induced by the sweeping of the excitation frequency of an active microsphere. Thanks to this approach, the quality factor of an optical resonator was measured and we obtained Q = 5.8 × 10 8 . (paper) 17. A high throughput DNA extraction method with high yield and quality Directory of Open Access Journals (Sweden) Xin Zhanguo 2012-07-01 Full Text Available Abstract Background Preparation of large quantity and high quality genomic DNA from a large number of plant samples is a major bottleneck for most genetic and genomic analyses, such as, genetic mapping, TILLING (Targeting Induced Local Lesion IN Genome, and next-generation sequencing directly from sheared genomic DNA. A variety of DNA preparation methods and commercial kits are available. However, they are either low throughput, low yield, or costly. Here, we describe a method for high throughput genomic DNA isolation from sorghum [Sorghum bicolor (L. Moench] leaves and dry seeds with high yield, high quality, and affordable cost. Results We developed a high throughput DNA isolation method by combining a high yield CTAB extraction method with an improved cleanup procedure based on MagAttract kit. The method yielded large quantity and high quality DNA from both lyophilized sorghum leaves and dry seeds. The DNA yield was improved by nearly 30 fold with 4 times less consumption of MagAttract beads. The method can also be used in other plant species, including cotton leaves and pine needles. Conclusion A high throughput system for DNA extraction from sorghum leaves and seeds was developed and validated. The main advantages of the method are low cost, high yield, high quality, and high throughput. One person can process two 96-well plates in a working day at a cost of $0.10 per sample of magnetic beads plus other consumables that other methods will also need. 18. Methods and systems for fabricating high quality superconducting tapes Energy Technology Data Exchange (ETDEWEB) Majkic, Goran; Selvamanickam, Venkat 2018-02-13 An MOCVD system fabricates high quality superconductor tapes with variable thicknesses. The MOCVD system can include a gas flow chamber between two parallel channels in a housing. A substrate tape is heated and then passed through the MOCVD housing such that the gas flow is perpendicular to the tape's surface. Precursors are injected into the gas flow for deposition on the substrate tape. In this way, superconductor tapes can be fabricated with variable thicknesses, uniform precursor deposition, and high critical current densities. 19. High quality mask storage in an advanced Logic-Fab Science.gov (United States) Jähnert, Carmen; Fritsche, Silvio 2012-02-01 High efficient mask logistics as well as safe and high quality mask storage are essential requirements within an advanced lithography area of a modern logic waferfab. Fast operational availability of the required masks at the exposure tool with excellent mask condition requires a safe mask handling, safeguarding of high mask quality over the whole mask usage time without any quality degradation and an intelligent mask logistics. One big challenge is the prevention of haze on high advanced phase shift masks used in a high volume production line for some thousands of 248nm or 193nm exposures. In 2008 Infineon Dresden qualified a customer specific developed semi-bare mask storage system from DMSDynamic Micro Systems in combination with a high advanced mask handling and an interconnected complex logistic system. This high-capacity mask storage system DMS M1900.22 for more than 3000 masks with fully automated mask and box handling as well as full-blown XCDA purge has been developed and adapted to the Infineon Lithotoollandscape using Nikon and SMIF reticle cases. Advanced features for ESD safety and mask security, mask tracking via RFID and interactions with the exposure tools were developed and implemented. The stocker is remote controlled by the iCADA-RSM system, ordering of the requested mask directly from the affected exposure tool allows fast access. This paper discusses the advantages and challenges for this approach as well as the practical experience gained during the implementation of the new system which improves the fab performance with respect to mask quality, security and throughput. Especially the realization of an extremely low and stable humidity level in addition with a well controlled air flow at each mask surface, preventing masks from haze degradation and particle contamination, turns out to be a notable technical achievement. The longterm stability of haze critical masks has been improved significantly. Relevant environmental parameters like 20. Construction of High-Quality Camel Immune Antibody Libraries. Science.gov (United States) Romão, Ema; Poignavent, Vianney; Vincke, Cécile; Ritzenthaler, Christophe; Muyldermans, Serge; Monsion, Baptiste 2018-01-01 Single-domain antibodies libraries of heavy-chain only immunoglobulins from camelids or shark are enriched for high-affinity antigen-specific binders by a short in vivo immunization. Thus, potent binders are readily retrieved from relatively small-sized libraries of 10 7 -10 8 individual transformants, mostly after phage display and panning on a purified target. However, the remaining drawback of this strategy arises from the need to generate a dedicated library, for nearly every envisaged target. Therefore, all the procedures that shorten and facilitate the construction of an immune library of best possible quality are definitely a step forward. In this chapter, we provide the protocol to generate a high-quality immune VHH library using the Golden Gate Cloning strategy employing an adapted phage display vector where a lethal ccdB gene has to be substituted by the VHH gene. With this procedure, the construction of the library can be shortened to less than a week starting from bleeding the animal. Our libraries exceed 10 8 individual transformants and close to 100% of the clones harbor a phage display vector having an insert with the length of a VHH gene. These libraries are also more economic to make than previous standard approaches using classical restriction enzymes and ligations. The quality of the Nanobodies that are retrieved from immune libraries obtained by Golden Gate Cloning is identical to those from immune libraries made according to the classical procedure. 1. High beam quality and high energy short-pulse laser with MOPA Science.gov (United States) Jin, Quanwei; Pang, Yu; Jiang, JianFeng; Tan, Liang; Cui, Lingling; Wei, Bin; Sun, Yinhong; Tang, Chun 2018-03-01 A high energy, high beam quality short-pulse diode-pumped Nd:YAG master oscillator power-amplifier (MOPA) laser with two amplifier stages is demonstrated. The two-rod birefringence compensation was used as beam quality controlling methods, which presents a short-pulse energy of 40 mJ with a beam quality value of M2 = 1.2 at a repetition rate of 400Hz. The MOPA system delivers a short-pulse energy of 712.5 mJ with a pulse width of 12.4 ns.The method of spherical aberration compensation is improved the beam quality, a M2 factor of 2.3 and an optical-to-optical efficiency of 27.7% is obtained at the maximum laser out power.The laser obtained 1.4J out energy with polarization integration. 2. Cytokine profile and lymphocyte subsets in type 2 diabetes Directory of Open Access Journals (Sweden) C.O. Francisco 2016-01-01 Full Text Available Type 2 diabetes mellitus (T2D is a metabolic disease with inflammation as an important pathogenic background. However, the pattern of immune cell subsets and the cytokine profile associated with development of T2D are unclear. The objective of this study was to evaluate different components of the immune system in T2D patients' peripheral blood by quantifying the frequency of lymphocyte subsets and intracellular pro- and anti-inflammatory cytokine production by T cells. Clinical data and blood samples were collected from 22 men (51.6±6.3 years old with T2D and 20 nonsmoking men (49.4±7.6 years old who were matched for age and sex as control subjects. Glycated hemoglobin, high-sensitivity C-reactive protein concentrations, and the lipid profile were measured by a commercially available automated system. Frequencies of lymphocyte subsets in peripheral blood and intracellular production of interleukin (IL-4, IL-10, IL-17, tumor necrosis factor-α, and interferon-γ cytokines by CD3+ T cells were assessed by flow cytometry. No differences were observed in the frequency of CD19+ B cells, CD3+CD8+ and CD3+CD4+ T cells, CD16+56+ NK cells, and CD4+CD25+Foxp3+ T regulatory cells in patients with T2D compared with controls. The numbers of IL-10- and IL-17-producing CD3+ T cells were significantly higher in patients with T2D than in controls (P<0.05. The frequency of interferon-γ-producing CD3+ T cells was positively correlated with body mass index (r=0.59; P=0.01. In conclusion, this study shows increased numbers of circulating IL-10- and IL-17-producing CD3+ T cells in patients with T2D, suggesting that these cytokines are involved in the immune pathology of this disease. 3. Development of nuclear quality high pressure valve bellows in Canada International Nuclear Information System (INIS) Janzen, P.; Astill, C.J. 1978-06-01 Concurrent with the decision to use bellows stem sealed nuclear valves where feasible in commercial-scale CANDU plants, AECL undertook to develop an indigenous high pressure valve bellows technology. This program included developing the capability to fabricate improved high pressure valve bellows in conjunction with a Canadian manufacturer. This paper describes the evolution of a two-stage bellows fabrication process involving: (1) manufacture of discrete lengths of precision thin wall telescoping tubes - from preparation of strip blanks through edge grinding and edge forming to longitudinal welding; (2) forming of bellows from tube assemblies using a novel combination of mechanical inward forming followed by hydraulic outward forming. Bellows of Inconel 600 and Inconel 625 have been manufactured and evaluated. Test results indicate comparable to improved performance over alternative high quality bellows. (author) 4. Fat Quality Influences the Obesogenic Effect of High Fat Diets Directory of Open Access Journals (Sweden) Raffaella Crescenzo 2015-11-01 Full Text Available High fat and/or carbohydrate intake are associated with an elevated risk for obesity and chronic diseases such as diabetes and cardiovascular diseases. The harmful effects of a high fat diet could be different, depending on dietary fat quality. In fact, high fat diets rich in unsaturated fatty acids are considered less deleterious for human health than those rich in saturated fat. In our previous studies, we have shown that rats fed a high fat diet developed obesity and exhibited a decrease in oxidative capacity and an increase in oxidative stress in liver mitochondria. To investigate whether polyunsaturated fats could attenuate the above deleterious effects of high fat diets, energy balance and body composition were assessed after two weeks in rats fed isocaloric amounts of a high-fat diet (58.2% by energy rich either in lard or safflower/linseed oil. Hepatic functionality, plasma parameters, and oxidative status were also measured. The results show that feeding on safflower/linseed oil diet attenuates the obesogenic effect of high fat diets and ameliorates the blood lipid profile. Conversely, hepatic steatosis and mitochondrial oxidative stress appear to be negatively affected by a diet rich in unsaturated fatty acids. 5. Isolation of Human Skin Dendritic Cell Subsets. Science.gov (United States) Gunawan, Merry; Jardine, Laura; Haniffa, Muzlifah 2016-01-01 Dendritic cells (DCs) are specialized leukocytes with antigen-processing and antigen-presenting functions. DCs can be divided into distinct subsets by anatomical location, phenotype and function. In human, the two most accessible tissues to study leukocytes are peripheral blood and skin. DCs are rare in human peripheral blood (skin covering an average total surface area of 1.8 m(2) has approximately tenfold more DCs than the average 5 L of total blood volume (Wang et al., J Invest Dermatol 134:965-974, 2014). DCs migrate spontaneously from skin explants cultured ex vivo, which provide an easy method of cell isolation (Larsen et al., J Exp Med 172:1483-1493, 1990; Lenz et al., J Clin Invest 92:2587-2596, 1993; Nestle et al., J Immunol 151:6535-6545, 1993). These factors led to the extensive use of skin DCs as the "prototype" migratory DCs in human studies. In this chapter, we detail the protocols to isolate DCs and resident macrophages from human skin. We also provide a multiparameter flow cytometry gating strategy to identify human skin DCs and to distinguish them from macrophages. 6. High-volume image quality assessment systems: tuning performance with an interactive data visualization tool Science.gov (United States) Bresnahan, Patricia A.; Pukinskis, Madeleine; Wiggins, Michael 1999-03-01 Image quality assessment systems differ greatly with respect to the number and types of mags they need to evaluate, and their overall architectures. Managers of these systems, however, all need to be able to tune and evaluate system performance, requirements often overlooked or under-designed during project planning. Performance tuning tools allow users to define acceptable quality standards for image features and attributes by adjusting parameter settings. Performance analysis tools allow users to evaluate and/or predict how well a system performs in a given parameter state. While image assessment algorithms are becoming quite sophisticated, duplicating or surpassing the human decision making process in their speed and reliability, they often require a greater investment in 'training' or fine tuning of parameters in order to achieve optimum performance. This process may involve the analysis of hundreds or thousands of images, generating a large database of files and statistics that can be difficult to sort through and interpret. Compounding the difficulty is the fact that personnel charged with tuning and maintaining the production system may not have the statistical or analytical background required for the task. Meanwhile, hardware innovations have greatly increased the volume of images that can be handled in a given time frame, magnifying the consequences of running a production site with an inadequately tuned system. In this paper, some general requirements for a performance evaluation and tuning data visualization system are discussed. A custom engineered solution to the tuning and evaluation problem is then presented, developed within the context of a high volume image quality assessment, data entry, OCR, and image archival system. A key factor influencing the design of the system was the context-dependent definition of image quality, as perceived by a human interpreter. This led to the development of a five-level, hierarchical approach to image quality 7. High quality electron beams from a laser wakefield accelerator Energy Technology Data Exchange (ETDEWEB) Wiggins, S M; Issac, R C; Welsh, G H; Brunetti, E; Shanks, R P; Anania, M P; Cipiccia, S; Manahan, G G; Aniculaesei, C; Ersfeld, B; Islam, M R; Burgess, R T L; Vieux, G; Jaroszynski, D A [SUPA, Department of Physics, University of Strathclyde, Glasgow (United Kingdom); Gillespie, W A [SUPA, Division of Electronic Engineering and Physics, University of Dundee, Dundee (United Kingdom); MacLeod, A M [School of Computing and Creative Technologies, University of Abertay Dundee, Dundee (United Kingdom); Van der Geer, S B; De Loos, M J, E-mail: [email protected] [Pulsar Physics, Burghstraat 47, 5614 BC Eindhoven (Netherlands) 2010-12-15 High quality electron beams have been produced in a laser-plasma accelerator driven by femtosecond laser pulses with a peak power of 26 TW. Electrons are produced with an energy up to 150 MeV from the 2 mm gas jet accelerator and the measured rms relative energy spread is less than 1%. Shot-to-shot stability in the central energy is 3%. Pepper-pot measurements have shown that the normalized transverse emittance is {approx}1{pi} mm mrad while the beam charge is in the range 2-10 pC. The generation of high quality electron beams is understood from simulations accounting for beam loading of the wakefield accelerating structure. Experiments and self-consistent simulations indicate that the beam peak current is several kiloamperes. Efficient transportation of the beam through an undulator is simulated and progress is being made towards the realization of a compact, high peak brilliance free-electron laser operating in the vacuum ultraviolet and soft x-ray wavelength ranges. 8. Present status of high quality beam facility at Waseda University International Nuclear Information System (INIS) Washio, M.; Kawai, H.; Hama, Y.; Kudo, N.; Kobayashi, M.; Kuribayasi, T.; Kawaguchi, M.; Kuroda, R.; Maeda, K.; Nagasawa, F.; Ueyama, D.; Hizume, K.; Wang, X.J.; Hayano, H.; Urakawa, J.; Kashiwagi, S. 2004-01-01 A research project named High-Tech Research Center Project has been conducted at Waseda University. In this project, an RF gun system has been used for production of low emittance and short bunched electron beam. The experiments for the electron beam quality measurement have been carried out by slit scan techniques, etc. Short pulsed x-ray with the energy range of so-called water window has been generation by the inverse compton scattering. Further, the pulse radiolysis system has been constructed, and the stroboscopic pulse radiolysis has been applied for the detection of hydrated electron in picosecond time region. (author) 9. Anti-Stokes Luminescence in High Quality Quantum Wells Science.gov (United States) Vinattieri, A.; Bogani, F.; Miotto, A.; Ceccherini, S. 1997-11-01 We present a detailed investigation of the anti-Stokes (AS) luminescence which originates from exciton recombination when below gap excitation is used, in a set of high quality quantum well structures. We observe strong excitonic resonances in the AS signal as measured from photoluminescence and photoluminescence excitation spectra. We demonstrate that neither the electromagnetic coupling between the wells nor the morphological disorder can explain this up-conversion effect. Time-resolved luminescence data after ps excitation and fs correlation spectroscopy results provide clear evidence of the occurrence of a two-step absorption which is assisted by the exciton population resonantly excited by the first photon. 10. The quality of high-energy X-ray beams International Nuclear Information System (INIS) LaRiviere, P.D. 1989-01-01 Supplement 17 of the British Journal of Radiology is a survey of central-axis depth doses for radiotherapy machines, patterned largely on BJR Supplement 11 (1972). Inspection of high-energy X-ray depth doses for a 10 x 10 cm field at an SSD of 100 cm disclosed large differences between the two sets of data, especially for qualities above 8 MV, e.g. a depth dose of 80% at 10 cm is rated at about 19 MV according to BJR Supplement 11, and 23 MV according to BJR Supplement 17. It was found that Supplement 17 depth-dose data above 8 MV were erratic, but Supplement 11 data could be represented by an analytical expression, providing a unique means of assigning MV quality. It was also found that dose-weighted average energy of the filtered beam plotted smoothly against depth dose. For dosimetric purposes, it is suggested that this parameter be used as a true measure of beam quality, removing discrepancies introduced by the use of nominal MV for this purpose. (author) 11. CCD Astrophotography High-Quality Imaging from the Suburbs CERN Document Server Stuart, Adam 2006-01-01 This is a reference book for amateur astronomers who have become interested in CCD imaging. Those glorious astronomical images found in astronomy magazines might seem out of reach to newcomers to CCD imaging, but this is not the case. Great pictures are attainable with modest equipment. Adam Stuart’s many beautiful images, reproduced in this book, attest to the quality of – initially – a beginner’s efforts. Chilled-chip astronomical CCD-cameras and software are also wonderful tools for cutting through seemingly impenetrable light-pollution. CCD Astrophotography from the Suburbs describes one man’s successful approach to the problem of getting high-quality astronomical images under some of the most light-polluted conditions. Here is a complete and thoroughly tested program that will help every CCD-beginner to work towards digital imaging of the highest quality. It is equally useful to astronomers who have perfect observing conditions, as to those who have to observe from light-polluted city skies. 12. SOCP relaxation bounds for the optimal subset selection problem applied to robust linear regression OpenAIRE Flores, Salvador 2015-01-01 This paper deals with the problem of finding the globally optimal subset of h elements from a larger set of n elements in d space dimensions so as to minimize a quadratic criterion, with an special emphasis on applications to computing the Least Trimmed Squares Estimator (LTSE) for robust regression. The computation of the LTSE is a challenging subset selection problem involving a nonlinear program with continuous and binary variables, linked in a highly nonlinear fashion. The selection of a ... 13. Percutaneous vertebroplasty with a high-quality rotational angiographic unit Energy Technology Data Exchange (ETDEWEB) Pedicelli, Alessandro [Department of Bioimaging and Radiological Sciences, Catholic University of Sacred Heart, Policl. A.Gemelli, l.go Gemelli 1, 00168 Rome (Italy)], E-mail: [email protected]; Rollo, Massimo [Department of Bioimaging and Radiological Sciences, Catholic University of Sacred Heart, Policl. A.Gemelli, l.go Gemelli 1, 00168 Rome (Italy)], E-mail: [email protected]; Piano, Mariangela [Department of Bioimaging and Radiological Sciences, Catholic University of Sacred Heart, Policl. A.Gemelli, l.go Gemelli 1, 00168 Rome (Italy)], E-mail: [email protected]; Re, Thomas J. [Department of Bioimaging and Radiological Sciences, Catholic University of Sacred Heart, Policl. A.Gemelli, l.go Gemelli 1, 00168 Rome (Italy)], E-mail: [email protected]; Cipriani, Maria C. [Department of Gerontology, Catholic University of Sacred Heart, Policl. A.Gemelli, l.go Gemelli 1, 00168 Rome (Italy)], E-mail: [email protected]; Colosimo, Cesare [Department of Bioimaging and Radiological Sciences, Catholic University of Sacred Heart, Policl. A.Gemelli, l.go Gemelli 1, 00168 Rome (Italy)], E-mail: [email protected]; Bonomo, Lorenzo [Department of Bioimaging and Radiological Sciences, Catholic University of Sacred Heart, Policl. A.Gemelli, l.go Gemelli 1, 00168 Rome (Italy)], E-mail: [email protected] 2009-02-15 We evaluated the reliability of a rotational angiographic unit (RA) with flat-panel detector as a single technique to guide percutaneous vertebroplasty (PVP) and for post-procedure assessment by 2D and 3D reformatted images. Fifty-five consecutive patients (104 vertebral bodies) were treated under RA fluoroscopy. Rotational acquisitions with 2D and 3D reconstruction were obtained in all patients for immediate post-procedure assessment. In complex cases, this technique was also used to evaluate the needle position during the procedure. All patients underwent CT scan after the procedure. RA and CT findings were compared. In all cases, a safe trans-pedicular access and an accurate control of the bone-cement injection were successfully performed with high-quality fluoroscopy, even at the thoracic levels and in case of vertebra plana. 2D and 3D rotational reconstructions permitted CT-like images that clearly showed needle position and were similar to CT findings in depicting intrasomatic implant-distribution. RA detected 40 cement leakages compared to 42 demonstrated by CT and showed overall 95% sensitivity and 100% specificity compared to CT for final post-procedure assessment. Our preliminary results suggest that high-quality RA is reliable and safe as a single technique for PVP guidance, control and post-procedure assessment. It permits fast and cost-effective procedures avoiding multi-modality imaging. 14. Computer-aided control of high-quality cast iron Directory of Open Access Journals (Sweden) S. Pietrowski 2008-04-01 Full Text Available The study discusses the possibility of control of the high-quality grey cast iron and ductile iron using the author’s genuine computer programs. The programs have been developed with the help of algorithms based on statistical relationships that are said to exist between the characteristic parameters of DTA curves and properties, like Rp0,2, Rm, A5 and HB. It has been proved that the spheroidisation and inoculation treatment of cast iron changes in an important way the characteristic parameters of DTA curves, thus enabling a control of these operations as regards their correctness and effectiveness, along with the related changes in microstructure and mechanical properties of cast iron. Moreover, some examples of statistical relationships existing between the typical properties of ductile iron and its control process were given for cases of the melts consistent and inconsistent with the adopted technology.A test stand for control of the high-quality cast iron and respective melts has been schematically depicted. 15. Automated high speed volume computed tomography for inline quality control International Nuclear Information System (INIS) Hanke, R.; Kugel, A.; Troup, P. 2004-01-01 Increasing complexity of innovative products as well as growing requirements on quality and reliability call for more detailed knowledge about internal structures of manufactured components rather by 100 % inspection than just by sampling test. A first-step solution, like radioscopic inline inspection machines, equipped with automated data evaluation software, have become state of the art in the production floor during the last years. However, these machines provide just ordinary two-dimensional information and deliver no volume data e.g. to evaluate exact position or shape of detected defects. One way to solve this problem is the application of X-ray computed tomography (CT). Compared to the performance of the first generation medical scanners (scanning times of many hours), today, modern Volume CT machines for industrial applications need about 5 minutes for a full object scan depending on the object size. Of course, this is still too long to introduce this powerful method into the inline production quality control. In order to gain acceptance, the scanning time including subsequent data evaluation must be decreased significantly and adapted to the manufacturing cycle times. This presentation demonstrates the new technical set up, reconstruction results and the methods for high-speed volume data evaluation of a new fully automated high-speed CT scanner with cycle times below one minute for an object size of less than 15 cm. This will directly create new opportunities in design and construction of more complex objects. (author) 16. Production of high quality water for oil sands application Energy Technology Data Exchange (ETDEWEB) Beaudette-Hodsman, C.; Macleod, B. [Pall Corp., Mississauga, ON (Canada); Venkatadri, R. [Pall Corp., East Hills, NY (United States) 2008-10-15 This paper described a pressurized microfiltration membrane system installed at an oil sands extraction site in Alberta. The system was designed to complement a reverse osmosis (RO) system installed at the site to produce the high quality feed water required by the system's boilers. Groundwater in the region exhibited moderate total suspended solids and high alkalinity and hardness levels, and the RO system required feed water with a silt density index of 3 or less. The conventional pretreatment system used at the site was slowing down production due to the severe fouling of the RO membranes. The new microfiltration system contained an automated PVDF hollow fiber microfiltration membrane system contained in a trailer. Suspended particles and bacteria were captured within the filter, and permeate was sent to the RO unit. Within 6 hours of being installed, the unit was producing water with SDI values in the range of 1.0 to 2.5. It was concluded that the microfiltration system performed reliably regardless of wide variations in feed water quality and flow rates. 3 refs., 1 tab., 8 figs. 17. Endothelial progenitor cell subsets and preeclampsia: Findings and controversies Directory of Open Access Journals (Sweden) Armin Attar 2017-10-01 Full Text Available Vascular remodeling is an essential component of gestation. Endothelial progenitor cells (EPCs play an important role in the regulation of vascular homeostasis. The results of studies measuring the number of EPCs in normal pregnancies and in preeclampsia have been highly controversial or even contradictory because of some variations in technical issues and different methodologies enumerating three distinct subsets of EPCs: circulating angiogenic cells (CAC, colony forming unit endothelial cells (CFU-ECs, and endothelial colony-forming cells (ECFCs. In general, most studies have shown an increase in the number of CACs in the maternal circulation with a progression in the gestational age in normal pregnancies, while functional capacities measured by CFU-ECs and ECFCs remain intact. In the case of preeclampsia, mobilization of CACs and ECFCs occurs in the peripheral blood of pregnant women, but the functional capacities shown by culture of the derived colony-forming assays (CFU-EC and ECFC assays are altered. Furthermore, the number of all EPC subsets will be reduced in umbilical cord blood in the case of preeclampsia. As EPCs play an important role in the homeostasis of vascular networks, the difference in their frequency and functionality in normal pregnancies and those with preeclampsia can be expected. In this review, there was an attempt to provide a justification for these controversies. 18. Monocyte Subsets in Schistosomiasis Patients with Periportal Fibrosis Directory of Open Access Journals (Sweden) Jamille Souza Fernandes 2014-01-01 Full Text Available A major issue with Schistosoma mansoni infection is the development of periportal fibrosis, which is predominantly caused by the host immune response to egg antigens. Experimental studies have pointed to the participation of monocytes in the pathogenesis of liver fibrosis. The aim of this study was to characterize the subsets of monocytes in individuals with different degrees of periportal fibrosis secondary to schistosomiasis. Monocytes were classified into classical (CD14++CD16−, intermediate (CD14++CD16+, and nonclassical (CD14+CD16++. The expressions of monocyte markers and cytokines were assessed using flow cytometry. The frequency of classical monocytes was higher than the other subsets. The expression of HLA-DR, IL-6, TNF-α, and TGF-β was higher in monocytes from individuals with moderate to severe fibrosis as compared to other groups. Although no differences were observed in receptors expression (IL-4R and IL-10R between groups of patients, the expression of IL-12 was lower in monocytes from individuals with moderate to severe fibrosis, suggesting a protective role of this cytokine in the development of fibrosis. Our data support the hypothesis that the three different monocyte populations participate in the immunopathogenesis of periportal fibrosis, since they express high levels of proinflammatory and profibrotic cytokines and low levels of regulatory markers. 19. Peripheral blood and milk leukocytes subsets of lactating Sarda ewes Directory of Open Access Journals (Sweden) Piero Bonelli 2013-05-01 Full Text Available Leukocytes subpopulations in blood and milk of lactating Sarda ewes were investigated. Animals characterized by a SSC level <500×103cells/mL and a negative bacteriological examination were sampled in early, mid and late lactation. Milk differential cell count evidenced that macrophage represented the main population (42.8%±3.5 followed by lymphocytes (40.2%±3.4 and neutrophils (8,6%±2.1. Flow cytometry analysis showed that lymphocytes subsets in milk were quite different from blood. High CD8+ and low CD4+ lymphocytes percentages determined a CD4/CD8 ratio inversion in milk compared to blood (0.3%±0.03 vs 1.8%±0.08. CD8+ decreased while, conversely, CD4+ increased in late lactation. γδ T cells were more represented in milk (12.6%±1.3 than in blood (6.8%±0.3 and their proportions appeared similar throughout lactation in both compartments. IL-2 receptor was mainly expressed in milk on T cytotoxic lymphocytes. Data obtained in uninfected mammary glands could allow an early discrimination between physiological and pathological changes occurring in ewe milk. Further phenotypical and functional studies on milk leukocytes subsets might help to understand defense mechanisms of the ovine mammary gland against IMI. 20. High Framingham risk score decreases quality of life in adults Directory of Open Access Journals (Sweden) Christian Yosaputra 2010-04-01 Full Text Available Cardiovascular disease (CVD risk factors, such as diabetes, hypertension, hypercholesterolemia, smoking, and obesity tend to occur together in the general population. Increasing prevalence of multiple CVD risk factors has been related to increased risk of death from coronary heart disease and stroke. Studies have suggested that people with several risk factors of CVD may have impaired health-related quality of life. The objective of this study was to assess the association of CVD risk factors with quality of life (QOL among adults aged 40 to 65 years. A cross-sectional study was conducted involving 220 subjects 40 - 65 years of age at a health center. The CVD risk factors were assessed using the Framingham risk score that is the standard instrument for assessment of the risk of a first cardiac event. The risk factors assessed were age, smoking, blood pressure, total cholesterol and high density lipoprotein cholesterol concentrations. QOL was assessed by means of the WHOQOL-BREF instrument that had been prevalidated. The results of the study showed that 28.2% of subjects were smokers, 56.4% had stage 1 hypertension, 42.8% high total cholesterol and 13.6% low HDL cholesterol. The high risk group amounted to 45.5% and 42.3% constitued an intermediate risk group. High CVD risk scores were significantly associated with a low QOL for all domains (physical, psychological, social and environment (p=0.000. Preventing or reducing the multiple CVD risk factors to improve QOL is necessary among adults. 1. Automated Theorem Proving in High-Quality Software Design Science.gov (United States) Schumann, Johann; Swanson, Keith (Technical Monitor) 2001-01-01 The amount and complexity of software developed during the last few years has increased tremendously. In particular, programs are being used more and more in embedded systems (from car-brakes to plant-control). Many of these applications are safety-relevant, i.e. a malfunction of hardware or software can cause severe damage or loss. Tremendous risks are typically present in the area of aviation, (nuclear) power plants or (chemical) plant control. Here, even small problems can lead to thousands of casualties and huge financial losses. Large financial risks also exist when computer systems are used in the area of telecommunication (telephone, electronic commerce) or space exploration. Computer applications in this area are not only subject to safety considerations, but also security issues are important. All these systems must be designed and developed to guarantee high quality with respect to safety and security. Even in an industrial setting which is (or at least should be) aware of the high requirements in Software Engineering, many incidents occur. For example, the Warshaw Airbus crash, was caused by an incomplete requirements specification. Uncontrolled reuse of an Ariane 4 software module was the reason for the Ariane 5 disaster. Some recent incidents in the telecommunication area, like illegal "cloning" of smart-cards of D2GSM handies, or the extraction of (secret) passwords from German T-online users show that also in this area serious flaws can happen. Due to the inherent complexity of computer systems, most authors claim that only a rigorous application of formal methods in all stages of the software life cycle can ensure high quality of the software and lead to real safe and secure systems. In this paper, we will have a look, in how far automated theorem proving can contribute to a more widespread application of formal methods and their tools, and what automated theorem provers (ATPs) must provide in order to be useful. 2. Recruiting and retaining high-quality teachers in rural areas. Science.gov (United States) Monk, David H 2007-01-01 In examining recruitment and retention of teachers in rural areas, David Monk begins by noting the numerous possible characteristics of rural communities--small size, sparse settlement, distance from population concentrations, and an economic reliance on agricultural industries that are increasingly using seasonal and immigrant workers to minimize labor costs. Many, though not all, rural areas, he says, are seriously impoverished. Classes in rural schools are relatively small, and teachers tend to report satisfaction with their work environments and relatively few problems with discipline. But teacher turnover is often high, and hiring can be difficult. Monk observes that rural schools have a below-average share of highly trained teachers. Compensation in rural schools tends to be low, perhaps because of a lower fiscal capacity in rural areas, thus complicating efforts to attract and retain teachers. Several student characteristics, including relatively large shares of students with special needs and with limited English skills and lower shares of students attending college, can also make it difficult to recruit and retain high-quality teachers. Other challenges include meeting the needs of highly mobile children of low-income migrant farm workers. With respect to public policy, Monk asserts a need to focus on a subcategory of what might be called hard-to-staff rural schools rather than to develop a blanket set of policies for all rural schools. In particular, he recommends a focus on such indicators as low teacher qualifications, teaching in fields far removed from the area of training, difficulty in hiring, high turnover, a lack of diversity among teachers in the school, and the presence of migrant farm workers' children. Successful efforts to stimulate economic growth in these areas would be highly beneficial. He also calls attention to the potential for modern telecommunication and computing technologies to offset some of the drawbacks associated with teaching 3. Routine quality control of high dose rate brachytherapy equipment International Nuclear Information System (INIS) Guzman Calcina, Carmen S.; Almeida, Adelaide de; Rocha, Jose R. Oliveira 2001-01-01 A Quality Assurance program should be installed also for High Dose Rate brachytherapy, in the order to achieve a correct dose administration to the patient and for the safety to those involved directly with the treatment. The work presented here has the following purposes: Analyze the types of equipment tests presented by the official protocols (TG40, TG56 e ARCAL XXX), evaluate the brachytherapy routine tests of protocols from various national and international radiotherapy services and compare the latter with those presented in the official protocols. As a result, we conclude the following: TG56 presents a higher number of tests when compared to the other official protocols and most of the tests presented by the analyzed services are present in TG56. A suggestion for a basic protocol is presented, emphasizing the periodicity and tolerance level of each of the tests. (author) 4. High quality diesel fuels by VO-LSGO hydrotreatment Energy Technology Data Exchange (ETDEWEB) Stanica-Ezeanu, Dorin; Juganaru, Traian [Petroleum and Gas Univ. of Ploiesti (Romania) 2013-06-01 The aim of the paper is to obtain a high quality Diesel fuel by hydro-deoxigenation of vegetable oils (VO) mixed with a low sulfur gasoil (LSGO). The process is possible by using a bi-functional catalyst Ni-Mo supported by an activated Al{sub 2}O{sub 3} containing 2% Ultrastable Y-zeolite. The experimental conditions were: T =340 - 380 C, Pressure = 50 bar, LHSV = 1,5 h{sup -1}, H{sub 2}/Feed ratio = 15 mole H{sub 2} /mole liquid feed. The liquid product was separated in two fractions: a light distillate (similar to gasoline) and a heavy distillate (boiling point > 200 C) with very good characteristics for Diesel engines. The reaction chemistry is very complex, but the de-oxygenation process is decisive for the chemical structure of hydrocarbons from final product. Finally, a schema for the reaction mechanism is proposed. (orig.) 5. Supercapacitors based on high-quality graphene scrolls Science.gov (United States) Zeng, Fanyan; Kuang, Yafei; Liu, Gaoqin; Liu, Rui; Huang, Zhongyuan; Fu, Chaopeng; Zhou, Haihui 2012-06-01 High-quality graphene scrolls (GSS) with a unique scrolled topography are designed using a microexplosion method. Their capacitance properties are investigated by cyclic voltammetry, galvanostatic charge-discharge and electrical impedance spectroscopy. Compared with the specific capacity of 110 F g-1 for graphene sheets, a remarkable capacity of 162.2 F g-1 is obtained at the current density of 1.0 A g-1 in 6 M KOH aqueous solution owing to the unique scrolled structure of GSS. The capacity value is increased by about 50% only because of the topological change of graphene sheets. Meanwhile, GSS exhibit excellent long-term cycling stability along with 96.8% retained after 1000 cycles at 1.0 A g-1. These encouraging results indicate that GSS based on the topological structure of graphene sheets are a kind of promising material for supercapacitors. 6. A roadmap to high quality chemically prepared graphene Energy Technology Data Exchange (ETDEWEB) Gengler, Regis Y N; Spyrou, Konstantinos; Rudolf, Petra, E-mail: [email protected], E-mail: [email protected] [Zernike Institute for Advanced Materials, University of Groningen, Nijenborgh 4, 9747AG Groningen (Netherlands) 2010-09-22 Graphene was discovered half a decade ago and proved the existence of a two-dimensional system which becomes stable as a result of 3D corrugation. It appeared very quickly that this exceptional material had truly outstanding electronic, mechanical, thermal and optical properties. Consequently a broad range of applications appeared, as the graphene science speedily moved forward. Since then, a lot of effort has been devoted not only to the study of graphene but also to its fabrication. Here we review the chemical approaches to graphene production, their advantages as well as their downsides. Our aim is to draw a roadmap of today's most reliable path to high quality graphene via chemical preparation. 7. High quality factor HTS Josephson junctions on low loss substrates Energy Technology Data Exchange (ETDEWEB) Stornaiuolo, D; Longobardi, L; Massarotti, D; Barone, A; Tafuri, F [CNR-SPIN Napoli, Complesso Universitario di Monte Sant' Angelo, via Cinthia, 80126 Napoli (Italy); Papari, G; Carillo, F [NEST, CNR-NANO and Scuola Normale Superiore, Piazza San Silvestro 12, 56127 Pisa (Italy); Cennamo, N [Dipartimento Ingegneria dell' Informazione, Seconda Universita degli Studi di Napoli, via Roma 29, 81031 Aversa (Italy) 2011-04-15 We have extended the off-axis biepitaxial technique to produce YBCO grain boundary junctions on low loss substrates. Excellent transport properties have been reproducibly found, with remarkable values of the quality factor I{sub c}R{sub n} (with I{sub c} the critical current and R{sub n} the normal state resistance) above 10 mV, far higher than the values commonly reported in the literature for high temperature superconductor (HTS) based Josephson junctions. The outcomes are consistent with a picture of a more uniform grain boundary region along the current path. This work supports a possible implementation of grain boundary junctions for various applications including terahertz sensors and HTS quantum circuits in the presence of microwaves. 8. A roadmap to high quality chemically prepared graphene International Nuclear Information System (INIS) Gengler, Regis Y N; Spyrou, Konstantinos; Rudolf, Petra 2010-01-01 Graphene was discovered half a decade ago and proved the existence of a two-dimensional system which becomes stable as a result of 3D corrugation. It appeared very quickly that this exceptional material had truly outstanding electronic, mechanical, thermal and optical properties. Consequently a broad range of applications appeared, as the graphene science speedily moved forward. Since then, a lot of effort has been devoted not only to the study of graphene but also to its fabrication. Here we review the chemical approaches to graphene production, their advantages as well as their downsides. Our aim is to draw a roadmap of today's most reliable path to high quality graphene via chemical preparation. 9. CHOREOGRAPHIC METHODS FOR CREATING NOVEL, HIGH QUALITY DANCE Directory of Open Access Journals (Sweden) David Kirsh 2016-02-01 Full Text Available We undertook a detailed ethnographic study of the dance creation process of a noted choreographer and his distinguished troupe. All choreographer dancer interactions were videoed, the choreographer and dancers were interviewed extensively each day, as well as other observations and tests performed. The choreographer used three main methods to produce high quality and novel content: showing, making-on, and tasking. We present, analyze and evaluate these methods, and show how these approaches allow the choreographer to increase the creative output of the dancers and him. His methods, although designed for dance, apply more generally to other creative endeavors, especially where brainstorming is involved, and where the creative process is distributed over many individuals. His approach is also a case study in multi-modal direction, owing to the range of mechanisms he uses to communicate and direct. 10. Biotransformation of Organic Waste into High Quality Fertilizer DEFF Research Database (Denmark) Bryndum, Sofie Agriculture faces several challenges of future provision of nutrients such as limited P reserves and increasing prices of synthetic fertilizers and recycling of nutrients from organic waste can be an important strategy for the long-term sustainability of the agricultural systems. Organically...... and S, is often low; and (3) the unbalanced composition of nutrients rarely matches crop demands. Therefore the objective of this project was to investigate the potential for (1) recycling nutrients from agro-industrial wastes and (2) compost biotransformation into high-quality organic fertilizers...... other uses into fertilizer use would be unlikely. An estimated ~50 % of the total organic waste pool, primarily consisting of animal manure and waste from the processing of sugar cane, coffee, oil palm and oranges, is currently being re-used as “fertilizers”, meaning it is eventually returned... 11. Randomized phase III trial of APF530 versus palonosetron in the prevention of chemotherapy-induced nausea and vomiting in a subset of patients with breast cancer receiving moderately or highly emetogenic chemotherapy International Nuclear Information System (INIS) Boccia, Ralph; O’Boyle, Erin; Cooper, William 2016-01-01 APF530 provides controlled, sustained-release granisetron for preventing acute (0–24 h) and delayed (24–120 h) chemotherapy-induced nausea and vomiting (CINV). In a phase III trial, APF530 was noninferior to palonosetron in preventing acute CINV following single-dose moderately (MEC) or highly emetogenic chemotherapy (HEC) and delayed CINV in MEC (MEC and HEC defined by Hesketh criteria). This exploratory subanalysis was conducted in the breast cancer subpopulation. Patients were randomized to subcutaneous APF530 250 or 500 mg (granisetron 5 or 10 mg) or intravenous palonosetron 0.25 mg during cycle 1. Palonosetron patients were randomized to APF530 for cycles 2 to 4. The primary efficacy end point was complete response (CR, no emesis or rescue medication) in cycle 1. Among breast cancer patients (n = 423 MEC, n = 185 HEC), > 70 % received anthracycline-containing regimens in each emetogenicity subgroup. There were no significant between-group differences in CRs in cycle 1 for acute (APF530 250 mg: MEC 71 %, HEC 77 %; 500 mg: MEC 73 %, HEC 73 %; palonosetron: MEC 68 %, HEC 66 %) and delayed (APF530 250 mg: MEC 46 %, HEC 58 %; 500 mg: MEC 48 %, HEC 63 %; palonosetron: MEC 52 %, HEC 52 %) CINV. There were no significant differences in within-cycle CRs between APF530 doses for acute and delayed CINV in MEC or HEC in cycles 2 to 4; CRs trended higher in later cycles, with no notable differences in adverse events between breast cancer and overall populations. APF530 effectively prevented acute and delayed CINV over 4 chemotherapy cycles in breast cancer patients receiving MEC or HEC. Clinicaltrials.gov identifier: NCT00343460 (June 22, 2006) 12. Urofacial syndrome: A subset of neurogenic bladder dysfunction syndromes? Directory of Open Access Journals (Sweden) K N Stamatiou 2010-01-01 Full Text Available The urofacial syndrome is probably a subset of neurogenic bladder dysfunction syndromes characterized by detrusor-sphincter discoordination along with a characteristic inversion of facial expression with laughing. This characteristic facial expression can facilitate early detection of this disorder, which leads to poor bladder emptying with high residual urine, hydro-nephrosis with vesico-ureteral reflux and potentially renal failure if left untreated. The etiology of the urofacial syndrome is unknown. In our case, a 12-year-old boy of Middle-Eastern origin presented to the Outpatient Department of our hospital with left pyelonephritis, hydronephrosis and bladder dilatation. Voiding cystourethrography performed 15 days later revealed left vesicoureteral reflux. Cystoscopy revealed bladder trabeculation however an anatomic urethral obstruction was not noticed. Both, neurological examination and radiography of the lumbosacral spine were normal. Urodynamic evaluation revealed the typical findings of detrusor-sphincter discoordination. 13. Statistical image reconstruction for transmission tomography using relaxed ordered subset algorithms International Nuclear Information System (INIS) Kole, J S 2005-01-01 Statistical reconstruction methods offer possibilities for improving image quality as compared to analytical methods, but current reconstruction times prohibit routine clinical applications in x-ray computed tomography (CT). To reduce reconstruction times, we have applied (under) relaxation to ordered subset algorithms. This enables us to use subsets consisting of only single projection angle, effectively increasing the number of image updates within an entire iteration. A second advantage of applying relaxation is that it can help improve convergence by removing the limit cycle behaviour of ordered subset algorithms, which normally do not converge to an optimal solution but rather a suboptimal limit cycle consisting of as many points as there are subsets. Relaxation suppresses the limit cycle behaviour by decreasing the stepsize for approaching the solution. A simulation study for a 2D mathematical phantom and three different ordered subset algorithms shows that all three algorithms benefit from relaxation: equal noise-to-resolution trade-off can be achieved using fewer iterations than the conventional algorithms, while a lower minimal normalized mean square error (NMSE) clearly indicates a better convergence. Two different schemes for setting the relaxation parameter are studied, and both schemes yield approximately the same minimal NMSE 14. High quality protein microarray using in situ protein purification Directory of Open Access Journals (Sweden) Fleischmann Robert D 2009-08-01 Full Text Available Abstract Background In the postgenomic era, high throughput protein expression and protein microarray technologies have progressed markedly permitting screening of therapeutic reagents and discovery of novel protein functions. Hexa-histidine is one of the most commonly used fusion tags for protein expression due to its small size and convenient purification via immobilized metal ion affinity chromatography (IMAC. This purification process has been adapted to the protein microarray format, but the quality of in situ His-tagged protein purification on slides has not been systematically evaluated. We established methods to determine the level of purification of such proteins on metal chelate-modified slide surfaces. Optimized in situ purification of His-tagged recombinant proteins has the potential to become the new gold standard for cost-effective generation of high-quality and high-density protein microarrays. Results Two slide surfaces were examined, chelated Cu2+ slides suspended on a polyethylene glycol (PEG coating and chelated Ni2+ slides immobilized on a support without PEG coating. Using PEG-coated chelated Cu2+ slides, consistently higher purities of recombinant proteins were measured. An optimized wash buffer (PBST composed of 10 mM phosphate buffer, 2.7 mM KCl, 140 mM NaCl and 0.05% Tween 20, pH 7.4, further improved protein purity levels. Using Escherichia coli cell lysates expressing 90 recombinant Streptococcus pneumoniae proteins, 73 proteins were successfully immobilized, and 66 proteins were in situ purified with greater than 90% purity. We identified several antigens among the in situ-purified proteins via assays with anti-S. pneumoniae rabbit antibodies and a human patient antiserum, as a demonstration project of large scale microarray-based immunoproteomics profiling. The methodology is compatible with higher throughput formats of in vivo protein expression, eliminates the need for resin-based purification and circumvents 15. The structure of high-quality aluminium cast iron Directory of Open Access Journals (Sweden) D. Kopyciński 2012-01-01 Full Text Available In this study presents the analyse of aluminium iron cast structure (as-cast condition which are used in high temperature. While producing the casts of aluminium iron major influence has been preserve the structure of technological process parameters. The addition to Fe-C-Al alloy V, Ti, Cr leads to the improvement of functional and mechanical cast qualities. In this study, a method was investigated to eliminate the presence of undesirable Al4C3 phases in a aluminium cast iron structure and thus improve the production process. V and Ti additions in aluminium cast iron allows to development of FeAl - VC or TiC alloys. In particular, V or Ti contents above 5 wt.% were found to totally eliminate the presence of Al4C3. In addition, preliminary work indicates that the alloy with the FeAl - VC or TiC structure reveals high oxidation resistance. The introduction of 5 wt.% chromium to aluminium cast iron strengthened Al4C3 precipitate. Thus, the resultant alloy can be considered an intermetallic FeAl matrix strengthened by VC and TiC or modified Al4C3 reinforcements. 16. High-Quality Seismic Observations of Sonic Booms Science.gov (United States) Wurman, Gilead; Haering, Edward A., Jr.; Price, Michael J. 2011-01-01 The SonicBREWS project (Sonic Boom Resistant Earthquake Warning Systems) is a collaborative effort between Seismic Warning Systems, Inc. and NASA Dryden Flight Research Center. This project aims to evaluate the effects of sonic booms on Earthquake Warning Systems in order to prevent such systems from experiencing false alarms due to sonic booms. The airspace above the Antelope Valley, California includes the High Altitude Supersonic Corridor and the Black Mountain Supersonic Corridor. These corridors are among the few places in the US where supersonic flight is permitted, and sonic booms are commonplace in the Antelope Valley. One result of this project is a rich dataset of high-quality accelerometer records of sonic booms which can shed light on the interaction between these atmospheric phenomena and the solid earth. Nearly 100 sonic booms were recorded with low-noise triaxial MEMS accelerometers recording 1000 samples per second. The sonic booms had peak overpressures ranging up to approximately 10 psf and were recorded in three flight series in 2010 and 2011. Each boom was recorded with up to four accelerometers in various array configurations up to 100 meter baseline lengths, both in the built environment and the free field. All sonic booms were also recorded by nearby microphones. We present the results of the project in terms of the potential for sonic-boom-induced false alarms in Earthquake Warning Systems, and highlight some of the interesting features of the dataset. 17. High-quality remote interactive imaging in the operating theatre Science.gov (United States) Grimstead, Ian J.; Avis, Nick J.; Evans, Peter L.; Bocca, Alan 2009-02-01 We present a high-quality display system that enables the remote access within an operating theatre of high-end medical imaging and surgical planning software. Currently, surgeons often use printouts from such software for reference during surgery; our system enables surgeons to access and review patient data in a sterile environment, viewing real-time renderings of MRI & CT data as required. Once calibrated, our system displays shades of grey in Operating Room lighting conditions (removing any gamma correction artefacts). Our system does not require any expensive display hardware, is unobtrusive to the remote workstation and works with any application without requiring additional software licenses. To extend the native 256 levels of grey supported by a standard LCD monitor, we have used the concept of "PseudoGrey" where slightly off-white shades of grey are used to extend the intensity range from 256 to 1,785 shades of grey. Remote access is facilitated by a customized version of UltraVNC, which corrects remote shades of grey for display in the Operating Room. The system is successfully deployed at Morriston Hospital, Swansea, UK, and is in daily use during Maxillofacial surgery. More formal user trials and quantitative assessments are being planned for the future. 18. Blueprint for action: steps toward a high-quality, high-value maternity care system. Science.gov (United States) Angood, Peter B; Armstrong, Elizabeth Mitchell; Ashton, Diane; Burstin, Helen; Corry, Maureen P; Delbanco, Suzanne F; Fildes, Barbara; Fox, Daniel M; Gluck, Paul A; Gullo, Sue Leavitt; Howes, Joanne; Jolivet, R Rima; Laube, Douglas W; Lynne, Donna; Main, Elliott; Markus, Anne Rossier; Mayberry, Linda; Mitchell, Lynn V; Ness, Debra L; Nuzum, Rachel; Quinlan, Jeffrey D; Sakala, Carol; Salganicoff, Alina 2010-01-01 Childbirth Connection hosted a 90th Anniversary national policy symposium, Transforming Maternity Care: A High Value Proposition, on April 3, 2009, in Washington, DC. Over 100 leaders from across the range of stakeholder perspectives were actively engaged in the symposium work to improve the quality and value of U.S. maternity care through broad system improvement. A multi-disciplinary symposium steering committee guided the strategy from its inception and contributed to every phase of the project. The "Blueprint for Action: Steps Toward a High Quality, High Value Maternity Care System", issued by the Transforming Maternity Care Symposium Steering Committee, answers the fundamental question, "Who needs to do what, to, for, and with whom to improve the quality of maternity care over the next five years?" Five stakeholder workgroups collaborated to propose actionable strategies in 11 critical focus areas for moving expeditiously toward the realization of the long term "2020 Vision for a High Quality, High Value Maternity Care System", also published in this issue. Following the symposium these workgroup reports and recommendations were synthesized into the current blueprint. For each critical focus area, the "Blueprint for Action" presents a brief problem statement, a set of system goals for improvement in that area, and major recommendations with proposed action steps to achieve them. This process created a clear sightline to action that if enacted could improve the structure, process, experiences of care, and outcomes of the maternity care system in ways that when anchored in the culture can indeed transform maternity care. Copyright 2010 Jacobs Institute of Women 19. Stress and sleep quality in high school brazilian adolescents Directory of Open Access Journals (Sweden) Gema Mesquita 2010-06-01 Full Text Available OBJECTIVE: The objective of the present study is to analyze the effect of stress on sleep quality in a group of adolescents. METHOD: Two high schools in Alfenas, southern Minas Gerais State, Brazil, were chosen to participate in the study. The sample consisted of both genders (n=160 with 65.63% females. The age range of participants was 15 to18 years. The Pittsburgh Sleep Quality Index (PSQI was applied for collection of data to quantify sleep quality. The Lipp Inventory of Stress Symptoms that objectively identifies symptoms of stress was applied. RESULTS: It was observed that 23.53% of stressed students and 45.33% of unstressed ones sleep well; 76.47% of stressed pupils and 54.67% of those unstressed do not sleep well. With regard to school performance, a mean of 0.65 was found for stressed students and 0.60 for those without stress, Mann-Whitney (p=0.0596. CONCLUSION: Stress contributed to raising the percentage of poor sleepers, as ell as increasing ean school performance.OBJETIVO: O objetivo do presente estudo foi analisar a influência do stress sobre a qualidade do sono em um grupo de adolescentes. MÉTODO: Foram escolhidas duas instituições educacionais do ensino médio, na cidade de Alfenas, sul de Minas Gerais, Brasil. A amostra foi composta por ambos os sexos (n=160, com 65,63% do sexo feminino. A faixa etária dos participantes foi de 15 a 18 anos. Para a coleta de dados aplicou-se: Índice de Qualidade de Sono de Pittsburgh (IQSP utilizado para quantificar a qualidade do sono; o Inventário de Sintomas de Stress para Adultos de Lipp (ISSL que identifica de modo objetivo a sintomatologia de stress foi aplicado. RESULTADOS: Observou-se que 23,53% dos estressados dormem bem e 45,33% dos não estressados dormem bem; 76,47% dos estressados não dormem bem e 54,67% dos não estressados não dormem bem. Quanto ao rendimento escolar têm-se as médias 0,65 para os alunos estressados e 0,60 para aqueles que não sofrem de stress, Mann 20. High-quality endoscope reprocessing decreases endoscope contamination. Science.gov (United States) Decristoforo, P; Kaltseis, J; Fritz, A; Edlinger, M; Posch, W; Wilflingseder, D; Lass-Flörl, C; Orth-Höller, D 2018-02-24 Several outbreaks of severe infections due to contamination of gastrointestinal (GI) endoscopes, mainly duodenoscopes, have been described. The rate of microbial endoscope contamination varies dramatically in literature. The aim of this multicentre prospective study was to evaluate the hygiene quality of endoscopes and automated endoscope reprocessors (AERs) in Tyrol/Austria. In 2015 and 2016, a total of 463 GI endoscopes and 105 AERs from 29 endoscopy centres were analysed by a routine (R) and a combined routine and advanced (CRA) sampling procedure and investigated for microbial contamination by culture-based and molecular-based analyses. The contamination rate of GI endoscopes was 1.3%-4.6% according to the national guideline, suggesting that 1.3-4.6 patients out of 100 could have had contacts with hygiene-relevant microorganisms through an endoscopic intervention. Comparison of R and CRA sampling showed 1.8% of R versus 4.6% of CRA failing the acceptance criteria in phase I and 1.3% of R versus 3.0% of CRA samples failing in phase II. The most commonly identified indicator organism was Pseudomonas spp., mainly Pseudomonas oleovorans. None of the tested viruses were detected in 40 samples. While AERs in phase I failed (n = 9, 17.6%) mainly due to technical faults, phase II revealed lapses (n = 6, 11.5%) only on account of microbial contamination of the last rinsing water, mainly with Pseudomonas spp. In the present study the contamination rate of endoscopes was low compared with results from other European countries, possibly due to the high quality of endoscope reprocessing, drying and storage. Copyright © 2018 European Society of Clinical Microbiology and Infectious Diseases. Published by Elsevier Ltd. All rights reserved. 1. 76 FR 45397 - Export Inspection and Weighing Waiver for High Quality Specialty Grain Transported in Containers Science.gov (United States) 2011-07-29 ...-AB18 Export Inspection and Weighing Waiver for High Quality Specialty Grain Transported in Containers... permanent a waiver due to expire on July 31, 2012, for high quality specialty grain exported in containers... of high quality specialty grain exported in containers are small entities that up until recently... 2. The efficiency of subset selection of an almost best treatment NARCIS (Netherlands) Laan, van der P. 1991-01-01 A generalized goal using subset selection is discussed for the location parameter case. This goal is to select a non-empty subset from a set of k (k \\geq 2) treatments that contains at least one \\epsilon-best treatment with confidence level P*. For a set of treatments an \\epsilon-best treatment is 3. Indirect Positive Evidence in the Acquisition of a Subset Grammar Science.gov (United States) Schwartz, Misha; Goad, Heather 2017-01-01 This article proposes that second language learners can use indirect positive evidence (IPE) to acquire a phonological grammar that is a subset of their L1 grammar. IPE is evidence from errors in the learner's L1 made by native speakers of the learner's L2. It has been assumed that subset grammars may be acquired using direct or indirect negative… 4. ';Best' Practices for Aggregating Subset Results from Archived Datasets Science.gov (United States) Baskin, W. E.; Perez, J. 2013-12-01 In response to the exponential growth in science data analysis and visualization capabilities Data Centers have been developing new delivery mechanisms to package and deliver large volumes of aggregated subsets of archived data. New standards are evolving to help data providers and application programmers deal with growing needs of the science community. These standards evolve from the best practices gleaned from new products and capabilities. The NASA Atmospheric Sciences Data Center (ASDC) has developed and deployed production provider-specific search and subset web applications for the CALIPSO, CERES, TES, and MOPITT missions. This presentation explores several use cases that leverage aggregated subset results and examines the standards and formats ASDC developers applied to the delivered files as well as the implementation strategies for subsetting and processing the aggregated products. The following topics will be addressed: - Applications of NetCDF CF conventions to aggregated level 2 satellite subsets - Data-Provider-Specific format requirements vs. generalized standards - Organization of the file structure of aggregated NetCDF subset output - Global Attributes of individual subsetted files vs. aggregated results - Specific applications and framework used for subsetting and delivering derivative data files 5. Bayesian Subset Modeling for High-Dimensional Generalized Linear Models KAUST Repository Liang, Faming; Song, Qifan; Yu, Kai 2013-01-01 criterion model. The consistency of the resulting posterior is established under mild conditions. Further, a variable screening procedure is proposed based on the marginal inclusion probability, which shares the same properties of sure screening 6. Circulating TFH subset distribution is strongly affected in lupus patients with an active disease. Directory of Open Access Journals (Sweden) Carole Le Coz Full Text Available Follicular helper T cells (TFH represent a distinct subset of CD4(+ T cells specialized in providing help to B lymphocytes, which may play a central role in autoimmune diseases having a major B cell component such as systemic lupus erythematosus. Recently, TFH subsets that share common phenotypic and functional characteristics with TFH cells from germinal centers, have been described in the peripheral blood from healthy individuals. The aim of this study was to analyze the distribution of such populations in lupus patients. Circulating TFH cell subsets were defined by multicolor flow cytometry as TFH17 (CXCR3(-CCR6(+, TFH1 (CXCR3 (+ CCR6(- or TFH2 (CXCR3(-CCR6(- cells among CXCR5 (+ CD45RA(-CD4(+ T cells in the peripheral blood of 23 SLE patients and 23 sex and age-matched healthy controls. IL-21 receptor expression by B cells was analyzed by flow cytometry and the serum levels of IL-21 and Igs were determined by ELISA tests. We found that the TFH2 cell subset frequency is strongly and significantly increased in lupus patients with an active disease (SLEDAI score>8, while the TFH1 cell subset percentage is greatly decreased. The TFH2 and TFH1 cell subset frequency alteration is associated with the presence of high Ig levels and autoantibodies in patient's sera. Moreover, the TFH2 cell subset enhancement correlates with an increased frequency of double negative memory B cells (CD27(-IgD(-CD19(+ cells expressing the IL-21R. Finally, we found that IgE levels in lupus patients' sera correlate with disease activity and seem to be associated with high TFH2 cell subset frequency. In conclusion, our study describes for the first time the distribution of circulating TFH cell subsets in lupus patients. Interestingly, we found an increased frequency of TFH2 cells, which correlates with disease activity. Our results suggest that this subset might play a key role in lupus pathogenesis. 7. Tachykinins stimulate a subset of mouse taste cells. Directory of Open Access Journals (Sweden) Jeff Grant Full Text Available The tachykinins substance P (SP and neurokinin A (NKA are present in nociceptive sensory fibers expressing transient receptor potential cation channel, subfamily V, member 1 (TRPV1. These fibers are found extensively in and around the taste buds of several species. Tachykinins are released from nociceptive fibers by irritants such as capsaicin, the active compound found in chili peppers commonly associated with the sensation of spiciness. Using real-time Ca(2+-imaging on isolated taste cells, it was observed that SP induces Ca(2+ -responses in a subset of taste cells at concentrations in the low nanomolar range. These responses were reversibly inhibited by blocking the SP receptor NK-1R. NKA also induced Ca(2+-responses in a subset of taste cells, but only at concentrations in the high nanomolar range. These responses were only partially inhibited by blocking the NKA receptor NK-2R, and were also inhibited by blocking NK-1R indicating that NKA is only active in taste cells at concentrations that activate both receptors. In addition, it was determined that tachykinin signaling in taste cells requires Ca(2+-release from endoplasmic reticulum stores. RT-PCR analysis further confirmed that mouse taste buds express NK-1R and NK-2R. Using Ca(2+-imaging and single cell RT-PCR, it was determined that the majority of tachykinin-responsive taste cells were Type I (Glial-like and umami-responsive Type II (Receptor cells. Importantly, stimulating NK-1R had an additive effect on Ca(2+ responses evoked by umami stimuli in Type II (Receptor cells. This data indicates that tachykinin release from nociceptive sensory fibers in and around taste buds may enhance umami and other taste modalities, providing a possible mechanism for the increased palatability of spicy foods. 8. Long-Run Benefits from Universal High-Quality Preschooling DEFF Research Database (Denmark) Bauchmüller, Robert; Gørtz, Mette; Rasmussen, Astrid Würtz 2014-01-01 This paper investigates the role of preschool quality for children's school performance at the end of primary school. We construct five structural quality indicators based on unique Danish administrative register data. 30,444 children finishing primary school's 9th grade in 2008 and who attended ...... in children's test results in Danish. Boys benefit more from preschool quality than girls. Ethnic minority children benefit from higher staff stability.... 9. Low-quality birds do not display high-quality signals: The cysteine-pheomelanin mechanism of honesty Science.gov (United States) Galván, Ismael; Wakamatsu, Kazumasa; Camarero, Pablo R; Mateo, Rafael; Alonso-Alvarez, Carlos 2015-01-01 The mechanisms that make that the costs of producing high-quality signals are unaffordable to low-quality signalers are a current issue in animal communication. The size of the melanin-based bib of male house sparrows Passer domesticus honestly signals quality. We induced the development of new bibs while treating males with buthionine-sulfoximine (BSO), a substance that depletes the levels of the antioxidant glutathione (GSH) and the amino acid cysteine, two elements that switch melanogenesis from eumelanin to pheomelanin. Final bib size is negatively related to pheomelanin levels in the bib feathers. BSO reduced cysteine and GSH levels in all birds, but improved phenotypes (bibs larger than controls) were only expressed by high-quality birds (BSO birds with largest bibs initially). Negative associations between final bib size and cysteine levels in erythrocytes, and between pheomelanin and cysteine levels, were observed in high-quality birds only. These findings suggest that a mechanism uncoupling pheomelanin and cysteine levels may have evolved in low-quality birds to avoid producing bibs of size not corresponding to their quality and greater relative costs. Indeed, greater oxidative stress in cells was not observed in low-quality birds. This may represent the first mechanism maintaining signal honesty without producing greater relative costs on low-quality signalers. PMID:25330349 10. Systems and processes that ensure high quality care. Science.gov (United States) Bassett, Sally; Westmore, Kathryn 2012-10-01 This is the second in a series of articles examining the components of good corporate governance. It considers how the structures and processes for quality governance can affect an organisation's ability to be assured about the quality of care. Complex information systems and procedures can lead to poor quality care, but sound structures and processes alone are insufficient to ensure good governance, and behavioural factors play a significant part in making sure that staff are enabled to provide good quality care. The next article in this series looks at how the information reporting of an organisation can affect its governance. 11. Characterization of non equilibrium effects on high quality critical flows International Nuclear Information System (INIS) Camelo, E.; Lemonnier, H.; Ochterbeck, J. 1995-01-01 The appropriate design of various pieces of safety equipment such as relief systems, relies on the accurate description of critical flow phenomena. Most of the systems of industrial interest are willing to be described by one-dimensional area-averaged models and a large fraction of them involves multi-component high gas quality flows. Within these circumstances, the flow is very likely to be of an annular dispersed nature and its description by two-fluid models requires various closure relations. Among the most sensitive closures, there is the interfacial area and the liquid entrained fraction. The critical flowrate depends tremendously on the accurate description of the non equilibrium which results from the correctness of the closure equations. In this study, two-component flows are emphasized and non equilibrium results mainly form the differences in the phase velocities. It is therefore of the utmost importance to have reliable data to characterize non equilibrium phenomena and to assess the validity of the closure models. A comprehensive description of air-water nozzle flows, with emphasis on the effect of the nozzle geometry, has been undertaken and some of the results are presented here which helps understanding the overall flow dynamics. Besides the critical flowrate, the presented material includes pressure profiles, droplet size and velocity, liquid film flowrate and liquid film thickness 12. Characterization of non equilibrium effects on high quality critical flows Energy Technology Data Exchange (ETDEWEB) Camelo, E.; Lemonnier, H.; Ochterbeck, J. [Commissariat a l Energie Atomique, Grenoble (France)] [and others 1995-09-01 The appropriate design of various pieces of safety equipment such as relief systems, relies on the accurate description of critical flow phenomena. Most of the systems of industrial interest are willing to be described by one-dimensional area-averaged models and a large fraction of them involves multi-component high gas quality flows. Within these circumstances, the flow is very likely to be of an annular dispersed nature and its description by two-fluid models requires various closure relations. Among the most sensitive closures, there is the interfacial area and the liquid entrained fraction. The critical flowrate depends tremendously on the accurate description of the non equilibrium which results from the correctness of the closure equations. In this study, two-component flows are emphasized and non equilibrium results mainly form the differences in the phase velocities. It is therefore of the utmost importance to have reliable data to characterize non equilibrium phenomena and to assess the validity of the closure models. A comprehensive description of air-water nozzle flows, with emphasis on the effect of the nozzle geometry, has been undertaken and some of the results are presented here which helps understanding the overall flow dynamics. Besides the critical flowrate, the presented material includes pressure profiles, droplet size and velocity, liquid film flowrate and liquid film thickness. 13. Image Quality in High-resolution and High-cadence Solar Imaging Science.gov (United States) Denker, C.; Dineva, E.; Balthasar, H.; Verma, M.; Kuckein, C.; Diercke, A.; González Manrique, S. J. 2018-03-01 Broad-band imaging and even imaging with a moderate bandpass (about 1 nm) provides a photon-rich environment, where frame selection (lucky imaging) becomes a helpful tool in image restoration, allowing us to perform a cost-benefit analysis on how to design observing sequences for imaging with high spatial resolution in combination with real-time correction provided by an adaptive optics (AO) system. This study presents high-cadence (160 Hz) G-band and blue continuum image sequences obtained with the High-resolution Fast Imager (HiFI) at the 1.5-meter GREGOR solar telescope, where the speckle-masking technique is used to restore images with nearly diffraction-limited resolution. The HiFI employs two synchronized large-format and high-cadence sCMOS detectors. The median filter gradient similarity (MFGS) image-quality metric is applied, among others, to AO-corrected image sequences of a pore and a small sunspot observed on 2017 June 4 and 5. A small region of interest, which was selected for fast-imaging performance, covered these contrast-rich features and their neighborhood, which were part of Active Region NOAA 12661. Modifications of the MFGS algorithm uncover the field- and structure-dependency of this image-quality metric. However, MFGS still remains a good choice for determining image quality without a priori knowledge, which is an important characteristic when classifying the huge number of high-resolution images contained in data archives. In addition, this investigation demonstrates that a fast cadence and millisecond exposure times are still insufficient to reach the coherence time of daytime seeing. Nonetheless, the analysis shows that data acquisition rates exceeding 50 Hz are required to capture a substantial fraction of the best seeing moments, significantly boosting the performance of post-facto image restoration. 14. Quality Assurance Roadmap for High Performance Residential Buildings Energy Technology Data Exchange (ETDEWEB) None 2008-10-05 This report outlines the approach to quality assurance in the construction process for new residential construction, including seven process steps from the assessment of current construction practice, through design and documentation changes, to training and quality control for on-site personnel. 15. Paying for High- and Low-Quality Teaching Science.gov (United States) Schacter, John; Thum, Yeow Meng 2004-01-01 The extensive research on teacher quality has led to two conclusions. First, there are large and significant differences among teachers in terms of their capacity to improve student achievement. Second, these differences are not captured by common measures of teacher qualifications (E.A. Hanushek, Teacher quality, in: L.T. Izumi, W.M. Evers… 16. Teaching quality: High school students' autonomy and competence. Science.gov (United States) León, Jaime; Medina-Garrido, Elena; Ortega, Miriam 2018-05-01 How teachers manage class learning and interact with students affects students’ motivation and engagement. However, it could be that the effect of students’ representation of teaching quality on the students’ motivation varies between classes. Students from 90 classes participated in the study. We used multilevel random structural equation modeling to analyze whether the relationship of the students’ perception of teaching quality (as an indicator of the students’ mental representation) and students’ motivation varies between classes, and if this variability depends on the class assessment of teaching quality (as an indicator of teaching quality). The effect of teachers’ structure on the regression slope of student perception of student competence was .127. The effect of teachers’ autonomy support on the regression slope of student perception of student autonomy was .066. With this study we contribute a more detailed description of the relationship between teaching quality, competence and autonomy. 17. New developments in high quality grey cast irons Directory of Open Access Journals (Sweden) Iulian Riposan 2014-07-01 Full Text Available The paper reviews original data obtained by the present authors, revealed in recent separate publications, describing specific procedures for high quality grey irons, and reflecting the forecast needs of the worldwide iron foundry industry. High power, medium frequency coreless induction furnaces are commonly used in electric melting grey iron foundries. This has resulted in low sulphur (1,500 °C, contributing to unfavourable conditions for graphite nucleation. Thin wall castings are increasingly produced by these electric melt shops with a risk of greater eutectic undercooling during solidification. The paper focused on two groups of grey cast irons and their specific problems: carbides and graphite morphology control in lower carbon equivalent high strength irons (CE=3.4%-3.8%, and austenite dendrite promotion in eutectic and slightly hypereutectic irons (CE=4.1%-4.5%, in order to increase their strength characteristics. There are 3 stages and 3 steps involving graphite formation, iron chemistry and iron processing that appear to be important. The concept in the present paper sustains a threestage model for nucleating flake graphite [(Mn,XS type nuclei]. There are three important groups of elements (deoxidizer, Mn/S, and inoculant and three technological stages in electric melting of iron (superheat, pre-conditioning of base iron, final inoculation. Attention is drawn to a control factor (%Mn x (%S ensuring it equals to 0.03 – 0.06, accompanied by 0.005wt.%–0.010wt.% Al and/or Zr content in inoculated irons. It was found that iron powder addition promotes austenite dendrite formation in eutectic and slightly eutectic, acting as reinforcement for the eutectic cells. But, there is an accompanying possible negative influence on the characteristics of the (Mn,XS type graphite nuclei (change the morphology of nuclei from polygonal compact to irregular polygonal, and therefore promote chill tendency in treated irons. A double addition (iron 18. A high yielding, better quality chickpea mutant variety 'NIFA-95' International Nuclear Information System (INIS) Hassan, S.; Javed, M.A.; Khattak, S.U.K.; Iqbal, M.M. 2001-01-01 Chickpea or gram (Cicer arietinum L.) is an important legume crop of Pakistan, grown on over one million hectares annually. The national average yield of the crop is very low (0.5 t/ha) and thus the country had to spent about 2 billion rupees ($ 50 million) on import of pulses. The main causes of low yield are non-availability of genetic sources for resistance to various diseases especially gram blight Ascochyta rabiei (Pass.) Lab., insect pest (Pod borer) and non-adoption of proper production technology by the farmers. This calls for earnest efforts of breeders to evolve high yielding and disease resistant varieties of chickpea for provision of quality seeds to the farming community to increase production of this important crop. Seeds of a highly blight susceptible variety '6153' were irradiated at 200 Gy dose of gamma radiation in 1985 and the promising mutant line CMN-446-4 was selected in M3 generation on the basis of disease resistance, greater number of pods and better plant type. After confirmation of its resistance to blight in M 4 and M 5 , the mutant line was evaluated in various trials at different locations. In the advanced and zonal yield trials during 1993-95, the line CMN-446-4 produced the highest grain yield of 2,600 kg/ha as compared to the rest of the mutants and varieties. The line was also evaluated in the chickpea national uniform yield trial, conducted on over 11 locations in the country during 1993-94. In this trial, the mutant line ranked 3rd by producing an average yield of 1,528 kg/ha as compared to the two check varieties 'Punjab-91' (1,316 kg/ha) and 'Paidar-91' (1,391 kg/ha). The mutant line CMN-446-4 is moderately resistant to gram blight, highly resistant to stored pest (pulse beetle), contains 25.3% more protein as compared to the parental variety 6153 and is also better in nitrogen fixing capacity.The proposal for release of the mutant line CMN-446-4 as a new variety under the name 'NIFA-95' for general cultivation in the rainfed 19. Comparison of Subset-Based Local and Finite Element-Based Global Digital Image Correlation KAUST Repository Pan, Bing 2015-02-12 Digital image correlation (DIC) techniques require an image matching algorithm to register the same physical points represented in different images. Subset-based local DIC and finite element-based (FE-based) global DIC are the two primary image matching methods that have been extensively investigated and regularly used in the field of experimental mechanics. Due to its straightforward implementation and high efficiency, subset-based local DIC has been used in almost all commercial DIC packages. However, it is argued by some researchers that FE-based global DIC offers better accuracy because of the enforced continuity between element nodes. We propose a detailed performance comparison between these different DIC algorithms both in terms of measurement accuracy and computational efficiency. Then, by measuring displacements of the same calculation points using the same calculation algorithms (e.g., correlation criterion, initial guess estimation, subpixel interpolation, optimization algorithm and convergence conditions) and identical calculation parameters (e.g., subset or element size), the performances of subset-based local DIC and two FE-based global DIC approaches are carefully compared in terms of measurement error and computational efficiency using both numerical tests and real experiments. A detailed examination of the experimental results reveals that, when subset (element) size is not very small and the local deformation within a subset (element) can be well approximated by the shape function used, standard subset-based local DIC approach not only provides better results in measured displacements, but also demonstrates much higher computation efficiency. However, several special merits of FE-based global DIC approaches are indicated. 20. Comparison of Subset-Based Local and Finite Element-Based Global Digital Image Correlation KAUST Repository Pan, Bing; Wang, B.; Lubineau, Gilles; Moussawi, Ali 2015-01-01 Digital image correlation (DIC) techniques require an image matching algorithm to register the same physical points represented in different images. Subset-based local DIC and finite element-based (FE-based) global DIC are the two primary image matching methods that have been extensively investigated and regularly used in the field of experimental mechanics. Due to its straightforward implementation and high efficiency, subset-based local DIC has been used in almost all commercial DIC packages. However, it is argued by some researchers that FE-based global DIC offers better accuracy because of the enforced continuity between element nodes. We propose a detailed performance comparison between these different DIC algorithms both in terms of measurement accuracy and computational efficiency. Then, by measuring displacements of the same calculation points using the same calculation algorithms (e.g., correlation criterion, initial guess estimation, subpixel interpolation, optimization algorithm and convergence conditions) and identical calculation parameters (e.g., subset or element size), the performances of subset-based local DIC and two FE-based global DIC approaches are carefully compared in terms of measurement error and computational efficiency using both numerical tests and real experiments. A detailed examination of the experimental results reveals that, when subset (element) size is not very small and the local deformation within a subset (element) can be well approximated by the shape function used, standard subset-based local DIC approach not only provides better results in measured displacements, but also demonstrates much higher computation efficiency. However, several special merits of FE-based global DIC approaches are indicated. 1. Non-suppressive regulatory T cell subset expansion in pulmonary arterial hypertension. Science.gov (United States) Sada, Yoshiharu; Dohi, Yoshihiro; Uga, Sayuri; Higashi, Akifumi; Kinoshita, Hiroki; Kihara, Yasuki 2016-08-01 Regulatory T cells (Tregs) have been reported to play a pivotal role in the vascular remodeling of pulmonary arterial hypertension (PAH). Recent studies have revealed that Tregs are heterogeneous and can be characterized by three phenotypically and functionally different subsets. In this study, we investigated the roles of Treg subsets in the pathogenesis of PAH in eight patients with PAH and 14 healthy controls. Tregs and their subsets in peripheral blood samples were analyzed by flow cytometry. Treg subsets were defined as CD4(+)CD45RA(+)FoxP3(low) resting Tregs (rTregs), CD4(+)CD45RA(-)FoxP3(high) activated Tregs (aTregs), and CD4(+)CD45RA(-)FoxP3(low) non-suppressive Tregs (non-Tregs). The proportion of Tregs among CD4(+) T cells was significantly higher in PAH patients than in controls (6.54 ± 1.10 vs. 3.81 ± 0.28 %, p < 0.05). Of the three subsets, the proportion of non-Tregs was significantly elevated in PAH patients compared with controls (4.06 ± 0.40 vs. 2.79 ± 0.14 %, p < 0.01), whereas those of rTregs and aTregs were not different between the two groups. Moreover, the expression levels of cytotoxic T lymphocyte antigen 4, a functional cell surface molecule, in aTregs (p < 0.05) and non-Tregs (p < 0.05) were significantly higher in PAH patients compared with controls. These results suggested the non-Treg subset was expanded and functionally activated in peripheral lymphocytes obtained from IPAH patients. We hypothesize that immunoreactions involving the specific activation of the non-Treg subset might play a role in the vascular remodeling of PAH. 2. On Maximal Non-Disjoint Families of Subsets Directory of Open Access Journals (Sweden) Yu. A. Zuev 2017-01-01 Full Text Available The paper studies maximal non-disjoint families of subsets of a finite set. Non-disjointness means that any two subsets of a family have a nonempty intersection. The maximality is expressed by the fact that adding a new subset to the family cannot increase its power without violating a non-disjointness condition. Studying the properties of such families is an important section of the extreme theory of sets. Along with purely combinatorial interest, the problems considered here play an important role in informatics, anti-noise coding, and cryptography.In 1961 this problem saw the light of day in the Erdos, Ko and Rado paper, which established a maximum power of the non-disjoint family of subsets of equal power. In 1974 the Erdos and Claytman publication estimated the number of maximal non-disjoint families of subsets without involving the equality of their power. These authors failed to establish an asymptotics of the logarithm of the number of such families when the power of a basic finite set tends to infinity. However, they suggested such an asymptotics as a hypothesis. A.D. Korshunov in two publications in 2003 and 2005 established the asymptotics for the number of non-disjoint families of the subsets of arbitrary powers without maximality condition of these families.The basis for the approach used in the paper to study the families of subsets is their description in the language of Boolean functions. A one-to-one correspondence between a family of subsets and a Boolean function is established by the fact that the characteristic vectors of subsets of a family are considered to be the unit sets of a Boolean function. The main theoretical result of the paper is that the maximal non-disjoint families are in one-to-one correspondence with the monotonic self-dual Boolean functions. When estimating the number of maximal non-disjoint families, this allowed us to use the result of A.A. Sapozhenko, who established the asymptotics of the number of the 3. Physics and quality assurance for high dose rate brachytherapy International Nuclear Information System (INIS) Anderson, Lowell L. 1995-01-01 Purpose: To review the physical aspects of high dose rate (HDR) brachytherapy, including commissioning and quality assurance, source calibration and dose distribution measurements, and treatment planning methods. Following the introduction of afterloading in brachytherapy, development efforts to make it 'remote' culminated in 1964 with the near-simultaneous appearance of remote afterloaders in five major medical centers. Four of these machines were 'high dose rate', three employing 60Co and one (the GammaMed) using a single, cable-mounted 192Ir source. Stepping-motor source control was added to the GammaMed in 1974, making it the precursor of modern remote afterloaders, which are now suitable for interstitial as well as intracavitary brachytherapy by virtue of small source-diameter and indexer-accessed multiple channels. Because the 192Ir sources currently used in HDR remote afterloaders are supplied at a nominal air-kerma strength of 11.4 cGy cm2 s-1 (10 Ci), are not collimated in clinical use, and emit a significant fraction (15%) of photons at energies greater than 600 keV, shielding and facility design must be undertaken as carefully and thoroughly as for external beam installations. Licensing requirements of regulatory agencies must be met with respect both to maximum permissible dose limits and to the existence and functionality of safety devices (door interlocks, radiation monitors, etc.). Commissioning and quality assurance procedures that must be documented for HDR remote afterloading relate to (1) machine, applicator, guide-tube, and facility functionality checks, (2) source calibration, (3) emergency response readiness, (4) planning software evaluation, and (5) independent checks of clinical dose calculations. Source calibration checks must be performed locally, either by in-air measurement of air kerma strength or with a well ionization chamber calibrated (by an accredited standards laboratory) against an in-air measurement of air kerma strength for the 4. Twin pregnancy possibly associated with high semen quality DEFF Research Database (Denmark) Asklund, Camilla; Jensen, Tina Kold; Jørgensen, Niels 2007-01-01 BACKGROUND: Recent studies found an association between a long waiting time to pregnancy (TTP) and reduced probability of twinning and a reduced dizygotic (DZ) twinning rate in subfertile men. However, it remains unsolved whether semen quality is associated with twin offspring. We therefore studied...... the semen quality in a group of fathers of naturally conceived twins. METHODS: In this study, 37 fathers of DZ twins and 15 fathers of monozygotic (MZ) twins participated, and 349 normal fertile men served as a reference group. All men delivered a semen sample, underwent a physical examination and completed...... points higher than the reference group (P semen quality than the reference group, which supports... 5. High quality ceramic coatings sprayed by high efficiency hypersonic plasma spraying gun International Nuclear Information System (INIS) Zhu Sheng; Xu Binshi; Yao JiuKun 2005-01-01 This paper introduced the structure of the high efficiency hypersonic plasma spraying gun and the effects of hypersonic plasma jet on the sprayed particles. The optimised spraying process parameters for several ceramic powders such as Al 2 O 3 , Cr 2 O 3 , ZrO 2 , Cr 3 C 2 and Co-WC were listed. The properties and microstructure of the sprayed ceramic coatings were investigated. Nano Al 2 O 3 -TiO 2 ceramic coating sprayed by using the high efficiency hypersonic plasma spraying was also studied. Compared with the conventional air plasma spraying, high efficiency hypersonic plasma spraying improves greatly the ceramic coatings quality but at low cost. (orig.) 6. Hippo Experiment Data Access and Subseting System Science.gov (United States) Krassovski, M.; Hook, L.; Boden, T. 2014-12-01 HIAPER Pole-to-Pole Observations (HIPPO) was an NSF- and NOAA-funded, multi-year global airborne research project to survey the latitudinal and vertical distribution of greenhouse and related gases, and aerosols. Project scientists and support staff flew five month-long missions over the Pacific Basin on the NSF/NCAR Gulfstream V, High-performance Instrumented Airborne Platform for Environmental Research (HIAPER) aircraft between January 2009 and September 2011, spread throughout the annual cycle, from the surface to 14 km in altitude, and from 87°N to 67°S. Data from the HIPPO study of greenhouse gases and aerosols are now available to the atmospheric research community and the public. This comprehensive dataset provides the first high-resolution vertically resolved measurements of over 90 unique atmospheric species from nearly pole-to-pole over the Pacific Ocean across all seasons. The suite of atmospheric trace gases and aerosols is pertinent to understanding the carbon cycle and challenging global climate models. This dataset will provide opportunities for research across a broad spectrum of Earth sciences, including those analyzing the evolution in time and space of the greenhouse gases that affect global climate. The Carbon Dioxide Information Analysis Center (CDIAC) at Oak Ridge National Laboratory (ORNL) provides data management support for the HIPPO experiment including long-term data storage and dissemination. CDIAC has developed a relational database to house HIPPO merged 10-second meteorology, atmospheric chemistry, and aerosol data. This data set provides measurements from all Missions, 1 through 5, that took place from January of 2009 to September 2011. This presentation introduces newly build database and web interface, reflects the present state and functionality of the HIPPO Database and Exploration System as well as future plans for expansion and inclusion of combined discrete flask and GC sample GHG, Halocarbon, and hydrocarbon data. 7. Safe and high quality food production using low quality waters and improved irrigation systems and management DEFF Research Database (Denmark) Plauborg, Finn; Andersen, Mathias Neumann; Liu, Fulai 2010-01-01 uneven irrigation patterns can increase the water use efficiency as well as the quality of vegetable crops. Furthermore, recent innovations in the water treatment and irrigation industry have shown potential for the use of low quality water resources, such as reclaimed water or surface water in peri... 8. High throughput electrospinning of high-quality nanofibers via an aluminum disk spinneret Science.gov (United States) Zheng, Guokuo In this work, a simple and efficient needleless high throughput electrospinning process using an aluminum disk spinneret with 24 holes is described. Electrospun mats produced by this setup consisted of fine fibers (nano-sized) of the highest quality while the productivity (yield) was many times that obtained from conventional single-needle electrospinning. The goal was to produce scaled-up amounts of the same or better quality nanofibers under variable concentration, voltage, and the working distance than those produced with the single needle lab setting. The fiber mats produced were either polymer or ceramic (such as molybdenum trioxide nanofibers). Through experimentation the optimum process conditions were defined to be: 24 kilovolt, a distance to collector of 15cm. More diluted solutions resulted in smaller diameter fibers. Comparing the morphologies of the nanofibers of MoO3 produced by both the traditional and the high throughput set up it was found that they were very similar. Moreover, the nanofibers production rate is nearly 10 times than that of traditional needle electrospinning. Thus, the high throughput process has the potential to become an industrial nanomanufacturing process and the materials processed by it may be used as filtration devices, in tissue engineering, and as sensors. 9. Improving the quality of pork and pork products for the consumer : development of innovative, integrated, and sustainable food production chains of high quality pork products matching consumer demands NARCIS (Netherlands) Heimann, B.; Christensen, M.; Rosendal Rasmussen, S.; Bonneau, M.; Grunert, K.G.; Arnau, J.; Trienekens, J.H.; Oksbjerg, N.; Greef, de K.H.; Petersen, B. 2012-01-01 Improving the quality of pork and pork products for the consumer: development of innovative, integrated, and sustainable food production chains of high quality pork products matching consumer demands. 10. Regulation of EMMPRIN (CD147) on monocyte subsets in patients with symptomatic coronary artery disease. Science.gov (United States) Sturhan, Henrik; Ungern-Sternberg, Saskia N I v; Langer, Harald; Gawaz, Meinrad; Geisler, Tobias; May, Andreas E; Seizer, Peter 2015-06-01 The role of individual monocyte subsets in inflammatory cardiovascular diseases is insufficiently understood. Although the Extracellular Matrix Metalloproteinase Inducer (EMMPRIN) regulates important processes for inflammation such as MMP-release, its expression and regulation on monocyte subsets has not been characterized. In this clinical study, blood was obtained from 80 patients with stable coronary artery disease (CAD), 49 with acute myocardial infarction (AMI) and 34 healthy controls. Monocytes were divided into 3 subsets: CD14(++)CD16(-) (low), CD14(++)CD16(+) (intermediate), CD14(+)CD16(++) (high) according to phenotypic markers analyzed by flow cytometry. Surface expression of EMMPRIN was evaluated and compared with CD36 and CD47 expression. In all patients, EMMPRIN expression was significantly different among monocyte subsets with the highest expression on "classical" CD14(++)CD16(-) monocytes. EMMPRIN was upregulated on all monocyte subsets in patients with AMI as compared to patients with stable CAD. Notably, neither CD47 nor CD36 revealed a significant difference in patients with AMI compared to patients with stable CAD. EMMPRIN could serve as a marker for classical monocytes, which is upregulated in patients with acute myocardial infarction. Copyright © 2015 Elsevier Ltd. All rights reserved. 11. Comprehensive Ocean - Atmosphere Data Set (COADS) LMRF Arctic Subset Data.gov (United States) National Aeronautics and Space Administration — The Comprehensive Ocean - Atmosphere Data Set (COADS) LMRF Arctic subset contains marine surface weather reports for the region north of 65 degrees N from ships,... 12. Subset Statistics in the linear IV regression model NARCIS (Netherlands) Kleibergen, F.R. 2005-01-01 We show that the limiting distributions of subset generalizations of the weak instrument robust instrumental variable statistics are boundedly similar when the remaining structural parameters are estimated using maximum likelihood. They are bounded from above by the limiting distributions which 13. High Yield Chemical Vapor Deposition Growth of High Quality Large-Area AB Stacked Bilayer Graphene Science.gov (United States) Liu, Lixin; Zhou, Hailong; Cheng, Rui; Yu, Woo Jong; Liu, Yuan; Chen, Yu; Shaw, Jonathan; Zhong, Xing; Huang, Yu; Duan, Xiangfeng 2012-01-01 Bernal stacked (AB stacked) bilayer graphene is of significant interest for functional electronic and photonic devices due to the feasibility to continuously tune its band gap with a vertical electrical field. Mechanical exfoliation can be used to produce AB stacked bilayer graphene flakes but typically with the sizes limited to a few micrometers. Chemical vapor deposition (CVD) has been recently explored for the synthesis of bilayer graphene but usually with limited coverage and a mixture of AB and randomly stacked structures. Herein we report a rational approach to produce large-area high quality AB stacked bilayer graphene. We show that the self-limiting effect of graphene growth on Cu foil can be broken by using a high H2/CH4 ratio in a low pressure CVD process to enable the continued growth of bilayer graphene. A high temperature and low pressure nucleation step is found to be critical for the formation of bilayer graphene nuclei with high AB stacking ratio. A rational design of a two-step CVD process is developed for the growth of bilayer graphene with high AB stacking ratio (up to 90 %) and high coverage (up to 99 %). The electrical transport studies demonstrated that devices made of the as-grown bilayer graphene exhibit typical characteristics of AB stacked bilayer graphene with the highest carrier mobility exceeding 4,000 cm2/V·s at room temperature, comparable to that of the exfoliated bilayer graphene. PMID:22906199 14. A Quantum Approach to Subset-Sum and Similar Problems OpenAIRE 2017-01-01 In this paper, we study the subset-sum problem by using a quantum heuristic approach similar to the verification circuit of quantum Arthur-Merlin games. Under described certain assumptions, we show that the exact solution of the subset sum problem my be obtained in polynomial time and the exponential speed-up over the classical algorithms may be possible. We give a numerical example and discuss the complexity of the approach and its further application to the knapsack problem. 15. Estimation of functional failure probability of passive systems based on subset simulation method International Nuclear Information System (INIS) Wang Dongqing; Wang Baosheng; Zhang Jianmin; Jiang Jing 2012-01-01 In order to solve the problem of multi-dimensional epistemic uncertainties and small functional failure probability of passive systems, an innovative reliability analysis algorithm called subset simulation based on Markov chain Monte Carlo was presented. The method is found on the idea that a small failure probability can be expressed as a product of larger conditional failure probabilities by introducing a proper choice of intermediate failure events. Markov chain Monte Carlo simulation was implemented to efficiently generate conditional samples for estimating the conditional failure probabilities. Taking the AP1000 passive residual heat removal system, for example, the uncertainties related to the model of a passive system and the numerical values of its input parameters were considered in this paper. And then the probability of functional failure was estimated with subset simulation method. The numerical results demonstrate that subset simulation method has the high computing efficiency and excellent computing accuracy compared with traditional probability analysis methods. (authors) 16. Stochastic subset selection for learning with kernel machines. Science.gov (United States) Rhinelander, Jason; Liu, Xiaoping P 2012-06-01 Kernel machines have gained much popularity in applications of machine learning. Support vector machines (SVMs) are a subset of kernel machines and generalize well for classification, regression, and anomaly detection tasks. The training procedure for traditional SVMs involves solving a quadratic programming (QP) problem. The QP problem scales super linearly in computational effort with the number of training samples and is often used for the offline batch processing of data. Kernel machines operate by retaining a subset of observed data during training. The data vectors contained within this subset are referred to as support vectors (SVs). The work presented in this paper introduces a subset selection method for the use of kernel machines in online, changing environments. Our algorithm works by using a stochastic indexing technique when selecting a subset of SVs when computing the kernel expansion. The work described here is novel because it separates the selection of kernel basis functions from the training algorithm used. The subset selection algorithm presented here can be used in conjunction with any online training technique. It is important for online kernel machines to be computationally efficient due to the real-time requirements of online environments. Our algorithm is an important contribution because it scales linearly with the number of training samples and is compatible with current training techniques. Our algorithm outperforms standard techniques in terms of computational efficiency and provides increased recognition accuracy in our experiments. We provide results from experiments using both simulated and real-world data sets to verify our algorithm. 17. Consumers' expected quality and intention to purchase high quality pork meat. Science.gov (United States) Papanagiotou, P; Tzimitra-Kalogianni, I; Melfou, K 2013-03-01 Expected quality is believed to be one of the most important factors that influence consumers' intention to purchase food. The present study seeks to explore the concept of pork meat expected quality and compare it with self-stated consumer intention to purchase pork meat. The aim is attempted by means of a field research conducted in Greece, following a conjoint analytic procedure. Results show that quality expectations comply with intention to buy pork, in many aspects. However, several differences have been identified. More specifically, country of origin and marbling appear to be more important for respondents' purchase decisions than they are for their quality evaluations, while the opposite appears to be true for price. Finally, socio-demographic factors such as gender, level of education, place of purchase and consumption habits seem to influence perceptions. Copyright © 2012 Elsevier Ltd. All rights reserved. 18. Safe and High Quality Food Production using Low Quality Waters and Improved Irrigation Systems and Management, EU Project DEFF Research Database (Denmark) Plauborg, Finn; Jensen, Christian Richardt; Dalsgaard, Anders 2009-01-01 : the safety and quality of food products, and the increasing competition for clean freshwater. SAFIR is funded for the period 2005-2009 under the Food Quality and Safety thematic area of the EU 6th Framework Research Programme. The challenge for the next years will be to produce safe and high quality foods...... a multi-disciplinary team, with food safety and quality experts, engineers, agronomists and economists from17 research institutes and private companies in Europe, Israel and China working together. The project assesses potential risks to farmers. Coupled with farm management and economic models, a new...... intelligent tool for efficient and safe use and re-use of low-quality water are being developed. Already published results indicate water saving in the order of 25-30% in agricultural crops as potatoes and tomatoes are possible without yield reduction. Slightly treated waste water can be used safely when... 19. Application of subset simulation methods to dynamic fault tree analysis International Nuclear Information System (INIS) Liu Mengyun; Liu Jingquan; She Ding 2015-01-01 Although fault tree analysis has been implemented in the nuclear safety field over the past few decades, it was recently criticized for the inability to model the time-dependent behaviors. Several methods are proposed to overcome this disadvantage, and dynamic fault tree (DFT) has become one of the research highlights. By introducing additional dynamic gates, DFT is able to describe the dynamic behaviors like the replacement of spare components or the priority of failure events. Using Monte Carlo simulation (MCS) approach to solve DFT has obtained rising attention, because it can model the authentic behaviors of systems and avoid the limitations in the analytical method. In this paper, it provides an overview and MCS information for DFT analysis, including the sampling of basic events and the propagation rule for logic gates. When calculating rare-event probability, large amount of simulations in standard MCS are required. To improve the weakness, subset simulation (SS) approach is applied. Using the concept of conditional probability and Markov Chain Monte Carlo (MCMC) technique, the SS method is able to accelerate the efficiency of exploring the failure region. Two cases are tested to illustrate the performance of SS approach, and the numerical results suggest that it gives high efficiency when calculating complicated systems with small failure probabilities. (author) 20. Quality assessment and potential utilization of high amylolytic ... African Journals Online (AJOL) STORAGESEVER 2008-12-03 Dec 3, 2008 ... This study was carried out to compare the qualities of two acclaimed Nigerian amylolytic maize cultivars; SPMAT ... growing demand for usage as a gluten-free cereal. (Sweeny ... The grain was malted and the malting loss was calculated using the .... also confirmed that there was no significant difference in. 1. Quality Alternative Certification Programs in Special Education Ensure High Retention Science.gov (United States) Karge, Belinda D.; McCabe, Marjorie 2014-01-01 Market driven alternative routes to teaching have evolved into a quality program option and not just an answer to the teacher shortage. Alternative certification is a viable means of recruiting, training, and certifying those who have a bachelor's degree and a strong desire to enter the field of teaching. California has been a leader in the… 2. Innovation in Business Education: Developing a High Quality Online MBA Science.gov (United States) Roe, C. William; Toma, Alfred G.; Yallapragada, RamMohan R. 2015-01-01 Online degree programs were probably pioneered by for-profit universities such as University of Phoenix. Many online degree programs were initially considered low quality academic programs compared to traditional programs. Therefore, many public and private universities were slow to adopt the online programs. However, gradually more and more… 3. Quality of life of people living with HIV/AIDS and on highly active ... African Journals Online (AJOL) Quality of life of people living with HIV/AIDS and on highly active ... the Short Form-36 Health Survey (SF-36) was used to measure quality of life among patients on ... instrument for use in Ethiopia to assess the quality of life of people living with ... 4. Does High School Facility Quality Affect Student Achievement? A Two-Level Hierarchical Linear Model Science.gov (United States) Bowers, Alex J.; Urick, Angela 2011-01-01 The purpose of this study is to isolate the independent effects of high school facility quality on student achievement using a large, nationally representative U.S. database of student achievement and school facility quality. Prior research on linking school facility quality to student achievement has been mixed. Studies that relate overall… 5. Manufacturing High-Quality Carbon Nanotubes at Lower Cost Science.gov (United States) Benavides, Jeanette M.; Lidecker, Henning 2004-01-01 A modified electric-arc welding process has been developed for manufacturing high-quality batches of carbon nanotubes at relatively low cost. Unlike in some other processes for making carbon nanotubes, metal catalysts are not used and, consequently, it is not necessary to perform extensive cleaning and purification. Also, unlike some other processes, this process is carried out at atmospheric pressure under a hood instead of in a closed, pressurized chamber; as a result, the present process can be implemented more easily. Although the present welding-based process includes an electric arc, it differs from a prior electric-arc nanotube-production process. The welding equipment used in this process includes an AC/DC welding power source with an integral helium-gas delivery system and circulating water for cooling an assembly that holds one of the welding electrodes (in this case, the anode). The cathode is a hollow carbon (optionally, graphite) rod having an outside diameter of 2 in. (approximately equal to 5.1 cm) and an inside diameter of 5/8 in. (approximately equal to 1.6 cm). The cathode is partly immersed in a water bath, such that it protrudes about 2 in. (about 5.1 cm) above the surface of the water. The bottom end of the cathode is held underwater by a clamp, to which is connected the grounding cable of the welding power source. The anode is a carbon rod 1/8 in. (approximately equal to 0.3 cm) in diameter. The assembly that holds the anode includes a thumbknob- driven mechanism for controlling the height of the anode. A small hood is placed over the anode to direct a flow of helium downward from the anode to the cathode during the welding process. A bell-shaped exhaust hood collects the helium and other gases from the process. During the process, as the anode is consumed, the height of the anode is adjusted to maintain an anode-to-cathode gap of 1 mm. The arc-welding process is continued until the upper end of the anode has been lowered to a specified height 6. Numerical determination of injector design for high beam quality International Nuclear Information System (INIS) Boyd, J.K. 1985-01-01 The performance of a free electron laser strongly depends on the electron beam quality or brightness. The electron beam is transported into the free electron laser after it has been accelerated to the desired energy. Typically the maximum beam brightness produced by an accelerator is constrained by the beam brightness deliverd by the accelerator injector. Thus it is important to design the accelerator injector to yield the required electron beam brightness. The DPC (Darwin Particle Code) computer code has been written to numerically model accelerator injectors. DPC solves for the transport of a beam from emission through acceleration up to the full energy of the injector. The relativistic force equation is solved to determine particle orbits. Field equations are solved for self consistent electric and magnetic fields in the Darwin approximation. DPC has been used to investigate the beam quality consequences of A-K gap, accelerating stress, electrode configuration and axial magnetic field profile 7. 'Radiooncology 2007: High quality of life, the ultimate goal'. Abstracts International Nuclear Information System (INIS) 2007-01-01 The 13th annual congress of Deutsche Gesellschaft fuer Radioonkologie e.V. discussed clinical radiotherapy, medical physics and radiation biology, as well as an extensive programme of working groups on radiology in medical technology, care, and office. The focus was on new equipment (from image-supported to image-guided radiotherapy) and new approaches of targeted tumour therapy. Aspects of life quality are coming to the fore and will get even more important in the future. (orig.) 8. High image quality sub 100 picosecond gated framing camera development International Nuclear Information System (INIS) Price, R.H.; Wiedwald, J.D. 1983-01-01 A major challenge for laser fusion is the study of the symmetry and hydrodynamic stability of imploding fuel capsules. Framed x-radiographs of 10-100 ps duration, excellent image quality, minimum geometrical distortion (< 1%), dynamic range greater than 1000, and more than 200 x 200 pixels are required for this application. Recent progress on a gated proximity focused intensifier which meets these requirements is presented 9. The High Flying Leadership Qualities: What Matters the Most Science.gov (United States) 2016-04-01 operational and functional background of its personnel steeped in both DoD and industry experience. Research Methodology Based on their experiences...DeLorean Motors, Levitz Furniture , Enron, and many other corporations like these learned what happens when key leadership qualities lose all lift...graduate of the Canadian Force Command and Staff College in Toronto, Ontario, Canada; and the U.S. Army War College in Carlisle Barracks 10. Water Quality: A Field-Based Quality Testing Program for Middle Schools and High Schools. Science.gov (United States) Massachusetts State Water Resources Authority, Boston. This manual contains background information, lesson ideas, procedures, data collection and reporting forms, suggestions for interpreting results, and extension activities to complement a water quality field testing program. Information on testing water temperature, water pH, dissolved oxygen content, biochemical oxygen demand, nitrates, total… 11. Real-World Solutions for Developing High-Quality PHP Frameworks and Applications CERN Document Server Bergmann, Sebastian 2011-01-01 Learn to develop high-quality applications and frameworks in PHP Packed with in-depth information and step-by-step guidance, this book escorts you through the process of creating, maintaining and extending sustainable software of high quality with PHP. World-renowned PHP experts present real-world case studies for developing high-quality applications and frameworks in PHP that can easily be adapted to changing business requirements. . They offer different approaches to solving  typical development and quality assurance problems that every developer needs to know and master.Details the process 12. Idea on patent ; It is high time to stress quality International Nuclear Information System (INIS) 2012-03-01 This book deals with patent stressing on the quality, which includes from idea to technical business, It's simple to register the computer program, why do patent lawyer appoint the patent attorney's office? construction of patent right range, a good patent and a bad patent, strong patent and weak patent. It doesn't allow for Dus to use as we like, each patent has different value, Let's write technical specifications, advice on talking for invention with a patent attorney's office and what kind of task do intellectual property division do? 13. High-quality PWO crystals for the PANDA-EMC International Nuclear Information System (INIS) Novotny, R W; Bremer, D; Dormenev, V; Drexler, P; Eissner, T; Kuske, T; Moritz, M 2011-01-01 The paper provides a status report on the crystal production and quality control of a major part of the PbWO 4 crystals for the PANDA-EMC. The results confirm the excellent performance of the new generation of PWO-II. The mechanism of stimulated recovery provides an additional tool to recover radiation damage at room and low temperatures by applying an external infrared light source. Even on-line recovery can be considered if the photo sensor is insensitive in that particular wavelength region. 14. Measurement system with high accuracy for laser beam quality. Science.gov (United States) Ke, Yi; Zeng, Ciling; Xie, Peiyuan; Jiang, Qingshan; Liang, Ke; Yang, Zhenyu; Zhao, Ming 2015-05-20 Presently, most of the laser beam quality measurement system collimates the optical path manually with low efficiency and low repeatability. To solve these problems, this paper proposed a new collimated method to improve the reliability and accuracy of the measurement results. The system accuracy controlled the position of the mirror to change laser beam propagation direction, which can realize the beam perpendicularly incident to the photosurface of camera. The experiment results show that the proposed system has good repeatability and the measuring deviation of M2 factor is less than 0.6%. 15. Rehosting of Bacterial Chaperones for High-Quality Protein Production▿ Science.gov (United States) Martínez-Alonso, Mónica; Toledo-Rubio, Verónica; Noad, Rob; Unzueta, Ugutz; Ferrer-Miralles, Neus; Roy, Polly; Villaverde, Antonio 2009-01-01 Coproduction of DnaK/DnaJ in Escherichia coli enhances solubility but promotes proteolytic degradation of their substrates, minimizing the yield of unstable polypeptides. Higher eukaryotes have orthologs of DnaK/DnaJ but lack the linked bacterial proteolytic system. By coexpression of DnaK and DnaJ in insect cells with inherently misfolding-prone recombinant proteins, we demonstrate simultaneous improvement of soluble protein yield and quality and proteolytic stability. Thus, undesired side effects of bacterial folding modulators can be avoided by appropriate rehosting in heterologous cell expression systems. PMID:19820142 16. Subset-sum phase transitions and data compression Science.gov (United States) Merhav, Neri 2011-09-01 We propose a rigorous analysis approach for the subset-sum problem in the context of lossless data compression, where the phase transition of the subset-sum problem is directly related to the passage between ambiguous and non-ambiguous decompression, for a compression scheme that is based on specifying the sequence composition. The proposed analysis lends itself to straightforward extensions in several directions of interest, including non-binary alphabets, incorporation of side information at the decoder (Slepian-Wolf coding), and coding schemes based on multiple subset sums. It is also demonstrated that the proposed technique can be used to analyze the critical behavior in a more involved situation where the sequence composition is not specified by the encoder. 17. Probabilistic quantum cloning of a subset of linearly dependent states Science.gov (United States) Rui, Pinshu; Zhang, Wen; Liao, Yanlin; Zhang, Ziyun 2018-02-01 It is well known that a quantum state, secretly chosen from a certain set, can be probabilistically cloned with positive cloning efficiencies if and only if all the states in the set are linearly independent. In this paper, we focus on probabilistic quantum cloning of a subset of linearly dependent states. We show that a linearly-independent subset of linearly-dependent quantum states {| Ψ 1⟩,| Ψ 2⟩,…,| Ψ n ⟩} can be probabilistically cloned if and only if any state in the subset cannot be expressed as a linear superposition of the other states in the set {| Ψ 1⟩,| Ψ 2⟩,…,| Ψ n ⟩}. The optimal cloning efficiencies are also investigated. 18. Health-Related Quality of Life in Children with High-Functioning Autism Science.gov (United States) Potvin, Marie-Christine; Snider, Laurie; Prelock, Patricia A.; Wood-Dauphinee, Sharon; Kehayia, Eva 2015-01-01 The health-related quality of life of school-aged children with high-functioning autism is poorly understood. The objectives of this study were to compare the health-related quality of life of children with high-functioning autism to that of typically developing peers and to compare child-self and parent-proxy reports of health-related quality of… 19. Publishing high-quality climate data on the semantic web Science.gov (United States) Woolf, Andrew; Haller, Armin; Lefort, Laurent; Taylor, Kerry 2013-04-01 20. Investigating evolutionary conservation of dendritic cell subset identity and functions Directory of Open Access Journals (Sweden) Thien-Phong eVu Manh 2015-06-01 Full Text Available Dendritic cells (DC were initially defined as mononuclear phagocytes with a dendritic morphology and an exquisite efficiency for naïve T cell activation. DC encompass several subsets initially identified by their expression of specific cell surface molecules and later shown to excel in distinct functions and to develop under the instruction of different transcription factors or cytokines. Very few cell surface molecules are expressed in a specific manner on any immune cell type. Hence, to identify cell types, the sole use of a small number of cell surface markers in classical flow cytometry can be deceiving. Moreover, the markers currently used to define mononuclear phagocyte subsets vary depending on the tissue and animal species studied and even between laboratories. This has led to confusion in the definition of DC subset identity and in their attribution of specific functions. There is a strong need to identify a rigorous and consensus way to define mononuclear phagocyte subsets, with precise guidelines potentially applicable throughout tissues and species. We will discuss the advantages, drawbacks and complementarities of different methodologies: cell surface phenotyping, ontogeny, functional characterization and molecular profiling. We will advocate that gene expression profiling is a very rigorous, largely unbiased and accessible method to define the identity of mononuclear phagocyte subsets, which strengthens and refines surface phenotyping. It is uniquely powerful to yield new, experimentally testable, hypotheses on the ontogeny or functions of mononuclear phagocyte subsets, their molecular regulation and their evolutionary conservation. We propose defining cell populations based on a combination of cell surface phenotyping, expression analysis of hallmark genes and robust functional assays, in order to reach a consensus and integrate faster the huge but scattered knowledge accumulated by different laboratories on different cell types 1. Quality Improvement, Inventory Management, Lead Time Reduction and Production Scheduling in High-Mix Manufacturing Environments Science.gov (United States) 2017-01-13 Quality Improvement , Inventory Management, Lead Time Reduction and Production Scheduling in High-mix Manufacturing Environments by Sean Daigle B.S...Mechanical Engineering Chairman, Department Committee on Graduate Theses 2 Quality Improvement , Inventory Management, Lead Time Reduction and... Production Scheduling in High-mix Manufacturing Environments by Sean Daigle Submitted to the Department of Mechanical Engineering on January 13, 2017, in 2. Optimization of nitridation conditions for high quality inter-polysilicon dielectric layers NARCIS (Netherlands) Klootwijk, J.H.; Bergveld, H.J.; van Kranenburg, H.; Woerlee, P.H.; Wallinga, Hans 1996-01-01 Nitridation of deposited high temperature oxides (HTO) was studied to form high quality inter-polysilicon dielectric layers for embedded non volatile memories. Good quality dielectric layers were obtained earlier by using an optimized deposition of polysilicon and by performing a post-dielectric 3. Bridges to Excellence--recognizing high-quality care: analysis of physician quality and resource use. Science.gov (United States) Rosenthal, Meredith B; de Brantes, Francois S; Sinaiko, Anna D; Frankel, Matthew; Robbins, Russell D; Young, Sara 2008-10-01 4. On the size of the subset partial order DEFF Research Database (Denmark) Elmasry, Amr Ahmed Abd Elmoneim 2012-01-01 Given a family of k sets with cardinalities S 1,S 2,⋯, S k and N=Σ k i=1S i, we show that the size of the partial order graph induced by the subset relation (called the subset graph) is O(Σ si≤B 2si+N/lgN·Σ si>Blg(s i/B)), 2 where B=lg(N/lg 2N). This implies a simpler proof to the O(N 2/lg 2N... 5. Academic coordination at university: Strategies for high quality education Directory of Open Access Journals (Sweden) Maria del Mar Durán Bellonch 2009-08-01 Full Text Available Most lecturers and professors involved in teaching School Management and Education Management courses have been engaged in developing some innovative actions to improve the training quality that we offer to students in the Pedagogy degree at the Autonomous University of Barcelona. These actions are aimed at increasing co-ordination among the teaching staff when designing and implementing the course curricula. Co-ordination meetings, exchanges of teaching experiences, discussions about what, how and when to teach the different contents, in which courses and at what level, methodological issues pointed out through technical description cards or the elaboration of study cases to be solved have become the basis of relevant actions during the last academic years. This paper explains each one of them, and provides useful information about the theoretical background, the process carried out, some of the results obtained, the output and the tools created. 6. Possibilities of obtaining and controlling high-quality pressure castings Directory of Open Access Journals (Sweden) S. Pietrowski 2011-07-01 Full Text Available The paper presents the influence of the type of furnace charging melting, refining and modification silumins 226 and 231 on the porosity and microstructure of castings. It was shown that in order to reduce or eliminate the porosity of the castings is necessary to the refining ECOSAL-AL113 of liquid silumin both in the melting furnace, and in the ladle and an additional nitrogen, in the heat furnace modified and refining with nitrogen. To control the effects of refining and modifying the TDA method was used. It was found that based on crystal- lization curve can be qualitatively assess the gas porosity of the castings. In order to control and quality control silumins author developed a computer program using the method of TDA, which sets out: Rm, A5, HB and casting porosity P and the concentration of hydrogen in them. The program also informs the technological procedures to be performed for liquid silumin improper preparation. 7. Identification of an Immunogenic Subset of Metastatic Uveal Melanoma. Science.gov (United States) Rothermel, Luke D; Sabesan, Arvind C; Stephens, Daniel J; Chandran, Smita S; Paria, Biman C; Srivastava, Abhishek K; Somerville, Robert; Wunderlich, John R; Lee, Chyi-Chia R; Xi, Liqiang; Pham, Trinh H; Raffeld, Mark; Jailwala, Parthav; Kasoji, Manjula; Kammula, Udai S 2016-05-01 Uveal melanoma is a rare melanoma variant with no effective therapies once metastases develop. Although durable cancer regression can be achieved in metastatic cutaneous melanoma with immunotherapies that augment naturally existing antitumor T-cell responses, the role of these treatments for metastatic uveal melanoma remains unclear. We sought to define the relative immunogenicity of these two melanoma variants and determine whether endogenous antitumor immune responses exist against uveal melanoma. We surgically procured liver metastases from uveal melanoma (n = 16) and cutaneous melanoma (n = 35) patients and compared the attributes of their respective tumor cell populations and their infiltrating T cells (TIL) using clinical radiology, histopathology, immune assays, and whole-exomic sequencing. Despite having common melanocytic lineage, uveal melanoma and cutaneous melanoma metastases differed in their melanin content, tumor differentiation antigen expression, and somatic mutational profile. Immunologic analysis of TIL cultures expanded from these divergent forms of melanoma revealed cutaneous melanoma TIL were predominantly composed of CD8(+) T cells, whereas uveal melanoma TIL were CD4(+) dominant. Reactivity against autologous tumor was significantly greater in cutaneous melanoma TIL compared with uveal melanoma TIL. However, we identified TIL from a subset of uveal melanoma patients which had robust antitumor reactivity comparable in magnitude with cutaneous melanoma TIL. Interestingly, the absence of melanin pigmentation in the parental tumor strongly correlated with the generation of highly reactive uveal melanoma TIL. The discovery of this immunogenic group of uveal melanoma metastases should prompt clinical efforts to determine whether patients who harbor these unique tumors can benefit from immunotherapies that exploit endogenous antitumor T-cell populations. Clin Cancer Res; 22(9); 2237-49. ©2015 AACR. ©2015 American Association for Cancer Research. 8. Changes in the host lymphocyte subsets during chemical carcinogenesis International Nuclear Information System (INIS) Brodt, P.; Lala, P.K. 1983-01-01 Changes in small lymphocyte subsets in the lymphoid organs of young C3H mice were studied following i.m. injection of a carcinogenic dose of 3-methylcholanthrene (mc). Using monoclonal anti-Lyt antibodies and a sandwich radiolabeling method with 125 I-labeled rabbit anti-mouse Immunoglobulin, the lymphocyte subpopulations in the thymus, spleen, and draining lymph node were examined by radioautography. During the fifth week following the administration of the carcinogen a sharp decrease in the level of Ly-1,2+ small lymphocyte population in the thymus was noted which coincided with a considerable increase (10-fold) in the Ly-2+. During the same period, a similar increase in the Ly-2+ population was also observed in the draining. The high levels of Ly-2+ cells lasted for more than 4 weeks in the thymus while, in the draining node, they lasted for 2 weeks and dropped to normal levels (0 to 2%) simultaneously with the appearance of tumor cells identified in histological preparations. These systemic increases coincided with the appearance of macroscopic tumor nodules. The mixed lymphocyte reaction response of the draining node cells, but not of the spleen, was suppressed during the period of increased level of Ly-2+ cells. Furthermore, during this period, s.c. transplantation of a syngeneic mammary tumor in the same leg resulted in enhanced local growth as well as metastatic spread of the tumor to the lungs in mc treated mice. These findings suggest that a localized immunosuppression associated with the rise in the Ly-2+ cells may be of functional significance during carcinogen-induced tumor development 9. prediction of bread-making quality using size exclusion high African Journals Online (AJOL) ACSS Variation in the distribution of protein molecular weight in wheat (Triticum aestivum), influences ... with high G/G ratios in SDS-insoluble protein fraction were Sceptre x Nata and Kariega x Sceptre. ...... Molecular characterisation and dynamic. 10. High-quality single crystals for neutron experiments studies and our collaborative research projects with other UK and international groups will be discussed. Keywords. Crystal growth; floating zone method; neutron scattering. ... of single crystals of new materials is a highly competitive business. 11. Two Distinct Myeloid Subsets at the Term Human Fetal–Maternal Interface Directory of Open Access Journals (Sweden) Maria Laura Costa 2017-10-01 Full Text Available During pregnancy, immune cells infiltrate the placenta at different stages of fetal development. NK cells and macrophages are the most predominant cell types. These immune cells play pleiotropic roles, as they control spiral artery remodeling to ensure appropriate blood supply and maintain long-term tolerance to a true allograft; yet, they must be able to mount appropriate immune defenses to pathogens that may threaten the fetus. Whether the same cell type accomplishes all these tasks or if there are dedicated subsets remains controversial. Here, we identify and characterize two distinct subsets of myeloid cells that differ in their pro-inflammatory/regulatory capacity. While one subset predominantly produces the immune-modulating cytokine IL-10, the second subset has superior capacity to secrete pro-inflammatory mediators, such as IL-1β and IL-6. The putative regulatory myeloid cells also express high levels of inhibitory receptors and their ligands, including programmed cell death 1 (PD1 ligands. Importantly, a large fraction of CD8 and CD4 cells in normal term human placenta are PD1 positive, suggesting that the PD1/PD1 ligands axis might be critical to maintain tolerance during pregnancy. 12. T-Cell Subsets Predict Mortality in Malnourished Zambian Adults Initiating Antiretroviral Therapy. Directory of Open Access Journals (Sweden) Caroline C Chisenga Full Text Available To estimate the prognostic value of T-cell subsets in Zambian patients initiating antiretroviral therapy (ART, and to assess the impact of a nutritional intervention on T-cell subsets.This was a sub-study of a randomised clinical trial of a nutritional intervention for malnourished adults initiating ART. Participants in a randomised controlled trial (NUSTART trial were enrolled between April and December 2012. Participants received lipid-based nutritional supplement either with or without additional vitamins and minerals. Immunophenotyping was undertaken at baseline and, in survivors, after 12 weeks of ART to characterize T-cell subsets using the markers CD3, CD4, CD8, CD45RA, CCR7, CD28, CD57, CD31, α4β7, Ki67, CD25 and HLA-DR. Univariate and multivariate survival analysis was performed, and responses to treatment were analysed using the Wicoxon rank-sum test.Among 181 adults, 36 (20% died by 12 weeks after starting ART. In univariate analysis, patients who died had fewer proliferating, more naïve and fewer gut homing CD4+ T-cells compared to survivors; and more senescent and fewer proliferating CD8+ T-cells. In a multivariate Cox regression model high naïve CD4+, low proliferating CD4+, high senescent CD8+ and low proliferating CD8+ subsets were independently associated with increased risk of death. Recent CD4+ thymic emigrants increased less between recruitment and 12 weeks of ART in the intervention group compared to the control group.Specific CD4+ T-cell subsets are of considerable prognostic significance for patients initiating ART in Zambia, but only thymic output responded to this nutritional intervention. 13. Quality International Nuclear Information System (INIS) Burnett, N.; Jeffries, J.; Mach, J.; Robson, M.; Pajot, D.; Harrigan, J.; Lebsack, T.; Mullen, D.; Rat, F.; Theys, P. 1993-01-01 What is quality? How do you achieve it? How do you keep it once you have got it. The answer for industry at large is the three-step hierarchy of quality control, quality assurance and Total quality Management. An overview is given of the history of quality movement, illustrated with examples from Schlumberger operations, as well as the oil industry's approach to quality. An introduction of the Schlumberger's quality-associated ClientLink program is presented. 15 figs., 4 ills., 16 refs 14. Range-reference determination of lymphocyte subsets in Moroccan African Journals Online (AJOL) Centre de Transfusion Sanguine, Hôpital Militaire d'Instruction Med V Rabat, Morocco. 3. Service ... The percentage of CD3-CD56+ subsets was affected by smoking (p < 0.01). Our analysis .... lymphocyte subpopulations in non-smokers and. 15. Subset selection for an epsilon-best population : efficiency results NARCIS (Netherlands) Laan, van der P. 1991-01-01 An almost best or an \\epsilon-best population is defined as a population with location parameter on a distance not larger than \\epsilon (\\geq 0) from the best population (with largest value of the location parameter). For the subset selection tables with the relative efficiency of selecting an 16. T-lymphocyte subsets, thymic size and breastfeeding in infancy DEFF Research Database (Denmark) Jeppesen, Dorthe Lisbeth; Hasselbalch, Helle; Lisse, Ida M 2004-01-01 We followed the changes in concentration of T-lymphocyte subsets (CD4+ and CD8+ cells) in peripheral blood and thymus size during infancy. Previous studies have found increased thymus size in breastfed infants. The present study analyzed the association between breastfeeding and the number of CD4... 17. T lymphocyte subsets in prostate cancer subjects in south eastern ... African Journals Online (AJOL) Humoral and cellular mechanisms play roles in immune response to foreign antigens. The present study was designed to determine the T lymphocyte subsets (CD4 + T cells, CD8 + T cells and CD4/CD8 ratio) in the prostate cancer subjects and control subjects. CD4 + T cells (l/count) and CD8 + T cells (l/count) were ... 18. High quality actively cooled plasma-facing components for fusion International Nuclear Information System (INIS) Nygren, R.E. 1995-01-01 This paper interweaves some suggestions for developing actively cooled plasma-facing components (PFCs) for future fusion devices, with supporting examples taken from the design, fabrication and operation of Tore Supra's Phase III outboard pump limiter (OPL). This actively cooled midplane limiter, designed for heat and particle removal during long-pulse operation, has been operated under essentially thermally steady state conditions. Testing to identify braze flaws, analysis of the impact of joining flaws on the thermal-hydraulic performance of the OPL, and the extensive calorimetry and IR thermography used to confirm and update safe operating limits for power handling of the OPL are reviewed. This experience suggests that, for PFCs in future fusion devices, flaw-tolerant designs are possible; analyses of the impacts of flaws on performance can provide criteria for quality assurance; and validating appropriate methods of inspection for such flaws early in the design development of PFCs is prudent. The need for in-service monitoring is also discussed. (orig.) 19. Health care in small prisons: incorporating high-quality standards. Science.gov (United States) Rieder, Jean-Pierre; Casillas, Alejandra; Mary, Gérard; Secretan, Anne-Dominique; Gaspoz, Jean-Michel; Wolff, Hans 2013-01-01 In the past, health management in Geneva's six post-trial prisons had been variable and inconsistent. In 2008, the unit of penitentiary medicine of the Geneva University Hospitals was mandated to re-organize and provide health care at all six prison facilities. The specific aim of this paper is to outline the example as a practical solution to some of the common challenges in unifying the structure and process of health services across multiple small facilities, while meeting European prison health and local quality standards. Geneva's post-trial prisons are small and close to one another in geographical proximity - ideal conditions for the construction of a health mobile team (HMT). This multidisciplinary mobile team operated like a community ambulatory care model; it was progressively launched in all prison facilities in Geneva. The authors incorporated an implementation strategy where health providers partnered with prison and community stakeholders in the health delivery model's development and adaption process. The model's strategic initiatives are described along the following areas, in light of other international prison health activity and prior care models: access to a health care professional, equivalence of care, patient consent, confidentiality, humanitarian interventions, and professional competence and independence. From the perspective of the HMT members, the authors provide the "lessons learned" through this experience, especially to providers who are working on prison health services reform and coordination improvement. The paper particularly stresses the importance of partnering with community health stakeholders and prison staff, a key component to the approach. 20. A high-quality catalog of the Drosophila melanogaster proteome DEFF Research Database (Denmark) Brunner, Erich; Ahrens, Christian H.; Mohanty, Sonaly 2007-01-01 % of the predicted Drosophila melanogaster proteome by detecting 9,124 proteins from 498,000 redundant and 72,281 distinct peptide identifications. This unprecedented high proteome coverage for a complex eukaryote was achieved by combining sample diversity, multidimensional biochemical fractionation and analysis... 1. Bust economics: foragers choose high quality habitats in lean times Directory of Open Access Journals (Sweden) Sonny S. Bleicher 2016-01-01 Full Text Available In environments where food resources are spatially variable and temporarily impoverished, consumers that encounter habitat patches with different food density should focus their foraging initially where food density is highest before they move to patches where food density is lower. Increasing missed opportunity costs should drive individuals progressively to patches with lower food density as resources in the initially high food density patches deplete. To test these expectations, we assessed the foraging decisions of two species of dasyurid marsupials (dunnarts: Sminthopsis hirtipes and S. youngsoni during a deep drought, or bust period, in the Simpson Desert of central Australia. Dunnarts were allowed access to three patches containing different food densities using an interview chamber experiment. Both species exhibited clear preference for the high density over the lower food density patches as measured in total harvested resources. Similarly, when measuring the proportion of resources harvested within the patches, we observed a marginal preference for patches with initially high densities. Models analyzing behavioral choices at the population level found no differences in behavior between the two species, but models analyzing choices at the individual level uncovered some variation. We conclude that dunnarts can distinguish between habitat patches with different densities of food and preferentially exploit the most valuable. As our observations were made during bust conditions, experiments should be repeated during boom times to assess the foraging economics of dunnarts when environmental resources are high. 2. Whole high-quality light environment for humans and plants Science.gov (United States) Sharakshane, Anton 2017-11-01 Plants sharing a single light environment on a spaceship with a human being and bearing a decorative function should look as natural and attractive as possible. And consequently they can be illuminated only with white light with a high color rendering index. Can lighting optimized for a human eye be effective and appropriate for plants? Spectrum-based effects have been compared under artificial lighting of plants by high-pressure sodium lamps and general-purpose white LEDs. It has been shown that for the survey sample phytochrome photo-equilibria does not depend significantly on the parameters of white LED light, while the share of phytoactive blue light grows significantly as the color temperature increases. It has been revealed that yield photon flux is proportional to luminous efficacy and increases as the color temperature decreases, general color rendering index Ra and the special color rendering index R14 (green leaf) increase. General-purpose white LED lamps with a color temperature of 2700 K, Ra > 90 and luminous efficacy of 100 lm/W are as efficient as the best high-pressure sodium lamps, and at a higher luminous efficacy their yield photon flux per joule is even bigger in proportion. Here we show that demand for high color rendering white LED light is not contradictory to the agro-technical objectives. 3. The influence of knowledge management implementation toward the quality of high schools Directory of Open Access Journals (Sweden) Hilda Ainissyifa 2012-09-01 Full Text Available The purpose of this study is to analyze the influence of knowledge management implementation toward the quality of high schools. This study was conducted among five high schools under an education foundation. The analysis models used in the study were correlation analysis and t-test. The respondents were used as the profession references are 86 teachers. The result of the study shows that knowledge management implementation has a positive and significant influence toward the quality of high schools. 4. Achieving high-quality care: a view from NICE. Science.gov (United States) Leng, Gillian; Partridge, Gemma 2018-01-01 The National Institute for Health and Care Excellence (NICE) was established in 1999 to provide evidence-based guidance. The task of producing guidance by reviewing primary research data and using an advisory committee to develop evidence-based recommendations, is not straightforward. Guidance production is, however, less challenging than the task of putting evidence-based recommendations into practice.NICE is very sensitive to this challenge as, since 1999, over 1500 pieces of NICE guidance have been published. A number of pieces of guidance relate to heart disease, including pharmaceutical agents, new medical technologies and clinical guidelines. Examples include guidelines on acute heart failure and atrial fibrillation, and advice on technologies including edoxaban and implantable cardioverter defibrillators.The research evidence is clear that a change in practice rarely comes about as a result of simply disseminating guidance on best practice. Simple dissemination is particularly ineffective if the guidance has not been produced by a well-respected, credible organisation. It is also clear from the literature that implementation is more successful when more than one approach is taken, and when there is alignment between efforts at organisational, local and national levels.At an organisational level, there should be support from the Board for quality improvement, with ongoing measurement of progress. Resources should be provided for targeted change programmes, particularly where new guidance suggests improvements are required. A systematic process for putting change in place should include identifying barriers to change, agreeing interventions to overcome the barriers and drive forward improvement and planning for implementation and evaluation. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2018. All rights reserved. No commercial use is permitted unless otherwise expressly granted. 5. Modern High Technology Solutions for Quality and Longterm Vegetable Preservation International Nuclear Information System (INIS) Nacheva, I.; Miteva, D.; Todorov, Y.; Loginovska, K.; Tsvetkov, Ts. 2012-01-01 In the publication the authors present the results of the applying of two modern technologies for long term and safe vegetable preservation – freeze-drying and gamma sterilization. The freeze-dried vegetables feature minimum moisture – from 2 – 5% and taste-aroma complex preserved to the highest degree. The carried out gamma sterilization ensures a high microbial purity of the vegetables and guarantees for their long term preservation - up to 5 years in polymer packing, under usual conditions 6. High quality factor gigahertz frequencies in nanomechanical diamond resonators OpenAIRE Gaidarzhy, Alexei; Imboden, Matthias; Mohanty, Pritiraj; Rankin, Janet; Sheldon, Brian W. 2007-01-01 We report actuation and detection of gigahertz-range resonance frequencies in nano-crystalline diamond mechanical resonators. High order transverse vibration modes are measured in coupled-beam resonators exhibiting frequencies up to 1.441 GHz. The cantilever-array design of the resonators translates the gigahertz-range resonant motion of micron-long cantilever elements to the displacement of the central supporting structure. Use of nano-crystalline diamond further increases the frequency comp... 7. Differential adipokine receptor expression on circulating leukocyte subsets in lean and obese children. Directory of Open Access Journals (Sweden) Genoveva Keustermans 8. Phenotypic, ultra-structural, and functional characterization of bovine peripheral blood dendritic cell subsets. Directory of Open Access Journals (Sweden) Janet J Sei Full Text Available Dendritic cells (DC are multi-functional cells that bridge the gap between innate and adaptive immune systems. In bovine, significant information is lacking on the precise identity and role of peripheral blood DC subsets. In this study, we identify and characterize bovine peripheral blood DC subsets directly ex vivo, without further in vitro manipulation. Multi-color flow cytometric analysis revealed that three DC subsets could be identified. Bovine plasmacytoid DC were phenotypically identified by a unique pattern of cell surface protein expression including CD4, exhibited an extensive endoplasmic reticulum and Golgi apparatus, efficiently internalized and degraded exogenous antigen, and were the only peripheral blood cells specialized in the production of type I IFN following activation with Toll-like receptor (TLR agonists. Conventional DC were identified by expression of a different pattern of cell surface proteins including CD11c, MHC class II, and CD80, among others, the display of extensive dendritic protrusions on their plasma membrane, expression of very high levels of MHC class II and co-stimulatory molecules, efficient internalization and degradation of exogenous antigen, and ready production of detectable levels of TNF-alpha in response to TLR activation. Our investigations also revealed a third novel DC subset that may be a precursor of conventional DC that were MHC class II+ and CD11c-. These cells exhibited a smooth plasma membrane with a rounded nucleus, produced TNF-alpha in response to TLR-activation (albeit lower than CD11c+ DC, and were the least efficient in internalization/degradation of exogenous antigen. These studies define three bovine blood DC subsets with distinct phenotypic and functional characteristics which can be analyzed during immune responses to pathogens and vaccinations of cattle. 9. Toolsets for Airborne Data (TAD): Enhanced Airborne Data Merging Functionality through Spatial and Temporal Subsetting Science.gov (United States) Early, A. B.; Chen, G.; Beach, A. L., III; Northup, E. A. 2016-12-01 NASA has conducted airborne tropospheric chemistry studies for over three decades. These field campaigns have generated a great wealth of observations, including a wide range of the trace gases and aerosol properties. The Atmospheric Science Data Center (ASDC) at NASA Langley Research Center in Hampton Virginia originally developed the Toolsets for Airborne Data (TAD) web application in September 2013 to meet the user community needs for manipulating aircraft data for scientific research on climate change and air quality relevant issues. The analysis of airborne data typically requires data subsetting, which can be challenging and resource intensive for end users. In an effort to streamline this process, the TAD toolset enhancements will include new data subsetting features and updates to the current database model. These will include two subsetters: temporal and spatial, and vertical profile. The temporal and spatial subsetter will allow users to both focus on data from a specific location and/or time period. The vertical profile subsetter will retrieve data collected during an individual aircraft ascent or descent spiral. This effort will allow for the automation of the typically labor-intensive manual data subsetting process, which will provide users with data tailored to their specific research interests. The development of these enhancements will be discussed in this presentation. 10. Health-related quality of life in children with high-functioning autism. Science.gov (United States) Potvin, Marie-Christine; Snider, Laurie; Prelock, Patricia A; Wood-Dauphinee, Sharon; Kehayia, Eva 2015-01-01 The health-related quality of life of school-aged children with high-functioning autism is poorly understood. The objectives of this study were to compare the health-related quality of life of children with high-functioning autism to that of typically developing peers and to compare child-self and parent-proxy reports of health-related quality of life of children. A cross-sectional study of children with high-functioning autism (n = 30) and peers (n = 31) was conducted using the Pediatric Quality of Life Inventory 4.0 Generic Core Scales. Children with high-functioning autism had significantly poorer health-related quality of life than peers whether reported by themselves (p children and parental scores suggested variance in points of view. This study specifically investigated health-related quality of life in children with high-functioning autism as compared to a sample of peers, from the child's perspective. It strengthens earlier findings that children with high-functioning autism experience poorer health-related quality of life than those without this disorder and points to the importance of clinicians working with families to identify areas in a child's life that promote or hinder their sense of well-being. © The Author(s) 2013. 11. Subsetting Tools for Enabling Easy Access to International Airborne Chemistry Data Science.gov (United States) Northup, E. A.; Chen, G.; Quam, B. M.; Beach, A. L., III; Silverman, M. L.; Early, A. B. 2017-12-01 In response to the Research Opportunities in Earth and Space Science (ROSES) 2015 release announcement for Advancing Collaborative Connections for Earth System Science (ACCESS), researchers at NASA Langley Research Center (LaRC) proposed to extend the capabilities of the existing Toolsets for Airborne Data (TAD) to include subsetting functionality to allow for easier access to international airborne field campaign data. Airborne field studies are commonly used to gain a detailed understanding of atmospheric processes for scientific research on international climate change and air quality issues. To accommodate the rigorous process for manipulating airborne field study chemistry data, and to lessen barriers for researchers, TAD was created with the ability to geolocate data from various sources measured on different time scales from a single flight. The analysis of airborne chemistry data typically requires data subsetting, which can be challenging and resource-intensive for end users. In an effort to streamline this process, new data subsetting features and updates to the current database model will be added to the TAD toolset. These will include two subsetters: temporal and spatial, and vertical profile. The temporal and spatial subsetter will allow users to both focus on data from a specific location and/or time period. The vertical profile subsetter will retrieve data collected during an individual aircraft ascent or descent spiral. These new web-based tools will allow for automation of the typically labor-intensive manual data subsetting process, which will provide users with data tailored to their specific research interests. The system has been designed to allow for new in-situ airborne missions to be added as they become available, with only minor pre-processing required. The development of these enhancements will be discussed in this presentation. 12. High oleic sunflower biodiesel: quality control and different purification methods Directory of Open Access Journals (Sweden) Pighinelli, Anna L.M.T. 2011-06-01 Full Text Available The objective of the present work is to evaluate the production of biodiesel using ethanol and sunflower oil. The extraction of the sunflower oil was evaluated first. An experimental design was used to estimate the influence of the independent variables grain temperature (25º to 110ºC and expeller rotation (85 to 119rpm on the crude oil. The best result obtained was 68.38%, achieved with a rotation from 100 to 115rpm, grain temperature ranging from 25º to 30ºC and moisture content of around 7%. The next study consisted of transesterification, evaluating the influence of the ethanol, oil molar ratio and the catalyst concentration (sodium methylate on the ester-rich phase yield. The highest yield was 98.39% obtained with a molar ratio of 9:1 and 3% catalyst. An experiment was then carried out on a small reactor and the biodiesel produced was purified by three different methods: acidified water, silica and distillation. The quality aspects of the purified biodiesel samples were evaluated according to the Brazilian specifications for biodiesel, and distillation was shown to be the best method of purification. El objetivo del presente trabajo es evaluar la producción de biodiesel usando etanol y aceite de girasol. La extracción del aceite de girasol fue evaluada primero. Un diseño experimental fue usado para estimar la influencia de las variables independientes: temperatura del grano (25º a 110ºC y rotación del expeller (85 a 119 rpm en la obtención del aceite crudo. El mejor resultado obtenido fue un 68,38%, conseguido con una rotación de 100 a 115 rpm, una temperatura del grano de 25º a 30ºC y un contenido de humedad de alrededor del 7%. El siguiente estudio mediante transesterificación, evaluó la influencia de la relación molar etanol: aceite y concentración de catalizador (metilato sódico en el rendimiento de la fase rica en esteres. El rendimiento más alto fue 98,39% obtenido con una relación molar de 9.1 y 3% de 13. A safe inexpensive method to isolate high quality plant and fungal ... African Journals Online (AJOL) STORAGESEVER 2008-08-18 Aug 18, 2008 ... quality DNA from plant and fungal species. This method uses potassium acetate to remove proteins and polysaccharides in an SDS extraction buffer. Further DNA purification is achieved using a low salt. CTAB treatment. This SDS/CTAB protocol was used to isolate high quality genomic DNA subject to. 14. Dual-Readout Calorimetry for High-Quality Energy CERN Multimedia During the past seven years, the DREAM collaboration has systematically investigated all factors that determine and limit the precision with which the properties of hadrons and jets can be measured in calorimeters. Using simultaneous detection of the deposited energy and the Cerenkov light produced in hadronic shower development ${(dual}$ ${readout}$), the fluctuations in the electromagnetic shower fraction could be measured event by event their effects on signal linearity, response function and energy resolution eliminated. Detailed measurement of the time structure of the signals made it possible to measure the contirbutions of nuclear evaporation neutrons to the signals and thus reduce the effects of fluctuations in "invisible energy". We are now embarking on the construction of a full-scale calorimeter which incorporates all these elements and which should make it possible to measure the four-vectors of both electrons, hadrons and jets with very high precision, in an instrument that can be simply calibrat... 15. Export is necessary to maintain high quality supply International Nuclear Information System (INIS) Sirugue, S. 2016-01-01 Nuclear industry achieves about 20% of its turnover with the export of equipment and services but nuclear trade is submitted to specific constrains like non-proliferation policies. According to IAEA the world capacity of nuclear power is expected to double allowing the development of emerging countries while reducing CO_2 emissions. The French commercial offer focuses on the high-end market with the EPR reactor and an intermediate power range reactor ATMEA-1 designed with MHI (Mitsubishi Heavy Industries). AREVA is the unique world operator able to master the entire chain of spent fuel reprocessing thanks to the La Hague plant. In the framework of the rebuilding of the French nuclear industry, discussions are ongoing to achieve partnerships with foreign partners in exchange of shares in AREVA and AREVA NP. Today's main challenges of AREVA are to reduce EPR cost, to speed EPR construction and to benefit fully of the move into the digital age. (A.C.) 16. Access High Quality Imagery from the NOAA View Portal Science.gov (United States) Pisut, D.; Powell, A. M.; Loomis, T.; Goel, V.; Mills, B.; Cowan, D. 2013-12-01 NOAA curates a vast treasure trove of environmental data, but one that is sometimes not easily accessed, especially for education, outreach, and media purposes. Traditional data portals in NOAA require extensive knowledge of the specific names of observation platforms, models, and analyses, along with nomenclature for variable outputs. A new website and web mapping service (WMS) from NOAA attempts to remedy such issues. The NOAA View data imagery portal provides a seamless entry point into data from across the agency: satellite, models, in-situ analysis, etc. The system provides the user with ability to browse, animate, and download high resolution (e.g., 4,000 x 2,000 pixel) imagery, Google Earth, and even proxy data files. The WMS architecture also allows the resources to be ingested into other software systems or applications. 17. High Quality Rapeseed Products as Feed for Sensitive Monogastrics DEFF Research Database (Denmark) Frandsen, Heidi Blok in plants of the order Brassicales (former known as Capparales), which include rapeseed, rype (Brassica campestris L.) Indian mustard (Brassica juncea L.), broccoli (Brassica oleracea L.var. italica) and many other plants. Glucosinolates have been studied widely for their biologic effects ranging from...... by xenobiotica enzymes in the liver. The last study (manuscript IV) deals with the novel processing techniques, pulsed electric field (PEF) and high pressure treatment (HPT) and the processing effects on glucosinolates in broccoli. The largest effects were observed to be a result of the different handling...... of the plant materials prior to the process treatment. It was thus found that a great amount of the glucosinolate loss has occurred in the broccoli juice and purée prior to PEF processing. Only a minor loss was observed in broccoli flowers prior to processing, and HP treatment at 700 MPa for 10 min. was found... 18. Accelerating an Ordered-Subset Low-Dose X-Ray Cone Beam Computed Tomography Image Reconstruction with a Power Factor and Total Variation Minimization. Science.gov (United States) Huang, Hsuan-Ming; Hsiao, Ing-Tsung 2016-01-01 In recent years, there has been increased interest in low-dose X-ray cone beam computed tomography (CBCT) in many fields, including dentistry, guided radiotherapy and small animal imaging. Despite reducing the radiation dose, low-dose CBCT has not gained widespread acceptance in routine clinical practice. In addition to performing more evaluation studies, developing a fast and high-quality reconstruction algorithm is required. In this work, we propose an iterative reconstruction method that accelerates ordered-subsets (OS) reconstruction using a power factor. Furthermore, we combine it with the total-variation (TV) minimization method. Both simulation and phantom studies were conducted to evaluate the performance of the proposed method. Results show that the proposed method can accelerate conventional OS methods, greatly increase the convergence speed in early iterations. Moreover, applying the TV minimization to the power acceleration scheme can further improve the image quality while preserving the fast convergence rate. 19. Recycle food wastes into high quality fish feeds for safe and quality fish production. Science.gov (United States) Wong, Ming-Hung; Mo, Wing-Yin; Choi, Wai-Ming; Cheng, Zhang; Man, Yu-Bon 2016-12-01 The amount of food waste generated from modern societies is increasing, which has imposed a tremendous pressure on its treatment and disposal. Food waste should be treated as a valuable resource rather than waste, and turning it into fish feeds would be a viable alternative. This paper attempts to review the feasibility of using food waste to formulate feed pellets to culture a few freshwater fish species, such as grass carp, grey mullet, and tilapia, under polyculture mode (growing different species in the same pond). These species occupy different ecological niches, with different feeding modes (i.e., herbivorous, filter feeding, etc.), and therefore all the nutrients derived from the food waste could be efficiently recycled within the ecosystem. The problems facing environmental pollution and fish contamination; the past and present situation of inland fish culture (focusing on South China); upgrade of food waste based feed pellets by adding enzymes, vitamin-mineral premix, probiotics (yeast), prebiotics, and Chinese medicinal herbs into feeds; and potential health risks of fish cultivated by food waste based pellets are discussed, citing some local examples. It can be concluded that appropriate portions of different types of food waste could satisfy basic nutritional requirements of lower trophic level fish species such as grass carp and tilapia. Upgrading the fish pellets by adding different supplements mentioned above could further elevated the quality of feeds, leading to higher growth rates, and enhanced immunity of fish. Health risk assessments based on the major environmental contaminants (mercury, PAHs and DDTs) in fish flesh showed that fish fed food waste based pellets are safer for consumption, when compared with those fed commercial feed pellets. Copyright © 2016 Elsevier Ltd. All rights reserved. 20. High-quality thorium TRISO fuel performance in HTGRs Energy Technology Data Exchange (ETDEWEB) Verfondern, Karl [Forschungszentrum Juelich GmbH (Germany); Allelein, Hans-Josef [Forschungszentrum Juelich GmbH (Germany); Technische Hochschule Aachen (Germany); Nabielek, Heinz; Kania, Michael J. 2013-11-01 Thorium as a nuclear fuel has received renewed interest, because of its widespread availability and the good irradiation performance of Th and mixed (Th,U) oxide compounds as fuels in nuclear power systems. Early HTGR development employed thorium together with high-enriched uranium (HEU). After 1980, HTGR fuel systems switched to low-enriched uranium (LEU). After completing fuel development for the AVR and the THTR with BISO coated particles, the German program expanded its efforts utilizing thorium and HEU TRISO coated particles in advanced HTGR concepts for process heat applications (PNP) and direct-cycle electricity production (HHT). The combination of a low-temperature isotropic (LTI) inner and outer pyrocarbon layers surrounding a strong, stable SiC layer greatly improved manufacturing conditions and the subsequent contamination and defective particle fractions in production fuel elements. In addition, this combination provided improved mechanical strength and a higher degree of solid fission product retention, not known previously with high-temperature isotropic (HTI) BISO coatings. The improved performance of the HEU (Th, U)O{sub 2} TRISO fuel system was successfully demonstrated in three primary areas of development: manufacturing, irradiation testing under normal operating conditions, and accident simulation testing. In terms of demonstrating performance for advanced HTGR applications, the experimental failure statistic from manufacture and irradiation testing are significantly below the coated particle requirements specified for PNP and HHT designs at the time. Covering a range to 1300 C in normal operations and 1600 C in accidents, with burnups to 13% FIMA and fast fluences to 8 x 10{sup 25} n/m{sup 2} (E> 16 fJ), the performance results exceed the design limits on manufacturing and operational requirements for the German HTR-Modul concept, which are 6.5 x 10{sup -5} for manufacturing, 2 x 10{sup -4} for normal operating conditions, and 5 x 10{sup -4 1. High-quality thorium TRISO fuel performance in HTGRs International Nuclear Information System (INIS) Verfondern, Karl; Allelein, Hans-Josef; Nabielek, Heinz; Kania, Michael J. 2013-01-01 Thorium as a nuclear fuel has received renewed interest, because of its widespread availability and the good irradiation performance of Th and mixed (Th,U) oxide compounds as fuels in nuclear power systems. Early HTGR development employed thorium together with high-enriched uranium (HEU). After 1980, HTGR fuel systems switched to low-enriched uranium (LEU). After completing fuel development for the AVR and the THTR with BISO coated particles, the German program expanded its efforts utilizing thorium and HEU TRISO coated particles in advanced HTGR concepts for process heat applications (PNP) and direct-cycle electricity production (HHT). The combination of a low-temperature isotropic (LTI) inner and outer pyrocarbon layers surrounding a strong, stable SiC layer greatly improved manufacturing conditions and the subsequent contamination and defective particle fractions in production fuel elements. In addition, this combination provided improved mechanical strength and a higher degree of solid fission product retention, not known previously with high-temperature isotropic (HTI) BISO coatings. The improved performance of the HEU (Th, U)O 2 TRISO fuel system was successfully demonstrated in three primary areas of development: manufacturing, irradiation testing under normal operating conditions, and accident simulation testing. In terms of demonstrating performance for advanced HTGR applications, the experimental failure statistic from manufacture and irradiation testing are significantly below the coated particle requirements specified for PNP and HHT designs at the time. Covering a range to 1300 C in normal operations and 1600 C in accidents, with burnups to 13% FIMA and fast fluences to 8 x 10 25 n/m 2 (E> 16 fJ), the performance results exceed the design limits on manufacturing and operational requirements for the German HTR-Modul concept, which are 6.5 x 10 -5 for manufacturing, 2 x 10 -4 for normal operating conditions, and 5 x 10 -4 for accident conditions. These 2. Towards a high quality high school workforce: A longitudinal, demographic analysis of U.S. public school physics teachers OpenAIRE Gregory T. Rushton; David Rosengrant; Andrew Dewar; Lisa Shah; Herman E. Ray; Keith Sheppard; Lynn Watanabe 2017-01-01 Efforts to improve the number and quality of the high school physics teaching workforce have taken several forms, including those sponsored by professional organizations. Using a series of large-scale teacher demographic data sets from the National Center for Education Statistics (NCES), this study sought to investigate trends in teacher quality at the national level in the two and a half decades between 1987 and 2012. Specifically, we investigated (i) details about the degree backgrounds, ma... 3. EXPERT-ANALITICAL MONITORING OF LEARNING PROCESS QUALITY IN HIGH SCHOOL Directory of Open Access Journals (Sweden) T. M. Korotun 2010-10-01 Full Text Available The technological model is proposed for monitoring process of learning process quality in high school compliant with current European and home standards. The mathematical methods are elaborated for diverse activities as to learning process objects quality determination unified support. They self-consistently combine: automatic expert evaluation with Bayesian net and Value tree models; Delphi technique enhancement; best practices for education quality assessment. Quality estimates’ consistency index is introduced for their choice and acceptability analysis. Its permanent increasing over monitoring stages is guaranteed. The tools for these stages’ automatic support are described. 4. Timing of high-quality child care and cognitive, language, and preacademic development. Science.gov (United States) Li, Weilin; Farkas, George; Duncan, Greg J; Burchinal, Margaret R; Vandell, Deborah Lowe 2013-08-01 The effects of high- versus low-quality child care during 2 developmental periods (infant-toddlerhood and preschool) were examined using data from the National Institute of Child Health and Human Development Study of Early Child Care. Propensity score matching was used to account for differences in families who used different combinations of child care quality during the 2 developmental periods. Findings indicated that cognitive, language, and preacademic skills prior to school entry were highest among children who experienced high-quality care in both the infant-toddler and preschool periods, somewhat lower among children who experienced high-quality child care during only 1 of these periods, and lowest among children who experienced low-quality care during both periods. Irrespective of the care received during infancy-toddlerhood, high-quality preschool care was related to better language and preacademic outcomes at the end of the preschool period; high-quality infant-toddler care, irrespective of preschool care, was related to better memory skills at the end of the preschool period. (PsycINFO Database Record (c) 2013 APA, all rights reserved). 5. An approach to high speed ship ride quality simulation Science.gov (United States) Malone, W. L.; Vickery, J. M. 1975-01-01 The high speeds attained by certain advanced surface ships result in a spectrum of motion which is higher in frequency than that of conventional ships. This fact along with the inclusion of advanced ride control features in the design of these ships resulted in an increased awareness of the need for ride criteria. Such criteria can be developed using data from actual ship operations in varied sea states or from clinical laboratory experiments. A third approach is to simulate ship conditions using measured or calculated ship motion data. Recent simulations have used data derived from a math model of Surface Effect Ship (SES) motion. The model in turn is based on equations of motion which have been refined with data from scale models and SES of up to 101 600-kg (100-ton) displacement. Employment of broad band motion emphasizes the use of the simulators as a design tool to evaluate a given ship configuration in several operational situations and also serves to provide data as to the overall effect of a given motion on crew performance and physiological status. 6. Identifying suitable substrates for high-quality graphene-based heterostructures Science.gov (United States) Banszerus, L.; Janssen, H.; Otto, M.; Epping, A.; Taniguchi, T.; Watanabe, K.; Beschoten, B.; Neumaier, D.; Stampfer, C. 2017-06-01 We report on a scanning confocal Raman spectroscopy study investigating the strain-uniformity and the overall strain and doping of high-quality chemical vapour deposited (CVD) graphene-based heterostuctures on a large number of different substrate materials, including hexagonal boron nitride (hBN), transition metal dichalcogenides, silicon, different oxides and nitrides, as well as polymers. By applying a hBN-assisted, contamination free, dry transfer process for CVD graphene, high-quality heterostructures with low doping densities and low strain variations are assembled. The Raman spectra of these pristine heterostructures are sensitive to substrate-induced doping and strain variations and are thus used to probe the suitability of the substrate material for potential high-quality graphene devices. We find that the flatness of the substrate material is a key figure for gaining, or preserving high-quality graphene. 7. Production of high quality sodium iodide preparations labelled with carrier free iodine-125 International Nuclear Information System (INIS) Abdukayumov, M.N.; Chistyakov, P.G.; Shilin, E.A. 2001-01-01 Work is related to the problem of high-quality Sodium Iodide preparation production and to the choice of the peptids iodination methods with the purpose of control test developing to determine the Biological activity of the above mentioned preparation 8. The Role of Dosimetry in High-Quality EMI Risk Assessment National Research Council Canada - National Science Library 2006-01-01 The Final Proceedings for The role of EMF dosimetry in high quality risk assessment 13 September 2006 - 15 September 2006 In the last three decades the use of devices that emit electromagnetic fields (EMF... 9. Diode pumped 1kHz high power Nd:YAG laser with excellent beam quality NARCIS (Netherlands) Godfried, Herman; Godfried, H.P; Offerhaus, Herman L. 1997-01-01 The design and operation of a one kilohertz diode pumped all solid-state Nd:YAG master oscillator power amplifier system with a phase conjugate mirror is presented. The setup allows high power scaling without reduction in beam quality. 10. Optimum unambiguous discrimination between subsets of nonorthogonal quantum states International Nuclear Information System (INIS) Sun Yuqing; Hillery, Mark; Bergou, Janos A. 2002-01-01 It is known that unambiguous discrimination among nonorthogonal but linearly independent quantum states is possible with a certain probability of success. Here, we consider a variant of that problem. Instead of discriminating among all of the different states, we shall only discriminate between two subsets of them. In particular, for the case of three nonorthogonal states, { vertical bar ψ 1 >, vertical bar ψ 2 >, vertical bar ψ 3 >}, we show that the optimal strategy to distinguish vertical bar ψ 1 > from the set { vertical bar ψ 2 >, vertical bar ψ 3 >} has a higher success rate than if we wish to discriminate among all three states. Somewhat surprisingly, for unambiguous discrimination the subsets need not be linearly independent. A fully analytical solution is presented, and we also show how to construct generalized interferometers (multiport) which provide an optical implementation of the optimal strategy 11. Auditing Complex Concepts in Overlapping Subsets of SNOMED Science.gov (United States) Wang, Yue; Wei, Duo; Xu, Junchuan; Elhanan, Gai; Perl, Yehoshua; Halper, Michael; Chen, Yan; Spackman, Kent A.; Hripcsak, George 2008-01-01 Limited resources and the sheer volume of concepts make auditing a large terminology, such as SNOMED CT, a daunting task. It is essential to devise techniques that can aid an auditor by automatically identifying concepts that deserve attention. A methodology for this purpose based on a previously introduced abstraction network (called the p-area taxonomy) for a SNOMED CT hierarchy is presented. The methodology algorithmically gathers concepts appearing in certain overlapping subsets, defined exclusively with respect to the p-area taxonomy, for review. The results of applying the methodology to SNOMED’s Specimen hierarchy are presented. These results are compared against a control sample composed of concepts residing in subsets without the overlaps. With the use of the double bootstrap, the concept group produced by our methodology is shown to yield a statistically significant higher proportion of error discoveries. PMID:18998838 12. Massive two-loop Bhabha scattering - the factorizable subset International Nuclear Information System (INIS) Fleischer, J.; Tarasov, O.V.; Werthenbach, A. 2002-11-01 The experimental precision that will be reached at the next generation of colliders makes it indispensable to improve theoretical predictions significantly. Bhabha scattering (e + e - → e + e - ) is one of the prime processes calling for a better theoretical precision, in particular for non-zero electron masses. We present a first subset of the full two-loop calculation, namely the factorizable subset. Our calculation is based on DIANA. We reduce tensor integrals to scalar integrals in shifted (increased) dimensions and additional powers of various propagators, so-called dots-on-lines. Recurrence relations remove those dots-on-lines as well as genuine dots-on-lines (originating from mass renormalization) and reduce the dimension of the integrals to the generic d=4-2ε dimensions. The resulting master integrals have to be expanded to O(ε) to ensure proper treatment of all finite terms. (orig.) 13. Individual discriminative face recognition models based on subsets of features DEFF Research Database (Denmark) Clemmensen, Line Katrine Harder; Gomez, David Delgado; Ersbøll, Bjarne Kjær 2007-01-01 The accuracy of data classification methods depends considerably on the data representation and on the selected features. In this work, the elastic net model selection is used to identify meaningful and important features in face recognition. Modelling the characteristics which distinguish one...... person from another using only subsets of features will both decrease the computational cost and increase the generalization capacity of the face recognition algorithm. Moreover, identifying which are the features that better discriminate between persons will also provide a deeper understanding...... of the face recognition problem. The elastic net model is able to select a subset of features with low computational effort compared to other state-of-the-art feature selection methods. Furthermore, the fact that the number of features usually is larger than the number of images in the data base makes feature... 14. High Decision Latitude at Work Combined with Traumatic Life Events in Private Life is Associated with Reduced Sleep Quality: Results from the GAZEL Study Directory of Open Access Journals (Sweden) 2018-04-01 Full Text Available Psychosocial stress, both at work and at home, plays a role in sleep disturbances. Theoretical models of stress underscore the cumulative effect of stress from different domains in affecting health and well-being. However, previous research has not considered how work stress and stress in private life might interact to affect sleep. The study investigated potential synergistic effects of work stress and traumatic events in private life on sleep quality. Participants ('N' = 6552, average baseline age = 52 were a subset of the French GAZEL cohort who were working in 1999. Work stress was operationalized as high job demands and low decision latitude. These working conditions were self-reported in 1999 (study baseline. Traumatic events in the past 12 months were reported in 2000. Sleep quality was assessed in 2000 as disturbed sleep in the past 12 months, and current sleep problems was indexed by the sleep subscale of the Nottingham Health Profile questionnaire. The results showed that recent events interacted with decision latitude, but not job demands, in predicting sleep quality. However, contrary to our expectations, it was 'high' and not low decision latitude at work that amplified the negative association between stressful events and sleep quality. Adjusted for baseline health, individuals with highest numbers of events and highest levels of decision latitude were at highest risk for impaired sleep. These results challenge the idea that high decision latitude always serves as a protective factor, and underscore the necessity for considering a broader life context when studying stress in a particular domain of life. 15. PKH26 staining defines distinct subsets of normal human colon epithelial cells at different maturation stages. Directory of Open Access Journals (Sweden) Anna Pastò Full Text Available BACKGROUND AND AIM: Colon crypts are characterized by a hierarchy of cells distributed along the crypt axis. Aim of this paper was to develop an in vitro system for separation of epithelial cell subsets in different maturation stages from normal human colon. METHODOLOGY AND MAJOR FINDINGS: Dissociated colonic epithelial cells were stained with PKH26, which allows identification of distinct populations based on their proliferation rate, and cultured in vitro in the absence of serum. The cytofluorimetric expression of CK20, Msi-1 and Lgr5 was studied. The mRNA levels of several stemness-associated genes were also compared in cultured cell populations and in three colon crypt populations isolated by microdissection. A PKH(pos population survived in culture and formed spheroids; this population included subsets with slow (PKH(high and rapid (PKH(low replicative rates. Molecular analysis revealed higher mRNA levels of both Msi-1 and Lgr-5 in PKH(high cells; by cytofluorimetric analysis, Msi-1(+/Lgr5(+ cells were only found within PKH(high cells, whereas Msi-1(+/Lgr5(- cells were also observed in the PKH(low population. As judged by qRT-PCR analysis, the expression of several stemness-associated markers (Bmi-1, EphB2, EpCAM, ALDH1 was highly enriched in Msi-1(+/Lgr5(+ cells. While CK20 expression was mainly found in PKH(low and PKH(neg cells, a small PKH(high subset co-expressed both CK20 and Msi-1, but not Lgr5; cells with these properties also expressed Mucin, and could be identified in vivo in colon crypts. These results mirrored those found in cells isolated from different crypt portions by microdissection, and based on proliferation rates and marker expression they allowed to define several subsets at different maturation stages: PKH(high/Lgr5(+/Msi-1(+/CK20(-, PKH(high/Lgr5(-/Msi-1(+/CK20(+, PKH(low/Lgr5(-/Msi-1(+/Ck20(-, and PKH(low/Lgr5(-/Msi-1(-/CK20(+ cells. CONCLUSIONS: Our data show the possibility of deriving in vitro, without any 16. Roquin Paralogs Differentially Regulate Functional NKT Cell Subsets. Science.gov (United States) Drees, Christoph; Vahl, J Christoph; Bortoluzzi, Sabrina; Heger, Klaus D; Fischer, Julius C; Wunderlich, F Thomas; Peschel, Christian; Schmidt-Supprian, Marc 2017-04-01 NKT cells represent a small subset of glycolipid-recognizing T cells that are heavily implicated in human allergic, autoimmune, and malignant diseases. In the thymus, precursor cells recognize self-glycolipids by virtue of their semi-invariant TCR, which triggers NKT cell lineage commitment and maturation. During their development, NKT cells are polarized into the NKT1, NKT2, and NKT17 subsets, defined through their cytokine-secretion patterns and the expression of key transcription factors. However, we have largely ignored how the differentiation into the NKT cell subsets is regulated. In this article, we describe the mRNA-binding Roquin-1 and -2 proteins as central regulators of murine NKT cell fate decisions. In the thymus, T cell-specific ablation of the Roquin paralogs leads to a dramatic expansion of NKT17 cells, whereas peripheral mature NKT cells are essentially absent. Roquin-1/2-deficient NKT17 cells show exaggerated lineage-specific expression of nearly all NKT17-defining proteins tested. We show through mixed bone marrow chimera experiments that NKT17 polarization is mediated through cell-intrinsic mechanisms early during NKT cell development. In contrast, the loss of peripheral NKT cells is due to cell-extrinsic factors. Surprisingly, Roquin paralog-deficient NKT cells are, in striking contrast to conventional T cells, compromised in their ability to secrete cytokines. Altogether, we show that Roquin paralogs regulate the development and function of NKT cell subsets in the thymus and periphery. Copyright © 2017 by The American Association of Immunologists, Inc. 17. Phenotypic and functional characterization of earthworm coelomocyte subsets DEFF Research Database (Denmark) Engelmann, Péter; Hayashi, Yuya; Bodo, Kornélia 2016-01-01 Flow cytometry is a common approach to study invertebrate immune cells including earthworm coelomocytes. However, the link between light-scatter- and microscopy-based phenotyping remains obscured. Here we show, by means of light scatter-based cell sorting, both subpopulations (amoebocytes...... amoebocytes and eleocytes, with the former being in favor of bacterial engulfment. This study has proved successful in linking flow cytometry and microscopy analysis and provides further experimental evidence of phenotypic and functional heterogeneity in earthworm coelomocyte subsets.... 18. Universal and rapid salt-extraction of high quality genomic DNA for PCR-based techniques. OpenAIRE Aljanabi, S M; Martinez, I 1997-01-01 A very simple, fast, universally applicable and reproducible method to extract high quality megabase genomic DNA from different organisms is described. We applied the same method to extract high quality complex genomic DNA from different tissues (wheat, barley, potato, beans, pear and almond leaves as well as fungi, insects and shrimps' fresh tissue) without any modification. The method does not require expensive and environmentally hazardous reagents and equipment. It can be performed even i... 19. The use of citation indicators to identify and support high-quality research in Poland. Science.gov (United States) Pilc, Andrzej 2008-01-01 In large, mostly English-speaking countries, where the "critical mass" of scientists working in different subfields of science is achieved, the peer review system may be sufficient to assess the quality of scientific research. However, in smaller countries, outside the Anglo-American circle, it is important to introduce different systems to identify research of high quality. In Poland, a parametric system for assessing the quality of research has been introduced. It was largely based on the impact factor of scientific journals. While the use of this indicator to assess research quality is highly questionable, the implementation of the system in the Polish reality is even worse. Therefore it is important to change and improve the system currently used by the Ministry of Science and Higher Education to both evaluate and, more importantly, finance science in Poland. Here, a system based on three factors, i.e. the impact factor, the institutional h-index, and the institutional number of citations, is proposed. The scientific quality of institutions in Division VI: Medical Sciences of the Polish Academy of Sciences were evaluated and the results were compared with the existing system. Moreover, a method to identify high-quality researchers and institutions at the national level based on the quantity of highly cited papers is shown. Additionally, an attempt to identify the highest quality Polish research on an international level is proposed. This is based on the number of individual citations, the individual h-index, the number of publications, and the priority of the discovery. 20. Reconfiguration of NKT Cell Subset Compartment Is Associated with Plaque Development in Patients with Carotid Artery Stenosis. Science.gov (United States) Cai, Lun; Yu, Lei; Liu, Sa; Li, Tongxun; Zhang, Xiaoping; Cui, Wei; Du, Jie; Zhang, Qinyi 2017-02-01 Accumulating evidence shows that immune cells play an important role in carotid atherosclerotic plaque development. In this study, we assessed the association of 6 different natural killer T (NKT) cell subsets, based on CD57 and CD8 expression, with risk for development of carotid atherosclerotic plaque (CAP). Molecular expression by peripheral NKT cells was evaluated in 13 patients with high-risk CAP and control without carotid stenosis (n = 18). High-risk CAP patients, compared with healthy subjects, had less percentage of CD57+CD8- NKT cell subsets (8.64 ± 10.15 versus 19.62 ± 10.8 %; P = 0.01) and CD57+CD8int NKT cell subsets (4.32 ± 3.04 versus 11.87 ± 8.56 %; P = 0.002), with a corresponding increase in the CD57-CD8high NKT cell subsets (33.22 ± 11.87 versus 18.66 ± 13.68 %; P = 0.007). Intracellular cytokine staining showed that CD8+ NKT cell subset was the main cytokine-producing NKT cell. Cytokine production in plasma was measured with Bio-Plex assay. The expression levels of pro-inflammatory mediators (IFN-γ, IL-17, IP-10) were significantly higher in CAP patients as compared to that from controls. These data provide evidence that NKT cell subset compartment reconfiguration in patients with carotid stenosis seems to be associated with the occurrence of carotid atherosclerotic plaque and suggest that both pathogenic and protective NKT cell subsets exist. 1. Selection of process parameters for producing high quality defatted sesame flour at pilot scale. Science.gov (United States) Manikantan, M R; Sharma, Rajiv; Yadav, D N; Gupta, R K 2015-03-01 The present work was undertaken to study the effect of pearling duration, soaking time, steaming duration and drying temperature on the quality of sesame seeds and mechanically extracted partially defatted sesame cake. On the basis of quality attributes i.e. high protein, low crude fibre, low residual oil and low oxalic acid, the optimum process parameters were selected. The combination of 20 min of pearling duration, 15 min of soaking, 15 min of steaming at 100 kPa pressure and drying at 50 °C yielded high quality partially defatted protein rich sesame flour as compared to untreated defatted sesame flour. The developed high quality partially defatted protein rich sesame flour may be used in various food applications as a vital ingredient to increase the nutritional significance of the prepared foodstuffs. 2. High Quality RNA Isolation from Leaf, Shell, Root Tissues and Callus of Hazelnut (Corylus avellana L. Directory of Open Access Journals (Sweden) Hossein Khosravi 2017-12-01 Full Text Available Extraction of high quality RNA is a critical step in molecular genetics studies. Hazelnut is one of the most important nuts plants in the world. The presence of the taxol and other taxanes in hazelnut plant necessitates explaining their biosynthesis pathway and identifying the candidate genes. Therefore, an easy and practical method is necessary for RNA extraction from hazelnuts. Hazelnut has high levels of phenolic compounds. High amounts of polyphenolic and polysaccharide compounds in plants could be causing problems in RNA extraction procedures.  To avoid these problems, a simple and efficient method can be used based on cetyltrimethylammonium bromide (CTAB extraction buffer and lithium chloride for extraction of high quality RNA from different parts of hazelnut plant. Using this method, a high-quality RNA sample (light absorbed in the A260/A280 was 2.04 3. Characterization of a chromosome-specific chimpanzee alpha satellite subset: Evolutionary relationship to subsets on human chromosomes Energy Technology Data Exchange (ETDEWEB) Warburton, P.E.; Gosden, J.; Lawson, D. [Western General Hospital, Edinburgh (United Kingdom)] [and others 1996-04-15 Alpha satellite DNA is a tandemly repeated DNA family found at the centromeres of all primate chromosomes examined. The fundamental repeat units of alpha satellite DNA are diverged 169- to 172-bp monomers, often found to be organized in chromosome-specific higher-order repeat units. The chromosomes of human (Homo sapiens (HSA)), chimpanzee (Pan troglodytes (PTR) and Pan paniscus), and gorilla (Gorilla gorilla) share a remarkable similarity and synteny. It is of interest to ask if alpha satellite arrays at centromeres of homologous chromosomes between these species are closely related (evolving in an orthologous manner) or if the evolutionary processes that homogenize and spread these arrays within and between chromosomes result in nonorthologous evolution of arrays. By using PCR primers specific for human chromosome 17-specific alpha satellite DNA, we have amplified, cloned, and characterized a chromosome-specific subset from the PTR chimpanzee genome. Hybridization both on Southern blots and in situ as well as sequence analysis show that this subset is most closely related, as expected, to sequences on HSA 17. However, in situ hybridization reveals that this subset is not found on the homologous chromosome in chimpanzee (PTR 19), but instead on PTR 12, which is homologous to HSA 2p. 40 refs., 3 figs. 4. Isolation of high quality RNA from pistachio (Pistacia vera L.) and other woody plants high in secondary metabolites. Science.gov (United States) Moazzam Jazi, Maryam; Rajaei, Saideh; Seyedi, Seyed Mahdi 2015-10-01 The quality and quantity of RNA are critical for successful downstream transcriptome-based studies such as microarrays and RNA sequencing (RNA-Seq). RNA isolation from woody plants, such as Pistacia vera, with very high amounts of polyphenols and polysaccharides is an enormous challenge. Here, we describe a highly efficient protocol that overcomes the limitations posed by poor quality and low yield of isolated RNA from pistachio and various recalcitrant woody plants. The key factors that resulted in a yield of 150 μg of high quality RNA per 200 mg of plant tissue include the elimination of phenol from the extraction buffer, raising the concentration of β-mercaptoethanol, long time incubation at 65 °C, and nucleic acid precipitation with optimized volume of NaCl and isopropyl alcohol. Also, the A260/A280 and A260/A230 of extracted RNA were about 1.9-2.1and 2.2-2.3, respectively, revealing the high purity. Since the isolated RNA passed highly stringent quality control standards for sensitive reactions, including RNA sequencing and real-time PCR, it can be considered as a reliable and cost-effective method for RNA extraction from woody plants. 5. Analysis of apple beverages treated with high-power ultrasound: a quality function deployment approach. Science.gov (United States) Režek Jambrak, Anet; Šimunek, Marina; Grbeš, Franjo; Mandura, Ana; Djekic, Ilija 2018-04-01 The objective of this paper was to demonstrate application of quality function deployment in analysing effects of high power ultrasound on quality properties of apple juices and nectars. In order to develop a quality function deployment model, joint with instrumental analysis of treated samples, a field survey was performed to identify consumer preferences towards quality characteristics of juices/nectar. Based on field research, the three most important characteristics were 'taste' and 'aroma' with 28.5% of relative absolute weight importance, followed by 'odour' (16.9%). The quality function deployment model showed that the top three 'quality scores' for apple juice were treatments with amplitude 90 µm, 9 min treatment time and sample temperature 40 °C; 60 µm, 9 min, 60 °C; and 90 µm, 6 min, 40 °C. For nectars, the top three were treatments 120 µm, 9 min, 20 °C; 60 µm, 9 min, 60 °C; and A2.16 60 µm, 9 min, 20 °C. This type of quality model enables a more complex measure of large scale of different quality parameters. Its simplicity should be understood as its practical advantage and, as such, this tool can be a part of design quality when using novel preservation technologies. © 2017 Society of Chemical Industry. © 2017 Society of Chemical Industry. 6. Bullying as a risk for poor sleep quality among high school students in China. Directory of Open Access Journals (Sweden) Ying Zhou Full Text Available To determine whether involvement in bullying as a bully, victim, or bully-victim was associated with a higher risk of poor sleep quality among high school students in China.A cross-sectional study was conducted. A total of 23,877 high school students were surveyed in six cities in Guangdong Province. All students were asked to complete the adolescent health status questionnaire, which included the Chinese version of the Pittsburgh Sleep Quality Index (PSQI and bullying involvement. Descriptive statistics were used to evaluate sleep quality and the prevalence of school bullying. Multi-level logistic regression analyses were conducted to examine the association between being victimized and bullying others with sleep quality.Among the 23,877 students, 6,127 (25.66% reported having poor sleep quality, and 10.89% reported being involved in bullying behaviors. Of the respondents, 1,410 (5.91% were pure victims of bullying, 401 (1.68% were bullies and 784 (3.28% were bully-victims. Frequently being involved in bullying behaviors (being bullied or bullying others was related to increased risks of poor sleep quality compared with adolescents who were not involved in bullying behaviors. After adjusting for age, sex, and other confounding factors, the students who were being bullied (OR=2.05, 95%CI=1.81-2.32, bullied others (OR=2.30, 95%CI=1.85-2.86 or both (OR=2.58, 95%CI=2.20-3.03 were at a higher risk for poor sleep quality.Poor sleep quality among high school students is highly prevalent, and school bullying is prevalent among adolescents in China. The present results suggested that being involved in school bullying might be a risk factor for poor sleep quality among adolescents. 7. Bullying as a risk for poor sleep quality among high school students in China. Science.gov (United States) Zhou, Ying; Guo, Lan; Lu, Ci-yong; Deng, Jian-xiong; He, Yuan; Huang, Jing-hui; Huang, Guo-liang; Deng, Xue-qing; Gao, Xue 2015-01-01 To determine whether involvement in bullying as a bully, victim, or bully-victim was associated with a higher risk of poor sleep quality among high school students in China. A cross-sectional study was conducted. A total of 23,877 high school students were surveyed in six cities in Guangdong Province. All students were asked to complete the adolescent health status questionnaire, which included the Chinese version of the Pittsburgh Sleep Quality Index (PSQI) and bullying involvement. Descriptive statistics were used to evaluate sleep quality and the prevalence of school bullying. Multi-level logistic regression analyses were conducted to examine the association between being victimized and bullying others with sleep quality. Among the 23,877 students, 6,127 (25.66%) reported having poor sleep quality, and 10.89% reported being involved in bullying behaviors. Of the respondents, 1,410 (5.91%) were pure victims of bullying, 401 (1.68%) were bullies and 784 (3.28%) were bully-victims. Frequently being involved in bullying behaviors (being bullied or bullying others) was related to increased risks of poor sleep quality compared with adolescents who were not involved in bullying behaviors. After adjusting for age, sex, and other confounding factors, the students who were being bullied (OR=2.05, 95%CI=1.81-2.32), bullied others (OR=2.30, 95%CI=1.85-2.86) or both (OR=2.58, 95%CI=2.20-3.03) were at a higher risk for poor sleep quality. Poor sleep quality among high school students is highly prevalent, and school bullying is prevalent among adolescents in China. The present results suggested that being involved in school bullying might be a risk factor for poor sleep quality among adolescents. 8. Accelerated time-of-flight (TOF) PET image reconstruction using TOF bin subsetization and TOF weighting matrix pre-computation International Nuclear Information System (INIS) Mehranian, Abolfazl; Kotasidis, Fotis; Zaidi, Habib 2016-01-01 Time-of-flight (TOF) positron emission tomography (PET) technology has recently regained popularity in clinical PET studies for improving image quality and lesion detectability. Using TOF information, the spatial location of annihilation events is confined to a number of image voxels along each line of response, thereby the cross-dependencies of image voxels are reduced, which in turns results in improved signal-to-noise ratio and convergence rate. In this work, we propose a novel approach to further improve the convergence of the expectation maximization (EM)-based TOF PET image reconstruction algorithm through subsetization of emission data over TOF bins as well as azimuthal bins. Given the prevalence of TOF PET, we elaborated the practical and efficient implementation of TOF PET image reconstruction through the pre-computation of TOF weighting coefficients while exploiting the same in-plane and axial symmetries used in pre-computation of geometric system matrix. In the proposed subsetization approach, TOF PET data were partitioned into a number of interleaved TOF subsets, with the aim of reducing the spatial coupling of TOF bins and therefore to improve the convergence of the standard maximum likelihood expectation maximization (MLEM) and ordered subsets EM (OSEM) algorithms. The comparison of on-the-fly and pre-computed TOF projections showed that the pre-computation of the TOF weighting coefficients can considerably reduce the computation time of TOF PET image reconstruction. The convergence rate and bias-variance performance of the proposed TOF subsetization scheme were evaluated using simulated, experimental phantom and clinical studies. Simulations demonstrated that as the number of TOF subsets is increased, the convergence rate of MLEM and OSEM algorithms is improved. It was also found that for the same computation time, the proposed subsetization gives rise to further convergence. The bias-variance analysis of the experimental NEMA phantom and a clinical 9. When high-volume PCI operators in high-volume hospitals move to lower volume hospitals-Do they still maintain high volume and quality of outcomes? Science.gov (United States) Lu, Tsung-Hsueh; Li, Sheng-Tun; Liang, Fu-Wen; Lee, Jo-Chi; Yin, Wei-Hsian 2017-10-31 The aim of this quasi-experimental study was to examine whether high-volume percutaneous coronary intervention (PCI) operators still maintain high volume and quality of outcomes when they moved to lower volume hospitals. Systematic reviews have indicated that high-volume PCI operators and hospitals have higher quality outcomes. However, little is known on whether high PCI volume and high quality outcomes are mainly due to operator characteristics (i.e., skill and experience) and is portable across organizations or whether it is due to hospital characteristics (i.e., equipment, team, and management system) and is less portable. We used Taiwan National Health Insurance claims data 2000-2012 to identify 98 high-volume PCI operators, 10 of whom moved from one hospital to another during the study period. We compared the PCI volume, risk-adjusted mortality ratio, and major adverse cardiovascular event (MACE) ratio before and after moving. Of the 10 high-volume operators who moved, 6 moved from high- to moderate- or low-volume hospitals, with median annual PCI volumes (interquartile range) of 130 (117-165) in prior hospitals and 54 (46-84) in subsequent hospitals (the hospital the operator moved to), and the remaining 4 moved from high to high-volume hospitals, with median annual PCI volumes (interquartile range) of 151 (133-162) in prior hospitals and 193 (178-239) in subsequent hospitals. No significant differences were observed in the risk-adjusted mortality ratios and MACE ratios between high-volume operators and matched controls before and after moving. High-volume operators cannot maintain high volume when they moved from high to moderate or low-volume hospitals; however, the quality of care is maintained. High PCI volume and high-quality outcomes are less portable and more hospital bound. © 2017 Wiley Periodicals, Inc. 10. A Narrative Review of High-Quality Literature on the Effects of Resident Duty Hours Reforms. Science.gov (United States) Lin, Henry; Lin, Emery; Auditore, Stephanie; Fanning, Jon 2016-01-01 To summarize current high-quality studies evaluating the effect and efficacy of resident duty hours reforms (DHRs) on patient safety and resident education and well-being. The authors searched PubMed and Medline in August 2012 and again in May 2013 for literature (1987-2013) about the effects of DHRs. They assessed the quality of articles using the Medical Education Research Study Quality Instrument (MERSQI) scoring system. They considered randomized controlled trials (RCTs), partial RCTs, and all studies with a MERSQI score ≥ 14 to be "high-quality" methodology studies. A total of 72 high-quality studies met inclusion criteria. Most studies showed no change or slight improvement in mortality and complication rates after DHRs. Resident well-being was generally improved, but there was a perceived negative impact on education (knowledge acquisition, skills, and cognitive performance) following DHRs. Eleven high-quality studies assessed the impact of DHR interventions; all reported a neutral to positive impact. Seven high-quality studies assessed costs associated with DHRs and demonstrated an increase in hospital costs. The results of most studies that allow enough time for DHR interventions to take effect suggest a benefit to patient safety and resident well-being, but the effect on the quality of training remains unknown. Additional methodologically sound studies on the impact of DHRs are necessary. Priorities for future research include approaches to optimizing education and clinical proficiency and studies on the effect of intervention strategies on both education and patient safety. Such studies will provide additional information to help improve duty hours policies. 11. Certifying leaders? high-quality management practices and healthy organisations: an ISO-9000 based standardisation approach. Science.gov (United States) Montano, Diego 2016-08-05 The present study proposes a set of quality requirements to management practices by taking into account the empirical evidence on their potential effects on health, the systemic nature of social organisations, and the current conceptualisations of management functions within the framework of comprehensive quality management systems. Systematic reviews and meta-analyses focusing on the associations between leadership and/or supervision and health in occupational settings are evaluated, and the core elements of an ISO 9001 standardisation approach are presented. Six major occupational health requirements to high-quality management practices are identified pertaining to communication processes, organisational justice, role clarity, decision making, social influence processes and management support. It is concluded that the quality of management practices may be improved by developing a quality management system of management practices that ensures not only conformity to product but also to occupational safety and health requirements. Further research may evaluate the practicability of the proposed approach. 12. Modeling the Subjective Quality of Highly Contrasted Videos Displayed on LCD With Local Backlight Dimming DEFF Research Database (Denmark) Mantel, Claire; Bech, Søren; Korhonen, Jari 2015-01-01 Local backlight dimming is a technology aiming at both saving energy and improving visual quality on television sets. As the rendition of the image is specified locally, the numerical signal corresponding to the displayed image needs to be computed through a model of the display. This simulated...... signal can then be used as input to objective quality metrics. The focus of this paper is on determining which characteristics of locally backlit displays influence quality assessment. A subjective experiment assessing the quality of highly contrasted videos displayed with various local backlight......-dimming algorithms is set up. Subjective results are then compared with both objective measures and objective quality metrics using different display models. The first analysis indicates that the most significant objective features are temporal variations, power consumption (probably representing leakage... 13. The effect of the amiodarone-warfarin interaction on anticoagulation quality in a single, high-quality anticoagulation center. Science.gov (United States) White, Ryan D; Riggs, Kyle W; Ege, Ed J; Petroski, Gregory F; Koerber, Scott M; Flaker, Greg 2016-03-01 Clinical trials have reported a low time in therapeutic range (TTR) in patients with atrial fibrillation treated with both warfarin andamiodarone. These trials included centers and countries with both high and low TTRs. What is the impact of amiodarone on the TTR in a single, high-quality anticoagulation clinic? TTR was assessed in amiodarone and nonamiodarone-treated patients from a University anticoagulation clinic. Baseline characteristics between patients ever-taking or never-taking amiodarone were similar, except more amiodarone patients were smokers (19.5 vs. 6.1%, P = 0.0031). The TTR calculated from 8901international normalized ratios (INRs) in 249 nonamiodarone patients with a mean follow-up of 34 ± 20 months (mean INR 36 ± 18) was 66 ± 16.6% compared with 61.3 ± 16.2% (P = 0.111) from 1455 INRs in 41 amiodarone-treated patients with a mean follow-up of 28 ± 20 months (mean INR 35 ± 22). Factors associated with a low TTR were male sex (P = 0.0013), smoker (P = 0.0048), and amiodarone use (P = 0.0374). A second on-treatment analysis, in which the TTR was calculated only during amiodarone therapy, resulted in similar findings; however, amiodarone did not emerge as a predictor of a low TTR. In 11 patients, the TTR prior to amiodarone (54.5 ± 22.2%) was not significantly different in the first 3 months (54.6 ± 33.4%) or after 3 months (67.2 ± 33.7%) of amiodarone. In a single high-quality anticoagulation center, anticoagulation quality, as measured by the TTR, can be comparable in amiodarone and nonamiodarone-treated patients. 14. Combined microfluidization and ultrasonication: a synergistic protocol for high-efficient processing of SWCNT dispersions with high quality Energy Technology Data Exchange (ETDEWEB) Luo, Sida, E-mail: [email protected] [Beihang University, School of Mechanical Engineering and Automation (China); Liu, Tao, E-mail: [email protected] [Florida State University, High-Performance Materials Institute (United States); Wang, Yong; Li, Liuhe [Beihang University, School of Mechanical Engineering and Automation (China); Wang, Guantao; Luo, Yun [China University of Geosciences, Center of Safety Research, School of Engineering and Technology (China) 2016-08-15 High-efficient and large-scale production of high-quality CNT dispersions is necessary for meeting the future needs to develop various CNT-based electronic devices. Herein, we have designed novel processing protocols by combining conventional ultrasonication process with a new microfluidization technique to produce high-quality SWCNT dispersions with improved processing efficiency. To judge the quality of SWCNT dispersions, one critical factor is the degree of exfoliation, which could be quantified by both geometrical dimension of the exfoliated nanotubes and percentage of individual tubes in a given dispersion. In this paper, the synergistic effect of the combined protocols was systematically investigated through evaluating SWCNT dispersions with newly developed characterization techniques, namely preparative ultracentrifuge method (PUM) and simultaneous Raman scattering and photoluminescence spectroscopy (SRSPL). The results of both techniques draw similar conclusions that as compared with either of the processes operated separately, a low-pass microfluidization followed by a reasonable duration of ultrasonication could substantially improve the processing efficiency to produce high-quality SWCNT dispersions with averaged particle length and diameter as small as ~600 and ~2 nm, respectively.Graphical abstract. 15. Towards a High Quality High School Workforce: A Longitudinal, Demographic Analysis of U.S. Public School Physics Teachers Science.gov (United States) Rushton, Gregory T.; Rosengrant, David; Dewar, Andrew; Shah, Lisa; Ray, Herman E.; Sheppard, Keith; Watanabe, Lynn 2017-01-01 Efforts to improve the number and quality of the high school physics teaching workforce have taken several forms, including those sponsored by professional organizations. Using a series of large-scale teacher demographic data sets from the National Center for Education Statistics (NCES), this study sought to investigate trends in teacher quality… 16. Examining Relational Engagement across the Transition to High Schools in Three US High Schools Reformed to Improve Relationship Quality Science.gov (United States) Davis, Heather A.; Chang, Mei-Lin; Andrzejewski, Carey E.; Poirier, Ryan R. 2014-01-01 The purpose of this study was to examine changes in students' relational engagement across the transition to high school in three schools reformed to improve the quality of student-teacher relationships. In order to analyze this data we employed latent growth curve (LGC) modeling techniques (n = 637). We ran three LGC models on three… 17. Selecting a climate model subset to optimise key ensemble properties Directory of Open Access Journals (Sweden) N. Herger 2018-02-01 Full Text Available End users studying impacts and risks caused by human-induced climate change are often presented with large multi-model ensembles of climate projections whose composition and size are arbitrarily determined. An efficient and versatile method that finds a subset which maintains certain key properties from the full ensemble is needed, but very little work has been done in this area. Therefore, users typically make their own somewhat subjective subset choices and commonly use the equally weighted model mean as a best estimate. However, different climate model simulations cannot necessarily be regarded as independent estimates due to the presence of duplicated code and shared development history. Here, we present an efficient and flexible tool that makes better use of the ensemble as a whole by finding a subset with improved mean performance compared to the multi-model mean while at the same time maintaining the spread and addressing the problem of model interdependence. Out-of-sample skill and reliability are demonstrated using model-as-truth experiments. This approach is illustrated with one set of optimisation criteria but we also highlight the flexibility of cost functions, depending on the focus of different users. The technique is useful for a range of applications that, for example, minimise present-day bias to obtain an accurate ensemble mean, reduce dependence in ensemble spread, maximise future spread, ensure good performance of individual models in an ensemble, reduce the ensemble size while maintaining important ensemble characteristics, or optimise several of these at the same time. As in any calibration exercise, the final ensemble is sensitive to the metric, observational product, and pre-processing steps used. 18. Selecting a climate model subset to optimise key ensemble properties Science.gov (United States) Herger, Nadja; Abramowitz, Gab; Knutti, Reto; Angélil, Oliver; Lehmann, Karsten; Sanderson, Benjamin M. 2018-02-01 End users studying impacts and risks caused by human-induced climate change are often presented with large multi-model ensembles of climate projections whose composition and size are arbitrarily determined. An efficient and versatile method that finds a subset which maintains certain key properties from the full ensemble is needed, but very little work has been done in this area. Therefore, users typically make their own somewhat subjective subset choices and commonly use the equally weighted model mean as a best estimate. However, different climate model simulations cannot necessarily be regarded as independent estimates due to the presence of duplicated code and shared development history. Here, we present an efficient and flexible tool that makes better use of the ensemble as a whole by finding a subset with improved mean performance compared to the multi-model mean while at the same time maintaining the spread and addressing the problem of model interdependence. Out-of-sample skill and reliability are demonstrated using model-as-truth experiments. This approach is illustrated with one set of optimisation criteria but we also highlight the flexibility of cost functions, depending on the focus of different users. The technique is useful for a range of applications that, for example, minimise present-day bias to obtain an accurate ensemble mean, reduce dependence in ensemble spread, maximise future spread, ensure good performance of individual models in an ensemble, reduce the ensemble size while maintaining important ensemble characteristics, or optimise several of these at the same time. As in any calibration exercise, the final ensemble is sensitive to the metric, observational product, and pre-processing steps used. 19. Psychological stress during exercise: lymphocyte subset redistribution in firefighters. Science.gov (United States) Huang, Chun-Jung; Webb, Heather E; Garten, Ryan S; Kamimori, Gary H; Acevedo, Edmund O 2010-10-05 The purpose of this study examined the changes in heart rate (HR), catecholamines (NE, EPI) and percentages of blood lymphocyte subsets (CD3+ T cells, CD3+CD4+ helper T cells, CD3+CD8+ cytotoxic T cells, CD3- CD56+ NK cells, CD4/CD8 ratio, CD19+ B cells, and total lymphocytes [NK cells+T cells+B cells]) in firefighters exposed to a computerized firefighting strategies and tactics decision-making challenge while participating in moderate intensity exercise. Furthermore, this study also examined the possible relationships between catecholamines (NE and EPI) and blood lymphocyte subsets following combined mental and physical challenge. Ten professional male firefighters participated in two counterbalanced exercise conditions on a cycle ergometer: (1) 37min of cycle ergometry at 60% VO(2max) (exercise alone condition; EAC) and (2) 37min of cycle ergometry at 60% VO(2max) along with 20min of a computerized firefighting strategies and tactics decision-making challenge (firefighting strategies condition; FSC). FSC elicited significantly greater HR, NE, and EPI when compared to EAC. Both EAC and FSC elicited increases in CD3- CD56+ NK cells. The percentages of CD3+ T cells, CD3+CD4+ helper T cells, CD4/CD8 ratio, CD19+ B cells, and total lymphocytes were lower immediately following both conditions. Following dual challenge NE AUC was negatively correlated with percentage of CD19+ B cells immediately post challenge, and HR was negatively associated with the percent change in the CD4/CD8 ratio from pre to post challenge. These elevations in NE and heart rate simultaneously in response to the dual challenge suggest greater sympathetic activation that in turn would possibly explain the alteration in the distribution of lymphocyte subsets. Published by Elsevier Inc. 20. Development of a high-quality cut-off wall using electrophoresis International Nuclear Information System (INIS) Kawachi, T.; Murahashi, H. 1991-01-01 Techniques to build a high-quality cut-off wall have been developed for storage facilities of low-level radioactive waste (LLW) as an emergency measures to prevent leakages. The cut-off wall is highly impermeable, nucleid-adsorptive and have long-term durability. Electrophoresis is used to form impermeable membrane of bentonite as main features of the cut-off wall. First of all, laboratory tests have been conducted to study ways of building barriers on site and to collect data on the barriers properties. Afterwards, on-site construction tests of a high-quality cut-off wall have been carried out. In this paper, we describe the process and results on the studies of the high-quality cut-off wall using electrophoresis 1. Analytical method for high resolution liquid chromatography for quality control French Macaw International Nuclear Information System (INIS) Garcia Penna, Caridad M; Torres Amaro, Leonid; Menendez Castillo, Rosa; Sanchez, Esther; Martinez Espinosa, Vivian; Gonzalez, Maria Lidia; Rodriguez, Carlos 2007-01-01 Was developed and validated an analytical method for high resolution liquid chromatography applicable to quality control of drugs dry French Macaw (Senna alata L. Roxb.) With ultraviolet detection at 340 nm. The method for high resolution liquid chromatography used to quantify the sennosides A and B, main components, was validated and proved to be specific, linear, precise and accurate. (Author) 2. High quality factor GaAs microcavity with buried bullseye defects DEFF Research Database (Denmark) Winkler, K.; Gregersen, Niels; Hayrynen, T. 2018-01-01 The development of high quality factor solid-state microcavities with low mode volumes has paved the way towards on-chip cavity quantum electrodynamics experiments and the development of high-performance nanophotonic devices. Here, we report on the implementation of a new kind of solid... 3. Quality and performance of laser cutting with a high power SM fiber laser DEFF Research Database (Denmark) Kristiansen, Morten; Selchau, Jacob; Olsen, F. O. 2013-01-01 The introduction of high power single mode fiber lasers allows for a beam of high power and a good beam quality factor (M2 ” 1.2), compared to the multimode fiber lasers often utilised in macro laser metal cutting. This paper describes fundamental studies of macro laser metal cutting with a singl... 4. Effects of High School Students' Perceptions of School Life Quality on Their Academic Motivation Levels Science.gov (United States) Akin Kösterelioglu, Meltem; Kösterelioglu, Ilker 2015-01-01 This study aims to identify the effects of high school students' perceptions of school life quality on their academic motivation levels. The study was conducted on a sample of high school students (n = 2371) in Amasya Province in the fall semester of 2013-2014 academic year. Study sample was selected with the help of cluster sampling method. Data… 5. NKT Cell Subsets Can Exert Opposing Effects in Autoimmunity, Tumor Surveillance and Inflammation Science.gov (United States) Viale, Rachael; Ware, Randle; Maricic, Igor; Chaturvedi, Varun; Kumar, Vipin 2014-01-01 The innate-like natural killer T (NKT) cells are essential regulators of immunity. These cells comprise at least two distinct subsets and recognize different lipid antigens presented by the MHC class I like molecules CD1d. The CD1d-dependent recognition pathway of NKT cells is highly conserved from mouse to humans. While most type I NKT cells can recognize αGalCer and express a semi-invariant T cell receptor (TCR), a major population of type II NKT cells reactive to sulfatide utilizes an oligoclonal TCR. Furthermore TCR recognition features of NKT subsets are also distinctive with almost parallel as opposed to perpendicular footprints on the CD1d molecules for the type I and type II NKT cells respectively. Here we present a view based upon the recent studies in different clinical and experimental settings that while type I NKT cells are more often pathogenic, they may also be regulatory. On the other hand, sulfatide-reactive type II NKT cells mostly play an inhibitory role in the control of autoimmune and inflammatory diseases. Since the activity and cytokine secretion profiles of NKT cell subsets can be modulated differently by lipid ligands or their analogs, novel immunotherapeutic strategies are being developed for their differential activation for potential intervention in inflammatory diseases. PMID:25288922 6. REST represses a subset of the pancreatic endocrine differentiation program DEFF Research Database (Denmark) Martin, David; Kim, Yung-Hae; Sever, Dror 2015-01-01 in neurons and in endocrine cells, which is necessary for their normal function. During development, REST represses a subset of genes in the neuronal differentiation program and Rest is down-regulated as neurons differentiate. Here, we investigate the role of REST in the differentiation of pancreatic...... endocrine cells, which are molecularly close to neurons. We show that Rest is widely expressed in pancreas progenitors and that it is down-regulated in differentiated endocrine cells. Sustained expression of REST in Pdx1(+) progenitors impairs the differentiation of endocrine-committed Neurog3... 7. Diagnosing Warm Frontal Cloud Formation in a GCM: A Novel Approach Using Conditional Subsetting Science.gov (United States) Booth, James F.; Naud, Catherine M.; DelGenio, Anthony D. 2013-01-01 This study analyzes characteristics of clouds and vertical motion across extratropical cyclone warm fronts in the NASA Goddard Institute for Space Studies general circulation model. The validity of the modeled clouds is assessed using a combination of satellite observations from CloudSat, Cloud-Aerosol Lidar and Infrared Pathfinder Satellite Observations (CALIPSO), Advanced Microwave Scanning Radiometer for Earth Observing System (AMSR-E), and the NASA Modern-Era Retrospective Analysis for Research and Applications (MERRA) reanalysis. The analysis focuses on developing cyclones, to test the model's ability to generate their initial structure. To begin, the extratropical cyclones and their warm fronts are objectively identified and cyclone-local fields are mapped into a vertical transect centered on the surface warm front. To further isolate specific physics, the cyclones are separated using conditional subsetting based on additional cyclone-local variables, and the differences between the subset means are analyzed. Conditional subsets are created based on 1) the transect clouds and 2) vertical motion; 3) the strength of the temperature gradient along the warm front, as well as the storm-local 4) wind speed and 5) precipitable water (PW). The analysis shows that the model does not generate enough frontal cloud, especially at low altitude. The subsetting results reveal that, compared to the observations, the model exhibits a decoupling between cloud formation at high and low altitudes across warm fronts and a weak sensitivity to moisture. These issues are caused in part by the parameterized convection and assumptions in the stratiform cloud scheme that are valid in the subtropics. On the other hand, the model generates proper covariability of low-altitude vertical motion and cloud at the warm front and a joint dependence of cloudiness on wind and PW. 8. High dose-rate brachytherapy source position quality assurance using radiochromic film International Nuclear Information System (INIS) Evans, M.D.C.; Devic, S.; Podgorsak, E.B. 2007-01-01 9. Production of high-quality polydisperse construction mixes for additive 3D technologies. Science.gov (United States) Gerasimov, M. D.; Brazhnik, Yu V.; Gorshkov, P. S.; Latyshev, S. S. 2018-03-01 The paper describes a new design of a mixer allowing production of high quality polydisperse powders, used in additive 3D technologies. A new principle of dry powder particle mixing is considered, implementing a possibility of a close-to-ideal distribution of such particles in common space. A mathematical model of the mixer is presented, allowing evaluating quality indicators of the produced mixture. Experimental results are shown and rational values of process parameters of the mixer are obtained. 10. Teaching Surgical Procedures with Movies: Tips for High-quality Video Clips OpenAIRE Jacquemart, Mathieu; Bouletreau, Pierre; Breton, Pierre; Mojallal, Ali; Sigaux, Nicolas 2016-01-01 Summary: Video must now be considered as a precious tool for learning surgery. However, the medium does present production challenges, and currently, quality movies are not always accessible. We developed a series of 7 surgical videos and made them available on a publicly accessible internet website. Our videos have been viewed by thousands of people worldwide. High-quality educational movies must respect strategic and technical points to be reliable. 11. Teaching Surgical Procedures with Movies: Tips for High-quality Video Clips. Science.gov (United States) Jacquemart, Mathieu; Bouletreau, Pierre; Breton, Pierre; Mojallal, Ali; Sigaux, Nicolas 2016-09-01 Video must now be considered as a precious tool for learning surgery. However, the medium does present production challenges, and currently, quality movies are not always accessible. We developed a series of 7 surgical videos and made them available on a publicly accessible internet website. Our videos have been viewed by thousands of people worldwide. High-quality educational movies must respect strategic and technical points to be reliable. 12. Timing of high-quality child care and cognitive, language, and preacademic development OpenAIRE Li, W; Farkas, G; Duncan, GJ; Burchinal, MR; Vandell, DL 2013-01-01 The effects of high- versus low-quality child care during 2 developmental periods (infant-toddlerhood and preschool) were examined using data from the National Institute of Child Health and Human Development Study of Early Child Care. Propensity score matching was used to account for differences in families who used different combinations of child care quality during the 2 developmental periods. Findings indicated that cognitive, language, and preacademic skills prior to school entry were hig... 13. High-quality poly-dispersed mixtures applied in additive 3D technologies. Science.gov (United States) Gerasimov, M. D.; Brazhnik, Yu V.; Gorshkov, P. S.; Latyshev, S. S. 2018-03-01 The paper describes the new mixer design to obtain high-quality poly-dispersed powders applied in additive 3D technologies. It also considers a new mixing principle of dry powder particles ensuring the distribution of such particles in the total volume, which is close to ideal. The paper presents the mathematical model of mixer operation providing for the quality assessment of the ready mixtures. Besides, it demonstrates experimental results and obtained rational values of mixer process parameters. 14. CD45RC isoform expression identifies functionally distinct T cell subsets differentially distributed between healthy individuals and AAV patients. Directory of Open Access Journals (Sweden) Laurence Ordonez Full Text Available In animal models of anti-neutrophil cytoplasmic antibody (ANCA-associated vasculitis (AAV, the proportion of CD45RC T cell subsets is important for disease susceptibility. Their human counterparts are, however, functionally ill defined. In this report, we studied their distribution in healthy controls (HC, AAV patients and in Systemic lupus erythematous (SLE patients as disease controls. We showed that CD45RC expression level on human CD4 and CD8 T cells identifies subsets that are highly variable among individuals. Interestingly, AAV patients exhibit an increased proportion of CD45RC(low CD4 T cells as compared to HC and SLE patients. This increase is stable over time and independent of AAV subtype, ANCA specificity, disease duration, or number of relapses. We also analyzed the cytokine profile of purified CD4 and CD8 CD45RC T cell subsets from HC, after stimulation with anti-CD3 and anti-CD28 mAbs. The CD45RC subsets exhibit different cytokine profiles. Type-1 cytokines (IL-2, IFN-gamma and TNF-alpha were produced by all CD45RC T cell subsets, while the production of IL-17, type-2 (IL-4, IL-5 and regulatory (IL-10 cytokines was restricted to the CD45RC(low subset. In conclusion, we have shown that CD45RC expression divides human T cells in functionally distinct subsets that are imbalanced in AAV. Since this imbalance is stable over time and independent of several disease parameters, we hypothesize that this is a pre-existing immune abnormality involved in the etiology of AAV. 15. Isolation of high-quality total RNA from leaves of Myrciaria dubia "CAMU CAMU". Science.gov (United States) Gómez, Juan Carlos Castro; Reátegui, Alina Del Carmen Egoavil; Flores, Julián Torres; Saavedra, Roberson Ramírez; Ruiz, Marianela Cobos; Correa, Sixto Alfredo Imán 2013-01-01 Myrciaria dubia is a main source of vitamin C for people in the Amazon region. Molecular studies of M. dubia require high-quality total RNA from different tissues. So far, no protocols have been reported for total RNA isolation from leaves of this species. The objective of this research was to develop protocols for extracting high-quality total RNA from leaves of M. dubia. Total RNA was purified following two modified protocols developed for leaves of other species (by Zeng and Yang, and by Reid et al.) and one modified protocol developed for fruits of the studied species (by Silva). Quantity and quality of purified total RNA were assessed by spectrophotometric and electrophoretic analysis. Additionally, quality of total RNA was evaluated with reverse-transcription polymerase chain reaction (RT-PCR). With these three modified protocols we were able to isolate high-quality RNA (A260nm/A280nm >1.9 and A260nm/A230nm >2.0). Highest yield was produced with the Zeng and Yang modified protocol (384±46µg ARN/g fresh weight). Furthermore, electrophoretic analysis showed the integrity of isolated RNA and the absence of DNA. Another proof of the high quality of our purified RNA was the successful cDNA synthesis and amplification of a segment of the M. dubia actin 1 gene. We report three modified protocols for isolation total RNA from leaves of M. dubia. The modified protocols are easy, rapid, low in cost, and effective for high-quality and quantity total RNA isolation suitable for cDNA synthesis and polymerase chain reaction. 16. A Hybrid Soft Computing Approach for Subset Problems Directory of Open Access Journals (Sweden) Broderick Crawford 2013-01-01 Full Text Available Subset problems (set partitioning, packing, and covering are formal models for many practical optimization problems. A set partitioning problem determines how the items in one set (S can be partitioned into smaller subsets. All items in S must be contained in one and only one partition. Related problems are set packing (all items must be contained in zero or one partitions and set covering (all items must be contained in at least one partition. Here, we present a hybrid solver based on ant colony optimization (ACO combined with arc consistency for solving this kind of problems. ACO is a swarm intelligence metaheuristic inspired on ants behavior when they search for food. It allows to solve complex combinatorial problems for which traditional mathematical techniques may fail. By other side, in constraint programming, the solving process of Constraint Satisfaction Problems can dramatically reduce the search space by means of arc consistency enforcing constraint consistencies either prior to or during search. Our hybrid approach was tested with set covering and set partitioning dataset benchmarks. It was observed that the performance of ACO had been improved embedding this filtering technique in its constructive phase. 17. Changes in T-cell subsets after radiation therapy International Nuclear Information System (INIS) Yang, S.J.; Rafla, S.; Youssef, E.; Selim, H.; Salloum, N.; Chuang, J.Y. 1988-01-01 The T-cell subsets of 129 patients with cancer were counted before and after radiation therapy. The cells were labeled with monoclonal antibodies that were specific for each type of T cell. Significant changes after therapy were decreases in the proportion of T-helper/inducer cells, pan-T cells, and in the ratio of T-helper/inducer to T-suppressor/cytotoxic cells. There was an increase in the percentage of T-suppressor/cytotoxic cells. When the site of the primary cancer was considered, genitourinary cancer and cancer of the head and neck both showed a decreased percentage of T-helper/inducer cells and a reduced ratio of T-helper/inducer to T-suppressor/cytotoxic cells. The percentage of pan-T cells in head and neck cancer and the ratio of T-helper/inducer to T-suppressor/cytotoxic cells in breast cancer were decreased. The percentage of T-helper cells was particularly decreased by radiation therapy in advanced stages of cancer, in higher grade tumors, and in larger tumors. The absolute numbers of various T-cell subsets were decreased in all groups 18. The regulatory roles of B cell subsets in transplantation. Science.gov (United States) Chu, Zhulang; Zou, Weilong; Xu, Yanan; Sun, Qiquan; Zhao, Yong 2018-02-01 B cells mediate allograft rejection through antigen presentation, and production of cytokines and antibodies. More and more immunosuppressive agents specifically targeting B cells and plasma cells have been applied in clinical transplantation. However, recent studies have indicated the regulatory roles of B cells. Therefore, it is vital to clarify the different effects of B cell subsets in organ transplantation so that we can completely understand the diverse functions of B cells in transplantation. Areas covered: This review focuses on the regulatory roles of B cells in transplantation. B cell subsets with immune modulation and factors mediating immunosuppressive functions of regulatory B (Breg) cells were analyzed. Therapies targeting B cells and the application of B cells for transplant tolerance induction were discussed. Expert commentary: Besides involving rejection, B cells could also play regulatory roles in transplantation. Breg cells and the related markers may be used to predict the immune tolerant state in transplant recipients. New therapeutic strategies targeting B cells should be explored to promote tolerance induction with less impact on the host's protective immunity in organ transplanted patients. 19. Mast cell subsets and neuropeptides in leprosy reactions Directory of Open Access Journals (Sweden) Antunes Sérgio Luiz Gomes 2003-01-01 Full Text Available The immunohistochemical identification of neuropeptides (calcitonin gene-related peptide, vasoactive intestinal polypeptide, substance P, alpha-melanocyte stimulating hormone and gamma-melanocyte stimulating hormone quantification of mast cells and their subsets (tryptase/chymase-immunoreactive mast cells = TCMC and tryptase-immunoreactive mast cells = TMC were determined in biopsies of six patients with leprosy reactions (three patients with type I reaction and three with type II. Biopsies were compared with those taken from the same body site in the remission stage of the same patient. We found a relative increase of TMC in the inflammatory infiltrate of the reactional biopsies compared to the post-reactional biopsy. Also, the total number of mast cells and the TMC/TCMC ratio in the inflammatory infiltrate was significantly higher than in the intervening dermis of the biopsies of both periods. No significant difference was found regarding neuroptide expression in the reactional and post-reactional biopsies. The relative increase of TMC in the reactional infiltrates could implicate this mast cell subset in the reported increase of the immune response in leprosy reactions. 20. Assessment Sleep Quality and its Relationship with Test Anxiety among High School Students in Qom- Iran Directory of Open Access Journals (Sweden) Saman barmeh ziyar 2016-10-01 Full Text Available Background Test anxiety is a special case of a general anxiety which is of particular importance in students, because students will be the future of the country and the society activists. On the other hand, sleep quality and sleep disorders, have correlation with ailments, poor performance, decreased quality of life and increase of associated costs; This study aimed to determine the quality of sleep and its relationship with test anxiety among students in Qom city, Iran. Materials and Methods This study was a cross-sectional study, which was performed among 250 students who were going to pass the exam preparation classes. In order to collect data Pittsburgh Sleep Quality Index (PSQI questionnaires and Test Anxiety Inventory (TAI questionnaire were used. Data were analyzed using SPSS-16 with descriptive statistics and statistical methods, independent t-test, ANOVA and Pearson correlation coefficient. Results In this study, 50% of participants were boys (n=125 and 50 percent were girls (n=125. 81.4% of subjects had poor sleep quality and 69.6% had average to high score for test anxiety. Based on the results of anxiety test and sleep quality index there was a significant correlation between anxiety and sleep quality with gender (P=0.003, r=0.447. Conclusion School children had poor sleep quality and high test anxiety, and due to their direct and significant correlation, attention to this category of students, especially for girls, is important. Therefore, anxiety and promoting sleep quality control programs are recommended in this group. 1. Soft X-ray generation via inverse compton scattering between high quality electron beam and high power laser International Nuclear Information System (INIS) Masakazu Washio; Kazuyuki Sakaue; Yoshimasa Hama; Yoshio Kamiya; Tomoko Gowa; Akihiko Masuda; Aki Murata; Ryo Moriyama; Shigeru Kashiwagi; Junji Urakawa 2007-01-01 High quality beam generation project based on High-Tech Research Center Project, which has been approved by Ministry of Education, Culture, Sports, Science and Technology in 1999, has been conducted by advance research institute for science and engineering, Waseda University. In the project, laser photo-cathode RF-gun has been selected for the high quality electron beam source. RF cavities with low dark current, which were made by diamond turning technique, have been successfully manufactured. The low emittance electron beam was realized by choosing the modified laser injection technique. The obtained normalized emmitance was about 3 m.mrad at 100 pC of electron charge. The soft x-ray beam generation with the energy of 370 eV, which is in the energy region of so-called water window, by inverse Compton scattering has been performed by the collision between IR laser and the low emmitance electron beams. (Author) 2. The Barriers to High-Quality Inpatient Pain Management: A Qualitative Study. Science.gov (United States) Lin, Richard J; Reid, M Carrington; Liu, Lydia L; Chused, Amy E; Evans, Arthur T 2015-09-01 The current literature suggests deficiencies in the quality of acute pain management among general medical inpatients. The aim of this qualitative study is to identify potential barriers to high-quality acute pain management among general medical inpatients at an urban academic medical center during a 2-year period. Data are collected using retrospective chart reviews, survey questionnaires, and semistructured, open-ended interviews of 40 general medical inpatients who have experienced pain during their hospitalization. Our results confirm high prevalence and disabling impacts of pain and significant patient- and provider-related barriers to high-quality acute pain management. We also identify unique system-related barriers such as time delay and pain management culture. Efforts to improve the pain management experience of general medical inpatients will need to address all these barriers. © The Author(s) 2014. 3. Quality Evaluation of Potentilla fruticosa L. by High Performance Liquid Chromatography Fingerprinting Associated with Chemometric Methods. Science.gov (United States) Liu, Wei; Wang, Dongmei; Liu, Jianjun; Li, Dengwu; Yin, Dongxue 2016-01-01 The present study was performed to assess the quality of Potentilla fruticosa L. sampled from distinct regions of China using high performance liquid chromatography (HPLC) fingerprinting coupled with a suite of chemometric methods. For this quantitative analysis, the main active phytochemical compositions and the antioxidant activity in P. fruticosa were also investigated. Considering the high percentages and antioxidant activities of phytochemicals, P. fruticosa samples from Kangding, Sichuan were selected as the most valuable raw materials. Similarity analysis (SA) of HPLC fingerprints, hierarchical cluster analysis (HCA), principle component analysis (PCA), and discriminant analysis (DA) were further employed to provide accurate classification and quality estimates of P. fruticosa. Two principal components (PCs) were collected by PCA. PC1 separated samples from Kangding, Sichuan, capturing 57.64% of the variance, whereas PC2 contributed to further separation, capturing 18.97% of the variance. Two kinds of discriminant functions with a 100% discrimination ratio were constructed. The results strongly supported the conclusion that the eight samples from different regions were clustered into three major groups, corresponding with their morphological classification, for which HPLC analysis confirmed the considerable variation in phytochemical compositions and that P. fruticosa samples from Kangding, Sichuan were of high quality. The results of SA, HCA, PCA, and DA were in agreement and performed well for the quality assessment of P. fruticosa. Consequently, HPLC fingerprinting coupled with chemometric techniques provides a highly flexible and reliable method for the quality evaluation of traditional Chinese medicines. 4. Quality Evaluation of Potentilla fruticosa L. by High Performance Liquid Chromatography Fingerprinting Associated with Chemometric Methods Science.gov (United States) Liu, Wei; Wang, Dongmei; Liu, Jianjun; Li, Dengwu; Yin, Dongxue 2016-01-01 The present study was performed to assess the quality of Potentilla fruticosa L. sampled from distinct regions of China using high performance liquid chromatography (HPLC) fingerprinting coupled with a suite of chemometric methods. For this quantitative analysis, the main active phytochemical compositions and the antioxidant activity in P. fruticosa were also investigated. Considering the high percentages and antioxidant activities of phytochemicals, P. fruticosa samples from Kangding, Sichuan were selected as the most valuable raw materials. Similarity analysis (SA) of HPLC fingerprints, hierarchical cluster analysis (HCA), principle component analysis (PCA), and discriminant analysis (DA) were further employed to provide accurate classification and quality estimates of P. fruticosa. Two principal components (PCs) were collected by PCA. PC1 separated samples from Kangding, Sichuan, capturing 57.64% of the variance, whereas PC2 contributed to further separation, capturing 18.97% of the variance. Two kinds of discriminant functions with a 100% discrimination ratio were constructed. The results strongly supported the conclusion that the eight samples from different regions were clustered into three major groups, corresponding with their morphological classification, for which HPLC analysis confirmed the considerable variation in phytochemical compositions and that P. fruticosa samples from Kangding, Sichuan were of high quality. The results of SA, HCA, PCA, and DA were in agreement and performed well for the quality assessment of P. fruticosa. Consequently, HPLC fingerprinting coupled with chemometric techniques provides a highly flexible and reliable method for the quality evaluation of traditional Chinese medicines. PMID:26890416 5. Integrated Pathology Informatics Enables High-Quality Personalized and Precision Medicine: Digital Pathology and Beyond. Science.gov (United States) Volynskaya, Zoya; Chow, Hung; Evans, Andrew; Wolff, Alan; Lagmay-Traya, Cecilia; Asa, Sylvia L 2018-03-01 - The critical role of pathology in diagnosis, prognosis, and prediction demands high-quality subspecialty diagnostics that integrates information from multiple laboratories. - To identify key requirements and to establish a systematic approach to providing high-quality pathology in a health care system that is responsible for services across a large geographic area. - This report focuses on the development of a multisite pathology informatics platform to support high-quality surgical pathology and hematopathology using a sophisticated laboratory information system and whole slide imaging for histology and immunohistochemistry, integrated with ancillary tools, including electron microscopy, flow cytometry, cytogenetics, and molecular diagnostics. - These tools enable patients in numerous geographic locations access to a model of subspecialty pathology that allows reporting of every specimen by the right pathologist at the right time. The use of whole slide imaging for multidisciplinary case conferences enables better communication among members of patient care teams. The system encourages data collection using a discrete data synoptic reporting module, has implemented documentation of quality assurance activities, and allows workload measurement, providing examples of additional benefits that can be gained by this electronic approach to pathology. - This approach builds the foundation for accurate big data collection and high-quality personalized and precision medicine. 6. Developing Quality Control Procedures to Sustain a Supply of High Quality Blood for Mass Rearing Tsetse Flies Energy Technology Data Exchange (ETDEWEB) De Beer, C J; Venter, G J; Potgieter, F T [ARC-Onderstepoort Veterinary Institute, Old Soutpans Road, Private Bag X05, 0110 Onderstepoort (South Africa) 2012-07-15 Mass rearing tsetse flies Glossina spp. is dependent on the sustained availability of a high quality blood diet. In any mass rearing facility, the logistics for obtaining sterile, high quality fresh blood is challenging. An added complication is the influence of potential chemical, physical and microbiological elements present in the blood of donors, as well as contamination during collection, handling and storage. Research at the Agricultural Research Council - Onderstepoort Veterinary institute (ARC-OVI) is directed towards the development of quality control procedures for the supply of the in vitro diet used to maintain productive colonies of Glossina brevipalpis Newstead and Glossina austeni Newstead. Factors that may influence the blood diet, e.g. defibrination, feeding times, collection of blood in anticoagulants, treatment of blood with taste stimuli, repeated freezing and thawing of blood, effect of bovine growth hormones, and also a preference for bovine or porcine blood were tested. A 25 day bioassay was used to determine the effects of these factors on tsetse survival and reproduction. Defibrination of the blood for 10 to 15 minutes gave the best results for both species. It was found that G. brevipalpis should be fed three times per week for 5 minutes each time, and G. austeni three times per week for 10 minutes. Heparin, acid citrate dextrose (ACD), citric acid, citrate phosphate dextrose adenine (CPDA) and a combination of sodium citrate and citric acid were effective anticoagulants in the blood diets of G. brevipalpis and G. austeni. Blood treated with inosine triphosphate (ITP) gave the highest quality factor (QFC) values for both G. austeni and G. brevipalpis. Repeated freezing and thawing of blood definitely affects pupal production negatively; G. brevipalpis especially produced significantly smaller pupae. A premixed diet of equal amounts of bovine and porcine blood was found to be best suited for G. brevipalpis, and for G. austeni a mixture of 7. Development of High Quality 4H-SiC Thick Epitaxy for Reliable High Power Electronics Using Halogenated Precursors Science.gov (United States) 2016-08-02 defects :=()llowed by a second buffer epilayer gro\\\\ th with ~: urn thickness with high n-type dopi:1g (- 5£17 cm-3) for the same C/Si ratio of ~1.4 at...gradient, pressure , etc.) can further reduce the parasitic deposition, especially in TFS-growth. • Thick epitaxy on-axis 4H-SiC Growth at High Growth...From - To) 08/02/2016 Final Technical Report 01-Apr-10 Through 31-Mar-14 4. TITLE AND SUBTITLE 5a. CONTRACT NUMBER Development of High Quality 4H 8. Study on quality assurance for high-level radioactive waste disposal project (2). Quality assurance system for the site characterization phase in the Yucca Mountain Project International Nuclear Information System (INIS) 2006-01-01 The objective of this report is to assist related organizations in the development of quality assurance systems for a high-level radioactive waste disposal system. This report presents detail information with which related organizations can begin the development of quality assurance systems at an initial phase of repository development for a high-level radioactive waste disposal program, including data qualification, model validation, systems and facilities for quality assurance (e.g., technical data management system, sample management facility, etc.), and QA program applicability (items and activities). These descriptions are based on information in QA program for the Yucca Mountain Project (YMP), such as the U.S. Department of Energy (DOE) Quality Assurance Requirements and Description (QARD), DOE/RW-0333P, quality implementing procedures, and reports implemented by the procedures. Additionally, this report includes some brief recommendations for developing of quality assurance systems, such as establishment of quality assurance requirements, measures for establishment of QA system. (author) 9. Highly purified water production technology. The influence of water purity on steam quality International Nuclear Information System (INIS) Ganter, J. 1975-01-01 The fundamental question related to high-pressure steam generation, intended for powering steam turbines, concerns steam production conditions based on constant quality standards. The characteristics of water (salinity, silica concentration) are indicated for a given steam quality as a function of the pressure. Two processes for the purification of feedwater for high pressure boilers are described: a treatment using precoated cellulose or resin filters and a treatment using mixed-bed ion exchangers. When ultrapure water is required, the demineralized water is filtred using microfiltration and ultrafiltration processes [fr 10. Stainless steel fabrication for high quality requirements in the nuclear industry International Nuclear Information System (INIS) Wareing, A.J. 1990-01-01 In this paper the author explains the welding procedures and practices adopted within the nuclear industry to achieve the high quality and standards of welds required. The changeover to mechanised welding, orbital TIG welding and synergic MIG welding, has resulted in consistent achievement of high quality standards as well as optimising the productivity. However, the use of mechanised welding machines does require the welder operating them to be fully trained and qualified. The formally organised training courses are described and the cost savings and production rates achieved by utilising the mechanised method are discussed. (author) 11. Different Polar Metabolites and Protein Profiles between High- and Low-Quality Japanese Ginjo Sake. Directory of Open Access Journals (Sweden) Kei Takahashi Full Text Available Japanese ginjo sake is a premium refined sake characterized by a pleasant fruity apple-like flavor and a sophisticated taste. Because of technical difficulties inherent in brewing ginjo sake, off-flavors sometimes occur. However, the metabolites responsible for off-flavors as well as those present or absent in higher quality ginjo sake remain uncertain. Here, the relationship between 202 polar chemical compounds in sake identified using capillary electrophoresis coupled with time-of-flight mass spectrometry and its organoleptic properties, such as quality and off-flavor, was examined. First, we found that some off-flavored sakes contained higher total amounts of metabolites than other sake samples. The results also identified that levels of 2-oxoglutaric acid and fumaric acid, metabolites in the tricarboxylic acid cycle, were highly but oppositely correlated with ginjo sake quality. Similarly, pyridoxine and pyridoxamine, co-enzymes for amino transferase, were also highly but oppositely correlated with ginjo sake quality. Additionally, pyruvic acid levels were associated with good quality as well. Compounds involved in the methionine salvage cycle, oxidative glutathione derivatives, and amino acid catabolites were correlated with low quality. Among off-flavors, an inharmonious bitter taste appeared attributable to polyamines. Furthermore, protein analysis displayed that a diversity of protein components and yeast protein (triosephosphate isomerase, TPI leakage was linked to the overall metabolite intensity in ginjo sake. This research provides insight into the relationship between sake components and organoleptic properties. 12. Effect of high-oxygen and high-carbon-dioxide atmospheres on strawberry flavor and other quality traits. Science.gov (United States) Pérez, A G; Sanz, C 2001-05-01 The effect of high-oxygen atmospheres on strawberry flavor was studied. Strawberry fruits (Fragariax ananassa Duch. cv. Camarosa) were stored at 8 degrees C in four different atmospheres: air, 5% O(2)/20% CO(2), 80% O(2)/20% CO(2), and 90% O(2)/10% CO(2). Changes in several quality parameters were evaluated. Atmospheres combining high O(2) and high CO(2) were the most effective in preventing fungal growth and enhancing strawberry firmness. Other quality parameters such as color, titrable acidity, sugars and organic acids distribution, off-flavor development, and aroma were only mildly affected by superatmospheric O(2) levels. After one week of storage, unexpected high contents of off-flavor related compounds were found in the 80% O(2)/20% CO(2) and 90% O(2)/10% CO(2) atmospheres. Evidence of an altered ester biosynthesis was also found in fruits stored under these high-O(2) atmospheres. Data obtained suggest that stress induced by high CO(2) and stress induced by high O(2) have an additive effect on strawberry flavor alteration. 13. Exome sequencing generates high quality data in non-target regions Directory of Open Access Journals (Sweden) Guo Yan 2012-05-01 Full Text Available Abstract Background Exome sequencing using next-generation sequencing technologies is a cost efficient approach to selectively sequencing coding regions of human genome for detection of disease variants. A significant amount of DNA fragments from the capture process fall outside target regions, and sequence data for positions outside target regions have been mostly ignored after alignment. Result We performed whole exome sequencing on 22 subjects using Agilent SureSelect capture reagent and 6 subjects using Illumina TrueSeq capture reagent. We also downloaded sequencing data for 6 subjects from the 1000 Genomes Project Pilot 3 study. Using these data, we examined the quality of SNPs detected outside target regions by computing consistency rate with genotypes obtained from SNP chips or the Hapmap database, transition-transversion (Ti/Tv ratio, and percentage of SNPs inside dbSNP. For all three platforms, we obtained high-quality SNPs outside target regions, and some far from target regions. In our Agilent SureSelect data, we obtained 84,049 high-quality SNPs outside target regions compared to 65,231 SNPs inside target regions (a 129% increase. For our Illumina TrueSeq data, we obtained 222,171 high-quality SNPs outside target regions compared to 95,818 SNPs inside target regions (a 232% increase. For the data from the 1000 Genomes Project, we obtained 7,139 high-quality SNPs outside target regions compared to 1,548 SNPs inside target regions (a 461% increase. Conclusions These results demonstrate that a significant amount of high quality genotypes outside target regions can be obtained from exome sequencing data. These data should not be ignored in genetic epidemiology studies. 14. Providing high-quality care in primary care settings: how to make trade-offs. Science.gov (United States) Beaulieu, Marie-Dominique; Geneau, Robert; Del Grande, Claudio; Denis, Jean-Louis; Hudon, Eveline; Haggerty, Jeannie L; Bonin, Lucie; Duplain, Réjean; Goudreau, Johanne; Hogg, William 2014-05-01 To gain a deeper understanding of how primary care (PC) practices belonging to different models manage resources to provide high-quality care. Multiple-case study embedded in a cross-sectional study of a random sample of 37 practices. Three regions of Quebec. Health care professionals and staff of 5 PC practices. Five cases showing above-average results on quality-of-care indicators were purposefully selected to contrast on region, practice size, and PC model. Data were collected using an organizational questionnaire; the Team Climate Inventory, which was completed by health care professionals and staff; and 33 individual interviews. Detailed case histories were written and thematic analysis was performed. The core common feature of these practices was their ongoing effort to make trade-offs to deliver services that met their vision of high-quality care. These compromises involved the same 3 areas, but to varying degrees depending on clinic characteristics: developing a shared vision of high-quality care; aligning resource use with that vision; and balancing professional aspirations and population needs. The leadership of the physician lead was crucial. The external environment was perceived as a source of pressure and dilemmas rather than as a source of support in these matters. Irrespective of their models, PC practices' pursuit of high-quality care is based on a vision in which accessibility is a key component, balanced by appropriate management of available resources and of external environment expectations. Current PC reforms often create tensions rather than support PC practices in their pursuit of high-quality care. Copyright© the College of Family Physicians of Canada. 15. Methods for slow axis beam quality improvement of high power broad area diode lasers Science.gov (United States) An, Haiyan; Xiong, Yihan; Jiang, Ching-Long J.; Schmidt, Berthold; Treusch, Georg 2014-03-01 For high brightness direct diode laser systems, it is of fundamental importance to improve the slow axis beam quality of the incorporated laser diodes regardless what beam combining technology is applied. To further advance our products in terms of increased brightness at a high power level, we must optimize the slow axis beam quality despite the far field blooming at high current levels. The later is caused predominantly by the built-in index step in combination with the thermal lens effect. Most of the methods for beam quality improvements reported in publications sacrifice the device efficiency and reliable output power. In order to improve the beam quality as well as maintain the efficiency and reliable output power, we investigated methods of influencing local heat generation to reduce the thermal gradient across the slow axis direction, optimizing the built-in index step and discriminating high order modes. Based on our findings, we have combined different methods in our new device design. Subsequently, the beam parameter product (BPP) of a 10% fill factor bar has improved by approximately 30% at 7 W/emitter without efficiency penalty. This technology has enabled fiber coupled high brightness multi-kilowatt direct diode laser systems. In this paper, we will elaborate on the methods used as well as the results achieved. 16. Microstructural and Material Quality Effects on Rolling Contact Fatigue of Highly Elastic Intermetallic Ball Bearings Science.gov (United States) DellaCorte, Christopher; Howard, S. Adam; Thomas, Fransua; Stanford, Malcolm K. 2016-01-01 Rolling element bearings made from highly-elastic intermetallic materials (HIM)s, such as 60NiTi, are under development for applications that require superior corrosion and shock resistance. Compared to steel, intermetallics have been shown to have much lower rolling contact fatigue (RCF) stress capability in simplified 3-ball on rod (ASTM STP 771) fatigue tests. In the 3-ball tests, poor material quality and microstructural flaws negatively affect fatigue life but such relationships have not been established for full-scale 60NiTi bearings. In this paper, 3-ball-on-rod fatigue behavior of two quality grades of 60NiTi are compared to the fatigue life of full-scale 50mm bore ball bearings made from the same materials. 60NiTi RCF rods with material or microstructural flaws suffered from infant mortality failures at all tested stress levels while high quality 60NiTi rods exhibited no failures at lower stress levels. Similarly, tests of full-scale bearings made from flawed materials exhibited early surface fatigue and through crack type failures while bearings made from high quality material did not fail even in long-term tests. Though the full-scale bearing test data is yet preliminary, the results suggest that the simplified RCF test is a good qualitative predictor of bearing performance. These results provide guidance for materials development and to establish minimum quality levels required for successful bearing operation and life. 17. Chemical library subset selection algorithms: a unified derivation using spatial statistics. Science.gov (United States) Hamprecht, Fred A; Thiel, Walter; van Gunsteren, Wilfred F 2002-01-01 If similar compounds have similar activity, rational subset selection becomes superior to random selection in screening for pharmacological lead discovery programs. Traditional approaches to this experimental design problem fall into two classes: (i) a linear or quadratic response function is assumed (ii) some space filling criterion is optimized. The assumptions underlying the first approach are clear but not always defendable; the second approach yields more intuitive designs but lacks a clear theoretical foundation. We model activity in a bioassay as realization of a stochastic process and use the best linear unbiased estimator to construct spatial sampling designs that optimize the integrated mean square prediction error, the maximum mean square prediction error, or the entropy. We argue that our approach constitutes a unifying framework encompassing most proposed techniques as limiting cases and sheds light on their underlying assumptions. In particular, vector quantization is obtained, in dimensions up to eight, in the limiting case of very smooth response surfaces for the integrated mean square error criterion. Closest packing is obtained for very rough surfaces under the integrated mean square error and entropy criteria. We suggest to use either the integrated mean square prediction error or the entropy as optimization criteria rather than approximations thereof and propose a scheme for direct iterative minimization of the integrated mean square prediction error. Finally, we discuss how the quality of chemical descriptors manifests itself and clarify the assumptions underlying the selection of diverse or representative subsets. 18. Multicentre evaluation of stable reference whole blood for enumeration of lymphocyte subsets by flow cytometry. Science.gov (United States) Edwards, Cherry; Belgrave, Danielle; Janossy, George; Bradley, Nicholas J; Stebbings, Richard; Gaines-Das, Rose; Thorpe, Robin; Sawle, Alex; Arroz, Maria Jorge; Brando, Bruno; Gratama, Jan Willem; Orfao de Matos, Alberto; Papa, Stephano; Papamichail, Michael; Lenkei, Rodica; Rothe, Gregor; Barnett, David 2005-06-22 BACKGROUND: Clinical indications for lymphocyte subset enumeration by flow cytometry include monitoring of disease progression and timing of therapeutic intervention in infection with human immunodeficiency virus. Until recently international standardisation has not been possible due to a lack of suitable stable reference material. METHODS: This study consisted of two trials of a stabilised whole blood preparation. Eleven participants were sent two standard protocols for staining plus gating strategy and asked to report absolute counts for lymphocyte subsets. RESULTS: No significant difference was detected between the two methods when results from the two assays and all partners were pooled. Significant differences in results from the different partners were observed. However, representative mean counts were obtained for geometric means, geometric coefficient of variation, and 95% confidence interval for CD3 910 cells/mul, 9%, and 888 to 933, respectively), CD4 (495 cells/mul, 12%, and 483 to 507), and CD8 (408 cells/mul, 13%, and 393 to 422). CONCLUSION: We have introduced a stabilised blood preparation and a well-characterized biological standard. The availability of this reference material greatly simplifies the validation of new techniques for CD4(+) T-cell enumeration and the expansion of external quality assurance programmes for clinical laboratories, including those that operate in resource-restricted environments. (c) 2005 Wiley-Liss, Inc. 19. [Differential diagnosis and atypical subsets of amyotrophic lateral sclerosis]. Science.gov (United States) 2006-06-01 Amyotrophic lateral sclerosis (ALS) is a progressive degeneration of upper and lower motor neurons. In the absence of any validated biological marker, the diagnosis of ALS depends upon recognition of characteristic symptoms and signs together with supportive electrophysiological findings. The diagnosis of ALS is easy to recognize in its fully developed form but during the early stages both false positive and false negative diagnoses are common. In clinical practice, diagnostic difficulties mostly arise with patients who present either with only upper motor neuron, or with only lower motor neuron signs. It may be difficult to distinguish ALS with clinically predominant lower motor neuron involvement from alternative diagnoses including spinal atrophies of adult onset, Kennedy's disease, inclusion body myositis and motor neuropathies with conduction blocks. The diagnosis of ALS related syndromes (progressive muscular atrophy, primary lateral sclerosis and progressive bulbar palsy) requires the elimination of alternate diagnoses. This paper reviews the main characteristics of diseases mimicking ALS and the atypical subsets of ALS. 20. T-lymphocyte subsets, thymic size and breastfeeding in infancy DEFF Research Database (Denmark) Jeppesen, Dorthe Lisbeth; Hasselbalch, Helle; Lisse, Ida M 2004-01-01 We followed the changes in concentration of T-lymphocyte subsets (CD4+ and CD8+ cells) in peripheral blood and thymus size during infancy. Previous studies have found increased thymus size in breastfed infants. The present study analyzed the association between breastfeeding and the number of CD4......+ and CD8+ cells. Two different populations of infants between birth and 1 year of age were examined. Study Group I: infants with a variable duration of breastfeeding. Study Group II: long-term breastfed infants. In both groups a correlation was found between CD8+ cells and the thymic index at 10 months...... to 10 months of age; and a positive correlation between the number of breastfeedings per day at 8 months of age, and an increase in CD4+ cells from 8 to 10 months of age (p Breastfeeding might have both a current and long... 1. Seronegative necrolytic acral erythema: A distinct clinical subset? Directory of Open Access Journals (Sweden) Panda S 2010-01-01 Full Text Available A patient was referred to us with asymptomatic, erythematous, nonitchy, scaly lesions present bilaterally on the dorsa of his feet and toes since the last 2 months. Both the legs had pitting edema as well. There were hyperkeratosis, focal parakeratosis, acanthosis and scattered spongiosis in the epidermis, and proliferation of capillaries with perivascular infiltration of lymphomononuclear cells in the dermis. There was no serological evidence of hepatitis C virus. Laboratory investigations revealed hypoalbuminemia and low-normal serum zinc. On clinicopathological correlation, we made a diagnosis of necrolytic acral erythema (NAE. The lesions responded dramatically to oral zinc sulfate and topical clobetasol propionate within 3 weeks with disappearance of edema and scaling and only a minimal residual erythema. This is the first reported case of NAE from Eastern India. NAE with negative serology for hepatitis C may be viewed as a distinct subset of the condition that had been originally described. 2. Effect of high hydrostatic pressure on overall quality parameters of watermelon juice. Science.gov (United States) Liu, Y; Zhao, X Y; Zou, L; Hu, X S 2013-06-01 High hydrostatic pressure as a kind of non-thermal processing might maintain the quality of thermo-sensitive watermelon juice. So, the effect of high hydrostatic pressure treatment on enzymes and quality of watermelon juice was investigated. After high hydrostatic pressure treatment, the activities of polyphenol oxidase, peroxidase, and pectin methylesterase of juice decreased significantly with the pressure (P  0.05). No significant difference was observed in lycopene and total phenolics after high hydrostatic pressure treatment when compared to the control (P > 0.05). Cloudiness and viscosity increased with pressure (P  0.05). a*- and b*-value both unchanged after high hydrostatic pressure treatment (P > 0.05) while L*-value increased but the values had no significant difference among treated juices. Browning degree after high hydrostatic pressure treatment decreased with increase in pressure and treatment time (P hydrostatic pressure had little effect on color of juice. The results of this study demonstrated the efficacy of high hydrostatic pressure in inactivating enzymes and maintaining the quality of watermelon juice. 3. Characteristics of primary care practices associated with high quality of care. Science.gov (United States) Beaulieu, Marie-Dominique; Haggerty, Jeannie; Tousignant, Pierre; Barnsley, Janet; Hogg, William; Geneau, Robert; Hudon, Éveline; Duplain, Réjean; Denis, Jean-Louis; Bonin, Lucie; Del Grande, Claudio; Dragieva, Natalyia 2013-09-03 No primary practice care model has been shown to be superior in achieving high-quality primary care. We aimed to identify the organizational characteristics of primary care practices that provide high-quality primary care. We performed a cross-sectional observational study involving a stratified random sample of 37 primary care practices from 3 regions of Quebec. We recruited 1457 patients who had 1 of 2 chronic care conditions or 1 of 6 episodic care conditions. The main outcome was the overall technical quality score. We measured organizational characteristics by use of a validated questionnaire and the Team Climate Inventory. Statistical analyses were based on multilevel regression modelling. The following characteristics were strongly associated with overall technical quality of care score: physician remuneration method (27.0; 95% confidence interval [CI] 19.0-35.0), extent of sharing of administrative resources (7.6; 95% CI 0.8-14.4), presence of allied health professionals (15.3; 95% CI 5.4-25.2) and/or specialist physicians (19.6; 95% CI 8.3-30.9), the presence of mechanisms for maintaining or evaluating competence (7.7; 95% CI 3.0-12.4) and average organizational access to the practice (4.9; 95% CI 2.6-7.2). The number of physicians (1.2; 95% CI 0.6-1.8) and the average Team Climate Inventory score (1.3; 95% CI 0.1-2.5) were modestly associated with high-quality care. We identified a common set of organizational characteristics associated with high-quality primary care. Many of these characteristics are amenable to change through practice-level organizational changes. 4. Evaluating Stability and Comparing Output of Feature Selectors that Optimize Feature Subset Cardinality Czech Academy of Sciences Publication Activity Database Somol, Petr; Novovičová, Jana 2010-01-01 Roč. 32, č. 11 (2010), s. 1921-1939 ISSN 0162-8828 R&D Projects: GA MŠk 1M0572; GA ČR GA102/08/0593; GA ČR GA102/07/1594 Grant - others:GA MŠk(CZ) 2C06019 Institutional research plan: CEZ:AV0Z10750506 Keywords : feature selection * feature stability * stability measures * similarity measures * sequential search * individual ranking * feature subset-size optimization * high dimensionality * small sample size Subject RIV: BD - Theory of Information Impact factor: 5.027, year: 2010 http://library.utia.cas.cz/separaty/2010/RO/somol-0348726.pdf 5. Recent developments in high-quality drying of vegetables, fruits, and aquatic products. Science.gov (United States) Zhang, Min; Chen, Huizhi; Mujumdar, Arun S; Tang, Juming; Miao, Song; Wang, Yuchuan 2017-04-13 Fresh foods like vegetables, fruits, and aquatic products have high water activity and they are highly heat-sensitive and easily degradable. Dehydration is one of the most common methods used to improve food shelf-life. However, drying methods used for food dehydration must not only be efficient and economic but also yield high-quality products based on flavor, nutrients, color, rehydration, uniformity, appearance, and texture. This paper reviews some new drying technologies developed for dehydration of vegetables, fruits, and aquatic products. These include: infrared drying, microwave drying, radio frequency drying, electrohydrodynamic drying, etc., as well as hybrid drying methods combining two or more different drying techniques. A comprehensive review of recent developments in high-quality drying of vegetables, fruits and aquatic products is presented and recommendations are made for future research. 6. Formation of a high quality electron beam using photo cathode RF electron gun International Nuclear Information System (INIS) Washio, Masakazu 2000-01-01 Formation of a high quality electron beam using photo cathode RF electron gun is expected for formation of a next generation high brilliant X-ray beam and a source for electron and positron collider. And, on a field of material science, as is possible to carry out an experiment under ultra short pulse and extremely high precision in time, it collects large expectation. Recently, formation of high quality beam possible to develop for multi directions and to use by everyone in future has been able to realize. Here were explained on electron beam source, principle and component on RF electron gun, working features on RF gun, features and simulation of RF gun under operation, and some views in near future. (G.K.) 7. Clinical evaluation of iterative reconstruction (ordered-subset expectation maximization) in dynamic positron emission tomography: quantitative effects on kinetic modeling with N-13 ammonia in healthy subjects DEFF Research Database (Denmark) Hove, Jens D; Rasmussen, Rune; Freiberg, Jacob 2008-01-01 BACKGROUND: The purpose of this study was to investigate the quantitative properties of ordered-subset expectation maximization (OSEM) on kinetic modeling with nitrogen 13 ammonia compared with filtered backprojection (FBP) in healthy subjects. METHODS AND RESULTS: Cardiac N-13 ammonia positron...... emission tomography (PET) studies from 20 normal volunteers at rest and during dipyridamole stimulation were analyzed. Image data were reconstructed with either FBP or OSEM. FBP- and OSEM-derived input functions and tissue curves were compared together with the myocardial blood flow and spillover values...... and OSEM flow values were observed with a flow underestimation of 45% (rest/dipyridamole) in the septum and of 5% (rest) and 15% (dipyridamole) in the lateral myocardial wall. CONCLUSIONS: OSEM reconstruction of myocardial perfusion images with N-13 ammonia and PET produces high-quality images for visual... 8. Health-related quality of life of long-term high-grade glioma survivors NARCIS (Netherlands) Bosma, Ingeborg; Reijneveld, Jaap C.; Douw, Linda; Vos, Maaike J.; Postma, Tjeerd J.; Aaronson, Neil K.; Muller, Martin; Vandertop, W. Peter; Slotman, Ben J.; Taphoorn, Martin J. B.; Heimans, Jan J.; Klein, Martin 2009-01-01 The objective of this study was to compare the health-related quality of life (HRQOL) of long-term to short-term high-grade glioma (HGG) survivors, determine the prognostic value of HRQOL for overall survival, and determine the effect of tumor recurrence on HRQOL for long-term survivors. Following 9. Water quality monitoring for high-priority water bodies in the Sonoran Desert network Science.gov (United States) Terry W. Sprouse; Robert M. Emanuel; Sara A. Strorrer 2005-01-01 This paper describes a network monitoring program for “high priority” water bodies in the Sonoran Desert Network of the National Park Service. Protocols were developed for monitoring selected waters for ten of the eleven parks in the Network. Park and network staff assisted in identifying potential locations of testing sites, local priorities, and how water quality... 10. Producing high-quality negatives from ERTS black-and-white transparancies Science.gov (United States) Richard J. Myhre 1973-01-01 A method has been devised for producing high-quality black-and-white negatives quickly and efficiently from dense transparencies orgininating from Earth Resources Technology Satellite imagery. Transparencies are evaluated on a standard light source to determine exposure and processing information needed for making negatives. A “System ASA Rating” was developed by... 11. Determining the Measurement Quality of a Montessori High School Teacher Evaluation Survey Science.gov (United States) Setari, Anthony Philip; Bradley, Kelly D. 2017-01-01 The purpose of this study was to conduct a psychometric validation of a course evaluation instrument, known as a student evaluation of teaching (SET), implemented in a Montessori high school. The authors demonstrate to the Montessori community how to rigorously examine the measurement and assessment quality of instruments used within Montessori… 12. High quality ZnO layers with adjustable refractive indices for integrated optics applications NARCIS (Netherlands) Heideman, Rene; Lambeck, Paul; Gardeniers, Johannes G.E. 1995-01-01 Thin (approx. 1 μm) crystalline ZnO films with a good optical quality and a good (0002) texture are grown under two considerably different process parameter sets using a r.f. planar magnetron sputtering unit. The optical parameters of the two corresponding ZnO layers are distinctly different: high 13. Study on the Introduction of High-Quality Educational Resources for Sino-Foreign Cooperative Education Science.gov (United States) Jinhui, Lin 2016-01-01 In Sino-foreign cooperative education, high-quality introduced educational resources must benefit the growth and development of students, facilitate the school's capacity building and the improvement of overall educational standards, and promote national socioeconomic development. It is necessary to establish and perfect the various working… 14. High quality ion channels recordings on an injection molded polymer chip DEFF Research Database (Denmark) Tanzi, Simone In this thesis we demonstrate high quality recordings of the ion channel activity across the cell membrane in a biological cell by employing the so called patch clamping technique on an injection molded polymer microfluidic device. Such recordings are traditionally made using glass micropipettes,... 15. Rapid chilling cannot prevent inferior pork quality caused by high preslaughter stress NARCIS (Netherlands) Hambrecht, E.; Eissen, J.J.; Klein, W.J.H.; Ducro, B.J.; Smits, C.H.M.; Verstegen, M.W.A.; Hartog, den L.A. 2004-01-01 The present experiment investigated whether increasing chilling rate could improve meat quality in pigs exposed to either minimal or high stress immediately preslaughter. Pigs (n = 192) were offspring of halothane-free lines. On various days, four groups of 48 pigs were processed at a commercial 16. High-quality bulk hybrid perovskite single crystals within minutes by inverse temperature crystallization KAUST Repository Saidaminov, Makhsud I.; Abdelhady, Ahmed L.; Banavoth, Murali; Alarousu, Erkki; Burlakov, Victor M.; Peng, Wei; Dursun, Ibrahim; Wang, Lingfei; He, Yao; Maculan, Giacomo; Goriely, Alain; Wu, Tao; Mohammed, Omar F.; Bakr, Osman 2015-01-01 Single crystals of methylammonium lead trihalide perovskites (MAPbX3; MA=CH3NH3+, X=Br− or I−) have shown remarkably low trap density and charge transport properties; however, growth of such high-quality semiconductors is a time-consuming process 17. Animated Cell Biology: A Quick and Easy Method for Making Effective, High-Quality Teaching Animations Science.gov (United States) O'Day, Danton H. 2006-01-01 There is accumulating evidence that animations aid learning of dynamic concepts in cell biology. However, existing animation packages are expensive and difficult to learn, and the subsequent production of even short animations can take weeks to months. Here I outline the principles and sequence of steps for producing high-quality PowerPoint… 18. Emotional Experience, Expression, and Regulation of High-Quality Japanese Elementary School Teachers Science.gov (United States) Hosotani, Rika; Imai-Matsumura, Kyoko 2011-01-01 The present study investigates the emotional experience, expression, and regulation processes of high-quality Japanese elementary school teachers while they interact with children, in terms of teachers' emotional competence. Qualitative analysis of interview data demonstrated that teachers had various emotional experiences including self-elicited… 19. Year impact of highly active antiretroviral therapy on quality of life of ... African Journals Online (AJOL) Objective: The availability of highly active antiretroviral therapy (HAART) has resulted in a number of achievements as well as challenges. The aim of this study was to assess the influence of 48 weeks HAART of stavudine, lamivudine and nevirapine on the quality of life of HIVinfected Nigerians. Materials and Method: ... 20. School District Wellness Policy Quality and Weight-Related Outcomes among High School Students in Minnesota Science.gov (United States) Hoffman, Pamela K.; Davey, Cynthia S.; Larson, Nicole; Grannon, Katherine Y.; Hanson, Carlie; Nanney, Marilyn S. 2016-01-01 Weight-related outcomes were examined among high school students in Minnesota public school districts according to the quality of district wellness policies. Wellness policy strength and comprehensiveness were scored using the Wellness School Assessment Tool (WellSAT) for 325 Minnesota public school districts in 2013. The associations between… 1. 75 FR 41693 - Export Inspection and Weighing Waiver for High Quality Specialty Grains Transported in Containers Science.gov (United States) 2010-07-19 ... Specialty Grains Transported in Containers AGENCY: Grain Inspection, Packers and Stockyards Administration... rule to potentially make permanent the current waiver for high quality grain exported in containers... exported in containers that was established by a final rule on December 13, 2005 (70 FR 73556). This... 2. Growth and characterization of high quality ZnS thin films by RF sputtering Science.gov (United States) Mukherjee, C.; Rajiv, K.; Gupta, P.; Sinha, A. K.; Abhinandan, L. 2012-06-01 High optical quality ZnS films are deposited on glass and Si wafer by RF sputtering from pure ZnS target. Optical transmittance, reflectance, ellipsometry, FTIR and AFM measurements are carried out. Effect of substrate temperature and chamber baking for long duration on film properties have been studied. Roughness of the films as measured by AFM are low (1-2Å). 3. Student Writing Accepted as High-Quality Responses to Analytic Text-Based Writing Tasks Science.gov (United States) Wang, Elaine; Matsumura, Lindsay Clare; Correnti, Richard 2018-01-01 Literacy standards increasingly emphasize the importance of analytic text-based writing. Little consensus exists, however, around what high-quality student responses should look like in this genre. In this study, we investigated fifth-grade students' writing in response to analytic text-based writing tasks (15 teachers, 44 writing tasks, 88 pieces… 4. Toward High Quality Family Day Care for Infants and Toddlers. Final Report. Science.gov (United States) Rauch, Marian D.; Crowell, Doris C. Reported were the results of a project which established a cluster of family day care homes in Hawaii in which caregivers were selected, trained, and provided with supportive services and salaries. The primary objective of the program was to provide a replicable, high quality program for preschool children that would maximize social, emotional,… 5. 10 Principles for Building a High-Quality System of Assessments Science.gov (United States) Jobs for the Future, 2018 2018-01-01 Many states and districts are working toward developing and implementing high-quality systems that align assessments with each other, and to college and career readiness, and a comprehensive set of higher-order thinking skills. In order to support states, districts, and communities in this, the following 10 principles as guidance and common… 6. Access, Participation, and Supports: The Defining Features of High-Quality Inclusion Science.gov (United States) 2011-01-01 This article describes current knowledge about early childhood inclusion, summarizing research and the DEC/NAEYC joint position statement on inclusion. The article also describes effective or promising educational practices that promote access, participation, and supports--the defining features of high-quality inclusion. Future efforts to improve… 7. Long pulse KrCl laser with a high discharge quality NARCIS (Netherlands) Casper, L.C.; Bastiaens, Hubertus M.J.; Peters, P.J.M.; Boller, Klaus J.; Hofstra, R.M. 2007-01-01 The discharge quality and optimum pump parameters of a long-pulse high-pressure gas discharge excited KrCl laser are investigated. A three-electrode prepulse–mainpulse excitation circuit is employed as pump source. The discharge volume contains a gas mixture of HCl/Kr/Ne operated at a total pressure 8. A Novel Approach to High-Quality Postmortem Tissue Procurement: The GTEx Project NARCIS (Netherlands) Carithers, Latarsha J.; Ardlie, Kristin; Barcus, Mary; Branton, Philip A.; Britton, Angela; Buia, Stephen A.; Compton, Carolyn C.; Deluca, David S.; Peter-Demchok, Joanne; Gelfand, Ellen T.; Guan, Ping; Korzeniewski, Greg E.; Lockhart, Nicole C.; Rabiner, Chana A.; Rao, Abhi K.; Robinson, Karna L.; Roche, Nancy V.; Sawyer, Sherilyn J.; Segrè, Ayellet V.; Shive, Charles E.; Smith, Anna M.; Sobin, Leslie H.; Undale, Anita H.; Valentino, Kimberly M.; Vaught, Jim; Young, Taylor R.; Moore, Helen M.; Barker, Laura; Basile, Margaret; Battle, Alexis; Boyer, Joy; Bradbury, Debra; Bridge, Jason P.; Brown, Amanda; Burges, Robin; Choi, Christopher; Colantuoni, Deborah; Cox, Nancy; Dermitzakis, Emmanouil T.; Derr, Leslie K.; Dinsmore, Michael J.; Erickson, Kenyon; Fleming, Johnelle; Flutre, Timothée; Foster, Barbara A.; Gamazon, Eric R.; Getz, Gad; Gillard, Bryan M.; Guigo, Roderic; Hambright, Kenneth W.; Hariharan, Pushpa; Hasz, Rick; Im, Hae K.; Jewell, Scott; Karasik, Ellen; Kellis, Manolis; Kheradpour, Pouya; Koester, Susan; Koller, Daphne; Konkashbaev, Anuar; Lappalainen, Tuuli; Little, Roger; Liu, Jun; Lo, Edmund; Lonsdale, John T.; Lu, Chunrong; MacArthur, Daniel G.; Magazine, Harold; Maller, Julian B.; Marcus, Yvonne; Mash, Deborah C.; McCarthy, Mark I.; McLean, Jeffrey; Mestichelli, Bernadette; Miklos, Mark; Monlong, Jean; Mosavel, Magboeba; Moser, Michael T.; Mostafavi, Sara; Nicolae, Dan L.; Pritchard, Jonathan; Qi, Liqun; Ramsey, Kimberly; Rivas, Manuel A.; Robles, Barnaby E.; Rohrer, Daniel C.; Salvatore, Mike; Sammeth, Michael; Seleski, John; Shad, Saboor; Siminoff, Laura A.; Stephens, Matthew; Struewing, Jeff; Sullivan, Timothy; Sullivan, Susan; Syron, John; Tabor, David; Taherian, Mehran; Tejada, Jorge; Temple, Gary F.; Thomas, Jeffrey A.; Thomson, Alexander W.; Tidwell, Denee; Traino, Heather M.; Tu, Zhidong; Valley, Dana R.; Volpi, Simona; Walters, Gary D.; Ward, Lucas D.; Wen, Xiaoquan; Winckler, Wendy; Wu, Shenpei; Zhu, Jun 2015-01-01 The Genotype-Tissue Expression (GTEx) project, sponsored by the NIH Common Fund, was established to study the correlation between human genetic variation and tissue-specific gene expression in non-diseased individuals. A significant challenge was the collection of high-quality biospecimens for 9. Effect of high hydrostatic pressure processing on the background microbial loads and quality of cantaloupe puree Science.gov (United States) The objective of this study was to investigate and evaluate the effects of high hydrostatic pressure (HHP) applied to cantaloupe puree (CP) on microbial loads and product quality during storage for 10 days at 4 degrees C. Freshly prepared, double sealed and double bagged CP (ca. 5 g) was pressure tr... 10. From journal to headline: the accuracy of climate science news in Danish high quality newspapers DEFF Research Database (Denmark) Vestergård, Gunver Lystbæk 2011-01-01 analysis to examine the accuracy of Danish high quality newspapers in quoting scientific publications from 1997 to 2009. Out of 88 articles, 46 contained inaccuracies though the majority was found to be insignificant and random. The study concludes that Danish broadsheet newspapers are ‘moderately... 11. Dare to share? How people share high-quality knowledge in online communities NARCIS (Netherlands) Rotundo, Enrico; Blank, M.; Moser, C.; Leopold, H. The Internet has fundamentally changed how work gets done in the 21st century. For example, people increasingly spend time on the Internet where they share and develop knowledge in online communities. Yet, little is known about how high-quality knowledge comes about in these communities. This is 12. Quality control of Chinese herbal tonic wine by high performance liquid chromatography fingerprint NARCIS (Netherlands) Wei, X.J.; Zhang, H.; Wang, W.F.; Li, B.; Yang Zhu, Yang 2007-01-01 Herbal tonic wines are alcoholic drinks in which medicinal herbs are soaked and extracted. These drinks are considered having various health functions. However, the quality of herbal products is largely influenced by the origin and harvest season of the herbs. Due to its high commercial value, 13. Project outline of high quality electron beam generation at Waseda University Energy Technology Data Exchange (ETDEWEB) Washio, M.; Hama, Y.; Kashiwagi, S.; Kuroda, R.; Kobuki, T. [Waseda Univ., Advanced Research Institute for Science and Engineering, Shinjuku, Tokyo (Japan); Hirose, T. [Tokyo Metropolitan Univ. (Japan). Dept. of Physics 2000-03-01 High quality electron beam generation project has been started at Waseda University under the grant of Ministry of Education, named High-Tech Research Center Project. In the project, we will install a laser photo-cathode RF Gun system with 1.6 accelerating structure cells of s-band and a stabilized RF power source. This RF Gun is expected to produce single electron bunch up to 1 or 2nC with around 10ps pulse duration. (author) 14. Project outline of high quality electron beam generation at Waseda University International Nuclear Information System (INIS) Washio, M.; Hama, Y.; Kashiwagi, S.; Kuroda, R.; Kobuki, T.; Hirose, T. 2000-01-01 High quality electron beam generation project has been started at Waseda University under the grant of Ministry of Education, named High-Tech Research Center Project. In the project, we will install a laser photo-cathode RF Gun system with 1.6 accelerating structure cells of s-band and a stabilized RF power source. This RF Gun is expected to produce single electron bunch up to 1 or 2nC with around 10ps pulse duration. (author) 15. Indoor Air Quality in 24 California Residences Designed as High-Performance Homes Energy Technology Data Exchange (ETDEWEB) Less, Brennan [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Mullen, Nasim [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Singer, Brett [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Walker, Iain [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States) 2015-07-01 Today’s high performance green homes are reaching previously unheard of levels of airtightness and are using new materials, technologies and strategies, whose impacts on Indoor Air Quality (IAQ) cannot be fully anticipated from prior studies. This research study used pollutant measurements, home inspections, diagnostic testing and occupant surveys to assess IAQ in 24 new or deeply retrofitted homes designed to be high performance green buildings in California. 16. High-quality AlN films grown on chemical vapor-deposited graphene films Directory of Open Access Journals (Sweden) Chen Bin-Hao 2016-01-01 Full Text Available We report the growth of high-quality AlN films on graphene. The graphene films were synthesized by CVD and then transferred onto silicon substrates. Epitaxial aluminum nitride films were deposited by DC magnetron sputtering on both graphene as an intermediate layer and silicon as a substrate. The structural characteristics of the AlN films and graphene were investigated. Highly c-axis-oriented AlN crystal structures are investigated based on the XRDpatterns observations. 17. Generating high-quality single droplets for optical particle characterization with an easy setup Science.gov (United States) Xu, Jie; Ge, Baozhen; Meng, Rui 2018-06-01 The high-performance and micro-sized single droplet is significant for optical particle characterization. We develop a single-droplet generator (SDG) based on a piezoelectric inkjet technique with advantages of low cost and easy setup. By optimizing the pulse parameters, we achieve various size single droplets. Further investigations reveal that SDG generates single droplets of high quality, demonstrating good sphericity, monodispersity and a stable length of several millimeters. 18. Quantitative phosphoproteomic analysis of caprine muscle with high and low meat quality. Science.gov (United States) Liu, Manshun; Wei, Yanchao; Li, Xin; Quek, Siew Young; Zhao, Jing; Zhong, Huazhen; Zhang, Dequan; Liu, Yongfeng 2018-07-01 During the conversion of muscle to meat, protein phosphorylation can regulate various biological processes that have important effects on meat quality. To investigate the phosphorylation pattern of protein on rigor mortis, goat longissimus thoracis and external intercostals were classified into two groups (high quality and low quality), and meat quality was evaluated according to meat quality attributes (Warner-Bratzler shear force, Color, pH and drip loss). A quantitative mass spectrometry-based phosphoproteomic study was conducted to analyze the caprine muscle at 12h postmortem applying the TiO 2 -SIMAC-HILIC (TiSH) phosphopeptide enrichment strategy. A total of 2125 phosphopeptides were identified from 750 phosphoproteins. Among them, 96 proteins had differed in phosphorylation levels. The majority of these proteins are involved in glucose metabolism and muscle contraction. The differential phosphorylation level of proteins (PFK, MYL2 and HSP27) in two groups may be the crucial factors of regulating muscle rigor mortis. This study provides a comprehensive view for the phosphorylation status of caprine muscle at rigor mortis, it also gives a better understanding of the regulation of protein phosphorylation on various biological processes that affect the final meat quality attributes. Copyright © 2018. Published by Elsevier Ltd. 19. Quality control of coated fuel particles for high temperature gas-cooled reactor International Nuclear Information System (INIS) Kaneko, Mitsunobu 1987-01-01 The quality control of the coated fuel particles for high temperature gas-cooled reactors is characterized by the fact that the size of the target product to be controlled is very small, and the quantity is very large. Accordingly, the sampling plan and the method of evaluating the population through satisfically treating the measured data of the samples are the important subjects to see and evaluate the quality of a batch or a lot. This paper shows the fabrication process and the quality control procedure for the coated fuel particles. The development work of a HTGR was started by Japan Atomic Energy Research Institute in 1969, and as for the production technology for coated fuel particles, Nuclear Fuel Industries, Ltd. has continued the development work. The pilot plan with the capacity of about 40 kg/year was established in 1972. The fuel product fabricated in this plant was put to the irradiation experiment and out-of-pile evaluation test. In 1983, the production capacity was expanded to 200 kg/year, and the fuel compacts for the VHTRC in JAERI were produced for two years. The basic fuel design, the fabrication process, the quality control, the process control and the quality assurance are reported. For the commercial product, the studies from the viewpoint of production and quality control costs are required. (Kako, I.) 20. Molecular Beam Epitaxy Growth of High Crystalline Quality LiNbO3 Science.gov (United States) Tellekamp, M. Brooks; Shank, Joshua C.; Goorsky, Mark S.; Doolittle, W. Alan 2016-12-01 Lithium niobate is a multi-functional material with wide reaching applications in acoustics, optics, and electronics. Commercial applications for lithium niobate require high crystalline quality currently limited to bulk and ion sliced material. Thin film lithium niobate is an attractive option for a variety of integrated devices, but the research effort has been stagnant due to poor material quality. Both lattice matched and mismatched lithium niobate are grown by molecular beam epitaxy and studied to understand the role of substrate and temperature on nucleation conditions and material quality. Growth on sapphire produces partially coalesced columnar grains with atomically flat plateaus and no twin planes. A symmetric rocking curve shows a narrow linewidth with a full width at half-maximum (FWHM) of 8.6 arcsec (0.0024°), which is comparable to the 5.8 arcsec rocking curve FWHM of the substrate, while the film asymmetric rocking curve is 510 arcsec FWHM. These values indicate that the individual grains are relatively free of long-range disorder detectable by x-ray diffraction with minimal measurable tilt and twist and represents the highest structural quality epitaxial material grown on lattice mismatched sapphire without twin planes. Lithium niobate is also grown on lithium tantalate producing high quality coalesced material without twin planes and with a symmetric rocking curve of 193 arcsec, which is nearly equal to the substrate rocking curve of 194 arcsec. The surface morphology of lithium niobate on lithium tantalate is shown to be atomically flat by atomic force microscopy. 1. Study on quality assurance for high-level radioactive waste disposal project International Nuclear Information System (INIS) 2005-01-01 The U.S. Department of Energy (DOE) has developed comparatively detailed quality assurance requirements for the high-level radioactive waste disposal systems. Quality assurance is recognized as a key issue for confidence building and smooth implementation of the HLW program in Japan, and Japan is at an initial phase of repository development. Then the quality assurance requirements at site research and site selection, site characterization, and site suitability analysis used in the Yucca Mountain project were examined in detail and comprehensive descriptions were developed using flow charts. Additionally, the applicability to the Japan high-level radioactive waste disposal project was studied. The examination and study were performed for the following QA requirements: The requirements that have the relative importance at site research and site selection, site characterization, and site suitability analysis (such as planning and performing scientific investigations, sample control, data control, model development and use, technical report review, software control, and control of the electric management of data). The requirements that have the relative importance at the whole repository phases (such as quality assurance program, document control, and control of quality assurance records). (author) 2. High cardiac vagal control is related to better subjective and objective sleep quality. Science.gov (United States) Werner, Gabriela G; Ford, Brett Q; Mauss, Iris B; Schabus, Manuel; Blechert, Jens; Wilhelm, Frank H 2015-03-01 Cardiac vagal control (CVC) has been linked to both physical and mental health. One critical aspect of health, that has not received much attention, is sleep. We hypothesized that adults with higher CVC--operationalized by high-frequency heart rate variability (HF-HRV)--will exhibit better sleep quality assessed both subjectively (i.e., with Pittsburgh Sleep Quality Index) and objectively (i.e., with polysomnography). HF-HRV was measured in 29 healthy young women during an extended neutral film clip. Participants then underwent full polysomnography to obtain objective measures of sleep quality and HF-HRV during a night of sleep. As expected, higher resting HF-HRV was associated with higher subjective and objective sleep quality (i.e., shorter sleep latency and fewer arousals). HF-HRV during sleep (overall or separated by sleep phases) showed less consistent relationships with sleep quality. These findings indicate that high waking CVC may be a key predictor of healthy sleep. Copyright © 2015 The Authors. Published by Elsevier B.V. All rights reserved. 3. Short time synthesis of high quality carbon nanotubes with high rates by CVD of methane on continuously emerged iron nanoparticles International Nuclear Information System (INIS) 2011-01-01 We report the variation of yield and quality of carbon nanotubes (CNTs) grown by chemical vapor deposition (CVD) of methane on iron oxide-MgO at 900-1000 deg. C for 1-60 min. The catalyst was prepared by impregnation of MgO powder with iron nitrate, dried, and calcined at 300 deg. C. As calcined and unreduced catalyst in quartz reactor was brought to the synthesis temperature in helium flow in a few minutes, and then the flow was switched to methane. The iron oxide was reduced to iron nanoparticles in methane, while the CNTs were growing. TEM micrographs, in accordance with Raman RBM peaks, indicate the formation of mostly single wall carbon nanotubes of about 1.0 nm size. High quality CNTs with I G /I D Raman peak ratio of 14.5 are formed in the first minute of CNTs synthesis with the highest rate. Both the rate and quality of CNTs degrades with increasing CNTs synthesis time. Also CNTs quality sharply declines with temperature in the range of 900-1000 deg. C, while the CNTs yield passes through a maximum at 950 deg. C. About the same CNTs lengths are formed for the whole range of the synthesis times. A model of continuous emergence of iron nanoparticle seeds for CNTs synthesis may explain the data. The data can also provide information for continuous production of CNTs in a fluidized bed reactor. 4. Short time synthesis of high quality carbon nanotubes with high rates by CVD of methane on continuously emerged iron nanoparticles Energy Technology Data Exchange (ETDEWEB) Bahrami, Behnam, E-mail: [email protected] [Catalysis and Nanostructured Materials Research Laboratory, School of Chemical Engineering, University of Tehran, Tehran (Iran, Islamic Republic of); Nanoelectronics Centre of Excellence, University of Tehran, Tehran (Iran, Islamic Republic of); Khodadadi, Abasali [Catalysis and Nanostructured Materials Research Laboratory, School of Chemical Engineering, University of Tehran, Tehran (Iran, Islamic Republic of); Mortazavi, Yadollah, E-mail: [email protected] [Nanoelectronics Centre of Excellence, University of Tehran, Tehran (Iran, Islamic Republic of); Esmaieli, Mohamad [Nanoelectronics Centre of Excellence, University of Tehran, Tehran (Iran, Islamic Republic of) 2011-09-15 We report the variation of yield and quality of carbon nanotubes (CNTs) grown by chemical vapor deposition (CVD) of methane on iron oxide-MgO at 900-1000 deg. C for 1-60 min. The catalyst was prepared by impregnation of MgO powder with iron nitrate, dried, and calcined at 300 deg. C. As calcined and unreduced catalyst in quartz reactor was brought to the synthesis temperature in helium flow in a few minutes, and then the flow was switched to methane. The iron oxide was reduced to iron nanoparticles in methane, while the CNTs were growing. TEM micrographs, in accordance with Raman RBM peaks, indicate the formation of mostly single wall carbon nanotubes of about 1.0 nm size. High quality CNTs with I{sub G}/I{sub D} Raman peak ratio of 14.5 are formed in the first minute of CNTs synthesis with the highest rate. Both the rate and quality of CNTs degrades with increasing CNTs synthesis time. Also CNTs quality sharply declines with temperature in the range of 900-1000 deg. C, while the CNTs yield passes through a maximum at 950 deg. C. About the same CNTs lengths are formed for the whole range of the synthesis times. A model of continuous emergence of iron nanoparticle seeds for CNTs synthesis may explain the data. The data can also provide information for continuous production of CNTs in a fluidized bed reactor. 5. CD14CD16 Monocyte Subset Levels in Heart Failure Patients Directory of Open Access Journals (Sweden) Chiara Barisione 2010-01-01 Full Text Available Our aim was to define the distribution of monocyte subsets in a cohort of congestive heart failure (CHF patients, to verify whether increased severity of CHF is linked to the expansion of specific monocyte subsets, and finally to investigate the relationship between monocyte subset relative frequencies, laboratory parameters of inflammation, and monocyte ACE expression. 6. The Vast Universe of T Cell Diversity: Subsets of Memory Cells and Their Differentiation. Science.gov (United States) Jandus, Camilla; Usatorre, Amaia Martínez; Viganò, Selena; Zhang, Lianjun; Romero, Pedro 2017-01-01 The T cell receptor confers specificity for antigen recognition to T cells. By the first encounter with the cognate antigen, reactive T cells initiate a program of expansion and differentiation that will define not only the ultimate quantity of specific cells that will be generated, but more importantly their quality and functional heterogeneity. Recent achievements using mouse model infection systems have helped to shed light into the complex network of factors that dictate and sustain memory T cell differentiation, ranging from antigen load, TCR signal strength, metabolic fitness, transcriptional programs, and proliferative potential. The different models of memory T cell differentiation are discussed in this chapter, and key phenotypic and functional attributes of memory T cell subsets are presented, both for mouse and human cells. Therapeutic manipulation of memory T cell generation is expected to provide novel unique ways to optimize current immunotherapies, both in infection and cancer. 7. Preparation of high-quality iodine-125-labeled pituitary luteinizing hormone for radioimmunoassay International Nuclear Information System (INIS) Pinto, H.; Wajchenberg, B.L.; Higa, O.Z.; Toledo e Souza, I.T. de; Werner, R.S.; Pieroni, R.R. 1974-01-01 High quality pituitary luteinizing hormone labeled with 125 I was obtained after separating out the more heavily iodinated fractions, through starch gel electrophoresis, using the cathodal component (fraction 1) which was further purified on Sephadex G-100, with the obtention of an almost pure 125 I-LH preparation, presenting excellent immunoreactivity and low levels of damage on incubation in plasma. The quality control of the steps of the technique was done with plasma-coated talc (200 mg) which compared favorably, as far indicating undamaged labeled LH, with the more time-consuming chromatoelectrophoresis 8. Extracting DNA from 'jaws': High yield and quality from archived tiger shark (Galeocerdo cuvier) skeletal material DEFF Research Database (Denmark) Eg Nielsen, Einar; Morgan, J. A T; Maher, S. L. 2017-01-01 of tiger sharks (Galeocerdo cuvier). Protocols were compared for DNA yield and quality using a qPCR approach. For jaw swarf, all methods provided relatively high DNA yield and quality, while large differences in yield between protocols were observed for vertebrae. Similar results were obtained from samples...... observed, likely reflecting different preparation and storage methods for the trophies. Trial sequencing of DNA capture genomic libraries using 20 000 baits revealed that a significant proportion of captured sequences were derived from tiger sharks. This study demonstrates that archived shark jaws... 9. High-quality fiber fabrication in buffered hydrofluoric acid solution with ultrasonic agitation. Science.gov (United States) Zhong, Nianbing; Liao, Qiang; Zhu, Xun; Wang, Yongzhong; Chen, Rong 2013-03-01 An etching method for preparing high-quality fiber-optic sensors using a buffered etchant with ultrasonic agitation is proposed. The effects of etching conditions on the etch rate and surface morphology of the etched fibers are investigated. The effect of surface roughness is discussed on the fibers' optical properties. Linear etching behavior and a smooth fiber surface can be repeatedly obtained by adjusting the ultrasonic power and etchant pH. The fibers' spectral quality is improved as the ratio of the pit depth to size decreases, and the fibers with smooth surfaces are more sensitive to a bacterial suspension than those with rough surfaces. 10. THE ACADEMIC PERSONNEL MOTIVATION - A FACTOR FOR HIGH QUALITY EDUCATIONAL PRODUCT Directory of Open Access Journals (Sweden) Viara Slavianska 2014-06-01 Full Text Available The present paper consecutively puts an accent on 1 the quality of higher education as a national priority, 2 the qualification and motivation of the academic staff as factors for offering an educational product of high quality, 3 the strategies, policies and practices for motivating the academic personnel. The necessity of education improvement is adduced, the strategies and politics in the field of academic personnel training are presented, and the possible effects from a wrong approach to employees’ motivation in academic environment are commented. 11. High frequency of sub-optimal semen quality in an unselected population of young men DEFF Research Database (Denmark) Andersen, A G; Jensen, T K; Carlsen, E 2000-01-01 for military service, this provided a unique opportunity to study the reproductive function in an unbiased population. Altogether 891 young men delivered a blood sample in which reproductive hormones were measured. From 708 of these men data were also obtained on semen quality and testis size. The median sperm...... immotile spermatozoa and follicle stimulating hormone. Possible causes for this high frequency of young men with suboptimal semen quality are obscure and need to be explored. Whether these findings apply for young male populations of comparable countries remains to be seen.... 12. An improved multi-exposure approach for high quality holographic femtosecond laser patterning International Nuclear Information System (INIS) Zhang, Chenchu; Hu, Yanlei; Li, Jiawen; Lao, Zhaoxin; Ni, Jincheng; Chu, Jiaru; Huang, Wenhao; Wu, Dong 2014-01-01 High efficiency two photon polymerization through single exposure via spatial light modulator (SLM) has been used to decrease the fabrication time and rapidly realize various micro/nanostructures, but the surface quality remains a big problem due to the speckle noise of optical intensity distribution at the defocused plane. Here, a multi-exposure approach which used tens of computer generate holograms successively loaded on SLM is presented to significantly improve the optical uniformity without losing efficiency. By applying multi-exposure, we found that the uniformity at the defocused plane was increased from ∼0.02 to ∼0.6 according to our simulation. The fabricated two series of letters “HELLO” and “USTC” under single-and multi-exposure in our experiment also verified that the surface quality was greatly improved. Moreover, by this method, several kinds of beam splitters with high quality, e.g., 2 × 2, 5 × 5 Daman, and complex nonseperate 5 × 5, gratings were fabricated with both of high quality and short time (<1 min, 95% time-saving). This multi-exposure SLM-two-photon polymerization method showed the promising prospect in rapidly fabricating and integrating various binary optical devices and their systems 13. Teolenn: an efficient and customizable workflow to design high-quality probes for microarray experiments Science.gov (United States) Jourdren, Laurent; Duclos, Aurélie; Brion, Christian; Portnoy, Thomas; Mathis, Hugues; Margeot, Antoine; Le Crom, Stéphane 2010-01-01 Despite the development of new high-throughput sequencing techniques, microarrays are still attractive tools to study small genome organisms, thanks to sample multiplexing and high-feature densities. However, the oligonucleotide design remains a delicate step for most users. A vast array of software is available to deal with this problem, but each program is developed with its own strategy, which makes the choice of the best solution difficult. Here we describe Teolenn, a universal probe design workflow developed with a flexible and customizable module organization allowing fixed or variable length oligonucleotide generation. In addition, our software is able to supply quality scores for each of the designed probes. In order to assess the relevance of these scores, we performed a real hybridization using a tiling array designed against the Trichoderma reesei fungus genome. We show that our scoring pipeline correlates with signal quality for 97.2% of all the designed probes, allowing for a posteriori comparisons between quality scores and signal intensities. This result is useful in discarding any bad scoring probes during the design step in order to get high-quality microarrays. Teolenn is available at http://transcriptome.ens.fr/teolenn/. PMID:20176570 14. Minimizing quality changes of cloudy apple juice: The use of kiwifruit puree and high pressure homogenization. Science.gov (United States) Yi, Junjie; Kebede, Biniam; Kristiani, Kristiani; Grauwet, Tara; Van Loey, Ann; Hendrickx, Marc 2018-05-30 Cloud loss, enzymatic browning, and flavor changes are important quality defects of cloudy fruit juices determining consumer acceptability. The development of clean label options to overcome such quality problems is currently of high interest. Therefore, this study investigated the effect of kiwifruit puree (clean label ingredient) and high pressure homogenization on quality changes of cloudy apple juice using a multivariate approach. The use of kiwifruit puree addition and high pressure homogenization resulted in a juice with improved uniformity and cloud stability by reducing particle size and increasing viscosity and yield stress (p < 0.01). Furthermore, kiwifruit puree addition reduced enzymatic browning (ΔE ∗  < 3), due to the increased ascorbic acid and contributed to a more saturated and bright yellow color, a better taste balance, and a more fruity aroma of juice. This work demonstrates that clean label options to control quality degradation of cloudy fruit juice might offer new opportunities. Copyright © 2018 Elsevier Ltd. All rights reserved. 15. An improved multi-exposure approach for high quality holographic femtosecond laser patterning Energy Technology Data Exchange (ETDEWEB) Zhang, Chenchu; Hu, Yanlei, E-mail: [email protected], E-mail: [email protected]; Li, Jiawen, E-mail: [email protected], E-mail: [email protected]; Lao, Zhaoxin; Ni, Jincheng; Chu, Jiaru; Huang, Wenhao; Wu, Dong [Department of Precision Machinery and Precision Instrumentation, University of Science and Technology of China, Hefei 230026 (China) 2014-12-01 High efficiency two photon polymerization through single exposure via spatial light modulator (SLM) has been used to decrease the fabrication time and rapidly realize various micro/nanostructures, but the surface quality remains a big problem due to the speckle noise of optical intensity distribution at the defocused plane. Here, a multi-exposure approach which used tens of computer generate holograms successively loaded on SLM is presented to significantly improve the optical uniformity without losing efficiency. By applying multi-exposure, we found that the uniformity at the defocused plane was increased from ∼0.02 to ∼0.6 according to our simulation. The fabricated two series of letters “HELLO” and “USTC” under single-and multi-exposure in our experiment also verified that the surface quality was greatly improved. Moreover, by this method, several kinds of beam splitters with high quality, e.g., 2 × 2, 5 × 5 Daman, and complex nonseperate 5 × 5, gratings were fabricated with both of high quality and short time (<1 min, 95% time-saving). This multi-exposure SLM-two-photon polymerization method showed the promising prospect in rapidly fabricating and integrating various binary optical devices and their systems. 16. A quality assurance initiative for commercial-scale production in high-throughput cryopreservation of blue catfish sperm. Science.gov (United States) Hu, E; Liao, T W; Tiersch, T R 2013-10-01 Cryopreservation of fish sperm has been studied for decades at a laboratory (research) scale. However, high-throughput cryopreservation of fish sperm has recently been developed to enable industrial-scale production. This study treated blue catfish (Ictalurus furcatus) sperm high-throughput cryopreservation as a manufacturing production line and initiated quality assurance plan development. The main objectives were to identify: (1) the main production quality characteristics; (2) the process features for quality assurance; (3) the internal quality characteristics and their specification designs; (4) the quality control and process capability evaluation methods, and (5) the directions for further improvements and applications. The essential product quality characteristics were identified as fertility-related characteristics. Specification design which established the tolerance levels according to demand and process constraints was performed based on these quality characteristics. Meanwhile, to ensure integrity throughout the process, internal quality characteristics (characteristics at each quality control point within process) that could affect fertility-related quality characteristics were defined with specifications. Due to the process feature of 100% inspection (quality inspection of every fish), a specific calculation method, use of cumulative sum (CUSUM) control charts, was applied to monitor each quality characteristic. An index of overall process evaluation, process capacity, was analyzed based on in-control process and the designed specifications, which further integrates the quality assurance plan. With the established quality assurance plan, the process could operate stably and quality of products would be reliable. Copyright © 2013 Elsevier Inc. All rights reserved. 17. (-201) β-Gallium oxide substrate for high quality GaN materials KAUST Repository Roqan, Iman S. 2015-03-13 (-201) oriented β-Ga2O3 has the potential to be used as a transparent and conductive substrate for GaN-growth. The key advantages of Ga2O3 are its small lattice mismatches (4.7%), appropriate structural, thermal and electrical properties and a competitive price compared to other substrates. Optical characterization show that GaN layers grown on (-201) oriented β-Ga2O3 are dominated by intense bandedge emission with a high luminescence efficiency. Atomic force microscopy studies show a modest threading dislocation density of ~108 cm-2, while complementary Raman spectroscopy indicates that the GaN epilayer is of high quality with slight compressive strain. Room temperature time-findings suggest that the limitation of the photoluminescence lifetime (~500 ps) is due to nonradiative recombination arising from threading dislocation. Therefore, by optimizing the growth conditions, high quality material with significant optical efficiency can be obtained. 18. Synthesis and Characterization of Mass Produced High Quality Few Layered Graphene Sheets via a Chemical Method KAUST Repository Khenfouch, Mohammed 2014-04-01 Graphene is a two-dimensional crystal of carbon atoms arranged in a honeycomb lattice. It is a zero band gap semimetal with very unique physical and chemical properties which make it useful for many applications such as ultra-high-speed field-effect transistors, p-n junction diodes, terahertz oscillators, and low-noise electronic, NEMS and sensors. When the high quality mass production of this nanomaterial is still a big challenge, we developed a process which will be an important step to achieve this goal. Atomic Force Microscopy, Scanning Electron Microscopy, Scanning tunneling microscopy, High Resolution Transmission Electron Microscopy, X-Ray Diffraction, Raman spectroscopy, Energy Dispersive X-ray system were investigated to characterize and examine the quality of this product. 19. A design technique of low cost but high quality peak stretcher Energy Technology Data Exchange (ETDEWEB) Lo, H Y; Su, C S; Hsu, J Y; Wang, L 1981-03-01 This paper presents the design of low cost but high quality pulse peak stretcher incorporated with a LSI of 12 bit ADC and SKD-85 microcomputer. The conflict between the capacitor high charging speed and longer holding time for realizing a high quality stretcher is discussed and solved. For a lager number of channels available in a dual ramp Wilkinson-type ADC, two-stage stretchers connected in series are designed. The first stage is a fast-discharge to keep the output stretched pulse follow-up the input quickly and the second-stage (main stretcher) is a slow-discharge to keep the transient of the circuit minimum. Both of these two peak stretchers are described and the experiment results are photographically recorded. 20. High quality factor GaAs microcavity with buried bullseye defects Science.gov (United States) Winkler, K.; Gregersen, N.; Häyrynen, T.; Bradel, B.; Schade, A.; Emmerling, M.; Kamp, M.; Höfling, S.; Schneider, C. 2018-05-01 The development of high quality factor solid-state microcavities with low mode volumes has paved the way towards on-chip cavity quantum electrodynamics experiments and the development of high-performance nanophotonic devices. Here, we report on the implementation of a new kind of solid-state vertical microcavity, which allows for confinement of the electromagnetic field in the lateral direction without deep etching. The confinement originates from a local elongation of the cavity layer imprinted in a shallow etch and epitaxial overgrowth technique. We show that it is possible to improve the quality factor of such microcavities by a specific in-plane bullseye geometry consisting of a set of concentric rings with subwavelength dimensions. This design results in a smooth effective lateral photonic potential and therefore in a reduction of lateral scattering losses, which makes it highly appealing for experiments in the framework of exciton-polariton physics demanding tight spatial confinement. 1. High quality atomically thin PtSe2 films grown by molecular beam epitaxy Science.gov (United States) Yan, Mingzhe; Wang, Eryin; Zhou, Xue; Zhang, Guangqi; Zhang, Hongyun; Zhang, Kenan; Yao, Wei; Lu, Nianpeng; Yang, Shuzhen; Wu, Shilong; Yoshikawa, Tomoki; Miyamoto, Koji; Okuda, Taichi; Wu, Yang; Yu, Pu; Duan, Wenhui; Zhou, Shuyun 2017-12-01 Atomically thin PtSe2 films have attracted extensive research interests for potential applications in high-speed electronics, spintronics and photodetectors. Obtaining high quality thin films with large size and controlled thickness is critical. Here we report the first successful epitaxial growth of high quality PtSe2 films by molecular beam epitaxy. Atomically thin films from 1 ML to 22 ML have been grown and characterized by low-energy electron diffraction, Raman spectroscopy and x-ray photoemission spectroscopy. Moreover, a systematic thickness dependent study of the electronic structure is revealed by angle-resolved photoemission spectroscopy (ARPES), and helical spin texture is revealed by spin-ARPES. Our work provides new opportunities for growing large size single crystalline films to investigate the physical properties and potential applications of PtSe2. 2. Fabrication of high-quality single-crystal Cu thin films using radio-frequency sputtering. Science.gov (United States) Lee, Seunghun; Kim, Ji Young; Lee, Tae-Woo; Kim, Won-Kyung; Kim, Bum-Su; Park, Ji Hun; Bae, Jong-Seong; Cho, Yong Chan; Kim, Jungdae; Oh, Min-Wook; Hwang, Cheol Seong; Jeong, Se-Young 2014-08-29 Copper (Cu) thin films have been widely used as electrodes and interconnection wires in integrated electronic circuits, and more recently as substrates for the synthesis of graphene. However, the ultra-high vacuum processes required for high-quality Cu film fabrication, such as molecular beam epitaxy (MBE), restricts mass production with low cost. In this work, we demonstrated high-quality Cu thin films using a single-crystal Cu target and radio-frequency (RF) sputtering technique; the resulting film quality was comparable to that produced using MBE, even under unfavorable conditions for pure Cu film growth. The Cu thin film was epitaxially grown on an Al2O3 (sapphire) (0001) substrate, and had high crystalline orientation along the (111) direction. Despite the 10(-3) Pa vacuum conditions, the resulting thin film was oxygen free due to the high chemical stability of the sputtered specimen from a single-crystal target; moreover, the deposited film had >5× higher adhesion force than that produced using a polycrystalline target. This fabrication method enabled Cu films to be obtained using a simple, manufacturing-friendly process on a large-area substrate, making our findings relevant for industrial applications. 3. Fundamental properties of high-quality carbon nanofoam: from low to high density Directory of Open Access Journals (Sweden) Natalie Frese 2016-12-01 Full Text Available Highly uniform samples of carbon nanofoam from hydrothermal sucrose carbonization were studied by helium ion microscopy (HIM, X-ray photoelectron spectroscopy (XPS, and Raman spectroscopy. Foams with different densities were produced by changing the process temperature in the autoclave reactor. This work illustrates how the geometrical structure, electron core levels, and the vibrational signatures change when the density of the foams is varied. We find that the low-density foams have very uniform structure consisting of micropearls with ≈2–3 μm average diameter. Higher density foams contain larger-sized micropearls (≈6–9 μm diameter which often coalesced to form nonspherical μm-sized units. Both, low- and high-density foams are comprised of predominantly sp2-type carbon. The higher density foams, however, show an advanced graphitization degree and a stronger sp3-type electronic contribution, related to the inclusion of sp3 connections in their surface network. 4. Health-related quality of life in patients with high-grade gliomas DEFF Research Database (Denmark) Piil, K; Jakobsen, J; Christensen, Karl Bang 2015-01-01 follow-up after 1 year. Scores of Karnofsky Performance Status (KPS), physical activity, anxiety and depression and health-related quality of life (FACT-Br) are obtained. Patients' physical activity level and KPS decrease during the disease- and treatment trajectory. The majority of patients did......The diagnosis of a high-grade glioma usual is followed by functional impairment(s), cognitive decline and an impaired psycho-social well-being. This might well have a significant and negative impact on the health related quality of life. The purpose of this study was to explore physical activity...... levels, prevalence and severity of anxiety and depressive symptoms and health-related quality of life among patients with a highgrade glioma. This paper is based on a longitudinal mixed methods study. Patients (n = 30) completed questionnaires at 5 time points from time of diagnosis until the final... 5. Improvement of the image quality of a high-temperature vision system International Nuclear Information System (INIS) Fabijańska, Anna; Sankowski, Dominik 2009-01-01 In this paper, the issues of controlling and improving the image quality of a high-temperature vision system are considered. The image quality improvement is needed to measure the surface properties of metals and alloys. Two levels of image quality control and improvement are defined in the system. The first level in hardware aims at adjusting the system configuration to obtain the highest contrast and weakest aura images. When optimal configuration is obtained, the second level in software is applied. In this stage, image enhancement algorithms are applied which have been developed with consideration of distortions arising from the vision system components and specificity of images acquired during the measurement process. The developed algorithms have been applied in the vision system to images. The influence on the accuracy of wetting angles and surface tension determination are considered 6. On beam quality and stopping power ratios for high-energy x-rays International Nuclear Information System (INIS) Johnsson, S.A.; Ceberg, C.P.; Knoeoes, T.; Nilsson, P. 2000-01-01 The aim of this work is to quantitatively compare two commonly used beam quality indices, TPR(20/10) and %dd(10) x , with respect to their ability to predict stopping power ratios (water to air), s w,air , for high-energy x-rays. In particular, effects due to a varied amount of filtration of the photon beam will be studied. A new method for characterizing beam quality is also presented, where the information we strive to obtain is the moments of the spectral distribution. We will show how the moments enter into a general description of the transmission curve and that it is possible to correlate the moments to s w,air with a unique and simple relationship. Comparisons with TPR(20/10) and %dd(10) x show that the moments are well suited for beam quality specification in terms of choosing the correct s w,air . (author) 7. Improving high resolution retinal image quality using speckle illumination HiLo imaging. Science.gov (United States) Zhou, Xiaolin; Bedggood, Phillip; Metha, Andrew 2014-08-01 Retinal image quality from flood illumination adaptive optics (AO) ophthalmoscopes is adversely affected by out-of-focus light scatter due to the lack of confocality. This effect is more pronounced in small eyes, such as that of rodents, because the requisite high optical power confers a large dioptric thickness to the retina. A recently-developed structured illumination microscopy (SIM) technique called HiLo imaging has been shown to reduce the effect of out-of-focus light scatter in flood illumination microscopes and produce pseudo-confocal images with significantly improved image quality. In this work, we adopted the HiLo technique to a flood AO ophthalmoscope and performed AO imaging in both (physical) model and live rat eyes. The improvement in image quality from HiLo imaging is shown both qualitatively and quantitatively by using spatial spectral analysis. 8. Low-cost high-quality crystalline germanium based flexible devices KAUST Repository Nassar, Joanna M. 2014-06-16 High performance flexible electronics promise innovative future technology for various interactive applications for the pursuit of low-cost, light-weight, and multi-functional devices. Thus, here we show a complementary metal oxide semiconductor (CMOS) compatible fabrication of flexible metal-oxide-semiconductor capacitors (MOSCAPs) with high-κ/metal gate stack, using a physical vapor deposition (PVD) cost-effective technique to obtain a high-quality Ge channel. We report outstanding bending radius ~1.25 mm and semi-transparency of 30%. 9. Low-cost high-quality crystalline germanium based flexible devices KAUST Repository Nassar, Joanna M.; Hussain, Aftab M.; Rojas, Jhonathan Prieto; Hussain, Muhammad Mustafa 2014-01-01 High performance flexible electronics promise innovative future technology for various interactive applications for the pursuit of low-cost, light-weight, and multi-functional devices. Thus, here we show a complementary metal oxide semiconductor (CMOS) compatible fabrication of flexible metal-oxide-semiconductor capacitors (MOSCAPs) with high-κ/metal gate stack, using a physical vapor deposition (PVD) cost-effective technique to obtain a high-quality Ge channel. We report outstanding bending radius ~1.25 mm and semi-transparency of 30%. 10. High quality-factor fano metasurface comprising a single resonator unit cell Science.gov (United States) Sinclair, Michael B.; Warne, Larry K.; Basilio, Lorena I.; Langston, William L.; Campione, Salvatore; Brener, Igal; Liu, Sheng 2017-06-20 A new monolithic resonator metasurface design achieves ultra-high Q-factors while using only one resonator per unit cell. The metasurface relies on breaking the symmetry of otherwise highly symmetric resonators to induce intra-resonator mixing of bright and dark modes (rather than inter-resonator couplings), and is scalable from the near-infrared to radio frequencies and can be easily implemented in dielectric materials. The resulting high-quality-factor Fano metasurface can be used in many sensing, spectral filtering, and modulation applications. 11. Production of high quality castile soap from high rancid olive oil Directory of Open Access Journals (Sweden) 2003-09-01 Full Text Available Non-edible olive oil, characterized by high acid and peroxide values as well as deep in color and unpleasant odor, was used to produce a fine castile soap (soap sample nº. 1. Semi-fine virgin olive oil was also used to produce the standard castile soap (soap sample nº. 2. The obtained results illustrated that the unpleasant odor was disappeared in soap nº. 1 compared to the standard soap (weakly like oil. Also, there were remarkable that no high differences were observed in all physical and chemical properties (appearance, smooth surface, erosion from hand-washing, consistency, moisture content, total fatty acids, free alkali and salt content in the two fresh soap samples. Whilst, the color in soap sample nº. 1 was fuscous green color compared to the standard soap (which was white to pale yellow. Soap samples were stored on a shelf at room temperature for 6 months showed some changes in their chemical properties. On the other hand, physical properties of the above two samples were improved after the storage period (6 months where their structures became very firm with high lather volume and rates of their erosions from hand-washing were retrenched except, the color in soap sample nº. 1 was not improved which was dark green color. Therefore, the present study recommend to use non-edible olive oil as unusually fatty material to produce a fine castile soap (high smooth surface, fairly lather and high glossy appearance as an alternative to edible olive oil (which is very expensive and also to reduce the cost of castile soap manufacturing.Aceite de oliva no comestible, caracterizado por su alta acidez e índice de peróxido así como de su elevada coloración y sabor desagradable se utilizó para la producción de jabón fino tipo ‘‘Castilla’’ (muestra de jabón nº 1. Otro aceite de oliva semifino se empleo para la fabricación de jabón estándar tipo ‘‘Castilla’’. Los resultados mostraron 12. Urological symptoms in a subset of patients with urological chronic pelvic pain syndrome and a polysymptomatic, polysyndromic pattern of presentation. Science.gov (United States) Lai, H Henry; North, Carol S; Andriole, Gerald L; Cupps, Lori; Song, David; Ness, Timothy J; Hong, Barry A 2014-06-01 We characterized urological symptoms in a subset of patients with urological chronic pelvic pain syndrome who have a high somatic symptom burden and a wide symptom distribution fitting a polysymptomatic, polysyndromic presentation pattern. A total of 81 patients with urological chronic pelvic pain syndrome enrolled in the National Institute of Diabetes and Digestive and Kidney Diseases MAPP Research Network Study at Washington University in St. Louis and University of Alabama at Birmingham sites. They completed a symptom questionnaire to assess the somatic symptom burden and its distribution, and GUPI (Genitourinary Pain Index) to assess urological chronic pelvic pain syndrome symptoms, impact on quality of life and self-reported treatment seeking behaviors for urological chronic pelvic pain symptoms. The polysymptomatic, polysyndromic symptom pattern was defined by self-report of numerous painful and nonpainful somatic symptoms across many organ systems and by symptom categories on the polysymptomatic, polysyndromic questionnaire. Patients with urological chronic pelvic pain syndrome and the symptom pattern reported more severe genitourinary pain on a Likert scale, more frequent pain in the last week and more widespread pain distribution in the genital and pelvic areas than patients with urological chronic pelvic pain syndrome without the pattern. Patients with the symptom pattern also had significantly higher scores on the GUPI pain subscale, quality of life subscale (worse) and total questionnaire scores than patients without the pattern. Patients with the pattern reported significantly more treatment seeking behavior than others. The polysymptomatic, polysyndromic pattern might be an important phenotypic factor to assess in the evaluation of urological chronic pelvic pain syndrome with clinical and research implications. This may be a distinct clinical subgroup among patients with urological chronic pelvic pain syndrome. Copyright © 2014 American Urological 13. Blood Monocyte Subsets and Selected Cardiovascular Risk Markers in Rheumatoid Arthritis of Short Duration in relation to Disease Activity Directory of Open Access Journals (Sweden) Ewa Klimek 2014-01-01 Full Text Available Objectives. To evaluate blood monocyte subsets and functional monocyte properties in patients with rheumatoid arthritis (RA of short duration in the context of cardiovascular (CV risk and disease activity. Methods. We studied conventional markers of CV risk, intima media thickness (IMT, and blood monocyte subsets in 27 patients aged 41 ± 10 years with RA of short duration (median 12 months and 22 healthy controls. The RA subjects were divided into low (DAS28: 2.6–5.1 and high (DAS28 > 5.1 disease activity. Results. RA patients exhibited increased levels of intermediate (CD14++CD16+ monocytes with decreased CD45RA expression compared to controls, increased counts of classical (CD14++CD16− monocytes, and decreased percentages of nonclassical (CD14+CD16++ monocytes. Patients with high disease activity had lower HLA DR expression on classical monocytes compared to low disease activity patients. There were no differences in monocyte subsets between subjects with DAS > 5.1 and DAS ≤ 5.1. There were no significant intergroup differences in IMT and the majority of classical CV risk factors. Conclusions. Patients with RA of short duration show alteration in peripheral blood monocyte subsets despite the fact that there is no evidence of subclinical atherosclerosis. Disease activity assessed with DAS28 was associated with impaired functional properties but not with a shift in monocyte subpopulations. 14. Cerebellins are differentially expressed in selective subsets of neurons throughout the brain. Science.gov (United States) Seigneur, Erica; Südhof, Thomas C 2017-10-15 Cerebellins are secreted hexameric proteins that form tripartite complexes with the presynaptic cell-adhesion molecules neurexins or 'deleted-in-colorectal-cancer', and the postsynaptic glutamate-receptor-related proteins GluD1 and GluD2. These tripartite complexes are thought to regulate synapses. However, cerebellins are expressed in multiple isoforms whose relative distributions and overall functions are not understood. Three of the four cerebellins, Cbln1, Cbln2, and Cbln4, autonomously assemble into homohexamers, whereas the Cbln3 requires Cbln1 for assembly and secretion. Here, we show that Cbln1, Cbln2, and Cbln4 are abundantly expressed in nearly all brain regions, but exhibit strikingly different expression patterns and developmental dynamics. Using newly generated knockin reporter mice for Cbln2 and Cbln4, we find that Cbln2 and Cbln4 are not universally expressed in all neurons, but only in specific subsets of neurons. For example, Cbln2 and Cbln4 are broadly expressed in largely non-overlapping subpopulations of excitatory cortical neurons, but only sparse expression was observed in excitatory hippocampal neurons of the CA1- or CA3-region. Similarly, Cbln2 and Cbln4 are selectively expressed, respectively, in inhibitory interneurons and excitatory mitral projection neurons of the main olfactory bulb; here, these two classes of neurons form dendrodendritic reciprocal synapses with each other. A few brain regions, such as the nucleus of the lateral olfactory tract, exhibit astoundingly high Cbln2 expression levels. Viewed together, our data show that cerebellins are abundantly expressed in relatively small subsets of neurons, suggesting specific roles restricted to subsets of synapses. © 2017 Wiley Periodicals, Inc. 15. Delivering high quality hip fracture rehabilitation: the perspective of occupational and physical therapy practitioners. Science.gov (United States) Leland, Natalie E; Lepore, Michael; Wong, Carin; Chang, Sun Hwa; Freeman, Lynn; Crum, Karen; Gillies, Heather; Nash, Paul 2018-03-01 The majority of post-acute hip fracture rehabilitation in the US is delivered in skilled nursing facilities (SNFs). Currently, there are limited guidelines that equip occupational and physical therapy practitioners with a summary of what constitutes evidence-based high quality rehabilitation. Thus, this study aimed to identify rehabilitation practitioners' perspectives on the practices that constitute high quality hip fracture rehabilitation. Focus groups were conducted with 99 occupational and physical therapy practitioners working in SNFs in southern California. Purposive sampling of facilities was conducted to capture variation in key characteristics known to impact care delivery for this patient population (e.g., financial resources, staffing, and patient case-mix). Questions aimed to elicit practitioners' perspectives on high quality hip fracture rehabilitation practices. Each session was audio-recorded and transcribed. Data were systematically analyzed using a modified grounded theory approach. Seven themes emerged: objectives of care; first 72 h; positioning, pain, and precautions; use of standardized assessments; episode of care practices; facilitating insight into progress; and interdisciplinary collaboration. Clinical guidelines are critical tools to facilitate clinical decision-making and achieve desired patient outcomes. The findings of this study highlight the practitioners' perspective on what constitutes high quality hip fracture rehabilitation. This work provides critical information to advance the development of stakeholder-driven rehabilitation clinical guidelines. Future research is needed to verify the findings from other stakeholders (e.g., patients), ensure the alignment of our findings with current evidence, and develop measures for evaluating their delivery and relationship to desired outcomes. Implications for Rehabilitation This study highlights occupational and physiotherapy therapy practitioners' perspectives on the cumulative best 16. Quality of care and patient satisfaction in hospitals with high concentrations of black patients. Science.gov (United States) Brooks-Carthon, J Margo; Kutney-Lee, Ann; Sloane, Douglas M; Cimiotti, Jeannie P; Aiken, Linda H 2011-09-01 To examine the influence of nursing-specifically nurse staffing and the nurse work environment-on quality of care and patient satisfaction in hospitals with varying concentrations of Black patients. Cross-sectional secondary analysis of 2006-2007 nurse survey data collected across four states (Florida, Pennsylvania, New Jersey, and California), the Hospital Consumer Assessment of Healthcare Providers and Systems survey, and administrative data. Global analysis of variance and linear regression models were used to examine the association between the concentration of Black patients on quality measures (readiness for discharge, patient or family complaints, health care-associated infections) and patient satisfaction, before and after accounting for nursing and hospital characteristics. Nurses working in hospitals with higher concentrations of Blacks reported poorer confidence in patients' readiness for discharge and more frequent complaints and infections. Patients treated in hospitals with higher concentrations of Blacks were less satisfied with their care. In the fully adjusted regression models for quality and patient satisfaction outcomes, the effects associated with the concentration of Blacks were explained in part by nursing and structural hospital characteristics. This study demonstrates a relationship between nursing, structural hospital characteristics, quality of care, and patient satisfaction in hospitals with high concentrations of Black patients. Consideration of nursing factors, in addition to other important hospital characteristics, is critical to understanding and improving quality of care and patient satisfaction in minority-serving hospitals. © 2011 Sigma Theta Tau International. 17. Biophysical subsets of embryonic stem cells display distinct phenotypic and morphological signatures. Directory of Open Access Journals (Sweden) Tom Bongiorno Full Text Available The highly proliferative and pluripotent characteristics of embryonic stem cells engender great promise for tissue engineering and regenerative medicine, but the rapid identification and isolation of target cell phenotypes remains challenging. Therefore, the objectives of this study were to characterize cell mechanics as a function of differentiation and to employ differences in cell stiffness to select population subsets with distinct mechanical, morphological, and biological properties. Biomechanical analysis with atomic force microscopy revealed that embryonic stem cells stiffened within one day of differentiation induced by leukemia inhibitory factor removal, with a lagging but pronounced change from spherical to spindle-shaped cell morphology. A microfluidic device was then employed to sort a differentially labeled mixture of pluripotent and differentiating cells based on stiffness, resulting in pluripotent cell enrichment in the soft device outlet. Furthermore, sorting an unlabeled population of partially differentiated cells produced a subset of "soft" cells that was enriched for the pluripotent phenotype, as assessed by post-sort characterization of cell mechanics, morphology, and gene expression. The results of this study indicate that intrinsic cell mechanical properties might serve as a basis for efficient, high-throughput, and label-free isolation of pluripotent stem cells, which will facilitate a greater biological understanding of pluripotency and advance the potential of pluripotent stem cell differentiated progeny as cell sources for tissue engineering and regenerative medicine. 18. Marine Corps Contract Supply Model for High Quality Male Enlistment Contracts at the Recruiting Sub Station Level National Research Council Canada - National Science Library Welsh, Brian K 2008-01-01 .... This research develops contract production models at the Recruiting Sub Station level to estimate the effects of local economic conditions, demographics, and recruiting resources on new high quality... 19. Fast Batch Production of High-Quality Graphene Films in a Sealed Thermal Molecular Movement System. Science.gov (United States) Xu, Jianbao; Hu, Junxiong; Li, Qi; Wang, Rubing; Li, Weiwei; Guo, Yufen; Zhu, Yongbo; Liu, Fengkui; Ullah, Zaka; Dong, Guocai; Zeng, Zhongming; Liu, Liwei 2017-07-01 Chemical vapor deposition (CVD) growth of high-quality graphene has emerged as the most promising technique in terms of its integrated manufacturing. However, there lacks a controllable growth method for producing high-quality and a large-quantity graphene films, simultaneously, at a fast growth rate, regardless of roll-to-roll (R2R) or batch-to-batch (B2B) methods. Here, a stationary-atmospheric-pressure CVD (SAPCVD) system based on thermal molecular movement, which enables fast B2B growth of continuous and uniform graphene films on tens of stacked Cu(111) foils, with a growth rate of 1.5 µm s -1 , is demonstrated. The monolayer graphene of batch production is found to nucleate from arrays of well-aligned domains, and the films possess few defects and exhibit high carrier mobility up to 6944 cm 2 V -1 s -1 at room temperature. The results indicate that the SAPCVD system combined with single-domain Cu(111) substrates makes it possible to realize fast batch-growth of high-quality graphene films, which opens up enormous opportunities to use this unique 2D material for industrial device applications. © 2017 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim. 20. Air Quality Standards for Particulate Matter (PM) at high altitude cities International Nuclear Information System (INIS) Bravo Alvarez, H.; Sosa Echeverria, R.; Sanchez Alvarez, P.; Krupa, S. 2013-01-01 The Air Quality Standards for Particulate Matter (PM) at high altitude urban areas in different countries, must consider the pressure and temperature due to the effect that these parameters have on the breath volume. This paper shows the importance to correct Air Quality Standards for PM considering pressure and temperature at different altitudes. Specific factors were suggested to convert the information concerning PM, from local to standard conditions, and adjust the Air Quality Standards for different high altitudes cities. The correction factors ranged from: 1.03 for Santiago de Chile to 1.47 for El Alto Bolivia. Other cities in this study include: Mexico City, México; La Paz, Bolivia; Bogota, Cali and Medellin, Colombia; Quito, Ecuador and Cuzco, Peru. If these corrections are not considered, the atmospheric concentrations will be underestimated. - Highlights: ► AQS for particulate matter concentrations adjusted by pressure and temperature. ► Particulate matter concentrations can be underestimated in high altitude Cities. ► Particulate matter concentrations must be compared under the same conditions. - In order to compare high altitude atmospheric PM concentrations with AQS, one must consider T and P of the sampling site. 1. High-quality GaN epitaxially grown on Si substrate with serpentine channels Science.gov (United States) Wei, Tiantian; Zong, Hua; Jiang, Shengxiang; Yang, Yue; Liao, Hui; Xie, Yahong; Wang, Wenjie; Li, Junze; Tang, Jun; Hu, Xiaodong 2018-06-01 A novel serpentine-channeled mask was introduced to Si substrate for low-dislocation GaN epitaxial growth and the fully coalesced GaN film on the masked Si substrate was achieved for the first time. Compared with the epitaxial lateral overgrowth (ELOG) growth method, this innovative mask only requires one-step epitaxial growth of GaN which has only one high-dislocation region per mask opening. This new growth method can effectively reduce dislocation density, thus improving the quality of GaN significantly. High-quality GaN with low dislocation density ∼2.4 × 107 cm-2 was obtained, which accounted for about eighty percent of the GaN film in area. This innovative technique is promising for the growth of high-quality GaN templates and the subsequent fabrication of high-performance GaN-based devices like transistors, laser diodes (LDs), and light-emitting diodes (LEDs) on Si substrate. 2. Personality and Relationship Quality During the Transition From High School to Early Adulthood Science.gov (United States) Parker, Philip D.; Lüdtke, Oliver; Trautwein, Ulrich; Roberts, Brent W. 2013-01-01 The post–high school transition period is believed to be associated with considerable changes in social networks, yet longitudinal studies documenting these changes are scarce. To address this gap, the current research explored 3 relevant issues. First, changes in participants’ relationship characteristics during the transition from high school were examined. Second, the roles of personality traits as antecedents of these changes were studied. Third, the association between change in relationship characteristics and personality during the transition was explored. A sample of over 2,000 German emerging adults, surveyed before leaving school and then 2 years after the transition from high school, was assessed on personality traits and a multidimensional assessment of the quality of their relationships. Findings indicated that participants experienced mostly positive changes in relationship quality during the transition from high school and that antecedent personality at school was an important predictor of the nature of this change. Finally, change in relationship quality was found to be associated with personality change during the post-school transition. Findings indicated that personality traits may influence transition success and that change in relationships during this transition may influence personality development. The implications of the research for post-school transition success are discussed. PMID:22224909 3. Aberrant T Cell Signaling and Subsets in Systemic Lupus Erythematosus Directory of Open Access Journals (Sweden) Takayuki Katsuyama 2018-05-01 Full Text Available Systemic lupus erythematosus (SLE is a chronic multi-organ debilitating autoimmune disease, which mainly afflicts women in the reproductive years. A complex interaction of genetics, environmental factors and hormones result in the breakdown of immune tolerance to “self” leading to damage and destruction of multiple organs, such as the skin, joints, kidneys, heart and brain. Both innate and adaptive immune systems are critically involved in the misguided immune response against self-antigens. Dendritic cells, neutrophils, and innate lymphoid cells are important in initiating antigen presentation and propagating inflammation at lymphoid and peripheral tissue sites. Autoantibodies produced by B lymphocytes and immune complex deposition in vital organs contribute to tissue damage. T lymphocytes are increasingly being recognized as key contributors to disease pathogenesis. CD4 T follicular helper cells enable autoantibody production, inflammatory Th17 subsets promote inflammation, while defects in regulatory T cells lead to unchecked immune responses. A better understanding of the molecular defects including signaling events and gene regulation underlying the dysfunctional T cells in SLE is necessary to pave the path for better management, therapy, and perhaps prevention of this complex disease. In this review, we focus on the aberrations in T cell signaling in SLE and highlight therapeutic advances in this field. 4. Application of subset simulation in reliability estimation of underground pipelines International Nuclear Information System (INIS) Tee, Kong Fah; Khan, Lutfor Rahman; Li, Hongshuang 2014-01-01 This paper presents a computational framework for implementing an advanced Monte Carlo simulation method, called Subset Simulation (SS) for time-dependent reliability prediction of underground flexible pipelines. The SS can provide better resolution for low failure probability level of rare failure events which are commonly encountered in pipeline engineering applications. Random samples of statistical variables are generated efficiently and used for computing probabilistic reliability model. It gains its efficiency by expressing a small probability event as a product of a sequence of intermediate events with larger conditional probabilities. The efficiency of SS has been demonstrated by numerical studies and attention in this work is devoted to scrutinise the robustness of the SS application in pipe reliability assessment and compared with direct Monte Carlo simulation (MCS) method. Reliability of a buried flexible steel pipe with time-dependent failure modes, namely, corrosion induced deflection, buckling, wall thrust and bending stress has been assessed in this study. The analysis indicates that corrosion induced excessive deflection is the most critical failure event whereas buckling is the least susceptible during the whole service life of the pipe. The study also shows that SS is robust method to estimate the reliability of buried pipelines and it is more efficient than MCS, especially in small failure probability prediction 5. Aberrant T Cell Signaling and Subsets in Systemic Lupus Erythematosus Science.gov (United States) Katsuyama, Takayuki; Tsokos, George C.; Moulton, Vaishali R. 2018-01-01 Systemic lupus erythematosus (SLE) is a chronic multi-organ debilitating autoimmune disease, which mainly afflicts women in the reproductive years. A complex interaction of genetics, environmental factors and hormones result in the breakdown of immune tolerance to “self” leading to damage and destruction of multiple organs, such as the skin, joints, kidneys, heart and brain. Both innate and adaptive immune systems are critically involved in the misguided immune response against self-antigens. Dendritic cells, neutrophils, and innate lymphoid cells are important in initiating antigen presentation and propagating inflammation at lymphoid and peripheral tissue sites. Autoantibodies produced by B lymphocytes and immune complex deposition in vital organs contribute to tissue damage. T lymphocytes are increasingly being recognized as key contributors to disease pathogenesis. CD4 T follicular helper cells enable autoantibody production, inflammatory Th17 subsets promote inflammation, while defects in regulatory T cells lead to unchecked immune responses. A better understanding of the molecular defects including signaling events and gene regulation underlying the dysfunctional T cells in SLE is necessary to pave the path for better management, therapy, and perhaps prevention of this complex disease. In this review, we focus on the aberrations in T cell signaling in SLE and highlight therapeutic advances in this field. PMID:29868033 6. Decontaminate feature for tracking: adaptive tracking via evolutionary feature subset Science.gov (United States) Liu, Qiaoyuan; Wang, Yuru; Yin, Minghao; Ren, Jinchang; Li, Ruizhi 2017-11-01 Although various visual tracking algorithms have been proposed in the last 2-3 decades, it remains a challenging problem for effective tracking with fast motion, deformation, occlusion, etc. Under complex tracking conditions, most tracking models are not discriminative and adaptive enough. When the combined feature vectors are inputted to the visual models, this may lead to redundancy causing low efficiency and ambiguity causing poor performance. An effective tracking algorithm is proposed to decontaminate features for each video sequence adaptively, where the visual modeling is treated as an optimization problem from the perspective of evolution. Every feature vector is compared to a biological individual and then decontaminated via classical evolutionary algorithms. With the optimized subsets of features, the "curse of dimensionality" has been avoided while the accuracy of the visual model has been improved. The proposed algorithm has been tested on several publicly available datasets with various tracking challenges and benchmarked with a number of state-of-the-art approaches. The comprehensive experiments have demonstrated the efficacy of the proposed methodology. 7. Lymphocyte subsets and response to PHA among atomic bomb survivors International Nuclear Information System (INIS) Nakao, Susumu; Noguchi, Kyouichi; Eida, Kazuyuki; Tashiro, Kazunori; Hayashida, Ken 1986-01-01 In an effort to elucidate the effect of radiation exposure on immune competence in man, the number of lymphocytes, lymphocyte subsets, and the percentage of phytohemagglutinin (PHA)-induced transformation of lymphocytes were determined in 66 cancer patients, 25 of whom were exposed to atomic radiation at ≤ 2,000 m from ground zero and 41 others were not exposed. The number of lymphocytes was decreased with increasing age at exposure. The percentage of OKT3-positive cells tended to be lower in exposed patients who were in their twenties at the time of exposure than the non-exposed patients. Among patients in their teens and twenties at the time of exposure, there was a tendency toward decreased percentage of OKT4-positive cells (T4) and increased percentage of OKT8-positive cells (T8). The T4/T8 ratio was reduced. Patients who were in their first decade of life at the time of exposure tended to have decreased OKIa 1-positive cells, and increased Leulla-positive cells. Patients exposed in their twenties and thirties had slightly decreased percentage of PHA-induced transformation of lymphocytes. (Namekawa, K.) 8. Key elements of high-quality practice organisation in primary health care: a systematic review. Science.gov (United States) Crossland, Lisa; Janamian, Tina; Jackson, Claire L 2014-08-04 To identify elements that are integral to high-quality practice and determine considerations relating to high-quality practice organisation in primary care. A narrative systematic review of published and grey literature. Electronic databases (PubMed, CINAHL, the Cochrane Library, Embase, Emerald Insight, PsycInfo, the Primary Health Care Research and Information Service website, Google Scholar) were searched in November 2013 and used to identify articles published in English from 2002 to 2013. Reference lists of included articles were searched for relevant unpublished articles and reports. Data were configured at the study level to allow for the inclusion of findings from a broad range of study types. Ten elements were most often included in the existing organisational assessment tools. A further three elements were identified from an inductive thematic analysis of descriptive articles, and were noted as important considerations in effective quality improvement in primary care settings. Although there are some validated tools available to primary care that identify and build quality, most are single-strategy approaches developed outside health care settings. There are currently no validated organisational improvement tools, designed specifically for primary health care, which combine all elements of practice improvement and whose use does not require extensive external facilitation. 9. A high efficiency, high quality and low cost internal regulated bioanalytical laboratory to support drug development needs. Science.gov (United States) Song, Yan; Dhodda, Raj; Zhang, Jun; Sydor, Jens 2014-05-01 In the recent past, we have seen an increase in the outsourcing of bioanalysis in pharmaceutical companies in support of their drug development pipeline. This trend is largely driven by the effort to reduce internal cost, especially in support of late-stage pipeline assets where established bioanalytical assays are used to analyze a large volume of samples. This article will highlight our perspective of how bioanalytical laboratories within pharmaceutical companies can be developed into the best partner in the advancement of drug development pipelines with high-quality support at competitive cost. 10. Optimization of laser parameters to obtain high-energy, high-quality electron beams through laser-plasma acceleration International Nuclear Information System (INIS) Samant, Sushil Arun; Sarkar, Deepangkar; Krishnagopal, Srinivas; Upadhyay, Ajay K.; Jha, Pallavi 2010-01-01 The propagation of an intense (a 0 =3), short-pulse (L∼λ p ) laser through a homogeneous plasma has been investigated. Using two-dimensional simulations for a 0 =3, the pulse-length and spot-size at three different plasma densities were optimized in order to get a better quality beam in laser wakefield accelerator. The study reveals that with increasing pulse-length the acceleration increases, but after a certain pulse-length (L>0.23λ p ) the emittance blows-up unacceptably. For spot-sizes less than that given by k p0 r s =2√(a 0 ), trapping is poor or nonexistent, and the optimal spot-size is larger. The deviation of the optimal spot-size from this formula increases as the density decreases. The efficacy of these two-dimensional simulations has been validated by running three-dimensional simulations at the highest density. It has been shown that good quality GeV-class beams can be obtained at plasma densities of ∼10 18 cm -3 . The quality of the beam can be substantially improved by selecting only the high-energy peak; in this fashion an energy-spread of better than 1% and a current in tens of kA can be achieved, which are important for applications such as free-electron lasers. 11. The use of induced mutation combined with crossing in high quality rice breeding Energy Technology Data Exchange (ETDEWEB) Do Huu At; Bui Huy Thuy; Nguyen Van Bich; Tran Duy Quy [Agricultural Genetics Institute, Division of Genetics and Hybrid Rice Technology, Hanoi (Viet Nam); Nguyen Minh Cong [Hanoi No. 1 Teacher Training Univ., Department of Genetics (Viet Nam) 2001-03-01 The high quality rice varieties: Tam thom mutant rice Var., DT17 rice Var, DT21 glutinous rice Var were formed by induced mutation combined with crossing. Tam thom mutant rice Var. lost photosensitivity, could be planted 2 crops/year. DT17 rice Var with high yielding capacity, suitable for growth on lowland in summer crop, is replacing step-by-step Moctuyen rice Var. in North Vietnam. DT21 glutinous rice Var. could be planted 2 crops/year and had short growth duration, average yield was 4.0-4.5 tons/ha. These three ones had good quality, soft and scent cooked rice, suitable for customers and export requirements. Tam thom mutant rice Var. DT17 rice Var., DT21 and glutinous rice Var. were adopted for regional production by Ministry of Agriculture and Rural Development and allowed to be in trial production. (author) 12. Attempt at cloning high-quality goldfish breed 'Ranchu' by fin-cultured cell nuclear transplantation. Science.gov (United States) Tanaka, Daisuke; Takahashi, Akito; Takai, Akinori; Ohta, Hiromi; Ueno, Koichi 2012-02-01 The viability of ornamental fish culture relies on the maintenance of high-quality breeds. To improve the profitability of culture operations we attempted to produce cloned fish from the somatic nucleus of the high-quality Japanese goldfish (Carassius auratus auratus) breed 'Ranchu'. We transplanted the nucleus of a cultured fin-cell from an adult Ranchu into the non-enucleated egg of the original goldfish breed 'Wakin'. Of the 2323 eggs we treated, 802 underwent cleavage, 321 reached the blastula stage, and 51 reached the gastrula stage. Two of the gastrulas developed until the hatching stage. A considerable number of nuclear transplants retained only the donor nucleus. Some of these had only a 2n nucleus derived from the same donor fish. Our results provide insights into the process of somatic cell nuclear transplantation in teleosts, and the cloning of Ranchu. 13. The use of induced mutation combined with crossing in high quality rice breeding International Nuclear Information System (INIS) Do Huu At; Bui Huy Thuy; Nguyen Van Bich; Tran Duy Quy; Nguyen Minh Cong 2001-01-01 The high quality rice varieties: Tam thom mutant rice Var., DT17 rice Var, DT21 glutinous rice Var were formed by induced mutation combined with crossing. Tam thom mutant rice Var. lost photosensitivity, could be planted 2 crops/year. DT17 rice Var with high yielding capacity, suitable for growth on lowland in summer crop, is replacing step-by-step Moctuyen rice Var. in North Vietnam. DT21 glutinous rice Var. could be planted 2 crops/year and had short growth duration, average yield was 4.0-4.5 tons/ha. These three ones had good quality, soft and scent cooked rice, suitable for customers and export requirements. Tam thom mutant rice Var. DT17 rice Var., DT21 and glutinous rice Var. were adopted for regional production by Ministry of Agriculture and Rural Development and allowed to be in trial production. (author) 14. High quality junctions by interpenetration of vapor liquid solid grown nanostructures for microchip integration Energy Technology Data Exchange (ETDEWEB) Jebril, Seid; Kuhlmann, Hanna; Adelung, Rainer [Funktionale Nanomaterialien, CAU Kiel (Germany); Mueller, Sven [Nanowires and Thin Films, II. Physikalisches Institut, Goettingen (Germany); Ronning, Carsten [Institute for Solid State Physics, Universitaet Jena (Germany); Kienle, Lorenz [Synthese und Realstruktur, CAU Kiel (Germany); Duppel, Viola [MPI fuer Festkoerperforschung, Stuttgart (Germany) 2009-07-01 The usability of nanostructures in electrical devices like gas sensors depends critically on the ability to form high quality contacts and junctions. For the fabrication of various nanostructures, vapor-liquid-solid (VLS) growth is a wide spread and very efficient technique. However, forming contacts with the VLS grown structures to utilize them in a device is still tedious, because either the substrate has to be epitaxial to the VLS material or a manual alignment is necessary. Here we demonstrate the contact formation by simply using the ability of individual crystals to interpenetrate each other during the straight forward VLS growth. This allows growing VLS structures directly on two neighboring gold circuit paths of a microchip; bridges over predefined gaps will be formed. Moreover, TEM investigations confirm the high quality of the crystalline junctions that allow demonstrations as UV and hydrogen-sensor. The VLS devices are compared with conventional produced. 15. Bundled Payments in Total Joint Replacement: Keeping Our Care Affordable and High in Quality. Science.gov (United States) McLawhorn, Alexander S; Buller, Leonard T 2017-09-01 The purpose of this review was to evaluate the literature regarding bundle payment reimbursement models for total joint arthroplasty (TJA). From an economic standpoint, TJA are cost-effective, but they represent a substantial expense to the Centers for Medicare & Medicaid Services (CMS). Historically, fee-for-service payment models resulted in highly variable cost and quality. CMS introduced Bundled Payments for Care Improvement (BPCI) in 2012 and subsequently the Comprehensive Care for Joint Replacement (CJR) reimbursement model in 2016 to improve the value of TJA from the perspectives of both CMS and patients, by improving quality via cost control. Early results of bundled payments are promising, but preserving access to care for patients with high comorbidity burdens and those requiring more complex care is a lingering concern. Hospitals, regardless of current participation in bundled payments, should develop care pathways for TJA to maximize efficiency and patient safety. 16. A Portable Low-Cost High Density Sensor Network for Air Quality at London Heathrow Airport Science.gov (United States) Popoola, Olalekan; Mead, Iq; Bright, Vivien; Baron, Ronan; Saffell, John; Stewart, Gregor; Kaye, Paul; Jones, Roderic 2013-04-01 Outdoor air quality and its impact on human health and the environment have been well studied and it has been projected that poor air quality will surpass poor sanitation as the major course of environmental premature mortality by 2050 (IGAC / IGBP, release statement, 2012). Transport-related pollution has been regulated at various levels by enactment of legislations at local, national, regional and global stages. As part of the mitigation measures, routine measurements of atmospheric pollutants such as carbon monoxide (CO), nitric oxide (NO) and nitrogen dioxide (NO2) have to be established in areas where air quality problems are identified. In addition, emission inventories are also generated for different atmospheric environments including urban areas and airport environments required for air quality models. Whilst recognising that most of the existing sparse monitoring networks provide high temporal measurements, spatial data of these highly variable pollutants are not captured, making it difficult to adequately characterise the highly heterogeneous air quality. Spatial information is often obtained from model data which can only be constrained using measurements from the sparse monitoring networks. The work presented here shows the application of low-cost sensor networks aimed at addressing this missing spatial information. We have shown in previous studies the application of low-cost electrochemical sensor network instruments in monitoring road transport pollutants including CO, NO and NO2 in an urban environment (Mead et. al. 2012, accepted Atmospheric Environment). Modified versions of these instruments which include additional species such as O3, SO2, VOCs and CO2 are currently deployed at London Heathrow Airport (LHR) as part of the Sensor Network for Air Quality (SNAQ) project. Meteorology data such as temperature, relative humidity, wind speed and direction are also measured as well as size-speciated particulates (0.38 to 17.4 µm). A network of 50 17. Manufacturing of High Quality Teachers for Chemistry Education at Higher Secondary Level in Current Era OpenAIRE R. Azmat 2013-01-01 Education is the most influential tool whose efficient use requires the power of determination, devoted work and sacrifice. As teachers are major handler of this tool therefore, they must possess qualities of high education and competency for deliver. Education extends attractive way of life, talent and manners which make an individual a fine civilian. Primary and secondary education is one of the foundation stone of development of children and country. It acts as a vital part in placing the ... 18. Fabrication of high quality GaN nanopillar arrays by dry and wet chemical etching OpenAIRE Paramanik, Dipak; Motayed, Abhishek; King, Matthew; Ha, Jong-Yoon; Kryluk, Sergi; Davydov, Albert V.; Talin, Alec 2013-01-01 We study strain relaxation and surface damage of GaN nanopillar arrays fabricated using inductively coupled plasma (ICP) etching and post etch wet chemical treatment. We controlled the shape and surface damage of such nanopillar structures through selection of etching parameters. We compared different substrate temperatures and different chlorine-based etch chemistries to fabricate high quality GaN nanopillars. Room temperature photoluminescence and Raman scattering measurements were carried ... 19. Quality control of 192Ir high dose rate after loading brachytherapy dose veracity International Nuclear Information System (INIS) Feng Zhongsu; Xu Xiao; Liu Fen 2008-01-01 Recently, 192 Ir high dose rate (HDR) afterloading are widely used in brachytherapy. The advantage of using HDR systems over low dose rate systems are shorter treatment time and higher fraction dose. To guarantee the veracity of the delivery dose, several quality control methods are deseribed in this work. With these we can improve the position precision, time precision and dose precision of the brachytherapy. (authors) 20. Public reporting on quality, waiting times and patient experience in 11 high-income countries OpenAIRE Rechel, Bernd; McKee, Martin; Haas, Marion; Marchildon, Gregory P; Bousquet, Frederic; Blümel, Miriam; Geissler, Alexander; van Ginneken, Ewout; Ashton, Toni; Saunes, Ingrid Sperre; Anell, Anders; Quentin, Wilm; Saltman, Richard; Culler, Steven; Barnes, Andrew 2016-01-01 : This article maps current approaches to public reporting on waiting times, patient experience and aggregate measures of quality and safety in 11 high-income countries (Australia, Canada, England, France, Germany, Netherlands, New Zealand, Norway, Sweden, Switzerland and the United States). Using a questionnaire-based survey of key national informants, we found that the data most commonly made available to the public are on waiting times for hospital treatment, being reported for major hospi... 1. Efficient quality-eactor estimation of a vertical cavity employing a high-contrast grating DEFF Research Database (Denmark) Taghizadeh, Alireza; Mørk, Jesper; Chung, Il-Sug 2017-01-01 Hybrid vertical cavity lasers employing high-contrast grating reflectors are attractive for Si-integrated light source applications. Here, a method for reducing a three-dimensional (3D) optical simulation of this laser structure to lower-dimensional simulations is suggested, which allows for very...... fast and approximate analysis of the quality-factor of the 3D cavity. This approach enables us to efficiently optimize the laser cavity design without performing cumbersome 3D simulations.... 2. Neglecting rice milling yield and quality underestimates economic losses from high-temperature stress. Directory of Open Access Journals (Sweden) Nathaniel B Lyman Full Text Available Future increases in global surface temperature threaten those worldwide who depend on rice production for their livelihoods and food security. Past analyses of high-temperature stress on rice production have focused on paddy yield and have failed to account for the detrimental impact of high temperatures on milling quality outcomes, which ultimately determine edible (marketable rice yield and market value. Using genotype specific rice yield and milling quality data on six common rice varieties from Arkansas, USA, combined with on-site, half-hourly and daily temperature observations, we show a nonlinear effect of high-temperature stress exposure on yield and milling quality. A 1 °C increase in average growing season temperature reduces paddy yield by 6.2%, total milled rice yield by 7.1% to 8.0%, head rice yield by 9.0% to 13.8%, and total milling revenue by 8.1% to 11.0%, across genotypes. Our results indicate that failure to account for changes in milling quality leads to understatement of the impacts of high temperatures on rice production outcomes. These dramatic losses result from reduced paddy yield and increased percentages of chalky and broken kernels, which together decrease the quantity and market value of milled rice. Recently published estimates show paddy yield reductions of up to 10% across the major rice-producing regions of South and Southeast Asia due to rising temperatures. The results of our study suggest that the often-cited 10% figure underestimates the economic implications of climate change for rice producers, thus potentially threatening future food security for global rice producers and consumers. 3. Quantitative Analysis of High-Quality Officer Selection by Commandants Career-Level Education Board Science.gov (United States) 2017-03-01 impact on the organization and allocate resources to improve the human capital of this select group. From 2011 onward, CCLEB revamped the application...ANALYSIS OF HIGH-QUALITY OFFICER SELECTION BY COMMANDANT’S CAREER - LEVEL EDUCATION BOARD by Clifton N. Rateike March 2017 Thesis Advisor...of Management and Budget, Paperwork Reduction Project (0704-0188) Washington, DC 20503. 1. AGENCY USE ONLY (Leave blank) 2. REPORT DATE March 4. Burnout in boiling heat transfer. Part III. High-quality forced-convection systems International Nuclear Information System (INIS) Bergles, A.E. 1979-01-01 This is the final part of a review of burnout during boiling heat transfer. The status of burnout in high-quality forced-convection systems is reviewed, and recent developments are summarized in detail. A general guide to the considerable literature is given. Parametric effects and correlations for water in circular and noncircular ducts are presented. Other topics discussed include transients, steam-generator applications, correlations for other fluids, fouling, and augmentation 5. The Total Quality Management philosophy: The Deming's principles in two Portuguese institutions of high education OpenAIRE Saraiva, Margarida 2006-01-01 Usually, the process of TQM is implemented in the industry and in services through the adoption of a group of principles. Deming that presented a set of principles to improve the quality process in a organisation. The study applied some of the Deming’s principles in two Portuguese institutions of high education. The objective was to understand if the Deming’s principles could be applied. The results showed a group of possible difficulties and benefits and concluded that can be ... 6. Different paths to high-quality care: three archetypes of top-performing practice sites. Science.gov (United States) Feifer, Chris; Nemeth, Lynne; Nietert, Paul J; Wessell, Andrea M; Jenkins, Ruth G; Roylance, Loraine; Ornstein, Steven M 2007-01-01 Primary care practices use different approaches in their quest for high-quality care. Previous work in the Practice Partner Research Network (PPRNet) found that improved outcomes are associated with strategies to prioritize performance, involve staff, redesign elements of the delivery system, make patients active partners in guideline adherence, and use tools embedded in the electronic medical record. The aim of this study was to examine variations in the adoption of improvements among sites achieving the best outcomes. This study used an observational case study design. A practice-level measure of adherence to clinical guidelines was used to identify the highest performing practices in a network of internal and family medicine practices participating in a national demonstration project. We analyzed qualitative and quantitative information derived from project documents, field notes, and evaluation questionnaires to develop and compare case studies. Nine cases are described. All use many of the same improvement strategies. Differences in the way improvements are organized define 3 distinct archetypes: the Technophiles, the Motivated Team, and the Care Enterprise. There is no single approach that explains the superior performance of high-performing practices, though each has adopted variations of PPRNet's improvement model. Practices will vary in their path to high-quality care. The archetypes could prove to be a useful guide to other practices selecting an overall quality improvement approach. 7. Quality assurance requirements and methods for high level waste package acceptability International Nuclear Information System (INIS) 1992-12-01 This document should serve as guidance for assigning the necessary items to control the conditioning process in such a way that waste packages are produced in compliance with the waste acceptance requirements. It is also provided to promote the exchange of information on quality assurance requirements and on the application of quality assurance methods associated with the production of high level waste packages, to ensure that these waste packages comply with the requirements for transportation, interim storage and waste disposal in deep geological formations. The document is intended to assist both the operators of conditioning facilities and repositories as well as national authorities and regulatory bodies, involved in the licensing of the conditioning of high level radioactive wastes or in the development of deep underground disposal systems. The document recommends the quality assurance requirements and methods which are necessary to generate data for these parameters identified in IAEA-TECDOC-560 on qualitative acceptance criteria, and indicates where and when the control methods can be applied, e.g. in the operation or commissioning of a process or in the development of a waste package design. Emphasis is on the control of the process and little reliance is placed on non-destructive or destructive testing. Qualitative criteria, relevant to disposal of high level waste, are repository dependent and are not addressed here. 37 refs, 3 figs, 2 tabs 8. A Systematic Process for Developing High Quality SaaS Cloud Services Science.gov (United States) La, Hyun Jung; Kim, Soo Dong Software-as-a-Service (SaaS) is a type of cloud service which provides software functionality through Internet. Its benefits are well received in academia and industry. To fully utilize the benefits, there should be effective methodologies to support the development of SaaS services which provide high reusability and applicability. Conventional approaches such as object-oriented methods do not effectively support SaaS-specific engineering activities such as modeling common features, variability, and designing quality services. In this paper, we present a systematic process for developing high quality SaaS and highlight the essentiality of commonality and variability (C&V) modeling to maximize the reusability. We first define criteria for designing the process model and provide a theoretical foundation for SaaS; its meta-model and C&V model. We clarify the notion of commonality and variability in SaaS, and propose a SaaS development process which is accompanied with engineering instructions. Using the proposed process, SaaS services with high quality can be effectively developed. 9. Establishing High-Quality Prostate Brachytherapy Using a Phantom Simulator Training Program Energy Technology Data Exchange (ETDEWEB) 2014-11-01 Purpose: To design and implement a unique training program that uses a phantom-based simulator to teach the process of prostate brachytherapy (PB) quality assurance and improve the quality of education. Methods and Materials: Trainees in our simulator program were practicing radiation oncologists, radiation oncology residents, and fellows of the American Brachytherapy Society. The program emphasized 6 core areas of quality assurance: patient selection, simulation, treatment planning, implant technique, treatment evaluation, and outcome assessment. Using the Iodine 125 ({sup 125}I) preoperative treatment planning technique, trainees implanted their ultrasound phantoms with dummy seeds (ie, seeds with no activity). Pre- and postimplant dosimetric parameters were compared and correlated using regression analysis. Results: Thirty-one trainees successfully completed the simulator program during the period under study. The mean phantom prostate size, number of seeds used, and total activity were generally consistent between trainees. All trainees met the V100 >95% objective both before and after implantation. Regardless of the initial volume of the prostate phantom, trainees' ability to cover the target volume with at least 100% of the dose (V100) was not compromised (R=0.99 pre- and postimplant). However, the V150 had lower concordance (R=0.37) and may better reflect heterogeneity control of the implant process. Conclusions: Analysis of implants from this phantom-based simulator shows a high degree of consistency between trainees and uniformly high-quality implants with respect to parameters used in clinical practice. This training program provides a valuable educational opportunity that improves the quality of PB training and likely accelerates the learning curve inherent in PB. Prostate phantom implantation can be a valuable first step in the acquisition of the required skills to safely perform PB. 10. Establishing High-Quality Prostate Brachytherapy Using a Phantom Simulator Training Program International Nuclear Information System (INIS) Thaker, Nikhil G.; Kudchadker, Rajat J.; Swanson, David A.; Albert, Jeffrey M.; Mahmood, Usama; Pugh, Thomas J.; Boehling, Nicholas S.; Bruno, Teresa L.; Prestidge, Bradley R.; Crook, Juanita M.; Cox, Brett W.; Potters, Louis; Moran, Brian J.; Keyes, Mira; Kuban, Deborah A.; Frank, Steven J. 2014-01-01 Purpose: To design and implement a unique training program that uses a phantom-based simulator to teach the process of prostate brachytherapy (PB) quality assurance and improve the quality of education. Methods and Materials: Trainees in our simulator program were practicing radiation oncologists, radiation oncology residents, and fellows of the American Brachytherapy Society. The program emphasized 6 core areas of quality assurance: patient selection, simulation, treatment planning, implant technique, treatment evaluation, and outcome assessment. Using the Iodine 125 ( 125 I) preoperative treatment planning technique, trainees implanted their ultrasound phantoms with dummy seeds (ie, seeds with no activity). Pre- and postimplant dosimetric parameters were compared and correlated using regression analysis. Results: Thirty-one trainees successfully completed the simulator program during the period under study. The mean phantom prostate size, number of seeds used, and total activity were generally consistent between trainees. All trainees met the V100 >95% objective both before and after implantation. Regardless of the initial volume of the prostate phantom, trainees' ability to cover the target volume with at least 100% of the dose (V100) was not compromised (R=0.99 pre- and postimplant). However, the V150 had lower concordance (R=0.37) and may better reflect heterogeneity control of the implant process. Conclusions: Analysis of implants from this phantom-based simulator shows a high degree of consistency between trainees and uniformly high-quality implants with respect to parameters used in clinical practice. This training program provides a valuable educational opportunity that improves the quality of PB training and likely accelerates the learning curve inherent in PB. Prostate phantom implantation can be a valuable first step in the acquisition of the required skills to safely perform PB 11. Adoption of high technology medical imaging and hospital quality and efficiency: Towards a conceptual framework. Science.gov (United States) Sandoval, Guillermo A; Brown, Adalsteinn D; Wodchis, Walter P; Anderson, Geoffrey M 2018-05-17 12. A guide to calculating habitat-quality metrics to inform conservation of highly mobile species Science.gov (United States) Bieri, Joanna A.; Sample, Christine; Thogmartin, Wayne E.; Diffendorfer, James E.; Earl, Julia E.; Erickson, Richard A.; Federico, Paula; Flockhart, D. T. Tyler; Nicol, Sam; Semmens, Darius J.; Skraber, T.; Wiederholt, Ruscena; Mattsson, Brady J. 2018-01-01 Many metrics exist for quantifying the relative value of habitats and pathways used by highly mobile species. Properly selecting and applying such metrics requires substantial background in mathematics and understanding the relevant management arena. To address this multidimensional challenge, we demonstrate and compare three measurements of habitat quality: graph-, occupancy-, and demographic-based metrics. Each metric provides insights into system dynamics, at the expense of increasing amounts and complexity of data and models. Our descriptions and comparisons of diverse habitat-quality metrics provide means for practitioners to overcome the modeling challenges associated with management or conservation of such highly mobile species. Whereas previous guidance for applying habitat-quality metrics has been scattered in diversified tracks of literature, we have brought this information together into an approachable format including accessible descriptions and a modeling case study for a typical example that conservation professionals can adapt for their own decision contexts and focal populations.Considerations for Resource ManagersManagement objectives, proposed actions, data availability and quality, and model assumptions are all relevant considerations when applying and interpreting habitat-quality metrics.Graph-based metrics answer questions related to habitat centrality and connectivity, are suitable for populations with any movement pattern, quantify basic spatial and temporal patterns of occupancy and movement, and require the least data.Occupancy-based metrics answer questions about likelihood of persistence or colonization, are suitable for populations that undergo localized extinctions, quantify spatial and temporal patterns of occupancy and movement, and require a moderate amount of data.Demographic-based metrics answer questions about relative or absolute population size, are suitable for populations with any movement pattern, quantify demographic 13. Exploratory subsetting of autism families based on savant skills improves evidence of genetic linkage to 15q11-q13. Science.gov (United States) Nurmi, Erika L; Dowd, Michael; Tadevosyan-Leyfer, Ovsanna; Haines, Jonathan L; Folstein, Susan E; Sutcliffe, James S 2003-07-01 Autism displays a remarkably high heritability but a complex genetic etiology. One approach to identifying susceptibility loci under these conditions is to define more homogeneous subsets of families on the basis of genetically relevant phenotypic or biological characteristics that vary from case to case. The authors performed a principal components analysis, using items from the Autism Diagnostic Interview, which resulted in six clusters of variables, five of which showed significant sib-sib correlation. The utility of these phenotypic subsets was tested in an exploratory genetic analysis of the autism candidate region on chromosome 15q11-q13. When the Collaborative Linkage Study of Autism sample was divided, on the basis of mean proband score for the "savant skills" cluster, the heterogeneity logarithm of the odds under a recessive model at D15S511, within the GABRB3 gene, increased from 0.6 to 2.6 in the subset of families in which probands had greater savant skills. These data are consistent with the genetic contribution of a 15q locus to autism susceptibility in a subset of affected individuals exhibiting savant skills. Similar types of skills have been noted in individuals with Prader-Willi syndrome, which results from deletions of this chromosomal region. 14. Molecular subsets in the gene expression signatures of scleroderma skin. Directory of Open Access Journals (Sweden) Ausra Milano 2008-07-01 Full Text Available Scleroderma is a clinically heterogeneous disease with a complex phenotype. The disease is characterized by vascular dysfunction, tissue fibrosis, internal organ dysfunction, and immune dysfunction resulting in autoantibody production.We analyzed the genome-wide patterns of gene expression with DNA microarrays in skin biopsies from distinct scleroderma subsets including 17 patients with systemic sclerosis (SSc with diffuse scleroderma (dSSc, 7 patients with SSc with limited scleroderma (lSSc, 3 patients with morphea, and 6 healthy controls. 61 skin biopsies were analyzed in a total of 75 microarray hybridizations. Analysis by hierarchical clustering demonstrates nearly identical patterns of gene expression in 17 out of 22 of the forearm and back skin pairs of SSc patients. Using this property of the gene expression, we selected a set of 'intrinsic' genes and analyzed the inherent data-driven groupings. Distinct patterns of gene expression separate patients with dSSc from those with lSSc and both are easily distinguished from normal controls. Our data show three distinct patient groups among the patients with dSSc and two groups among patients with lSSc. Each group can be distinguished by unique gene expression signatures indicative of proliferating cells, immune infiltrates and a fibrotic program. The intrinsic groups are statistically significant (p<0.001 and each has been mapped to clinical covariates of modified Rodnan skin score, interstitial lung disease, gastrointestinal involvement, digital ulcers, Raynaud's phenomenon and disease duration. We report a 177-gene signature that is associated with severity of skin disease in dSSc.Genome-wide gene expression profiling of skin biopsies demonstrates that the heterogeneity in scleroderma can be measured quantitatively with DNA microarrays. The diversity in gene expression demonstrates multiple distinct gene expression programs in the skin of patients with scleroderma. 15. Microstructure, microbial profile and quality characteristics of high-pressure-treated chicken nuggets. Science.gov (United States) Devatkal, Suresh; Anurag, Rahul; Jaganath, Bindu; Rao, Srinivasa 2015-10-01 High-pressure processing (300 MPa for 5 min) as a non-thermal post-processing intervention was employed to improve the shelf life and qualities of cooked refrigerated chicken nuggets. Pomegranate peel extract (1%) was also used as a source of natural antioxidant and antimicrobial in chicken nuggets. Microstructure, microbial profile, instrumental colour, texture profile and lipid oxidation were evaluated. High-pressure treatment and pomegranate peel extract did not influence significantly the colour and textural properties of cooked chicken nuggets. Thiobarbituric acid reactive substance values significantly (p chicken nuggets were the major spoilage bacteria. © The Author(s) 2014. 16. Life quality and living standards in big cities under conditions of high-rise construction development Science.gov (United States) Avdeeva, Elena; Averina, Tatiana; Kochetova, Larisa 2018-03-01 Modern urbanization processes occurring on a global scale inevitably lead to an increase in population density in large cities. People assess the state of life quality and living standards of megalopolises under conditions of high-rise construction development ambiguously. Using SWOT analysis, the authors distinguished positive and negative aspects of high-rise construction, highlighted threats to its development and its opportunities. The article considers the model of development of the city's industry and infrastructure, which enables determining the optimal volume of production by sectors and branches of city economy in order to increase its innovative, production and economic potential and business activity. 17. Fast Synthesis of High Quality Biodiesel from ‘Waste Fish Oil’ by Single Step Transesterification Directory of Open Access Journals (Sweden) Yogesh C. Sharma 2014-09-01 Full Text Available A large volume of fish wastes is produced on a daily basis in the Indian sub-continent. This abundant waste source could serve as an economic feedstock for bioenergy generation. In the present study, oil extracted from discarded fish parts was used for high quality biodiesel production. More specifically, a single step transesterification of ‘waste fishoil’ with methanol using sodium methoxide (CH3ONa as homogeneous catalyst under moderate operational conditions resulted in highly pure biodiesel of > 98% of fatty acid methyl ester (FAME content. Characterization was performed by Fourier Transform-Nuclear Magnetic Resonance (FT-NMR. 18. Ultra-High-Resolution Computed Tomography of the Lung: Image Quality of a Prototype Scanner OpenAIRE Kakinuma, Ryutaro; Moriyama, Noriyuki; Muramatsu, Yukio; Gomi, Shiho; Suzuki, Masahiro; Nagasawa, Hirobumi; Kusumoto, Masahiko; Aso, Tomohiko; Muramatsu, Yoshihisa; Tsuchida, Takaaki; Tsuta, Koji; Maeshima, Akiko Miyagi; Tochigi, Naobumi; Watanabe, Shun-ichi; Sugihara, Naoki 2015-01-01 Purpose: The image noise and image quality of a prototype ultra-high-resolution computed tomography (U-HRCT) scanner was evaluated and compared with those of conventional high-resolution CT (C-HRCT) scanners. Materials and Methods: This study was approved by the institutional review board. A U-HRCT scanner prototype with 0.25 mm × 4 rows and operating at 120 mAs was used. The C-HRCT images were obtained using a 0.5 mm × 16 or 0.5 mm × 64 detector-row CT scanner operating at 150 mAs. Images fr... 19. Low index contrast heterostructure photonic crystal cavities with high quality factors and vertical radiation coupling Science.gov (United States) Ge, Xiaochen; Minkov, Momchil; Fan, Shanhui; Li, Xiuling; Zhou, Weidong 2018-04-01 We report here design and experimental demonstration of heterostructure photonic crystal cavities resonating near the Γ point with simultaneous strong lateral confinement and highly directional vertical radiation patterns. The lateral confinement is provided by a mode gap originating from a gradual modulation of the hole radii. High quality factor resonance is realized with a low index contrast between silicon nitride and quartz. The near surface-normal directional emission is preserved when the size of the core region is scaled down. The influence of the cavity size parameters on the resonant modes is also investigated theoretically and experimentally. 20. Random-subset fitting of digital holograms for fast three-dimensional particle tracking [invited]. Science.gov (United States) Dimiduk, Thomas G; Perry, Rebecca W; Fung, Jerome; Manoharan, Vinothan N 2014-09-20 Fitting scattering solutions to time series of digital holograms is a precise way to measure three-dimensional dynamics of microscale objects such as colloidal particles. However, this inverse-problem approach is computationally expensive. We show that the computational time can be reduced by an order of magnitude or more by fitting to a random subset of the pixels in a hologram. We demonstrate our algorithm on experimentally measured holograms of micrometer-scale colloidal particles, and we show that 20-fold increases in speed, relative to fitting full frames, can be attained while introducing errors in the particle positions of 10 nm or less. The method is straightforward to implement and works for any scattering model. It also enables a parallelization strategy wherein random-subset fitting is used to quickly determine initial guesses that are subsequently used to fit full frames in parallel. This approach may prove particularly useful for studying rare events, such as nucleation, that can only be captured with high frame rates over long times. 1. Lymphocyte subsets are influenced by positivity levels in healthy subjects before and after mild acute stress. Science.gov (United States) Caprara, Gian Vittorio; Nisini, Roberto; Castellani, Valeria; Vittorio, Pasquali; Alessandri, Guido; Vincenzo, Ziparo; Claudia, Ferlito; Valentina, Germano; Andrea, Picchianti Diamanti; Biondo, Michela Ileen; Milanetti, Francesca; Salerno, Gerardo; Vincenzo, Visco; Mario, Pietrosanti; Aniballi, Eros; Simonetta, Salemi; Angela, Santoni; D'Amelio, Raffaele 2017-08-01 In the current study, the possible association of positivity (POS), recently defined as general disposition to view life under positive outlook, with immune markers and post-stress modifications, was analyzed. Circulating lymphocyte subsets and serum cytokine levels were evaluated before and after a standard mild acute stress test, in 41 healthy students, previously selected by a questionnaire for their level of POS (high [POS-H] and low [POS-L]). The CD3 + and CD4 + cell frequency was higher in the POS-H students before and after acute stress. CD4 + subpopulation analysis revealed baseline higher terminally differentiated frequency in the POS-H, whereas higher effector memory frequency was present in the POS-L students. Moreover, the frequency of post-stress B cells was higher in the POS-H students. The mild-stress test was associated to an increase of the IL-10 mean values, while mean values of the other cytokines tested did not change significantly. It is tempting to speculate that IL-10 may work as biomarker of response to acute mild stress and that POS-H may be associated to a better capacity of the immune system to contrast the disturbing effects of mild acute stress. Yet further studies on lymphocyte subset absolute number and function of larger and different populations are needed to definitively prove these preliminary observations. Copyright © 2017 European Federation of Immunological Societies. Published by Elsevier B.V. All rights reserved. 2. Comparison of high pressure and high temperature short time processing on quality of carambola juice during cold storage. Science.gov (United States) Huang, Hsiao-Wen; Chen, Bang-Yuan; Wang, Chung-Yi 2018-05-01 This study validated high hydrostatic pressure processing (HPP) for achieving greater than 5-log reductions of Escherichia coli O157:H7 in carambola juice and determined shelf life of processed juice stored at 4 °C. Carambola juice processed at 600 MPa for 150 s was identified capable of achieving greater than 5.15-log reductions of E. coli O157:H7, and the quality was compared with that of high temperature short time (HTST)-pasteurized juice at 110 °C for 8.6 s. Aerobic, psychrotrophic, E. coli /coliform, and yeasts and moulds in the juice were reduced by HPP or HTST to levels below the minimum detection limit (HTST juices. However, HTST treatment significantly changed the color of juice, while no significant difference was observed between the control and HPP samples. HPP and HTST treatments reduced the total soluble solids in the juice, but maintained higher sucrose, glucose, fructose, and total sugar contents than untreated juice. The total phenolic and ascorbic acid contents were higher in juice treated with HPP than untreated and HTST juice, but there was no significant difference in the flavonoid content. Aroma score analysis showed that HPP had no effect on aroma, maintaining the highest score during cold storage. The results of this study suggest that appropriate HPP conditions can achieve the same microbial safety as HTST, while maintaining the quality and extending the shelf life of carambola juice. 3. A high-throughput system for high-quality tomographic reconstruction of large datasets at Diamond Light Source. Science.gov (United States) Atwood, Robert C; Bodey, Andrew J; Price, Stephen W T; Basham, Mark; Drakopoulos, Michael 2015-06-13 Tomographic datasets collected at synchrotrons are becoming very large and complex, and, therefore, need to be managed efficiently. Raw images may have high pixel counts, and each pixel can be multidimensional and associated with additional data such as those derived from spectroscopy. In time-resolved studies, hundreds of tomographic datasets can be collected in sequence, yielding terabytes of data. Users of tomographic beamlines are drawn from various scientific disciplines, and many are keen to use tomographic reconstruction software that does not require a deep understanding of reconstruction principles. We have developed Savu, a reconstruction pipeline that enables users to rapidly reconstruct data to consistently create high-quality results. Savu is designed to work in an 'orthogonal' fashion, meaning that data can be converted between projection and sinogram space throughout the processing workflow as required. The Savu pipeline is modular and allows processing strategies to be optimized for users' purposes. In addition to the reconstruction algorithms themselves, it can include modules for identification of experimental problems, artefact correction, general image processing and data quality assessment. Savu is open source, open licensed and 'facility-independent': it can run on standard cluster infrastructure at any institution. 4. Towards a high quality high school workforce: A longitudinal, demographic analysis of U.S. public school physics teachers Directory of Open Access Journals (Sweden) Gregory T. Rushton 2017-10-01 Full Text Available Efforts to improve the number and quality of the high school physics teaching workforce have taken several forms, including those sponsored by professional organizations. Using a series of large-scale teacher demographic data sets from the National Center for Education Statistics (NCES, this study sought to investigate trends in teacher quality at the national level in the two and a half decades between 1987 and 2012. Specifically, we investigated (i details about the degree backgrounds, main teaching assignments, and experience levels of those assigned to teach physics; (ii whether the proportion of those with certifications in physics as a fraction of the entire physics teaching workforce had changed; and (iii if workforce diversity (with respect to race and gender had changed over time. Our data indicate that trends in these domains have generally been positive, but still fall short of having a highly qualified physics teacher in each classroom. Additionally, the population of physics teachers has more novices and fewer veterans than it did 10 years ago, although veteran physics teachers are not as rare as those in other branches of high school STEM fields. We also analyzed trends in physics teacher race and gender diversity and found them to lag behind other STEM and non-STEM teacher communities. High school physics is still mostly taught by white males with backgrounds from outside of physics. Implications for future policy decisions at the local and national levels are discussed, including attending to the specific needs of degree-holding and non-degree-holding physics teachers separately and localizing teacher recruitment and preparation efforts in regional centers. 5. Towards a high quality high school workforce: A longitudinal, demographic analysis of U.S. public school physics teachers Science.gov (United States) Rushton, Gregory T.; Rosengrant, David; Dewar, Andrew; Shah, Lisa; Ray, Herman E.; Sheppard, Keith; Watanabe, Lynn 2017-12-01 Efforts to improve the number and quality of the high school physics teaching workforce have taken several forms, including those sponsored by professional organizations. Using a series of large-scale teacher demographic data sets from the National Center for Education Statistics (NCES), this study sought to investigate trends in teacher quality at the national level in the two and a half decades between 1987 and 2012. Specifically, we investigated (i) details about the degree backgrounds, main teaching assignments, and experience levels of those assigned to teach physics; (ii) whether the proportion of those with certifications in physics as a fraction of the entire physics teaching workforce had changed; and (iii) if workforce diversity (with respect to race and gender) had changed over time. Our data indicate that trends in these domains have generally been positive, but still fall short of having a highly qualified physics teacher in each classroom. Additionally, the population of physics teachers has more novices and fewer veterans than it did 10 years ago, although veteran physics teachers are not as rare as those in other branches of high school STEM fields. We also analyzed trends in physics teacher race and gender diversity and found them to lag behind other STEM and non-STEM teacher communities. High school physics is still mostly taught by white males with backgrounds from outside of physics. Implications for future policy decisions at the local and national levels are discussed, including attending to the specific needs of degree-holding and non-degree-holding physics teachers separately and localizing teacher recruitment and preparation efforts in regional centers. 6. Practical considerations for obtaining high quality quantitative computed tomography data of the skeletal system. Science.gov (United States) Troy, Karen L; Edwards, W Brent 2018-05-01 Quantitative CT (QCT) analysis involves the calculation of specific parameters such as bone volume and density from CT image data, and can be a powerful tool for understanding bone quality and quantity. However, without careful attention to detail during all steps of the acquisition and analysis process, data can be of poor- to unusable-quality. Good quality QCT for research requires meticulous attention to detail and standardization of all aspects of data collection and analysis to a degree that is uncommon in a clinical setting. Here, we review the literature to summarize practical and technical considerations for obtaining high quality QCT data, and provide examples of how each recommendation affects calculated variables. We also provide an overview of the QCT analysis technique to illustrate additional opportunities to improve data reproducibility and reliability. Key recommendations include: standardizing the scanner and data acquisition settings, minimizing image artifacts, selecting an appropriate reconstruction algorithm, and maximizing repeatability and objectivity during QCT analysis. The goal of the recommendations is to reduce potential sources of error throughout the analysis, from scan acquisition to the interpretation of results. Copyright © 2018 Elsevier Inc. All rights reserved. 7. How to Achieve High-Quality Oocytes? The Key Role of Myo-Inositol and Melatonin Directory of Open Access Journals (Sweden) Salvatore Giovanni Vitale 2016-01-01 Full Text Available Assisted reproductive technologies (ART have experienced growing interest from infertile patients seeking to become pregnant. The quality of oocytes plays a pivotal role in determining ART outcomes. Although many authors have studied how supplementation therapy may affect this important parameter for both in vivo and in vitro models, data are not yet robust enough to support firm conclusions. Regarding this last point, in this review our objective has been to evaluate the state of the art regarding supplementation with melatonin and myo-inositol in order to improve oocyte quality during ART. On the one hand, the antioxidant effect of melatonin is well known as being useful during ovulation and oocyte incubation, two occasions with a high level of oxidative stress. On the other hand, myo-inositol is important in cellular structure and in cellular signaling pathways. Our analysis suggests that the use of these two molecules may significantly improve the quality of oocytes and the quality of embryos: melatonin seems to raise the fertilization rate, and myo-inositol improves the pregnancy rate, although all published studies do not fully agree with these conclusions. However, previous studies have demonstrated that cotreatment improves these results compared with melatonin alone or myo-inositol alone. We recommend that further studies be performed in order to confirm these positive outcomes in routine ART treatment. 8. Evaluation of Hole Quality in Hardened Steel with High-Speed Drilling Using Different Cooling Systems Directory of Open Access Journals (Sweden) Lincoln Cardoso Brandão 2011-01-01 Full Text Available This work evaluates the hole quality on AISI H13 hardened steel using high-speed drilling. Specimens were machined with new and worn out drills with 8.6 mm diameter and (TiAlN coating. Two levels of cutting speed and three levels of cooling/lubrication systems (flooded, minimum lubrication quantity, and dry were used. The hole quality is evaluated on surface roughness (Ra parameter, diameter error, circularity, and cylindricity error. A statistical analysis of the results shows that the cooling/lubrication system significantly affects the hole quality for all measured variables. This analysis indicates that dry machining produces the worst results. Higher cutting speeds not only prove beneficial to diameter error and circularity errors, but also show no significant difference on surface roughness and cylindricity errors. The effects of the interaction between the cooling/lubrication systems, tool wear, and cutting speed indicate that only cylindricity error is influenced. Thus, the conclusion is that the best hole quality is produced with a higher cutting speed using flooded or minimum lubrication quantity independent of drill wear. 9. High perfomance liquid chromatography fingerprint analysis for quality control of brotowali (Tinospora crispa) Science.gov (United States) Syarifah, V. B.; Rafi, M.; Wahyuni, W. T. 2017-05-01 Brotowali (Tinospora crispa) is widely used in Indonesia as ingredient of herbal medicine formulation. To ensure the quality, safety, and efficacy of herbal medicine products, its chemical constituents should be continuously evaluated. High performance liquid chromatography (HPLC) fingerprint is one of powerful technique for this quality control process. In this study, HPLC fingerprint analysis method was developed for quality control of brotowali. HPLC analysis was performed in C18 column and detection was performed using photodiode array detector. The optimum mobile phase for brotowali fingerprint was acetonitrile (ACN) and 0.1% formic acid in gradient elution mode at a flow rate of 1 mL/min. The number of peaks detected in HPLC fingerprint of brotowali was 32 peaks and 23 peaks for stems and leaves, respectively. Berberine as marker compound was detected at retention time of 20.525 minutes. Evaluation of analytical performance including precision, reproducibility, and stability prove that this HPLC fingerprint analysis was reliable and could be applied for quality control of brotowali. 10. Applying Machine Learning and High Performance Computing to Water Quality Assessment and Prediction Directory of Open Access Journals (Sweden) Ruijian Zhang 2017-12-01 Full Text Available Water quality assessment and prediction is a more and more important issue. Traditional ways either take lots of time or they can only do assessments. In this research, by applying machine learning algorithm to a long period time of water attributes’ data; we can generate a decision tree so that it can predict the future day’s water quality in an easy and efficient way. The idea is to combine the traditional ways and the computer algorithms together. Using machine learning algorithms, the assessment of water quality will be far more efficient, and by generating the decision tree, the prediction will be quite accurate. The drawback of the machine learning modeling is that the execution takes quite long time, especially when we employ a better accuracy but more time-consuming algorithm in clustering. Therefore, we applied the high performance computing (HPC System to deal with this problem. Up to now, the pilot experiments have achieved very promising preliminary results. The visualized water quality assessment and prediction obtained from this project would be published in an interactive website so that the public and the environmental managers could use the information for their decision making. 11. Quality properties of pre- and post-rigor beef muscle after interventions with high frequency ultrasound. Science.gov (United States) Sikes, Anita L; Mawson, Raymond; Stark, Janet; Warner, Robyn 2014-11-01 12. Human Innate Lymphoid Cell Subsets Possess Tissue-Type Based Heterogeneity in Phenotype and Frequency DEFF Research Database (Denmark) Simoni, Yannick; Fehlings, Michael; Kloverpris, Henrik N. 2017-01-01 Animal models have highlighted the importance of innate lymphoid cells (ILCs) in multiple immune responses. However, technical limitations have hampered adequate characterization of ILCs in humans. Here, we used mass cytometry including a broad range of surface markers and transcription factors...... to accurately identify and profile ILCs across healthy and inflamed tissue types. High dimensional analysis allowed for clear phenotypic delineation of ILC2 and ILC3 subsets. We were not able to detect ILC1 cells in any of the tissues assessed, however, we identified intra-epithelial (ie)ILC1-like cells...... that represent a broader category of NK cells in mucosal and non-mucosal pathological tissues. In addition, we have revealed the expression of phenotypic molecules that have not been previously described for ILCs. Our analysis shows that human ILCs are highly heterogeneous cell types between individuals... 13. 75 FR 70289 - Certain Coated Paper Suitable For High-Quality Print Graphics Using Sheet-Fed Presses From China... Science.gov (United States) 2010-11-17 ...)] Certain Coated Paper Suitable For High-Quality Print Graphics Using Sheet-Fed Presses From China and... paper suitable for high-quality print graphics using sheet-fed presses (certain coated paper'') from... paper industry is materially injured by reason of imports of the subject merchandise from China and... 14. Bridging existing governance gaps: five evidence-based actions that boards can take to pursue high quality care. Science.gov (United States) Leggat, Sandra G; Balding, Cathy 2017-11-13 Objective To explore the impact of the organisational quality systems on quality of care in Victorian health services. Methods During 2015 a total of 55 focus groups were conducted with more than 350 managers, clinical staff and board members in eight Victorian health services to explore the effectiveness of health service quality systems. A review of the quality and safety goals and strategies outlined in the strategic and operating plans of the participating health services was also undertaken. Results This paper focuses on the data related to the leadership role of health service boards in ensuring safe, high-quality care. The findings suggest that health service boards are not fully meeting their governance accountability to ensure consistently high-quality care. The data uncovered major clinical governance gaps between stated board and executive aspirations for quality and safety and the implementation of these expectations at point of care. These gaps were further compounded by quality system confusion, over-reliance on compliance, and inadequate staff engagement. Conclusion Based on the existing evidence we propose five specific actions boards can take to close the gaps, thereby supporting improved care for all consumers. What is known about this topic? Effective governance is essential for high-quality healthcare delivery. Boards are required to play an active role in their organisation's pursuit of high quality care. What does this paper add? Recent government reports suggest that Australian health service boards are not fully meeting their governance requirements for high quality, safe care delivery, and our research pinpoints key governance gaps. What are the implications for practitioners? Based on our research findings we outline five evidence-based actions for boards to improve their governance of quality care delivery. These actions focus on an organisational strategy for high-quality care, with the chief executive officer held accountable for 15. Prodigious Effects of Concentration Intensification on Nanoparticle Synthesis: A High-Quality, Scalable Approach KAUST Repository Williamson, Curtis B. 2015-12-23 © 2015 American Chemical Society. Realizing the promise of nanoparticle-based technologies demands more efficient, robust synthesis methods (i.e., process intensification) that consistently produce large quantities of high-quality nanoparticles (NPs). We explored NP synthesis via the heat-up method in a regime of previously unexplored high concentrations near the solubility limit of the precursors. We discovered that in this highly concentrated and viscous regime the NP synthesis parameters are less sensitive to experimental variability and thereby provide a robust, scalable, and size-focusing NP synthesis. Specifically, we synthesize high-quality metal sulfide NPs (<7% relative standard deviation for Cu2-xS and CdS), and demonstrate a 10-1000-fold increase in Cu2-xS NP production (>200 g) relative to the current field of large-scale (0.1-5 g yields) and laboratory-scale (<0.1 g) efforts. Compared to conventional synthesis methods (hot injection with dilute precursor concentration) characterized by rapid growth and low yield, our highly concentrated NP system supplies remarkably controlled growth rates and a 10-fold increase in NP volumetric production capacity (86 g/L). The controlled growth, high yield, and robust nature of highly concentrated solutions can facilitate large-scale nanomanufacturing of NPs by relaxing the synthesis requirements to achieve monodisperse products. Mechanistically, our investigation of the thermal and rheological properties and growth rates reveals that this high concentration regime has reduced mass diffusion (a 5-fold increase in solution viscosity), is stable to thermal perturbations (64% increase in heat capacity), and is resistant to Ostwald ripening. 16. The perspective awareness model - Eliciting multiple perspectives to formulate high quality decisions International Nuclear Information System (INIS) Boucher, Laurel 2013-01-01 A great deal of attention is given to the importance of communication in environmental remediation and radioactive waste management. However, very little attention is given to eliciting multiple perspectives so as to formulate high quality decisions. Plans that are based on a limited number of perspectives tend to be narrowly focused whereas those that are based on a wide variety of perspectives tend to be comprehensive, higher quality, and more apt to be put into application. In addition, existing methods of dialogue have built-in limitations in that they typically draw from the predominant thinking patterns which focus in some areas but ignore others. This can result in clarity but a lack of comprehensiveness. This paper presents a Perspective Awareness Model which helps groups such as partnering teams, interagency teams, steering committees, and working groups elicit a wide net of perspectives and viewpoints. The paper begins by describing five factors that makes cooperation among such groups challenging. Next, a Perspective Awareness Model that makes it possible to manage these five factors is presented. The two primary components of this model --- the eight 'Thinking Directions' and the 'Shared Documentation' --- are described in detail. Several examples are given to illustrate how the Perspective Awareness Model can be used to elicit multiple perspectives to formulate high quality decisions in the area of environmental remediation and radioactive waste management. (authors) 17. Development of a biotechnological process for the production of high quality linen fibers. Science.gov (United States) Valladares Juárez, Ana Gabriela; Rost, Gernot; Heitmann, Uwe; Heger, Egon; Müller, Rudolf 2011-10-01 A novel biotechnological process for the production of high-quality flax fibers was developed. In this process, decorticated fibers from green flax were washed with 0.5% soda solution and treated with the pectinolytic strain Geobacillus thermoglucosidasius PB94A. Before drying the fibers, they were treated with the textile softener Adulcinol BUN. If the fibers contained contaminant shives, a bleaching step with hydrogen peroxide was performed before the softener treatment. In experiments where fibers were treated by the new process, and in which the bacterial solutions were reused seven times, the fiber quality was similar in all batches. The resolution of the treated fibers was 2.7 ± 0.4 and the fineness was 11.1 ± 1.1 dtex, while the starting material had a resolution of 7.3 and a fineness of 37 dtex. The new biotechnological treatment eliminates the weather-associated risks of the traditional fiber retting completely and produces consistently high-quality fibers that can be used to produce fine linen yarns. 18. High quality voice synthesis middle ware; Kohinshitsu onsei gosei middle war Energy Technology Data Exchange (ETDEWEB) NONE 2000-03-01 Toshiba Corp. newly developed a natural voice synthesis system, TOS Drive TTS (TOtally speaker Driven Text-To-Speech) system, in which natural high-quality read-aloud is greatly improved, and also developed as its application a voice synthesis middle ware. In the newly developed system, using as a model a narrator's voice recorded preliminarily, a metrical control dictionary is automatically learned that reproduces the characteristics of metrical patters such as intonation or rhythm of a human voice, as is a voice bases dictionary that reproduces the characteristics of a voice quality, enabling natural voice synthesis to be realized that picks up human voice characteristics. The system is high quality and also very compact, while the voice synthesis middle ware utilizing this technology is adaptable to various platforms such as MPU or OS. The system is very suitable for audio response in the ITS field having car navigation systems as the core; besides, expanded application is expected to an audio response system that used to employ a sound recording and reproducing system. (translated by NEDO) 19. CBE growth of high-quality ZnO epitaxial layers Energy Technology Data Exchange (ETDEWEB) El-Shaer, A.; Bakin, A.; Mofor, A.C.; Kreye, M.; Waag, A. [Institute of Semiconductor Technology, Technical University Braunschweig, Hans-Sommer-Strasse 66, 38106 Braunschweig (Germany); Blaesing, J.; Krost, A. [Institute of Experimental Physics, Otto-von-Guericke-University Magdeburg (Germany); Stoimenos, J. [Physics Department, Aristotele University, Univ. Campus, 54006 Thessaloniki (Greece); Pecz, B. [Research Institute for Technical Physics and Materials Science, Hungarian Academy of Sciences, P.O. Box 49, 1525 Budapest (Hungary); Heuken, M. [Aixtron AG, Kackertstr. 15-17, 52072 Aachen (Germany) 2006-03-15 Further improvements on the recently reported novel approach to zinc oxide Chemical Beam Epitaxy (CBE) are presented. Hydrogen peroxide is employed as a very efficient novel oxidant. ZnO layers with a thickness from 100 nm to 600 nm were grown on c-sapphire using a MgO buffer. PL-mapping as well as conductivity mapping shows a good uniformity across the 2 inch ZnO-on-sapphire epiwafers. The measured surface roughness for the best layers is as low as 0.26 nm. HRXRD measurements of the obtained ZnO layers show excellent quality of the single crystalline ZnO. The FWHM of the HRXRD (0002) rocking curves measured for the 2 inch ZnO-on-sapphire wafers is as low as 27 arcsec with a very high lateral homogeneity across the whole wafer. Plane view HRTEM observations reveal the very good quality of the ZnO films. The results indicate that CBE is a suitable technique to fabricate ZnO of very high structural quality, which can eventually be used as an alternative to bulk ZnO substrates. (copyright 2006 WILEY-VCH Verlag GmbH and Co. KGaA, Weinheim) (orig.) 20. Effect of osmotic dehydration and vacuum-frying parameters to produce high-quality mango chips. Science.gov (United States) Nunes, Yolanda; Moreira, Rosana G 2009-09-01 Mango (Mangifera indica L.) is a fruit rich in flavor and nutritional values, which is an excellent candidate for producing chips. The objective of this study was to develop high-quality mango chips using vacuum frying. Mango ("Tommy Atkins") slices were pretreated with different maltodextrin concentrations (40, 50, and 65, w/v), osmotic dehydration times (45, 60, and 70 min), and solution temperatures (22 and 40 degrees C). Pretreated slices were vacuum fried at 120, 130, and 138 degrees C and product quality attributes (oil content, texture, color, carotenoid content) determined. The effect of frying temperatures at optimum osmotic dehydration times (65 [w/v] at 40 degrees C) was assessed. All samples were acceptable (scores > 5) to consumer panelists. The best mango chips were those pretreated with 65 (w/v) concentration for 60 min and vacuum fried at 120 degrees C. Mango chips under atmospheric frying had less carotenoid retention (32%) than those under vacuum frying (up to 65%). These results may help further optimize vacuum-frying processing of high-quality fruit-based snacks. 1. The production of the heavy pig for high quality processed products Directory of Open Access Journals (Sweden) Paolo Bosi 2010-01-01 Full Text Available The review discusses the topics presented below, with experimental data and more evidence for some aspects which areproblematic for the pig producer. To obtain high quality processed pig products, heavy pig production in Italy is subjectedto rules fixed by several Consortia, on the basis of the experience of producers and the results from scientific research.Slaughter animals of at least 9 months of age and 160 kg live weight are required to provide heavy cuts with excellentmeat. Breeds are limited to Italian Large White and Landrace, selected for specific parameters such as loss at 1st saltingof the ham, and their crosses. Crosses with Duroc and hybrids can be used if obtained from selection and crossingschemes that fit the objectives of Italian selection. Backfat thickness must be “sufficient” to obtain retailed fresh hamswith fat cover ranging from 20 to 30 mm, depending on ham weight, and the content of linoleic acid in ham fat covercannot exceed 15%. Fat quantity and quality in feeds should be carefully controlled to satisfy the quality of fat in theham. The feeding level is also relevant, due to the imposed minimum age at slaughter. Feed restriction is a necessarypractice to respect the rules for ham production. For castrate males of medium genetic value, the restriction shouldreduce fat deposition, but the diet composition can be adapted to maintain the maximum growth of protein. For strainswith high lean deposition potential and particularly for gilts, protein deposition should also be limited to maintain a sufficientfat cover of ham. 2. Formation of high-quality self-assembled monolayers of conjugated dithiols on gold: base matters. Science.gov (United States) Valkenier, Hennie; Huisman, Everardus H; van Hal, Paul A; de Leeuw, Dago M; Chiechi, Ryan C; Hummelen, Jan C 2011-04-06 This Article reports a systematic study on the formation of self-assembled monolayers (SAMs) of conjugated molecules for molecular electronic (ME) devices. We monitored the deprotection reaction of acetyl protected dithiols of oligophenylene ethynylenes (OPEs) in solution using two different bases and studied the quality of the resulting SAMs on gold. We found that the optimal conditions to reproducibly form dense, high-quality monolayers are 9-15% triethylamine (Et(3)N) in THF. The deprotection base tetrabutylammonium hydroxide (Bu(4)NOH) leads to less dense SAMs and the incorporation of Bu(4)N into the monolayer. Furthermore, our results show the importance of the equilibrium concentrations of (di)thiolate in solution on the quality of the SAM. To demonstrate the relevance of these results for molecular electronics applications, large-area molecular junctions were fabricated using no base, Et(3)N, and Bu(4)NOH. The magnitude of the current-densities in these devices is highly dependent on the base. A value of β=0.15 Å(-1) for the exponential decay of the current-density of OPEs of varying length formed using Et(3)N was obtained. © 2011 American Chemical Society 3. Air quality high resolution simulations of Italian urban areas with WRF-CHIMERE Science.gov (United States) Falasca, Serena; Curci, Gabriele 2017-04-01 The new European Directive on ambient air quality and cleaner air for Europe (2008/50/EC) encourages the use of modeling techniques to support the observations in the assessment and forecasting of air quality. The modelling system based on the combination of the WRF meteorological model and the CHIMERE chemistry-transport model is used to perform simulations at high resolution over the main Italian cities (e.g. Milan, Rome). Three domains covering Europe, Italy and the urban areas are nested with a decreasing grid size up to 1 km. Numerical results are produced for a winter month and a summer month of the year 2010 and are validated using ground-based observations (e.g. from the European air quality database AirBase). A sensitivity study is performed using different physics options, domain resolution and grid ratio; different urban parameterization schemes are tested using also characteristic morphology parameters for the cities considered. A spatial reallocation of anthropogenic emissions derived from international (e.g. EMEP, TNO, HTAP) and national (e.g. CTN-ACE) emissions inventories and based on the land cover datasets (Global Land Cover Facility and GlobCover) and the OpenStreetMap tool is also included. Preliminary results indicate that the introduction of the spatial redistribution at high-resolution allows a more realistic reproduction of the distribution of the emission flows and thus the concentrations of the pollutants, with significant advantages especially for the urban environments. 4. Argumentation Quality of Socio-scientific Issue between High School Students and Postgraduate Students about Cancer Science.gov (United States) Anisa, A.; Widodo, A.; Riandi, R. 2017-09-01 Argumentation is one factor that can help improve critical thinking skills. Arguing means to defend statements with the various data, denials, evidence, and reinforcement that support the statement. The research aimed to capture the quality of argument skills by students in grade 12 high school students and in postgraduate student on social-scientific issues of cancer. Both group subjects are not in the same school or institution, chosen purposively with the subject of 39 high school students of grade 12 in one district of West Java and 13 students of Biology education postgraduate in one of University in West Java - Indonesia. The results of the quality structure of arguments in both subject groups show the same pattern, which is claim - warrant - and ground, with the quality of counterclaim aspects on the postgraduate students look better than grade 12 students. This provides an illustration that the ability in argumentation between students and teachers in the socio-scientific issue of cancer should be evaluate so that the learning process would be more refined in schools. 5. Void fraction in horizontal bulk flow boiling at high flow qualities International Nuclear Information System (INIS) Collado, Fancisco J.; Monne, Carlos; Pascau, Antonio 2008-01-01 In this work, a new thermodynamic prediction of the vapor void fraction in bulk flow boiling, which is the core process of many energy conversion systems, is analyzed. The current heat balance is based on the flow quality, which is closely related to the measured void fraction, although some correlation for the vapor-liquid velocity ratio is needed. So here, it is suggested to work with the 'static' or thermodynamic quality, which is directly connected to the void fraction through the densities of the phases. Thus, the relation between heat and the mixture enthalpy (here based on the thermodynamic quality instead of the flow one) should be analyzed in depth. The careful void fraction data taken by Thom during the 'Cambridge project' for horizontal saturated flow boiling with high flow qualities (≤0.8) have been used for this analysis. As main results, first, we have found that the applied heat and the increment of the proposed thermodynamic enthalpy mixture throughout the heated duct do not agree, and for closure, a parameter is needed. Second, it has been checked that this parameter is practically equal to the classic velocity ratio or 'slip' ratio, suggesting that it should be included in a true thermodynamic heat balance. Furthermore, it has been clearly possible to improve the 'Cambridge project' correlations for the 'slip' ratio, here based on inlet pressure and water velocity, and heat flux. The calculated void fractions compare quite well with the measured ones. Finally, the equivalence of the suggested new heat balance with the current one through the 'slip' ratio is addressed. Highlighted is the same new energetic relation for saturated flow boiling that has been recently confirmed by the authors for Knights data, also taken during the 'Cambridge project', which include not only horizontal but also vertical upwards flows with moderate outlet flow quality (≤0.2) 6. Void fraction in horizontal bulk flow boiling at high flow qualities Energy Technology Data Exchange (ETDEWEB) Collado, Fancisco J.; Monne, Carlos [Dpto. de Ingenieria Mecanica, Universidad de Zaragoza-CPS, Maria de Luna 3, 50018-Zaragoza (Spain); Pascau, Antonio [Dpto. de Ciencia de los Materiales y Fluidos, Universidad de Zaragoza-CPS, Maria de Luna 3, 50018-Zaragoza (Spain) 2008-04-15 In this work, a new thermodynamic prediction of the vapor void fraction in bulk flow boiling, which is the core process of many energy conversion systems, is analyzed. The current heat balance is based on the flow quality, which is closely related to the measured void fraction, although some correlation for the vapor-liquid velocity ratio is needed. So here, it is suggested to work with the 'static' or thermodynamic quality, which is directly connected to the void fraction through the densities of the phases. Thus, the relation between heat and the mixture enthalpy (here based on the thermodynamic quality instead of the flow one) should be analyzed in depth. The careful void fraction data taken by Thom during the 'Cambridge project' for horizontal saturated flow boiling with high flow qualities ({<=}0.8) have been used for this analysis. As main results, first, we have found that the applied heat and the increment of the proposed thermodynamic enthalpy mixture throughout the heated duct do not agree, and for closure, a parameter is needed. Second, it has been checked that this parameter is practically equal to the classic velocity ratio or 'slip' ratio, suggesting that it should be included in a true thermodynamic heat balance. Furthermore, it has been clearly possible to improve the 'Cambridge project' correlations for the 'slip' ratio, here based on inlet pressure and water velocity, and heat flux. The calculated void fractions compare quite well with the measured ones. Finally, the equivalence of the suggested new heat balance with the current one through the 'slip' ratio is addressed. Highlighted is the same new energetic relation for saturated flow boiling that has been recently confirmed by the authors for Knights data, also taken during the 'Cambridge project', which include not only horizontal but also vertical upwards flows with moderate outlet flow quality ({<=}0.2). (author) 7. Determining the Measurement Quality of a Montessori High School Teacher Evaluation Survey Directory of Open Access Journals (Sweden) Anthony Philip Setari 2017-05-01 Full Text Available The purpose of this study was to conduct a psychometric validation of a course evaluation instrument, known as a student evaluation of teaching (SET, implemented in a Montessori high school. The authors demonstrate to the Montessori community how to rigorously examine the measurement and assessment quality of instruments used within Montessori schools. The Montessori high school community needs an SET that has been rigorously examined for measurement issues. The examined SET was developed by a Montessori high school, and the sample data were collected from Montessori high school students. Using a Rasch partial credit model, the results of the analysis identified several measurement issues, including multidimensionality, misfit items, and inappropriate item difficulty levels. A revised version of the SET underwent the same analysis procedure, and the results indicated that measurement issues persisted. The authors suggest several ways to improve the overall measurement quality of the instrument while keeping the Montessori foundation. Additional validation studies with a revised version of the SET will be needed before the instrument can be endorsed for full implementation in a Montessori setting. 8. Mechanically stable tuning fork sensor with high quality factor for the atomic force microscope. Science.gov (United States) Kim, Kwangyoon; Park, Jun-Young; Kim, K B; Lee, Naesung; Seo, Yongho 2014-01-01 A quartz tuning fork was used instead of cantilever as a force sensor for the atomic force microscope. A tungsten tip was made by electrochemical etching from a wire of 50 µm diameter. In order to have mechanical stability of the tuning fork, it was attached on an alumina plate. The tungsten tip was attached on the inside end of a prong of a tuning fork. The phase shift was used as a feedback signal to control the distance between the tip and sample, and the amplitude was kept constant using a lock-in amplifier and a homemade automatic gain controller. Due to the mechanical stability, the sensor shows a high quality factor (∼10(3)), and the image quality obtained with this sensor was equivalent to that of the cantilever-based AFM. © 2014 Wiley Periodicals, Inc. 9. The Desire for (Danish) Quality in High and Low Income Countries DEFF Research Database (Denmark) Nguyen, Daniel Xuyen with sales. These correlations are significantly different across destination countries within product categories, but across years for a given product-destination pair. While some existing theories perform better than others at predicting these patterns, none can reconcile the variation across countries......We estimate the correlation between firm prices and sales within a CN8 product-country-year market. We do this for every market to which at least 16 different Danish firms exported between 1999 and 2006. Approximately 60% of Danish exports are to markets in which the price is negatively correlated....... To fully explain the patterns, We introduce a model in which the price-sales correlation can be interpreted as the market's desire for high quality goods over low cost substitutes. We discover an inverted U shaped relation between a country's desire for quality and its per capita GDP, which we term... 10. High-quality bulk hybrid perovskite single crystals within minutes by inverse temperature crystallization KAUST Repository Saidaminov, Makhsud I. 2015-07-06 Single crystals of methylammonium lead trihalide perovskites (MAPbX3; MA=CH3NH3+, X=Br− or I−) have shown remarkably low trap density and charge transport properties; however, growth of such high-quality semiconductors is a time-consuming process. Here we present a rapid crystal growth process to obtain MAPbX3 single crystals, an order of magnitude faster than previous reports. The process is based on our observation of the substantial decrease of MAPbX3 solubility, in certain solvents, at elevated temperatures. The crystals can be both size- and shape-controlled by manipulating the different crystallization parameters. Despite the rapidity of the method, the grown crystals exhibit transport properties and trap densities comparable to the highest quality MAPbX3 reported to date. The phenomenon of inverse or retrograde solubility and its correlated inverse temperature crystallization strategy present a major step forward for advancing the field on perovskite crystallization. 11. Strategies for Promoting High-Quality Care and Personal Resilience in Palliative Care. Science.gov (United States) Heinze, Katherine E; Holtz, Heidi K; Rushton, Cynda H 2017-06-01 Palliative care (PC) clinicians are faced with ever-expanding pressures, which can make it difficult to fulfill their duties to self and others and lead to moral distress. Understanding the pressures that PC clinicians face and the resources that could be employed to ease their moral distress is crucial to maintaining a healthy PC workforce and to providing necessary PC services to patients. In this paper, we discuss recommendations related to two promising pathways for supporting PC clinicians in providing high-quality PC: (1) improving systemic PC delivery and (2) strategies to promote ethical practice environments and individual resilience. Enacting these recommendations holds promise for sustaining higher-quality and accessible PC and a more engaged PC workforce. © 2017 American Medical Association. All Rights Reserved. 12. A protocol for generating a high-quality genome-scale metabolic reconstruction. Science.gov (United States) Thiele, Ines; Palsson, Bernhard Ø 2010-01-01 Network reconstructions are a common denominator in systems biology. Bottom-up metabolic network reconstructions have been developed over the last 10 years. These reconstructions represent structured knowledge bases that abstract pertinent information on the biochemical transformations taking place within specific target organisms. The conversion of a reconstruction into a mathematical format facilitates a myriad of computational biological studies, including evaluation of network content, hypothesis testing and generation, analysis of phenotypic characteristics and metabolic engineering. To date, genome-scale metabolic reconstructions for more than 30 organisms have been published and this number is expected to increase rapidly. However, these reconstructions differ in quality and coverage that may minimize their predictive potential and use as knowledge bases. Here we present a comprehensive protocol describing each step necessary to build a high-quality genome-scale metabolic reconstruction, as well as the common trials and tribulations. Therefore, this protocol provides a helpful manual for all stages of the reconstruction process. 13. Quality assurance during fabrication of high-damping rubber isolation bearings Energy Technology Data Exchange (ETDEWEB) Way, D.; Greaves, W.C. [Base Isolation Consultants, Inc., San Francisco, CA (United States) 1995-12-01 Successful implementation of a high-damping rubber (HDR) base isolation project requires the application of Quality Assurance/Quality Control (QA/QC) methodology through all phases of the bearing fabrication process. HDR base isolation bearings must be fabricated with uniform physical characteristics while being produced in large quantities. To satisfy this requirement, manufacturing processes must be controlled. Prototype tests that include dynamic testing of small samples of rubber are necessary. Stringent full scale bearing testing must be carried out prior to beginning production, during which manufacturing is strictly regulated by small rubber sample and production bearing testing. All such activities should be supervised and continuously inspected by independent and experienced QA/QC personnel. 14. Impact of individually controlled facially applied air movement on perceived air quality at high humidity Energy Technology Data Exchange (ETDEWEB) Skwarczynski, M.A. [Faculty of Environmental Engineering, Institute of Environmental Protection Engineering, Department of Indoor Environment Engineering, Lublin University of Technology, Lublin (Poland); International Centre for Indoor Environment and Energy, Department of Civil Engineering, Technical University of Denmark, Copenhagen (Denmark); Melikov, A.K.; Lyubenova, V. [International Centre for Indoor Environment and Energy, Department of Civil Engineering, Technical University of Denmark, Copenhagen (Denmark); Kaczmarczyk, J. [Faculty of Energy and Environmental Engineering, Department of Heating, Ventilation and Dust Removal Technology, Silesian University of Technology, Gliwice (Poland) 2010-10-15 The effect of facially applied air movement on perceived air quality (PAQ) at high humidity was studied. Thirty subjects (21 males and 9 females) participated in three, 3-h experiments performed in a climate chamber. The experimental conditions covered three combinations of relative humidity and local air velocity under a constant air temperature of 26 C, namely: 70% relative humidity without air movement, 30% relative humidity without air movement and 70% relative humidity with air movement under isothermal conditions. Personalized ventilation was used to supply room air from the front toward the upper part of the body (upper chest, head). The subjects could control the flow rate (velocity) of the supplied air in the vicinity of their bodies. The results indicate an airflow with elevated velocity applied to the face significantly improves the acceptability of the air quality at the room air temperature of 26 C and relative humidity of 70%. (author) 15. THE LEAN SIX SIGMAAPPROACH FOR PROCESS IMPROVEMENT: A CASE STUDY IN A HIGH QUALITY TUSCANY WINERY Directory of Open Access Journals (Sweden) Riccardo Bettini 2012-06-01 Full Text Available This paper describes the application of a Lean Six Sigma (LSS project to a winemaking process in a high-quality, Italian winery. LSS is used to focus on the problem through a quantitative analysis of waste and quality performances. The LSS basic algorithm (called “DMAIC” helps to detect and quantify critical aspects of the process for transferring liquid used in the cellar. The improvement solution is developed and applied through the modification of the cellar system and the process procedure. The results obtained with this solution are shown and discussed in this paper, so too the long term reliability of the improved process analyzed. The results obtained by this case study can help to understand the importance of the LSS method to drive the improvement of agricultural and agrofood productions also in terms of environmental impact which is strongly connected to waste reduction. 16. A top-down approach to fabrication of high quality vertical heterostructure nanowire arrays. Science.gov (United States) Wang, Hua; Sun, Minghua; Ding, Kang; Hill, Martin T; Ning, Cun-Zheng 2011-04-13 We demonstrate a novel top-down approach for fabricating nanowires with unprecedented complexity and optical quality by taking advantage of a nanoscale self-masking effect. We realized vertical arrays of nanowires of 20-40 nm in diameter with 16 segments of complex longitudinal InGaAsP/InP structures. The unprecedented high quality of etched wires is evidenced by the narrowest photoluminescence linewidth ever produced in similar wavelengths, indistinguishable from that of the corresponding wafer. This top-down, mask-free, large scale approach is compatible with the established device fabrication processes and could serve as an important alternative to the bottom-up approach, significantly expanding ranges and varieties of applications of nanowire technology. 17. Accelerator Magnet Quench Heater Technology and Quality Control Tests for the LHC High Luminosity Upgrade CERN Document Server AUTHOR|(CDS)2132435; Seifert, Thomas The High Luminosity upgrade of the Large Hadron Collider (HL-LHC) foresees the installation of new superconducting Nb$_{3}$Sn magnets. For the protection of these magnets, quench heaters are placed on the magnet coils. The quench heater circuits are chemically etched from a stainless steel foil that is glued onto a flexible Polyimide film, using flexible printed circuit production technology. Approximately 500 quench heaters with a total length of about 3000 m are needed for the HL-LHC magnets. In order to keep the heater circuit electrical resistance in acceptable limits, an approximately 10 µm-thick Cu coating is applied onto the steel foil. The quality of this Cu coating has been found critical in the quench heater production. The work described in this thesis focuses on the characterisation of Cu coatings produced by electrolytic deposition, sputtering and electron beam evaporation. The quality of the Cu coatings from different manufacturers has been assessed for instance by ambient temperature electrica... 18. NEW METHOD FOR ACTIVE SURFACES QUALITY INSPECTION FOR HIGH DIMENSIONS BEARINGS Directory of Open Access Journals (Sweden) BRAUN Barbu 2016-07-01 Full Text Available The paper presents a stage of an ample research on high quality control for different type of large dimension bearings, with application in wind turbines for electrical energy production. There is presented a new and efficient method for bearing’s active surfaces quality control, in terms of roughness, to ensure a good and safety functioning and a long life of these. The research involved the use of a powerful digital microscope for surface analysis followed by an assisted by PC algorithm for results obtaining in terms of roughness values. It has been observed that the proposed analyzing method could in the future to be successfully applied for a very large range of bearings, different dimensions and applications. 19. High quality copy number and genotype data from FFPE samples using Molecular Inversion Probe (MIP) microarrays Energy Technology Data Exchange (ETDEWEB) Wang, Yuker; Carlton, Victoria E.H.; Karlin-Neumann, George; Sapolsky, Ronald; Zhang, Li; Moorhead, Martin; Wang, Zhigang C.; Richardson, Andrea L.; Warren, Robert; Walther, Axel; Bondy, Melissa; Sahin, Aysegul; Krahe, Ralf; Tuna, Musaffe; Thompson, Patricia A.; Spellman, Paul T.; Gray, Joe W.; Mills, Gordon B.; Faham, Malek 2009-02-24 A major challenge facing DNA copy number (CN) studies of tumors is that most banked samples with extensive clinical follow-up information are Formalin-Fixed Paraffin Embedded (FFPE). DNA from FFPE samples generally underperforms or suffers high failure rates compared to fresh frozen samples because of DNA degradation and cross-linking during FFPE fixation and processing. As FFPE protocols may vary widely between labs and samples may be stored for decades at room temperature, an ideal FFPE CN technology should work on diverse sample sets. Molecular Inversion Probe (MIP) technology has been applied successfully to obtain high quality CN and genotype data from cell line and frozen tumor DNA. Since the MIP probes require only a small ({approx}40 bp) target binding site, we reasoned they may be well suited to assess degraded FFPE DNA. We assessed CN with a MIP panel of 50,000 markers in 93 FFPE tumor samples from 7 diverse collections. For 38 FFPE samples from three collections we were also able to asses CN in matched fresh frozen tumor tissue. Using an input of 37 ng genomic DNA, we generated high quality CN data with MIP technology in 88% of FFPE samples from seven diverse collections. When matched fresh frozen tissue was available, the performance of FFPE DNA was comparable to that of DNA obtained from matched frozen tumor (genotype concordance averaged 99.9%), with only a modest loss in performance in FFPE. MIP technology can be used to generate high quality CN and genotype data in FFPE as well as fresh frozen samples. 20. High quality antireflective ZnS thin films prepared by chemical bath deposition International Nuclear Information System (INIS) Tec-Yam, S.; Rojas, J.; Rejón, V.; Oliva, A.I. 2012-01-01 Zinc sulfide (ZnS) thin films for antireflective applications were deposited on glass substrates by chemical bath deposition (CBD). Chemical analysis of the soluble species permits to predict the optimal pH conditions to obtain high quality ZnS films. For the CBD, the ZnCl 2 , NH 4 NO 3 , and CS(NH 2 ) 2 were fixed components, whereas the KOH concentration was varied from 0.8 to 1.4 M. Groups of samples with deposition times from 60 to 120 min were prepared in a bath with magnetic agitation and heated at 90 °C. ZnS films obtained from optimal KOH concentrations of 0.9 M and 1.0 M exhibited high transparency, homogeneity, adherence, and crystalline. The ZnS films presented a band gap energy of 3.84 eV, an atomic Zn:S stoichiometry ratio of 49:51, a transmittance above 85% in the 300–800 nm wavelength range, and a reflectance below 25% in the UV–Vis range. X-ray diffraction analysis revealed a cubic structure in the (111) orientation for the films. The thickness of the films was tuned between 60 nm and 135 nm by controlling the deposition time and KOH concentration. The incorporation of the CBD-ZnS films into ITO/ZnS/CdS/CdTe and glass/Mo/ZnS heterostructures as antireflective layer confirms their high optical quality. -- Highlights: ► High quality ZnS thin films were prepared by chemical bath deposition (CBD). ► Better CBD-ZnS films were achieved by using 0.9 M-KOH concentration. ► Reduction in the reflectance was obtained for ZnS films used as buffer layers.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5290395021438599, "perplexity": 6748.879709665485}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585371807538.83/warc/CC-MAIN-20200408010207-20200408040707-00137.warc.gz"}
http://imi.cas.sc.edu/events/720/
IMI Interdisciplinary Mathematics InstituteCollege of Arts and Sciences ## Optimal sampling in weighted least-squares methods- Application to high-dimensional approximation • Nov. 21, 2017 • 4:15 p.m. • LeConte 312 ## Abstract Least squares methods are of common use when one needs to approximate a function based on its noiseless or noisy observation at n scattered points by a simpler function chosen in an m dimensional space with m less than n. Depending on the context, these points may be randomly drawn according to some distribution, or deterministically selected by the user. This talk will survey some recent results on the stability and approximation properties of weighted least squares method, in relation with the spatial distribution of the sampling. One of our main finding is that an optimal random sampling strategy can be defined by means of the Christoffel function associated with the space where the approximation is searched. Here, optimal means that near-best approximation properties are achieved at the prize of a number of sample n larger than m only by a logarithmic factor. One principal motivation is the application to high-dimensional approximation problems, such as coming PDEs with random input parameters. Motivated by this setting, we also discuss how the optimal sampling can be practically generated and incrementally updated when the approximation spaces are adaptively selected. This is a joint work with Giovanni Migliorati. Brief Bio: Albert Cohen earned his PhD in 1990 under the supervision of Yves Meyer. Since 1995 he has been a full professor at Laboratoire Jacques-Louis Lions, Sorbonne Universite, Paris. He is the author of over 100 papers in journals and 3 books. He was an invited speaker at ICM 2002 and plenary speaker at ICIAM 2006. He received the Vasil Popov, Jacques Herbrant, and Blaise Pascal prizes. He is currently a member of the Institut Universitaire de France. His interests include approximation theory, statistics, signal-image-data processing, and numerical analysis. © Interdisciplinary Mathematics Institute | The University of South Carolina Board of Trustees | Webmaster
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8388592600822449, "perplexity": 893.2381694518238}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-09/segments/1518891815318.53/warc/CC-MAIN-20180224033332-20180224053332-00204.warc.gz"}
https://www.hpmuseum.org/forum/thread-8091-post-71140.html
Important Request 04-02-2017, 08:33 PM Post: #1 math7 Member Posts: 182 Joined: Feb 2017 Important Request In the statistical application there are many options to do the regression analysis, but I think a very important option is missing which is the best fit option that does appear on the HP 50g, it is an important option so you do not have to search for all the options Of the calculator which is the curve that best fits the plotted data. 04-02-2017, 09:03 PM Post: #2 Tim Wessman Senior Member Posts: 2,239 Joined: Dec 2013 RE: Important Request So you are saying you should "numerically" decide which is the best fit on the data without doing any sort of analysis regarding whether the fit is actually apropriate for the situation? The 50g only has linear fits. Once you move into different fits, you have no equivalent to "R" to do comparison with. What would you propose comparing when there is no equivalent? TW Although I work for the HP calculator group, the views and opinions I post here are my own. 04-02-2017, 11:25 PM Post: #3 math7 Member Posts: 182 Joined: Feb 2017 RE: Important Request No, I mean that in the HP 50g there is an option when doing regression analysis that is called "Best Fit" that is not present in the Prime, this option in the Prime finds the lowest deviation of the statistical data; And as in the Prime is not this option because in reality it is difficult to realize what the best fit curve. 04-02-2017, 11:32 PM Post: #4 Paul Dale Senior Member Posts: 1,670 Joined: Dec 2013 RE: Important Request I with Tim here. It is far better to know enough about your data to choose an appropriate fit. One of the students I did honours statistics with did an exhaustive search for GLIMs on one assignment. All the way up to four factor interaction terms. The resulting best fit was utterly impossible to interpret and a terrible model for the data. - Pauli 04-03-2017, 04:16 AM Post: #5 math7 Member Posts: 182 Joined: Feb 2017 RE: Important Request Maybe I have not explained myself well. On the HP 50g for a data set or points with the "Best fit" option the calculator checks the best fit curve that passes through the given points, I think internally it calculates the least deviation using the least squares method for Each regression model, and then choose the one with the lowest deviation, and all that does in the "Best Fit" option of the statistics application, it would be good for the Prime too. 04-03-2017, 04:56 AM (This post was last modified: 04-03-2017 05:00 AM by Joe Horn.) Post: #6 Joe Horn Senior Member Posts: 1,686 Joined: Dec 2013 RE: Important Request I hesitate to disagree with the teaching experts here, but I agree with math7 on this. In fact, it would be very useful if Prime could simply create -- in one step -- a table of ALL the correlation coefficients for ALL the available curve fit types. My reason is that the BEST teacher is not the expert at the front of the classroom, or the expert who wrote the textbook, but the student's own CURIOSITY which leads him or her to PLAY with the concepts being taught. The best way to learn what correlation really means is by EXPLORING its value generated by different data sets for different curve fit equations. Easy access to that information would stimulate such exploration. But such exploration FAILS to be student-motivated if the calculator makes it a tedious procedure, as Prime currently does, requiring lots of keystrokes (as well as writing the results down on paper!) for each correlation coefficient. YES, understanding of the function that generated the data is certainly more important than blindly using an automated "best fit" button. Nobody disagrees with that. But that's not a good reason to omit a time-proven tool that stimulates student exploration, especially for the HP Prime whose primary market is students. In brief: Please stop thinking of "Best Fit" as a way for lazy cheaters to avoid thinking, and start thinking of it as a kind of "Explorer" that curious students will play with and learn from. I've watched many students play with it and learn well from it in the RPL models for over 25 years, and have never once seen it hinder learning. <0|ɸ|0> -Joe- 04-03-2017, 05:43 AM Post: #7 cyrille de brébisson Senior Member Posts: 1,009 Joined: Dec 2013 RE: Important Request Hello, Unfortunately, the "usual" best fit method, based on least (pred(x)-realy)² is actually incorrect as it calculates distances between data and curve on a vertical axis.... The "right" way to do is would be to calculate the distance based on "the" closest perpendicular to the curve that passes through the point... But this is hard and very computationally intensive to do. Cyrille Although I work for the HP calculator group, the views and opinions I post here are my own. I do not speak for HP. 04-03-2017, 01:18 PM Post: #8 mark4flies Member Posts: 112 Joined: Jan 2015 RE: Important Request Model selection is a long-standing problem in statistics and machine learning. It is still an very active research topic. There is no superior method, approach, or criterion. (That is, one that always returns the correct answer.) One of the best criteria today is Akaike's Information Criterion, corrected for small sample sizes (AICc). It is a combination of a measure of the reduction in the bias (under-fitting) and a measure of the complexity of the model (over-fitting, which leads to variance). The best model is a trade off between bias and variance that is achieved by minimizing AICc. AICc can be computed directly from the likelihood function of the model parameters when using maximum likelihood estimation or from the sum of the squared errors when using a least squares method. 04-03-2017, 03:24 PM (This post was last modified: 04-03-2017 03:25 PM by Tim Wessman.) Post: #9 Tim Wessman Senior Member Posts: 2,239 Joined: Dec 2013 RE: Important Request When there is no equivalent calculation of "error" for ALL fit types currently presented, what would I compare? I'd be fine to implement something, but I have not seen any math that applies to all fit types! Using one type of error calculation mixed with a different type simply messes things up due to bias in the various types of calculating things. TW Although I work for the HP calculator group, the views and opinions I post here are my own. 04-03-2017, 06:49 PM (This post was last modified: 04-03-2017 06:59 PM by math7.) Post: #10 math7 Member Posts: 182 Joined: Feb 2017 RE: Important Request The programmers of the HP 50g incorporated that and other tools into the calculator, if they did it was because they thought it was useful and beneficial for statistical calculations and based on it realize what curve is best fit a series of data dices. In the HP Prime there are 13 options for adjusting curves for the series of data entered in the numerical view. Please see attached pictures. A "simple view" can be observed that the "cubic" model is better suited to most points than the "exponential" model. This is seen with the naked eye, but you should not rely on the view. Then the dispersion of the data the calculator could calculate it for the 13 internal models that it has instead of having to do it one at a time, and of all those with the lowest dispersion will be the "best fit" curve. It is cumbersome for the user to have to test the 13 different models to see which fits better, for that is the calculator. Derivatives or integrals can also be done by hand, but for that the calculator also brings it in, not for the user to do it but it. A system of 2 x 2 equations can be solved one in paper, but the calculator also does it by itself, for that it comes from factory. It can not be possible for the user to have to test all 13 different models and point on a paper as Joe Horn says to decide next which is the curve that best fits and therefore have his equation. The Prime must calculate for the 13 models which is the one that best approaches all the data entered, if the HP 50g does, with more reason HP Prime with a processor of 400 MHz. Attached File(s) Thumbnail(s) 04-03-2017, 07:52 PM Post: #11 johnet123 Junior Member Posts: 13 Joined: Jul 2015 RE: Important Request I agree with Joe Horn's opinion with regard to student's curiosity. Years ago, a work colleague wrote a BASIC language regression program that tested for many types of models (I believe it was 20 or thirty) and ranked them by their correlation coefficient. Although it can easily be argued that it was a shotgun approach to analyzing data, it did cause one to explore and exam data points. Obviously, some of the regression fits were weird or silly but, still, it was all very informative. Anything that causes a student to explore the subject they are learning can not be a bad thing. Obviously, it is incumbent upon a student's teacher to teach a student not to blindly rely on a cookbook or shotgun approach to the subject they are learning. Lastly, the aforementioned program I was referring to, was written on an old, slow IBM computer. How hard would it be to write or do something similar on HP's whizbang HP Prime? 04-03-2017, 08:25 PM (This post was last modified: 04-03-2017 08:27 PM by Tim Wessman.) Post: #12 Tim Wessman Senior Member Posts: 2,239 Joined: Dec 2013 RE: Important Request The problem here is that you cannot compare "error" in all the models as there is not a mathematical way to do so except in the case of a *LINEAR* model. Every fit in the 50g is a linear fit. The options other then "linear" are just data that has some transformations to the data, but they are in reality simple linear fits. This makes comparing them trivial. Until someone can point out a unifying error algorithm that applies to all regressions equally, I don't think there is anything I can do here. There is not an equivalent "error" value for polynomial fits (there is a value that is close, but not directly comparable to the linear "error"), logistic fits, or trigonometric fits. In addition, once you allow an Nth polynomial fit, your error by definition will not matter as simply adjusting to N+1 for your polynomial with perfectly match the data. I am also completely unaware of *any* statistical package that offers an "auto-fit" for anything but simple, linear regressions, and even those are extremely rare. The other major problem is that unless your data is very small, the logistic and sinusoidal fitting is VERY slow and consumes VERY large amounts of memory. This is because you have to do a lot of matrix calculations, inversions, and iterations to attempt to minimize errors. They are not "simple" calculations like the linear or polynomial fits and are very noticeable in calculation time when you have anything other then small data sets. If you attempt to run these on data that is not already a decent fit, you can see seconds of calculation time with poor results at the end. Running that constantly for each display would not be a very good UI experience. Well then, why not just do the linear fits? That was a possibility. However, explaining "this is a best fit, but only for these and not those fits" doesn't really seem like a very understandable explanation and creates other problems in the workflow/ui. Nor is a "best fit" a feature that has been in high demand (in fact, you are the first person to bring it up). There is also a very strong argument that doing a "Best Fit" is in fact a bad practice that should NOT be encouraged or taught for several important reasons. The great thing is that the Prime is programmable! You can very easily make your own "best fit" with a few simple commands. Stick this in your app program: Code: VIEW "Best-Fit",BestFit() BEGIN LOCAL j,r,fit; FOR j FROM 1 TO 6 DO   S1(4):=j; //set fit type   STARTVIEW(0,1); //load settings, redraw   Do2VStats(); //recalculate      IF CoefDet > r THEN      r:=CoefDet; //found a better fit      fit:=j;    END; END; MSGBOX("CoefDet:" + r + "\n"    + "Fit Selected:" + fit); S1(4):=fit; //set our selected STARTVIEW(1,1); //see the plot END; To run it, press your VIEW key on the app and you have a nice interface doing the best fit. I am not against adding a "best fit" type of option, I just would want to do it correctly and right now I don't see any way to do that. TW Although I work for the HP calculator group, the views and opinions I post here are my own. 04-03-2017, 11:30 PM (This post was last modified: 04-03-2017 11:33 PM by math7.) Post: #13 math7 Member Posts: 182 Joined: Feb 2017 RE: Important Request Yes, something like that is what i mean Tim, I copied your code into the statistical application and it works fine, that's what I mean by doing the HP Prime, that she calculates and chooses the best approach, I think it should not be so difficult if you already did it for a Model as the complete program would be for the 13 models, a little more code, and the calculator would choose the one that has a better fit to the data, it is simply something like that, if the HP 50g exists I do not see why not in the Prime. Note: I suppose for a lot of points as it does take more processing and amount of memory in the calculator, but the option would be viable for practical life problems for about 20 or 30 points I think the calculator is quite capable and speedy to calculate something so. 04-03-2017, 11:43 PM Post: #14 math7 Member Posts: 182 Joined: Feb 2017 RE: Important Request I imagine that perhaps in the development team of the Prime there are people who know a lot of mathematical and numerical analysis, which is the area that studies methods of approach to solve equations, functions, algorithms of different types, etc. 04-04-2017, 03:31 AM Post: #15 Tim Wessman Senior Member Posts: 2,239 Joined: Dec 2013 RE: Important Request (04-03-2017 11:30 PM)math7 Wrote:  I think it should not be so difficult if you already did it for a Model as the complete program would be for the 13 models, a little more code, and the calculator would choose the one that has a better fit to the data, it is simply something like that, if the HP 50g exists I do not see why not in the Prime. Please re-read carefully my post. I explained in detail several times and reasons why this request is neither simple, nor mathematically possible to the best of my knowledge. There is no such thing as "correlation" for ANYTHING except a linear model. Other methods of measuring error are not compatible. Also, please note that you are talking to the person that wrote 95+% of the statistics code in Prime. I did spend probably at least a month of time over the course of the 39gII/Prime researching and investigating if this would be possible. I'd love it if I were corrected here, but so far I've not seen anything that would be possible to implement. I do agree with Joe that a "Best fit" opens up some interesting learning options and investigation paths, but as I've shown a simple app function gives that option. I'm not sure how to build it in, nor the appropriateness of doing so. A lot of statistics educators actually were vehemently against having such an option built in. TW Although I work for the HP calculator group, the views and opinions I post here are my own. 04-04-2017, 03:41 AM (This post was last modified: 04-04-2017 03:45 AM by Han.) Post: #16 Han Senior Member Posts: 1,842 Joined: Dec 2013 RE: Important Request (04-03-2017 11:30 PM)math7 Wrote:  Yes, something like that is what i mean Tim, I copied your code into the statistical application and it works fine, that's what I mean by doing the HP Prime, that she calculates and chooses the best approach, I think it should not be so difficult if you already did it for a Model as the complete program would be for the 13 models, a little more code, and the calculator would choose the one that has a better fit to the data, it is simply something like that, if the HP 50g exists I do not see why not in the Prime. Note: I suppose for a lot of points as it does take more processing and amount of memory in the calculator, but the option would be viable for practical life problems for about 20 or 30 points I think the calculator is quite capable and speedy to calculate something so. You did not read Tim's response very carefully. The program he posted only works on regression models whose unknown parameters are linear. Even then, given any n distinct points, a polynomial fit (of degree n-1) will always be exact (i.e. "best fit") since it will pass through every single one of the n points. Thus, the "answer" to the "best fit" question is always a polynomial fit. Of course, this is also nonsense because polynomials are not always the best models of data (consider any set of data corresponding to periodic phenomenon). It would NOT be easy to generalize Tim's program to all the other fits because there is not "standard" for comparison. It would be like comparing apples to oranges (because only some of the regression models are linear; others are not). You cannot define "best" when there is no standard for comparison. You cannot even use a "least amount of error" type of comparison because a polynomial fit will always have 0 error unless your points have repeated input values. Graph 3D | QPI | SolveSys 04-04-2017, 04:49 AM (This post was last modified: 04-04-2017 05:01 AM by math7.) Post: #17 math7 Member Posts: 182 Joined: Feb 2017 RE: Important Request What I mean is that the calculator gives the lowest value of R^2 and r as seen in the figures, not having to do the user one by one for each model adjustment, but the calculator does it automatically, it is very little Processing for it. The closer to one is the R value the better fit the model will have and therefore the curve. Attached File(s) Thumbnail(s) 04-04-2017, 05:48 AM (This post was last modified: 04-04-2017 05:51 AM by Joe Horn.) Post: #18 Joe Horn Senior Member Posts: 1,686 Joined: Dec 2013 RE: Important Request Thanks, Tim and Han, for explaining how it's simply not possible to have a "Best Fit" for ALL the curve fit models offered by Prime. Therefore I'd be happy with an R^2 table for just those models which DO have a value for R^2 (viz. Linear, Logarithmic, Exponential, Power, Exponent, Inverse, Quadratic, Cubic, and Quartic)... unless implementing such a feature would reduce Prime's adoption by the Education Establishment, of course. It's good to know that a user program can create such a table, if the teacher or student want to stroll down that avenue of exploration. <0|ɸ|0> -Joe- 04-04-2017, 05:12 PM Post: #19 Tim Wessman Senior Member Posts: 2,239 Joined: Dec 2013 RE: Important Request The R^2 values in polynomial fits are not "technically" equivalent to the linear ones. They are roughly equivalent though, which can make them "mostly comparable". I've been thinking about how to do this. I think a table that shows differences of some key values would be a good compromise. TW Although I work for the HP calculator group, the views and opinions I post here are my own. 04-05-2017, 04:46 AM (This post was last modified: 04-05-2017 04:53 AM by math7.) Post: #20 math7 Member Posts: 182 Joined: Feb 2017 RE: Important Request Of course it would be a good thing. A table with the values of R ^ 2 or r and thus to be able to compare the different models. I did not do the Prime statistics app. But the calculator, as seen in the images I attached above, calculates when the statistics button is given several things, including the value of R ^ 2 and r. These values should work for something, say Tim, so what I'm saying is that the calculator makes a table with those values for the 13 models it has and then the user will know which is the best model for the data plotted. « Next Oldest | Next Newest » User(s) browsing this thread: 1 Guest(s)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.435149222612381, "perplexity": 978.362352413467}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610703506832.21/warc/CC-MAIN-20210116165621-20210116195621-00147.warc.gz"}
https://euro-math-soc.eu/review/linear-canonical-transforms
# Linear Canonical Transforms The Fourier transform rotates the time-frequency content of a signal over 90 degrees from the time axis to the frequency axis. Already in the 1970s it was observed that certain optical systems rotated the signal over an arbitrary angle, which became known as the fractional Fourier transform since it it acts like a fractional (i.e., a real) power of the Fourier operator. The fractional Fourier transform is thus depending on this rotation angle forms a one-parameter family of transforms. Around the same time, quantum physicists realized that certain systems could transform the position and momentum as a vector into their new values leaving the quantum mechanics invariant. All such transforms were obtained by multiplying the position-momentum vector with a unimodular matrix. In the scalar case such a matrix depends on 3 free parameters. In a general N-dimensional space we have a 2 by 2 block matrix that form a real symplectic group Sp(2N). These transforms are the essence of the family of linear canonical transforms (LCT). When in 2D optics the vector of position and momentum of the optical ray is considered, the same kind of transformation can be used, and the fractional Fourier transform like many other (fractional) transforms can be seen as a special case of the LCT. The reader less familiar with the subject should be warned that these fractional transforms are not directly related to the equally important and equally flourishing domain known as fractional calculus which studies fractional derivatives and fractional integrals. Since the early days many papers appeared on all kinds of fractional transforms, and even several books, among which a basic one on The Fractional Fourier Transform by Haldun M. Ozaktas, Zeev Zalevsky, and M. Alper Kutay in 2001. There the emphasis was on definitions, mathematical properties and computation and their applications in optics and signal processing. The LCT is already there but it is not the main focus. Two of the authors of that book are now also editor of the present one. One could think of it as the LCT analog of their fractional Fourier transform book, but less extensive, and it is not a monograph. Several experts are contributing to the present book. It gives an up-to-date overview of the many aspects of the LCT. As it appears as a volume of the Springer Series in Optical Sciences, there is an understandable bias towards the optical viewpoint and applications, with less emphasis on the quantum physics. The fifteen chapters are subdivided into three parts: (1) Fundamentals, (2) Discretization and computation, (3) Applications. All the aspects are covered: operator theory, theoretical physics, analysis, and group theory in the early chapters, discrete approximations and digital implementation in the second part, and it ends with some applications in the last part. In the first part one gets the basics in some 100 pages: some history and of course the definition and properties, the kernels when written as an integral transform, all the types and special cases, and the effects of the transform in phase-space. The eigenmodes are as important as the Gauss-Hermite eigenfunctions of the Fourier transform. So there is a separate chapter dealing with the eigenfunctions. Also uncertainty principles play an important role when it comes to sampling theory for these transforms. The first part is completed with two chapters covering extensively the optical aspects of the LCT. The second part deals with the computational aspects. While the fractional Fourier transform resulted in a rotation in the time-frequency plane, the LCT will result in oblique transforms. It is then important to obtain some analogs of bandwidth, sampling theorems, and degrees of freedom in the signal when one wants to come to an implementation of a fast discrete LCT transform. Analyzing these effects is related to a decomposition of the LCT in a sequence of elementary transforms which boils down to a sequence of chirp multiplications and Fourier transforms. Several possibilities are proposed to come to a fast digital LCT implementation, although no software is provided. The approach taken is from a signal processing viewpoint. One chapter gives an alternative that is based on optical interpretation. That alternative approach relies on coherent self-imaging, known as Talbot effect which describes Fresnel diffraction of a strictly periodic wavefront. Knowing that Fresnel diffraction is a special case of LCT. Therefore various generalizations of self-imaging in the wider LCT context can also lead to a practical implementation of discrete LCT. The application part illustrates how LCT can be used to solve certain problems like deterministic phase retrieval, analyzing holographic systems, double random phase encoding, speckle metrology and quantum states of light. This book is a most welcome addition to the literature. The subjects discussed appear in quite diverse contexts and it is therefore difficult to get the same overview as it is presented here. The general approach is the same as was used in the fractional Fourier book that I mentioned above, but it is less thorough. Moreover, since the different chapters are written by different authors, and because they highlight different aspects, the notation is not always strictly uniform, but that should not be hindering too much. The part on the algorithmic implementation is rather detailed but no software is provided, so it is a challenge for computer scientists to design an optimal implementation so that it can become standard software in signal processing packages. Reviewer: Book details The linear canonical transform is a phase space transform with roots in optics and quantum mechanics. This is a collection of survey papers written by renowned specialists that give a state-of-the-art of the many aspects of the linear canonical transform with an emphasis on the optical interpretation and applications. The quantum mechanical aspects are less present. It covers the basic mathematical and optical aspects, the possible implementation of a fast discrete algorithm, and several of the possible optical and signal processing applications. Author:  Publisher: Published: 2015 ISBN: 978-1-4939-3027-2 (hbk) Price: 137,79 € (hbk) Pages: 264 Categorisation
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8644567131996155, "perplexity": 491.61482377913273}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875145534.11/warc/CC-MAIN-20200221172509-20200221202509-00516.warc.gz"}
https://sbainvent.com/fluid-mechanics/compressibility-of-a-fluid/
# Bulk Modulus When analyzing a fluid it is important to consider how easily the fluids volume will change due to a pressure change.  The reason why is because as that fluids volume changes its density will change.  To determine the compressibility of a fluid you will need to determine what the bulk modulus is.  The bulk modulus is found by finding the ratio of the change in pressure over the volume differential. (Eq 1) $E_v=-\frac{dP}{dV/V}$ Also, since the density of a fluid will change if its volume changes, the density differential can also be used to determine the bulk modulus. (Eq 2)  $E_v=-\frac{dP}{dρ/ρ}$ The resulting dimensions of the bulk modulus will be $FL^{-2}$.  For BG units this will be $lb/in^2$ (psi).  While for SI units it will be $N/m^2$ (Pa).  Notice that these are the same units that are used for the modulus of elasticity for a solid. When the bulk modulus has a large value you can normally consider the fluid to be incompressible.  This is because it will take a large pressure differential to cause a relatively small change in fluid volume or density.  Liquids are normally always considered incompressible due to their large bulk modulus.  The table below shows the bulk modulus of some select liquid.  From this table you can see that a large pressure change will be required to create a volume or density change in these liquids. Liquid Temperature Bulk Modulus Fahrenheit Celsius psi Pa Ethyl Alcohol 68 20 1.54e5 1.06e9 Gasoline 60 15.6 1.60e5 1.30e9 Mercury 68 20 4.14e6 2.85e10 SAE 30 oil 60 15.6 2.20e5 1.50e9 Sea Water 60 15.6 3.39e5 2.34e9 Water 60 15.6 3.12e5 2.15e9 ### Compression and Expansion of a Gas The relationship between a gas’s pressure and density when the gas expands or contracts is dependent on the type of process. ##### Isothermal Process The first process that i’m going to talk about is the isothermal process.  For this process to take place the temperature must remain constant.  This will result in the following pressure and density relationship. (Eq 3) $\frac{P}{ρ}=constant$ ##### Isentropic Process The next process that I would like to mention is the isentropic process.  For the isentripic process to occur heat cannot be exchanged with the surroundings.  The pressure and density relationship would be almost the same as the isothermal process, except you will need to consider the ratio of the specific heat at constant pressure $c_p$ and the specific heat at constant volume $c_V$. (Eq 4) $k=\frac{c_p}{c_V}$ Note the two specific heat values can be used to find the ideal gas constant of the gas. (Eq 5) $R= c_p-c_V$ Finally, once the ratio between the two specific heat values is determined you can now calculate the relationship between the pressure and density. (Eq 6) $\frac{P}{ρ^k}=constant$ ##### Bulk Modulus of a Gas Unlike a liquid, a gas is considered compressible, since a relatively low pressure change will noticeably change the volume that a gas occupies.  Due to this fact the bulk modulus will change directly with pressure.  The bulk modulus can be found by taking the derivative $\frac{dp}{dρ}$. The resulting bulk modulus for the isothermal process will be as following. (Eq 7) $E_v=P$ The resulting bulk modulus for the isentropic process will be as follows. (Eq 8) $E_v=kP$ ### Example If a gas was being compressed from one state to another determine the following equations.  Write the an equation that can be used to find the final pressure of the gas if it were being compressed isothermally. Write the an equation that will determine the final pressure of the gas if it were compressed isentropically. ### Solution Isothermal Process $\frac{p_i}{ρ_i}=\frac{p_f}{ρ_f}$ $p_f = p_i\left(\frac{ρ_f}{ρ_i}\right)$ Isentropic Process $\frac{p_i}{{ρ_i}^k}=\frac{p_f}{{ρ_f}^k}$ $p_f = p_i\left(\frac{ρ_f}{ρ_i}\right)^k$
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9188816547393799, "perplexity": 686.834123480453}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-29/segments/1593655887319.41/warc/CC-MAIN-20200705090648-20200705120648-00263.warc.gz"}
https://eprints.soton.ac.uk/425561/
The University of Southampton University of Southampton Institutional Repository ePrints Soton is experiencing an issue with some file downloads not being available. We are working hard to fix this. Please bear with us. NICER discovers mHz oscillations in the "clocked" burster GS 1826-238 Strohmayer, T. E., Gendreau, K. C., Altamirano, D., Arzoumanian, Z., Bult, P. M., Chakrabarty, D., Chenevez, J., Guillot, S., Guver, T., Homan, J., Jaisawal, G. K., Keek, L., Mahmoodifar, S., Miller, J. M. and Ozel, F. (2018) NICER discovers mHz oscillations in the "clocked" burster GS 1826-238. Astrophysical Journal, 865 (1), 1-12, [63]. Record type: Article Abstract We report the discovery with the Neutron Star Interior Composition Explorer (NICER) of mHz X-ray brightness oscillations from the "clocked burster" GS 1826-238. NICER observed the source in the periods 2017 June 20-29, July 11-13, and September 9-15, for a total useful exposure of 34 ks. Two consecutive dwells obtained on 2017 September 9 revealed highly significant oscillations at a frequency of 8 mHz. The fractional, sinusoidal modulation amplitude increases from 0.7% at 1 keV to ≈2% at 6 keV. Similar oscillations were also detected at lower significance in three additional dwells. The oscillation frequency and amplitude are consistent with those of mHz QPOs reported in other accreting neutron star systems. A thermonuclear X-ray burst was also observed on 2017 June 22. The burst properties and X-ray colors are both consistent with GS 1826 being in a soft spectral state during these observations, findings that are confirmed by ongoing monitoring with MAXI and SWIFT-BAT. Assuming that the mHz oscillations are associated with blackbody emission from the neutron star surface, modeling of the phase-resolved spectra shows that the oscillation is consistent with being produced by modulation of the temperature component of this emission. In this interpretation, the blackbody normalization, proportional to the emitting surface area, is consistent with being constant through the oscillation cycle. We place the observations in the context of the current theory of marginally stable burning and briefly discuss the potential for constraining neutron star properties using mHz oscillations. Text NICER Discovers mHz Oscillations in the "clocked" Burster GS 1826-238 - Accepted Manuscript Accepted/In Press date: 11 August 2018 e-pub ahead of print date: 21 September 2018 Keywords: stars: neutron, stars: oscillations, stars: rotation, X-rays: binaries, X-rays: individual (GS 1826,238) Identifiers Local EPrints ID: 425561 URI: http://eprints.soton.ac.uk/id/eprint/425561 ISSN: 0004-637X PURE UUID: 35176ce9-6786-4302-9068-014b3780a203 ORCID for D. Altamirano: orcid.org/0000-0002-3422-0074 Catalogue record Date deposited: 24 Oct 2018 16:30 Contributors Author: T. E. Strohmayer Author: K. C. Gendreau Author: D. Altamirano Author: Z. Arzoumanian Author: P. M. Bult Author: D. Chakrabarty Author: J. Chenevez Author: S. Guillot Author: T. Guver Author: J. Homan Author: G. K. Jaisawal Author: L. Keek Author: S. Mahmoodifar Author: J. M. Miller Author: F. Ozel
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9054571390151978, "perplexity": 15059.180059529912}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-49/segments/1637964362969.51/warc/CC-MAIN-20211204094103-20211204124103-00095.warc.gz"}
https://answers.ros.org/question/312736/extend-colcon-workspace-without-sourcing-setupsh/
Extend colcon workspace without sourcing setup.sh edit Using colcon, is there a way to Explicitly Specifying Workspace Chaining without sourcing the parent workspace directly via setup.sh? With catkin_tools, there exists the --extend; is there a colcon equivalent? I didn't see anything of link in the migration guide: Or is sourcing install/setup.sh the conventional approach going forward? edit retag close merge delete Sort by » oldest newest most voted The --extend option is define within catkin which generates the setup files (not catkin_tools). The chaining of workspaces basically works as follows: • If you build a workspace A without sourcing any setup file before the workspace A is not chained (so it can't use anything from other workspaces) • When you source the setup file of workspace A and then build a workspace B that workspace B will overlay the sourced workspace A). • When you then source the setup file of workspace B and then build a workspace C that workspace C will overlay the sourced workspace B and the workspace A). What you can commonly not do is source two setup files. At least you won't get the result you might expect. Lets consider the following example: • Workspace A was built without sourcing anything else before. • Workspace B was built without sourcing anything else before. • When sourcing the setup file of workspace A and then of workspace B the resulting environment will contain only the stuff from B. The reason is that the setup files will try their best to undo the environment changes from workspace A when sourcing the setup file of workspace B. • Their are two options to get a "merged" environment: • Rebuild one of the workspaces with the other workspace sourced to chain them. • Source the second setup file with the option --extend. After describing all this for ROS 1 finally getting to the difference in ROS 2: The setup files in ROS 2 / ament do not contain any logic to undo changes of previously sourced workspaces. Simply because that logic is very fragile and likely to be incomplete in non-trivial cases. As such they also don't provide the --extend logic since you can simply source two setup files from independent workspaces and the result will be an environment containing both. more
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5986207723617554, "perplexity": 4881.289655486559}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243991812.46/warc/CC-MAIN-20210515004936-20210515034936-00049.warc.gz"}
https://docs.blender.org/manual/ja/2.82/physics/particles/emitter/render.html
# Render¶ Reference Panel Particle System ‣ Render The Render Panel controls how particles appear when they are rendered. Cycles supports only Object and Collection render types. ## Common Settings¶ Scale Todo. Scale Randomness Todo. Material Set which of the object's materials is used to shade the particles. Coordinates System Use a different object's coordinates to determine the birth of particles. Show Emitter When disabled, the emitter is no longer rendered. Activate the button Emitter to also render the mesh. ## Render As¶ ### None¶ When set to None, particles are not rendered. This is useful if you are using the particles to duplicate objects. ### Halo¶ Halo are rendered as glowing dots or a little cloud of light. Although they are not really lights because they do not cast light into the scene like a light object. They are called Halos because you can see them, but they do not have any substance. ### Path¶ The Path visualization needs a Hair particle system or Keyed particles. B-Spline Interpolate hair using B-splines. This may be an option for you if you want to use low Render values. You loose a bit of control but gain smoother paths. Steps Set the number of subdivisions of the rendered paths (the value is a power of 2). You should set this value carefully, because if you increase the render value by two you need four times more memory to render. Also the rendering is faster if you use low render values (sometimes drastically). But how low you can go with this value depends on the waviness of the hair (the value is a power of 2). This means 0 steps give 1 subdivision, 1 give 2 subdivisions, 2 --> 4, 3 --> 8, 4 --> 16, ... n --> n2. ### Timing¶ Reference Panel Particle System ‣ Render ‣ Timing Type Hair Absolute Path Time Path timing is in absolute frames. End End time of the practical path. Random Give the path length a random variation. ### Object¶ Reference Panel Particle System ‣ Render ‣ Object Instance Object The specified object is instanced in place of each particle. Global Coordinates Use object's global coordinates for instancing. Object Rotation Use the rotation of the object. Object Scale Use the size of the object. ### Collection¶ Reference Panel Particle System ‣ Render ‣ Collection Instance Collection The objects that belong to a collection are instanced sequentially in the place of the particles. Whole Collection Use the whole group at once, instead of one of its elements, the group being displayed in place of each particle. Pick Random The objects in the group are selected in a random order, and only one object is displayed in place of a particle. Please note that this mechanism fully replaces old Blender particles system using parentage and Instancing Verts to replace particles with actual geometry. This method is fully deprecated and does not work anymore. Global Coordinates Use object's global coordinates for instancing. Object Rotation Use the rotation of the objects. Object Scale Use the size of the objects. #### Use Count¶ Reference Panel Particle System ‣ Render ‣ Collection ‣ Use Count Use objects multiple times in the same groups. Specify the order and number of times to repeat each object with the list view that appears. You can duplicate an object in the list with the + button, or remove a duplicate with the - button. ## Extra¶ Reference Panel Particle System ‣ Render ‣ Extra Parents Particles Render also parent particles if child particles are used. Children have a lot of different deformation options, so the straight parents would stand between their curly children. So by default Parents are not rendered if you activate Children. See Children. Unborn Render particles before they are born. Dead Render particles after they have died. This is very useful if particles die in a collision Die on hit, so you can cover objects with particles.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.1834516078233719, "perplexity": 3246.7498704393365}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656103328647.18/warc/CC-MAIN-20220627043200-20220627073200-00620.warc.gz"}
http://jiangtanghu.com/blog/2012/09/16/statistical-notes-4-dragonrsquos-teeth-and-fleas-hypothesis-testing-in-plain-english/
# Statistical Notes (4): Dragon&rsquo;s Teeth and Fleas: Hypothesis Testing in Plain English I was asked in several different occasions to explain hypothesis testing to non-technical people in plain English. Now I think I got a pretty neat one while honors belong to three great Germans, Friedrich Engels, Karl Marx, and Heinrich Heine. Engels wrote that Marx once quoted a saying from Heine that “I have sown dragon’s teeth and harvested fleas” (but I didn’t find the original source): In terms of hypothesis testing, say, if you plant dragon’s teeth, you are not supposed to harvest the fleas; but if you do get the fleas, then most likely they are not dragon’s teeth at all! To make up the story as simple as possible while keep most of the key messages, here take a one-sample T-test example from SAS TTEST Procedure User Guide, where the investigated data is the Degree of Reading Power (DRP, a measurement of children’s reading skill) from 44 third-grade children. The goal is to test if the mean score of DRP is equal to 30: Null        Hypothesis (H0): μ = 30 Alternative Hypothesis (H1): μ ≠ 30 For the following discussion, three statistical measures needed (number of cases, mean and standard error): proc means data=read maxdec=4 n mean stderr; var score; freq count; run; and the output: Here we go: 1. Suppose H0 is true that the mean score is equal to 30 (population mean μ = 30, we assume they are dragon’s teeth!). 2.  We know the sample mean is the best estimate of the population mean, here it is 34.8636. 3. This sample mean is derived from a sample (namely, the 44 children). We’d like to know if this sample (with mean of 34.8636) really come from the population (with mean of 30). 4. To get to know this, we can repeat this survey (or trial, experiment), for example, take another 99 samples (also with 44 children in each sample) and calculate their sample means (we then have 100 different means from the whole 100 samples). 5.  We may know these means (got from such mechanism), with some necessary mathematical transformation,  follow a t-distribution with degree of freedom of 43 (df = 44 – 1): The denominator as a whole is the standard error. 6. Now we can check our interested sample (with mean of 34.8636) in this t-distribution. There is a corresponding t-value of this sample in the distribution, (34.8636 – 30) / 1.6930 =  2.8728: 7. In this distribution, there are only 0.63% (0.0063) of t-values either above 2.8728 or below -2.8728 (symmetry of t-distribution). What does this mean?  Obviously, our interested sample (with mean of 34.8636 and t-value of 2.8728) is not in the mainstream of the whole distribution. We can even think it is  very “extreme”  because, there are only 0.63% of elements in the whole that are more (or at least as equal) extreme than it. 8. We call this number, 0.63% (0.0063) the P-value. According to wikipedia: the p-value is the probability of obtaining a test statistic at least as extreme as the one that was actually observed, assuming that the null hypothesis is true. Here the test statistic that was actually observed is 2.8728, the t-value from our interested sample. And we got from this t-distribution that P{|t|>2.8728} = 0.0063. 9. A question is, how extreme is extreme? We can’t define “extreme” exactly, but probably, we can. The threshold is the significance level α (usually 0.05). As a rule of thumb, if p-value is less than the predefine significance level α, then  we think the event associated  with this p-value is pretty extreme. 10. Retell our story: first we assume the population mean is 30 just as the null hypothesis claims (assume they are dragon’s teeth), and under this assumption,  we built a t-distribution; we then got that our interested sample is an extreme case (it is a flea!). It is so extreme (much less than 0.05) that we can even think about that our assumption (the null hypothesis) is far away from true (most likely they are not dragon’s teeth at all!). We reject the null hypothesis and take the alternative that the mean score is different from 30. 11. Formally, it just follows a simple logic (taken from Glenn Walker and Jack Shostak, P.18): In our story, P is the null hypothesis, while Q is the extreme case: if H0 is true, most likely such extreme case should not happen; now we observe the extreme case (a t-value of 2.8728 in a t-distribution with degree of freedom of 43), then it is reasonable to question the null hypothesis itself. 12. To end this analysis, we can take a look at the output of SAS PROC TTEST: var score; freq count; run; The same result (P-value 0.0063 < 0.05, reject H0): # SAS Notes 12. In this example, input data take a form of cell count data, not as raw as the case-record data, so in both PROC MEANS and PROC TTEST, a “FREQ” statement added. 13. You can compute this P-value in step-6 with SAS using ProbT function: data; t    = 2.8727938571; df   = 43; tail = 2; P    = tail*(1-probt(t,df)); put P = ; run; According to the symmetry of t-distribution, you should multiple 2 to get the two-sided probability: 14. Long long ago, a so called “critical value” is calculated to support the decision. That’s the calculation method using TINV function: data; alpha = 0.05; tail  = 2; df    = 43; tCritic = tinv(1-alpha/tail,df); put tCritic=; run; We get this critical value of 2.0166921992: These two shaded zones are called “rejection zones”. The rule is, if the t-value we observed lies within the rejection zones, we should then reject the null hypothesis and take the alternative. In step-6, the t-value we calculated is 2.8728, which is bigger than this critical value, so we should reject H0. Critical value approach is equivalent with the p-value approach we discussed before, but slightly old fashioned. P-value is pretty intuitive and can be also easily calculated by computer. In days when computing power was not wildly available, people just took the t-distribution table to look up the critical values. Most statistical packages including SAS software don’t report the critical value (but still alive in statistics text books). # Reference Common Statistical Methods for Clinical Research with SAS Examples by Glenn Walker and Jack Shostak _Hypothesis Testing: A Primer_ (in Chinese) Two sides T-distribution calculator Single side t-distribution calculator Online Latex Equation Editor
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8741306662559509, "perplexity": 2159.2403874495717}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-30/segments/1531676591683.78/warc/CC-MAIN-20180720135213-20180720155213-00039.warc.gz"}
https://immunologyphd.hms.harvard.edu/browse/people?f%5B0%5D=sm_og_vocabulary%3Ataxonomy_term%3A121840&f%5B1%5D=sm_og_vocabulary%3Ataxonomy_term%3A121852&f%5B2%5D=sm_og_vocabulary%3Ataxonomy_term%3A121881&f%5B3%5D=sm_og_vocabulary%3Ataxonomy_term%3A121859&f%5B4%5D=sm_og_vocabulary%3Ataxonomy_term%3A121829
# Jack L. Strominger Higgins Professor of Biochemistry The study of histocompatibility in man and in other vertebrates led to the understanding of the mechanisms of immune recognition and to the discovery of...
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8771323561668396, "perplexity": 4434.130812557658}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656104495692.77/warc/CC-MAIN-20220707154329-20220707184329-00104.warc.gz"}
http://tex.stackexchange.com/questions/4152/how-do-i-prevent-widow-orphan-lines?answertab=active
How do I prevent widow/orphan lines? How do I prevent a line from appearing by itself: Orphan: at the bottom of the page, or Widow: at the top of the page? - As Brent points out, you cannot always do this. The best you can do is to tell TeX that it's infinitely bad for these to appear: \widowpenalty10000 \clubpenalty10000 One thing to keep in mind is that when presented with multiple infinitely bad options, TeX just picks one of them so you can still get widows or orphans. - I looked up somewhere else and they had \widowpenalty=10000. It didn't work then. Thanks! :) –  Kit Oct 16 '10 at 1:30 @Kit (it's a late comment, but anyway) You don't need the = in the assignment, but it doesn't hurt. \widowpenalty=10000 and \widowpenalty 10000 are the same. –  topskip Oct 4 '11 at 8:38 You can now use the nowidow package to make this task easier: \usepackage[all]{nowidow} - Where do I find nowindow.sty, I don't seem to have it in my standard (Mac) TeX install... Thanks –  Emit Taste Oct 25 at 9:03 The Memoir manual, in section 3.5 "Sloppybottom" discusses this in some detail, which I won't reproduce here. Be prepared even to re-word in the most intractable cases. Update: I think the specific commands like \enlargethispage and \sloppybottom are exclusively for the memoir package, but here's a snippet extracted from the aforementioned that you may care to adjust (you can see the extensive comments in the original): \clubpenalty=9996 \widowpenalty=9999 \brokenpenalty=4991 \predisplaypenalty=10000 \postdisplaypenalty=1549 \displaywidowpenalty=1602 Personally, I tend to avoid this TinXering with Plain TeX internals; although I don't know how to do it specifically for newlfm, I'd probably opt for adjusting the textheight on a case-by-case basis, as a final tidy-up before publishing. - I'm using newlfm. Is this applicable, too? –  Kit Oct 16 '10 at 1:18 while \sloppybottom is indeed memoir-specific (\raggedbottom is the comparable "plain" command), \enlargethispage is defined in base latex, so should be usable with any document class. –  barbara beeton Sep 27 '11 at 13:27 This FAQ answer gives some tips, including enlarging/reducing the (double-)page, setting the paragraph tighter, using \raggedbottom (for which, see also this FAQ answer which discusses putting some stretch in the \topskip).
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9577876329421997, "perplexity": 2708.8040008701546}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2013-48/segments/1387345758566/warc/CC-MAIN-20131218054918-00038-ip-10-33-133-15.ec2.internal.warc.gz"}
https://www.physicsforums.com/threads/rates-of-change.409808/
# Rates of Change 1. Jun 13, 2010 ### xX-Cyanide-Xx Hey everyone, im trying to study some questions from the internet and i came across this one. i think this will help me heeps with my study but i just cant figure it out. can anyone els??? ......................The picture of the graph is in the attachment........................... Using the graph above, find: a) The average rate of change from x = 2 to x = 4 b) The instantaneous rate of change of this function at x = -3 c) The above function is given by y = (x=3)(x - 1)(x - 4). Using your graphics calculator: I. What is the instantaneous rate of change at x = -2 II. Sketch your screen for how you got this solution. III. Write down your View Window settings. #### Attached Files: • ###### New Picture (1).png File size: 20.9 KB Views: 81 Last edited: Jun 13, 2010 2. Jun 13, 2010 ### Mentallic I can help you with the maths part of this problem, but not the calculator side of it. I can't see the attachment yet but it's not a problem. The average rate of change is given by the slope of the line between the points $$(2,f(2))$$ and $$(4,f(4))$$. The instantaneous rate of change is given by the tangent at the point $$(3,f(3))$$. 3. Jun 13, 2010 ### xX-Cyanide-Xx So they're both determind by the gradient. Are they??? 4. Jun 13, 2010 ### Mentallic Yes that's right. 5. Jun 13, 2010 ### xX-Cyanide-Xx yay. so i just need to find out how to use the graphics calculator to do the rest. but how will i do that? 6. Jun 13, 2010 ### Mentallic I don't see why you would even need to do those questions. You are studying on the net and if you're learning this for school, I doubt they will ever ask you to grab your graphics calculator and answer it in this way. So why even bother? 7. Jun 13, 2010 ### xX-Cyanide-Xx i guess its just so im prepared, if i can learn the hard stuff, i should be able to do everything they throw at me. andwho knows wat they might give us. 8. Jun 13, 2010 ### Mentallic This calculator business is hard stuff in its own way, not in the way a maths test works. If your school is anything like most other high schools out there, they won't be asking about how to use your graphics calculator. Unless... you do that regularly in class? Oh by the way, you need polishing up on those other skills before you should even be thinking of trying to learn about calculators. Calculators seriously aren't helping you at all, you need to become a little more independent with your maths. But if you're still intent on doing those questions, maybe someone else can help you because I can't. 9. Jun 13, 2010 ### xX-Cyanide-Xx Well, we have been using the graphics calculators alot in class, so the test does involve them. but thanks for the help ive learnt heeps....:)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4714275002479553, "perplexity": 914.0531247442043}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-13/segments/1521257650188.31/warc/CC-MAIN-20180324093251-20180324113251-00666.warc.gz"}
https://www.physicsforums.com/threads/unbanked-curve-problem.543022/
# Unbanked curve problem 1. Oct 22, 2011 ### physics_ 1. The problem statement, all variables and given/known data A block is hung by a string from the inside roof of a van. When the van goes straight ahead at a speed of 28m/s, the block hands vertically down. But when the van maintains this same speed around an unbanked curve (radius=150m), the block swings toward the outside of the curve. Then the stringmakes an angle A with the vertical. Find angle A. 2. Relevant equations F=mv^2/r??? FsMax = (mew)s Fn ??? Mew = coefficient of static friction. 3. The attempt at a solution Managed to find coefficient of static friction by equating the two equations above and substituting w=mg=Fn. Coefficient based on my calculatinos was 0.53. Didn't help me answer the question. Then I reasoned that based on a drawing, which may be flawed, that tanA = Fn/r. Fn= normal force which would lead to equating FsMax = Mv^2/r. Couldn't find the mass though. If I had found the mass, I could find one of sides of hte triangle the angle is in and since radius is given, it would be easy to determine angle A. Last edited: Oct 22, 2011 2. Oct 22, 2011 ### ehild When the block is stationary with respect to the van it moves around the same circle and with the same speed as the van. Where does the centripetal force come from? ehild 3. Oct 22, 2011 ### HallsofIvy Staff Emeritus And what does "static friction" have to do with this? I see no mention of friction in the problem. The block is not sitting on anything to have friction with. 4. Oct 22, 2011 ### physics_ Anyone have a solution? The answer in the book is 28 degrees. I was merely guessing with teh work I posted above. Last edited: Oct 22, 2011 5. Oct 22, 2011 ### ehild Make a drawing like the one attached, and find out what forces act on the hanging box. The resultant has to be the centripetal force. ehild #### Attached Files: • ###### vanbox.JPG File size: 4.7 KB Views: 111 6. Oct 22, 2011 ### physics_ I seem to have made the drawing fairly accurately already. The problem is the reasoning to solving the problem. 7. Oct 22, 2011 ### ehild You see the right triangle with sides mg and Fcp? What is tan(theta)? ehild Similar Discussions: Unbanked curve problem
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.86373370885849, "perplexity": 1821.6464106700341}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-47/segments/1510934808254.76/warc/CC-MAIN-20171124142303-20171124162303-00221.warc.gz"}
https://zbmath.org/?q=an:0779.35067
× # zbMATH — the first resource for mathematics Global solutions of systems of conservation laws by wave-front tracking. (English) Zbl 0779.35067 The author considers the Cauchy problem for a nonlinear strictly hyperbolic $$n\times n$$ system of conservation laws in one space variable $$u_ t+ F(u)x=0$$, $$x\in\mathbb{R}$$, $$t\geq 0$$; $$u(0,x)=v(x)$$ with sufficiently small total variation of $$v$$ and proves existence of weak global solutions satisfying entropy admissible conditions. Approximate solutions are constructed based on wave-front tracking – the approximate solutions are piecewise constant functions generated by solutions of Riemann problems, starting from piecewise constant approximation of $$v$$. It is proved, that total variation remains small and that the number of lines of discontinuities remains finite. Consequently a subsequence converges to a desired solution. Reviewer: A.Doktor (Praha) ##### MSC: 35L65 Hyperbolic conservation laws 35A05 General existence and uniqueness theorems (PDE) (MSC2000) 35A35 Theoretical approximation in context of PDEs Full Text: ##### References: [1] {\scA. Bressan}, A contractive metric for systems of conservation laws with coinciding shock and rarefaction curves, J. Differential Equations, in press. · Zbl 0802.35095 [2] DiPerna, R, Global existence of solutións to nonlinear hyperbolic systems of conservation laws, J. differential equations, 20, 187-212, (1976) · Zbl 0314.58010 [3] Glimm, J, Solutions in the large for nonlinear hyperbolic systems of equations, Comm. pure appl. math., 18, 95-105, (1965) · Zbl 0141.28902 [4] Lax, P, Hyperbolic systems of conservation laws, II, Comm. pure appl. math., 10, 537-567, (1957) · Zbl 0081.08803 [5] Lia, T.P, The deterministic version of the glimm scheme, Comm. math. phys., 57, 135-148, (1977) · Zbl 0376.35042 [6] Smoller, J, Shock waves and reaction-diffusion equations, (1983), Springer-Verlag New York · Zbl 0508.35002 [7] Rozdestvenskii, B.L; Yanenko, N, Systems of quasilinear equations, () [8] {\scB. Temple}, Systems of conservation laws with coinciding shock and rarefaction curves, in “Nonlinear Partial Differential Equations” (J. Smoller, Ed.), pp. 143-151, Contemporary Math Series, Vol. 17, Amer. Math. Soc., Providence, RI. · Zbl 0538.35050 This reference list is based on information provided by the publisher or from digital mathematics libraries. Its items are heuristically matched to zbMATH identifiers and may contain data conversion errors. It attempts to reflect the references listed in the original paper as accurately as possible without claiming the completeness or perfect precision of the matching.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8858626484870911, "perplexity": 1392.3159328157947}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780056900.32/warc/CC-MAIN-20210919190128-20210919220128-00613.warc.gz"}
https://www.hackmath.net/en/math-problem/6880
# Integer Find the integer whose distance on the numerical axis from number 1 is two times smaller as the distance from number 6. Result x =  -4 #### Solution: $|x-1|=|x-6|/2 \ \\ \ \\ x ... (-inf, 1) \ \\ \ \\ -(a-1)=-(a-6)/2 \ \\ \ \\ a=-4 \ \\ \ \\ a=-4 \ \\ \ \\ x ...... (1,6) \ \\ \ \\ (b-1)=-(b-6)/2 \ \\ \ \\ 3b=8 \ \\ \ \\ b=\dfrac{ 8 }{ 3 } \doteq 2.666667 \ \\ \ \\ x ...... (6,+inf) \ \\ \ \\ (c-1)=(c-6)/2 \ \\ \ \\ c=-4 \ \\ \ \\ c=-4 \ \\ \ \\ \ \\ x=a=(-4)=-4$ Our examples were largely sent or created by pupils and students themselves. Therefore, we would be pleased if you could send us any errors you found, spelling mistakes, or rephasing the example. Thank you! Leave us a comment of this math problem and its solution (i.e. if it is still somewhat unclear...): Be the first to comment! Tips to related online calculators Do you have a linear equation or system of equations and looking for its solution? Or do you have quadratic equation? ## Next similar math problems: 1. Intercept with axis F(x)=log(x+4)-2, what is the x intercept 2. Theorem prove We want to prove the sentence: If the natural number n is divisible by six, then n is divisible by three. From what assumption we started? 3. Ball game Richard, Denis and Denise together scored 932 goals. Denis scored 4 goals over Denise but Denis scored 24 goals less than Richard. Determine the number of goals for each player. 4. Stones 3 Simiyu and Nasike each collected a number of stones in an arithmetic lesson. If Simiyu gave Nasike 5 stones, Nasike would have twice as many stones as Simiyu. If initially, Simiyu had five stones less than Nasike how many stones did each have? 5. Football match 4 In a football match with the Italy lost 3 goals with Germans. Totally fell 5 goals in the match. Determine the number of goals of Italy and Germany. 6. Equation with abs value How many solutions has the equation ? in the real numbers? 7. Three workshops There are 2743 people working in three workshops. In the second workshop works 140 people more than in the first and in third works 4.2 times more than the second one. How many people work in each workshop? 8. Null points Calculate the roots of the equation: ? 9. Linsys2 Solve two equations with two unknowns: 400x+120y=147.2 350x+200y=144 10. Warehouses In the three warehouses, a total of 70 tons of grain was stored. In the second warehouse was stored 8.5t less and in the third 3.5t more than in the first. How many tons of grain was stored in each warehouse? 11. Men, women and children On the trip went men, women and children in the ratio 2:3:5 by bus. Children pay 60 crowns and adults 150. How many women were on the bus when a bus was paid 4,200 crowns? 12. Geometric sequence 5 About members of geometric sequence we know: ? ? Calculate a1 (first member) and q (common ratio or q-coefficient) 13. Children The group has 42 children. There are 4 more boys than girls. How many boys and girls are in the group? 14. Elimination method Solve system of linear equations by elimination method: 5/2x + 3/5y= 4/15 1/2x + 2/5y= 2/15 15. Solve 3 Solve quadratic equation: (6n+1) (4n-1) = 3n2 16. Legs Cancer has 5 pairs of legs. The insect has 6 legs. 60 animals have a total of 500 legs. How much more are cancers than insects? 17. Asymptote What is the vertical asymptote of ?
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 1, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5400117039680481, "perplexity": 2872.4123467437216}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875147628.27/warc/CC-MAIN-20200228170007-20200228200007-00233.warc.gz"}
https://www.physicsforums.com/threads/absence-of-gravity-at-atomic-level-at-all.118327/
# Absence of gravity at atomic level at all? 1. Apr 21, 2006 ### heartless Hello, I'm not familiar with what scientists working at accelerators say, but almost every book, and documentary movie say that Quantum Mechanics don't cover the gravity. I just gave a quick thought, well, maybe the gravity isn't there at all? Atomic level simply lacks the gravity and is only based upon 3 forces, electro-magnetism, strong and weak force? If the gravity, exists there, it must be very weak, but wouldn't atoms fall into the gravitational field of one another and start circling around each other? Hey, so how is it with this gravity at atomic level? And well, maybe I'm just out of time, and every one knows the answer except for me :) P.S, just a quick question, what keeps the electrons tied up to an atom? Thanks for all the help :shy: 2. Apr 21, 2006 ### chroot Staff Emeritus Gravity certainly exists at atomic scales, but its strength is negligible when compared to the other three forces. - Warren 3. Apr 21, 2006 ### Staff: Mentor Let's calculate the gravitational potential energy of two protons or neutrons (mass about $m = 2.7 \times 10^{-27}$ kg) in a nucleus, separated by about $r = 10^{-15}$ m. The gravitational constant is about $G = 6.7 \times 10^{-11}$N-m^2/kg^2. The potential energy is $$U = - \frac{Gm^2}{r}[/itex] which gives about $-4.9 \times 10^{-49}$ J, or about $-3.0 \times 10^{-30}$ eV. To put this in perspective, the binding energy per proton or neutron in most nuclei is in the general ballpark of a few million ($10^{+6}$) eV. So, unless gravity behaves wildly differently on a nuclear scale than on a macroscopic scale, its effects are insignificant compared to those of the other forces. Last edited: Apr 21, 2006 4. Apr 23, 2006 ### heartless Thanks, now let me ask you another question, suppose the gravity at atomic level is stronger than nuclear strong force. How would if affect atoms, and all the particles? 5. Apr 23, 2006 ### -Job- You used $r = 10^{-15}$. I don't know if, from a particle perspective, that is close or not, but the closer the better. If we were to take r arbitrarily close to 0 we would get a significant value. I'm not suggesting we take the limit of the gravitational potential as r goes to 0, but i wonder why you used the value of r that you used. How small can r be, in theory? Naturally, at this small scale things would have to be extremely close in order for gravity to have a chance, but in the event that they did, it sounds like it would be hard to get them apart. What are some of the main reasons why the nuclear forces are not caused by gravity? Last edited: Apr 23, 2006 6. Apr 26, 2006 ### dextercioby Well, the smaller "r" gets in gravity, the more you'd have to use GR... Daniel. 7. Apr 26, 2006 ### jtbell ### Staff: Mentor Atomic nuclei have diameters in the range of $10^{-15}$ to $10^{-14}$ m, so $10^{-15}$ is a reasonable order of magnitude estimate for the separation of two "neighboring" nucleons in a nucleus. Last edited: Apr 27, 2006 8. Apr 26, 2006 ### arivero More than "absence of gravity", I like to think of "presence of gauge forces". The point is, suppose you put a general cut-off asking that no force between two particles can be greater than the Force of Planck [tex]F_P \propto {\hbar c \over l_P^2$$ . If we had only gravity, this cut-off happens to be milder than the Planck Length cut-off because two electrons have a gravity force $$F_G=\hbar c ({m_e\over m_P})^2 {1\over r^2}$$ so they can approach a lot without violating the Planck Force. Fortunately they also happen to have an electromagnetic force $$F_E=\alpha(r) \hbar c {1\over r^2}$$ so that at distances of order Planck Length, the *electromagnetic* force is the one causing a force of order Planck Force, and then hitting the cut-off. Thus the gauge forces are needed to make the Force cutoff and the Distance cutoff compatible. Best said, Force cut-off + Gauge forces imply the length cut-off. Without gauge forces, the Force cut-off and the length (area, if you prefer) cutoff are two separate hypothesis. Last edited: Apr 26, 2006
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9704051613807678, "perplexity": 1502.6249161905944}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-26/segments/1529267863463.3/warc/CC-MAIN-20180620050428-20180620070428-00398.warc.gz"}
http://www.ntg.nl/pipermail/ntg-context/2008/036639.html
# [NTG-context] Times New Roman Alan Bowen acbowen at princeton.edu Mon Dec 1 15:51:09 CET 2008 ```That is good to know, I will use it in talking with the publisher. Many thanks. Alan On Dec 1, 2008, at 09;28,08 , Yue Wang wrote: > Termes is a TeX Gyre version of Times Roman (almost the same glyphs). > Times New Roman is a Microsoft version of Times Roman. > Times New Roman is more illegible than Times Roman. > > On Mon, Dec 1, 2008 at 10:19 PM, Alan Bowen <acbowen at princeton.edu> > wrote: >> Hi, Mojca! >> >> Thanks for the clarification. I had thought that I was getting Latin >> Modern instead but wanted to be sure. >> >> Termes sounds like a good alternative to me, but I had best consult >> the publisher, since he originally specified Times New Roman. It is >> apaprently no longer a typesetting problem. >> >> Again, thanks for your help. >> >> Alan >> >> On Dec 1, 2008, at 09;04,49 , Mojca Miklavec wrote: >> >>> On Mon, Dec 1, 2008 at 2:46 PM, Alan Bowen wrote: >>>> I am running the ConTeXt Minimals of 11/10 and am trying to access >>>> Times New Roman using the script >>>> >>>> \usetypescript [timesnewroman][ec] >>>> \definetypeface[timesnewroman][rm][serif][timesnewroman][default] >>>> [encoding=ec] >>>> \definebodyfont[9pt,11pt][rm][default] >>>> \setupbodyfont[timesnewroman,rm,11pt] >>>> >>>> I am not sure what I getting and I am puzzled by the log file , >>>> hence, >>>> my question: Is Times New Roman distributed with the minimals >>>> and is >>>> this the way to access it? >>> >>> Hello Alan, >>> >>> Do you need "something-like-times" or Times New Roman? >>> >>> This works: >>> % possibly \usetypescriptfile[type-gyr] >>> \usetypescript[times][ec] >>> \setupbodyfont[times,12pt] >>> >>> It will give you times-like typeface (TeX Gyre Termes), but not >>> Times >>> New Roman. Times New Roman is not distributied (and possibly >>> >>> Mojca >>> ___________________________________________________________________________________ >>> entry to the Wiki! >>> >>> maillist : ntg-context at ntg.nl / http://www.ntg.nl/mailman/listinfo/ntg-context >>> webpage : http://www.pragma-ade.nl / http://tex.aanhet.net >>> archive : https://foundry.supelec.fr/projects/contextrev/ >>> wiki : http://contextgarden.net >>> ___________________________________________________________________________________ >> >> ___________________________________________________________________________________ >> entry to the Wiki! >> >> maillist : ntg-context at ntg.nl / http://www.ntg.nl/mailman/listinfo/ntg-context >> webpage : http://www.pragma-ade.nl / http://tex.aanhet.net >> archive : https://foundry.supelec.fr/projects/contextrev/ >> wiki : http://contextgarden.net >> ___________________________________________________________________________________ >> > ___________________________________________________________________________________
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9270151853561401, "perplexity": 20830.287553607377}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-35/segments/1409535924501.17/warc/CC-MAIN-20140909014403-00467-ip-10-180-136-8.ec2.internal.warc.gz"}
https://math.uiowa.edu/calendar/algebra-seminar-113
# Algebra Seminar Speaker: Erich Jauch Topic: An Alternating Analogue of $U(\mathfrak{gl}_n)$ and Its Representations Abstract: The universal enveloping algebra of a Lie algebra $\mathfrak{g}$ is of utmost importance when studying representations of $\mathfrak{g}$. In 2010, V. Futorny and S. Ovsienko gave a realization of $U(\mathfrak{gl}_n)$ as a subalgebra of the ring of invariants of a certain noncommutative ring with respect to the action of $S_1\times S_2\times\cdots\times S_n$ where $S_j$ is the symmetric group on $j$ variables. With some connections to Galois Theory, an interesting question is what would a similar object be in the invariant ring with respect to a product of Alternating groups? We will discuss such an object and some results about its representations and how they relate to $U(\mathfrak{gl}_n)$ representations. This is based on the work in my recent preprint arXiv:1907.13254 Event Date: October 21, 2019 - 3:30pm to 4:20pm Location: 217 MLH Calendar Category: Seminar Seminar Category: Algebra
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.827111005783081, "perplexity": 248.9143879151092}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439736057.87/warc/CC-MAIN-20200810145103-20200810175103-00073.warc.gz"}
https://answers.opencv.org/question/33452/how-do-i-calculates-the-angle-of-orientation-of-the-image-with-the-fourier-transform/?answer=58579
# How Do I Calculates The Angle Of Orientation Of The Image With The Fourier Transform? edit #include "opencv2/imgproc/imgproc.hpp" #include "opencv2/highgui/highgui.hpp" #include <stdlib.h> #include <stdio.h> #include <opencv2\opencv.hpp> #include <opencv2/highgui/highgui.hpp> #include <opencv2/core/core.hpp> #include <stdint.h> #include <opencv2/imgproc/imgproc.hpp> #include <iostream> using namespace cv; using namespace std; /// Global variables int edgeThresh = 1; int lowThreshold; int const max_lowThreshold = 100; int ratio = 3; int kernel_size = 3; char* window_name = "Edge Map"; void CannyThreshold(int, void*) { blur( src_gray, detected_edges, Size(3,3) ); Canny( detected_edges, detected_edges, lowThreshold, lowThreshold*ratio, kernel_size ); dst = Scalar::all(0); src.copyTo( dst, detected_edges); imshow( window_name, dst ); int m = getOptimalDFTSize( dst.rows ); int n = getOptimalDFTSize( dst.cols ); copyMakeBorder(dst, padded, 0, m - dst.rows, 0, n - dst.cols, BORDER_CONSTANT, Scalar::all(0)); Mat complexdst; merge(planes, 2, complexdst); dft(complexdst, complexdst); split(complexdst, planes); magnitude(planes[0], planes[1], planes[0]); Mat magdst = planes[0]; magdst += Scalar::all(1); log(magdst, magdst); magdst = magdst(Rect(0, 0, magdst.cols & -2, magdst.rows & -2)); int cx = magdst.cols/2; int cy = magdst.rows/2; Mat q0(magdst, Rect(0, 0, cx, cy)); Mat q1(magdst, Rect(cx, 0, cx, cy)); Mat q2(magdst, Rect(0, cy, cx, cy)); Mat q3(magdst, Rect(cx, cy, cx, cy)); Mat tmp; q0.copyTo(tmp); q3.copyTo(q0); tmp.copyTo(q3); q1.copyTo(tmp); q2.copyTo(q1); tmp.copyTo(q2); normalize(magdst, magdst, 0, 1, CV_MINMAX); imshow("spectrum magnitude", magdst); } int main( int argc, char** argv ) { resize(src, src, Size(), 0.2, 0.2, INTER_LANCZOS4); if( !src.data ) { return -1; } dst.create( src.size(), src.type() ); cvtColor( src, src_gray, CV_BGR2GRAY ); namedWindow( window_name, CV_WINDOW_AUTOSIZE ); createTrackbar( "Min Threshold:", window_name, &lowThreshold, max_lowThreshold, CannyThreshold ); CannyThreshold(0, 0); waitKey(0); return 0; } edit retag close merge delete Hello, I have an image I applied the fourier transform for the orientation angle,but i face here channels problem.When i build this program,it shows success but it's not run.It is showing "Open cv Error: Assertion failed <_tp>::channels==m.channels<>>in unknown function, file c:\opencv\opencv\build\include\opencv2/core/mat.hpp,line890" i used Breakpoint in this code when it comes this line "Mat planes[] = {Mat_(padded), Mat::zeros(padded.size(), CV_32F)};"then it break.Personally I do not know how i solve this problem because I have no error! Please help me ( 2014-05-15 18:58:29 -0500 )edit Sort by » oldest newest most voted The error you got was due to the fact that you worked with src image, which is a multi channels image. If you change the line: src.copyTo( dst, detected_edges); to: src_gray.copyTo( dst, detected_edges); the error gone. In case you wish to do the dft with a multi channels image, you have to proceed with each channel individually, and merge the results to get what you want. Plus, you should check if the image was opened before carrying out any operation on it: if( !src.data ) { return -1; } resize(src, src, Size(), 0.2, 0.2, INTER_LANCZOS4); more So now the OP has a DFT that works (as I do), how do you find the angle from the result? ( 2015-03-30 12:52:04 -0500 )edit 1 For computing the orientation part, right after the split statement: split(complexdst, planes); phase(planes[0], planes[1], phasePart); Then, the other needed code is the same as you did with the magnitude part. Additionally, do not overwrite the planes[0] component as you wrote: magnitude(planes[0], planes[1], planes[0]); Instead, you should do like that: magnitude(planes[0], planes[1], magPart); ( 2015-03-31 00:02:39 -0500 )edit OK, pulled out phasePart, but it is still just snow (see above). I am thinking of doing a polar histogram on just magnitude of one quadrant, the biggest pile will be in the direction I care about (maybe). ( 2015-03-31 16:01:01 -0500 )edit Well, the Assertion is saying that the number of channels is wrong. Best guess is that you still have color when you need to go either grey (or extract one of the BGR components). I thought the imread should have handled that but the flag to convert to grey is not explicit. Try src = imread("Filename", 0); //0 for grey, 1 for color. The step you reference (line numbers please!) is where the complex dimension gets added to the image data. It looks like you copied the tutorial verbatim. Is padded bigger than I at this point as it should be? Otherwise, I'm just trying to read through the tutorial. It looks like you copied that anyway, why not run that with your image? Then let me know how you get the angle! EDIT I think the spirit of the original question was how to interpret results like these: EDIT Pulling out phase as suggested. Similar quadrant swap but not log scale of course. more The best way is to create a orientated band pass filter and filter the image with it. I made an online app where you can play around with FFT filters called 'Fourifier' http://www.ejectamenta.com/Imaging-Ex... This approach is also used with Steerable Filters, see the work of Freeman, Adelson, Castleman for example http://people.csail.mit.edu/billf/pub... http://www.tka4.org/materials/lib/Art... You can also get the orientations directly from a steerable filter, look at page 17 of http://persci.mit.edu/pub_pdfs/freema... This is what I did with them, before I ran out of postdoc funding! http://www.ejectamenta.com/Imaging-Ex... more Official site GitHub Wiki Documentation
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.15038800239562988, "perplexity": 9048.520999510707}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046154032.75/warc/CC-MAIN-20210730220317-20210731010317-00643.warc.gz"}
https://www.aimsciences.org/article/doi/10.3934/dcdss.2017028
American Institute of Mathematical Sciences • Previous Article Quasi-periodic solutions for a class of second order differential equations with a nonlinear damping term • DCDS-S Home • This Issue • Next Article Stability of non-monotone non-critical traveling waves in discrete reaction-diffusion equations with time delay June  2017, 10(3): 557-579. doi: 10.3934/dcdss.2017028 Exponential stability of 1-d wave equation with the boundary time delay based on the interior control Department of Mathematics, Tianjin University, Haihe Education Park, Tianjin, Tianjin 300350, China Received  June 2016 Revised  January 2017 Published  February 2017 Fund Project: This work is supported by Science Foundation of China under Grant Nos.61174080, 61503275 and 61573252 In this paper, the stability problem of 1-d wave equation with the boundary delay and the interior control is considered. The well-posedness of the closed-loop system is investigated by the linear operator. Based on the idea of Lyapunov functional technology, we give the condition on the relationship between the control parameter α and the delay parameter k to guarantee the exponential stability of the system. Citation: Yaru Xie, Genqi Xu. Exponential stability of 1-d wave equation with the boundary time delay based on the interior control. Discrete & Continuous Dynamical Systems - S, 2017, 10 (3) : 557-579. doi: 10.3934/dcdss.2017028 References: show all references References: [1] Serge Nicaise, Cristina Pignotti, Julie Valein. Exponential stability of the wave equation with boundary time-varying delay. Discrete & Continuous Dynamical Systems - S, 2011, 4 (3) : 693-722. doi: 10.3934/dcdss.2011.4.693 [2] Martin Gugat, Günter Leugering, Ke Wang. Neumann boundary feedback stabilization for a nonlinear wave equation: A strict $H^2$-lyapunov function. Mathematical Control & Related Fields, 2017, 7 (3) : 419-448. doi: 10.3934/mcrf.2017015 [3] Serge Nicaise, Cristina Pignotti. Stability of the wave equation with localized Kelvin-Voigt damping and boundary delay feedback. Discrete & Continuous Dynamical Systems - S, 2016, 9 (3) : 791-813. doi: 10.3934/dcdss.2016029 [4] Bopeng Rao, Laila Toufayli, Ali Wehbe. Stability and controllability of a wave equation with dynamical boundary control. Mathematical Control & Related Fields, 2015, 5 (2) : 305-320. doi: 10.3934/mcrf.2015.5.305 [5] Le Thi Phuong Ngoc, Nguyen Thanh Long. Existence and exponential decay for a nonlinear wave equation with nonlocal boundary conditions. Communications on Pure & Applied Analysis, 2013, 12 (5) : 2001-2029. doi: 10.3934/cpaa.2013.12.2001 [6] Luis Barreira, Claudia Valls. Delay equations and nonuniform exponential stability. Discrete & Continuous Dynamical Systems - S, 2008, 1 (2) : 219-223. doi: 10.3934/dcdss.2008.1.219 [7] Sigurdur Freyr Hafstein. A constructive converse Lyapunov theorem on exponential stability. Discrete & Continuous Dynamical Systems - A, 2004, 10 (3) : 657-678. doi: 10.3934/dcds.2004.10.657 [8] Nicolas Fourrier, Irena Lasiecka. Regularity and stability of a wave equation with a strong damping and dynamic boundary conditions. Evolution Equations & Control Theory, 2013, 2 (4) : 631-667. doi: 10.3934/eect.2013.2.631 [9] Andrzej Nowakowski. Variational approach to stability of semilinear wave equation with nonlinear boundary conditions. Discrete & Continuous Dynamical Systems - B, 2014, 19 (8) : 2603-2616. doi: 10.3934/dcdsb.2014.19.2603 [10] István Györi, Ferenc Hartung. Exponential stability of a state-dependent delay system. Discrete & Continuous Dynamical Systems - A, 2007, 18 (4) : 773-791. doi: 10.3934/dcds.2007.18.773 [11] Yanni Guo, Genqi Xu, Yansha Guo. Stabilization of the wave equation with interior input delay and mixed Neumann-Dirichlet boundary. Discrete & Continuous Dynamical Systems - B, 2016, 21 (8) : 2491-2507. doi: 10.3934/dcdsb.2016057 [12] Stéphane Gerbi, Belkacem Said-Houari. Exponential decay for solutions to semilinear damped wave equation. Discrete & Continuous Dynamical Systems - S, 2012, 5 (3) : 559-566. doi: 10.3934/dcdss.2012.5.559 [13] Tomás Caraballo, José Real, T. Taniguchi. The exponential stability of neutral stochastic delay partial differential equations. Discrete & Continuous Dynamical Systems - A, 2007, 18 (2&3) : 295-313. doi: 10.3934/dcds.2007.18.295 [14] Xiang Xie, Honglei Xu, Xinming Cheng, Yilun Yu. Improved results on exponential stability of discrete-time switched delay systems. Discrete & Continuous Dynamical Systems - B, 2017, 22 (1) : 199-208. doi: 10.3934/dcdsb.2017010 [15] Ismael Maroto, Carmen Núñez, Rafael Obaya. Exponential stability for nonautonomous functional differential equations with state-dependent delay. Discrete & Continuous Dynamical Systems - B, 2017, 22 (8) : 3167-3197. doi: 10.3934/dcdsb.2017169 [16] Linfang Liu, Tomás Caraballo, Xianlong Fu. Exponential stability of an incompressible non-Newtonian fluid with delay. Discrete & Continuous Dynamical Systems - B, 2018, 23 (10) : 4285-4303. doi: 10.3934/dcdsb.2018138 [17] Ábel Garab. Unique periodic orbits of a delay differential equation with piecewise linear feedback function. Discrete & Continuous Dynamical Systems - A, 2013, 33 (6) : 2369-2387. doi: 10.3934/dcds.2013.33.2369 [18] Leif Arkeryd, Raffaele Esposito, Rossana Marra, Anne Nouri. Exponential stability of the solutions to the Boltzmann equation for the Benard problem. Kinetic & Related Models, 2012, 5 (4) : 673-695. doi: 10.3934/krm.2012.5.673 [19] Michael Scheutzow. Exponential growth rate for a singular linear stochastic delay differential equation. Discrete & Continuous Dynamical Systems - B, 2013, 18 (6) : 1683-1696. doi: 10.3934/dcdsb.2013.18.1683 [20] Stéphane Junca, Bruno Lombard. Stability of neutral delay differential equations modeling wave propagation in cracked media. Conference Publications, 2015, 2015 (special) : 678-685. doi: 10.3934/proc.2015.0678 2018 Impact Factor: 0.545
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.468166321516037, "perplexity": 4158.468890356327}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-35/segments/1566027314638.49/warc/CC-MAIN-20190819011034-20190819033034-00394.warc.gz"}
https://www.famnit.upr.si/sl/seminars/1510
Univerza na Primorskem Fakulteta za matematiko, naravoslovje in informacijske tehnologije SI | EN # Raziskovalni matematični seminar - Arhiv 2020 2019 2018 2017 2016 2015 2014 2013 2012 2011 2010 1 2 3 4 5 6 7 8 9 10 11 12 Datum in ura / Date and time: 2.11.15 (10:00 - 11:00) Predavalnica / Location: FAMNIT-POŠTA Predavatelj / Lecturer: Štefko Miklavič (UP IAM, UP FAMNIT) Naslov / Title: Distance-regular cayley graphs Vsebina / Abstract: In this talk we consider the following problem: for a given class of groups ${\cal G}$, classify distance-regular Cayley graphs $Cay(G;S)$, where $G \in {\cal G}$. Datum in ura / Date and time: 26.10.15 (11:00-12:00) Predavalnica / Location: Famnit-VP Predavatelj / Lecturer: Wilfried Imrich (Department Mathematics and Information Technology, Montanuniversität Leoben, Austria) Naslov / Title: Symmetry breaking in graphs Vsebina / Abstract: In a graph, a set of vertices that is stabilized setwise by only the trivial automorphism is called a distinguishing set. Tom Tucker conjectured that every connected, infinite locally finite graph G has such a set if each nontrivial automorphism of G moves infinitely many vertices. The conjecture is know as the Infinite Motion Conjecture, which is still open despite the fact that numerous large classes of graphs have been shown to satisfy it. In finite graphs distinguishing sets, if they exist, can be very small in comparison to the size of the graph, and in infinite graphs such sets can be finite. If they are not finite, their density can be zero. This talk introduces to  the subject and presents new results about the density of distinguishing sets and the  Infinite Motion Conjecture. Datum in ura / Date and time: 26.10.15 (10:00-11:00) Predavalnica / Location: Famnit-VP Predavatelj / Lecturer: Mats Gyllenberg (Finnish Centre of Excellence in Analysis and Dynamics Research, Department of Mathematics and Statistics, University of Helsinki) Naslov / Title: Finite dimensional state representation of nonlinear delay systems Vsebina / Abstract: In this talk I give necessary and sufficient conditions for when a system of nonlinear delay equations can be reduced to a system of ordinary differential equations. The class of systems considered contains nonlinear models of physiologically structured populations. My talk is based on joint work (in progress) with Odo Diekmann and Hans Metz. Datum in ura / Date and time: 19.10.15 (10:00 - 11:00) Predavalnica / Location: FAMNIT-POSTA Predavatelj / Lecturer: Dragan Marušič (UP FAMNIT, UP IAM) Naslov / Title: Odd automorphisms Vsebina / Abstract: An automorphism (or symmetry) of a combinatorial graph may be called even or odd according to whether it acts as an even or odd permutation on the vertices of the graph. In this talk, I will present a partial result about the following question: When does the existence of even automorphisms of a graph force existence of odd automorphisms? In particular, I will present complete information the on existence of odd automorphisms for cubic symmetric graphs. Datum in ura / Date and time: 12.10.15 (10:00-11:00) Predavalnica / Location: FAMNIT-POSTA Predavatelj / Lecturer: Stephen E. Wilson (UP IAM, UP FAMNIT) Naslov / Title: Operartors on Maniplexes (Part II) Vsebina / Abstract: 'Maniplex' is a generalization of 'polytope' and is at the same time a generalization of 'map'.  We are interested in those which have large symmetry groups, and these come in two kinds, 'rotary' and 'reflexible', both called 'regular' by some faction of the community.  In this talk, we will introduce manilexes, talk about the motivations for the definition, give some basic examples of several dimensions, and discuss the distinction between maniplexes and polytopes.  Our main goal is to understand a handful of operators (Dual, Petrie, Bubble, Twist) which transform one regular maniplex into another. Datum in ura / Date and time: 5.10.15 (10:00-11:00) Predavalnica / Location: FAMNIT-POSTA Predavatelj / Lecturer: Stephen E. Wilson (UP IAM, UP FAMNIT) Naslov / Title: Operartors on Maniplexes Vsebina / Abstract: 'Maniplex' is a generalization of 'polytope' and is at the same time a generalization of 'map'.  We are interested in those which have large symmetry groups, and these come in two kinds, 'rotary' and 'reflexible', both called 'regular' by some faction of the community.  In this talk, we will introduce manilexes, talk about the motivations for the definition, give some basic examples of several dimensions, and discuss the distinction between maniplexes and polytopes.  Our main goal is to understand a handful of operators (Dual, Petrie, Bubble, Twist) which transform one regular maniplex into another. fa
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7558192610740662, "perplexity": 2639.79757846661}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-50/segments/1606141672314.55/warc/CC-MAIN-20201201074047-20201201104047-00556.warc.gz"}
https://arxiv.org/abs/0709.3107
astro-ph (what is this?) # Title: Formation of Kuiper Belt Binaries Abstract: The discovery that a substantial fraction of Kuiper Belt objects (KBOs) exists in binaries with wide separations and roughly equal masses, has motivated a variety of new theories explaining their formation. Goldreich et al. (2002) proposed two formation scenarios: In the first, a transient binary is formed, which becomes bound with the aid of dynamical friction from the sea of small bodies (L^2s mechanism); in the second, a binary is formed by three body gravitational deflection (L^3 mechanism). Here, we accurately calculate the L^2s and L^3 formation rates for sub-Hill velocities. While the L^2s formation rate is close to previous order of magnitude estimates, the L^3 formation rate is about a factor of 4 smaller. For sub-Hill KBO velocities (v << v_H) the ratio of the L^3 to the L^2s formation rate is 0.05 (v/v_H) independent of the small bodies' velocity dispersion, their surface density or their mutual collisions. For Super-Hill velocities (v >> v_H) the L^3 mechanism dominates over the L^2s mechanism. Binary formation via the L^3 mechanism competes with binary destruction by passing bodies. Given sufficient time, a statistical equilibrium abundance of binaries forms. We show that the frequency of long-lived transient binaries drops exponentially with the system's lifetime and that such transient binaries are not important for binary formation via the L^3 mechanism, contrary to Lee et al. (2007). For the L^2s mechanism we find that the typical time, transient binaries must last, to form Kuiper Belt binaries (KBBs) for a given strength of dynamical friction, D, increases only logarithmically with D. Longevity of transient binaries only becomes important for very weak dynamical friction (i.e. D \lesssim 0.002) and is most likely not crucial for KBB formation. Comments: 20 pages, 3 figures, Accepted for publication in ApJ, correction of minor typos Subjects: Astrophysics (astro-ph) DOI: 10.1086/524930 Cite as: arXiv:0709.3107 [astro-ph] (or arXiv:0709.3107v2 [astro-ph] for this version) ## Submission history From: Hilke Schlichting [view email] [v1] Wed, 19 Sep 2007 21:24:46 GMT (20kb) [v2] Tue, 6 Nov 2007 19:40:39 GMT (19kb)
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8509989380836487, "perplexity": 5090.3426955830755}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-26/segments/1498128321938.75/warc/CC-MAIN-20170627221726-20170628001726-00337.warc.gz"}
http://mathinsight.org/applet/angled_line_or_plane
# Math Insight ### Applet: An angled line or a plane The graph of the equation $y=3x-2$ looks like a line, which it would be if were an equation in two dimensions, i.e., in the $xy$-plane. However, rotate the graph with the mouse to give you a new perspective on the graph. Since in this case, the graph is really in three dimensions, the graph of the equation becomes a plane. The equation just happens to not depend on the variable $z$, so the plane is parallel to the $z$-axis.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9075852632522583, "perplexity": 164.1141312252917}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-30/segments/1500549423903.35/warc/CC-MAIN-20170722062617-20170722082617-00344.warc.gz"}
https://ftp.aimsciences.org/article/doi/10.3934/dcds.2015.35.4439
# American Institute of Mathematical Sciences September  2015, 35(9): 4439-4453. doi: 10.3934/dcds.2015.35.4439 ## Dynamic programming using radial basis functions 1 Center for Mathematics, Technische Universität München, 85747 Garching bei München, Germany, Germany Received  May 2014 Revised  October 2014 Published  April 2015 We propose a discretization of the optimality principle in dynamic programming based on radial basis functions and Shepard's moving least squares approximation method. We prove convergence of the value iteration scheme, derive a statement about the stability region of the closed loop system using the corresponding approximate optimal feedback law and present several numerical experiments. Citation: Oliver Junge, Alex Schreiber. Dynamic programming using radial basis functions. Discrete and Continuous Dynamical Systems, 2015, 35 (9) : 4439-4453. doi: 10.3934/dcds.2015.35.4439 ##### References: show all references ##### References: [1] Martin D. Buhmann, Slawomir Dinew. Limits of radial basis function interpolants. Communications on Pure and Applied Analysis, 2007, 6 (3) : 569-585. doi: 10.3934/cpaa.2007.6.569 [2] Runchang Lin, Huiqing Zhu. A discontinuous Galerkin least-squares finite element method for solving Fisher's equation. Conference Publications, 2013, 2013 (special) : 489-497. doi: 10.3934/proc.2013.2013.489 [3] Peter Giesl. Construction of a global Lyapunov function using radial basis functions with a single operator. Discrete and Continuous Dynamical Systems - B, 2007, 7 (1) : 101-124. doi: 10.3934/dcdsb.2007.7.101 [4] JaEun Ku. Maximum norm error estimates for Div least-squares method for Darcy flows. Discrete and Continuous Dynamical Systems, 2010, 26 (4) : 1305-1318. doi: 10.3934/dcds.2010.26.1305 [5] H. D. Scolnik, N. E. Echebest, M. T. Guardarucci. Extensions of incomplete oblique projections method for solving rank-deficient least-squares problems. Journal of Industrial and Management Optimization, 2009, 5 (2) : 175-191. doi: 10.3934/jimo.2009.5.175 [6] Yongjian Yang, Zhiyou Wu, Fusheng Bai. A filled function method for constrained nonlinear integer programming. Journal of Industrial and Management Optimization, 2008, 4 (2) : 353-362. doi: 10.3934/jimo.2008.4.353 [7] Sigurdur Hafstein, Skuli Gudmundsson, Peter Giesl, Enrico Scalas. Lyapunov function computation for autonomous linear stochastic differential equations using sum-of-squares programming. Discrete and Continuous Dynamical Systems - B, 2018, 23 (2) : 939-956. doi: 10.3934/dcdsb.2018049 [8] Rein Luus. Optimal control of oscillatory systems by iterative dynamic programming. Journal of Industrial and Management Optimization, 2008, 4 (1) : 1-15. doi: 10.3934/jimo.2008.4.1 [9] Yunhai Xiao, Soon-Yi Wu, Bing-Sheng He. A proximal alternating direction method for $\ell_{2,1}$-norm least squares problem in multi-task feature learning. Journal of Industrial and Management Optimization, 2012, 8 (4) : 1057-1069. doi: 10.3934/jimo.2012.8.1057 [10] Zhuoyi Xu, Yong Xia, Deren Han. On box-constrained total least squares problem. Numerical Algebra, Control and Optimization, 2020, 10 (4) : 439-449. doi: 10.3934/naco.2020043 [11] Xiao-Wen Chang, David Titley-Peloquin. An improved algorithm for generalized least squares estimation. Numerical Algebra, Control and Optimization, 2020, 10 (4) : 451-461. doi: 10.3934/naco.2020044 [12] Cheng Ma, Xun Li, Ka-Fai Cedric Yiu, Yongjian Yang, Liansheng Zhang. On an exact penalty function method for semi-infinite programming problems. Journal of Industrial and Management Optimization, 2012, 8 (3) : 705-726. doi: 10.3934/jimo.2012.8.705 [13] Sohana Jahan, Hou-Duo Qi. Regularized multidimensional scaling with radial basis functions. Journal of Industrial and Management Optimization, 2016, 12 (2) : 543-563. doi: 10.3934/jimo.2016.12.543 [14] Quyen Tran, Harbir Antil, Hugo Díaz. Optimal control of parameterized stationary Maxwell's system: Reduced basis, convergence analysis, and a posteriori error estimates. Mathematical Control and Related Fields, 2022  doi: 10.3934/mcrf.2022003 [15] Tengfei Yan, Qunying Liu, Bowen Dou, Qing Li, Bowen Li. An adaptive dynamic programming method for torque ripple minimization of PMSM. Journal of Industrial and Management Optimization, 2021, 17 (2) : 827-839. doi: 10.3934/jimo.2019136 [16] Ryan Loxton, Qun Lin. Optimal fleet composition via dynamic programming and golden section search. Journal of Industrial and Management Optimization, 2011, 7 (4) : 875-890. doi: 10.3934/jimo.2011.7.875 [17] Guy Barles, Ariela Briani, Emmanuel Trélat. Value function for regional control problems via dynamic programming and Pontryagin maximum principle. Mathematical Control and Related Fields, 2018, 8 (3&4) : 509-533. doi: 10.3934/mcrf.2018021 [18] Hassan Mohammad, Mohammed Yusuf Waziri, Sandra Augusta Santos. A brief survey of methods for solving nonlinear least-squares problems. Numerical Algebra, Control and Optimization, 2019, 9 (1) : 1-13. doi: 10.3934/naco.2019001 [19] Ya-Xiang Yuan. Recent advances in numerical methods for nonlinear equations and nonlinear least squares. Numerical Algebra, Control and Optimization, 2011, 1 (1) : 15-34. doi: 10.3934/naco.2011.1.15 [20] Mila Nikolova. Analytical bounds on the minimizers of (nonconvex) regularized least-squares. Inverse Problems and Imaging, 2008, 2 (1) : 133-149. doi: 10.3934/ipi.2008.2.133 2021 Impact Factor: 1.588
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.2921209931373596, "perplexity": 6677.595173399294}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656104244535.68/warc/CC-MAIN-20220703134535-20220703164535-00162.warc.gz"}
https://bio.libretexts.org/Courses/University_of_California_Davis/BIS_2B%3A_Introduction_to_Biology_-_Ecology_and_Evolution/2%3A_Biodiversity/2.2_Measuring_Species_Diversity
# 2.2 Measuring Species Diversity How do we measure species diversity within a habitat? How do we compare diversity across different types of habitats containing very different numbers and types of organisms? There are many mathematical models that have been developed to quantify species diversity in different habitats. While these models differ in the exact method of diversity estimation, they all include two important components: species richness and species evenness. Species richness is a measure of the number of different types of species in an ecosystem. A large number of different species in a habitat represents a higher species richness, and an overall more diverse ecosystem. Species evenness is a measure of the relative abundance of each species. More evenly represented species (evidenced by similar population sizes) illustrate a higher species evenness and an overall more diverse ecosystem. Mathematically, we can distill species richness and species evenness of a habitat into a single measure of overall diversity using the following equation $D = (p_1^{-p1}) (p_2^{-p2}) (p_3^{-p3})…(p_n^{-n})$ where D is a measure of the total diversity of the ecosystem, and pn is the proportion of species n.  You will practice calculating D values and comparing diversity for different ecosystems in Lab 1: Discovering Diversity. Exercise $$\PageIndex{1}$$ For the following exercise, consider the four insect communities shown below. Fictional insect communities in four different habitats. Figure created by L Gerhart-Barley with biorender.com To determine the diversity value for a given habitat, we must first calculate the number and proportion of species in the habitat and then enter those numbers into the equation provided above. Habitat A: Habitat A has two species, blue insects and red insects. There are 10 total insects in this habitat, 1 of which is blue and 9 of which are red. Consequently, the proportion of blue insects is 1/10 or 0.1 and the proportion of red insects is 9/10 or 0.9. The D value for Habitat A is then D = (0.1-0.1)(0.9-0.9) = 1.384 Calculate the diversity values for each of the other habitats to determine which habitat has the highest diversity. Habitat B: D = (0.1-0.1) (0.1-0.1) (0.1-0.1) (0.3-0.3) (0.4-0.4) = 4.13 Habitat C: D = (0.5-0.5) (0.5-0.5) = 2 Habitat D: D = (0.2-0.2) (0.2-0.2) (0.2-0.2) (0.2-0.2) (0.2-0.2) = 5 Habitat D has the highest diversity The methodology above is easy to use with small ecosystems that contain a small number of individuals and only a few species. Real ecosystems, however, are very large and can contain hundreds or thousands of species and tens of thousands of individual organisms. In such cases, it is not possible to count every single individual and determine the exact proportion of each species in the entire ecosystem. Instead, we need a method to determine how much sampling is required in order to estimate species diversity with a reasonable degree of accuracy. In this section, we will consider two such methods: species-area curves and rarefaction curves. Species-area curves show the relationship between the area sampled in an ecosystem (on the x axis) and the number of species found in that area (on the y axis). Figure 1 illustrates how a species-area curve is developed. In the figure, the top box represents the ecosystem being studied, with letters representing individuals of different species. Suppose the researcher begins in the top left corner of the habitat and counts the number of individuals in the 1-meter square box shaded red. There are two individuals, both of Species A, in this area. As the researcher expands the area sampled (into the orange, then yellow, then green, etc areas) the number of individuals counted increases, as does the number of species identified. A summary table of the total area sampled and number of species identified is provided next to the map. In this particular sampling method, the area sampled doubles at each step: the red box is 1 m2, sampling the orange box brings the total area 2 m2, the yellow box to 4 m2, the green box to 8 m2 and so on. The number of species identified in the habitat follows a slightly different pattern. Initially, the number of species found increases rapidly, however, few new species are found as the area increases, to the point that the line in the graph begins to taper off after about 16 m2. The resulting species-area curve suggests that the researcher has likely already found all (or at least the vast majority of) species present in this ecosystem and that additional sampling would not discover more species. Figure $$\PageIndex{1}$$: Left: Map of a fictional ecosystem to be sampled with letters denoting individuals of different species; Middle: A summary table of area sampled and number of species found; Right: A species-area curve developed from the summary table. Colored boxes in the ecosystem correspond to colored bars in the graph. Figure by L Gerhart-Barley A similar method for estimating ecosystem diversity is a rarefaction curve, which is similar to a species-area curve, but focuses on the number of individuals sampled as opposed to the area. Figure 2.2.3 illustrates how a rarefaction curve is developed. Suppose another researcher samples the same area shown in Figure 2 and also begins in the top left corner, following the sampling path represented by the gray arrows. The researcher works their way through the ecosystem noting each individual organism and whether or not it represents a species already found in the ecosystem. A summary table of the number of individuals and species counted is provided next to the map. The number of individuals counted progresses rapidly, while the number of species identified tapers off after about 45 individuals sampled. The resulting rarefaction curve suggests that the researcher has likely already found all (or at least the vast majority of) species present in this ecosystem and that additional sampling would not discover more species. Figure $$\PageIndex{2}$$: Left: Map of a fictional ecosystem to be sampled with letters denoting individuals of different species; Middle: A summary table of number of individuals sampled and number of species found; Right: A rarefaction curve developed from the summary table. Figure by L Gerhart-Barley The species-area curve and rarefaction curves look quite similar, and indeed they should, since the primary difference in the two methodologies is how we define sampling ‘effort’ on the x axis – through area surveyed (species-area curve) or through number of individuals counted (rarefaction curve). In both cases, the researcher sampled the entire area shown in the map, which should result in a similar estimate of species diversity for the ecosystem. Which sampling method is used depends on the characteristics of the ecosystem and species to be sampled.  You will practice sampling the diversity of an ecosystem and developing a rarefaction curve of your data in Lab 1: Discovering Diversity. ## Earth's Biodiversity How many species exist on Earth? This is a question that scientists have attempted to answer for centuries, beginning with Carl Linnaeus’ first steps towards naming and classifying organisms in Systema Naturae in 1735. Linneaus himself described over 12,000 species and in the nearly 300 years since, scientists have formally described and studied over a million more. Have scientists already discovered all, or most, of the species that live on Earth? If not, how many more remain to be discovered? These questions are inherently difficult to answer because they require us to estimate how much we don’t know; how many species scientists have not discovered. This difficulty is compounded by the fact that certain regions of Earth and certain group of organisms have been much more heavily studied than others and so we know the diversity of organisms that is currently documented is somewhat biased. Many studies have attempted to estimate the total biodiversity of Earth and, with varying methodologies, have produced estimates anywhere from 2 million to over 100 million species. Recently, Dr. Camilo Mora (at the University of Hawai’i, Mānoa) and his colleagues reviewed the various estimates and tested a new methodology to narrow down this large range to a plausible estimate. In the study, Mora used a similar strategy to the species-area and rarefaction curves described above; however, the estimate of ‘effort’ on the x-axis was not the area or number of individuals sampled, but time. Scientists have been described approximately 1.2 million species since the mid 1700’s; has the number of new species described every year begun to level off? If so, that might indicate that most species have been discovered. If not, perhaps we can use the trend in new species discovered through time to predict where the graph might begin to level off. Mora’s team also considered higher taxonomic levels; we can be relatively certain that scientists have not yet discovered every single species on Earth, but might scientists have already described all the genera, families, orders, classes, or phyla? The results of Mora’s analysis are shown in Fig 3.  By using the pattern of discovery of new organisms at different taxonomic levels (Fig 3A-F) and the relationship between maximum diversity of each level (Fig 3G), Mora’s team arrived at an overall estimate of 8.7 million species on Earth. Since scientists have currently described approximately 1.2 million species, this estimate indicates that approximately 86% of species on Earth have not yet been discovered. At our current rate of discovery, Mora’s team estimated that it will take scientists another 1,200 years to identify all species on Earth. Figure $$\PageIndex{3}$$: Discovery of new animal taxa through time. Gray lines represent accumulation of discovered groups and colored ranges are multimodel agreement; horizontal gray lines are the consensus asymptote of the models. Panel G shows the relationship between the consensus asymptote for each taxonomic level. Figure from Mora et al 2011.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5688332319259644, "perplexity": 1299.4307871294088}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-25/segments/1623488528979.69/warc/CC-MAIN-20210623011557-20210623041557-00418.warc.gz"}
http://arxiv.org/abs/math/0701917
math (what is this?) # Title: Proliferating parasites in dividing cells : Kimmel's branching model revisited Authors: Vincent Bansaye (PMA) Abstract: We consider a branching model introduced by Kimmel for cell division with parasite infection. Cells contain proliferating parasites which are shared randomly between the two daughter cells when they divide. We determine the probability that the organism recovers, meaning that the asymptotic proportion of contaminated cells vanishes. We study the tree of contaminated cells, give the asymptotic number of contaminated cells and the asymptotic proportions of contaminated cells with a given number of parasites. This depends on domains inherited from the behavior of branching processes in random environment (BPRE) and given by the bivariate value of the means of parasite offsprings. In one of these domains, the convergence of proportions holds in probability, the limit is deterministic and given by the Yaglom quasistationary distribution. Moreover, we get an interpretation of the limit of the Q-process as the size-biased quasistationary distribution. Subjects: Probability (math.PR); Populations and Evolution (q-bio.PE) MSC classes: 60J80, 60J85, 60K37, 92C37, 92D25, 92D30 Cite as: arXiv:math/0701917 [math.PR] (or arXiv:math/0701917v2 [math.PR] for this version) ## Submission history From: Vincent Bansaye [view email] [via CCSD proxy] [v1] Wed, 31 Jan 2007 10:33:43 GMT (55kb) [v2] Sat, 28 Jun 2008 07:41:56 GMT (108kb)
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9454482793807983, "perplexity": 4786.915872613106}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-30/segments/1469257824756.90/warc/CC-MAIN-20160723071024-00261-ip-10-185-27-174.ec2.internal.warc.gz"}
http://math.stackexchange.com/questions/165941/finding-a-splitting-field-of-x3-x-1-over-mathbbz-2
# Finding a splitting field of $x^3 + x +1$ over $\mathbb{Z}_2$ Finding a splitting field of $x^3 + x +1$ over $\mathbb{Z}_2$. Ok so originally I messed around with $x^3 + x +1$ for a bit looking for an easy way to factor it and eventually decided that the factors are probably made up of really messy nested roots. So then I tried looking at the quotient field $\mathbb{Z}_2[x]/x^3 + x + 1$ to see if I would get lucky and it would contain all three roots but it doesn't. Is there a clever way to easily find this splitting field besides using the cubic formula to find the roots and then just directly adjoining them to $\mathbb{Z}_2$? Edit: Ok it turns out I miscalculated in my quotient field, $\mathbb{Z}_2[x]/x^3 + x + 1$ does contain all three roots. - If degree is not $3$ it is $6$. – André Nicolas Jul 3 '12 at 2:08 Actually, your quotient $\mathbb{Z}_2[x]/(x^3+x+1)$ does contain at least one root: $x$. This is because $x^3+x+1\equiv 0\pmod{x^3+x+1}$. Can you find the others? – roninpro Jul 3 '12 at 2:08 The splitting field is either degree $3$ or degree $6$ over $\mathbb{Z}_2$, hence it is either $\mathbb{F}_8$ or $\mathbb{F}_{64}$. Let $\alpha$ be a root, so that $\mathbb{F}_8 = \mathbb{F}(\alpha)$. The elements are of the form $a+b\alpha+c\alpha^2$, with $\alpha^3=\alpha+1$. Now, the question is whether any of these elements besides $\alpha$ is a root of the original polynomial $x^3+x+1$. Note that $(\alpha^2)^3 = (\alpha^3)^2 = (\alpha+1)^2 = \alpha^2+1$, and so if we plug in $\alpha^2$ into the polynomial we have $$\alpha^6 + \alpha^2 + 1 = \alpha^2+1+\alpha^2+1= 0.$$ Thus, $\alpha^2$ is also a root. So the polynomial has at least two roots in $\mathbb{F}_8$, and so splits there. - oh shoot I must have miscaculated in my quotient field, ok cool, thanks! – Thoth Jul 3 '12 at 2:11 If your cubic were to factor, the factorization must have at least one linear term, which corresponds to a root. It is easy to check it has no roots in $\mathbb{Z}_2$ - check them both! Plugging in $0$ and $1$ both give $1$ so are not roots, so your polynomial is irreducible over $\mathbb{Z}_2.$ A way to write the splitting field is $\mathbb{Z}_2(\alpha)$ where $\alpha$ is any one of the roots of $x^3+x+1.$ This is because $\alpha^2$ and $\alpha^4= \alpha^2+\alpha$ must then also be distinct roots of $x^3+x+1$, and these 3 then comprise a full list. Note, this isn't a special trick to notice for this problem. In fields of characteristic $p$, if $\beta$ is a root of a polynomial then $\beta^p$ will automatically also be a root, due to properties of the Frobenius endomorphism. Or as roninpro pointed out in the comments, the quotient ring you considered (which is essentially the same thing as adjoining these roots) does contain at least one root, and by this same trick, all the roots. - I thought forming the quotient field adjoined just a single root? And that sometimes you would get lucky and the remaining roots could be formed within the quotient field as well, but that sometimes they couldn't, is this not correct? – Thoth Jul 3 '12 at 2:17 @NollieTré You're correct. But extensions of finite fields are all Galois and in particular normal. So if $K/k$ are finite fields and $f \in k[X]$ is irreducible and has a root in $K$, then it splits in $K$. Frobenius is a great thing! – Dylan Moreland Jul 3 '12 at 2:25 @NollieTré You are correct, sometimes the extra roots we find with the Frobenius trick just correspond to the same roots. But here they are distinct. You can check yourself (by carrying out the division) that here $x^2$ is also a root, since $x^6+x^2+1 = (x^3+x+1)(x^3-x-1)+2(x^2+x+1) = 0$, and similarly $x^4$ is also a root. So the quotient formed actually has all the roots. – Ragib Zaman Jul 3 '12 at 2:26 Ok interesting, so is there a canonical example of when the quotient field does not contain all the roots of the polynomial? – Thoth Jul 3 '12 at 2:29 I think it's easier to note that if $\alpha$ is a root of the polynomial then squaring gives $0^2 = (\alpha^3 + \alpha + 1)^2 = (\alpha^2)^3 + \alpha^2 + 1$. – Dylan Moreland Jul 3 '12 at 2:30 Two good answers already here, but I wanted to emphasize the usefulness of the Frobenius endomorphism. Let $f(X) = X^3 + X + 1$. If I let $\alpha$ denote the image of $X$ in $k = \mathbb F_2[X]/(f(X))$ then applying Frobenius to $0 = f(\alpha)$ gives $0 = (\alpha^3 + \alpha + 1)^2 = (\alpha^2)^3 + \alpha^2 + 1.$ Hence $\alpha^2$ is also a root of $f$, and $\alpha^2 \neq \alpha$ because $\alpha \neq 0, 1$. Since $k$ contains two roots of the cubic $f$, it contains three. To bring in slightly more technology: extensions of finite fields are Galois and in particular normal, and this implies that if you adjoin one root of an irreducible polynomial over a finite field then you've actually adjoined all of them. This is, of course, not true in general! - Just wanted to add here, in $\Bbb{Z}_2(\alpha)[x]$, one can factor $x^3 + x + 1 = (x + \alpha)(x^2 + \alpha x + (\alpha^2 + 1))$. Given that $\alpha^2$ is also a root (as Dylan shows above), using ordinary high-school factorization techniques the third root is easily seen to be $\alpha^2 + \alpha$ (a somewhat preferable form than $\alpha^4$). That is: $x^2 + \alpha x + (\alpha^2 + 1) = (x + \alpha^2)(x + (\alpha^2 + \alpha))$ – David Wheeler Jul 6 '12 at 10:20
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 2, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9060187339782715, "perplexity": 186.81982532539902}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-07/segments/1454701150206.8/warc/CC-MAIN-20160205193910-00182-ip-10-236-182-209.ec2.internal.warc.gz"}
https://arxiv.org/abs/1807.02069
math.DS (what is this?) # Title:Singular perturbation analysis of a regularized MEMS model Abstract: Micro-Electro Mechanical Systems (MEMS) are defined as very small structures that combine electrical and mechanical components on a common substrate. Here, the electrostatic-elastic case is considered, where an elastic membrane is allowed to deflect above a ground plate under the action of an electric potential, whose strength is proportional to a parameter $\lambda$. Such devices are commonly described by a parabolic partial differential equation that contains a singular nonlinear source term. The singularity in that term corresponds to the so-called "touchdown" phenomenon, where the membrane establishes contact with the ground plate. Touchdown is known to imply the non-existence of steady state solutions and blow-up of solutions in finite time. We study a recently proposed extension of that canonical model, where such singularities are avoided due to the introduction of a regularizing term involving a small "regularization" parameter $\varepsilon$. Methods from dynamical systems and geometric singular perturbation theory, in particular the desingularization technique known as "blow-up", allow for a precise description of steady-state solutions of the regularized model, as well as for a detailed resolution of the resulting bifurcation diagram. The interplay between the two main model parameters $\varepsilon$ and $\lambda$ is emphasized; in particular, the focus is on the singular limit as both parameters tend to zero. Subjects: Dynamical Systems (math.DS) MSC classes: 34B16, 34C23, 34E05, 34E15, 34L30, 35K67, 74G10 Cite as: arXiv:1807.02069 [math.DS] (or arXiv:1807.02069v1 [math.DS] for this version) ## Submission history From: Annalisa Iuorio [view email] [v1] Thu, 5 Jul 2018 16:03:43 UTC (305 KB)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9080376625061035, "perplexity": 818.8537188746083}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-09/segments/1550247505838.65/warc/CC-MAIN-20190221152543-20190221174543-00419.warc.gz"}
https://edwardfhughes.wordpress.com/2012/08/
# The Theorem of The Existence of Zeroes It’s time to prove the central result of elementary algebraic geometry. Mostly it’s referred to as Hilbert’s Nullstellensatz. This German term translates precisely to the title of this post. Indeed ‘Null’ means ‘zero’, ‘stellen’ means to exist and ‘Satz’ means theorem. But referring to it merely as an existence theorem for zeroes is inadequate. Its real power is in setting up a correspondence between algebra and geometry. Are you sitting comfortably? Grab a glass of water (or wine if you prefer). Settle back and have a peruse of these theorems. This is your first glance into the heart of a magical subject. (In many texts these theorems are all referred to as the Nullstellensatz. I think this is both pointless and confusing, so have renamed them! If you have any comments or suggestions about these names please let me know). Theorem 4.1 (Hilbert’s Nullstellensatz) Let $J\subsetneq k[\mathbb{A}^n]$ be a proper ideal of the polynomial ring. Then $V(J)\neq \emptyset$. In other words, for every nontrivial ideal there exists a point which simulataneously zeroes all of its elements. Theorem 4.2 (Maximal Ideal Theorem) Every maximal ideal $\mathfrak{m}\subset k[\mathbb{A}^n]$ is of the form $(x-a_1,\dots,x-a_n)$ for some $(a_1,\dots,a_n)\in \mathbb{A}^n$. In other words every maximal ideal is the ideal of some single point in affine space. Theorem 4.3 (Correspondence Theorem) For every ideal $J\subset k[\mathbb{A}^n]$ we have $I(V(J))=\sqrt{J}$. We’ll prove all of these shortly. Before that let’s have a look at some particular consequences. First note that 4.1 is manifestly false if $k$ is not algebraically closed. Consider for example $k=\mathbb{R}$ and $n=1$. Then certainly $V(x^2+1)=\emptyset$. Right then. From here on in we really must stick just to algebraically closed fields. Despite having the famous name, 4.1 not really immediately useful. In fact we’ll see its main role is as a convenient stopping point in the proof of 4.3 from 4.2. The maximal ideal theorem is much more important. It precisely provides the converse to Theorem 3.10. But it is the correspondence theorem that is of greatest merit. As an immediate corollary of 4.3, 3.8 and 3.10 (recalling that prime and maximal ideals are radical) we have Corollary 4.4 The maps $V,I$ as defined in 1.2 and 2.4 give rise to the following bijections $\{\textrm{affine varieties in }\mathbb{A}^n\} \leftrightarrow \{\textrm{radical ideals in } k[\mathbb{A}^n]\}$ $\{\textrm{irreducible varieties in }\mathbb{A}^n\} \leftrightarrow \{\textrm{prime ideals in } k[\mathbb{A}^n]\}$ $\{\textrm{points in }\mathbb{A}^n\} \leftrightarrow \{\textrm{maximal ideals in } k[\mathbb{A}^n]\}$ Proof We’ll prove the first bijection explicitly, for it is so rarely done in the literature. The second and third bijections follow from the argument for the first and 3.8, 3.10. Let $J$ be a radical ideal in $k[\mathbb{A}^n]$. Then $V(J)$ certainly an affine variety so $V$ well defined. Moreover $V$ is injective. For suppose $\exists J'$ radical with $V(J')=V(J)$. Then $I(V(J'))=I(V(J))$ and thus by 4.3 $J = J'$. It remains to prove that $V$ surjective. Take $X$ an affine variety. Then $J'=I(X)$ an ideal with $V(J')=X$ by Lemma 2.5. But $J'$ not necessarily radical. Let $J=\sqrt{J'}$ a radical ideal. Then by 4.3 $I(V(J'))=J$. So $V(J) = V(I(V(J')) = V(J') = X$ by 2.5. This completes the proof. $\blacksquare$ We’ll see in the next post that we need not restrict our attention to $\mathbb{A}^n$. In fact using the coordinate ring we can gain a similar correspondence for the subvarieties of any given variety. This will lead to an advanced introduction to the language of schemes. With these promising results on the horizon, let’s get down to business. We’ll begin by recalling a definition and a theorem. Definition 4.5 A finitely generated $k$-algebra is a ring $R$ s.t. $R \cong k[a_1,\dots,a_n]$ for some $a_i \in R$. A finite $k$-algebra is a ring $R$ s.t. $R\cong ka_1 + \dots ka_n$. Observe how this definition might be confusing when compared to a finitely generated $k$-module. But applying a broader notion of ‘finitely generated’ to both algebras and modules clears up the issue. You can check that the following definition is equivalent to those we’ve seen for algebras and modules. A finitely generated algebra is richer than a finitely generated module because an algebra has an extra operation – multiplication. Definition 4.6 We say an algebra (module) $A$ is finitely generated if there exists a finite set of generators $F$ s.t. $A$ is the smallest algebra (module) containing $F$. We then say that $A$ is generated by $F$. Theorem 4.7 Let $k$ be a general field and $A$ a finitely generated $k$-algebra. If $A$ is a field then $A$ is algebraic over $k$. Okay I cheated a bit saying ‘recall’ Theorem 4.7. You probably haven’t seen it anywhere before. And you might think that it’s a teensy bit abstract! Nevertheless we shall see that it has immediate practical consequences. If you are itching for a proof, don’t worry. We’ll in fact present two. The first will be due to Zariski, and the second an idea of Noether. But before we come to those we must deduce 4.1 – 4.3 from 4.7. Proof of 4.2 Let $m \subset k[\mathbb{A}^n]$ be a maximal ideal. Then $F = k[\mathbb{A}^n]/m$ a field. Define the natural homomorphism $\pi: k[\mathbb{A}^n] \ni x \mapsto x+m \in F$. Note $F$ is a finitely generated $k$-algebra, generated by the $x_i+m$ certainly. Thus by 4.7 $F/k$ is an algebraic extension. But $k$ was algebraically closed. Hence $k$ is isomorphic to $F$ via $\phi : k \rightarrowtail k[\mathbb{A}^n] \xrightarrow{\pi} F$. Let $a_i = \phi^{-1}(x_i+m)$. Then $\pi(x_i - a_i) = 0$ so $x_i - a_i \in \textrm{ker}\pi = m$. Hence $(x_1-a_1, \dots, x_n-a_n) \subset m$. But $(x_1-a_1, \dots, x_n-a_n)$ is itself maximal by 3.10. Hence $m = (x_1-a_1, \dots, x_n-a_n)$ as required. $\blacksquare$ That was really quite easy! We just worked through the definitions, making good use of our stipulation that $k$ is algebraically closed. We’ll soon see that all the algebraic content is squeezed into the proof of 4.7 Proof of 4.1 Let $J$ be a proper ideal in the polynomial ring. Since $k[\mathbb{A}^n]$ Noetherian $J\subset m$ some maximal ideal. From 4.2 we know that $m=I(P)$ some point $P\in \mathbb{A}^n$. Recall from 2.5 that $V(I(P)) = \{P\} \subset V(J)$ so $V(J) \neq \emptyset$. $\blacksquare$ The following proof is lengthier but still not difficult. Our argument uses a method known as the Rabinowitsch trick. Proof of 4.3 Let $J\triangleleft k[\mathbb{A}^n]$ and $f\in I(V(J))$. We want to prove that $\exists N$ s.t. $f^N \in J$. We start by introducing a new variable $t$. Define an ideal $J_f \supset J$ by $J_f = (J, ft - 1) \subset k[x_1,\dots,x_n,t]$. By definition $V(J_f) = \{(P,b) \in \mathbb{A}^{n+1} : P\in V(J), \ f(P)b = 1\}$. Note that $f \in I(V(J))$ so $V(J_f) = \emptyset$. Now by 4.1 we must have that $J_f$ improper. In other words $J_f = k[x_1,\dots, x_n, t]$. In particular $1 \in J_f$. Since $k[x_1,\dots, x_n, t]$ is Noetherian we know that $J$ finitely generated by some $\{f_1,\dots,f_r\}$ say. Thus we can write $1 = \sum_{i=1}^r g_i f_i + g_o (ft - 1)$ where $g_i\in k[x_1,\dots , x_n, t]$ (*). Let $N$ be such that $t^N$ is the highest power of $t$ appearing among the $g_i$ for $=\leq i \leq r$. Now multiplying (*) above by $f^N$ yields $f^N = \sum_{i=1}^r G_i(x_1,\dots, x_n, ft) f_i + G_0(x_1,\dots,x_n,ft)(ft-1)$ where we define $G_i = f^N g_i$. This equation is valid in $k[x_1,\dots,x_n, t]$. Consider its reduction in the ring $k[x_1,\dots,x_n,t]/(ft - 1)$. We have the congruence $f_N\equiv \sum_{i=1}^r h_i (x_1,\dots,x_n) f_i \ \textrm{mod}\ (ft-1)$ where $h_i = G_i(x_1,\dots,x_n,1)$. Now consider the map $\phi:k[x_1,\dots, x_n]\rightarrowtail k[x_n,\dots, x_n,t]\xrightarrow{\pi} k[x_n,\dots, x_n,t]/(ft-1)$. Certainly nothing in the image of the injection can possibly be in the ideal $(ft - 1)$, not having any $t$ dependence. Hence $\phi$ must be injective. But then we see that $f^N = \sum_{i=1}^r h_i(x_1,\dots, x_n) f_i$ holds in the ring $k[\mathbb{A}^n]$. Recalling that the $f_i$ generate $J$ gives the result. $\blacksquare$ We shall devote the rest of this post to establishing 4.7. To do so we’ll need a number of lemmas. You might be unable to see the wood for the trees! If so, you can safely skim over much of this. The important exception is Noether normalisation, which we’ll come to later. I’ll link the ideas of our lemmas to geometrical concepts at our next meeting. Definition 4.8 Let $A,B$ be rings with $B \subset A$. Let $a\in A$. We say that $a$ is integral over $B$ if $a$ is the root of some monic polynomial with roots in $B$. That is to say $\exists b_i \in B$ s.t. $a^n + b_{n-1}a^{n-1} + \dots + b_0 = 0$. If every $a \in A$ is integral over $B$ we say that $A$  is integral over $B$ or $A$ is an integral extension of $B$. Let’s note some obvious facts. Firstly we can immediately talk about $A$ being integral over $B$ when $A,B$ are algebras with $B$ a subalgebra of $A$. Remember an algebra is still a ring! It’s rather pedantic to stress this now, but hopefully it’ll prevent confusion if I mix my termin0logy later. Secondly observe that when $A$ and $B$ are fields “integral over” means exactly the same as “algebraic over”. We’ll begin by proving some results that will be of use in both our approaches. We’ll see that there’s a subtle interplay between finite $k$-algebras, integral extensions and fields. Lemma 4.9 Let $F$ be a field and $R\subset F$ a subring. Suppose $F$ is an integral extension of $R$. Then $R$ is itself a field. Proof Let $r \in R$. Then certainly $r \in F$ so $r^{-1} \in F$ since $F$ a field. Now $r^{-1}$ integral over $R$ so satisfies an equation $r^-n = b_{n-1} r^{-n+1} +\dots + b_0$ with$b_i \in R$. But now multiplying through by $r^{n-1}$ yields $r^{-1} = b_{n-1} + \dots + b_0 r^{n-1} \in R$. $\blacksquare$ Note that this isn’t obvious a priori. The property that an extension is integral contains sufficient information to percolate the property of inverses down to the base ring. Lemma 4.10 If $A$ is a finite $B$ algebra then $A$ is integral over $B$. Proof Write $A = Ba_1 + \dots +Ba_n$. Let $x \in A$. We want to prove that $x$ satisfies some equation $x^n + b_{n-1}x^n{n-1} + \dots + b_0 = 0$. We’ll do so by appealing to our knowledge about determinants. For each $a_i$ we may clearly write $xa_i = \sum_{i=1}^{n} b_{ij}a_j$ for some $b_ij \in B$. Writing $\vec{a} = (a_1, \dots, a_n)$ and defining the matrix $(\beta)_{ij} = b_{ij}$ we can express our equation as $\beta a = xa$. We recognise this as an eigenvalue problem. In particular $x$ satisfies the characteristic polynomial of $\beta$, a polynomial of degree $n$ with coefficients in $B$. But this is precisely what we wanted to show. $\blacksquare$ Corollary 4.11 Let $A$ be a field and $B\subset A$ a subring. If $A$ is a finite $B$-algebra then $B$ is itself a field. Proof Immediate from 4.9 and 4.10. $\blacksquare$ We now focus our attention on Zariski’s proof of the Nullstellensatz. I take as a source Daniel Grayson’s excellent exposition. Lemma 4.12 Let $R$ be a ring an $F$ a $R$-algebra generated by $x \in F$. Suppose further that $F$ a field. Then $\exists s \in R$ s.t. $S = R[s^{-1}]$ a field.  Moreover $x$ is algebraic over $S$. Proof Let $R'$ be the fraction field of $R$. Now recall that $x$ is algebraic over $R'$ iff $R'[x] \supset R'(x)$. Thus $x$ is algebraic over $R'$ iff $R'[x]$ is a field. So certainly our $x$ is algebraic over $R'$ for we are given that $F$ a field. Let $x^n + f_{n-1}x^{n-1} + \dots + f_0$ be the minimal polynomial of $x$. Now define $s\in R$ to be the common denominator of the $f_i$, so that $f_0,\dots, f_{n-1} \in R[s^{-1}] = S$. Now $x$ is integral over $S$ so $F/S$ an integral extension. But then by 4.9 $S$ a field, and $x$ algebraic over it. $\blacksquare$ Observe that this result is extremely close to 4.7. Indeed if we take $R$ to be a field we have $S = R$ in 4.12. Then lemma then says that $R[x]$ is algebraic as a field extension of $R$. Morally this proof mostly just used definitions. The only nontrivial fact was the relationship between $R'(x)$ and $R'[x]$. Even this is not hard to show rigorously from first principles, and I leave it as an exercise for the reader. We’ll now attempt to generalise 4.12 to $R[x_1,\dots,x_n]$. The argument is essentially inductive, though quite laborious. 4.7 will be immediate once we have succeeded. Lemma 4.13 Let $R = F[x]$ be a polynomial ring over a field $F$. Let $u\in R$. Then $R[u^{-1}]$ is not a field. Proof By Euclid, $R$ has infinitely many prime elements. Let $p$ be a prime not dividing $u$. Suppose $\exists q \in R[u^{-1}]$ s.t. $qp = 1$. Then $q = f(u^{-1})$ where $f$ a polynomial of degree $n$ with coefficients in $R$. Hence in particular $u^n = u^n f(u^{-1}) p$ holds in $R$ for $u^n f(u^{-1}) \in R$. Thus $p | u^n$ but $p$ prime so $p | u$. This is a contradiction. $\blacksquare$ Corollary 4.14 Let $K$ be a field, $F\subset K$ a subfield, and $x \in K$. Let $R = F[x]$. Suppose $\exists u\in R$ s.t. $R[u^{-1}] = K$. Then $x$ is algebraic over $F$. Moreover $R = K$. Proof Suppose $x$ were transcendental over $F$. Then $R=F[x]$ would be a polynomial ring, so by 4.12 $R[u^{-1}]$ couldn’t be a field. Hence $x$ is algebraic over $F$ so $R$ is a field. Hence $R=R[u{-1}]=K$. $\blacksquare$ The following fairly abstract theorem is the key to unlocking the Nullstellensatz. It’s essentially a slight extension of 4.14, applying 4.12 in the process. I’d recommend skipping the proof first time, focussing instead on how it’s useful for the induction of 4.16. Theorem 4.15 Take $K$ a field, $F \subset K$ a subring, $x \in K$. Let $R = F[x]$. Suppose $\exists u\in R$ s.t. $R[u^{-1}] = K$. Then $\exists 0\neq s \in F s.t. F[s^{-1}]$ is a field. Moreover $F[s^{-1}][x] = K$ and $x$ is algebraic over $F[s^{-1}]$. Proof Let $L=\textrm{Frac}(F)$. Now by 4.14 we can immediately say that $L[x]=K$, with $x$ algebraic over $L$. Now we seek our element $s$ with the desired properties. Looking back at 4.12, we might expect it to be useful. But to use 4.12 for our purposes we’ll need to apply it to some $F' = F[t^{-1}]$ with $F'[x] = K$, where $t \in F$. Suppose we’ve found such a $t$. Then 4.12 gives us $s' \in F'$ s.t. $F'[s'^{-1}]$ a field with $x$ algebraic over it. But now $s' = qt^{-m}$ some $q \in F, \ m \in \mathbb{N}$. Now $F'[s'^{-1}]=F[t^{-1}][s'^{-1}]=F[(qt)^{-1}]$, so setting $=qt$ completes the proof. (You might want to think about that last equality for a second. It’s perhaps not immediately obvious). So all we need to do is find $t$. We do this using our first observation in the proof. Observe that $u^{-1}\in K=L[x]$ so we can write $u^{-1}=l_0+\dots +l_{n-1}x^{n-1}$, $l_i \in L$. Now let $t \in F$ be a common denominator for all the $l_i$. Then $u^{-1} \in F'=F[t^{-1}]$ so $F'[x]=K$ as required. $\blacksquare$ Corollary 4.16 Let $k$ a ring, $A$ a field, finitely generated as a $k$-algebra by $x_1,\dots,x_n$. Then $\exists 0\neq s\in k$ s.t. $k[s^{-1}]$ a field, with $A$ a finite algebraic extension of $k[s^{-1}]$. Trivially if $k$ a field, then $A$ is algebraic over $k$, establishing 4.7. Proof Apply Lemma 4.15 with $F=k[x_1,\dots,x_{n-1}]$, $x=x_n$, $u=1$ to get $s'\in F$ s.t. $A' = k[x_1,\dots,x_{n-1}][s'^{-1}]$ is a field with $x_n$ algebraic over it. But now apply 4.15 again with $F=k[x_1,\dots,x_{n-2}]$, $u = s'$ to deduce that $A''=k[x_1,\dots, x_{n-2}][s''^{-1}]$ is a field, with $A'$ algebraic over $A''$, for some $s'' \in F$. Applying the lemma a further $(n-2)$ times gives the result. $\blacksquare$ This proof of the Nullstellensatz is pleasingly direct and algebraic. However it has taken us a long way away from the geometric content of the subject. Moreover 4.13-4.15 are pretty arcane in the current setting. (I’m not sure whether they become more meaningful with a better knowledge of the subject. Do comment if you happen to know)! Our second proof sticks closer to the geometric roots. We’ll introduce an important idea called Noether Normalisation along the way. For that you’ll have to come back next time! # Gravity for Beginners Everyone knows about gravity. Every time you drop a plate or trip on the stairs you’re painfully aware of it. It’s responsible for the thrills and spills at theme parks and the heady hysteria of a plummeting toboggan. But gravity is not merely restricted to us small fry on Earth. It is a truly universal force, clumping mass together to form stars and planets, keeping worlds in orbit around their Suns. Physicists call gravity a fundamental force – it cannot be explained in terms of any other interaction. The most accurate theory of gravity is Einstein’s General Theory of Relativity. The presence of mass warps space and time, creating the physical effects we observe. The larger the mass, the more curved space and time become, so the greater the gravitational pull. Space and time are a rubber sheet, which a large body like the Sun distorts. Gravity is very weak compared to other fundamental forces. This might be a bit of a surprise – after all it takes a very powerful rocket to leave Earth’s orbit. But this is just because Earth is so very huge. Things on an everyday scale don’t seem to be pulled together by gravity. But small magnets certainly are attracted to each other by magnetism. So magnetism is stronger than gravity. The fact that gravity acts by changing the geometry of space and time sets it apart from all other forces. In fact our best theories of particles assume that there is some cosmic blank canvas on which interactions happen. This dichotomy and the weakness of gravity give rise to the conflicts at the heart of physics. It’s All Relative, After All “You can’t just let nature run wild.” Walt Disney Throughout history, scientists have employed principles of relativity to understand nature. In broad terms these say that different people doing the same scientific experiment should get the same answer. This stands to reason – we travel the world and experience the same laws of nature everywhere. Any differences we might think we observe can be explained by a change in our experimental conditions. Looking for polar bears in the Sahara is obvious madness. In the early 17th century Galileo formulated a specific version of relativity that applied to physics. He said that the laws of physics are the same for everyone moving at constant speed, regardless of the speed they are going. We notice this in everyday life. If you do an egg and spoon race walking at a steady pace the egg will stay on the spoon just as if you were standing still. If you try to speed up or slow down, though, the egg will likely go crashing to the ground. This shows that for an accelerating observer the laws of physics are not the same. Newton noticed this and posited that an accelerating object feels a force that grows with its mass and acceleration. Sitting in a plane on takeoff this force pushes you backwards into your seat. Physicists call this Newton’s Second Law. Newton’s Second Law can be used the opposite way round too. Let’s take an example. Suppose we drop an orange in the kitchen. As it travels through the air its mass certainly stays the same and the only significant force it feels is gravity. Using Newton’s Second Law we can calculate the acceleration of the orange. Now acceleration is just change in speed over time. So given any time in the future we can predict the speed of the orange. But speed is change in distance over time. So we know the distance the orange has travelled towards the ground after any given time. In particular we can say exactly how much time it’ll be before the orange goes splat on the kitchen floor. One of the remarkable features of the human brain is that it can do approximations to these calculations very quickly, enabling us to catch the orange before disaster strikes. The power of the principle of relativity is now apparent. Suppose we drop the orange in a lift, while steadily travelling upwards. We can instantly calculate how long it’ll take to hit the lift floor. Indeed by the principle of relativity it must take exactly the same amount of time as it did when we were in the kitchen. There’s a hidden subtlety here. We’ve secretly assumed that there is some kind of universal clock, ticking away behind the scenes. In other words, everyone measures time the same, no matter how they’re moving. There’s also a mysterious cosmic tape measure somewhere offstage. That is, everyone agrees on distances, regardless of their motion. These hypotheses are seemingly valid for everyday life. But somehow these notions of absolute space and time are a little unsettling. It would seem that Galileo’s relativity principle applies not only to physics, but also to all of space and time. Newton’s ideas force the universe to exist against the fixed backdrop of graph paper. Quite why the clock ticks and the ruler measures precisely as they do is not up for discussion. And the mysteries only deepen with Newton’s theory of gravity. A Tale of Two Forces “That one body may act upon another at a distance through a vacuum without the mediation of anything else […] is to me so great an absurdity that […] no man who has […] a competent faculty of thinking could ever fall into it.” Sir Isaac Newton Newton was arguably the first man to formulate a consistent theory of gravitation. He claimed that two masses attract each other with a force related to their masses and separation. The heavier the objects and the closer they are, the bigger the force of gravity. This description of gravity was astoundingly successful.  It successively accounted for the curvature of the Earth, explained the motion of the planets around the Sun, and predicted the precise time of appearance of comets. Contemporary tests of Newton’s theory returned a single verdict – it was right. Nevertheless Newton was troubled by his theory. According to his calculations, changes in the gravitational force must be propagated instantaneously throughout the universe. Naturally he sought a mechanism for such a phenomenon. Surely something must carry this force and effect its changes. Cause and effect is an ubiquitous feature of everyday physics. Indeed Newton’s Second Law says motion and force and inextricably linked. Consider the force which opens a door. It has a cause – our hand pushing against the wood – and an effect – the door swinging open. But Newton couldn’t come up with an analogous explanation for gravity. He had solved the “what”, but the “why” and “how” evaded him entirely. For more than a century Newton’s theories reigned supreme. It was not until the early 1800s that physicists turned their attention firmly towards another mystery – electricity. Michael Faraday lead the charge with his 1721 invention of an electric motor. By placing a coil of wire in a magnetic field and connecting a battery he could make it rotate. The race was on to explain this curious phenomenon. Faraday’s work implied that electricity and magnetism were two sides of the same coin. The strange force he had observed became known, appropriately, as electromagnetism. The scientific community quickly settled on an idea. Electricity and magnetism were examples of a force field – at every point in space surrounding a magnet or current, there were invisible lines of force which affected the motion of nearby objects. All they needed now were some equations. This would allow them to predict the behaviour of currents near magnets, and verify that this revolutionary force field idea was correct. Without a firm mathematical footing the theory was worthless. Several preeminent figures tried their hand at deriving a complete description, but in 1861 the problem remained open. In that year a young Scotsman named James Clerk Maxwell finally cracked the issue. By modifying the findings of those before him he arrived at a set of equations which completely described electromagnetism. He even went one better than Newton had with gravity. He found a mechanism for the transmission of electromagnetic energy. Surprisingly, Maxwell’s equations suggested that light was an electromagnetic wave. To wit, solutions showed that electricity and magnetism could spread out in a wavelike manner. Moreover the speed of these waves was determined by a constant in his equations. This constant turned out to be very close to the expected speed of light in a vacuum. It wasn’t a giant leap to suppose this wave was light itself. At first glance it seems like these theories solve all of physics. To learn about gravity and the motion of uncharged objects we use Newton’s theory. To predict electromagnetic phenomena we use Maxwell’s. Presumably to understand the motion of magnets under gravity we need a bit of both. But trying to get Maxwell’s equations to play nice with Galileo’s relativity throws a big spanner in the works. A Breath of Fresh Air “So the darkness shall be the light, and the stillness the dancing.” T.S. Eliot Galileo’s relativity provides a solid bedrock for Newtonian physics. It renders relative speeds completely irrelevant, allowing us to concentrate on the effects of acceleration and force. Newton’s mechanics and Galileo’s philosophy reinforce each other. In fact Newton’s equations look the same no matter what speed you are travelling at. Physicists say they are invariant for Galileo’s relativity. We might naively hope that Maxwell’s equations are invariant. Indeed a magnet behaves the same whether you are sitting still at home or running around looking for buried treasure. Currents seem unaffected by how fast you are travelling – an iPod still works on a train. It would be convenient if electromagnetism had the same equations everywhere. Physicists initially shared this hope. But it became immediately apparent that Maxwell’s equations were not invariant. A change in the speed you were travelling caused a change in the equations. Plus there was only one speed at which solutions to the equations gave the right answer! Things weren’t looking good for Galileo’s ideology. To solve this paradox, physicists made clever use of a simple concept. Since antiquity we’ve had a sense of perspective – the world looks different from another point of view. Here’s a simple example in physics. Imagine you’re running away from a statue. From your viewpoint the statue is moving backwards. From the statue’s viewpoint you are moving forwards. Both descriptions are right. They merely describe the same motion in contrary ways. In physics we give perspective a special name. Any observer has a frame of reference from which they see the world. In your frame of reference the statue moves backwards. In the statue’s frame of reference you move forwards. We’ve seen that Newton’s physics is the same in every frame of reference which moves with a constant speed. We can now rephrase our discovery about Maxwell’s equations. Physicists found that there was one frame of reference in which they were correct. Maxwell’s equations somehow prefer this frame of reference over all other ones! Any calculations in electromagnetism must be done relative to this fixed frame. But why is this frame singled out? Faced with this question, physicists spotted the chance to kill two birds with one stone. The discovery that light is a wave of electromagnetism raises an immediate question. What medium carries the light waves? We’re used to waves travelling through some definite substance. Water waves need a sea or ocean, sound waves require air, and earthquakes move through rock. Light waves must travel through a fixed mysterious fog pervading all of space. It was called aether. The aether naturally has it’s own frame of reference, just as you and I do. When we measure the speed of light in vacuum, we’re really measuring it relative to the aether. So the aether is a special reference frame for light. But light is an electromagnetic wave. It’s quite sensible to suggest that Maxwell’s preferred reference frame is precisely the aether! Phew, we’ve sorted it. Newtonian mechanics works with Galilean relativity because there’s nothing to specify a particular reference frame. Maxwell’s equations don’t follow relativity because light waves exist in the aether, which is a special frame of reference. Once we’ve found good evidence for the aether we’ll be home and dry. So thought physicists in the late 19th century. The gauntlet was down. Provide reliable experimental proof and win instant fame. Two ambitious men took up the challenge – Albert Michelson and Edward Morley. They reasoned that the Earth must be moving relative to the aether. Therefore from the perspective of Earth the aether must move. Just as sound moves faster when carried by a gust of wind, light must move faster when carried by a gust of aether. This prediction became known as the aether wind. By measuring the speed of light in different directions, Michelson and Morley could determine the direction and strength of the aether wind. The speed of light would be greatest when they aligned their measuring apparatus with the wind. Indeed the gusts of aether would carry light more swiftly. The speed of light would be slowest when they aligned their apparatus at right angles to the wind. The expected changes in speed were so minute that it required great ingenuity to measure them. Nevertheless by 1887 they had perfected a cunning technique. With physicists waiting eagerly for confirmation, Michelson and Morley’s experiment failed spectacularly. They measured no aether wind at all. The experiment was repeated hundreds of times in the subsequent years. The results were conclusive. It didn’t matter which orientation you chose, the speed of light was the same. This devastating realisation blew aether theory to smithereens. Try as they might, no man nor woman could paper over the cracks. Physics needed a revolution. How Time Flies “Put your hand on a hot stove for a minute, and it seems like an hour. Sit with a pretty girl for an hour, and it seems like a minute. That’s relativity.” Albert Einstein It is the most romanticised of fables. The task of setting physics right again fell to a little known patent clerk in Bern. The life and work of Albert Einstein has become a paradigm for genius. But yet the idea that sparked his reworking of reality was beautifully simple. For a generation, physicists had been struggling to reconcile Galileo and Maxwell. Einstein claimed that they had missed the point. Galileo and Maxwell are both right. We’ve just misunderstood the very nature of space and time. This seems a ridiculously bold assumption. It’s best appreciated by way of an analogy. Suppose you live in a rickety old house and buy a nice new chair. Placing it in your lounge you notice that it wobbles slightly. You wedge a newspaper under one leg. At first glance you’ve solved the problem. But when you sit on the chair it starts wobbling again. After a few more abortive attempts with newspapers, rags and other household items you decide to put the chair in another room. But the wobble won’t go away. Even when you sand down the legs you can’t make it stand firm. The chair and the house seem completely incompatible. One day a rogue builder turns up. He promises to fix your problem. You don’t believe he can. Neither changing the house nor the chair has made the slightest difference to you. When you return from work you are aghast to see he’s knocked the whole house down. He shouts up to you from a hole in the ground, “just fixing your foundations”! More concretely Einstein said that there’s no problem with Galileo’s relativity. The laws of physics really are the same in every frame moving at constant speed. Physicists often rename this idea Einstein’s special principle of relativity even though it is Galileo’s invention! Einstein also had no beef with Maxwell’s equations. In particular they are the same in every frame. To get around the fact that Maxwell’s equations are not invariant for Galileo’s relativity, Einstein claimed that we don’t understand space and time. He claimed that there is no universal tape measure, or cosmic timepiece. Everybody is responsible for their own measurements of space and time. This clears up some of the issues that annoyed Newton. In order to get Maxwell’s equations to be invariant when we change perspective, Einstein had to alter the foundations of physics. He postulated that each person’s measurements of space and time were different, depending on how fast they were going. This correction magically made Maxwell’s equations work with his principle of special relativity. There’s an easy way to understand how Einstein modified space and time. We’re used to thinking that we move at a constant speed through time. Clocks, timers and watches all attest to our obsession with measuring time consistently. The feeling of time dragging on or whizzing by is merely a psychological curiosity. Before Einstein space didn’t have this privilege. We only move through space as we choose. And moving through space has no effect on moving through time. Einstein made everything much more symmetrical. He said that we are always moving through both space and time. We can’t do anything about it. Everyone always moves through space and time at the same speed – the speed of light. To move faster through space you must move slower through time to compensate. The slower you trudge through space, the faster you whizz through time. Simple as. Einstein simply put space and time on a more even footing. We call the whole construct spacetime. Intuitively spacetime is four dimensional. That is, you can move in four independent directions. Three of these are in space, up-down, left-right, forward-backward. One of these is in time. You probably haven’t noticed it yet, but special relativity has some weird effects. First and foremost, the speed of light is the same however fast you are going. This is because it is a constant in Maxwell’s equations, which are the same in all frames. You can never catch up with a beam of light! This is not something we are used to from everyday life. Nevertheless it can be explained using Einstein’s special relativity. Suppose you measure the speed of light when you are stationery. You do this by measuring the amount of time it takes for light to go a certain distance. For the sake of argument assume you get the answer 10 mph. Now imagine speeding up to 5 mph. You measure the speed of light again. Without Einstein you’d expect the result to be 5 mph. But because you’re moving faster through space you must be moving slower through time. That means it’ll take the light less time to go the same distance. In fact the warping of spacetime precisely accounts for the speed you’ve reached. The answer is again 10 mph. Einstein’s spacetime also forces us to forget our usual notions of simultaneity. In everyday physics we can say with precision whether two events happen at the same time. But this concept relies precisely on Newton’s absolute time. Without his convenient divine timepiece we can’t talk about exact time. We can only work with relative perspectives. Let’s take an example. Imagine watching a Harrison Ford thriller. He’s standing on the top of a train as it rushes through the station. He positions himself precisely in the center of the train. At either end is a rapier wielding bad guy eager to kill him. Ford is equipped with two guns which he fires simultaneously at his two nemeses. These guns fire beams of light that kill the men instantly. In Ford’s frame both men die at the same time. The speed of light is the same in both directions and he’s equidistant from the men when he shoots. Therefore the beams hit at the same moment according to Ford. But for Ford’s sweetheart on the platform the story is different. Let’s assume she is aligned with Ford at the moment he shoots. She sees the back of the train catching up with the point where Ford took the shot. Moreover she sees the front of the train moving away from the point of firing. Now the speed of light is constant in all directions for her. Therefore she’ll see the man at the back of the train get hit before the man at the front! Remarkably all of these strange effects have been experimentally confirmed. Special relativity and spacetime really do describe our universe. But with our present understanding it seems that making genuine measurements would be nigh on impossible. We’ve only seen examples of things we can’t measure with certainty! Thankfully there is a spacetime measurement all observers can agree on. This quantity is known as proper time. It’s very easy to calculate the proper time between events A and B. Take a clock and put it on a spaceship. Set your spaceship moving at a constant speed so that it goes through the spacetime points corresponding to A and B. The proper time between A and B is the time that elapses on the spaceship clock between A and B. Everyone is forced to agree on the proper time between two events. After all it only depends on the motion of the spaceship. Observers moving at different speeds will all see the same time elapsed according to the spaceship clock. The speed that they are moving has absolutely no effect on the speed of your spaceship! You might have spotted a potential flaw. What if somebody else sets off a spaceship which takes a different route from A to B? Wouldn’t it measure a different proper time? Indeed it would. But it turns out this situation is impossible. Remember that we had to set our spaceship off at a constant speed. This means it is going in a straight line through spacetime. It’s easy to see that there’s only one possible straight line route joining any two points. (Draw a picture)! That’s it. You now understand special relativity. In just a few paragraphs we’ve made a huge conceptual leap. Forget about absolute space and time – it’s plain wrong. Instead use Einstein’s new magic measurement of proper time. Proper time doesn’t determine conventional time or length. Rather it tells us about distances in spacetime. Einstein had cracked the biggest problem in physics. But he wasn’t done yet. Armed with his ideas about relativity and proper time, he turned to the Holy Grail. Could he go one better than Newton? It was time to explain gravity. The Shape Of Space “[…] the great questions about gravitation. Does it require time? […] Has it any reference to electricity? Or does it stand on the very foundation of matter – mass or inertia?” James Clerk Maxwell Physicists are always happy when they are going at constant speed. Thanks to Galileo and Einstein they don’t need to worry exactly how fast they are going. They can just observe the world and be sure that their observations are true interpretations of physics. We can all empathise with this. It’s much more pleasant going on a calm cruise across the Aegean than a rocky boat crossing the English Channel. This is because the choppy waters cause the boat to accelerate from side to side. We’re no longer travelling at a constant speed, so the laws of physics appear unusual. This can have unpleasant consequences if we don’t find our sea legs. Given this overwhelming evidence it seems madness to alter Einstein and Galileo’s relativity. But this is exactly what Einstein did. He postulated a general principle of relativity – the laws of physics are the same in any frame of reference. That is, no matter what your speed or acceleration. To explain this brazen statement, we’ll take an example. Imagine you go to Disneyland and queue for the Freefall Ride. In this terrifying experience you are hoisted 60 metres in the air and then dropped. As you plunge towards the ground you notice that you can’t feel your weight. For a few moments you are completely weightless! In other words there is no gravity in your accelerating frame. Let’s try a similar thought experiment. Suppose you wake up and find yourself in a bare room with no windows or doors. You stick to the floor as you usually would on Earth. You might be forgiven for thinking that you were under the influence of terrestrial gravity. In fact you could equally be trapped in a spaceship, accelerating at precisely the correct rate to emulate the force of gravity. Just think back to Newton’s Second Law and this is obviously true. Hopefully you’ve spotted a pattern. Acceleration and gravity are bound together. In fact there seems to be no way of unpicking the knots. Immediately Einstein’s general relativity becomes more credible. An accelerating frame is precisely a constant speed frame with gravity around it. So far so good. But we haven’t said anything about gravity that couldn’t be said about other forces. If we carry on thinking of gravity as a conventional interaction the argument becomes quite circular. We need a description of gravity that is independent of frames. Here special relativity comes to our rescue. Remember that proper time gives us a universally agreed distance between two points. It does so by finding the length of a straight line in spacetime from one event to the other. This property characterises spacetime as flat. To see this imagine you are a millipede, living on a piece of string. You can only move forwards or backwards. In other words you have one dimension of space. You also move through time. Therefore you exist in a two dimensional spacetime. One day you decide to make a map of all spacetime. You travel to every point in spacetime and write down the proper time it took to get there. Naturally you are very efficient, and always travel by the shortest possible route. When you return you try to make a scale drawing of all your findings on a flat sheet of paper. Suppose your spacetime is exactly as we’d expect from special relativity. Then you have no problem making your map. Indeed you always travelled by the shortest route to each point, which is a straight line in the spacetime of special relativity. On your flat sheet of paper, the shortest distance between two points is also a straight line. So your map will definitely work out. You can now see why special relativistic spacetime is flat. Just for fun, let’s assume that you had trouble making your map. Just as you’ve drawn a few points on the map you realise there’s a point which doesn’t fit. Thinking you must be mistaken you try again. The same thing happens. You go out for another trek round spacetime. Exhausted you return with the same results. How immensely puzzling! Eventually you start to realise that something fundamental is wrong. What if spacetime isn’t flat? You grab a nearby bowling ball and start drawing your map on its surface. Suddenly everything adds up. The distances you measured between each point work out perfectly. Your spacetime isn’t flat, it’s spherical. Let’s look a bit closer at why this works. When you went out measuring your spacetime you took the shortest route to every point. On the surface of a sphere this is not a straight line. Rather it is part of a circle – a line of longitude or latitude. If you aren’t convinced, find a globe and measure it yourself on the Earth’s surface. When you try to make a flat map, it doesn’t work. This is because the shortest distance between two points on the flat paper is a straight line. The two systems of measurement are just incompatible. We’re used to seeing this every time we look at a flat map of the world. The distances on it are all wrong because it has to be stretched and squished to fit the flat page. Gallivanting millipedes aside, what has this got to do with reality? If we truly live in the universe of special relativity then we don’t need to worry about such complications. Everything is always flat! But Einstein realised that these curiosities were exactly the key to the treasure chest of gravity. In his landmark paper of 1915, Einstein told us to forget gravity as a force. Instead, he claimed, gravity modifies the proper time between events. In doing so, it changes the geometry of spacetime. The flat, featureless landscape of special relativity was instantly replaced by cliffs and ravines. More precisely Einstein came up with a series of equations which described how the presence of mass changes the calculation of proper time. Large masses can warp spacetime, causing planets to orbit their suns and stars to form galaxies. Nothing is immune to the change in geometry. Even light bends around massive stars. This has been verified countless times during solar eclipses. Einstein had devised a frame independent theory of gravity which elegantly explained all gravitational phenomena. His general principle of relativity was more than vindicated; it became the title of a momentous new perspective on the cosmos. The quest to decipher Newton’s gravity was complete. In the past century, despite huge quantities of research, we have made little progress beyond Einstein’s insights. The experimental evidence for general relativity is tumultuous. But yet it resists all efforts to express it as a quantum theory. We stand at an impasse, looking desperately for a way across. I’ll start this post by tying up some loose ends from last time. Before we get going there’s no better recommendation for uplifting listening than this marvellous recording. Hopefully it’ll help motivate and inspire you (and I) as we journey deeper into the weird and wonderful world of algebra and geometry. I promised a proof that for algebraically closed fields $k$ every Zariski open set is dense in the Zariski topology. Quite a mouthful at this stage of a post, I admit. Basically what I’m showing is that Zariski open sets are really damn big, only in a mathematically precise way. But what of this ‘algebraically closed’ nonsense? Time for a definition. Definition 3.1 A field $k$ is algebraically closed if every nonconstant polynomial in $k[x]$ has a root in $k$. Let’s look at a few examples. Certainly $\mathbb{R}$ isn’t algebraically closed. Indeed the polynomial $x^2 + 1$ has no root in $\mathbb{R}$. By contrast $\mathbb{C}$ is algebraically closed, by virtue of the Fundamental Theorem of Algebra. Clearly no finite field is algebraically closed. Indeed suppose $k=\{p_1,\dots ,p_n\}$ then $(x-p_1)\dots (x-p_n) +1$ has no root in $k$. We’ll take a short detour to exhibit another large class of algebraically closed fields. Definition 3.2 Let $k,\ l$ be fields with $k\subset l$. We say that $l$ is a field extension of $k$ and write $l/k$ for this situation. If every element of $l$ is the root of a polynomial in $k[x]$ we call $l/k$ an algebraic extension. Finally we say that the algebraic closure of $k$ is the algebraic extension $\bar{k}$ of $k$ which is itself algebraically closed. (For those with a more technical background, recall that the algebraic closure is unique up to $k$-isomorphisms, provided one is willing to apply Zorn’s Lemma). The idea of algebraic closure gives us a pleasant way to construct algebraically closed fields. However it gives us little intuition about what these fields ‘look like’. An illustrative example is provided by the algebraic closure of the finite field of order $p^d$ for $p$ prime. We’ll write $\mathbb{F}_{p^d}$ for this field, as is common practice. It’s not too hard to prove the following Theorem 3.3 $\mathbb{F}_{p^d}=\bigcup_{n=1}^{\infty}\mathbb{F}_{p^{n!}}$ Proof Read this PlanetMath article for details. Now we’ve got a little bit of an idea what algebraically closed fields might look like! In particular we’ve constructed such fields with characteristic $p$ for all $p$. From now on we shall boldly assume that for our purposes every field $k$ is algebraically closed I imagine that you may have an immediate objection. After all, I’ve been recommending that you use $\mathbb{R}^n$ to gain an intuition about $\mathbb{A}^n$. But we’ve just seen that $\mathbb{R}$ is not algebraically closed. Seems like we have an issue. At this point I have to wave my hands a bit. Since $\mathbb{R}^n$ is a subset of $\mathbb{C}^n$ we can recover many (all?) of the geometrical properties we want to study in $\mathbb{R}^n$ by examining them in $\mathbb{C}^n$ and projecting appropriately. Moreover since $\mathbb{C}^n$ can be identified with $\mathbb{R}^{2n}$ in the Euclidean topology, our knowledge of $\mathbb{R}^n$ is still a useful intuitive guide. However we should be aware that when we are examining affine plane curves with $k=\mathbb{C}$ they are in some sense $4$ dimensional objects – subsets of $\mathbb{C}^2$. If you can imagine $4$ dimensional space then you are a better person than I! That’s not to say that these basic varieties are completely intractable though. By looking at projections in $\mathbb{R}^3$ and $\mathbb{R}^2$ we can gain a pretty complete geometric insight. And this will soon be complemented by our burgeoning algebraic understanding. Now that I’ve finished rambling, here’s the promised proof! Lemma 3.4 Every nonempty Zariski open subset of $\mathbb{A}^1$ is dense. Proof Recall that $k[x]$ is a principal ideal domain. Thus any ideal $I\subset k[x]$ may be written $I=(f)$. But $k$ algebraically closed so $f$ splits into linear factors. In other words $I = ((x-a_1)\dots (x-a_n))$. Hence the nontrivial Zariski closed subsets of $\mathbb{A}^1$ are finite, so certainly the Zariski open subsets of $\mathbb{A}^1$ are dense. $\blacksquare$ I believe that the general case is true for the complement of an irreducible variety, a concept which will be introduced next. However I haven’t been able to find a proof, so have asked here. How do varieties split apart? This is a perfectly natural question. Indeed many objects, both in mathematics and the everyday world, are made of some fundamental building block. Understanding this ‘irreducible atom’  gives us an insight into the properties of the object itself. We’ll thus need a notion for what constitutes an ‘irreducible’ or ‘atomic’ variety. Definition 3.5 An affine variety $X$ is called reducible if one can write $X=Y\cup Z$ with $Y,\ Z$ proper subsets of $X$. If $X$ is not reducible, we call it irreducible. This seems like a good and intuitive way of defining irreducibility. But we don’t yet know that every variety can be constructed from irreducible building blocks. We’ll use the next few minutes to pursue such a theorem. As an aside, I’d better tell you about some notational confusion that sometimes creeps in. Some authors use the term algebraic set for  my term affine variety. Such books will often use the term affine variety to mean irreducible algebraic set. For the time being I’ll stick to my guns, and use the word irreducible when it’s necessary! Before we go theorem hunting, let’s get an idea about what irreducible varieties look like by examining some examples. The ‘preschool’ example is that $V(x_1 x_2)\subset \mathbb{A}^2$ is reducible, for indeed $V(x_1 x_2) = V(x_1)\cup V(x_2)$. This is neither very interesting nor really very informative, however. A better example is the fact that $\mathbb{A}^1$ is irreducible. To see this, recall that earlier we found that the only proper subvarieties of $\mathbb{A}^1$ are finite. But $k$ is algebraically closed, so infinite. Hence we cannot write $\mathbb{A}^1$ as the union of two proper subvarities! What about the obvious generalization of this to $\mathbb{A}^n$? Turns out that it is indeed true, as we might expect. For the sake of formality I’ll write it up as a lemma. Lemma 3.6 $\mathbb{A}^n$ is irreducible Proof Suppose we could write $\mathbb{A}^n=V(f)\cup V(g)$. By Lemma 2.5 we know that $V(f)\cup V(g) = V((f)\cap (g))$. But $(f)\cap(g)\supset (fg)$ so $V((f)\cap(g))\subset V(fg)$ again by Lemma 2.5. Conversely if $x\in V(fg)$ then either $f(x) = 0$ or $g(x) = 0$, so $x \in V(f)\cup V(g)$. This shows that $V(f)\cup V(g)=V(fg)$. Now $V(fg)=\mathbb{A}^n$ immediately tells us $fg(x) = 0 \ \forall x\in k$. Suppose that $f$ is nonzero. We’ll prove that $g$ is the zero polynomial by induction on $n$. Then $V(g)=\mathbb{A}^n$ so $\mathbb{A}^n$ not irreducible, as required. We first note that since $k$ algebraically closed $k$ infinite. For $n=1$ suppose $f,\ g \neq 0$. Then $f,\ g$ are each zero at finite sets of points. Thus since $k$ infinite, $fg$ is not the zero polynomial, a contradiction. Now let $n>1$.  Consider $f,\ g$ nonzero polynomials in $k[\mathbb{A}^n]$. Fix $x_n \in k$. Then $f,\ g$ polynomials in $k[\mathbb{A}^{n-1}]$. For some $x_n$, $f,\ g$ nonzero as polynomials in $k[\mathbb{A}^{n-1}]$. By the induction hypothesis $fg\neq 0$. This completes the induction. $\blacksquare$ I’ll quickly demonstrate that $\mathbb{A}^n$ is quite strange, when considered as a topological space with the Zariski topology! Indeed let $U$ and $V$ be two nonempty open subsets. Then $U\cap V\neq \emptyset$. Otherwise $\mathbb{A}^n\setminus U,\ \mathbb{A}^n\setminus V$ would be proper closed subsets (affine subvarieties) which covered $\mathbb{A}^n$, violating irreducibility. This is very much not what happens in the Euclidean topology! Similarly we now have a rigorous proof that an open subset $U$ of $\mathbb{A}^n$ is dense. Otherwise $\bar{U}$ and $\mathbb{A}^n\setminus U$ would be proper subvarieties covering $\mathbb{A}^n$. It’s all very well looking for direct examples of irreducible varieties, but in doing so we’ve forgotten about algebra! In fact algebra gives us a big helping hand, as the following theorem shows. For completeness we first recall the definition of a prime ideal. Definition 3.7 $\mathfrak{p}$ is a prime ideal in $R$ iff whenever $fg \in \mathfrak{p}$ we have $f\in \mathfrak{p}$ or $g \in \mathfrak{p}$. Equivalently $\mathfrak{p}$ is prime iff $R/\mathfrak{p}$ is an integral domain. Theorem 3.8 Let $X$ be a nonempty affine variety. Then $X$ irreducible iff $I(X)$ a prime ideal. Proof [“$\Rightarrow$“] Suppose $I(X)$ not prime. Then $\exists f,g \in k[\mathbb{A}^n]$ with $fg \in I(X)$ but $f,\ g \notin I(X)$. Let $J_1 = (I(X),f)$ and $J_2 = (I(X),g)$. Further define $X_1 = V(J_1), \ X_2 = V(J_2)$. Then $V(X_1), \ V(X_2) \subset X$ so proper subsets of $\mathbb{A}^n$. On the other hand $X\subset X_1 \cup X_2$. Indeed if $P\in X$ then $fg(P)=0$ so $f(P)=0$ or $g(P)=0$ so $P \in X_1\cup X_2$. [“$\Leftarrow$“] Suppose $X$ is reducible, that is $\exists X_1,\ X_2$ proper subvarieties of $X$ with $X=X_1\cup X_2$. Since $X_1$ a proper subvariety of $X$ there must exist some element $f \in I(X_1)\setminus I(X)$. Similarly we find $g\in I(X_2)\setminus I(X)$. Hence $fg(P) = 0$ for all $P$ in $X_1\cup X_2 = X$, so certainly $fg \in I(X)$. But this means that $I(X)$ is not prime. $\blacksquare$ This easy theorem is our first real taste of the power that abstract algebra lends to the study of geometry. Let’s see it in action. Recall that a nonzero principal ideal of the ring $k[\mathbb{A}^n]$ is prime iff it is generated by an irreducible polynomial. This is an easy consequence of the fact that $k[\mathbb{A}^n]$ is a UFD. Indeed a nonzero principal ideal is prime iff it is generated by a prime element. But in a UFD every prime is irreducible, and every irreducible is prime! Using the theorem we can say that every irreducible polynomial $f$ gives rise to an irreducible affine hypersurface $X$ s.t. $I(X)=(f)$. Note that we cannot get a converse to this – there’s nothing to say that $I(X)$ must be principal in general. Does this generalise to ideals generated by several irreducible polynomials? We quickly see the answer is no. Indeed take $f = x\, g = x^2 + y^2 -1$ in $k[\mathbb{A}^2]$. These are both clearly irreducible, but $(f,g)$ is not prime. We can see this in two ways. Algebraically $y^2 \in (f,g)$ but $y \notin (f,g)$. Geometrically, recall Lemma 2.5 (3). Also note that by definition $(f,g) = (f)+(g)$. Hence $V(f,g) = V(f)\cap V(g)$. But $V(f) \cap V(g)$ is clearly just two distinct points (the intersection of the line with the circle). Hence it is reducible, and by our theorem $(f,g)$ cannot be prime. We can also use the theorem to exhibit a more informative example of a reducible variety. Consider $X = V(X^2Y - Y^2)$. Clearly $\mathfrak{a}=(X^2Y-Y^2)$ is not prime for $Y(X^2 - Y) \in \mathfrak{a}$ but $Y\notin \mathfrak{a}, \ X^2 - Y \notin \mathfrak{a}$. Noting that $\mathfrak{a}=(X^2-Y)\cap Y$ we see that geometrically $X$ is the union of the $X$-axis and the parabola $Y=X^2$, by Lemma 2.5. Having had such success with prime ideals and irreducible varieties, we might think – what about maximal ideals? Turns out that they have a role to play too. Note that maximal ideals are automatically prime, so any varieties they generate will certainly be irreducible. Definition 3.9 An ideal $\mathfrak{m}$ of $R$ is said to be maximal if whenever $\mathfrak{m}\subset\mathfrak{a}\subset R$ either $\mathfrak{a} = \mathfrak{m}$ or $\mathfrak{a} = R$. Equivalently $\mathfrak{m}$ is maximal iff $R/\mathfrak{m}$ is a field. Theorem 3.10 An affine variety $X$ in $\mathbb{A}^n$ is a point iff $I(X)$ is a maximal ideal. Proof  [“$\Rightarrow$“] Let $X = \{(a_1, \dots , a_n)\}$ be a single point. Then clearly $I(X) = (X_1-a_1,\dots ,X_n-a_n)$. But $k[\mathbb{A}^n]/I(X)$ a field. Indeed $k[\mathbb{A}^n]/I(X)$ isomorphic to $k$ itself, via the isomorphism $X_i \mapsto a_i$. Hence $I(X)$ maximal. [“$\Leftarrow$“] We’ll see this next time. In fact all we need to show is that $(X_1-a_1,\dots,X_n-a_n)$ are the only maximal ideals. $\blacksquare$ Theorems 3.8 and 3.10 are a promising start to our search for a dictionary between algebra and geometry. But they are unsatisfying in two ways. Firstly they tell us nothing about the behaviour of reducible affine varieties – a very large class! Secondly it is not obvious how to use 3.8 to construct irreducibly varieties in general. Indeed there is an inherent asymmetry in our knowledge at present, as I shall now demonstrate. Given an irreducible variety $X$ we can construct it’s ideal $I(X)$ and be sure it is prime, by Theorem 3.8. Moreover we know by Lemma 2.5 that $V(I(X))=X$, a pleasing correspondence. However, given a prime ideal $J$ we cannot immediately say that $V(J)$ is prime. For in Lemma 2.5 there was nothing to say that $I(V(J))=J$, so Theorem 3.8 is useless. We clearly need to find a set of ideals for which $I(V(J))=J$ holds, and hope that prime ideals are a subset of this. It turns out that such a condition is satisfied by a class called radical ideals. Next time we shall prove this, and demonstrate that radical ideals correspond exactly to algebraic varieties. This will provide us with the basic dictionary of algebraic geometry, allowing us to proceed to deeper results. The remainder of this post shall be devoted to radical ideals, and the promised proof of an irreducible decomposition. Definition 3.11 Let $J$ be an ideal in a ring $R$. We define the radical of $J$ to be the ideal $\sqrt{J}=\{f\in R : f^m\in J \ \textrm{some} \ m\in \mathbb{N}\}$. We say that $J$ is a radical ideal if $J=\sqrt{J}$. (That $\sqrt{J}$ is a genuine ideal needs proof, but this is merely a trivial check of the axioms). At first glance this appears to be a somewhat arbitrary definition, though the nomenclature should seem sensible enough. To get a more rounded perspective let’s introduce some other concepts that will become important later. Definition 3.12polynomial function or regular function on an affine variety $X$ is a map $X\rightarrow k$ which is defined by the restriction of a polynomial in $k[\mathbb{A}^n]$ to $X$. More explicitly it is a map $f:X\rightarrow k$ with $f(P)=F(P)$ for all $P\in X$ where $F\in k[\mathbb{A}^n]$ some polynomial. These are eminently reasonable quantities to be interested in. In many ways they are the most obvious functions to define on affine varieties. Regular functions are the analogues of smooth functions in differential geometry, or continuous functions in topology. They are the canonical maps. It is obvious that a regular function $f$ cannot in general uniquely define the polynomial $F$ giving rise to it. In fact suppose $f(P)=F(P)=G(P) \ \forall P \in X$. Then $F-G = 0$ on $X$ so $F-G\in I(X)$. This simple observation explains the implicit claim in the following definition. Definition 3.13 Let $X$ be an affine variety. The coordinate ring $k[X]$ is the ring $k[\mathbb{A}^n]|_X=k[\mathbb{A}^n]/I(X)$. In other words the coordinate ring is the ring of all regular functions on $X$. This definition should also appear logical. Indeed we define the space of continuous functions in topology and the space of smooth functions in differential geometry. The coordinate ring is merely the same notion in algebraic geometry.  The name  ‘coordinate ring’ arises since clearly $k[X]$ is generated by the coordinate functions $x_1,\dots ,x_n$ restricted to $X$. The reason for our notation $k[x_1,\dots ,x_n]=k[\mathbb{A}^n]$ should now be obvious. Note that the coordinate ring is trivially a finitely generated $k$-algebra. The coordinate ring might seem a little useless at present. We’ll see in a later post that it has a vital role in allowing us to apply our dictionary of algebra and geometry to subvarieties. To avoid confusion we’ll stick to $k[\mathbb{A}^n]$ for the near future. The reason for introducing coordinate rings was to link them to radical ideals. We’ll do this via two final definitions. Definition 3.14 An element $x$ of a ring $R$ is called nilpotent if $\exists$ some positive integer $n$ s.t. $x^n=0$. Definition 3.15 A ring $R$ is reduced if $0$ is its only nilpotent element. Lemma 3.16 $R/I$ is reduced iff $I$ is radical. Proof Let $x+I$ be a nilpotent element of $R/I$ i.e. $(x^n + I) = 0$. Hence $x^n \in I$ so by definition $x\in \sqrt{I}=I$. Conversely let $x\in R s.t. x^m \in I$. Then $x^m + I = 0$ in $R/I$ so $x+I = 0+I$ i.e. $x \in I$. $/blacksquare$ Putting this all together we immediately see that the coordinate ring $k[X]$ is a reduced, finitely generated $k$-algebra. That is, provided we assume that for an affine variety $X$, $I(X)$ is radical, which we’ll prove next time. It’s useful to quickly see that these properties characterise coordinate rings of varieties. In fact given any reduced, finitely generated $k$-algebra $A$ we can construct a variety $X$ with $k[X]=A$ as follows. Write $A=k[a_1,\dots ,a_n]$ and define a surjective homomorphism $\pi:k[\mathbb{A}^n]\rightarrow A, \ x_i\mapsto a_i$. Let $I=\textrm{ker}(\pi)$ and $X=V(I)$. By the isomorphism theorem $A = k[\mathbb{A}^n]/I$ so $I$ is radical since $A$ reduced. But then by our theorem next time $X$ an affine variety, with coordinate ring $A$. We’ve come a long way in this post, and congratulations if you’ve stayed with me through all of it! Let’s survey the landscape. In the background we have abstract algebra – systems of equations whose solutions we want to study. In the foreground are our geometrical ideas – affine varieties which represent solutions to the equations. These varieties are built out of irreducible blocks, like Lego. We can match up ideals and varieties according to various criteria. We can also study maps from geometrical varieties down to the ground field using the coordinate ring. Before I go here’s the promised proof that irreducible varieties really are the building blocks we’ve been talking about. Theorem 3.17 Every affine variety $X$ has a unique decomposition as $X_1\cup\dots\cup X_n$ up to ordering, where the $X_i$ are irreducible components and $X_i\not\subset X_j$ for $i\neq j$. Proof (Existence) An affine variety $X$ is either irreducible or $X=Y\cup Z$ with $Y,Z$ proper subset of $X$. We similarly may decompose $Y$ and $Z$ if they are reducible, and so on. We claim that this process stops after finitely many steps. Suppose otherwise, then $X$ contains an infinite sequence of subvarieties $X\supsetneq X_1 \supsetneq X_2 \supsetneq \dots$. By Lemma 2.5 (5) & (7) we have $I(X)\subsetneq I(X_1) \subsetneq I(X_2) \subsetneq \dots$. But $k[\mathbb{A}^n]$ a Noetherian ring by Hilbert’s Basis Theorem, and this contradicts the ascending chain condition! To satisfy the $X_i \not\subset X_j$ condition we simply remove any such $X_i$ that exist in the decomposition we’ve found. (Uniqueness) Suppose we have another decomposition $X=Y_1\cup Y_m$ with $Y_i\not\subset Y_j$ for $i\neq j$. Then $X_i = X_i\cap X = \bigcup_{j=1}^{m}( X_i\cap Y_j)$. Since $X_i$ is irreducible we must have $X_i\cap Y_j = X_i$ for some $j$. In particular $X_i \subset Y_j$. But now by doing the same with the $X$ and $Y$ reversed we find $X_k$ width $X_i \subset Y_j \subset X_k$. But this forces $i=k$ and $Y_j = X_i$. But $i$ was arbitrary, so we are done. $\blacksquare$ If you’re interested in calculating some specific examples of ideals and their associated varieties have a read about Groebner Bases. This will probably become a topic for a post at some point, loosely based on the ideas in Hassett’s excellent book. This question is also worth a skim. I leave you with this enlightening MathOverflow discussion , tackling the irreducibility of polynomials in two variables. Although some of the material is a tad dense, it’s nevertheless interesting, and may be a useful future reference! # Invariant Theory and David Hilbert Health warning: this post is part of a more advanced series on commutative algebra. It may be a little tricky for the layman to understand! David Hilbert was perhaps the greatest mathematicians of the late 19th century. Much of his work laid the foundations for our modern study of commutative algebra. In doing so, he was sometimes said to have killed the study of invariants by solving the central problem in the field. In this post I’ll give a sketch of how he did so. Motivated by Galois Theory we ask the following question. Given a polynomial ring $S = k[x_1,\dots,x_n]$ and a group $G$ acting on $S$ as a $k$-automorphism, what are the elements of $S$ that are invariant under the action of $G$? Following familiar notation we denote this set $S^G$ and note that it certainly forms a subalgebra of $S$. In the late 19th century it was found that $S^G$ could be described fully by a finite set of generators for several suggestive special cases of $G$. It soon became clear that the fundamental problem of invariant theory was to find necessary and sufficient conditions for $S^G$ to be finitely generated. Hilbert’s contribution was an incredibly general sufficient condition, as we shall soon see. To begin with we shall recall the alternative definition of a Noetherian ring. It is a standard proof that this definition is equivalent to that which invokes the ascending chain condition on ideals. As an aside, also recall that the ascending chain condition can be restated by saying that every nonempty collection of ideals has a maximal element. Definition A.1 A ring $R$ is Noetherian if every ideal of $R$ is finitely generated. We shall also recall without proof Hilbert’s Basis Theorem, and draw an easy corollary. Theorem A.2 If $R$ Noetherian then $R[x]$ Noetherian. Corollary A.3 If $S$ is a finitely generated algebra over $R$, with $R$ Noetherian, then $S$ Noetherian. Proof We’ll first show that any homomorphic image of $R$ is Noetherian. Let $I$ be an ideal in the image under than homomorphism $f$. Then $f^{-1}(I)$ an ideal in $R$. Indeed if $k\in f^{-1}(I)$ and $r\in R$ then $f(rk)=f(r)f(k)\in I$ so $rk \in f^{-1}(I)$. Hence $f^{-1}(I)$ finitely generated, so certainly $I$ finitely generated, by the images of the generators of $f^{-1}(I)$. Now we’ll prove the corollary. Since $S$ is a finitely generated algebra over $R$, $S$ is a homomorphic image of $R[x_1,\dots,x_n]$ for some $n$, by the obvious homomorphism that takes each $x_i$ to a generator of $S$. By Theorem A.2 and induction we know that $R[x_1,\dots,x_n]$ is Noetherian. But then by the above, $S$ is Noetherian. $\blacksquare$ Since we’re on the subject of Noetherian things, it’s probably worthwhile introducing the concept of a Noetherian module. The subsequent theorem is analogous to A.3 for general modules. This question ensures that the theorem has content. Definition A.4 An $R$-module $M$ is Noetherian if every submodule $N$ is finitely generated, that is, if every element of $N$ can be written as a polynomial in some generators $\{f_1,\dots,f_n\}\subset N$ with coefficients in $R$. Theorem A.5 If $R$ Noetherian and $M$ a finitely generated $R$-module then $M$ Noetherian. Proof Suppose $M$ generated by $f_1,\dots,f_t$, and let $N$ be a submodule. We show $N$ finitely generated by induction on $t$. If $t=1$ then clearly the map $h:R\rightarrow M$ defined by $1\mapsto f_1$ is surjective. Then the preimage of $N$ is an ideal, just as in A.3, so is finitely generated. Hence $N$ is finitely generated by the images of the generators of $h^{-1}(N)$.  (*) Now suppose $t>1$. Consider the quotient map $h:M \to M/Rf_1$. Let $\tilde{N}$ be the image of $N$ under this map. Then by the induction hypothesis $\tilde{N}$ is finitely generated as it is a submodule of $M/Rf_1$. Let $g_1,\dots,g_s$ be elements of $N$ whose images generate $\tilde{N}$. Since $Rf_1$ is a submodule of $M$ generated by a single element, we have by (*) that it’s submodule $Rf_1\cap N$ is finitely generated, by $h_1,\dots,h_r$ say. We claim that $\{g_1,\dots,g_s,h_1,\dots,h_r\}$ generate $N$. Indeed given $n \in N$ the image of $n \in N$ is a linear combination of the images of the $g_i$. Hence subtracting the relevant linear combination of the $g_i$ from $n$ produces an element of $N \cap Rf_1$ which is precisely a linear combination of the $h_i$ by construction. This completes the induction. $\blacksquare$ We’re now ready to talk about the concrete problem that Hilbert solved using these ideas, namely the existence of finite bases for invariants. We’ll take $k$ to be a field of characteristic $0$ and $G$ to be a finite group, or one of the linear groups $\textrm{ GL}_n(k),\ \textrm{SL}_n(k)$. As in our notation above, we take $S=k[x_1,\dots,x_n]$. Suppose also we are given a group homomorphism $\phi:G \to \textrm{GL}_r(k)$, which of course can naturally be seen as the group of invertible linear transformations of the vector space $V$ over $k$ with basis $x_1,\dots,x_r$. This is in fact the definition of a representation of $G$ on the vector space $V$. As is common practice in representation theory, we view $G$ as acting on $V$ via $(g,v)\mapsto \phi(g)v$. If $G$ is $\textrm{SL}_n(k)$ or $\textrm{GL}_n(k)$ we shall further suppose that our representation of $G$ is rational. That is, the matrices $g \in G$ act on $V$ as matrices whose entries are rational functions in the entries of $g$. (If you’re new to representation theory like me, you might want to read that sentence twice)! We now extend the action of $g\in G$ from $V$ to the whole of $S$ by defining $(g,f)\mapsto f(g^{-1}(x_1),\dots,g^{-1}(x_r),x_{r+1},\dots,x_n)$. Thus we may view $G$ as an automorphism group of $S$. The invariants under $G$ are those polynomials left unchanged by the action of every $g \in G$, and these form a subring of $S$ which we’ll denote $S^G$. Enough set up. To proceed to more interesting territory we’ll need to make another definition. Definition A.6 A polynomial is called homogeneous, homogeneous form, or merely a form, if each of its monomials with nonzero coefficient has the same total degree. Hilbert noticed that the following totally obvious fact about $S^G$ was critically important to the theory of invariants. We may write $S^G$ as a direct sum of the vector spaces $R_i$ of homogeneous forms of degree $i$ that are invariant under $G$. We say that $S^G$ may be graded by degree and use this to motivate our next definition. Definition A.7 A graded ring is a ring $R$ together with a direct sum decomposition as abelian groups $R = R_0 \oplus R_1 \oplus \dots$, such that $R_i R_j \subset R_{i+j}$. This allows us to generalise our notion of homogeneous also. Definition A.8 A homogeneous element of a graded ring $R$ is an element of one of the groups $R_i$. A homogeneous ideal of $R$ is an ideal generated by homogeneous elements. Be warned that clearly homogeneous ideals may contain many inhomogeneous elements! It’s worth mentioning that there was no special reason for taking $\mathbb{N}$ as our indexing set for the $R_i$. We can generalise this easily to $\mathbb{Z}$, and such graded rings are often called $\mathbb{Z}$-graded rings. We won’t need this today, however. Note that if $f \in R$ we have a unique expression for $f$ of the form $f = f_0 + f_1 + \dots + f_n$ with $f_i \in R_i$. (I have yet to convince myself why this terminates generally, any thoughts? I’ve also asked here.) We call the $f_i$homogeneous component of $f$. The next definition is motivated by algebraic geometry, specifically the study of projective varieties. When we arrive at these in the main blog (probably towards the end of this month) it shall make a good deal more sense! Definition A.9 The ideal in a graded ring $R$ generated by all forms of degree greater than $0$ is called the irrelevant ideal and notated $R_+$. Now we return to our earlier example. We may grade the polynomial ring $S=k[x_1,\dots,x_n]$ by degree. In other words we write $S=S_0\oplus S_1 \oplus \dots$ with $S_i$ containing all the forms (homogeneous polynomials) of degree $i$. To see how graded rings are subtly useful, we’ll draw a surprisingly powerful lemma. Lemma A.10 Let $I$ be a homogeneous ideal of a graded ring $R$, with $I$ generated by $f_1,\dots,f_r$. Let $f\in I$ be a homogeneous element. Then we may write $f = \sum f_i g_i$ with $g_i$ homogeneous of degree $\textrm{deg}(f)-\textrm{deg}(f_i)$. Proof We can certainly write $f = \sum f_i G_i$ with $G_i \in R$. Take $g_i$ to be the homogeneous components of $G_i$ of degree $\textrm{deg}(f)-\textrm{deg}(f_i)$. Then all other terms in the sum must cancel, for $f$ is homogeneous by assumption. $\blacksquare$ Now we return to our attempt to emulate Hilbert. We saw earlier that he spotted that grading $S^G$ by degree may be useful. His second observation was this. $\exists$ maps $\phi:S\to S^G$ of $S^G$-modules s.t. (1) $\phi$ preserves degrees and (2) $\phi$ fixes every element of $S^G$. It is easy to see that this abstract concept corresponds intuitively to the condition that $S^G$ be a summand of the graded ring $S$. This is trivial to see in the case that $G$ is a finite group. Indeed let $\phi (f) = \frac{1}{|G|}\sum_{g\in G} g(f)$. Note that we have implicitly used that $k$ has characteristic zero to ensure that the multiplicative inverse to $|G|$ exists. In the case that $G$ is a linear group acting rationally, then the technique is to replace the sum by an integral. The particulars of this are well beyond the scope of this post however! We finally prove the following general theorem. We immediately get Hilbert’s result on the finite generation of classes of invariants by taking $R=S^G$. Theorem A.11 Take$k$ a field and $S=k[x_1,\dots,x_n]$ a polynomial ring graded by degree. Let $R$ be a $k$-subalgebra of $S$. Suppose $R$ is a summand of $S$, in the sense described above. Then $R$ is finitely generated as a $k$-algebra. Proof Let $I\subset R$ be the ideal of $R$ generated by all homogeneous elements of degree $> 0$. By the Basis Theorem $S$ is Noetherian, and $IS$ an ideal of $S$, so finitely generated. By splitting each generator into its homogeneous components we may assume that $IS$ is generated by some homogeneous elements $f_1,\dots,f_s$ which we may wlog assume lie in $I$. We’ll prove that these elements precisely generate $R$ as a $k$-algebra. Now let $R'$ be the $k$-subalgebra of $S$ generated by $f_1,\dots,f_s$ and take $f\in R$ a general homogeneous polynomial. Suppose we have shown $f\in R'$. Let $g$ be a general element of $R$. Then certainly $g\in S$ a sum of homogeneous components. But $R$ a summand of $S$, so applying the given map $\phi$ we have that the homogeneous components are wlog in $R$. Thus $g\in R'$ also, and we are done. It only remains to prove $f \in R'$ which we’ll do by induction on the degree of $f$. If $\textrm{deg}(f)=0$ then $f\in K\subset R'$. Suppose $\textrm{deg}(f)>0$ so $f\in I$. Since the $f_i$ generate $IS$ as a homogeneous ideal of $S$ we may write $f = \sum f_i g_i$ with $g_i$ homogeneous of degree $\textrm{deg}(f)-\textrm{deg}(f_i)<\textrm{deg}(f)$ by Lemma A.10. But again we may use the map $\phi$ obtained from our observation that $R$ a summand of $S$. Indeed then $f=\sum \phi(g_i)f_i$ for $f,\ f_i \in R$. But $\phi$ preserves degrees so $\phi(g_i)$ 0f lower degree than $f$. Thus by the induction hypothesis $\phi(g_i) \in R'$ and hence $f\in R'$ as required. $\blacksquare$ It’s worth noting that such an indirect proof caused quite a furore when it was originally published in the late 19th century. However the passage of time has provided us with a broader view of commutative algebra, and techniques such as this are much more acceptable to modern tastes! Nevertheless I shall finish by making explicit two facts that help to explain the success of our argument. We’ll first remind ourselves of a useful definition of an algebra. Definition A.12 An $R$-algebra $S$ is a ring $S$ which has the compatible structure of a module over $R$ in such a way that ring multiplication is $R$-bilinear. It’s worth checking that this intuitive definition completely agrees with that we provided in the Background section, as is clearly outlined on the Wikipedia page.  The following provide an extension and converse to Corollary A.3 (that finitely generated algebras over fields are Noetherian) in the special case that $R$ a graded ring. Lemma A.13 $S=R_0\oplus R_1 \oplus \dots$ a Noetherian graded ring iff $R_0$ Noetherian and $S$ a finitely generated $R_0$ algebra. Lemma A.14 Let $S$ be a Noetherian graded ring, $R$ a summand of $S$. Then $R$ Noetherian. We’ll prove these both next time. Note that they certainly aren’t true in general when $S$ isn’t graded!
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 873, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.921130359172821, "perplexity": 277.93976921091735}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-34/segments/1502886117519.92/warc/CC-MAIN-20170823035753-20170823055753-00625.warc.gz"}
http://gmatclub.com/forum/at-a-certain-hospital-75-of-the-interns-receive-fewer-than-108658.html?kudos=1
Find all School-related info fast with the new School-Specific MBA Forum It is currently 25 May 2016, 14:05 ### GMAT Club Daily Prep #### Thank you for using the timer - this advanced tool can estimate your performance and suggest more practice questions. We have subscribed you to Daily Prep Questions via email. Customized for You we will pick new questions that match your level based on your Timer History Track every week, we’ll send you an estimated GMAT score based on your performance Practice Pays we will pick new questions that match your level based on your Timer History # Events & Promotions ###### Events & Promotions in June Open Detailed Calendar # At a certain hospital, 75% of the interns receive fewer than Author Message TAGS: ### Hide Tags Current Student Status: Up again. Joined: 31 Oct 2010 Posts: 541 Concentration: Strategy, Operations GMAT 1: 710 Q48 V40 GMAT 2: 740 Q49 V42 Followers: 18 Kudos [?]: 270 [4] , given: 75 At a certain hospital, 75% of the interns receive fewer than [#permalink] ### Show Tags 03 Feb 2011, 09:23 4 KUDOS 19 This post was BOOKMARKED 00:00 Difficulty: 55% (hard) Question Stats: 61% (02:43) correct 39% (01:33) wrong based on 679 sessions ### HideShow timer Statistics At a certain hospital, 75% of the interns receive fewer than 6 hours of sleep and report feeling tired during their shifts. At the same time, 70% of the interns who receive 6 or more hours of sleep report no feelings of tiredness. If 80% of the interns receive fewer than 6 hours of sleep, what percent of the interns report no feelings of tiredness during their shifts? A. 6 B. 14 C. 19 D. 20 E. 81 [Reveal] Spoiler: OA _________________ My GMAT debrief: from-620-to-710-my-gmat-journey-114437.html Math Expert Joined: 02 Sep 2009 Posts: 32990 Followers: 5751 Kudos [?]: 70473 [7] , given: 9845 ### Show Tags 03 Feb 2011, 09:49 7 KUDOS Expert's post gmatpapa wrote: At a certain hospital, 75% of the interns receive fewer than 6 hours of sleep and report feeling tired during their shifts. At the same time, 70% of the interns who receive 6 or more hours of sleep report no feelings of tiredness. If 80% of the interns receive fewer than 6 hours of sleep, what percent of the interns report no feelings of tiredness during their shifts? A. 6 B. 14 C. 19 D. 20 E. 81 Given: < 6 hours sleep >= 6 hours sleep TOTAL Tired 75 - - Not Tired - 0.7*X ? TOTAL 80 X 100 Interns who receive < than 6 hours sleep and Not Tired = 80 - 75 = 5; TOTAL interns who receive >= 6 hours sleep = 100 - 80 = 20, so interns who receive >= 6 hours sleep and are Not Tired = 0.7 * 20 =14; Interns who are Not Tired = 5 + 14 = 19. _________________ Senior Manager Joined: 24 Mar 2011 Posts: 457 Location: Texas Followers: 5 Kudos [?]: 113 [3] , given: 20 Re: Sets problem - Tired [#permalink] ### Show Tags 25 Jul 2011, 22:11 3 KUDOS the solution is in the table - Attachments table.JPG [ 8 KiB | Viewed 13615 times ] Math Expert Joined: 02 Sep 2009 Posts: 32990 Followers: 5751 Kudos [?]: 70473 [2] , given: 9845 ### Show Tags 03 Feb 2011, 10:38 2 KUDOS Expert's post 1 This post was BOOKMARKED gmatpapa wrote: Thanks! Bunuel, this tabular method for solving Venn Diagram problems: is it explained anywhere with other examples? It's not a Venn Diagram problem, it's an overlapping sets problem which can be solved with Venn Diagram as well as with Double Matrix Method (which is often better). You can check overlapping sets tag to see that some of the problems are solved with this method. Also check this link for video explanation: http://www.youtube.com/watch?v=uE_aMa_W0o0 _________________ Veritas Prep GMAT Instructor Joined: 16 Oct 2010 Posts: 6565 Location: Pune, India Followers: 1789 Kudos [?]: 10757 [1] , given: 210 Re: Sets problem - Tired [#permalink] ### Show Tags 25 Jul 2011, 21:38 1 KUDOS Expert's post vatsas wrote: At a certain hospital, 75% of the interns receive fewer than 6 hours of sleep and report feeling tired during their shifts. At the same time, 70% of the interns who receive 6 or more hours of sleep report no feelings of tiredness. If 80% of the interns receive fewer than 6 hours of sleep, what percent of the interns report no feelings of tiredness during their shifts? a. 6 b. 14 c. 19 d. 20 e. 81 I need to understand the method to solve this problem. I tried the matrix method but failed. Is there any other method? Thanks Sumeet You can just use logic or a Venn diagram, whatever works for you. I will give the logical solution below. Try to figure out the Venn diagram for it. Get back if it gets tricky. Look at the given data line by line. At a certain hospital, 75% of the interns receive fewer than 6 hours of sleep and report feeling tired during their shifts. n(FEW and TIRED) = 75% At the same time, 70% of the interns who receive 6 or more hours of sleep report no feelings of tiredness. .... We will come back... If 80% of the interns receive fewer than 6 hours of sleep, n(FEW) = 80% Therefore, n(Not FEW) = 20% Also since n(FEW and TIRED) = 75%, n(FEW and Not TIRED) = 5% Now we go back to the second sentence. We know that n(Not FEW) = 20%. 70% of these are not tired. So n(Not TIRED and Not FEW) = 70% of 20% = 14% what percent of the interns report no feelings of tiredness during their shifts? n(Not TIRED) = n(FEW and Not TIRED) + n(Not TIRED and Not FEW) = 5% + 14% = 19% The notation is used only to explain it to you. You obviously don't need to do that. Just remember that when you get some data '80% people get less sleep,' it also implies that 20% get enough or more. _________________ Karishma Veritas Prep | GMAT Instructor My Blog Get started with Veritas Prep GMAT On Demand for $199 Veritas Prep Reviews Veritas Prep GMAT Instructor Joined: 16 Oct 2010 Posts: 6565 Location: Pune, India Followers: 1789 Kudos [?]: 10757 [1] , given: 210 Re: Tuesday Q2 - Interns [#permalink] ### Show Tags 21 Jun 2012, 21:58 1 This post received KUDOS Expert's post RichaMidha2012 wrote: 5% of 80 =4 but why you guys took it as 5% am I mistaking something asterixmatrix wrote: 80% receive less than 6 hours of sleep 75% report feeling tired so 5% dont feel tired 20% receive 6 or more hours of sleep out of this 70% report no tiredness so it become 14% total 5%+14% = 19%...C A very important thing in percentages is to note what is the % of i.e. when I say 5%, it is 5% of what? 5% of total interns or 5% of interns who receive less than 6 hrs of sleep or 5% of interns who report tiredness etc The question says: 75% of the interns receive fewer than 6 hours of sleep and report feeling tired during their shifts (75% of total interns) If 80% of the interns receive fewer than 6 hours of sleep (again, 80% of the total interns) So out of 80% of total interns i.e. if total interns are 100, 80 of them receive fewer than 6 hrs of sleep. Out of 100, 75 receive fewer than 6 hrs of sleep and report tiredness. Therefore, 5 receive fewer than 6 hrs of sleep but report no tiredness. _________________ Karishma Veritas Prep | GMAT Instructor My Blog Get started with Veritas Prep GMAT On Demand for$199 Veritas Prep Reviews MBA Section Director Affiliations: GMAT Club Joined: 22 Feb 2012 Posts: 2559 Location: India City: Pune GPA: 3.4 Followers: 309 Kudos [?]: 2231 [1] , given: 1773 Re: At a certain hospital, 75% of the interns receive fewer than [#permalink] ### Show Tags 16 Feb 2013, 12:39 1 KUDOS Expert's post thinktank wrote: Guys, I am like SUPER CONFUSED right now .. I used the tabular method to solve the problem but it was just the opposite to the explanations (ie). Tired and not tired in the X axis, Receive < 6 hrs , >6 hours in the Y axis .. I know this would yield the same answer but im confused.. If someone could explain by drawing the grid the way i did.. id be really really grateful. And also, how do u choose which attribute to go where.. does that really matter?? for explanation, Pls go thru the attachment. see if it makes sense. Regards, Abhijit Attachments Book1.xlsx [10.95 KiB] _________________ e-GMAT Representative Joined: 04 Jan 2015 Posts: 338 Followers: 92 Kudos [?]: 707 [1] , given: 84 Re: At a certain hospital, 75% of the interns receive fewer than [#permalink] ### Show Tags 11 Jun 2015, 22:36 1 KUDOS Expert's post ankushbagwale wrote: How this queston can be done with venn diagram. Can anyone suggest. I have tried the following combinations for the circles: a. Tired one circle & Few than 6 hrs another circle. Here How do I represent the balance more or equal to 6 hrs ( by another circle??) 2. Non tired one & more than 6 hrs another circle. Still not gettign there. Hi ankushbagwale, Presenting the Venn diagram approach for the question. Please refer the diagram below: Given Number of interns who receive fewer than 6 hours of sleep and report feeling tired = $$b$$ = 75% of total interns. Number of interns who receive 6 or more hours of sleep and report no feeling of tiredness = $$d$$ = 70% of interns who receive more than 6 hours of sleep Number of interns who receive less than 6 hours of sleep = $$a + b = 80$$% of total interns We are asked to find the percent of interns who report no feeling of tiredness. Number of interns who report no feeling of tiredness = $$a + d$$. Hence we need to find $$\frac{a +d}{a + b + c+ d}$$. Approach As we are not given any info about the number of interns, to find the ratio of of interns who report no feeling of tiredness to the total interns we need to express them in the same terms. With this understanding let's proceed to the working out. Working Out Let's assume the total number of interns to be $$x$$. i.e. $$a + b + c + d = x$$ $$b = 75$$% of total interns $$= 0.75x$$ $$a + b = 80$$% of total interns $$= 0.8x$$ which would give us $$a = 0.05x$$ It also tells us that $$c + d = 0.2x$$. $$d = 70$$% of interns who receive more than 6 hours of sleep Interns who received more than 6 hours of sleep $$= c +d$$ So, we can write $$d = 70$$% of $$(c + d) = 70$$% of $$0.2x = 0.14x$$ Hence $$a + d = 0.05x + 0.14x = 0.19x$$ So $$\frac{a +d}{a + b + c+ d} = \frac{0.19x}{x} = 19$$% Hope it's clear Regards Harsh _________________ Current Student Status: Up again. Joined: 31 Oct 2010 Posts: 541 Concentration: Strategy, Operations GMAT 1: 710 Q48 V40 GMAT 2: 740 Q49 V42 Followers: 18 Kudos [?]: 270 [0], given: 75 ### Show Tags 03 Feb 2011, 10:25 1 This post was BOOKMARKED Bunuel wrote: gmatpapa wrote: At a certain hospital, 75% of the interns receive fewer than 6 hours of sleep and report feeling tired during their shifts. At the same time, 70% of the interns who receive 6 or more hours of sleep report no feelings of tiredness. If 80% of the interns receive fewer than 6 hours of sleep, what percent of the interns report no feelings of tiredness during their shifts? A. 6 B. 14 C. 19 D. 20 E. 81 Given: < 6 hours sleep >= 6 hours sleep TOTAL Tired 75 - - Not Tired - 0.7*X ? TOTAL 80 X 100 Interns who receive < than 6 hours sleep and Not Tired = 80 - 75 = 5; TOTAL interns who receive >= 6 hours sleep = 100 - 80 = 20, so interns who receive >= 6 hours sleep and are Not Tired = 0.7 * 20 =14; Interns who are Not Tired = 5 + 14 = 19. Thanks! Bunuel, this tabular method for solving Venn Diagram problems: is it explained anywhere with other examples? _________________ My GMAT debrief: from-620-to-710-my-gmat-journey-114437.html Intern Joined: 11 Sep 2011 Posts: 26 Followers: 1 Kudos [?]: 28 [0], given: 0 Re: At a certain hospital, 75% of the interns receive fewer than [#permalink] ### Show Tags 01 Apr 2012, 00:41 Let's assume total number of students are 1000. 80% of 1000 students gets <6 hrs of sleep = 800 (both tired and not tired) 75% of 1000 students gets <6 hrs and tired = 750 Therefore, students who get <6 hrs and not tired = 50..........(1) 20% of 1000 students gets >= 6 hrs of sleep = 200 (both tired and not tired) 70% of 200 students gets >=6 and not tired = 140.......(2) (1) + (2) = 190 or 19% of total students. Hence C. Manager Joined: 27 Oct 2011 Posts: 191 Location: United States Concentration: Finance, Strategy GMAT 1: Q V GPA: 3.7 WE: Account Management (Consumer Products) Followers: 4 Kudos [?]: 119 [0], given: 4 At a certain hospital, 75% of the interns receive fewer than [#permalink] ### Show Tags 15 Apr 2012, 15:31 option c. we figure out at that 20% get more than 6 hours of sleep. then from there we know that 70% of those who get more than 6 hrs of sleep feel tired which is 14% + 5% = 19% 5% = 80-75 _________________ DETERMINED TO BREAK 700!!! Senior Manager Joined: 30 Jun 2011 Posts: 274 Followers: 0 Kudos [?]: 43 [0], given: 20 Re: At a certain hospital, 75% of the interns receive fewer than [#permalink] ### Show Tags 17 Apr 2012, 06:54 Thanks Bunuel for the link to the video Intern Joined: 24 Oct 2011 Posts: 2 GMAT Date: 07-27-2012 Followers: 1 Kudos [?]: 0 [0], given: 4 ### Show Tags 21 Jun 2012, 18:53 5% of 80 =4 but why you guys took it as 5% am I mistaking something hogann wrote: asterixmatrix wrote: 80% receive less than 6 hours of sleep 75% report feeling tired so 5% dont feel tired 20% receive 6 or more hours of sleep out of this 70% report no tiredness so it become 14% total 5%+14% = 19%...C Great Explanation - OA is 19% Moderator Joined: 02 Jul 2012 Posts: 1230 Location: India Concentration: Strategy GMAT 1: 740 Q49 V42 GPA: 3.8 WE: Engineering (Energy and Utilities) Followers: 105 Kudos [?]: 1159 [0], given: 116 ### Show Tags 28 Oct 2012, 23:56 AartiPajan wrote: At a certain hospital, 75% of the interns receive fewer than 6 hours of sleep and report feeling tired during their shifts. At the same time, 70% of interns who receive 6 hours or more of sleep report no feeling of tiredness. If 80% of the interns receive fewer than 6 hrs sleep, what percentage of interns report no feelings of tiredness during their shift? A. 6 B. 14 C. 19 D. 20 E. 81 Plz help... My solution says D. 20 while correct answer is C. 19 .. Creating a two overlapping set grid ____________Less Than 6________6 or more__________________Total Tired__________75________________6______________________81 Not Tired________5_______________14______________________19 Total__________80_______________20_____________________100 Kudos Please... If my post helped. _________________ Did you find this post helpful?... Please let me know through the Kudos button. Thanks To The Almighty - My GMAT Debrief GMAT Reading Comprehension: 7 Most Common Passage Types Manager Joined: 28 Aug 2012 Posts: 52 Concentration: Operations, Marketing GMAT 1: 510 Q36 V25 GPA: 4 WE: Information Technology (Other) Followers: 0 Kudos [?]: 95 [0], given: 105 Re: At a certain hospital, 75% of the interns receive fewer than [#permalink] ### Show Tags 16 Feb 2013, 12:00 Guys, I am like SUPER CONFUSED right now .. I used the tabular method to solve the problem but it was just the opposite to the explanations (ie). Tired and not tired in the X axis, Receive < 6 hrs , >6 hours in the Y axis .. I know this would yield the same answer but im confused.. If someone could explain by drawing the grid the way i did.. id be really really grateful. And also, how do u choose which attribute to go where.. does that really matter?? Senior Manager Joined: 13 Aug 2012 Posts: 464 Concentration: Marketing, Finance GMAT 1: Q V0 GPA: 3.23 Followers: 22 Kudos [?]: 346 [0], given: 11 Re: At a certain hospital, 75% of the interns receive fewer than [#permalink] ### Show Tags 03 Mar 2013, 21:26 -------------Fewer than 6 / 6 or more --not tired-- .75x. --tired------ .05x. /------.14x-------/---.19x---- Total-----------.80x-------/------.20x-------/---x------- _________________ Impossible is nothing to God. Intern Joined: 23 Feb 2013 Posts: 8 Followers: 0 Kudos [?]: 0 [0], given: 22 Re: Sets problem - Tired [#permalink] ### Show Tags 07 Apr 2013, 17:19 agdimple333 wrote: the solution is in the table - I'm trying to understand why you had to calculate "70% of 20" but at the same time you left "5" untouched? Why would we take % of a %? Math Expert Joined: 02 Sep 2009 Posts: 32990 Followers: 5751 Kudos [?]: 70473 [0], given: 9845 Re: Sets problem - Tired [#permalink] ### Show Tags 07 Apr 2013, 22:45 Expert's post gmat978 wrote: agdimple333 wrote: the solution is in the table - I'm trying to understand why you had to calculate "70% of 20" but at the same time you left "5" untouched? Why would we take % of a %? _________________ Math Expert Joined: 02 Sep 2009 Posts: 32990 Followers: 5751 Kudos [?]: 70473 [0], given: 9845 Re: At a certain hospital, 75% of the interns receive fewer than [#permalink] ### Show Tags 05 Jul 2013, 02:27 Expert's post 1 This post was BOOKMARKED Bumping for review and further discussion*. Get a kudos point for an alternative solution! *New project from GMAT Club!!! Check HERE _________________ Current Student Joined: 05 Aug 2012 Posts: 42 Concentration: Technology, General Management GMAT Date: 09-16-2013 GPA: 3.5 WE: Analyst (Consulting) Followers: 0 Kudos [?]: 22 [0], given: 11 ### Show Tags 12 Sep 2013, 11:25 sidmohan84 wrote: I am havign difficulty understanding teh below problem: At a certain hospital, 75% of the interns receive fewer than 6 hours of sleep and report feeling tired during their shifts. At the same time, 70% of the interns who receive 6 or more hours of sleep report no feelings of tiredness. If 80% of the interns receive fewer than 6 hours of sleep, what percent of the interns report no feelings of tiredness during their shifts? A) 6 B) 14 C) 19 D) 20 E) 81 What is the relation between the 75% and the 80% stated. Please explain with solution. Hello, There are two things that you need to understand: One is set of intern which receive less than 6 hours of sleep and at the same time they feel tired. As per the question, the % of such interns is 75%. Second set talks about intern who receive less than 6 hours of sleep, however it does not talk anything about whether this set of intern felt tired or not. As per the question, the % of such intern is 80%. Since it is a overlapping set question, we can figure the % of intern who slept of less than 6 Hrs, but didn't felt tired. Hence the % comes out to be 80-75 = 5%. Rest is well explained in earlier posts. _________________ Regards, Suyash I want to live in a world where emails are short, love letters are brave, and every "Thank you" note is scribbled by hand. GO GREEN Re: Unsettling Sets   [#permalink] 12 Sep 2013, 11:25 Go to page    1   2    Next  [ 24 posts ] Similar topics Replies Last post Similar Topics: 8 Of the final grades received by the students in a certain 6 26 Dec 2012, 07:04 A certain Social Security recipient will receive an annual b 4 17 Mar 2011, 14:10 7 If x is a positive integer with fewer than 3 digits, what is 13 10 Dec 2010, 07:44 1 A certain auto manufacturer sold 3% fewer vehicles in 2007 4 16 Oct 2010, 13:29 6 At a certain hospital, 75% of the interns receive fewer than 8 22 Sep 2008, 02:09 Display posts from previous: Sort by
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.20188608765602112, "perplexity": 6702.388494056356}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-22/segments/1464049275328.63/warc/CC-MAIN-20160524002115-00202-ip-10-185-217-139.ec2.internal.warc.gz"}
https://www.dealii.org/current/doxygen/deal.II/structPETScWrappers_1_1PreconditionParaSails_1_1AdditionalData.html
Reference documentation for deal.II version 9.4.1 Searching... No Matches #include <deal.II/lac/petsc_precondition.h> ## Public Member Functions AdditionalData (const unsigned int symmetric=1, const unsigned int n_levels=1, const double threshold=0.1, const double filter=0.05, const bool output_details=false) ## Public Attributes unsigned int symmetric unsigned int n_levels double threshold double filter bool output_details ## Detailed Description Standardized data struct to pipe additional flags to the preconditioner. Definition at line 815 of file petsc_precondition.h. ## Constructor & Destructor Documentation PETScWrappers::PreconditionParaSails::AdditionalData::AdditionalData ( const unsigned int symmetric = 1, const unsigned int n_levels = 1, const double threshold = 0.1, const double filter = 0.05, const bool output_details = false ) Constructor. Definition at line 664 of file petsc_precondition.cc. ## ◆ symmetric This parameter specifies the type of problem to solve: • 0: nonsymmetric and/or indefinite problem, and nonsymmetric preconditioner • 1: SPD problem, and SPD (factored) preconditioner • 2: nonsymmetric, definite problem, and SPD (factored) preconditioner Default is symmetric = 1. Definition at line 837 of file petsc_precondition.h. ## ◆ n_levels The sparsity pattern used for the approximate inverse is the pattern of a power B^m where B has been sparsified from the given matrix A, n_level is equal to m+1. Default value is n_levels = 1. Definition at line 845 of file petsc_precondition.h. ## ◆ threshold Sparsification is performed by dropping nonzeros which are smaller than thresh in magnitude. Lower values of thresh lead to more accurate, but also more expensive preconditioners. Default value is thresh = 0.1. Setting thresh < 0 a threshold is selected automatically, such that -thresh represents the fraction of nonzero elements that are dropped. For example, if thresh = -0.9, then B will contain about ten percent of the nonzeros of the given matrix A. Definition at line 857 of file petsc_precondition.h. ## ◆ filter Filtering is a post-processing procedure, filter represents a fraction of nonzero elements that are dropped after creating the approximate inverse sparsity pattern. Default value is filter = 0.05. Setting filter < 0 a value is selected automatically, such that -filter represents the fraction of nonzero elements that are dropped. For example, if thresh = -0.9, then about 90 percent of the entries in the computed approximate inverse are dropped. Definition at line 869 of file petsc_precondition.h.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3049266040325165, "perplexity": 4851.715617494715}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-49/segments/1669446711150.61/warc/CC-MAIN-20221207053157-20221207083157-00388.warc.gz"}
https://ask.sagemath.org/answers/13437/revisions/
# Revision history [back] Note that the type of output depends on the base ring of the matrix, for example a floating point (double) matrix sage: matrix(RDF,[[1/2,-1/4],[-1/2,1/3]]).eigenvalues() [0.779908245295, 0.0534250880383] yields a floating-point result. A matrix over the rationals sage: matrix(QQ,[[1/2,-1/4],[-1/2,1/3]]).eigenvalues() [0.05342508803827721?, 0.779908245295057?] returns eigenvalues in the splitting field which uses interval arithmetic. For arbitrary precision floating point numbers, there is no special implementation, sage: matrix(RR,[[1/2,-1/4],[-1/2,1/3]]).eigenvalues() /home/vbraun/opt/sage-5.0.beta11/local/bin/sage-ipython:1: UserWarning: Using generic algorithm for an inexact ring, which will probably give incorrect results due to numerical precision issues. #!/usr/bin/env python [0.779908245295056, 0.0534250880382772] so Sage falls back to the lapack implementation which doesn't track precision. Finally, over the symbolic ring sage: matrix(SR,[[1/2,-1/4],[-1/2,1/3]]).eigenvalues() [-1/12*sqrt(19) + 5/12, 1/12*sqrt(19) + 5/12]
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.44980505108833313, "perplexity": 3305.4249044256585}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780057119.85/warc/CC-MAIN-20210920221430-20210921011430-00426.warc.gz"}
http://openstudy.com/updates/505d14f1e4b02e13940fdb50
## Got Homework? ### Connect with other students for help. It's a free community. • across Online now • laura* Helped 1,000 students Online now • Hero College Math Guru Online now Here's the question you clicked on: 55 members online • 0 viewing ## MathSofiya Group Title I'm try to strip out a term to have the sum start at n=0 but it turns out that I don't have to anything...I still don't understand why =/ $0+\sum_{n=1}^\infty a_n nx^n=\sum_{n=0}^\infty a_n nx^n$ one year ago one year ago Edit Question Delete Cancel Submit • This Question is Closed 1. MathSofiya Group Title Best Response You've already chosen the best response. 0 @lgbasallote • one year ago 2. MathSofiya Group Title Best Response You've already chosen the best response. 0 @Algebraic! • one year ago 3. MathSofiya Group Title Best Response You've already chosen the best response. 0 @Callisto • one year ago 4. MathSofiya Group Title Best Response You've already chosen the best response. 0 @saifoo.khan • one year ago 5. mahmit2012 Group Title Best Response You've already chosen the best response. 1 |dw:1348282976204:dw| • one year ago 6. MathSofiya Group Title Best Response You've already chosen the best response. 0 • one year ago 7. mahmit2012 Group Title Best Response You've already chosen the best response. 1 |dw:1348283328593:dw| • one year ago 8. mahmit2012 Group Title Best Response You've already chosen the best response. 1 |dw:1348283382718:dw| • one year ago 9. mahmit2012 Group Title Best Response You've already chosen the best response. 1 |dw:1348283519218:dw| • one year ago 10. mahmit2012 Group Title Best Response You've already chosen the best response. 1 Did you get my patern to solve DE with series?. • one year ago 11. mahmit2012 Group Title Best Response You've already chosen the best response. 1 I have never used sigma ! • one year ago 12. MathSofiya Group Title Best Response You've already chosen the best response. 0 Yeah the series is solved • one year ago 13. MathSofiya Group Title Best Response You've already chosen the best response. 0 I was just stuck on understanding that particular step but now I get it. • one year ago 14. mahmit2012 Group Title Best Response You've already chosen the best response. 1 • one year ago • Attachments: ## See more questions >>> ##### spraguer (Moderator) 5→ View Detailed Profile 23 • Teamwork 19 Teammate • Problem Solving 19 Hero • You have blocked this person. • ✔ You're a fan Checking fan status... Thanks for being so helpful in mathematics. If you are getting quality help, make sure you spread the word about OpenStudy.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9995585083961487, "perplexity": 29891.47527666683}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-23/segments/1405997858892.28/warc/CC-MAIN-20140722025738-00183-ip-10-33-131-23.ec2.internal.warc.gz"}
http://mathhelpforum.com/algebra/118066-algebra-2-word-problem-2-a-print.html
Algebra 2 Word Problem #2 • Dec 2nd 2009, 02:42 PM kelsikels Algebra 2 Word Problem #2 1.) The speed of a stream is 5 mph. If a boat travels 82 miles downstream in the same time that it takes to travel 41 miles upstream, what is the speed of the boat in still water? THANK YOU! • Dec 2nd 2009, 02:48 PM skeeter Quote: Originally Posted by kelsikels 1.) The speed of a stream is 5 mph. If a boat travels 82 miles downstream in the same time that it takes to travel 41 miles upstream, what is the speed of the boat in still water? THANK YOU! basic idea for setting up equations is (rate)(time) = distance let $v$ = speed of the boat in mph with no current downstream ... $(v+5)t = 82$ upstream ... $(v-5)t = 41$ solve the system for $v$
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 4, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8536911606788635, "perplexity": 1250.603054794017}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917119120.22/warc/CC-MAIN-20170423031159-00121-ip-10-145-167-34.ec2.internal.warc.gz"}
http://tex.stackexchange.com/questions/5019/table-of-contents-using-article-and-report-classes
When I make a table of contents with the article and report classes the results are different. I have two questions. 1. Why does this happen? 2. How can I make the report ToC look like the article one? Here are a pair of examples: \documentclass[]{report} \usepackage{lipsum} \begin{document} \tableofcontents \section{Section} \lipsum \end{document} \documentclass[]{article} \usepackage{lipsum} \begin{document} \tableofcontents \section{Section} \lipsum \end{document} - By default, the "top-level" entries in a table of contents will be typeset in bold and without a dotted line between entry name and page number. For the article class, the "top level" is \section, while for the report (and the book) class it's \chapter. Simply changing \section to \chapter in your first example will result in (roughly) the same TOC layout. 2. Consider to use a package to design the table of contents, for instance tocloft or titletoc. By default, in the report class, the tocdepth counter is set to 2 (so that the "lowest-ranking" heading going into the TOC is \subsection), while in the article class tocdepth is 3 (so that a \subsubsection will make it into the TOC). It seems Leslie Lamport didn't have different TOC layouts for report vs. article in mind. –  lockstep Nov 6 '10 at 21:40
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7109805941581726, "perplexity": 1932.8338299827471}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-06/segments/1422121934081.85/warc/CC-MAIN-20150124175214-00044-ip-10-180-212-252.ec2.internal.warc.gz"}
https://www.projecteuclid.org/euclid.aaa/1425048229
## Abstract and Applied Analysis ### Dynamics of an Almost Periodic Food Chain System with Impulsive Effects #### Abstract In order to obtain a more accurate description of the ecological system perturbed by human exploitation activities such as planting and harvesting, we need to consider the impulsive differential equations. Therefore, by applying the comparison theorem and the Lyapunov method of the impulsive differential equations, this paper gives some new sufficient conditions for the permanence and existence of a unique uniformly asymptotically stable positive almost periodic solution in a food chain system with almost periodic impulsive perturbations. The method used in this paper provides a possible method to study the permanence and existence of a unique uniformly asymptotically stable positive almost periodic solution of the models with impulsive perturbations in biological populations. Finally, an example and numerical simulations are given to illustrate that our results are feasible. #### Article information Source Abstr. Appl. Anal., Volume 2014, Special Issue (2014), Article ID 324912, 10 pages. Dates First available in Project Euclid: 27 February 2015 https://projecteuclid.org/euclid.aaa/1425048229 Digital Object Identifier doi:10.1155/2014/324912 Mathematical Reviews number (MathSciNet) MR3246329 Zentralblatt MATH identifier 07022172 #### Citation Li, Yaqin; Wu, Wenquan; Zhang, Tianwei. Dynamics of an Almost Periodic Food Chain System with Impulsive Effects. Abstr. Appl. Anal. 2014, Special Issue (2014), Article ID 324912, 10 pages. doi:10.1155/2014/324912. https://projecteuclid.org/euclid.aaa/1425048229 #### References • A. A. Berryman, “The origins and evolution of predator-prey theory,” Ecology, vol. 73, no. 5, pp. 1530–1535, 1992. • Y. G. Sun and S. H. Saker, “Positive periodic solutions of discrete three-level food-chain model of Holling type II,” Applied Mathematics and Computation, vol. 180, no. 1, pp. 353–365, 2006. • Y. K. Li, L. H. Lu, and X. Y. Zhu, “Existence of periodic solutions in n-species food-chain system with impulsive,” Nonlinear Analysis: Real World Applications, vol. 7, no. 3, pp. 414–431, 2006. • R. Xu, L. Chen, and F. Hao, “Periodic solutions of a discrete time Lotka-Volterra type food-chain model with delays,” Applied Mathematics and Computation, vol. 171, no. 1, pp. 91–103, 2005. • C. X. Shen, “Permanence and global attractivity of the food-chain system with Holling IV type functional response,” Applied Mathematics and Computation, vol. 194, no. 1, pp. 179–185, 2007. • H. Baek, “A food chain system with Holling type IV functional response and impulsive perturbations,” Computers & Mathematics with Applications, vol. 60, no. 5, pp. 1152–1163, 2010. • G. H. Cui and X. P. Yan, “Stability and bifurcation analysis on a three-species food chain system with two delays,” Communications in Nonlinear Science and Numerical Simulation, vol. 16, no. 9, pp. 3704–3720, 2011. • Z. Liu, S. Zhong, and X. Liu, “Permanence and periodic solu-tions for an impulsive reaction-diffusion food-chain system with ratio-dependent functional response,” Communications in Nonlinear Science and Numerical Simulation, vol. 19, no. 1, pp. 173–188, 2014. • F. Brauer and A. C. Soudack, “Coexistence properties of some predator-prey systems under constant rate harvesting and stocking,” Journal of Mathematical Biology, vol. 12, no. 1, pp. 101–114, 1981. • V. Lakshmikantham, D. D. Bainov, and P. S. Simeonov, Theory of Impulsive Differential Equations, World Scientific, 1989. • G. T. Stamov, Almost Periodic Solutions of Impulsive Differential Equations, Lecture Notes in Mathematics, Springer, Heidelberg, Germany, 2012. • D. D. Bainov and P. S. Simeonov, Impulsive Differential Equations: Periodic Solutions and Applications, Longman Scientific and Technical, 1993. • A. M. Samoilenko and N. A. Perestyuk, Impulsive Differential Equations, World Scientific, Singapore, 1995. • S. W. Zhang and D. J. Tan, “Permanence in a food chain system with impulsive perturbations,” Chaos, Solitons and Fractals, vol. 40, no. 1, pp. 392–400, 2009. • L. Bai and K. Wang, “Almost periodic solution of a non-autonomous food chains system with Holling's type II functional response,” Soochow Journal of Mathematics, vol. 28, no. 3, pp. 267–279, 2002. • Y. Xia, J. Cao, H. Zhang, and F. Chen, “Almost periodic solutions of $n$-species competitive system with feedback controls,” Journal of Mathematical Analysis and Applications, vol. 294, no. 2, pp. 503–522, 2004. • X. Z. Meng and L. S. Chen, “Almost periodic solution of non-autonomous Lotka-Volterra predator-prey dispersal system with delays,” Journal of Theoretical Biology, vol. 243, no. 4, pp. 562–574, 2006. • X. Lin and F. Chen, “Almost periodic solution for a Volterra model with mutual interference and Beddington-DeAngelis functional response,” Applied Mathematics and Computation, vol. 214, no. 2, pp. 548–556, 2009. • Y. Li and T. Zhang, “Permanence and almost periodic sequence solution for a discrete delay logistic equation with feedback control,” Nonlinear Analysis: Real World Applications, vol. 12, no. 3, pp. 1850–1864, 2011. • T. Zhang, Y. Li, and Y. Ye, “On the existence and stability of a unique almost periodic solution of Schoener's competition model with pure-delays and impulsive effects,” Communications in Nonlinear Science and Numerical Simulation, vol. 17, no. 3, pp. 1408–1422, 2012. • Z. Tian-Wei-Tian, “Multiplicity of positive almost periodic solutions in a delayed Hassell-Varley-type predator-prey model with harvesting on prey,” Mathematical Methods in the Applied Sciences, vol. 37, no. 5, pp. 686–697, 2014. • T. W. Zhang, Y. K. Li, and Y. Ye, “Persistence and almost periodic solutions for a discrete fishing model with feedback control,” Communications in Nonlinear Science and Numerical Simulation, vol. 16, no. 3, pp. 1564–1573, 2011. • T. Zhang and X. Gan, “Existence and permanence of almost periodic solutions for Leslie-Gower predator-prey model with variable delays,” Electronic Journal of Differential Equations, vol. 2013, no. 105, pp. 1–21, 2013. • T. W. Zhang and X. R. Gan, “Almost periodic solutions for a discrete fishing model with feedback control and time delays,” Communications in Nonlinear Science and Numerical Simulation, vol. 19, no. 1, pp. 150–163, 2014. • T. W. Zhang, “Almost periodic oscillations in a generalized Mackey-Glass model of respiratory dynamics with several delays,” International Journal of Biomathematics, vol. 7, no. 3, Article ID 1450029, 22 pages, 2014. \endinput
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.37165406346321106, "perplexity": 2048.808822870304}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-05/segments/1579250595282.35/warc/CC-MAIN-20200119205448-20200119233448-00448.warc.gz"}
http://www.computer.org/csdl/trans/tc/1998/07/t0750-abs.html
Subscribe Issue No.07 - July (1998 vol.47) pp: 750-756 ABSTRACT <p><b>Abstract</b>—A new technique for the a priori calculation of rigorous error bounds for floating-point computations is introduced. The theorems given in the paper combined with interval arithmetic lead to the implementation of reliable software routines, which enable the user to compute the desired error bounds automatically by a suitable computer program. As a prominent example, a table-lookup algorithm for calculating the function exp(<it>x</it>) <tmath>$-$</tmath> 1 that has been published by Tang [<ref rid="bibt075016" type="bib">16</ref>] is analyzed using these new tools. The result shows the high quality of the new approach.</p> INDEX TERMS Rigorous error bounds, table-lookup algorithms, elementary function algorithms, automation of error analysis. CITATION Walter Krämer, "A Priori Worst Case Error Bounds for Floating-Point Computations", IEEE Transactions on Computers, vol.47, no. 7, pp. 750-756, July 1998, doi:10.1109/12.709374
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6766301989555359, "perplexity": 1894.6204977538428}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-14/segments/1427131298538.29/warc/CC-MAIN-20150323172138-00270-ip-10-168-14-71.ec2.internal.warc.gz"}
http://blekko.com/wiki/Coherence_(physics)?source=672620ff
# Coherence (physics) In physics, coherence is an ideal property of waves that enables stationary (i.e. temporally and spatially constant) interference. It contains several distinct concepts, which are limit cases that never occur in reality but allow an understanding of the physics of waves, and has become a very important concept in quantum physics. More generally, coherence describes all properties of the correlation between physical quantities of a single wave, or between several waves or wave packets. Interference is nothing more than the addition, in the mathematical sense, of wave functions. In quantum mechanics, a single wave can interfere with itself, but this is due to its quantum behavior and is still an addition of two waves (see Young's slits experiment). This implies that constructive or destructive interferences are limit cases, and that waves can always interfere, even if the result of the addition is complicated or not remarkable. When interfering, two waves can add together to create a wave of greater amplitude than either one (constructive interference) or subtract from each other to create a wave of lesser amplitude than either one (destructive interference), depending on their relative phase. Two waves are said to be coherent if they have a constant relative phase. The degree of coherence is measured by the interference visibility, a measure of how perfectly the waves can cancel due to destructive interference. Spatial coherence describes the correlation between waves at different points in space. Temporal coherence describes the correlation or predictable relationship between waves observed at different moments in time. Both are observed in the Michelson–Morley experiment and Young's interference experiment. Once the fringes are obtained in the Michelson–Morley experiment, when one of the mirrors is moved away gradually, the time for the beam to travel increases and the fringes become dull and finally are lost, showing temporal coherence. Similarly, if in Young's double slit experiment the space between the two slits is increased, the coherence dies gradually and finally the fringes disappear, showing spatial coherence. ## Introduction Coherence was originally conceived in connection with Thomas Young's double-slit experiment in optics but is now used in any field that involves waves, such as acoustics, electrical engineering, neuroscience, and quantum mechanics. The property of coherence is the basis for commercial applications such as holography, the Sagnac gyroscope, radio antenna arrays, optical coherence tomography and telescope interferometers (astronomical optical interferometers and radio telescopes). ## Coherence and correlation The coherence of two waves follows from how well correlated the waves are as quantified by the cross-correlation function.[1][2][3][4][5] The cross-correlation quantifies the ability to predict the value of the second wave by knowing the value of the first. As an example, consider two waves perfectly correlated for all times. At any time, if the first wave changes, the second will change in the same way. If combined they can exhibit complete constructive interference/superposition at all times, then it follows that they are perfectly coherent. As will be discussed below, the second wave need not be a separate entity. It could be the first wave at a different time or position. In this case, the measure of correlation is the autocorrelation function (sometimes called self-coherence). Degree of correlation involves correlation functions. ## Examples of wave-like states These states are unified by the fact that their behavior is described by a wave equation or some generalization thereof. In most of these systems, one can measure the wave directly. Consequently, its correlation with another wave can simply be calculated. However, in optics one cannot measure the electric field directly as it oscillates much faster than any detector's time resolution.[6] Instead, we measure the intensity of the light. Most of the concepts involving coherence which will be introduced below were developed in the field of optics and then used in other fields. Therefore, many of the standard measurements of coherence are indirect measurements, even in fields where the wave can be measured directly. ## Temporal coherence Figure 1: The amplitude of a single frequency wave as a function of time t (red) and a copy of the same wave delayed by τ(green). The coherence time of the wave is infinite since it is perfectly correlated with itself for all delays τ. Figure 2: The amplitude of a wave whose phase drifts significantly in time τc as a function of time t (red) and a copy of the same wave delayed by 2τc(green). At any particular time t the wave can interfere perfectly with its delayed copy. But, since half the time the red and green waves are in phase and half the time out of phase, when averaged over t any interference disappears at this delay. Temporal coherence is the measure of the average correlation between the value of a wave and itself delayed by τ, at any pair of times. Temporal coherence tells us how monochromatic a source is. In other words, it characterizes how well a wave can interfere with itself at a different time. The delay over which the phase or amplitude wanders by a significant amount (and hence the correlation decreases by significant amount) is defined as the coherence time τc. At τ=0 the degree of coherence is perfect whereas it drops significantly by delay τc. The coherence length Lc is defined as the distance the wave travels in time τc. One should be careful not to confuse the coherence time with the time duration of the signal, nor the coherence length with the coherence area (see below). ### The relationship between coherence time and bandwidth It can be shown that the faster a wave decorrelates (and hence the smaller τc is) the larger the range of frequencies Δf the wave contains. Thus there is a tradeoff: $\tau_c \Delta f \approx 1$. Formally, this follows from the convolution theorem in mathematics, which relates the Fourier transform of the power spectrum (the intensity of each frequency) to its autocorrelation. ### Examples of temporal coherence We consider four examples of temporal coherence. • A wave containing only a single frequency (monochromatic) is perfectly correlated at all times according to the above relation. (See Figure 1) • Conversely, a wave whose phase drifts quickly will have a short coherence time. (See Figure 2) • Similarly, pulses (wave packets) of waves, which naturally have a broad range of frequencies, also have a short coherence time since the amplitude of the wave changes quickly. (See Figure 3) • Finally, white light, which has a very broad range of frequencies, is a wave which varies quickly in both amplitude and phase. Since it consequently has a very short coherence time (just 10 periods or so), it is often called incoherent. Monochromatic sources are usually lasers; such high monochromaticity implies long coherence lengths (up to hundreds of meters). For example, a stabilized and monomode helium–neon laser can easily produce light with coherence lengths of 300 m.[7] Not all lasers are monochromatic, however (e.g. for a mode-locked Ti-sapphire laser, Δλ ≈ 2 nm - 70 nm). LEDs are characterized by Δλ ≈ 50 nm, and tungsten filament lights exhibit Δλ ≈ 600 nm, so these sources have shorter coherence times than the most monochromatic lasers. Holography requires light with a long coherence time. In contrast, optical coherence tomography uses light with a short coherence time. ### Measurement of temporal coherence Figure 3: The amplitude of a wavepacket whose amplitude changes significantly in time τc (red) and a copy of the same wave delayed by 2τc(green) plotted as a function of time t. At any particular time the red and green waves are uncorrelated; one oscillates while the other is constant and so there will be no interference at this delay. Another way of looking at this is the wavepackets are not overlapped in time and so at any particular time there is only one nonzero field so no interference can occur. Figure 4: The time-averaged intensity (blue) detected at the output of an interferometer plotted as a function of delay τ for the example waves in Figures 2 and 3. As the delay is changed by half a period, the interference switches between constructive and destructive. The black lines indicate the interference envelope, which gives the degree of coherence. Although the waves in Figures 2 and 3 have different time durations, they have the same coherence time. In optics, temporal coherence is measured in an interferometer such as the Michelson interferometer or Mach–Zehnder interferometer. In these devices, a wave is combined with a copy of itself that is delayed by time τ. A detector measures the time-averaged intensity of the light exiting the interferometer. The resulting interference visibility (e.g. see Figure 4) gives the temporal coherence at delay τ. Since for most natural light sources, the coherence time is much shorter than the time resolution of any detector, the detector itself does the time averaging. Consider the example shown in Figure 3. At a fixed delay, here 2τc, an infinitely fast detector would measure an intensity that fluctuates significantly over a time t equal to τc. In this case, to find the temporal coherence at 2τc, one would manually time-average the intensity. ## Spatial coherence In some systems, such as water waves or optics, wave-like states can extend over one or two dimensions. Spatial coherence describes the ability for two points in space, x1 and x2, in the extent of a wave to interfere, when averaged over time. More precisely, the spatial coherence is the cross-correlation between two points in a wave for all times. If a wave has only 1 value of amplitude over an infinite length, it is perfectly spatially coherent. The range of separation between the two points over which there is significant interference is called the coherence area, Ac. This is the relevant type of coherence for the Young's double-slit interferometer. It is also used in optical imaging systems and particularly in various types of astronomy telescopes. Sometimes people also use "spatial coherence" to refer to the visibility when a wave-like state is combined with a spatially shifted copy of itself. ### Examples of spatial coherence Consider a tungsten light-bulb filament. Different points in the filament emit light independently and have no fixed phase-relationship. In detail, at any point in time the profile of the emitted light is going to be distorted. The profile will change randomly over the coherence time $\tau_c$. Since for a white-light source such as a light-bulb $\tau_c$ is small, the filament is considered a spatially incoherent source. In contrast, a radio antenna array, has large spatial coherence because antennas at opposite ends of the array emit with a fixed phase-relationship. Light waves produced by a laser often have high temporal and spatial coherence (though the degree of coherence depends strongly on the exact properties of the laser). Spatial coherence of laser beams also manifests itself as speckle patterns and diffraction fringes seen at the edges of shadow. Holography requires temporally and spatially coherent light. Its inventor, Dennis Gabor, produced successful holograms more than ten years before lasers were invented. To produce coherent light he passed the monochromatic light from an emission line of a mercury-vapor lamp through a pinhole spatial filter. In February 2011, Dr Andrew Truscott, leader of a research team at the ARC Centre of Excellence for Quantum-Atom Optics at Australian National University in Canberra, Australian Capital Territory, showed that helium atoms cooled to near absolute zero / Bose–Einstein condensate state, can be made to flow and behave as a coherent beam as occurs in a laser.[8][9] ## Spectral coherence Figure 10: Waves of different frequencies (i.e. colors) interfere to form a pulse if they are coherent. Figure 11: Spectrally incoherent light interferes to form continuous light with a randomly varying phase and amplitude Waves of different frequencies (in light these are different colours) can interfere to form a pulse if they have a fixed relative phase-relationship (see Fourier transform). Conversely, if waves of different frequencies are not coherent, then, when combined, they create a wave that is continuous in time (e.g. white light or white noise). The temporal duration of the pulse $\Delta t$ is limited by the spectral bandwidth of the light $\Delta f$ according to: $\Delta f\Delta t \ge 1$, which follows from the properties of the Fourier transform and results in Küpfmüller's uncertainty principle (for quantum particles it also results in the Heisenberg uncertainty principle). If the phase depends linearly on the frequency (i.e. $\theta (f) \propto f$) then the pulse will have the minimum time duration for its bandwidth (a transform-limited pulse), otherwise it is chirped (see dispersion). ### Measurement of spectral coherence Measurement of the spectral coherence of light requires a nonlinear optical interferometer, such as an intensity optical correlator, frequency-resolved optical gating (FROG), or spectral phase interferometry for direct electric-field reconstruction (SPIDER). ## Polarization coherence Light also has a polarization, which is the direction in which the electric field oscillates. Unpolarized light is composed of incoherent light waves with random polarization angles. The electric field of the unpolarized light wanders in every direction and changes in phase over the coherence time of the two light waves. An absorbing polarizer rotated to any angle will always transmit half the incident intensity when averaged over time. If the electric field wanders by a smaller amount the light will be partially polarized so that at some angle, the polarizer will transmit more than half the intensity. If a wave is combined with an orthogonally polarized copy of itself delayed by less than the coherence time, partially polarized light is created. The polarization of a light beam is represented by a vector in the Poincaré sphere. For polarized light the end of the vector lies on the surface of the sphere, whereas the vector has zero length for unpolarized light. The vector for partially polarized light lies within the sphere ## Applications ### Holography Coherent superpositions of optical wave fields include holography. Holographic objects are used frequently in daily life in bank notes and credit cards. ### Non-optical wave fields Further applications concern the coherent superposition of non-optical wave fields. In quantum mechanics for example one considers a probability field, which is related to the wave function $\psi (\mathbf r)$ (interpretation: density of the probability amplitude). Here the applications concern, among others, the future technologies of quantum computing and the already available technology of quantum cryptography. Additionally the problems of the following subchapter are treated. ## Quantum coherence In quantum mechanics, all objects have wave-like properties (see de Broglie waves). For instance, in Young's double-slit experiment electrons can be used in the place of light waves. Each electron's wave-function goes through both slits, and hence has two separate split-beams that contribute to the intensity pattern on a screen. According to standard wave theory (Fresnel, Huygens) these two contributions give rise to an intensity pattern of bright bands due to constructive interference, interlaced with dark bands due to destructive interference, on a downstream screen. (Each split-beam, by itself, generates a diffraction pattern with less noticeable, more widely spaced dark and light bands.) This ability to interfere and diffract is related to coherence (classical or quantum) of the wave. The association of an electron with a wave is unique to quantum theory. When the incident beam is represented by a quantum pure state, the split beams downstream of the two slits are represented as a superposition of the pure states representing each split beam. (This has nothing to do with two particles or Bell's inequalities relevant to an entangled state: a 2-body state, a kind of coherence between two 1-body states.) The quantum description of imperfectly coherent paths is called a mixed state. A perfectly coherent state has a density matrix (also called the "statistical operator") that is a projection onto the pure coherent state, while a mixed state is described by a classical probability distribution for the pure states that make up the mixture. Large-scale (macroscopic) quantum coherence leads to novel phenomena, the so-called macroscopic quantum phenomena. For instance, the laser, superconductivity and superfluidity are examples of highly coherent quantum systems whose effects are evident at the macroscopic scale. The macroscopic quantum coherence (Off-Diagonal Long-Range Order, ODLRO) [Penrose & Onsager (1957), C. N. Yang (1962)] for laser light, and superfluidity, is related to first-order (1-body) coherence/ODLRO, while superconductivity is related to second-order coherence/ODLRO. (For fermions, such as electrons, only even orders of coherence/ODLRO are possible.) Superfluidity in liquid He4 is related to a partial Bose–Einstein condensate. Here, the condensate portion is described by a multiply-occupied single-particle state. [e.g., Cummings & Johnston (1966)] Regarding the occurrence of quantum coherence at a macroscopic level, it is interesting to note that the classical electromagnetic field exhibits macroscopic quantum coherence. The most obvious example is the carrier signal for radio and TV. They satisfy Glauber's quantum description of coherence.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 8, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9156023859977722, "perplexity": 643.3527386270264}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-06/segments/1422122087108.30/warc/CC-MAIN-20150124175447-00008-ip-10-180-212-252.ec2.internal.warc.gz"}
https://byjus.com/mass-flow-rate-formula/
# Mass Flow Rate Formula ## Mass Flow Rate Formula The mass flow rate is the mass of a liquid substance passing per unit time. In other words, Mass flow rate is defined as the rate of movement of liquid mass through a unit area. The mass flow is directly depended on the density, velocity of the liquid and area of cross section. It is the movement of mass per unit time.The mass flow is denoted by m and the units in kg/s. The mass flow formula is given by, $m=\rho&space;VA$ Where, ρ = density of fluid, V = velocity of the liquid, and A = area of cross section Example 1 Determine the mass flow rate of a given fluid whose density is 700 kg/m3, velocity and area of cross section are 20m/s and 20cm2 respectively. Solution: Given values are, ρ = 700kg/m3, V = 20m/s, and A = 20cm2 = 0.20m2 The mass flow rate formula is given by, m = ρVA = 700 × 20 × 0.20 = 2800 kg/s Example 2 The density and velocity of a given fluid is 900kg/m3 and 5m/s. If this fluid flows through an area of 20cm2, determine the mass flow rate. Solution: Given values are, ρ = 900kg/m3, V = 5m/s, and A = 20cm2 = 0.20m2 The mass flow rate formula is given by, m = ρVA = 900 × 5 ×0.20 = 900 kg/s
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 1, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9851858615875244, "perplexity": 1572.7047383302308}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-35/segments/1566027318986.84/warc/CC-MAIN-20190823192831-20190823214831-00021.warc.gz"}
https://astarmathsandphysics.com/ib-physics-notes/thermal-physics/1462-evaporation.html
## Evaporation Evaporation is the process by which a liquid becomes a gas. Evaporation may take place at any temperature, but occurs most quickly when a liquid boils. Below the boiling point, the molecules in the surface of a liquid do not have enough kinetic energy on average to leave the liquid, but a fraction of molecules do and evaporation can still take place. Of course, it is typically the faster molecules that leave the liquid. The lower energy molecules remain as liquid, and higher energy molecules may leave escape the liquid but be pulled back, with even higher energy molecules able to escap altogether. If the system is isolated then the most energetic molecules remove from the liquid whatever energy they have. The molecules left behind have less energy on average, so the rate of evaporation will decrease. In practice no system is ever isolated. The liquid will be at the ambient temperature usually, and energy may be given to a liquid from it's surrounding to replace that energy taken away by the liquid that has evaporated. The net effect is that in the absence of the amount of liquid being replenished by other means, all the liquid will eventually evaporate. The rate at which evaporation takes place depends on: • The surface area of the liquid. Increased surface area means increased rate of evaporation. • The temperature og the liquid. Higher temperatures mean higher evaporation rates. • Air pressure. The evaporation rate increasing with decreasing atmospheric pressure. • The amount of vapour present in the air. The net rate of evaporation is a balance between vapour molecules rejoining the liquid and liquid molecules leaving. If there is a lot of vapour present, the rate at which molecules rejoin the liquid will increase and net evaporation will decrease. • A draught above the liquid surface will increase the evaporation rate by giving kinetic energy to molecules near the surface.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9256595373153687, "perplexity": 545.7540516678519}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-47/segments/1510934806070.53/warc/CC-MAIN-20171120145722-20171120165722-00019.warc.gz"}
https://www.intechopen.com/books/mental-disorders-theoretical-and-empirical-perspectives/parent-child-attachment-parental-depression-and-perception-of-child-behavioural-emotional-problems
Open access peer-reviewed chapter # Parent-Child Attachment, Parental Depression, and Perception of Child Behavioural/Emotional Problems By Lawrence T. Lam Submitted: April 27th 2012Reviewed: June 29th 2012Published: January 16th 2013 DOI: 10.5772/51170 ## 1. Introduction The issue of whether parents who have experienced symptoms of psychopathology, particularly depression, would be able to provide an accurate report on the behavioural/emotional problems of their children has long been raised. [1] It has been suggested that parental depression plays an important role in their perception of their children’s behavioural. [2-7] The early review by Breslau (1988) on the available studies in 1988 found that there was no evidence for any distortion of child behavioural problems using depressed mothers as informants. [1] However, more recent studies have found a positive relationship between parental depression, particularly maternal depression, and report of increased behavioural/emotional problems of their children. For example, in the study by Fergussen et al. on the effect of maternal depression on their ratings of children behaviour found a significant association between their depression and children’s conduct disorder and attention deficit behaviour. [3] A recent study by Hall et al. also found that depressive symptoms in mothers contributed significantly to the perception of both internalising and externalising problems of their children when they were asked to report on their children’s behaviour [7]. Parent-child relationships, particularly attachment or connectedness between the parent and child, have been reported to have an effect on the behavioural and emotional health of children and adolescents. [8-9] In terms of the parent-child relationship and parental perception of their children’s behavioural and emotional problems, it has been noted that few studies have been conducted. [5] The study by Kolko et al. found that low parental acceptance of the child was significantly associated with the difference between parents’ and teachers’ perception of children’s externalising but not internalising behaviour. [10] Another study by Mosley et al. also found that better parent-child relationships in terms of increased involvement also related to a decrease of parental perception on both internalising and externalising behavioural problems. [11] A more recent study by Treulter & Epkins also found that the parent-child relationship, measured as parental involvement and time spent with children, was significantly related to both parents’ rating of externalising behaviour. [5] In sum, these studies suggested an overall negative association between parent-child relationships and their perception on children’s behavioural/emotional problems such that the better the parent-child relationship, the less the degree of parental perception of children’s behavioural/emotional problems. The above-mentioned studies have suggested both parental depressive symptoms and parent-child relationships are associated with parent’s perception of children’s behavioural/emotional problems. However, the foci of all these studies are on individual variables, either parental depressive symptoms or parent-child relationships, and their associations with parental perception. All of the above-mentioned studies considered only one of the two factors with the exception of the study by Treulter & Epkins. [5] It has been known that parental depression, particularly maternal depression, exerts a significant influence on the parent-child relationship. [12] Therefore, it is logical to consider that there could be an interaction effect of parental depression and parent-child relationship on parent’s perception of children’s behavioural/emotional problems. A search of the current literature has revealed no related studies on this topic. The aim of this study is to further examine the relationships between parental depression, parent-child relationship and parental perception of children’s behavioural/emotional problems. Of particular interest of the study is the possible interaction effect between parental depression and parent-child relationship, specifically parent-child attachment or connectedness, on the perception of children’s behavioural/emotional problems. It is hypothesised that the parent-child relationship acts as an effect modifier in the relationship between parental depression and the perception of children’s behavioural/emotional problems in a manner that depressed parents who also have inadequate attachment or connectedness with their children would have a significantly increased rating on their children’s behavioural/emotional problems. ## 2. Methods This cross-sectional survey is part of the Guangzhou city primary school students’ behavioural problems intervention longitudinal cohort study. The survey serves as the base-line pre-intervention data collected on participating school children and their parents as well as a screening for behavioural problems among children. The cohort study is an on-going study conducted in the YueXiu district, the biggest district of the Guangzhou city in Guangdong Province, since December 2008. Guangdong Province is located in Southern China, and is the most populous province in China of which Guangzhou city is the capital. It is the biggest and most populated city of the Province with an estimated population of nearly 10 million in 2006. Institute ethics approval for the study was granted by the Human Ethics Committee of the Sun Yat-Sen University, GuangZhou, PR China. The sample was generated from the total student population of grade 4 primary school children within the study district. All students were registered with the Guangzhou city primary school registry. According the latest information from the registry, there were about 8400 grade 4 primary school children enrolled within the study district in 2008. A stratified random sampling method with stratification according to proportions of students in different type of schools: schools run by provincial government, local government, or private schools, were used for sample generation. The base-line survey was conducted at different schools within the same week. Parents of the selected students from different schools were invited to participate in the longitudinal study via school principals and their teachers, and were encouraged to fill in the baseline self-reported questionnaire designed specifically for the study. Informed consent was sought from each parent by signing a consent form indicating willing participation in the longitudinal study. The outcome of the study, namely parental perception of child behavioural problems, was assessed using the Child Behavioural Checklist for Ages 6-18 (CBCL/6-18), which was a validated, standardised, and a widely used scale for assessing behavioural problems amongst children and adolescents internationally. [13] Parents were asked to respond to 113 items that described specific behavioural and emotional problems that might occur in children and adolescents. These items were rated by parents on their perception of how true each item described their children now or within the past 6 months using a scale ranging from 0 to 2 (0=not true; 1=somewhat true or sometimes true; 2=very true or often true). The main content of the CBCL/6-18 covers 8 different domains including aggressive behaviour, anxious/depressed, attention problems, rule-breaking behaviour, social problems, somatic complaints, thought problems, and withdrawal. These eight domains are then further summarised into two main clusters of behavioural or emotive problems: internalisation (CBCL-inter) and externalisation (CBCL-exter) problems. A total score was also calculated for each child as an overall summary of the individual’s behavioural problems (CBCL-total). Examples of these items included: “Can’t concentrate, can’t pay attention for long” for attention problems; “Cruel to animals” for aggressive behaviour; and “Unhappy, sad, or depressed” for anxious/depressed emotions. Higher total scores, internalisation, and externalisation scores, indicate a greater severity of behavioural or emotive problems. Ample studies have been conducted on the validity and reliability of the CBCL/6-18 with full information available on the Achenbach System of Empirically Based Assessment (ASEBA) official website [14]. Parental depression was measured using the depression subscale of the Hospital Anxiety and Depression Scale (HADS) which was designed as a self-reported screening instrument for depression. [15] The instrument has been used in many studies and its validity is well proven. [16] The depression subscale consists of 7 items assessing depressive affects. Examples of these items include: “I look forward with enjoyment to things” and “I have lost interest in my appearance”. Respondents were asked to respond to questions of how often they had experienced these emotional states in the month prior to the survey. On a Likert scale, responses were rated from 0 to 3 for not at all to all of the time or some equivalent responses with total scores ranging from 0 to 21. Responses were graded in a negative direction in a manner that higher scores represent greater depressive affects. Attachment or connectedness between parent and child was assessed using the Attachment Problems subscale of the Parental Stress Index (PSI). [17] The PSI was designed specifically to evaluate and to diagnose individual parent-child dyads under stress. Of different domains included in the instrument, attachment between parent and child is one “designed to measure the intrinsic investment the parent in the role of parent”. [17] It was used to assess the degree of willingness parents would commit and attach in the relationship with their children. Parents were asked to rate how much they agree to a series of statements regarding their relationship with their children on a Likert scale ranging from 0 to 4. Responses were graded in a direction with higher scores indicating greater attachment problems. An example item is “I expected to have closer and warmer feelings for my child than I do and this bothers me”. The reliability and validity of the PSI have been demonstrated in many studies [17-19]. Other information collected in the survey included age and sex of the child, respondents’ age and sex, parental education levels, occupation, family structure whether intact or divorced, family monthly incomes, and whether the parent suffered any chronic illnesses or hospitalisation. Included in the study were only parents of target children. Other responding relatives, such as grandparents, were excluded. Data were analysed using the Stata V10.0 statistical software program. [20] Descriptive statistics on the sample including frequencies, percentages, means, standard deviations, and median were presented according to the nature of variables. All CBCL scores were treated as continuous variables. Bivariate analyses were conducted to examine unadjusted relationships between parental depression, attachment problems, demographics, and parental perception of child behavioural problems including the total, internalisation, and externalisation scores. Owing to the fact that CBCL-total, CBCL-inter, and CBCL-exter scores were highly skewed and exhibited problems of over dispersion, it was considered more appropriate to analyse these data with negative binomial regression models. Hence, this analytical technique was applied to all unadjusted and adjusted analyses. Bivariate associations between all variables of interest and parental perception of child behavioural problems were analysed with simple negative binominal regression models and were tested using the Likelihood Ratio Chi-squared tests. All significant variables identified from the bivariate analyses were included in further analyses for the adjusted relationship between parental depression, attachment problems and their perception of their children’s behavioural problems. Of particular interest to the study was the interaction effect of parental depression and attachment problems between parent and child, on parental perception of children behavioural problems. Hence, the significance of interaction terms was tested in the final models for CBCL-total, CBCL-inter, and CBCL-exter scores. ## 3. Results A total of 1839 parent-child dyads were recruited and provided usable information. This represented 97.1% of the total participants of the longitudinal study. Fifty-six respondents were not parents of the targeted child. The characteristics of children and their parents including parental depression, attachment problems, and parental perception of behavioural problems were summarised in Table 1. In terms of the outcome variables, the mean values of all CBCL scores were larger than their corresponding median values with large standard deviations suggesting all three scores were skewed and had problems of over dispersion. Variables Frequency (%) or mean (s.d.), median Demographics Age of the child (years) Mean=9.5 (0.61), median=9.5 Sex of childMaleFemale 958 (52.2)876 (47.8) Age of parent (years) Mean=38.7 (4.11), median=38.1 Relation of parent to childFatherMother 767 (42.7)1031 (57.3) Family structureLiving with both biological parentsOthers 1624 (89.8)184 (10.2) Parent’s education levelHigh school or lowerPost secondaryUniversity or higher 920 (51.1)418 (23.2)462 (25.7) Parent’s occupationUnemployed /retiredFarmer/labourerGovernment officialProfessional/technicalBusiness 232 (13.1)465 (26.4)425 (24.1)201 (11.4) Monthly income (RMB$)*LowMiddleHigh 601 (35.2)522 (30.5)586 (34.3) Parent having chronic disease or hospitalisedYesNo 292 (15.9)1547 (84.1) Parental Stress Index -Attachment Problems scores Mean=18.4 (3.35), median=18.0 Parental depression HADS depression scores Mean=4.7 (3.04), median=4.0 CBCL-Total scores Mean=21.1 (17.66), median=18.0 CBCL Internalisation scores Mean=5.8 (5.79), median=4.0 CBCL Externalisation scores Mean=7.9 (7.06), median 6.0 ### Table 1. Description information on the characteristics of children, their parents, parent-child attachment problems, parental depression, and parental perception of child behavioural problems (N=1839) *Low <RMB$3000; middle 3000-4900; high>5000 Results on association Variables CBCL total Internalisation Externalisation Demographics Age of the child χ2 1=2.52, p=0.112 χ2 1=0.49, p=0.485 χ2 1=3.81, p=0.051 Sex of child χ2 1=23.77, p<0.001 χ2 1=4.50, p=0.034 χ2 1=50.18, p<0.001 Age of parent χ2 1=16.01, p<0.001 χ2 1=8.09, p=0.0042 χ2 1=23.69, p<0.001 Relation of parent to child χ2 1=1.13, p=0.288 χ2 1=0.62, p=0.432 χ2 1=1.14, p=0.216 Family structure χ2 1=2.16, p=0.141 χ2 1=2.50 p=0.114 χ2 1=1.85, p=0.174 Parent’s education level χ2 2=5.90, p=0.052 χ2 2=4.32, p=0.115 χ2 2=3.95, p=0.139 Parent’s occupation χ2 4=5.34, p=0.254 χ2 4=7.21, p=0.125 χ2 4=3.70, p=0.449 Monthly income χ2 2=5.32, p=0.007 χ2 2=2.67, p=0.263 χ2 2=7.55, p=0.023 Parent having chronic disease or hospitalised χ2 1=22.65, p<0.001 χ2 1=27.31, p<0.001 χ2 1=25.36, p<0.001 Parental Stress Index -Attachment Problems scores χ2 1=12.13, p<0.001 χ2 1=19.22, p<0.001 χ2 1=12.38, p<0.001 Parental depression HADS depression scores χ2 1=42.52, p<0.001 χ2 1=50.62, p<0.001 χ2 1=28.83, p<0.001 ### Table 2. Unadjusted associations between child and parent characteristics and parental perception on child behavioural problems: CBCL Total; CBCL-Internalisation: and CBCL-Externalisation The bivariate relationships between demographics, parental depression, attachment problems and parental perception of child behavioural problems were examined. The results were summarised in Table 2. As shown, among the demographic variables, child’s sex, parents’ age, monthly income, and parental illness were significantly associated with all three CBCL scores. Results also suggested the unadjusted association between parental depression and the perception of their children’s behavioural problems was highly significant for all CBCL scores (CBCL-total: χ21=42.52, p<0.001; CBCL-inter: χ21=50.62, p<0.001; CBCL-exter: χ21=28.83, p<0.001). This also applied to attachment problems (CBCL-total: χ21=12.13, p<0.001; CBCL-inter: χ21=19.22, p<0.001; CBCL-exter: χ21=12.38, p<0.001). Results Variables in the final model β SE(β) Z-value p-value CBCL-Total Female -0.21 0.047 -4.45 <0.001 Parent’s age -0.02 0.006 -3.66 <0.001 Parent having chronic disease or hospitalised 0.25 0.062 3.94 <0.001 Attachment problems 0.03 0.007 3.32 0.001 Parental depression 0.05 0.008 5.71 <0.001 CBCL-Internalisation Parent’s age -0.02 0.007 -2.49 0.013 Parent having chronic disease or hospitalised 0.31 0.070 4.50 <0.001 Attachment problems 0.03 0.008 3.97 <0.001 Parental depression 0.06 0.001 5.96 <0.001 CBCL-Externalisation Female -0.31 0.047 -6.605 <0.001 Parent’s age -0.03 0.006 -4.72 <0.001 Parent having chronic disease or hospitalised 0.24 0.063 3.79 <0.001 Attachment problems 0.03 0.008 3.37 0.001 Parental depression 0.04 0.008 4.25 <0.001 ### Table 3. Results obtained from the multiple negative binomial regression analyses on CBCL Total, CBCL-Internalisation, and CBCL-Externalisation without interaction terms. The results obtained from the multiple negative binomial regression analyses were also presented in Table 3 and Table 4. Table 3 presented results on final regression models of CBCL scores, including the total, internalisation, and externalisation, regressing on various study variables with terms of interaction between parental depression and attachment problems included. Results indicated the interaction term in the final model for CBCL-exter was significant (Z=2.19, p=0.029), suggesting that there was a significant interaction effect of parental depression and attachment problems between parent and child on the perception of their children’s behavioural problems. Results Variables in the final model with interaction term β SE(β) Z-value p-value CBCL-Total Female -0.21 0.047 -4.45 <0.001 Parent’s age -0.02 0.006 -3.61 <0.001 Parent having chronic disease or hospitalised 0.25 0.062 3.99 <0.001 Attachment problems 0.04 0.012 3.18 0.001 Parental depression 0.11 0.046 2.41 0.016 Interaction: Attachment*Parental depression 0.01 0.002 1.42 0.155 CBCL-Internalisation Parent’s age -0.02 0.007 -2.45 0.014 Parent having chronic disease or hospitalised 0.32 0.070 4.51 <0.001 Attachment problems 0.04 0.014 2.84 0.004 Parental depression 0.08 0.053 1.54 0.124 Interaction: Attachment*Parental depression 0.01 0.003 0.51 0.611 CBCL-Externalisation Female -0.34 0.048 -7.01 <0.001 Parent’s age -0.03 0.006 -4.09 <0.001 Parent having chronic disease or hospitalised 0.25 0.063 4.00 <0.001 Attachment problems 0.05 0.013 4.02 <0.001 Parental depression 0.13 0.046 2.90 0.004 Interaction: Attachment*Parental depression 0.01 0.002 2.19 0.029 ### Table 4. Results obtained from the multiple negative binomial regression analyses on CBCL Total, CBCL-Internalisation, and CBCL-Externalisation with interaction terms. Figure 1 depicted the unadjusted mean scores of externalisation behaviour for different parental depression status by different levels of attachment problems. The figure exhibited a lack of parallelism suggesting an interaction effect between parental depression and parent-child attachment on the CBCL-exter scores. The interaction terms in models for CBCL-total and CBCL-inter were not significant. Table 4 presented results on the final models of CBCL-total and CBCL-inter after removal of the interaction terms. As shown, parental depression and attachment problems between parent and child were significantly related to parental perception of their children’s behavioural problems for both the total (depression: Z=5.71, p<0.001: attachment: Z=3.32, p=0.001) and internalisation (depression: Z=5.96, p<0.001: attachment: Z=3.97, p<0.001) scores after adjusting for sex, parental age, and parental illness. ## 4. Discussion and conclusion This study aims to examine the relationship between parental depression, parent-child relationship and parental perception of children’s behavioural/emotional problems. Particularly, it aims to test the hypothesis of the interaction effect of parental depression and parent-child attachment on parental perception of children’s behavioural/emotional problems. The results provided evidence of a significant interaction effect between parental depression and parent-child attachment problems on parental perception on the externalising of children’s problems, but no effects on internalising problems and total problems. The main effects of parental depression and parent-child attachment problems on parental perception of children behavioural/emotional problems were found significant. Due to the lack of a similar study on the interaction between parental depression and attachment problem and parental perception on children’s behavioural/emotional problems, a comparison of results obtained from this study with others reported in literature would be difficult. However, other results obtained from the study are consistent with those obtained in the general literature of parental depression, parent-child relationships and the distortion of parental perception of children’s behavioural/emotional problems as highlighted in the introduction section. For example, in the study by Treutler et al, it was found that parental depression and parent-child relationship were both associated with their ratings on internalising behaviours. [5] The results of the current study have rendered further support to this finding. Furthermore, it has shown that parental perception of children’s externalising behavioural problems is a function of their own depressive symptomatologies and the attachment to their children. The results indicated that attachment between parent and child acts as an effect modifier in the relationship between parental depression and their perception of children’s externalising behavioural problems. For parents who have a close relationship with their children, it could be considered natural that depression would influence their perception of children’s behavioural problems. However, for parents who have experienced attachment problems with their children as well as depression, the alienated relationship with their children tends to enhance their perception of children’s behavioural problems to such a degree that is much more than the usual effect brought on by depression alone. In other words, the alienated relationship enhances the effect of depression on parental perception of children’s behavioural problems. It is interesting to note that such an interactive effect is observed for externalising behavioural problems, but not for internalising problems. One possible explanation for this phenomenon is that externalising behavioural problems, including aggressive behaviour, rule-breaking, and attention problems, are observable manifestations of childhood problems. Parents can identify these behaviours and perceive them as problematic easily for both depressive and non-depressive parents, with depressive parents labelling them with a greater degree of severity. However, if the parent-child dyad has been experiencing a detrimental relationship, it is likely that the parent would have a heightened sensitivity to the child’s observable behavioural problems. This heightening of awareness or sensitivity is further enhanced in a parent who is also experiencing depression. This heightening of awareness may be absent for those less observable behavioural problems such as withdrawal and depression. The results obtained from this study have a direct clinical implication on the validity of using parents as informants for children’s behavioural assessments. As above-mentioned, parental distortion in their report on children’s behavioural assessments has long been identified and established as a potential source of error. [7] The results of this study further indicate that parents’ perception of their children’s behaviour is a function of their own internal state as well as their relationship with their children. The potential for distortion in reporting of child behaviour is great if a single parent is used as the sole informant. Should this aspect be overlooked, systematic biases would result due to distortion errors. This may occur in research as well as in clinical assessment. In fact, it has been advocated that multiple informants, especially non-familial informants, should be used for any childhood behavioural problems in both research and clinical settings. [21-22] If possible, observational tools instead of self-reported instruments should be employed and non-family observers should be used as informants. [23] In the current longitudinal study, children who have been identified with scores higher than the recommended cut-off will be assessed by a child psychiatrist using another clinical assessment tool. In terms of prevention and early intervention of developmental psychopathology in children, the results of the study can also shed some lights. Tackling attachment problems between parents and their children may provide an inroad to, not only the change of perception of child behavioural problems but also, the actual developmental problems during early childhood and behavioural problems throughout the entire child and adolescence period. A recent study has identified that the quality of interaction between mother and child, particularly in the area of emotion availability of the mother, has a direct bearing on the functioning, behavioural problems, and depressive symptomatologies of the child. [24] In this longitudinal study, it has been found that dysfunctional mother-child relationship and poor attachment at infancy is predictive of developmental and behavioural problems at later years of childhood. [24] However, early intervention program in enhancing attachment between parent and child could have a positive effect in improving the quality of parent-child relationship and reinforcing a secure attachment. [25] This could, in turn, have a direct impact on the on-going development of the child. Furthermore, it could also be possible that the problematic parent-child relationship may have contributed to the depression of the parent. An improvement in parent-child attachment could have an alleviating effect of the parent’s depression and, in turn, reinforcing a better relationship with the child and resulting in a better behavioural outcome. As in all studies, there are strengths and weaknesses in this study. This is a population-based study that includes a random sample of students. The use of a standardised and validated assessment instrument for the outcome measure minimised some measurement biases. The assessments on parental depression, parent-child relationship, and parental perception of children’s behavioural problems were concurrent and captured the current status on all these variables, hence minimising biases due to the lapse of time. A potential limitation has been identified in this study. Information on the exposure variables, namely parental depression and attachment problems is obtained via a self-reported questionnaire. Hence this will constitute a report bias in the exposure variables and it would likely present as a differential bias due to depressive parents reporting more severe symptomatologies as well as parent-child relationship problems. To improve the quality of the study, parental depression and parent-child relationships are best to be assessed by multiple informants as well as by observers as suggested by [23]. ## More © 2013 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms of the Creative Commons Attribution 3.0 License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. ## How to cite and reference ### Cite this chapter Copy to clipboard Lawrence T. Lam (January 16th 2013). Parent-Child Attachment, Parental Depression, and Perception of Child Behavioural/Emotional Problems, Mental Disorders - Theoretical and Empirical Perspectives, Robert Woolfolk and Lesley Allen, IntechOpen, DOI: 10.5772/51170. Available from: ### Related Content Next chapter #### Current Advances in the Treatment of Major Depression: Shift Towards Receptor Specific Drugs By Ashok Kumar Jainer, Rajkumar Kamatchi, Marek Marzanski and Bettahalasoor Somashekar First chapter #### What is the Link Between Protein Aggregation and Interneuronal Lesion Propagation in Neurodegenerative Disease? By Garth F. Hall We are IntechOpen, the world's leading publisher of Open Access books. Built by scientists, for scientists. Our readership spans scientists, professors, researchers, librarians, and students, as well as business professionals. We share our knowledge and peer-reveiwed research papers with libraries, scientific and engineering societies, and also work with corporate R&D departments and government entities. View all Books
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3447893559932709, "perplexity": 6876.324287743493}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-51/segments/1575540542644.69/warc/CC-MAIN-20191212074623-20191212102623-00359.warc.gz"}
https://www.quantinsti.com/blog/r-weekly-bulletin-vol-ii
# R Weekly Bulletin Vol – II This week’s R bulletin will cover functions calls, sorting data frame, creating time series object, and functions like is.na, na.omit, paste, help, rep, and seq function. Hope you like this R weekly bulletin. Enjoy reading! ### Shortcut Keys 1. To show files – Ctrl+5 2. To show plots – Ctrl+6 3. To show packages – Ctrl+7 ### Problem Solving Ideas #### Calling a function in an R script If you want to call a custom-built function in your R script from another script, one can use the “exists” function along with the “source” function. See the example below: Example: if(exists("daily_price_data", mode="function")) source("Stock price data.R") In this case, the expression will check whether a function called “daily_price_data” exists in the “Stock price data.R” script, and if it does, it will load the function in the current script. We can then use the function any number of times in our script by providing the relevant arguments. #### Convert dates from Google finance to a time series object When we download stock price data from Google finance, the “DATE” column shows a date in the yyymmdd format. This format is not considered as a time series object in R. To convert the dates from Google Finance into a time series object, one can use the ymd function from the lubridate package. The ymd function accepts dates in the form year, month, day. In the case of dates in other formats, the lubridate package has functions like ydm, mdy, myd, dmy, and dym, which can be used to convert it into a time series object. Example: library(lubridate) dt = ymd(20160523) print(dt) [1] “2016-05-23” #### Sorting a data frame in an ascending or descending order The arrange function from the dplyr package can be used to sort a data frame. The first argument is the data.frame and the next argument is the variable to sort by, either in an ascending or in a descending order. In the example below, we create a two column data frame comprising of stock symbols and their respective percentage price change. We then sort the Percent change column first in an ascending order, and in the second instance in a descending order. Example: library(dplyr) # Create a dataframe Ticker = c("UNITECH", "RCOM", "VEDL", "CANBK") Percent_Change = c(2.3, -0.25, 0.5, 1.24) df = data.frame(Ticker, Percent_Change) print(df) Ticker          Percent_Change 1  UNITECH    2.30 2      RCOM   -0.25 3        VEDL    0.50 4     CANBK    1.24 # Sort in an ascending order df_descending = arrange(df, Percent_Change) print(df_descending) Ticker     Percent_Change 1     RCOM    -0.25 2       VEDL    0.50 3    CANBK    1.24 4 UNITECH    2.30 # Sort in a descending order df_descending = arrange(df, desc(Percent_Change)) print(df_descending) Ticker         Percent_Change 1 UNITECH   2.30 2    CANBK   1.24 3       VEDL   0.50 4     RCOM   -0.25 ### Functions Demystified #### paste function The paste is a very useful function in R and is used to concatenate (join) the arguments supplied to it. To include or remove the space between the arguments use the “sep” argument. Example 1: Combining a string of words and a function using paste x = c(20:45) paste("Mean of x is", mean(x), sep = " ") [1] “Mean of x is 32.5” Example 2: Creating a filename using the dirPath, symbol, and the file extension name as the arguments to the paste function. dirPath = "C:/Users/MyFolder/" symbol = "INFY" filename = paste(dirPath, symbol, ".csv", sep = "") print(filename) [1] “C:/Users/MyFolder/INFY.csv” #### is.na and na.omit function The is.na functions checks whether there are any NA values in the given data set, whereas, the na.omit function will remove all the NA values from the given data set. Example: Consider a data frame comprising of open and close prices for a stock corresponding to each date. date = c(20160501, 20160502, 20160503, 20160504) open = c(234, NA, 236.85, 237.45) close = c(236, 237, NA, 238) df = data.frame(date, open, close) print(df) date           open        close 1  20160501  234.00     236 2  20160502        NA     237 3  20160503  236.85     NA 4  20160504  237.45     238 Let us check whether the data frame has any NA values using the is.na function. is.na(df) date      open      close [1,]  FALSE  FALSE   FALSE [2,]  FALSE  TRUE     FALSE [3,]  FALSE  FALSE   TRUE [4,]  FALSE  FALSE   FALSE As you can see from the result, it has two NA values. Let us now use the na.omit function, and view the results. na.omit(df) date           open         close 1  20160501  234.00      236 4  20160504  237.45      238 As can be seen from the result, the rows having NA values got omitted, and the resultant data frame now comprises of non-NA values only. These functions can be used to check for any NA values in large data sets on which we wish to apply some computations. The presence of NA values can cause the computations to give unwanted results, and hence such NA values need to be either removed or replaced by relevant values. #### rep and seq function The rep function repeats the arguments for the specified number of times, while the sequence function is used to form a required sequence of numbers. Note that in the sequence function we use a comma and not a colon. Example 1: rep("Strategy", times = 3) [1] “Strategy” “Strategy” “Strategy” rep(1:3, 2) [1] 1 2 3 1 2 3 Example 2: seq(1, 5) [1] 1 2 3 4 5 seq(1, 5, 0.5) [1] 1.0 1.5 2.0 2.5 3.0 3.5 4.0 4.5 5.0 #### help and example function The help function provides information on the topic sort, while the example function provides examples on the given topic. help(sum) example(sum) To access the R help files associated with specific functions within a particular package, include the function name as the first argument to the help function along with the package name mentioned in the second argument. Example: help(barplot, package="graphics") Alternatively, one can also type a question mark followed by the function name (e.g. ?barplot) and execute the command to know more about the function. ### Next Step We hope you liked this bulletin. In the next weekly bulletin, we will list more interesting ways and methods plus R functions for our readers. Update We have noticed that some users are facing challenges while downloading the market data from Yahoo and Google Finance platforms. In case you are looking for an alternative source for market data, you can use Quandl for the same. ### Learn Algorithmic trading from Experienced Market Practitioners • This field is for validation purposes and should be left unchanged.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.22539809346199036, "perplexity": 8044.9473868569685}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-22/segments/1558232256040.41/warc/CC-MAIN-20190520142005-20190520164005-00359.warc.gz"}
http://www.gradesaver.com/textbooks/math/precalculus/precalculus-mathematics-for-calculus-7th-edition/chapter-1-section-1-3-algebraic-expressions-1-3-exercises-page-34/77
## Precalculus: Mathematics for Calculus, 7th Edition $A^2$ - $B^2$ = (A + B)(A - B) 9$a^2$ - 16 = (3a + 4)(3a - 4)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5094496607780457, "perplexity": 1343.042960275943}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-43/segments/1508187826840.85/warc/CC-MAIN-20171023221059-20171024001059-00745.warc.gz"}
http://qnqk.de-piscines.fr/postgres-relation-does-not-exist-sequence.html
This Oracle tutorial explains how to use the Oracle EXISTS condition with syntax and examples. Do not assign code B95. Not only is this change true of the in-narrative post-apocalyptic wasteland that sees mankind making its final stand against machines, but it’s also true of the Terminator franchise, which. Most executable programs also support a --version option; at least postgres --version and psql --version should work. First, check the link_tmp table data:. It includes a comprehensive collection of TAP-emitting assertion functions, as well as the ability to integrate with other TAP-emitting test frameworks. 18, but sees errors:. A schema can also contain views, indexes, sequences, data types, operators, and functions. If you are an NBAM (Next-Best-Action-Marketing) customer, you can upload customer interaction history and other related information that changes daily. Some features may not be supported in earlier. object directory does not exist - Git operations fail after upgrade or STASH_HOME update. Even if your bug is not specific to Postgres Pro, do not send bug reports to any of the user mailing lists, such as or. All species can be organized into a sequence of increasing complexity. Any parameters not specifically set in the ALTER SEQUENCE command retain their prior settings. PgBadger is a tool that analyzes PostgreSQL log files and generates reports on execution of SQL and server operation. A disadvantage is that all tables share the same sequence. The specified database is set as the user's default database in SQL Server. Ranch Hand Posts: 424. Management Reporter with Data Mart will read from AA tables if AA has been activated. 21) by Greg Sabino Mullane, and contains information for 405 versions of Postgres. so : undefined symbol: SetUserId. On a new ubuntu 12. - Postgres has a "text" datatype which does not require a size specification. A relation could be dropped by a concurrent session, while one was doing a pg_relation_size on it, leading to a SQL exception. Defining your own range types¶. It seems like the best solution to fix the core code so that PostgreSQL can emulate UPDATE IGNORE by the use of savepoints, which I have now done by copying the method from the existing emulation used for INSERT. The SQL standard also distinguishes between global and local temporary tables, where a local temporary table has a separate set of contents for each SQL module within each session, though its definition is still shared across sessions. postgresで新規にテーブル作ってselectした時のこと。 普通に select * from テーブル名; でやると表示されるのですが、 select 列名 from テーブル名; とすると 「column "列名" does not exist」 というエラーが出てしまう。 なんで?(・ω・)?. You may have to register before you can post: click the register link above to proceed. Not surprisingly, you can get away with simply setting ECHO_HIDDEN instead of setting its value to 'on'. In both the command line and WSGI database user interfaces that I presented previously, there is an overlap (or redundancy) between the application and the database. As the same (unix) user, I execute psql -h localhost -U non-privileged-user table_name, and \dt reports NO relations found. ahmed yehia. Thus, we jump to the double-pipe symbol and "it is not a cow" prints to standard output. It is not the most efficient thing to work with in SQL! Since there is no magical built-in way to get the database to assign the next unused value in the sequence, we keep a table with all $$36^3$$ legal values, and do an exclusion join against the list of legal values. So I proceed by following steps, Before that I want to share my Rails application stack which is Rails 3. "hibernate_sequence" does not exist. Both are PostGIS enabled. fromloc to inventory. inventory table joining t. Dark energy does not exist and there is much more dark matter than we are aware of to date. You wish to introduce a new element (or maybe two elements), infinity, which you wish to append to the real number system. adrelid) should be used instead of adsrc when finding. 4 and PostgreSQL 11. Using a negative operator will cause the search to find every single row to identify that they are ALL not belong or exist within the table. If this entity - which is not a substance - was continually manifested everywhere, why and how it would move and what would correspond our nights? For us, on Earth, the space is black, out of the light sent by our Sun. CVE-2018-1058 : A flaw was found in the way Postgresql allowed a user to modify the behavior of a query for other users. It is much faster to search for an exact match (positive operator) such as using the LIKE, IN, EXIST or = symbol operator instead of a negative operator such as NOT LIKE, NOT IN, NOT EXIST or != symbol. It is not just that Dapper is at least twice as fast as EF when it comes to insert performance. What is the cause of the problem? A. It does not work with many-to-many relationships. 2 allows remote POP3 servers to cause a denial of service and possibly execute ar. Do not spend all your time to figure out which changes in the input make the problem go away. The only host entries in pg_hba. A disadvantage is that all tables share the same sequence. If specified, the table is created as a temporary table. Other versions of make use a simpler algorithm: if the file does not exist, and it is found via directory search, then that pathname is always used whether or not the target needs to be built. This child has not yet mastered the concept of object permanence. Looking back over the sequence of graphs (and perhaps generating some others) the graph of h(x) always has a line of symmetry parallel to the y-axis. 1 also with postgis 1. Need help? Post your question and get tips & solutions from a community of 445,869 IT Pros & Developers. clients RENAME TO customers; Messages. NET data provider. If it did, it would have to converge to the pointwise limit 0, but nf (1 2n) = n,. Textbook solution for Calculus (MindTap Course List) 8th Edition James Stewart Chapter 2. A common problem when copying or recreating a database is that database sequences are not updated just by inserting records in the table that sequence is used in. TEMPORARY or TEMP. 4 and PostgreSQL 11. And this is achieved without loss of efficiency. Protein assembly continues until the ribosome encounters a “stop” codon (a sequence of three bases that does not code for an amino acid). Tags: cygwin, heroku, postgres, rails 3 trackback. But I want it to rename a table with the same name if it finds an existing one. This shows for example that in Examples 2 and 3 above, lim x→0 f(x) does not exist. n = 1 1 2 + 1 3. This physiological entity brings closer together some specific phenomena participating in the activities of living beings, aware or not of the presence of this spirit. so : undefined symbol: SetUserId. Unlike InnoDB tables, MySQL does not create subdirectories that correspond to the database name when creating a MyISAM table with a DATA DIRECTORY or INDEX DIRECTORY option. 1057 The account name is invalid or does not exist, or the password is invalid for the account name specified. Facts are different than beliefs, rumors, and opinions. For example a schema will have classid set to pg_namespace. One of the reasons we love it so much is how easily you can dig into Postgres' implementation when there are issues. Numeric promotions Integral promotion. You do not want the PostgreSQL Server to start automatically after the restore operation is complete. It is enabled by adding the bracket pairs to be completed (e. In SQL this is extremely fast and returns me a bit that tells me very simply if the record exists or not. Consider the sequence ( 1)n+1. If there is no sequence owned by a column, all DDLs for a sequence behind an identity column do not raise errors but do nothing:. ERROR: relation "table1_id_seq" does not exist ERROR: relation "table2_id_seq" does not exist I renamed a couple tables, and the names of their corresponding sequences remained the same. Remove the spclocation field from pg_tablespace. PostgreSQL JOINS are used to retrieve data from multiple tables. The sequence generator generates sequential numbers, which can help to generate unique primary keys automatically, and to coordinate keys across multiple rows or tables. # run contents of "my_file" as a program perl my_file # run debugger "stand-alone". 0] Missing hibernate sequence in fresh installation on postgreSQL Jun 30, It seems that there could not exist 2 postgresql sequences with the same name in the same database. It is important to understand which part is to migrate. If the database does not exist, then it will be created and finally a database object will be returned. A connection can be established between two machines only if a connection between the two sockets does not exist, both machines agree to the connection, and both machines have adequate TCP resources to service the connection. Be careful to not misuse this theorem. Postgres in Cygwin for Rails 3 November 14, 2011 Posted by ficial in Blogroll, cygwin, ruby on rails, techy. library=# \dt users No matching relations found. Operator does not exist: timestamp without timezone and double precision Oracle supports addition operation between timestamp without time zone and double precision datatypes in SQL operations or in VIEW/MVIEW creation. Here is how you can do that, using the method of finite differences -- without having to determine the formula for the quadratic sequence. PostgreSQL allows the definition of custom range types. Is it a better way of dealing with it?. MH is a directory-based mailbox format invented for the MH Message Handling System, a mail user agent. This is the same as the definition of a sequential limit in the preceding section obtained by regarding the subset Dm ( f). In simple terms, the rules state that unique keys and events in the rows should say something about it, facts that do not relate to the key belongs into different tables, and the tables should not imply relationships that do not exist. Now, it merely returns NULL for this record. The identity_options clause has the same parameters as the CREATE SEQUENCE statement. Let's now take a look at the CSV file we wish to load into the database (note that the CSV does not contain real users but are randomly generated users using a Python library called faker). The only real difference between one-sided limits and normal limits is the range of $$x$$’s that we look at when determining the value of the limit. 推荐:PostgreSQL常用命令 昨天安装好PostgreSQL,今天开始学习使用。安装好以后,会附带有说明文档。 我打开psql,登陆如下界面: 按照说明文档中使用createdb命令创建数据库,但是提示语. I may, or may not care for the EF unit-of-work pattern. It does not work with child models in a multi-table inheritance scenario. > この"relation does not exist"とは一体どんなエラーなんでしょうか? 該当するテーブル、この場合は"tblbooks"が無いということです。 該当のSQLにタイプミスなどはありませんでしょうか? また、psql を使ってテーブル一覧を表示させたときにtblbooksは有りますか?. Any thoughts? This is a similar iss. 5 introduced INSERT ON CONFLICT [DO UPDATE] [DO NOTHING]. 1357 An attempt was made to exceed the limit on the number of domains per server. Do not assign code B95. We have used SELECT 1 in the subquery to increase performance since the column result set is not relevant to the EXISTS condition (only the existence of a returned row. PhysicalPath value is a Windows-style path. A normalized database is one such as the relationship among the tables matches the relationship that is really there among the data. PostgreSQL Database Forums on Bytes. So the check would be: if this country does not have postal codes, the postal code field should be null. I noticed almost all the resulting errors were due to references to these partitiioned tables, such as "relation "table_name" does not exist". As a user with admin privileges in postgres, I can psql database_name and then invoke \dt to see a list of relations. PostgreSQL NOT IN operator. However, if an infimum or supremum does exist, it is unique. Consequently, programs must not perform blocking operations while holding a lock. Restore postgres Dump gives Errors: types do not exist. 0 Mb) - this makes nucleotide numbering based on a genomic reference sequence rather impractical (e. I may have given up on LINQ as an interface to my DB, for example. Cualquiera sea la razón, migrar de MySQL a PostgreSQL no es muy complicado si se tienen a la mano algunos scripts que hagan el trabajo sucio por nosotros, pero aun así hay que hacer modificaciones manuales durante la migración. The solution was to restore databases using the. def isconsistent (self, other)-> bool: """ Whether the fields of the other Message object is consistent with the fields of self. For Plesk on Windows see this. The default behavior of setfacl is to recalculate the ACL mask entry, unless a mask entry was explicitly given. com is a community for Developers and IT Professionals. Consider the sequence ( 1)n+1. Unable to restore postgresql data dump for Django app hosted on Azure VM. Infima and suprema do not necessarily exist. Hello! I created database dump on server version 8. relation Same as table. We use your LinkedIn profile and activity data to personalize ads and to show you more relevant ads. We will use ! which is used by programming languages for Boolean reversing. It seems like the best solution to fix the core code so that PostgreSQL can emulate UPDATE IGNORE by the use of savepoints, which I have now done by copying the method from the existing emulation used for INSERT. If zero is given as the quantity, an empty sequence must be returned. Up-to-date, thousands of database applications has been designed using PostgreSQL and it. If you want to make your sequences all start just after whatever values are already there, it's possible to do that for most common configurations like this:. clients RENAME TO customers; Messages. All versions 9. If it does not exist, and the TEMPORARY keyword is not used, it drops a non-temporary table with the same name, if it exists. However the charmap should be provided by the locales packages. NET application, for example, in C# application using Npgsql. We do not have to worry. $dropdb library$ createdb library We now have an empty library database, and the table users does not exist in it. The performance boost occurs, since SSAS can navigate the relationship in order find and aggregate the measure without having to go all the way back down to the dimension key, if an appropriate relationship exists. OIDs in POSTGRESQL have 4 byte unique integers. Numeric promotions Integral promotion. Table %1 does not exist. This is the same as the definition of a sequential limit in the preceding section obtained by regarding the subset Dm ( f). If the login does not exist in the SQL Server instance or PostgreSQL database cluster, the Create Database User tool adds the login, creates a user in the database specified for the Input Database Connection, and creates a schema for the user in the database. General definition. How precise is this statement knowing that this limit is $\infty$?. To avoid this situation and make the drop command delete a user if present and quietly do nothing if the user does not exist, use the following syntax: DROP ROLE IF EXISTS role_name; As you can see, with this option specified, the command will complete successfully regardless of the validity of the role:. The pointwise convergent sequence in Example 5. First, check the link_tmp table data:. 6 or earlier version This commit addresses by removing increment_by or min_value - setval 3rd argument needs to set to false only when the table has no rows to avoid nextval() returns 2 where 1 is. person" does not exist。在创建table时,ca. In other words, with the use of quotes, the statement has explicitly requested that PostgreSQL interpret the identifier name literally. However if we drop the table public. When you validate the acquisition journal for the new fixed asset, the system displays the following message: Account for the transaction type, value model, does not exist for fixed asset FUR-ELC000136. Why compare to java, just leave it at C# does things C++ can't and C++ does things that C# can't. It has more than 15 years of active development phase and a proven architecture that ha. The sequence generator generates sequential numbers, which can help to generate unique primary keys automatically, and to coordinate keys across multiple rows or tables. Any cookies that may not be particularly necessary for the website to function and is used specifically to collect user personal data via analytics, ads, other embedded contents are termed as non-necessary cookies. 23 but does in all later versions. This PostgreSQL tutorial explains how to use PostgreSQL JOINS (inner and outer) with syntax, visual illustrations, and examples. So, if you need to perform a DELETE operation based on the values in another table, easiest way is to use the EXISTS or NOT EXISTS clauses. To come back to the address example: if a country does not have postal codes, it would be a problem if a country in this list would have a postal code. In this tutorial, read about how you can use psql to list databases and tables in PostgreSQL. Postgres in Cygwin for Rails 3 November 14, 2011 Posted by ficial in Blogroll, cygwin, ruby on rails, techy. Some have argued that common-sense flow-of-time theories can be compatible with eternalism, for example John G. pl -h yourwebserver # Securely edit the sudo file over the network visudo # Securely look at the group file over the network vigr # Securely seeing. The fastest to create a sequence called hibernate_sequence. Schema support permits the logical grouping of tables, views, and other all the data - as if the partition tables do not exist - creating a useful abstraction. Do not spend all your time to figure out which changes in the input make the problem go away. It then built the SQL statement "select * from public. But I want it to rename a table with the same name if it finds an existing one. This shows for example that in Examples 2 and 3 above, lim x→0 f(x) does not exist. Creating a Sequence: Example. If the file does not exist, a new zero-byte file is created, which is often useful to signify that an event has occurred. (created the postgis_template with postgis. Also being a system admin i checked under outgoing mail log and even notification is also sent to the assignee for ticket creation. Hyperbolic functions The abbreviations arcsinh, arccosh, etc. Postgres Help. This query uses WITH construction that works similar to sub-queries. If it exists, then we can update it with options to keep the historical records or not. Debian Bug report logs - #475069 postgresql-8. So I proceed by following steps, Before that I want to share my Rails application stack which is Rails 3. It has more than 15 years of active development phase and a proven architecture that ha. For our test case, the table in schema SCOTT would use the original sequence in schema DANI. A connection can be established between two machines only if a connection between the two sockets does not exist, both machines agree to the connection, and both machines have adequate TCP resources to service the connection. Sign up for Docker Hub Browse Popular Images. Why does it take an hour or more to come back and tell me that it can't drop the table? I think that the remote database is reachable because other lines in the job that reference it work fine. Not only does it power our API, Postgres is used extensively in a number of internal services. First, it was selecting sequences from the system metadata tables. 2 is still months away from being "production quality". I may have given up on LINQ as an interface to my DB, for example. C C Figure 1: Left: The limit exists because the left and right limits are both L. in [IP01, IPZ01], which postulates that such algorithms do not exist2. For example, if all schist rocks were produced by the same kind of metamorphism. If this happens, we say that this limit is the sum of the series. Both are PostGIS enabled. I noticed almost all the resulting errors were due to references to these partitiioned tables, such as "relation "table_name" does not exist". How closely related are humans to apes and other animals? How do scientists measure that? This changed in the 1950s and 1960s when protein sequence data and DNA sequence data, respectively. With the UTF-8 encoding,. Read does not directly correlate to FETCH. Welcome! Check out our FAQ and read through the Community Guidelines. Steve Klabnik gives an overview of Rust’s history, diving into the technical details of how the design has changed, and talks about the difficulties of adding a major new feature to a. Appendix A presents a review of Circular Time-Reversal and explains why the word circular is used. If the source operation was an update and the row does not exist, then an insert will be attempted. Operator does not exist: timestamp without timezone and double precision Oracle supports addition operation between timestamp without time zone and double precision datatypes in SQL operations or in VIEW/MVIEW creation. In Postgres, every table requires at least one PRIMARY KEY column that contains a unique set of values. For example: SELECT product_name, product_description FROM products WHERE product_name NOT LIKE 'H%'; By placing the PostgreSQL NOT Operator in front of the LIKE condition, you are able to retrieve all products whose product_name does not start with 'H'. Issue happening in porting postgresql database to newly setup database. Do not dump data for any tables matching the table pattern. These segments show the changing of, and relationship between, the client's and server's sequence and acknowledgement numbers. A series can have a sum only if the individual terms tend to zero. I recently did a backup of production and restored it to development. The relation does not exist in postgresql 2013-12-18 09:03:27 0; PostgreSQL: Exception when relation does not exist 2014-05-23 10:38:38 0; Common table expression,WITH clause in PostgreSQL ;ERROR: relation "stkpos" does not exist 2014-06-05 05:35:32 0. Not only does it power our API, Postgres is used extensively in a number of internal services. May also contain other transcript types that do not exist in their own individual files. In difference to ADO, ADO. At this point in time, the control file had only knowledge of the fact that the USERS tablespace has one data file and not two and the restore and recovery process will not try to do anything with regards to the newly added datafile. Column of relation already exists postgres. CVE-2018-1058 : A flaw was found in the way Postgresql allowed a user to modify the behavior of a query for other users. NET data provider. Josephus' language here (and throughout the passage, where it is not regarded as interpolated) is a "middle ground" between Christian acclaim of Jesus as divine and Jewish referral to Jesus as a magician and a deceiver -- against this spectrum, the words of Josephus are neutral and noncommital (and also parallel his treatment of John the Baptist). Oracle does not support WITH RECURSIVE, but has its own CONNECT BY syntax. Postgres sequence problem after DB restore - Help please. virendersharma Tuesday, September 27, 2011. not believe that bottom lines exist in history. You wish to introduce a new element (or maybe two elements), infinity, which you wish to append to the real number system. 5 support, and much more! Showing 1-1 of 1 messages. 11, Resistance to penicillin, as an additional diagnosis. Suppose you want to remove all rows from the linktable that have values of the id columns are in the link_tmptable. He struggled as a man condemned to death struggles in the hands of the executioner, knowing that he cannot save himself. For example, you can use it to upload changes to your product catalog. Note that the value of the function at x = c is not the same is the limit itself. (a)This is possible. Existing permanent tables with the same name are not visible to the current session while the temporary table exists, unless they are referenced with schema-qualified names. No other user is able to log into PostgreSQL. This sequence could be used to provide customer ID numbers when rows are added to the customers table. This method is provided as the alternative to overriding equality; often, pointer equality is the desirable means for comparison, but equality of the fields is also necessary. If one or both of the following are true, then clear the Recover Database Server check box. PostgreSQL Sequence: The sequence is a feature by some database products from which multiple users can generate unique integers. (Think plumbing). These options will force PostgreSQL parser to accept non-ANSI-SQL-compatible escape sequences (Postgre will still issue HINTs on it; you can safely ignore them). Multiple Death states can be used depending on the type of damage that kills the actor. PSQLException: ERROR: relation "AO_B9A0F0_APPLIED_TEMPLATE" does not exist. As a user with admin privileges in postgres, I can psql database_name and then invoke \dt to see a list of relations. This includes low-level protocol tools, a driver(PG-API and DB-API), and cluster management tools. If not, we say that the series has no sum. to the "SEQUENCE NAME"), why it is not allowed? But there is another DDL which works OK, but not reflected in the docs: test=# ALTER TABLE itest3 ALTER a RESTART 2; ALTER TABLE. You wish to introduce a new element (or maybe two elements), infinity, which you wish to append to the real number system. According to the definition, a function will be differentiable at x if a certain limit exists there. The fact that a coin lands heads does not fix that the coin landed heads by chance, because if it was simply placed heads up, as opposed to tossed in a normal fashion, we have the same outcome not by chance. If the login does not exist in the SQL Server instance or PostgreSQL database cluster, the Create Database User tool adds the login, creates a user in the database specified for the Input Database, and creates a schema for the user in the database. company schema names always start with 'company' and end with the company number. hibernate_sequence" does not exist By default, Hibernate will try to work off of a single sequence called hibernate_sequence , so to use the default behaviour, we should create this sequence before running our application (maybe in a migration). Examples You may find it useful to know how long it has been since a certain event occurred or a condition was true. We do not have a code list for this; and the codes only tell you about gauge cluster faults, not powertrain problems. Postgres sequence problem after DB restore - Help please. Our test server is running Apache 2. pgAdmin will otherwise continue to treat the server as a PostgreSQL 11 server resulting in these. Test­ing if some­thing exists: is defined, length, is not null, is not empty. Waiting for PostgreSQL 10 – Add pg_sequence system catalog On 20th of December, Peter Eisentraut committed patch: Add pg_sequence system catalog Move sequence metadata (start, increment, etc. PostgreSQL Tutorial; Relation does not exist; Constraint of relation does not exist; Constraint for relation already exists; Multiple primary keys for table are not allowed; Column specified more than once; Column must appear in the GROUP BY clause; INSERT has more target columns than expressions; Null value in column violates not-null constraint. Note that you do not need to know the name of the cursor to process the result set. Because EXISTS and NOT EXISTS only return TRUE or FALSE in the subquery, the SELECT list in the subquery does not need to contain actual column name(s). On a new ubuntu 12. This is an incident happened of vCenter Appliance with Postgresql and not uncommon. PostgreSQL - Python Interface - The PostgreSQL can be integrated with Python using psycopg2 module. If the login does not exist in the SQL Server instance or PostgreSQL database cluster, the Create Database User tool adds the login, creates a user in the database specified for the Input Database Connection, and creates a schema for the user in the database. It seems like the best solution to fix the core code so that PostgreSQL can emulate UPDATE IGNORE by the use of savepoints, which I have now done by copying the method from the existing emulation used for INSERT. LEFT JOIN / IS NULL uses exactly same execution plan as NOT EXISTS does and yields same results in same time. There are currently two built-in set types, set and frozenset. The only host entries in pg_hba. Specify NOORDER if you do not want to guarantee sequence numbers are generated in order of request. However the charmap should be provided by the locales packages. Welcome! Check out our FAQ and read through the Community Guidelines. postgresql. it tries to get the created id by calling currval on mantis_email_table_id_seq and it fails as the sequence doesn't exist (and if we create it by hand, it's not the incremented sequence so it fails). In SQL this is extremely fast and returns me a bit that tells me very simply if the record exists or not. Is this a safe state?. to keep track if the table exist yet or not and he can insert his data with confidence. For floating-point and complex numbers, +x is the same as x, while -x is the negation of x. The table or view entered does not exist, a synonym that is not allowed here was used, or a view was referenced where a table is required. But you can send reports to < [email protected] pg_restore -C -d postgres db. 2 is still months away from being "production quality". C C Figure 1: Left: The limit exists because the left and right limits are both L. It is the application that creates the relationship, and then only by virtue of what the application code is doing. Restore postgres Dump gives Errors: types do not exist. While there is no longer any significant performance advantage, using NOT EXISTS will avoid unexpected results when the subquery’s source data contains NULL values. It includes a comprehensive collection of TAP-emitting assertion functions, as well as the ability to integrate with other TAP-emitting test frameworks. In young infants, when a toy is covered by a piece of paper, the infant immediately stops and appears to lose interest in the toy (see figure above). Unable to restore postgresql data dump for Django app hosted on Azure VM. So if Red Hill contained a valuable mineral and in removing the mineral the hill was replaced with a giant pit, Red Hill would cease to exist. On some systems, this locale category does not exist. Assignment or comparison loses precision. item to inventory. I had a table as source, but I modificated it and rename it (the old name was not intuitive) and then I change the table at app as it needed. Do not output commands to set ownership of objects to match the original database. Since you didn't specify a sequence table name, it will look for a sequence table named hibernate_sequence and use it for default. If no table is specified, then the sample_percent value applies to the entire export job. It is that the one-size-fits-all approach does not offer a snug fit. Note: Foreign Key Constraint prohibits the creation of values in: 'messagescategories' that DO NOT exist in table: 'messages' d. clients RENAME TO customers; Messages. Steve Klabnik gives an overview of Rust’s history, diving into the technical details of how the design has changed, and talks about the difficulties of adding a major new feature to a. This will probably not help solving it. Also, once again, do not waste your time guessing why the bug exists. Above are two examples. In conclusion, the NOT EXISTS and NOT IN behave differently when there are null values involved. The sequence generator generates sequential numbers, which can help to generate unique primary keys automatically, and to coordinate keys across multiple rows or tables. loc and create recship if inventory. Using createdb a command-line executable. This is a complete, one-page listing of changes across all Postgres versions. If the file does not exist, a new zero-byte file is created, which is often useful to signify that an event has occurred. Not surprisingly, interpreting IRR results in a way that makes practical sense for investors and decision makers is a challenge. NET application, for example, in C# application using Npgsql. This is for tables, indexes, constraints, and views. Watch 439 Star 6. 1057 The account name is invalid or does not exist, or the password is invalid for the account name specified. In order to do that first, we drop the database if it exists and then create it. They are still only concerned with what is going on around the point. If not, we say that the series has no sum. 28 Why are my table and column names not recognized in my query? Why is capitalization not preserved?. Restore postgres Dump gives Errors: types do not exist. // Start a transaction NpgsqlTransaction t = conn. I have a large polygon table (30+million rows) and a smaller polygon table (45,000 rows). fromloc to inventory. You either want the createdb command or create it manually from within psql. customers" does not exist. Do not assign a code from subcategory Z16. Looking back over the sequence of graphs (and perhaps generating some others) the graph of h(x) always has a line of symmetry parallel to the y-axis. If the database does not exist, then it will be created and finally a database object will be returned.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.2311094105243683, "perplexity": 1741.9888062257692}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370497171.9/warc/CC-MAIN-20200330150913-20200330180913-00216.warc.gz"}
https://community.powerbi.com/t5/Desktop/Avg-Hrs-Worked/m-p/420829
cancel Showing results for Did you mean: Frequent Visitor ## Avg Hrs Worked Hi! Need help please on the following calculation that I need to translate to PBI. Here's what I got in Excel. I need to be able to compute for the overall avg hours work per site and account using the formulas noted on the screenshot. Thanks for all the help! 1 ACCEPTED SOLUTION Accepted Solutions New Contributor ## Re: Avg Hrs Worked Hi @minapot, Try this: ```Avg = IF ( HASONEVALUE ( table[emp_id] ), DIVIDE ( SUM ( table[total hrs worked] ), COUNT ( table[wknum] ) ), DIVIDE ( DIVIDE ( COUNT ( table[wknum] ), [measure to divide] ), DISTINCTCOUNT ( table[emp_id] ) ) )``` Replace the text in red with the appropriate measure since you did not mention which column to divide the count of weeks by for the average. "Tell me and I’ll forget; show me and I may remember; involve me and I’ll understand." 3 REPLIES 3 Established Member ## Re: Avg Hrs Worked Hello, CountOfWeeks=Count(Table[wknum]) Total_hrs=Sum(Table[total hrs worked]) avg_hrs=[Total_hrs]/[CountOfWeeks] Best regards. Highlighted Frequent Visitor ## Re: Avg Hrs Worked Hi Floriankx! Thanks for your reply! I tried your suggestions but failed to get the correct output. For the overall avg hrs worked, I need to have the total worked hrs for the site/account and then divide that by the average number of weeks the employees for that site/account worked and then divide the answer to the distinct # of employees for that site/account. so... site/account avg hrs worked = (total hours worked per site/ave # of weeks with schedule)/distinctcount(HC) New Contributor ## Re: Avg Hrs Worked Hi @minapot, Try this: ```Avg = IF ( HASONEVALUE ( table[emp_id] ), DIVIDE ( SUM ( table[total hrs worked] ), COUNT ( table[wknum] ) ), DIVIDE ( DIVIDE ( COUNT ( table[wknum] ), [measure to divide] ), DISTINCTCOUNT ( table[emp_id] ) ) )``` Replace the text in red with the appropriate measure since you did not mention which column to divide the count of weeks by for the average. "Tell me and I’ll forget; show me and I may remember; involve me and I’ll understand."
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8689180612564087, "perplexity": 4777.3576913244615}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-30/segments/1563195525659.27/warc/CC-MAIN-20190718145614-20190718171614-00000.warc.gz"}
https://cs.stackexchange.com/questions/12147/variations-of-omega-and-omega-infinity
# Variations of Omega and Omega infinity Some authors define $\Omega$ in a slightly different way: let’s use $\overset{\infty}{\Omega}$ (read “omega infinity”) for this alternative definition. We say that $f(n) = \overset{\infty}{\Omega}(g(n))$ if there exists a positive constant $c$ such that $f(n) \geq c\cdot g(n) \geq 0$ for infinitely many integers $n$, whereas the usual $\Omega$ requires that this holds for all integers greater than a certain $n_0$. Show that for any two functions $f(n)$ and $g(n)$ that are asymptotically nonnegative, either $f(n) = O(g(n))$ or $f(n)= \overset{\infty}{\Omega}(g(n))$ or both, whereas this is not true if we use $\Omega$ in place of $\overset{\infty}{\Omega}$. I am trying learn Algorithms. But I am unable to prove this. Can the experts help me ? • Try to use the definitions, keeping in mind that for every property $P$, either $P$ holds for infinitely many integers, or $P$ does not hold for almost all integers. Observe that $\Omega^\infty$ is the negation of $O$. – Shaull May 20 '13 at 4:04 • See here or here. – Raphael May 20 '13 at 19:45 Hint: If $f(n) \notin \overset{\infty}{\Omega}(g(n))$ and $g(n)$ is asymptotically non-negative, then for all positive constants $c$, $f(n) \leq c \cdot g(n)$ for large enough $n$. This follows by ignoring the condition $c \cdot g(n) \geq 0$ and negating the definition of $f(n) \in \overset{\infty}{\Omega}(g(n))$. In fact, this way you get the stronger result that either $f(n) \in \overset{\infty}{\Omega}(g(n))$ or $f(n) \in o(g(n))$ (but not both). Further hint: You can start by showing that the negation of "$P(n)$ for infinitely many $n$" is "$\lnot P(n)$ for large enough $n$". • I can't understand the difference of infinitely many & for large enough n. do you know a source that can help me?? – Fatemeh Karimi Jul 16 '17 at 5:37 • I suggest a good grounding in set theory and formal logic. Something holds for all large enough $n$ if there exists $n_0$ such that it holds for all $n\geq n_0$. It holds for infinitely many $n$ if the set of $n$ for which it holds is infinite. – Yuval Filmus Jul 16 '17 at 6:00 I can give you an example so that you can better understand $\overset{\infty}{\Omega}(g(n))$. Imagine a binomial heap. The insert operation is $O(logN)$, but is it ${\Omega}(logN)$? In cases when we have tree ranks of 4-3-2-1-0 and inserting a tree with rank 0 will be a ${\Omega}(logN)$ operation. But inserting a tree with rank 0 on the resulting heap from the previous operation (heap with having tree rank of 5) will be a $O(1)$ operation, since only pointers should be added and no extra merge work is necessary. This is the essential difference between ${\Omega}$ and $\overset{\infty}{\Omega}$. For example the binomial heap insert operation is $\overset{\infty}{\Omega}(logN)$ for set of $n = \{{1,3,7,...,2^k - 1}\}$. It doesn't state that when $n \geq n_0$ the complexity is ${\Omega}(logN)$ but rather for some infinite set of n, but not for all $n \geq n_0$ • @YuvalFilmus please correct me If I'm wrong – denis631 Jan 14 '18 at 11:06
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9181870818138123, "perplexity": 152.6413055077678}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875143805.13/warc/CC-MAIN-20200218180919-20200218210919-00172.warc.gz"}
https://link.springer.com/article/10.1186/s13673-017-0121-6
## Introduction Technology is the foundation of modern-day society. The Internet, for instance, is critical in facilitating large scale communication, trade, information sharing and learning. While there are numerous ways in which humans can interact with the Internet, websites play one of the most central roles. Current estimates indicate that there are more than 1 billion websites today [1], many of which act as the first point of contact for numerous individuals as they navigate through cyberspace. These sites offer a mixture of static and dynamic informational content, typically with the aim of encouraging user engagement and maintaining repeat visits. Throughout the literature, the topic of websites has been researched in several contexts and for a variety of purposes. One of the most noteworthy areas of work has been on enhancing user experience and building trust in sites. For instance, articles have identified and examined factors crucial to website design [2], defined frameworks for evaluating the quality of websites [3], and conducted user studies (e.g., via eye-tracking) to assess how websites are used and any innate complexities in perceptions [4]. Apart from usability, there is a significant strand of research exploring the behind the scenes activities, particularly as it relates to information processing and, the online tracking and profiling that occurs as users browse websites. These relate to the passive collection of user information often without user knowledge [5], the unsanctioned tracking of users and user web-surfing behaviour across websites by third-party trackers [6], and the proliferation of complicated privacy policies which make it difficult for users to understand conditions of website use [7]. Such work highlights the clear tension between users concerned about maintaining privacy and anonymity online, and advertising companies seeking to better track and profile users across the web. This article engages in research in the area of online web references. Specifically, we consider the topic of websites in the context of third-party references and the online advertising ecosystem. Through an analysis of the top-accessed 250 websites in the UK, USA, Germany, Russia and Japan, we empirically examine the extent to which third-party sites are referenced by these popular domains. Our research objectives in this article are: (a) to investigate these ‘behind the scenes’ references; (b) to explore the size and characteristics of their ecosystem (i.e., viewing the ecosystem as an entity that can be studied), and (c) to consider the perception and understanding that users have of this ecosystem. While previous literature has explored this general topic (e.g., [6, 8, 9]), we believe our work is novel in several regards as highlighted below. Firstly, as we analyse the links to these third-party websites, we also construct a model graph of the relationships between sites directly and the third-party organisations that run them; this allows detailed insight into the ecosystem and its connectivity. The second contribution is a comparison of the information on localised third-party references through assessing the most used sites in five of the world’s leading economies. From this we can consider the top domains referred to and assess how these compare in prevalence across nations of different cultures and languages. Lastly, we conduct a user study to determine the extent to which users understand these ‘behind the scene’ activities of websites including the origin of third-party sites and how often they are referenced. The remainder of this paper is structured as follows. The "State of the art" section reviews the current literature in online advertising and web tracking, most of which can be found in the online privacy field. In the "Methodology" section we detail the methodology that we adopt for our research study. This explains the sample of websites selected for analysis, the process through which we analyse them, and how we incorporate a user study to enrich our research. We then present and discuss our results in the "Results and discussion" section; these cover both the website analysis and the user-based study. Finally, we conclude and outline avenues for future work in the final section. ## State of the art Online advertising is an enormous industry, with advertising revenues reaching $27.5 billion in the first half of 2015 [10]. This figure climbed to$59.6 billion by the end of 2015, with both Google (with $30 billion) and Facebook (with$8 billion) being the biggest benefactors [11]. One of the key differences between online and offline advertising is the ease at which individual users online can be profiled, and then targeted by marketers based on their browsing history, personal information and location. There are various approaches that have been used to gather user information and track their movements online. Traditionally, tracking cookies and scripts have been the main way of monitoring who is visiting a particular website. As sites are visited, browser cookies act as a unique identifier, or scripts—many of which are provided by third-parties—execute and attempt to collect user data or provide customised content (e.g., information or advertisements). Good examples of the proliferation of these trackers and how they work can be found in existing studies [6, 9], including analyses of various tracker properties. Due to changing attitudes to user privacy, the increased usage of mobile phones, and recent legislation [12], other approaches to track users have surfaced. Browser fingerprinting [13, 14] is becoming an increasingly popular means of user identification in place of the more traditional cookie-based approaches. This is a technique that uses web scripts (often originating from third-party websites) to collect various information about a remote computer in order to try and (re)identify that particular machine (as a conduit to the user). The collected information can incorporate a wide variety of features including screen resolution, time zone or the availability of specific font sets. The established field of information science [15] has been fundamental to the processes mentioned above. This is linked to the fact that companies have continued to pursue novel ways to gather information (and metadata) on prospective customers online, and to process and manage that information for optimum use. Common activities include leveraging big data to optimise digital marketing [16], and developing approaches to extract individual user profiles from online surfing information for purposes of behavioural targeting [17]. These concerns will almost certainly increase in the future as technology becomes more ubiquitous [18]. There have been various approaches proposed to elucidate and, in the interest of privacy, to block the range of external scripts and web tracking attempts. One stark reality however is that fully preventing third-party references is non-trivial, as often scripts are required in order for sites and services to function properly. For example, a number of social media sites will use widgets (e.g., the Facebook ‘Like’ button) to identify individuals but also as a means of the user interacting with content. If all references (via their respective scripts) are blocked then this will limit the usefulness of many core services and sites. Moreover, an increasing amount of websites (e.g., Forbes) are preventing site access if they detect that the user is using software (e.g., Adblock [19]) that blocks digital advertisements. Ad-blocking software also comes with its own risk as seen recently when a fake Adblock Plus extension made its way on to the official Chrome Web Store, and was downloaded by thousands of users [20]. Wu et al. describe a solution called ‘DMTrackerDetector’ which detects third-party trackers automatically using structural hole theory and supervised machine learning [9]. Nikiforakis et al. identify that one of the most significant issues is not that the generated fingerprints are unique but that they can be used to link the device across multiple visits [21]. Their work proposes ‘PriVaricator’, a solution which aims to add an element of randomisation to make the fingerprinting process non-deterministic and thus making it more difficult to link fingerprints across multiple user visits. A salient point regarding privacy is that the way in which Internet users perceive privacy online is something that is still not fully understood. This is particularly related to attitudes towards online advertising and third-party website referencing and tracking. Melicher et al. interviewed Internet users to better understand their concerns about web tracking [22]. Interestingly, their results suggest that users want more control of tracking but are unwilling to invest the effort to control tracking and often distrust current tools. This point can be seen in much earlier works [23, 24], and typifies the existence of the well-researched Privacy Paradox [25]. In a recent TRUSTe study, 92% of British Internet users were concerned about their online privacy, but in contrast only 25% actually understand how organisations share their personal information [26]. In another context, Melicher et al. collected browsing histories and interviewed 35 people in USA about the benefits and dangers of online tracking [22]. The research found that, as with similar studies, individuals would like greater control over tracking but also that users were able to see benefits to controlled tracking. This study found that users were largely unwilling to dedicate the effort to control the tracking of their data. In this article, we aim to build on the research of many of the articles mentioned above, to examine the size and characteristics of the current third-party website (including online advertising) ecosystem. This includes drawing out common networks and associations of sites, and any third-party networks. ## Methodology To briefly reiterate, our research objectives are threefold. Firstly, we aim to investigate the ‘behind the scenes’ references made in popular websites today. Secondly, this work explores the size and characteristics of the ecosystem in which these sites reside and the links that are made between other entities (i.e., viewing the ecosystem as an informational entity that can be studied). Finally we consider the perception and understanding that users have of this ecosystem, and thus explore the human-computer interaction perspective in more detail. To structure our research, we adopted a methodology that involved a combination of technical and user-focused tasks. We felt that this was the most appropriate approach given the technical nature of analysing website source content, but also the importance of considering the understanding human users may have on third-party references and the online advertising ecosystem. There are four main steps in our approach. The first step was the selection of websites that would be most appropriate for our research. Given our interest in also conducting a cross-country comparison of sites and ecosystems, we decided to focus on five of the major economies today and those with a significant number of citizens online. These were likely to be countries where third-party references were the most active. The countries that met our criteria were the UK, USA, Germany, Russia and Japan; China, India and Brazil were also viable, however we were unable to reach other selection criteria as will be outlined below. We decided to scope our assessment to the top 250 sites in each of these countries, and used the rating index as as defined by Alexa [27], the commercial web traffic data and analytics company. The second step comprised of analysing each website’s HTML source and extracting the references to third-party scripts. These can typically be found within the <script> elements of the page, and included either as a src (source) attribute (e.g., <script type=“text/javascript” src=“http://...” ></script>) or in the script body itself (e.g., <script>.. js.src=“//...” .. url=“//..”</script>). We identified third parties as those references which did not match the current domain. As the number of sites and potential variety of external references could be quite high [28], we opted for a semi-automated approach to this task. Specifically, we developed a python script based on a combination of modules (e.g., urllib, beautifulsoup) and tailored regular expressions, to allow extraction of the desired external links. To ensure that we captured all of the third-party references, we adopted an iterative method to development and testing, by comparing our script’s output to that of manual analyses; we conducted our analysis on a site-by-site basis. Script extraction tools, such as LightBeam [29], were also very useful at supplying additional evidence and support. In situations where third-party websites or links made follow-on references to other sites, we recorded those as well; this allowed us to build a more complete picture of the range of third-parties involved. Most importantly, and unlike in many other studies in the field, we ensured that each site that was visited was accessed as if we were physically in that country—Virtual Private Network (VPN) services and a physical presence in two cases, allowed for this. This acted to increase the authenticity of the third-parties referenced as initial sites were accessed; unfortunately, we were unable to find reliable VPN services for China, India and Brazil hence the exclusion of these sites. The output of this task was a list of third-party domains associated with each of the top websites. Having gathered the listing of third-party website references of all sites, the next step concentrated on its analysis. We engaged in two levels of assessment. The first level sought to apply social network analysis (SNA) techniques to build a visual network that would better allow the identification of relationships within the site-domain listings. SNA can be described as a set of methods that allow the links between elements (e.g., people, devices, or things) to be studied [30]. We were keen on applying this technique to identify the third-party domains that were the most common across sites generally, or specifically when considering the site’s locale. Moreover, we could identify the domains central to the created network, how cohesive the network is, and clusters of highly connected domains. This was our information processing and interpretation stage. Based in part on the analysis from the first level, we then considered the companies behind the third-party website domains. This attempts to demystify the domains referenced and elucidate the organisations orchestrating them. As highlighted in other works (e.g., [8]), we expect that while there will be hundreds of domains, there will only be a small subset of core sources representing the giants in the online advertising and analytics ecosystem. The last step in our methodology was to conduct a user study to determine the extent to which average web users understand the ‘behind the scenes’ activities of websites they visit, including origins of online advertisements and the extent to which third-party scripts are referenced. We aim to build on previous work (e.g., [24]) that has examined the user’s perspective and pose additional research questions to gather perceptions on topics such as which types of sites users think are the most private and estimations about third-party website references. ## Results and discussion ### Analysis of third-party website references and the online advertising ecosystem #### Overview of websites and size of the ecosystem Our approach to analysing the ‘behind the scenes’ ecosystem began with an assessment of the 1250 websites selected. The websites examined were of various types including search engines, news agencies, social media, commerce and finance, professional networks, video streaming, and cloud services. While each country had its unique set of websites, there were several common domains present, and with similar degrees of popularity. For instance, well-known sites such as google.com, youtube.com and wikipedia.org all featured in the top 10 sites of the five countries. The largest similarity in web domains was between the UK and USA (at 46%), but this is to be expected given the likeness of the two countries (e.g., in society and language). This was followed by Europe’s UK and Germany (at 35%), and then the USA and Germany (at 29%). Unsurprisingly, the most dissimilar countries were Russia and Japan (at 9%) and the USA and Russia (at 11%). Language spoken and location are likely to be reasons for these differences. It is important to note such similarities and differences as they may help explain differences in ecosystems across countries in the subsequent sections. From our analysis of each country’s websites, we were able to extract a comprehensive list of third-party sites that were referenced. We used a simple “source-to-reference” list format which also catered for third-party sites making their own references; where source was the initial site and reference was the site referenced. The real advantage of this format however, was the ease at which social network analysis (SNA) tools could be applied to allow the characteristics of the third-party ecosystem to be explored and visualised. As an example, we present Fig. 1, which was created using the SNA tool, Gephi.Footnote 1 This figure displays a network constituted of the UK websites, the third-party sites referenced, and the various connections between them. The first point of significance regarding Fig. 1 was the substantial size of the network created based on only 250 initial sites; we do not intend this image to be legible but instead to use it to exemplify the network’s vastness. Overall, the network contained 1078 unique domains (or in graph terms, nodes), with 4242 connections (also known as directed graph edges) between them. In this context, connections denote references that indicate third-party website calls, potential links to advertisement networks, or associations to legitimate sites (e.g., bbc.com making references to bbci.co.uk for images). From a user perspective two key aspects to highlight here are, the amount of additional domains referenced (4 times the initial number), many of which are likely to be unknown to users, and the large number of connections between these various sites. Moreover, as can be seen, some of the domains are very well connected and may act as hubs that gather user information from various sites—this point will be discussed further in the next section. With regards to the characteristics of the other networks, the USA possessed the largest and most dense network with 1135 domains and 4455 references. It was followed by Germany with 1034 domains and 3808 references, Russia with 976 domains and 3582 references, and Japan with 796 domains and 2967 references. Each of these networks highlights a notable number of third-party sites that are referenced as individuals browse webpages. Although the size of these networks was a noteworthy factor in itself, our most intriguing finding was the positioning of Japan given their large market size (126 million citizens) and high Internet adoption rate (at 91%) [31]. In particular, Japan has the 5th largest number of Internet users in the world (second only to USA in our study) and with a 91% Internet adoption rate is only behind the UK (at 92.6%) in our study. A market such as this would be ideal for web advertising networks, hence our expectation of a larger third-party website prevalence. Upon further investigation, we found that the reason for this disparity may not due to lack of attempts, but rather because Japan still primarily focuses on television advertising [32]. This is in contrast to markets such as the USA and UK where online advertisements—as might be inferred from the size of the networks discussed—are the dominant way to reach consumers. #### The big players: third-party sites referenced the most The networks discussed above provide a ‘bird’s eye view’ of the website ecosystem and its complexities. Our aim in this section is to further understand that ecosystem, and in particular, the most referenced sites. This would allow us to empirically identify the central third-party websites in each country, the key players behind these domains, and how they may differ across locales. To determine the most referenced websites in the country networks that were created, we calculated the in-degree network centrality (i.e., an SNA metric that measures the number of incoming connections) for all of the websites. We then used this in-degree measure to proportionally size the various nodes in each of the networks such that the larger the node the more times it was referenced. Figure 2 shows the visual of the UK network from Fig. 1 updated to account for this metric. For purposes of readability, we also present an alternative visualisation in Fig. 3 which focuses on the top 50 domains. From an analysis of each website’s in-degree, the top ten were: doubleclick.net (189 references to it), google-analytics.com (134 references), google.com (117 references), facebook.com (107 references), google.co.uk (90 references), adnxs.com (82 references), scorecardresearch.com (71 references), googletagmanager.com (56 references), twitter.com (55 references) and quantserve.com (52 references). AppNexus (the owner of adnxs.com), for instance, regards itself as an independent technology company that provides trading solutions and powers marketplaces for Internet advertising—they partner with, and are backed by, companies such as Microsoft [33]. The aim of the next company, Scorecard Research (scorecardresearch.com), is to collect data (in this instance, information on users’ website visitation patterns) to assists their clients in better meeting the needs of consumers.Footnote 2 Finally, Quantcast (quantserve.com) engages in website audience measurement for publishers and tailored online advertisement delivery.Footnote 3 Broadly speaking, a central goal of these and other third-party organisations is to facilitate a better understanding of website users (including their profiles and browsing patterns across sites) to allow for more unique and targeted marketing. This is intended to benefit advertisers by enabling them to reach their ideal consumers, and consumers in ensuring they are shown the more suitable ads. The networks of the other countries were similar to that of the UK with Google dominating in the top 10 most referenced sites. As shown in Table 1, Google domains, doubleclick.net, google-analytics.com and google.com were permanent features, with only facebook.com also appearing in all countries. There were also a range of new domains found, many of which were associated with other digital advertising firms. These include: DemDex (demdex.net), now owned by the American multinational Adobe Systems Inc., which captures behavioural data on users to allow for better targeting of online ads [34]. InfoOnline (ioam.de) is a German-based organisation focused on digital audience measurement. In Russia, Yandex N.V. (yandex.ru) is a large technology firm which also engages in online advertising, while the site tns-counter.ru appears to be part of a project by a Russian enterprise (TNS) seeking to understand behaviour of Russian citizens on websites. Lastly, the technology company OpenX (openx.net), specialises in online advertising marketplaces. For those domains not directly associated with advertising firms, one reason for their prevalence may be social plug-ins (as discussed with Facebook and Twitter earlier). For instance, VK (vk.com) is a Russian social networking service that has site widgets and plug-ins for sharing and ‘liking’ similar to Facebook. We also have found plug-ins for auctions and marketplaces, which may explain the popularity of sites such as yahoo.co.jp (though we should note that Yahoo! does also engage in the online advertising industry [35]). While Table 1 is useful at highlighting the key domains in the ecosystem, we were also keen to explore the websites that were well referenced but not present in the initial 250 for each country. This would more clearly depict the prominent websites that were in the ecosystem only as a result of being referenced by other sites. Figure 4 presents our findings when considering the top 10 websites of each country. We also have noted the prevalence of these third-party sites in other countries to allow for comparison. Looking beyond the websites that are well-known to us based on earlier findings, Fig. 4 uncovers several additional domains. These include adition.com (from the German-based, ADITION Technologies), bluekai.com (BlueKai, which was acquired by Oracle in 2014), criteo.com (Criteo) and rubiconproject.com (Rubicon Project). These are all companies that engage in digital advertising of some form. One noteworthy point here is the fact that although certain organisations (e.g., AppNexus) are established in multiple countries, others appear to concentrate only on local networks. This seems particularly relevant in the German and Russian cases with sites such as adition.com, adriver.ru, ioam.de and yadro.ru. While culture may generally be a factor, it could also be a business decision to focus on core markets and their dominance. With respect to the US and UK, they have similar distributions across each of the various domains, with the US typically leading the UK. Figure 4, therefore, also mirrors the structure of those networks more broadly. Although Japan has the smallest network overall, it should be noted that several of the advertising domains are present which may suggest a growing interest by marketers generally. The dominance of Google in Japan however, is not to be overlooked. We compared the percentage of references made to the main Google domains (i.e., those in the top 10 in Table 1) across the countries and found that 20.42% of all (2967) third-party references in Japan sites are to Google domains. For the other countries, the US, UK, Germany and Russia are at 11.76, 13.81, 14.36 and 14.48% of all references respectively. It will be interesting to monitor this trend in the future and examine whether Google’s dominance expands, or other local advertising companies such as Dentsu and Hakuhodo establish a substantial online presence. #### The inbetweeners: third-party sites acting as a bridge In addition to analysing websites that were referenced the most, we also engaged in a brief assessment of the third-party sites that made the most references. In SNA terms, this is called the out-degree network centrality. Identifying these sites could indicate central points in the ecosystem which act as a connector or bridge for other sites, even if they themselves are not heavily referenced. In Fig. 5, we present a comparison of the top ten third-party websites that make the most references in each country (this is similar in concept to the depiction in Fig. 4). The first point of note as indicated in Fig. 5 was that several of the domains (e.g., adnxs.com, casalemedia.com, and doubleclick.net) and organisations uncovered in the previous section, remain present. This means that not only were these domains heavily referenced, but that they were also responsible for several references to other third-party sites themselves. A reason for this could be partnering third-party services, shared advertising networks or advertising marketplaces. Moreover, although we found that Google domains were again prevalent, they were not as dominant as when examining sites by references made. For instance, other domains such as those by AppNexus (adnxs.com) and Rubicon Project (rubiconproject.com) featured reasonably well here. Across countries, there were nuances in the domains present as might be expected. AdRiver (adriver.ru) for example, is a Russian company that specialises in Internet advertising technology; this explains their presence in Russia as opposed to in other nations. The US-based company AppNexus (adnxs.com), was responsible for a significant number of out-going references in the US, UK and Germany, but none in Russia or Japan. This finding echoes the low in-degree centrality measures for these countries in this domain (as shown in Fig. 4). In Japan, the Rubicon Project (rubiconproject.com) has the highest number of out-going references and could indicate a hub or main ‘bridge’ platform in the Japanese advertisement marketplace. This presence is the opposite in the US and Russia, where the Project is hardly prominent. While understanding the domains with the highest out-degrees allows some insight into the main intermediate third-parties, we were also interested in the specific sites that they reference. This could enable for more insight into the ecosystem. We therefore analysed each of the top ten domains for each country to determine what domains they referenced. As an example of our findings, we present Fig. 6 which is a directed network graph of the top UK third-party domains. To follow up on the potential association between these various third-party networks, we decided to conduct a brief exploratory study. For this, we returned to our initial network graphs of the full sites and references (e.g., as shown in Fig. 2). We then applied a clustering algorithm [36] to these graphs to determine the main associated groups and clusters that can be uncovered. Figure 7 depicts the clusters discovered in the UK network graph. From Fig. 7, we can see that the clustering algorithm identified three large clusters as represented in purple, green and light blue. The purple cluster is the largest one and accounts for $$\sim$$ 25% of the sites. It contains many of Google’s sites such as google.com, doubleclick.net, google-analytics.com and googletagmanager.com. An interesting addition to this cluster is facebook.com, given its own advertising ambitions, but this may be explained by the presense of these third-party references on many of the same sites. The green cluster is the second in size at $$\sim$$ 20% of the network. The prominent sites within this cluster include adnxs.com, rubiconproject.com, mathtag.com and yahoo.com. Visually analysing the cluster, many of the second-tier third-party sites and advertising companies (e.g., AppNexus and the Rubicon Project) can be seen. This better hints to the reality that advertising networks and organisations, particularly smaller ones, may work together in the third-party ecosystem. Finally, in the light blue and covering $$\sim$$ 13% of the network, is the third-largest cluster. Here there are more domains from advertisers (or enabling the provision of advertisements) such as googlesyndication.com, moatads.com, adsafeprotected.com, googletagservices.com. This cluster was somewhat understandable given the associations between the included sites, but it we might have expected a closer association with the first cluster (e.g., potentially the main Google domains being present in the same cluster). Possibly the most significant finding from this exploratory study was the close network (or grouping) between some sites that almost certainly alludes to a common advertising marketplace. This is an area we will seek to investigate more in our future work. ### User perceptions and understanding The results presented to date have highlighted the significance of advertising networks within popular websites across five countries. In fact it is clear to see that there are a number of networks that are prevalent across the world, with a core set of companies providing a large percentage of third-party content on websites. The next stage of our research involved understanding the perceptions of Internet users as it pertains to such networks. To facilitate our investigation, an online survey was carried out over the course of 3 months. The specific aim was to consider topics such as personal privacy, and also, to understand how aware users are of the prevalence of advertising networks and other third-party content on popular websites. The survey received ethical approval from our local university review board, and we also ensured to gain informed consent from participants. In total we received 109 responses, with questions covering standard biographic information (age, gender, education and technical understanding), perceptions of online privacy, and finally perceptions regarding the amount of third-party content that was linked to by a selection of popular websites (drawn from the Alexa Top websites). The majority of respondents to the survey were based in the United Kingdom, however, the survey was not limited by a user’s location with a range of countries being represented. The majority of responses to the survey came from females (54%) and 48% of the respondents were under the age of 34. The younger profile of the respondents suggested our sample had grown up with the Internet and associated technologies, therefore might be more knowledgeable in this area. This was further underlined by the participants’ high knowledge of computers and the Internet: over 67% of participants ranked themselves at 7 or above, with no-one ranking themselves below a 3 (this is based on self-reports using a scale of 1 to 10, where 10 was extremely knowledgeable and 1 was very little knowledge). After completing the biographic information, we queried participants about their own perceptions of online privacy and its importance to them; this would be especially important when considering knowledge of third-party networks. Specifically, they were asked to rate the importance of online privacy, using a 5-point Likert scale. In total, 80% of respondents expressed that online privacy was either very important (52%) or extremely important to them (28%). A Pearson’s Chi-Squared test was used to determine any correlation between the importance of online privacy to a participant and their perceived level of technical knowledge. This test resulted in a p-value of 0.85, which indicates that there was no correlation between perceived technical knowledge and the importance of online privacy. This independence could be interpreted as online privacy being an important concept regardless of the level of technical knowledge and understanding of the individual, which in itself is a positive finding. To assess participant’s perceptions and understanding of online advertisements, we asked them to indicate the category of sites they regularly visit and which of these categories they thought contained the most advertisements. Figure 8 shows the comparison between the types of sites that are commonly used and those which participants thought contained the most advertisements. The results show that the participants were of the opinion that news (15%) and social media (60%) sites were likely to contain the most advertisements. Interestingly, although that was felt to be the case, these were still the top two types of sites most commonly used by participants (with news at 75% of respondents and social media used by 86% of respondents). While there are several other variables that could be in play here, this could highlight the value of site utility over concerns of ads or online tracking. Our next aim was to understand how much participants were aware of some components of the underlying advertising ecosystem. For this, participants were asked to identify where they thought that the majority of advertisements on web pages originated from. A large proportion of respondents (90%) correctly identified that advertisements typically originate from a third-party, as opposed to the company that owns the site (as demonstrated in our earlier website analysis section). We also found that over half of participants were either ‘moderately aware’ (42%) or ‘extremely aware’ (21%) that when a website was visited several third-party sites were also automatically referenced. While this was good to see as it demonstrated awareness, approximately a third of participants (37%) were in the lower bracket of ‘not at all aware’, ‘slightly aware’, and ‘somewhat aware’. This is some cause for concern as it highlights that the activities that occur when users visit websites may not be transparent for all users involved. In addition to exploring participants’ awareness, we asked them to estimate the number of third-party sites that particular websites might make reference to. This would allow us to assess how well participants’ perceptions match reality (as presented in earlier sections). We used a subset of the main-site (or home) pages from the Alexa Top 250 such that they would cover a range of different categories and would be generally recognisable by individuals. While there were a number of UK-based sites included they were those that have a global reach. As previously discussed, while the majority of participants were indeed based in the UK the sites were specifically chosen to be recognisable by a wider, global audience. Table 2 contains a summary of the survey responses and for each site, it includes the minimum number of third-party sites stated, the maximum number of third-party sites stated, and the mean, median and mode number of sites stated across the entire set of responses. It should be noted here that we only asked individuals about the number of third-party website references on the main page of the site (e.g., google.com), and did not focus on sub-pages (e.g., google.com/news). A point that is immediately obvious from the table is the extremely large estimates that participants expressed about third-party links. Overall, we found that in 38% of responses users estimated site references of above 500 third-party links (far above the actual cases as mentioned in earlier). This highlights a sizeable disparity between perceptions and reality (as will be discussed below). To take Google as an example, it was the site that participants felt would contain the most references to third-party sites. One factor in this ranking could be that Google is a company built on data. Therefore, as much of Google’s business relates to collecting and analysing information, it is reasonable that participants would rank Google highly. In fact, from our earlier findings, we know that Google is one of the largest companies behind advertising networks. The comparison between the number of third-party links on a website (the ‘actual value’ column in Table 2) and participants’ perceptions of these links, is an interesting one. As can be seen, the estimates given by respondents are a long way from the values that were actually recorded in the course of our research. The mean values presented can be directly attributed to a number of outlying, and extremely high values for each of the websites as mentioned earlier. These values have acted to skew the mean and exemplify why the mean would not an appropriate centroid for comparison in such distributions. In Amazon’s case for instance, there is a mean value of 71,974, as values ranged from 1 to 5,000,000. In reality however, there were only 3 third-party references made on the site. Comparing these would be unhelpful, and would allow no insight into the difference between perceptions and reality. To address the issue of the extreme values entered by participants and the resulting skewed mean, we also examined the median (this is a common use of the median [37]) and mode values for each site. In Table 2 we present the median values and mode values (after value grouping), along with the total number of times (and respective percentage that) those values occur. While the median identifies the central value in the list of values supplied by participants, the modal groups attempt to categorise the values (into 25 value groups, e.g., 1–25 links, 26–50 links, and so on) and give an indication of which value-groups appear the most in the sample. Comparing the medians to the actual observed values, some were close (e.g., Facebook, Huffington Post and BBC), and overall, they were certainly within the correct order of magnitude—Google and Amazon being the main outliers. An interesting point to note here is that through the median, participants were largely correct at identifying the site with the highest number of links to third-party sites i.e., Facebook. Facebook generates a large percentage of their income through data and through advertising revenues, as alluded to earlier in our article. The fact that the survey participants were able to correctly identify this site as the one with the most third-party content demonstrates some understanding of which organisations are more likely to rely on third-party material. Conversely, Google was identified as a site that users would consider to reference numerous third-party websites. In reality, Google (at least google.com) ranks as one of the lowest sites in terms of links to third-party content. From assessing the mode groups, we could see that the most common values supplied by participants were actually within the 0–25 range for all sites. This inclination was somewhat accurate for a few cases, with the BBC, Spotify and TalkTalk, but not for others, such as Facebook and Expedia. The total percentage (shown in Table 2) is useful here as it highlights the proportion of participants with a value within the modal range. From this we can see that for the BBC for instance, most participants (67%) were relatively close to the correct value. However, for Google this was not the case as less than 1/3 of individuals were within the modal range. To investigate this further, we examined the second most common value ranges and found that Google and Facebook were both in the range of 76–100 with 16 and 16% of participants respectively. This acts to support our other findings above, and emphasise that while participants do have some understanding of the ‘behind the scenes’ activities, there are still misconceptions about sites and how much third-party referencing is conducted. ## Conclusions and future work In this paper we have investigated the various activities that occur when an individual visits a webpage, specifically focusing on third-party references and their proliferation across the web. These references often automatically execute in the background of a website without the express knowledge of a user. This paper has explored the nature of these references and the degree to which a website will link to third-party content, as its information set for study. Further to this, a survey of user perceptions has provided insight into how aware users actually are about this third-party content and its potential privacy implications. Our analysis of the top 250 websites in the UK, USA, Germany, Russia and Japan has shown, perhaps unsurprisingly, that the USA has the largest and densest advertising network. A more interesting observation was that Japan had the smallest advertising network even though it has a substantial market size of 126 million and a high Internet adoption rate of 91% [31]. When considering the different parties involved in advertising networks, our research found Google was the largest player thanks to DoubleClick, a Google owned company that specialises in developing and providing advertising services and was prevalent across all five of the countries studied. Similarly, Google Analytics, a web platform designed to offer website analytics was commonly seen across a wide-range of sites from all five countries. To reflect on the companies engaged in advertising networks, it was not just Google who had a significant interest. Facebook are another company who had a notable presence across all of the countries that were included within our research. That is not to say that advertising networks are solely controlled by a few large organisations. There were a reasonable number of smaller organisations within the space, for example, ScoreCard and ADNXS. Our findings indicated that some of these smaller sites, such as ADNXS, could be thought of as bridging sites. That is to say, these sites were identified as part of the ecosystem that act as a bridge to other sites, with the original site perhaps not being heavily referenced. The survey conducted on user perceptions of privacy on the web provided some interesting results and insights. One of the most notable findings was that comparing the median values to the observed values there were some very close results between the user estimates and the recorded values. The fact that participants were correctly able to identify Facebook as one of the sites with the most third-party link shows that there is some level of understanding of what types of organisations rely on third-party content. When assessing the modal groups of participant estimates, it became apparent that while there was some understanding of third-party references and the ‘behind the scenes’ website activity, there are still misconceptions about the true prevalence of third-party sites. There are a number of avenues through which we intend to engage in future work. The first of these is to broaden our study to encompass more countries, and particularly those with different cultures or those considered as developing states. This would strengthen our cross-country research and enable us to better compare and contrast a variety of country types. China for instance, would be interesting to study given the state’s control over the Internet, and India would be ideal to examine as the Internet is set to reach larger parts of the population. With a better appreciation of the prevalence of third-party website references across various nations, we would then seek to thoroughly explore the topic of online privacy in the wider context. As a society we are becoming increasingly reliant on the Internet and as such privacy should be an increasingly important consideration. The research discussed in this paper highlights the extent to which third-party organisations are being referenced often without the express permission of the user. Another important question that should be raised going forward is the extent to which user data is shared with such third-parties, especially if users are not made aware of the implications of this data being shared. The issue of online privacy is only going to become more pressing as our reliance on the digital world increases, and as such there are a number of questions about the extent to which gathered user data (and metadata) is passed between these third-party sites, and just how this data is used. The work in this paper provides an initial investigation into the extent that this data is gathered and shared without the explicit knowledge or consent of the user. In addition to this, we are keen to undertake a deeper user study by presenting the full third-party networks to the users—ideally in an interactive format-to help better understand their impressions of them, especially as it relates to privacy. While we note that many advertising companies offer ‘Opt out’ options on their websites, it would be useful to see if there was some effective way at creating a tool that would: firstly, identify sites where this option is possible; and secondly, streamline the process from the stage of visualising a third-party referencing network to selecting and opting out from sites.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3534216284751892, "perplexity": 1652.7474872761547}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-49/segments/1669446710417.25/warc/CC-MAIN-20221127173917-20221127203917-00087.warc.gz"}
https://www.nature.com/articles/s41467-017-01516-1?WT.feed_name=subjects_diseases&error=cookies_not_supported&code=b277b56d-6947-42f1-a52d-47dfb0bec801
Article | Open | Published: # Spatial competition constrains resistance to targeted cancer therapy ## Abstract Adaptive therapy (AT) aims to control tumour burden by maintaining therapy-sensitive cells to exploit their competition with resistant cells. This relies on the assumption that resistant cells have impaired cellular fitness. Here, using a model of resistance to a pharmacological cyclin-dependent kinase inhibitor (CDKi), we show that this assumption is valid when competition between cells is spatially structured. We generate CDKi-resistant cancer cells and find that they have reduced proliferative fitness and stably rewired cell cycle control pathways. Low-dose CDKi outperforms high-dose CDKi in controlling tumour burden and resistance in tumour spheroids, but not in monolayer culture. Mathematical modelling indicates that tumour spatial structure amplifies the fitness penalty of resistant cells, and identifies their relative fitness as a critical determinant of the clinical benefit of AT. Our results justify further investigation of AT with kinase inhibitors. ## Introduction Kinase inhibitors targeting signaling pathways have shown major value in targeted cancer therapies but generally fail due to acquired resistance1, 2. Numerous studies have identified activation of alternative signaling pathways as possible resistance mechanisms (e.g., ref. 3), suggesting that combination therapies directed against multiple pathways would be beneficial. As an alternative strategy, adaptive therapy (AT) is proposed to be advantageous in such settings, and more effective at controlling resistance than conventional maximal tolerated dose (MTD) approaches4,5,6,7,8. In AT, therapeutics are used at low-dose, adjusted to maintain tumour burden constant rather than eradicating all tumour cells. This in theory preserves therapy-sensitive cells that will outcompete resistant cells, due to the reduced proliferative fitness of the latter. This assumption has not been validated. Furthermore, whereas previous mathematical modelling7 indicated that AT should confer a large survival benefit, this model assumed that the relative fitness of resistant cells is proportional to their frequency in the population. As such, the relative fitness of rare resistant cells would approach zero, which is unlikely. Crucially, experimental investigations of AT did not monitor resistance frequency nor measure cell fitness. In mouse xenograft models using cytotoxic chemotherapy, combining one MTD dose followed by lower doses resulted in better long-term tumour control than the MTD treatment alone4, 6. Although this result might indeed reflect reduced selection for resistance, alternatively, it may have been due to the higher cumulative drug dose applied. The principles underlying AT thus remain unproven. To test the assumptions of AT, we developed a new mathematical model of the population dynamics of therapy-sensitive and resistant cells, and an experimental system allowing us to test its predictions. We hypothesised that resistance to inhibitors of cell cycle regulators would likely incur a fitness cost, potentially fulfilling the assumptions of AT and allowing us to test which parameters are critical. We focused on cyclin-dependent kinases (CDKs), which control the cell cycle and whose pathways are universally deregulated in cancer9. Small molecule CDK inhibitors (CDKi) have been developed as agents for cancer therapy. Early clinical trials with non-specific CDKi showed promising responses but were hindered by toxicity10. In 2015, palbociclib (PD0332991), which targets CDK4 and CDK6, was approved for use in cancer therapy11, 12. However, not all cancer cells respond to CDK4/6 inhibition, and loss of RB1 renders cells insensitive13,14,15,16. Yet probably all cancer cells have active CDK1 and CDK2. CDK1 is essential for cell proliferation17, 18, whereas CDK2 knockout mice are viable19, 20 and CDK2 knockdown is tolerated by most cancer cells21. Nevertheless, acute pharmacological or peptide-based inhibition of CDK2 strongly inhibits cancer cell proliferation22,23,24,25, CDK2 counteracts Myc-induced cellular senescence26 and CDK2-knockout mouse cells are resistant to oncogenic transformation19. Thus, CDK1 or CDK2 inhibition will likely have therapeutic benefits. We predicted that resistance to CDK1/CDK2 inhibitors might arise through alteration of cell cycle pathways, reducing proliferative fitness. We therefore generate colorectal cancer cells with acquired resistance to a CDK1/CDK2-selective inhibitor, and identify mechanisms of resistance. These involve stable rewiring of cell cycle pathways, resulting in compromised cellular fitness. Based on competition experiments with different treatment regimes and computer simulations, we find that tumour spatial structure is a critical parameter for AT. Competition for space increases fitness differentials, allowing effective suppression of resistant populations with low-dose treatments. ## Results ### Mathematical modelling of tumour evolution under AT To investigate the hypothesis that AT might control tumour growth more effectively than MTD, we first developed a new minimally complex mathematical model of tumour evolutionary dynamics during therapy to capture the fundamental dynamics of AT and MTD. Previous mathematical modelling7 indicated that AT could confer very large survival benefit, that strongly depended on the fraction of resistant cells in the population (frequency) when treatment begins. However, relative fitness of resistant cells was assumed to be proportional to their frequency (Fig. 1a, solid line), a probable oversimplification of dynamics in situ. The premise underlying AT is that, on average, resistant cells proliferate more slowly when surrounded by sensitive cells than other resistant cells. Yet competition for diffusion-limited resources is generally confined to relatively small neighbourhoods, and a change in frequency below or above certain thresholds should not much affect resistant cell fitness. From these considerations and a geometrical analysis of resistant subclone growth within a three-dimensional tumour (Supplementary Methods; Supplementary Fig. 1), we propose that the relationship between the relative fitness and the frequency of resistant cells can be more realistically represented by a sigmoidal function, with its lower asymptote greater than zero (Fig. 1a, dashed lines). The two asymptotes correspond to the relative fitness of resistant cells when they are either (i) surrounded by drug-sensitive tumour cells that constrain their population growth (lower), or (ii) more abundant and have escaped from competition with drug-sensitive cells (upper). We assume that the transition between fitness levels is relatively abrupt as resistant cells that escape competition with sensitive cells will rapidly expand. Yet our model predicts that the relative benefits of AT are insensitive to the exact frequency of the transition (Supplementary Methods; Supplementary Fig. 1). We used coupled differential equations to model population dynamics. We obtained approximate analytical solutions to determine how treatment outcomes depend on biological parameters (Supplementary Methods). To confirm their validity and to facilitate comparisons between studies, we also ran numerical simulations for AT and MTD regimes that were examined in previous analysis7. In simulations of MTD, therapy was applied as a constant-dose bolus at regular intervals. For AT, we began with half the MTD dose and adjusted it by 20% if the total population size had increased or decreased since the previous treatment. The AT dose was not allowed to exceed that of MTD. The cell number at treatment onset was set at 109, approximating the cell population in a typical human tumour at first detection (1 cm3). We compared the predicted survival time (defined as the time taken for the tumour cell population to reach 1012), and progression-free survival time (defined as the time taken for the tumour to regain its pre-therapy size) between AT and MTD regimes. With the previously-described linear fitness function7, AT can stall tumour growth indefinitely, provided resistance is sufficiently rare at the start of treatment and resistant cells maintain some sensitivity (Fig. 1b; Fig. 1c, black solid line). Even if some cells are 100% resistant, the benefit of AT, relative to MTD, is unbounded and increases rapidly with decreasing initial resistance frequency (Fig. 1c, black dashed line). Conversely, with our more plausible sigmoidal fitness function, the benefit of AT is predicted to be more modest (Fig. 1b). The relative benefit of AT for progression-free survival (Fig. 1c) or overall survival (Supplementary Fig. 2a), is limited by an upper bound that is independent of the initial resistance frequency but varies with the maximum growth rate of resistant cells (Fig. 1d) and the degree of resistance (Supplementary Fig. 2b). For realistic parameter values, the upper bound can be approximated as $$\frac{1}{{f_{{\mathrm{min}}}}} - \frac{{\lambda _W}}{{\mu _W}}\left( {\frac{1}{{f_{{\mathrm{min}}}}} - 1} \right),$$ (1) where f min is the relative fitness of resistant cells when they are rare; λ W is the division rate of sensitive cells without treatment; μ W is the death rate of sensitive cells under MTD (Supplementary Methods). The maximum benefit of AT is thus strongly dependent on f min, and increases as f min decreases. The same pattern holds for less frequent dosing (Supplementary Fig. 2c) and for sigmoidal rather than exponential growth curves (Supplementary Fig. 2d). In summary, this model predicts that AT will only have limited survival benefit over MTD unless the fitness differential between therapy-sensitive and resistant cells is large (at least a factor of two). ### Cell cycle rewiring in CDKi-resistant cells To test how the results of mathematical modeling compare with experimental data, we first generated CDK1/CDK2 inhibitor-resistant (ir) cells and analysed their fitness. We chose the CDKi NU6102 since we found that it is CDK2-selective at low doses and we could generate NU6102-resistant alleles of CDK2 by engineering combinatorial mutations in the kinase domain27. We confirmed that NU6102 had similar growth-inhibitory effects in several cancer cell lines of different origins and in non-transformed fibroblasts (Supplementary Fig. 3a). In all cell lines, growth was arrested at 20 μM, whereas 50 μM caused major cell death. We chose the colorectal cancer cell line HCT-116 which can rapidly evolve resistance to cell cycle kinase inhibitors28. We could also compare CDK inhibition by NU6102 with CDK2 gene deletion (CDK2−/−)29. We found that CDK2−/− cells were less sensitive to NU6102 than wild-type (WT) cells (Supplementary Fig. 3b), confirming that loss of CDK2 confers partial resistance to a CDK1/CDK2-selective inhibitor. We stably expressed eGFP in the parental HCT-116 cells, to distinguish them from resistant cells in mixed cultures. To obtain NU6102-resistant cells, we applied either escalating concentrations of NU6102 (up to 10, 20 or 50 μM), or maintained the same concentrations from the start. We obtained NU6102-resistant colonies, which we designated R10, R20 or R50 (e.g., R50 is resistant to 50 μM NU6102), but no line was totally impervious to the inhibitor (Supplementary Fig. 3c). Without NU6102, populations resistant to higher concentrations grew more slowly, indicating compromised fitness (Supplementary Fig. 3c). We next profiled the expression of cell cycle regulators in control, CDK2−/− and resistant cells (Supplementary Fig. 3d). In resistant and control cells, CDK2, CDK1, cyclin A2, cyclin E1, cyclin B1, cyclin D1, p21 and the CDK2 substrates RB and CDC6 were expressed at similar levels, while CDK6 was slightly increased in R20 cells. Cyclin A2, cyclin B1, CDK6, CDC6 and RB were strongly reduced in R50 cells, in accordance with their poor proliferation in the presence of inhibitor. However, after serial passaging in 50 μM NU6102, R50 cell proliferation became totally refractory to the inhibitor (Fig. 2a, compare R50-early and -late). Thus, resistance can evolve. Without inhibitor, R50-early and R50-late cells both proliferated less well than control cells, indicating that resistance indeed incurred a fitness penalty, in agreement with our original prediction. We tested the fraction of cells synthesising their DNA by analysing 5-ethynyl deoxyuridine (EdU) incorporation after either pulse-labelling or 24-hour exposure. This revealed that all R50 and WT cells replicated within a 24-h period, but 20% fewer R50 cells were in S-phase at a given time, indicating an altered cell cycle distribution (Supplementary Fig. 3e). R50 did not have increased apoptosis, as determined by 7-aminoactinomycin D staining and western blotting for cleaved caspase-3 (Supplementary Fig. 3f). To test whether resistance was reversible, we withdrew the inhibitor for 2 or 6 months (“drug holidays”). Re-exposure to NU6102 did not affect the cell growth, demonstrating that resistance was irreversible (Supplementary Fig. 3g). We next tested the sensitivity of R50 to other CDKi, and found that they were more sensitive than controls to another tri-substituted purine, purvalanol A30 (Supplementary Fig. 4a), that also strongly inhibits CDK1 and CDK227. This suggests that cells were not generally drug-resistant and that a specific alteration in CDK pathways contributed to NU6102 resistance. We cloned and sequenced the CDK2 gene from R50 cells and found no mutations. We then measured CDK2 kinase activity by assaying phosphorylation of histone H1 on immunoprecipitates of CDK2 or cyclin A from R50 cells, CDK2−/− cells, WT cells, and WT cells treated with 20 μM NU6102 (Fig. 2b). As expected, only background CDK2 activity was detectable in CDK2−/− cells, though cyclin A-associated activity (which includes CDK1) was less affected. WT cells treated with 20 μM NU6102 had higher maximal CDK2 activity than control cells, an expected result of their arrest in G2/M. Despite having similar levels of cyclin A-CDK2 complexes to WT (Supplementary Fig. 4b), CDK2 activity from R50 cells was reduced (Fig. 2b), perhaps contributing to their NU6102 resistance. To investigate CDK2 activity directly in cells, we used a recently-developed sensor, DHB-Venus31. This probe translocates from the nucleus to the cytoplasm upon phosphorylation by CDK2. However, the probe behaved similarly in CDK2−/−, R50 and WT cells (Supplementary Fig. 4c), indicating that another kinase can substitute for CDK2 in G1/S. Altogether, these results suggest that while CDK2 activity was reduced in R50, overall CDK activity was comparable, implying rewiring of CDK pathways. Upregulating Ras signaling pathways and cyclin E expression is involved in resistance to CDKi in HEY ovarian cancer cells16. Cyclin E expression was normal in R50 cells (Supplementary Fig. 3d). We probed the activity of major cell signaling pathways by protein phosphorylation array analysis, but found no major alterations between WT cells, WT cells treated with 20 μM NU6102, R50 cells or CDK2−/− cells (Supplementary Fig. 4d). This further highlights the CDK-specificity of NU6102 and is consistent with specific resistance mechanisms. We next investigated gene expression genome-wide by microarray analysis. R50 cells clustered with CDK2−/− cells, and the effects of loss of CDK2 gene or activity (NU6102 resistance) on the transcriptome were distinct from, and generally opposite to, the effects of CDK inhibition (Fig. 2c and Supplementary Data 1). Gene ontology analysis showed that the most altered pathways in R50 cells compared to WT cells ± NU6102 included pathways in cancer, the cell cycle, and RNA transport (Supplementary Fig. 4e). We confirmed by qRT-PCR that CDK6 was highly upregulated in R50 cells (but not CDK2−/− cells) (Fig. 2d). While CDK4 levels were slightly reduced, both CDK6 protein and kinase activity were strongly increased in proliferating R50 cells (Fig. 2e, f). As determined by western blotting immunoprecipitated cyclin D1 and cyclin D3 from R50 cells, CDK6 showed no cyclin D specificity, whereas CDK4 preferentially complexed with cyclin D1 (Supplementary Fig. 4f). These results suggest that upregulation of CDK6 is involved in resistance to CDK2 inhibition. Determination of CDK2, CDK4 and CDK6 enzyme kinetic parameters in vitro (Methods and Supplementary Fig. 4g) showed that CDK6 affinities for ATP and NU6102 were intermediate between those of CDK2 and CDK4. The K m(ATP)/K i(NU6102) ratio, which reflects inhibitor sensitivity, was 2-3-fold lower for CDK6-cyclin D1 than for CDK2-cyclin A2. R50 and CDK2−/− cells were more sensitive than control cells to the CDK4/CDK6 inhibitor PD0332991 (Fig. 2g). CDK6 knockdown by siRNA reduced cell proliferation similarly in both WT and R50 cells, as reflected by a decrease in cyclin A and phospho-RB levels (Fig. 2h). However, in R50, the CDK6 level remaining after siRNA was comparable to that of untreated WT cells. Thus, resistance to a CDK2-selective inhibitor led to increased dependency on CDK6, suggesting that CDK2 and CDK6 can perform similar cellular functions in HCT-116 cells. Next, we tested the physiological relevance for tumourigenesis of this rewiring of CDK pathways. In vivo, tumour formation will depend on sensitivity to limiting oxygen and nutrients. We therefore compared growth curves of GFP + and R50 under conditions of low nutrients or hypoxia. R50 cells were markedly sensitive to culture in 1% oxygen (Fig. 3a), and low serum, but were less sensitive to low glucose (Fig. 3b), indicating possible effects of CDK rewiring on metabolism. We next subcutaneously injected parental (WT), R50 and CDK2−/− cells in nude mice. All three cell lines could form tumours in mice, but R50 grew markedly more slowly than WT and CDK2−/− tumours (Fig. 3c). Lower mitotic index indicated impaired cell proliferation (Fig. 3d). Thus, whereas CDK2 gene deletion did not affect tumour cell growth, R50 cells have lower proliferative fitness both in vitro and in vivo. These results confirmed our initial hypothesis and suggested that, when mixed together in the absence of inhibitor, resistant cells should be outcompeted by sensitive cells. We adjusted results to compensate for loss of eGFP expression (as quantified in control experiments). When seeded at a 1:1 ratio, eGFP-expressing cells were outcompeted by eGFP-negative control cells, indicating impairment of growth due to eGFP expression (Supplementary Fig. 5a). We then competed R50 cells with eGFP-expressing control cells in equal initial proportions, in the absence of inhibitor. Although R50 cells initially grew faster, they were later outcompeted by control cells (Supplementary Fig. 5a). This indicates that resistant cells are significantly less fit than control cells in long-term co-culture. Their initial higher competitiveness suggested a possible addiction to inhibitor, so to test this we competed eGFP control cells with R50 cells that had been grown without inhibitor for 2 or 6 months (“drug holiday”). Indeed, R50 drug-holiday cells were even less fit and had no initial growth advantage (Fig. 3e). For each competition we estimated the selection coefficient, which measures the fitness difference between the two cell lines and corresponds to the difference between the growth rates of the two cell lines grown separately. R50 cells growing with eGFP cells had a selection coefficient similar to that predicted from the difference between the monoculture growth rates (Supplementary Fig. 5a). In contrast, the selection coefficient of drug-holiday R50 cells was substantially lower than expected (Fig. 3e), indicating further adaptation that reduces their competitiveness, i.e., drug addiction. To offset the fitness differential caused by expression of eGFP and to allow direct quantitation of the resistant and sensitive populations, we generated R50 cells stably expressing mCherry. We grew each cell line either alone or in mixed cultures at different initial ratios, and quantified populations over a one-month period by flow cytometry. R50-mCherry cells (R50-mCh) were outcompeted by WT-eGFP cells, and the selection coefficient was approximately equal to the difference in monoculture growth rates (Fig. 3f, Supplementary Fig. 5b). ### Effective AT requires spatially structured growth Our mathematical model predicts that AT performs substantially better than MTD only if resistant cells are relatively much less fit (>2-fold) than sensitive cells when resistance is rare. Our CDKi R50 cells had a relative fitness of 90–95%, regardless of their frequency, suggesting that AT would not be advantageous over MTD in slowing the growth of resistant cells in monolayer culture. To test this prediction, we first mimicked AT and MTD therapy by treating mixed monolayer cultures of R50-mCh and control GFP+ cells (seeded at 1:99 initial ratio) with either 50 μM NU6102 to kill sensitive cells (MTD, with or without [MTDx] a 2-day break between drug treatments), or an initial concentration of 15 μM NU6102, i.e., just below the concentration required to maintain a stable population (AT). We modified the concentration of NU6102 in AT by ± 20% at 3-day intervals to maintain a constant cell density. We determined the proportions of R50-mCh and GFP + at different time points by flow cytometry. As expected, MTD simply eliminated GFP+ cells while R50-mCh cells grew freely (Fig. 4a; Supplementary Fig. 6a). Although AT arrested growth of GFP+ cells (while not proliferating, they increased in size, reminiscent of senescent cells; Fig. 4b; Supplementary Fig. 6b), it had negligible effect on R50-mCh cell growth. We tested different concentrations of NU6102 for AT treatment, but under no condition did GFP+ cells strongly compete with, and hinder, growth of resistant cells, thus confirming the predictions of our mathematical model. In the above experiments, cells were grown in monolayers and it was essential to avoid confluency to allow continued measurement of cell growth and adjustment of AT doses. Competition for space was therefore minimal, and the relative lower fitness of resistant cells and their dependence on higher oxygen and growth factors might not have been fully exploited. We reasoned that fitness differentials might be further accentuated by spatial structure. To test this hypothesis, we created a computational model (Supplementary Methods, Supplementary Fig. 7a, b). We devised a more sophisticated model to investigate how fitness differentials can arise from competition for space and oxygen in tumour spheroids, a system that recapitulates the spatial cellular interactions and the resource gradients found in solid tumours. The model comprises subpopulations of sensitive and resistant cells that divide and die stochastically at rates dependent on local concentrations of oxygen and CDKi, which diffuse from the surrounding medium. To account for crowding effects, cells are able to divide only if there is sufficient nearby space. In the absence of CDKi, these factors decrease cellular fitness along a gradient from the tumour spheroid periphery toward the core. Based on our results, we assumed that, without CDKi, the proliferation rate of resistant cells is 90% of that of sensitive cells. CDKi effects on cell proliferation and death rates were also derived from our experimental data, whereas parameter values related to oxygen were obtained from the literature (Supplementary Methods; Supplementary Fig. 7c). To test the general validity of the model, we disregarded the higher sensitivity of R50 to hypoxia. The code used for tumour spheroid simulations is available under a permissive free software license32. We started the model with 1% resistant cells randomly distributed within the tumour spheroid and allowed an initial competition period before introducing CDKi, during which the spheroid increased in volume from 0.07 mm3 to 0.3 mm3. During treatment for 24 days at 20 μM CDKi, the median increase of the tumour spheroid volume was 38%, whereas at 50 μM it was 930% (Fig. 5a, b; Supplementary Movies 15). The median frequency of resistance was 0.8% after 24 days at 20 μM but reached 100% over the same period at 50 μM CDKi (Fig. 5c). At 20 μM CDKi, resistant cells were much more likely to increase in frequency if they were initially located close to the spheroid periphery (Fig. 5d), where abundant space and oxygen confer high relative fitness. To further investigate the effects of spatial structure, we varied the model’s parameter values and the time of treatment initiation. Removing the fitness cost of resistance had a relatively small effect, compared to the effect of the initial location of resistant cells (Fig. 5e). In contrast, removing the competition during the initial growth period up to the spheroid volume of 0.3 mm3 when the treatment was started, and with a random spatial distribution of resistant cells, resulted in much faster population growth (median 950% increase in volume over 24 days) and higher final resistance frequency (median 74% after 24 days) (Fig. 5f). This is because resistant cells located close to the periphery have higher relative fitness. Finally, when we removed crowding inhibition of proliferation, so that even cells far from the periphery were able to proliferate (at rates dependent on local oxygen and CDKi concentrations), tumour spheroids grew faster at the 20 μM CDKi dose than at 50 μM (Supplementary Fig. 7d), and the frequency of resistance after 24 days typically reached 100% (Fig. 5g). Thus, in summary, lower doses better control tumour growth and resistance than high doses, but only when resistant cells are spatially constrained by sensitive cells. To test the model experimentally, we set up competition experiments in 3D by growing tumour spheroids. We mixed R50-mCh and control GFP + cells at a 1:99 ratio and maintained them in various concentrations of NU6102. Without CDKi, tumour spheroids grew exponentially to a diameter of around 400 μm, and R50-mCh cells were undetectable, confirming their reduced fitness (Fig. 6a, b; Supplementary Fig. 8). Since we could not wash out the inhibitor without disrupting or losing spheroids, we mimicked AT and MTD using constant low, intermediate or high concentrations of the drug. There was good qualitative agreement between the results of the computational model and those of the experimental system (Figs. 5a, b, and 6a, c; Supplementary Movies 15), except that the simulated tumour spheroids were able to grow beyond the maximum size attainable in vitro. At 50 μM NU6102 (MTD), growth of spheroids was initially arrested, but then, as expected, spheroids rapidly grew back and became entirely composed of R50-mCh cells (Fig. 6a–c; Supplementary Fig. 8). Importantly, at lower NU6102 concentrations, overall tumour growth was slightly (10 μM) or strongly (20 μM) inhibited, and spheroids were overwhelmingly composed of GFP + cells (Fig. 6c; Supplementary Fig. 8). Thus, low concentrations of the drug were more effective than high concentrations in restraining both tumour growth and proliferation of resistant cells. This result provides strong support for the AT hypothesis as it shows that by reducing the selective pressure imposed by therapy to allow tumour maintenance rather than eradication, fitness differentials can be exploited to limit emergence of resistance. That this result could be achieved only in tumour spheroids, and not in monolayer culture, supports the conclusion drawn from our computational modelling: spatial structure that reduces the average relative fitness of resistant cells is a critical factor for effective AT. ## Discussion Conventional cytotoxic chemotherapies generate unsustainable DNA damage in highly proliferative tissues, killing cancer cells that are sensitised to apoptosis. These regimes are usually based on the MTD, which by definition engenders severe side effects, necessitating treatment-free recovery periods. The discontinuous nature of such treatments may well be a factor in disease relapse and resistance, which is almost invariably encountered in the clinic. Furthermore, the DNA damage resulting from disrupted S-phase or mitosis might contribute indirectly to resistance mechanisms by increasing mutation rates. Targeted therapies are more specific, and include inhibitors of protein kinases that are deregulated in certain cancers. Kinase inhibitors are increasingly used in cancer treatment, but also consistently confront resistance2. To limit development of resistance, which emerges as a result of selective pressure of treatment, alternative approaches, such as metronomic chemotherapy and adaptive therapy, have been proposed. In their seminal description of the AT hypothesis, Gatenby et al. stated4, “The goal of adaptive therapy is to enforce a stable tumour burden by permitting a significant population of chemosensitive cells to survive so that they, in turn, suppress proliferation of the less fit but chemoresistant subpopulations.” The questions we set out to answer were: (1) Does resistance to a candidate AT drug truly incur a cost? (2) Can sensitive cells indeed suppress proliferation of resistant cells? (3) Which parameters are decisive for AT to be effective? Protocols resembling AT, using human breast and ovarian cancer cell lines, were previously tested in immunodeficient mice using conventional chemotherapeutics4, 6, and resulted in long-term stabilisation of tumour burden. However, over the course of the experiment, more of the drug was given in “AT” than in the “MTD” arm, and the cumulative dose will likely directly affect the outcome. Furthermore, the treatments used might further have indirect impact through inhibition of proliferation of endothelial cells responsible for angiogenesis, as previously seen in metronomic chemotherapy33, 34. MTD-based therapies may work less well in this regard since they cannot be tolerated indefinitely, unlike the lower doses of metronomic or adaptive regimes. In any case, whether the benefit of AT indeed depends on increased overall dose directly affecting tumour cells, competition between therapy-sensitive and therapy-resistant cells5, or, like in metronomic chemotherapy, effects on angiogenesis, is not known. To maximize the chances of resistance reducing cell fitness, we chose CDK inhibitors, an emerging class of cancer drugs. We characterised resistant cells in terms of both biochemical mechanisms and fitness in vitro and in vivo. Our results provide strong support for the adaptive therapy hypothesis. We show that tumour cell-intrinsic resistance mechanisms can reduce fitness, and this difference in fitness is amplified by spatially structured tumour growth to a point where lower drug doses are better than higher doses at controlling tumour burden and resistance. This fitness penalty was additionally context-dependent as resistant cells were sensitised to hypoxia and low serum. Nutrients and oxygen become limiting in the interior of tumour spheroids and even more so in vivo. Context-dependent fitness penalties might be generally true for resistance to targeted therapies, as a recent study found that resistance to the EGFR kinase inhibitor erlotinib in lung cancer cells incurred a fitness penalty that varied according to nutrient and oxygen availability35. However, modelling shows that even without any context specificity, the combination of spatial confinement and the inherent fitness penalty of resistance allows effective competition between sensitive and resistant populations. In this context, AT is advantageous. That our results were obtained in vitro (and reproduced in silico) is important since further confounding issues of the tumour environment are avoided. We demonstrated that the reduced fitness of resistant cells holds both in cell culture and in mouse xenografts. As vascularised tumours are much larger than tumour spheroids, have more complex microenvironmental heterogeneity, and much more limited substrate, we would expect the fitness cost of resistance to be amplified in vivo, thus favouring AT. Further investigation is needed to test this prediction. Clinical outcomes will also depend on the pharmacodynamics of individual drugs, cancer cell type, microenvironment, and mechanisms of resistance36. The CDKi that we used is the most specific inihibitor available for CDK2 but does not have sufficient pharmacodynamics to be tested in mice37. In tumours, the local microenvironment might contribute to drug resistance, e.g., due to insufficient drug perfusion. Drug gradients in large tumours may be steeper than in our spheroid models, and prolonged exposure to low drug concentrations may facilitate the evolution of intrinsic resistance. Yet the expansion of intrinsically resistant cell populations will always be subject to competition with sensitive cells. Thus, our results are not unduly affected by the presence of environmentally-mediated resistance. Whereas treatment schedules have been compared in previous computational models of AT4, 7, 38, 39 and in models incorporating microenvironmental feedback (reviewed in40), the relationship between AT and metronomic therapy has received less attention. We thus integrated our general mathematical model of adaptive therapy with an experimentally validated metronomic model accounting for interplay between the tumour and its vascular support41 (Supplementary Methods). Interesting dynamics arise from this extended model, indicating that AT may be more effective when administered in frequent, low doses than at longer intervals and higher doses (Supplementary Fig. 9a). This is because the tumour’s vascular support recovers, promoting the growth of resistant cells during the breaks in treatment (Supplementary Fig. 9a). When the treatment’s antiangiogenic effect is large and the dose frequency is elevated, metronomic therapy with high drug doses compares favourably with AT, but in all other circumstances AT performs best (Supplementary Fig. 9b). Further research is needed to characterise the effects of additional factors, such as immune responses and modulation of resistant cell fitness by environment. For example, whereas our CDKi-resistant cells were more sensitive to hypoxia, this was not the case in erlotinib resistance35. Generally, however, AT and metronomic therapy probably both exploit competition between therapy-sensitive and resistant cells as well as effects on the microenvironment. Our results support the argument that targeting non-essential CDKs that control the cell cycle might be a useful approach for cancer therapy. While specific inhibition of a single CDK is unlikely to be a realistic aim, we have shown that inhibitors with selectivity for CDK2 can effectively limit tumour cell proliferation. We further demonstrated that upregulated CDK6 can compensate for compromised CDK2 functions and the two kinases have similar kinetic parameters. That resistance of cells was maintained for a six-month drug holiday suggests that it was stably encoded in epigenetic modifications. Resistant cells were sensitised to the CDK4/CDK6 inhibitor palbociclib, which is currently approved for therapy of certain breast cancers and, like other CDK4/CDK6 inhibitors, is undergoing clinical trials for other types of cancer42. We suggest that, reciprocally, upregulated CDK2 might contribute to palbociclib resistance, a scenario already discovered in acute myeloid leukemia cells with a mutated Flt3 receptor tyrosine kinase43. Since CDK1/2 inhibitor-resistant cells are sensitised to CDK4/6 inhibitors, combining both inhibitors could be an advantaegous strategy, exploiting a double-bind whereby cells might be unable to evolve resistance to both inhibitors of CDK1/CDK2 and CDK4/CDK6 without drastic reductions in fitness. This need not involve sequential administration of two drugs, as a recent study44 determined that simultaneous treatment is more effective provided there is no cross-resistance to both drugs. Collateral sensitivity of resistant cells to alternative drugs has recently been validated in experimental models of acute lymphoblastic leukemia and shown by modelling to exploit evolutionary trajectories, much like AT45. CDKi-resistant cells were also sensitive to hypoxia and low serum, suggesting other collateral sensitivities that could be exploited by an additional double-bind38, potentially aiming for cure rather than long-term tumour maintenance. While empirical therapeutic approaches avoid making untested assumptions and will continue to be the mainstay of cancer therapy for the immediate future, mathematical modelling of evolutionary trajectories will take on increasing importance46. ## Methods ### Cell lines The parental and CDK2−/− HCT116 (human colon cancer) cell lines, were purchased from Horizon, UK (HD R02-015). Other cell lines used were: U2OS (human osteosarcoma; purchased from ATCC); BJ-hTERT (human foreskin fibroblasts, immortalized with hTERT; obtained from Dr J. Piette); SK-MEL-28 (human melanoma; gift from Dr Ch. Theillet, IRCM Montpellier). Cells were not authenticated subsequently but were tested for mycoplasma contamination on a weekly basis. All cells were grown in Dulbecco modified Eagle medium (DMEM–high glucose, pyruvate, GlutaMAX–Gibco® LifeTechnologies) supplemented with 10% fetal bovine serum (SIGMA, HyClone or Pan-Biotech). Cells were grown under standard conditions at 37 °C in a humidified incubator containing 5% CO2. Cell lines were not authenticated in-house but were tested on a weekly basis for mycoplasma contamination. ### Establishing resistant cell lines Adaptive-resistant RA10 and RA20 cell lines were obtained by treating HCT116 cells with NU6102 at the initial 2 μM concentration, increased every three days by 2 μM (2>4>6 μM, etc.) until 10 and 20 μM, respectively. R10, R20 and R50 cell lines were grown from the beginning in 10 μM, 20 μM and 50 μM NU6102. Cells were passaged every three days (1/10 dilution) with adding fresh inhibitor. ### Establishing fluorescent cell lines HCT116 cells were plated at 1.5 × 104/cm2 density and transfected 24 h later with eGFP-N1 (1 μg/ml) or pmCherry-N1 (1 μg/ml) vectors, using JetPEI (Polyplus) or lipofectamine 3000 (Invitrogen), respectively, according to the manufacturer’s protocol. eGFP-transfected cells were selected with 1 mg/ml G418 for 10 days. GFP and mCherry-expressing cells were sorted with FACS Aria (BD Biosciences, SanJose, CA). Cell sorting was repeated every 6 months due to the loss of GFP expression. ### Drug treatments CDK inhibitors were dissolved in DMSO and used at concentrations indicated in figure legends: NU6102 (1–50 μM; Enzo Life Sciences); PD03320991 (0.5–10 μM; Selleckchem); Purvalanol A (0.5–5 μM; Enzo Life Sciences). For analyzing the effects of CDK inhibitors on cell proliferation, cells were plated at density of 150000 or 250000 cells per well in 6-well plates. 6 h later, the medium was replaced with medium containing appropriate concentration of inhibitor. For AT and MTD/MTDx treatments, mixed cultures of R50-mCh (1%) and WT-GFP + cells were plated at 200000 in 60mm dishes. Cells were treated with either 50 μM NU6102 (MTD: 1 day treatment, 2 days without drug; MTDx: continuous drug treatment), or an initial concentration of 15 μM NU6102 for AT, i.e. just below the concentration required to maintain a stable population. The concentration of NU6102 in AT arm was subsequently modified by ± 20% at 3-day intervals to maintain a constant cell density (70–80%). The proportions of R50-mCh and WT-GFP + were determined by flow cytometry at different time points. ### Cell proliferation assays Cells were counted using Muse Cell Analyzer and Muse Cell counting reagent (Millipore) according to manufacturer instructions. Briefly, cells were trypsinised, washed in PBS and resuspended in 1 ml of PBS. Pre-warmed counting reagent (380 μl) was mixed with 20 μl of cell suspension and incubated at RT for 5 min. processing on Cell Analyzer. ### Cell extracts and Western-blotting Frozen pellets (harvested by trypsinisation, washed with cold PBS) were lysed with lysis buffer (150 mM NaCl, 50 mM Tris pH 7.5, 0.2% Triton, 1 mM EDTA; freshly added: 1 mM DTT, 0.1 mM NaVO4, protease inhibitors cocktail (Roche)) and incubated on ice for 30 min. Samples were centrifuged for 10 min at 13000 rpm and supernatant collected. Protein concentrations were determined by BCA protein assay (Pierce Biotechnology). Samples were boiled for 5 min in Laemmlli buffer. Equivalent amounts of proteins were separated by SDS–PAGE (usually on 12 cm × 14.5 cm; 7.5% or 12.5% gels). The proteins were semi-dry transferred onto Immobilon membranes (Milipore). Secondary antibodies were either goat antibodies to mouse IgG-HRP (DACO) or donkey antibodies to rabbit IgG-HRP (GE Healthcare). The detection system was Western Lightning Plus-ECL (PerkinElmer) and Amersham Hyperfilm (GE Healthcare). Primary antibodies used were: pRB (G3-245; BD Pharmingen); Rb phospho-S795 (Abcam, ab47474); cyclin A (6E6; Novocastra); cyclin E1 (clone HE12, Santa Cruz Bio.); cyclin D1 (DSC6; Cell Signaling); cyclin D3 (D-7 and B-10; Santa Cruz Bio.); cyclin B1 (GNS1; Santa Cruz Bio.); CDK2 (D-12 and M-2; Santa Cruz Bio.); CDK1 (clone 17; Santa Cruz Bio.); CDK4 (C-22; Santa Cruz Bio.); CDK6 (C-22 and B-10; Santa Cruz Bio.); Cdc6 (180.2; Santa Cruz Bio.); p21 (C-19; Santa Cruz Bio.); Caspase-3 (Cell Signaling); cleaved Caspase-3 (Asp175; Cell Signaling). Uncropped scans are provided in the Supplementary Figs 1012. ### siRNA transfections The SMARTpool: ON-TARGETplus siRNAs (Cdk6, L-003240-00-0005; non-targeting, D-001810-10) were purchased from GE Dharmacon (Lafayette, CO, USA). Cells were transfected with siRNA at 100 nM by calcium phosphate transfection method. Briefly, cells were plated at 1.5 × 104/cm2 density. 24 h later, medium was changed for medium without antibiotics. Calcium phosphate–DNA coprecipitate was prepared (44 µl H2O, 5 µl 2.5 M CaCl2 and 1 µl 100 μM siRNA). 50 µl CaCl2-siRNA solution was combined with equal volume of 2xHBS buffer (50 mM HEPES, 280 mM NaCl, 1.5 mM Na2HPO4, 10 mM KCl; pH 7,04). Coprecipitates were incubated at room temperature for 1 min, mixed by pipetting, added drop by drop into medium above cells and gently mixed. ### RT-PCR Total cellular RNA (1 µg in total volume of 10 µl), extracted by RNeasy Mini Kit (Qiagen), was mixed with 1 µl of 10 mM dNTPs (2.5 mM of each; LifeTechnologies) and 1 µl of 50 µM random hexaprimers (New England Biolabs). Samples were incubated at 65 °C for 5 min., then immediately transferred on ice, followed by addition of 5 µl of 5xFirst Strand Buffer, 2 µl 100 mM DTT and 1 µl RNasin® Plus RNase Inhibitor (Promega). Samples were incubated at 25 °C for 10 min. and at 42 °C for 2 min. 1 µl of M-MLV reverse transcriptase (Thermo Fisher Scientific ref. 28025-013) was added to each sample, and incubated at 42 °C for 60 min., then at 70 °C for 15 min. ### qPCR qPCR was performed using LightCycler 480 SYBR Green I Master (Roche) and LightCycler 480 qPCR machine. The reaction contained 5 ng cDNA, 2 µl 1 μM qPCR primer pair (final concentration of each primer 200 nM), 5 µl 2x Master Mix, and final volume made up to 10 µl with DNase free water. qPCR was conducted at 95 °C for 10 min, followed by 40 cycles of 95 °C for 20 s, 58 °C for 20 s, and 72 °C for 20 s. The specificity of the reaction was verified by melt curve analysis. Each reaction was performed in three replicates. qPCR primers (Tm −60 °C): human CDK6 5′-TCAGCTTCTCCGAGGTCTGG-3′, 5′-TAGGTCTTTGCCTAGTTCATCG-3′. ### Flow cytometry - cell cycle analysis Cells were harvested, washed once with cold PBS, resuspended in 300 µl cold PBS and fixed with 700 µL chilled 100% ethanol. Cells were kept at −20 °C, at least overnight. On the day of analysis, cells were pelleted by centrifugation at 5000 r.p.m. for 5 min. After washing once with 1% BSA in PBS, cells were stained with Propidium Iodide (PI) solution (10 µg/ml PI, 1% BSA, 200 µg/ml RNase A in PBS) for 30 min. at room temperature, and analysed with BD FACS Calibur (BD Biosciences, SanJose, CA). ### EdU/BrdU incorporation To analyse the fraction of replicating cells, cells were either pulse-labeled (15 min.) or incubated for 24 h with either bromodeoxyuridine (BrdU, 200 μM; Sigma Aldrich) or 5-ethynyl-2′-deoxyuridine (EdU, 20 µM; LifeTechnologies). Cells were harvested, washed once with cold PBS, resuspended in 300 µl cold PBS and fixed with 700 µl ice-cold 100% ethanol. EdU incorporation was detected using the Click-iT® EdU Alexa Fluor® 488 Imaging Kit (LifeTechnologies), and analysed with BD FACS Calibur. For BrdU detection, cells were washed with cold PBS and permeabilised with 2 N HCl and 0.5% Triton X-100 for 30 min. at room temperature, with occasional vortexing. After adding 5 ml of PBS, cells were pelleted and resuspended in 200 μl anti-BrdU antibody (BD 347580, No.408) at 1:30 in PBS-0.5% Tween 20 and 1% BSA, and incubated 2 h at room temperature. Cells were washed with PBS and incubated with anti-mouse Alexa Fluor 488 (LifeTechnologies) for 2 h at room temperature. After wash in with PBS, cells were resuspended in 500 μl PBS containing 3 μg/ml 7AAD (LifeTechnologies), 200 μg/ml RNAse A (Sigma Aldrich). Samples were incubated for 2 h at room temperature and analysed on FL-1 and FL-3 channel with FACS Calibur. ### Cell competition experiments WT and R50 cells (extensively washed prior to the experiment to eliminate the inhibitor) were plated at the indicated ratios (2 million cells in total) in 10cm-dish, without the inhibitor. Cells were harvested every three days, and 1/10 of the mixed cell population was plated again. After harvesting, 1 million cells were washed once with cold PBS, resuspended in 1 ml of PI solution (1% BSA in PBS, 10 µg/ml PI) and analysed on Fortessa flow cytometer (BD Biosciences, SanJose, CA) for the percentage of GFP/mCherry positive cells. ### Microarray analysis–transcriptome RNA was prepared from HCT116 WT, CDK2 KO, R50 and WT cells treated with 20 µM NU6102 for 24 h, in duplicates, using RNeasy Mini Kit (Qiagen) following the manufacturer’s instructions. RNA was labelled with Cyanin 3 and complementary RNA (cRNA) was synthesized. Cy3-labelled cRNA was amplified and hybridized on the Agilent SurePrint G3 Human GE 8 × 60k Microarray according to the procedures by Hybrigenics Company (Paris, France). Raw data were processed using GeneSpring GX software (Agilent Technologies) to define differently expressed genes, using one-way ANOVA, with a Benjamini-Hochberg corrected p-value < 0.001 and post hoc Student Newman Keuls. ### Human phospho-kinase antibody array Human Phospho-Kinase Antibody Array (R&D Systems) is a set of nitrocellulose membranes on which capture and control antibodies for 43 kinases and 2 total proteins have been spotted in duplicates. Cell lysates from WT, Cdk2 KO, R50 and WT cells treated with 20 µM of NU6102 for 24 h, were added to array membranes and processed according to the protocol of R&D Systems. Signal from the membranes was imaged with ECL camera and the intensity of the signal quantified with ImageJ software. ### Live-cell CDK2 activity sensor The CDK2 activity sensor was a gift from Sabrina Spencer (Stanford University, CA, USA). The sensor includes 994–1087 amino acids of human DNA helicase B fused to the yellow fluorescent protein mVenus (DHB-Ven) and contains four CDK consensus phosphorylation sites, a nuclear localisation signal and a nuclear export signal31. Sensor was transduced into WT, CDK2 KO and R50 cells by lentiviral infection (see below). To obtain stable cell lines, cells were selected for YFP using cell sorter (FACS Aria). ### Lentiviral infection Viral particles were produced by transfecting packaging cells HEK293 cells with tat, rev, gag/pol, vsv-g vectors (provided by Dr E. Bertrand, IGMM Montpellier) by calcium phosphate transfection (see above). Cells were plated the day before transfection at density of 4 × 106 cells in 10 cm plates. Vectors were transfected in the following proportions: $$\begin{array}{l} 20 \quad \quad \;\;\, \,\quad : \quad 1 \;\;\quad : \quad \;1 \;\,\quad :\quad 1 \quad \quad \;\;\, \quad :\quad \quad 2\hfill \\ {\mathrm{backbone}}\quad :\quad {\mathrm{tat}}\quad :\quad {\mathrm{rev}}\quad :\quad {\mathrm{gag/pol}}\quad : \quad {\mathrm{vsv - g}} \hfill \\ 20\,\mathrm{\mu g} \quad \quad \quad \,\quad {\mathrm{1}}\,{\mathrm{\mu g}} \quad \;\quad 1\,{\mathrm{\mu g}} \quad \quad \,\, \quad 1\,{\mathrm{\mu g}}\quad \quad \quad \quad \;2\,{\mathrm{\mu g}}\quad = 25\,{\mathrm{\mu g}}\,{\mathrm{total}}\,{\mathrm{DNA}}\hfill\end{array}$$ The day after transfection, the supernatant from the virus producing cells was recovered, filtered with 0.45 µm filter and centrifuged in 2 ml eppendorfs at 4 °C for 3 h at maximum speed. The supernatant was collected in 50 ml Falcon tube tightly closed on ice in the cold room. The procedure was repeated the following day. For lentiviral infection, HCT-116 WT cells were plated in a 12-well plate, 5 × 104 cells/well. Cells were rinsed once with fresh medium and supplemented with infection mix (300 μl of medium without serum, 6 μg/ml polybrene). Cells were incubated for 2 h at 37 °C, 5% CO2 with occasional tilting (every 20 min). After 2 h, 1 ml of fresh medium was added and cells were left to recover overnight. The next day medium was changed and culture was expanded. ### Immunoprecipitation Cell lysates were prepared as described above and 100 µl was used for every immunoprecipitation reaction. Each sample was incubated with 3 µl of appropriate antibody on ice for 2 h, followed by incubation with 50 µl of Sepharose beads (Protein A Sepharose or Protein G Sepharose 4 Fast Flow, GE Healthcare (previously Amersham Biosciences)) on Adams Nutator Mixer at 4 °C for 30 min. Supernatants were collected and saved for analysis. Beads were washed three times with 900 µl of lysis buffer, incubated with 30 µl Laemmlli buffer at 37 °C for 15 min., and immunoprecipitated proteins were analysed by Western-blotting. ### In vitro kinase assays Wash buffer I: 25 mM Tris pH 7.5, 150 mM NaCl, 0.1% Triton X-100, 1 mM EDTA, 1 mM EGTA, 1 mM DTT, protease inhibitors. Wash buffer II: 25 mM Tris pH 7.5, 10 mM MgCl2, 1 mM DTT. Kinase buffer 2x: 100 mM HEPES pH 7.5, 20 mM MgCl2, 2 mM DTT, 0.04% Triton X-100. Kinase mix 1x (20 µl): 10 µl kinase mix 2×, 2 µl histone H1 (1 mg/ml) (Calbiochem-Merck Millipore) or Rb-CTF peptide (0.678 mg/ml; ProQinase), 1 µl ATP (1 mM), 0.25 µl γATPP33 (Perkin Elmer), 6.75 µl H2O. ### Kinase assays on immunoprecipitated CDK complexes Immunoprecipitations was performed as described above. Beads were washed two times with 500 µl of Wash buffer I and once with the same volume of Wash buffer II. Beads were incubated with kinase mix at 37 °C for 20 min. with occasional tapping the tube. Kinase mix contained either histone H1 (for cyclin A and CDK2) or Rb-CTF peptide (for CDK6/4) as a substrate. Negative controls were kinase mix without kinase (beads only) and IP beads incubated with kinase mix without substrate. The reaction mix was spotted on P81 phosphocellulose (Millipore) paper and washed three times in 1% orthophosphoric acid (10 ml per sample). Papers were air-dried at room temperature and counted with scintillation counter. ### Evaluation of Km and Ki of recombinant CDK complexes Recombinant CDK/cyclin complexes (CDK2/cyclin A2; CDK4/cyclin D1; CDK4/cyclin D3; CDK6/cylin D1; CDK6/cyclin D3) were obtained from ProQinase and used at 6.8ng/µl. Kinase assays were performed in triplicates, with Rb-CTF peptide as a substrate (final concentration 34ng/µl; ProQinase). For measuring K m, assays were performed with ATP at different concentrations (0, 100 M, 200 µM, 500 µM, 1 mM, 2 mM). For K i, kinase assays included NU6102 at 0, 0.2 µM, 0.5 µM, 1 µM, 3 µM, and 10 µM. Assay time was 8 min. Kinetic parameters were calculated using GraphPad Prism software. ### Sensitivity to glucose depletion and hypoxia 150000 or 250000 of cells per well were plated in 6-well plates. For measuring the response to hypoxia, the plates were placed in the incubator with 1% O2 (37 °C, 5% CO2) and the number of cells was analysed every 24 h for 3 days. For low glucose and low serum sensitivity analysis, the cells were plated in medium with low glucose (1 g/L), or medium with 1% of FBS, respectively, and the number of cells was counted every 24 h for 3 days. ### Tumour xenografts 36 female athymic nude mice (Envigo) of 5 weeks were injected subcutaneously into the right flank with 1.5 × 106 WT, CDK2 KO or R50 cells, in total volume of 150 µl (12 mice per each cell type). Tumour size and animal weight were measured weekly; mice were sacrificed when tumours reached the size of 1500 mm3. Tumours were dissected and samples were frozen for further protein and DNA analysis. Parts of tumours were embedded into paraffin blocks for immunohistochemistry analysis. ### Multicellular tumour spheroids Competitions in 3D were initiated by mixing sensitive (WT) HCT-116 GFP-positive cells with 1% R50-mCherry cells at day 0. 100% WT-GFP and 100% R50-mCherry spheroid cultures were prepared in parallel as controls. Spheroids were initiated in 96-well plates according to Friedrich et al.47. Each well was coated with 50 μl of 1.5% sterile agarose (wt/vol; Sigma, France) in DMEM. Spheroids were initiated by seeding 1500 cells in 200 μl of complete culture medium (DMEM, 10% FBS) per well. After 96 h, spheroids with a mean diameter of about 350 µm were formed. At day 4, treatment was started by exchanging 50% of the 200 µl medium with 100 µl of fresh medium containing NU6102 (2× final concentration). For control conditions, untreated and DMSO-treated spheroids, 100 µl of media were replaced with fresh culture medium or DMSO-containing fresh medium, respectively. Media were changed every three days as described above. At least 4 spheroids were treated and analysed per condition. Volume, integrity and fluorescence composition (GFP/mCherry) of each spheroid were monitored starting at day 4 and, every 3 days up to day 28. Phase contrast, GFP and mCherry images were acquired with 2.5x objective using Zeiss Inverted Axiovert 200 M microscope (Carl Zeiss, Germany). All phase contrast images of spheroids were checked and any deformed or irregular shaped spheroids were eliminated from calculations. Analysis of spheroid volume was performed using ImageJ (v 1.44) with a macro automating size measurements for a folder of phase-contrast spheroid images48. The measured area (S) of spheroids 2D projection was used to calculate the radius (R) and the volume (V) of an equivalent sphere. For flow cytometry, 4–5 spheroids of each condition were collected in 1.5 ml Eppendorf tube and dissociated enzymatically (using 0.05% trypsin, 15 min. treatment), and mechanically (by pipetting up and down three times). The cell suspension was washed with PBS and fixed in 3% paraformaldehyde. Flow cytometry analysis was performed using a Fortessa flow cytometer (BD Biosciences, SanJose, CA) equipped with blue laser (488 nm) and yellow laser (530 nm). Flow cytometry data were analysed with FlowJo v.10.2 software (LLC 2006–2016). Forward and Side scatter of aggregates of cells were determined using log scale SSC/FSC plots. In general, samples were analysed at a medium flow rate and 10000 events were acquired for each sample. ### Statistical analysis Significant differences between experimental groups were determined using an unpaired two-tailed Student t-test in Prism 5 (GraphPad). For all analyses, p-values < 0.05 were considered statistically significant. ### Calculation of selection coefficients Competition between two cell lines can be described using the selection coefficient. If the population sizes are P and Q then the corresponding frequencies are defined as $$p = \frac{P}{{P + Q}},\,q = \frac{Q}{{P + Q}} = 1 - p.$$ The selection coefficient of one cell line relative to another is then defined as the rate of change of the ratio of the frequencies. That is, $$s = \frac{\mathrm{d}}{{\mathrm{d}t}}\log \frac{p}{q} = \frac{\mathrm{d}}{{\mathrm{d}t}}\log \frac{p}{{1 - p}}.$$ (4) If s is positive then P will increase relative to Q; if s is negative then P will decrease relative to Q. If two cell lines in competition grow exponentially and do not interact then the selection coefficient can be predicted from their growth rates49. This is because $$\log \frac{p}{{1 - p}} = \log \frac{P}{Q} = \log \frac{{P\left( 0 \right){\mathrm{exp}}(r_Pt)}}{{Q\left( 0 \right){\mathrm{exp}}(r_Qt)}} = \log \frac{{P(0)}}{{Q(0)}} + \left( {r_P - r_Q} \right)t,$$ where r P and r Q are the growth rates and P(0) and Q(0) are the initial sizes of populations P and Q, respectively. Hence $$s = r_P - r_Q.$$ Accordingly, we predicted the selection coefficient for each competition assay as the difference between the exponential growth rates of the competing cell lines. We then compared this prediction to the selection coefficients calculated from the com- petition assay frequency dynamics. We made a single prediction whenever growth curves were measured at the same time as competitions were conducted; other- wise we made maximum and minimum predictions based on non-contemporaneous growth curves. We estimated the growth rate of each cell type as the mean slope of log-transformed growth curves during the first 72 h in monolayer culture. ### Adjusting for loss of the GFP marker To adjust data for loss of the GFP marker in competitions between GFP + and GFP- subpopulations, we began by normalizing the data (so the two subpopulation sizes summed to unity). We then estimated the rate of loss of the GFP marker by fitting a regression curve to the log-transformed frequency of GFP + cells in the GFP + control assay. We calculated adjustment factors as $$c\left( t \right) = \exp \left( {at + b} \right),$$ where a is the slope of the regression line, b is the intercept, and t is time. We adjusted all GFP+ cell frequencies for loss of the GFP marker by multiplying by 1/c, and all GFP- cell frequencies by multiplying by (1−c)/c. Finally, we renormalized the data. In no case did this adjustment change a qualitative outcome. ### Non-spatial mathematical model Our non-spatial mathematical model of cancer adaptive therapy is inspired by that of Silva et al.7 However, whereas that paper described population dynamics using recurrence relations, we instead use coupled differential equations, which are easier to parameterize and analyze. We begin with growth equations $$\frac{{\mathrm{d}}W}{{\mathrm{d}}t} = \lambda _WW,\,\frac{{\mathrm{d}}R}{{\mathrm{d}}t} = \lambda _Rf\left( {R,W} \right)R,$$ where W and R are the chemosensitive and resistant populations, respectively; λ W and λ R are the maximum growth rates; and f is a frequency-dependent relative fitness function. We also consider a model using a Gompertz growth function, which is the most widely-used function for modelling sigmoidal tumour growth curves50: $$\frac{{\mathrm{d}}W}{{\mathrm{d}}t} = \lambda _WW\frac{{\log (K/N)}}{{\log (K/N_0)}},\,\frac{{\mathrm{d}}R}{{\mathrm{d}}t} = \lambda _Rf\left( {R,W} \right)R\frac{{\log \left( {K/N} \right)}}{{\log \left( {K/N_0} \right)}},$$ where N = W + R is the total population size, N 0 is the initial population size, and K is the carrying capacity. In our numerical simulations (using the R programming language package deSolve51), therapy is applied as a bolus at regular intervals and causes instantaneous cell death. The treatment effect is simulated by multiplying each subpopulation at the time of treatment by $$\frac{1}{{1 + \rho /{\mathrm{IC}}50_X}},$$ where ρ is the dose, and IC50 X is either IC50 W (the half maximal inhibitory concentration for sensitive cells) or IC50 R (the corresponding value for resistant cells). We are interested in the relative benefits of two types of therapy. For maximum tolerated dose (MTD) therapy, every bolus dose is the same. For adaptive therapy (AT) in numerical simulations, the dose is increased (respectively decreased) by 20% if the total population size has increased (respectively decreased) since the previous treatment. Model analysis and further justification for the choice of frequency-dependent fitness function can be found in Supplementary Methods. ### Non-spatial model with microenvironmental feedback To examine how microenvironmental feedback might affect adaptive therapy outcomes, we integrated our frequency-dependent fitness model with an experimentally-validated model of tumour vascularisation developed by Hahnfeldt and colleagues52. As before, we assumed that tumour growth is limited by a carrying capacity according to a Gompertz growth function: $$\frac{{\mathrm{d}}W}{{\mathrm{d}}t} = \lambda _WW\log \frac{K}{N},\,\frac{{\mathrm{d}}R}{{\mathrm{d}}t} = \lambda _Rf\left( {R,W} \right)R\log \frac{K}{N},$$ where N = W + R is the total population size, and K is the carrying capacity. We further assumed that the carrying capacity is linked to the degree of tumour vascularisation, which can change over time due to an interplay of stimulatory and inhibitory factors: $$\frac{{\mathrm{d}}K}{{\mathrm{d}}t} = bN - dN^{\frac{2}{3}}K,$$ where b and d represent how strongly the tumour stimulates and inhibits vascularisation, respectively. The second term in the above equation derives from a mathematical analysis of inhibitory factor diffusion from the surface of a three-dimensional tumour52. Finally, we assumed that treatment not only kills tumour cells but also inhibits vascularisation. Like in the case of cell death, the treatment effect was simulated by multiplying K at the time of treatment by $$\frac{1}{{1 + \rho /{\mathrm{IC}}50_K}},$$ where ρ is the dose, and IC50 K is the half maximal inhibitory concentration. ### Spatial computational model To simulate the tumour spheroid experimental system, we created a so-called hybrid cellular automaton computational model53 (written in the C language) in which each cell inhabits a point on a two-dimensional square grid, which represents a cross-section through a three-dimensional tumour spheroid. Cells proliferate and die at rates that depend on local chemical concentrations and cell density. At the start of each updating loop, all cells that have insufficient oxygen to survive undergo cell death. These dead cells persist (unless replaced by living cells) and form the necrotic core of the tumour spheroid. Next, cells attempt proliferation or (due to effects of the CDK inhibitor) undergo cell death. We use the well-established Gillespie algorithm54 to select cells and event types, and to determine the periods between events. According to this algorithm, the probability that cell k will be chosen for either proliferation or death is $$(P_k + M_k)/\mathop {\sum }\limits_i \left( {P_i + M_i} \right),$$ where P i and M i are proliferation and death rates, respectively. The selection process amounts to sampling with replacement, so the choice of cell is independent of which cells have been selected previously. The chosen cell attempts proliferation with probability $$P_k/(P_k + M_k)$$, or else undergoes cell death. The time between events is calculated by drawing from an exponential distribution with mean $$1/\mathop {\sum }\limits_i \left( {P_i + M_i} \right)$$. For computational efficiency, the diffusion equations are not re-solved after every cell proliferation or death event. Instead, events occur sequentially until the number of cells that have undergone division or death reaches 10% of the population size, at which time the diffusion equations are re-solved and the proliferation and death rates are recalculated for each cell. Further details of the computational model are in Supplementary Methods. ### Code availability Code used for tumour spheroid simulations is available under a permissive free software license.32. ### Data availability Microarray gene expression data is available on the NCBI GEO database with the accession number GSE102165. ## References 1. 1. Gross, S., Rahal, R., Stransky, N., Lengauer, C. & Hoeflich, K. P. Targeting cancer with kinase inhibitors. J. Clin. Invest. 125, 1780–1789 (2015). 2. 2. Wu, P., Nielsen, T. E. & Clausen, M. H. FDA-approved small-molecule kinase inhibitors. Trends Pharmacol. Sci. 36, 422–439 (2015). 3. 3. Wilson, T. R. et al. Widespread potential for growth-factor-driven resistance to anticancer kinase inhibitors. Nature 487, 505–509 (2012). 4. 4. Gatenby, R. A., Silva, A. S., Gillies, R. J. & Frieden, B. R. Adaptive therapy. Cancer Res. 69, 4894–4903 (2009). 5. 5. Gatenby, R. A. A change of strategy in the war on cancer. Nature 459, 508–509 (2009). 6. 6. Enriquez-Navas, P. M. et al. Exploiting evolutionary principles to prolong tumor control in preclinical models of breast cancer. Sci. Transl. Med. 8, 327ra24 (2016). 7. 7. Silva, A. S. et al. Evolutionary approaches to prolong progression-free survival in breast cancer. Cancer Res. 72, 6362–6370 (2012). 8. 8. McGranahan, N. & Swanton, C. Biological and therapeutic impact of intratumor heterogeneity in cancer evolution. Cancer Cell 27, 15–26 (2015). 9. 9. Sherr, C. J. & McCormick, F. The RB and p53 pathways in cancer. Cancer Cell 2, 103–112 (2002). 10. 10. Cicenas, J. & Valius, M. The CDK inhibitors in cancer research and therapy. J. Cancer Res. Clin. Oncol. 137, 1409–1418 (2011). 11. 11. DeMichele, A. et al. CDK 4/6 inhibitor Palbociclib (PD0332991) in Rb + advanced breast cancer: phase II activity, safety, and predictive biomarker assessment. Clin. Cancer Res. Preprint at https://doi.org/10.1158/1078-0432.CCR-14-2258 (2014). 12. 12. Finn, R. S. et al. The cyclin-dependent kinase 4/6 inhibitor palbociclib in combination with letrozole versus letrozole alone as first-line treatment of oestrogen receptor-positive, HER2-negative, advanced breast cancer (PALOMA-1/TRIO-18): a randomised phase 2 study. Lancet Oncol. 16, 25–35 (2015). 13. 13. Konecny, G. E. et al. Expression of p16 and retinoblastoma determines response to CDK4/6 inhibition in ovarian cancer. Clin. Cancer Res 17, 1591–1602 (2011). 14. 14. Dean, J. L. et al. Therapeutic response to CDK4/6 inhibition in breast cancer defined by ex vivo analyses of human tumors. Cell Cycle Georget. Tex. 11, 2756–2761 (2012). 15. 15. Wiedemeyer, W. R. et al. Pattern of retinoblastoma pathway inactivation dictates response to CDK4/6 inhibition in GBM. Proc. Natl Acad. Sci. USA 107, 11501–11506 (2010). 16. 16. Taylor-Harding, B. et al. Cyclin E1 and RTK/RAS signaling drive CDK inhibitor resistance via activation of E2F and ETS. Oncotarget 6, 696–714 (2015). 17. 17. Santamaria, D. et al. Cdk1 is sufficient to drive the mammalian cell cycle. Nature 448, 811–815 (2007). 18. 18. Diril, M. K. et al. Cyclin-dependent kinase 1 (Cdk1) is essential for cell division and suppression of DNA re-replication but not for liver regeneration. Proc. Natl Acad. Sci. USA 109, 3826–3831 (2012). 19. 19. Ortega, S. et al. Cyclin-dependent kinase 2 is essential for meiosis but not for mitotic cell division in mice. Nat. Genet. 35, 25–31 (2003). 20. 20. Berthet, C., Aleem, E., Coppola, V., Tessarollo, L. & Kaldis, P. Cdk2 knockout mice are viable. Curr. Biol. 13, 1775–1785 (2003). 21. 21. Tetsu, O. & McCormick, F. Proliferation of cancer cells despite CDK2 inhibition. Cancer Cell 3, 233–245 (2003). 22. 22. Chen, Y. N. et al. Selective killing of transformed cells by cyclin/cyclin-dependent kinase 2 antagonists. Proc. Natl Acad. Sci. USA 96, 4325–4329 (1999). 23. 23. Horiuchi, D. et al. Chemical-genetic analysis of cyclin dependent kinase 2 function reveals an important role in cellular transformation by multiple oncogenic pathways. Proc. Natl Acad. Sci. USA 109, 1019–1027 (2012). 24. 24. Krasinska, L., Cot, E. & Fisher, D. Selective chemical inhibition as a tool to study Cdk1 and Cdk2 functions in the cell cycle. Cell Cycle 7, 1702–1708 (2008). 25. 25. Merrick, K. A. et al. Switching cdk2 on or off with small molecules to reveal requirements in human cell proliferation. Mol. Cell 42, 624–636 (2011). 26. 26. Campaner, S. et al. Cdk2 suppresses cellular senescence induced by the c-myc oncogene. Nat. Cell Biol. 12, 54–59 (2010). 27. 27. Echalier, A. et al. An integrated chemical biology approach provides insight into Cdk2 functional redundancy and inhibitor sensitivity. Chem. Biol. 19, 1028–1040 (2012). 28. 28. Girdler, F. et al. Molecular basis of drug resistance in aurora kinases. Chem. Biol. 15, 552–562 (2008). 29. 29. Chung, J. H. & Bunz, F. Cdk2 is required for p53-independent G2/M checkpoint control. PLoS Genet. 6, e1000863 (2010). 30. 30. Gray, N. S. et al. Exploiting chemical libraries, structure, and genomics in the search for kinase inhibitors. Science 281, 533–538 (1998). 31. 31. Spencer, S. L. et al. The proliferation-quiescence decision is controlled by a bifurcation in CDK2 activity at mitotic exit. Cell 155, 369–383 (2013). 32. 32. Noble, R. HyCAT.c. Preprint at 10.5281/zenodo.838190 (2017). 33. 33. Browder, T. et al. Antiangiogenic scheduling of chemotherapy improves efficacy against experimental drug-resistant cancer. Cancer Res. 60, 1878–1886 (2000). 34. 34. Klement, G. et al. Differences in therapeutic indexes of combination metronomic chemotherapy and an anti-VEGFR-2 antibody in multidrug-resistant human breast cancer xenografts. Clin. Cancer Res. 8, 221–232 (2002). 35. 35. Mumenthaler, S. M. et al. The Impact of Microenvironmental Heterogeneity on the Evolution of Drug Resistance in Cancer Cells. Cancer Inform. 14, 19–31 (2015). 36. 36. Hansen, E., Woods, R. J. & Read, A. F. How to Use a Chemotherapeutic Agent When Resistance to It Threatens the Patient. PLoS Biol. 15, e2001110 (2017). 37. 37. Thomas, H. D. et al. Preclinical in vitro and in vivo evaluation of the potent and specific cyclin-dependent kinase 2 inhibitor NU6102 and a water soluble prodrug NU6301. Eur. J. Cancer 47, 2052–2059 (2011). 38. 38. Silva, A. S. & Gatenby, R. A. A theoretical quantitative model for evolution of cancer chemotherapy resistance. Biol. Direct. 5, 25 (2010). 39. 39. Gallaher, J. A., Enriquez-Navas, P. M., Luddy, K. A., Gatenby, R. A. & Anderson, A. R. A. Adaptive vs continuous cancer therapy: Exploiting space and trade-offs in drug scheduling. bioRxiv Preprint at https://doi.org/10.1101/128959 (2017). 40. 40. Ledzewicz, U. & Schättler, H. Application of mathematical models to metronomic chemotherapy: What can be inferred from minimal parameterized models? Cancer Lett. 401, 74–80 (2017). 41. 41. Hahnfeldt, P., Panigrahy, D., Folkman, J. & Hlatky, L. Tumor development under angiogenic signaling: a dynamical theory of tumor growth, treatment response, and postvascular dormancy. Cancer Res. 59, 4770–4775 (1999). 42. 42. Sherr, C. J., Beach, D. & Shapiro, G. I. Targeting CDK4 and CDK6: From Discovery to Therapy. Cancer Discov. 6, 353–367 (2016). 43. 43. Wang, L. et al. Pharmacologic inhibition of CDK4/6: mechanistic evidence for selective activity or acquired resistance in acute myeloid leukemia. Blood 110, 2075–2083 (2007). 44. 44. Bozic, I. et al. Evolutionary dynamics of cancer in response to targeted combination therapy. eLife 2, e00747 (2013). 45. 45. Zhao, B. et al. Exploiting temporal collateral sensitivity in tumor clonal evolution. Cell 165, 234–246 (2016). 46. 46. Michor, F. & Beal, K. Improving cancer treatment via mathematical modeling: surmounting the challenges is worth the effort. Cell 163, 1059–1063 (2015). 47. 47. Friedrich, J., Seidel, C., Ebner, R. & Kunz-Schughart, L. A. Spheroid-based drug screen: considerations and practical approach. Nat. Protoc. 4, 309–324 (2009). 48. 48. Ivanov, D. P. et al. Multiplexing spheroid volume, resazurin and acid phosphatase viability assays for high-throughput screening of tumour spheroids and stem cell neurospheres. PLoS ONE 9, e103817 (2014). 49. 49. Chevin, L.-M. On measuring selection in experimental evolution. Biol. Lett. 7, 210–213 (2011). 50. 50. Gerlee, P. The model muddle: in search of tumor growth laws. Cancer Res. 73, 2407–2411 (2013). 51. 51. Soetaert, K., Petzoldt, T. & Setzer, R. W. Solving differential equations in R: package deSolve. J. Stat. Softw. 33, 1–25 (2010). 52. 52. Hahnfeldt, P., Panigrahy, D., Folkman, J. & Hlatky, L. Tumor development under angiogenic signaling: a dynamical theory of tumor growth, treatment response, and postvascular dormancy. Cancer Res. 59, 4770–4775 (1999). 53. 53. Patel, A. A., Gawlinski, E. T., Lemieux, S. K. & Gatenby, R. A. A cellular automaton model of early tumor growth and invasion. J. Theor. Biol. 213, 315–331 (2001). 54. 54. Gillespie, D. T. Exact stochastic simulation of coupled chemical reactions. J. Phys. Chem. 81, 2340–2361 (1977). ## Acknowledgements This work was supported by Inserm grant No PC201306 and Ligue Nationale Contre le Cancer grant EL2013.LNCC/DF. K.B. was supported by a fellowship from the Ligue Nationale Contre le Cancer. A.S. was supported by a cooperation grant from the French Institute Cairo and Egyptian Science and Technology Development Fund. MH, RN and D.F. were supported by ITMO ‘Physique Cancer’ (CanEvolve PC201306). Thanks to V. Georget of MRI Montpellier imaging facilities for assistance with light microscopy, and F. Bernex and N. Pirot of the Montpellier Histology Facility (RHEM). ## Author information D.F. and M.E.H. conceived the project. L.K. R.N. and D.F. designed and analysed the experiments and wrote the manuscript. R.N. performed the mathematical modeling. K.B., A.S., O.W.A., B.B. performed the cell biology experiments. S.P. supervised B.B. C.V. performed mouse xenograft experiments. Correspondence to Daniel Fisher. ## Ethics declarations ### Competing interests The authors declare no competing financial interests. Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. ## Rights and permissions Reprints and Permissions • ### A call for integrated metastatic management • Jessica J. Cunningham Nature Ecology & Evolution (2019) • ### The 2019 mathematical oncology roadmap • Russell C Rockne • , Andrea Hawkins-Daarud • , Kristin R Swanson • , James P Sluka • , James A Glazier • , Paul Macklin • , David A Hormuth • , Angela M Jarrett • , Ernesto A B F Lima • , J Tinsley Oden • , George Biros • , Thomas E Yankeelov • , Kit Curtius • , Ibrahim Al Bakir • , Dominik Wodarz • , Natalia Komarova • , Luis Aparicio • , Mykola Bordyuh • , Stacey D Finley • , Heiko Enderling • , Jimmy Caudell • , Eduardo G Moros • , Alexander R A Anderson • , Robert A Gatenby • , Artem Kaznatcheev • , Peter Jeavons • , Nikhil Krishnan • , Julia Pelesko • , Nara Yoon • , Daniel Nichol • , Andriy Marusyk • , Michael Hinczewski •  & Jacob G Scott Physical Biology (2019) • ### Modelling bistable tumour population dynamics to design effective treatment strategies • Andrei R. Akhmetzhanov • , Jong Wook Kim • , Ryan Sullivan • , Robert A. Beckman • , Pablo Tamayo •  & Chen-Hsiang Yeang Journal of Theoretical Biology (2019) • ### Solid stress, competition for space and cancer: The opposing roles of mechanical cell competition in tumour initiation and growth • Romain Levayer Seminars in Cancer Biology (2019) • ### PhysiBoSS: a multi-scale agent-based modelling framework integrating physical dimension and cell signalling • Gaelle Letort • , Arnau Montagud • , Gautier Stoll • , Randy Heiland • , Emmanuel Barillot • , Paul Macklin • , Andrei Zinovyev • , Laurence Calzone •  & Jonathan Wren Bioinformatics (2019)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 2, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8092377185821533, "perplexity": 10898.880317209834}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-30/segments/1563195528635.94/warc/CC-MAIN-20190723002417-20190723024417-00437.warc.gz"}
https://yutsumura.com/use-coordinate-vectors-to-show-a-set-is-a-basis-for-the-vector-space-of-polynomials-of-degree-2-or-less/
# Use Coordinate Vectors to Show a Set is a Basis for the Vector Space of Polynomials of Degree 2 or Less ## Problem 588 Let $P_2$ be the vector space over $\R$ of all polynomials of degree $2$ or less. Let $S=\{p_1(x), p_2(x), p_3(x)\}$, where $p_1(x)=x^2+1, \quad p_2(x)=6x^2+x+2, \quad p_3(x)=3x^2+x.$ (a) Use the basis $B=\{x^2, x, 1\}$ of $P_2$ to prove that the set $S$ is a basis for $P_2$. (b) Find the coordinate vector of $p(x)=x^2+2x+3\in P_2$ with respect to the basis $S$. ## Solution. ### (a) Prove that the set $S$ is a basis for $P_2$. The coordinate vectors of $p_1(x)$ with respect to the basis $B=\{x^2, x, 1\}$ is given by $[p_1(x)]_B=\begin{bmatrix} 1 \\ 0 \\ 1 \end{bmatrix}$ as $p_(x)$ can be written as a linear combination $p_1(x)=1\cdot x^2+0\cdot x+1\cdot 1$ of the basis vectors in $B$. Similarly, we have $[p_2(x)]_B=\begin{bmatrix} 6 \\ 1 \\ 2 \end{bmatrix} \text{ and } [p_3(x)]_B=\begin{bmatrix} 3 \\ 1 \\ 0 \end{bmatrix}.$ Let $T=\{[p_1(x)]_B, [p_2(x)]_B, [p_3(x)]_B\}$. Then $S$ is a basis for $P_2$ if and only if $T$ is a basis for $\R^3$. Thus it remains to show that $T$ is a basis for $\R^3$. Consider the matrix whose column vectors are vectors in $T$. We have \begin{align*} \begin{bmatrix} 1 & 6 & 3 \\ 0 &1 &1 \\ 1 & 2 & 0 \end{bmatrix} \xrightarrow{R_3-R_1} \begin{bmatrix} 1 & 6 & 3 \\ 0 &1 &1 \\ 0 & -4 & -3 \end{bmatrix} \xrightarrow{\substack{R_1-6R_2\\R_3+4R_2}}\6pt] \begin{bmatrix} 1 & 0 & -3 \\ 0 &1 &1 \\ 0 & 0 & 1 \end{bmatrix} \xrightarrow{\substack{R_1+3R_3\\R_2-R_3}} \begin{bmatrix} 1 & 0 & 0 \\ 0 &1 &0 \\ 0 & 0 & 1 \end{bmatrix}. \end{align*} It follows that T is linearly independent. As T consists of three linearly independent vectors in \R^3, it is a basis for \R^3. Hence S is a basis for P_2. ### (b) Find the coordinate vector of p(x)=x^2+2x+3\in P_2 with respect to the basis S. To find the coordinate vector [p(x)]_S with respect to the basis S, we express p(x) as a linear combination of the basis vectors in S. Thus we want to find scalars c_1, c_2, c_3 such that \[p(x)=c_1p_1(x)+c_2p_2(x)+c_3p_3(x). Considering the coordinate vectors of both sides, this is equivalent to $\begin{bmatrix} 1 \\ 2 \\ 3 \end{bmatrix}=c_1\begin{bmatrix} 1 \\ 0 \\ 1 \end{bmatrix}+c_2\begin{bmatrix} 6 \\ 1 \\ 2 \end{bmatrix}+c_3\begin{bmatrix} 3 \\ 1 \\ 0 \end{bmatrix}.$ We write this equation as a matrix equation $\begin{bmatrix} 1 & 6 & 3 \\ 0 &1 &1 \\ 1 & 2 & 0 \end{bmatrix}\begin{bmatrix} c_1 \\ c_2 \\ c_3 \end{bmatrix}=\begin{bmatrix} 1 \\ 2 \\ 3 \end{bmatrix}.$ To solve this, we apply elementary row operations to the augmented matrix as follows. \begin{align*} \left[\begin{array}{rrr|r} 1 & 6 & 3 & 1 \\ 0 &1 & 1 & 2 \\ 1 & 2 & 0 & 3 \end{array} \right] \xrightarrow{R_3-R_1} \left[\begin{array}{rrr|r} 1 & 6 & 3 & 1 \\ 0 &1 & 1 & 2 \\ 0 & -4 & -3 & 2 \end{array} \right] \xrightarrow{\substack{R_1-6R_2\\R_3+4R_2}}\6pt] \left[\begin{array}{rrr|r} 1 & 0 & -3 & -11 \\ 0 &1 & 1 & 2 \\ 0 & 0 & 1 & 10 \end{array} \right] \xrightarrow{\substack{R_1+3R_3\\R_2-R_3}} \left[\begin{array}{rrr|r} 1 & 0 & 0 & 19 \\ 0 &1 & 0 & -8 \\ 0 & 0 & 1 & 10 \end{array} \right]. \end{align*} (Note that elementary row operations are exactly the same as before.) Hence the solution is c_1=19, c_2=-8, c_3=10. Thus, we have the linear combination \[p(x)=19p_1(x)-8p_2(x)+10p_3(x) and the coordinate vector of $p(x)$ with respect to the basis $S$ is $[p(x)]_S=\begin{bmatrix} 19 \\ -8 \\ 10 \end{bmatrix}.$ ### More from my site #### You may also like... ##### Commuting Matrices $AB=BA$ such that $A-B$ is Nilpotent Have the Same Eigenvalues Let $A$ and $B$ be square matrices such that they commute each other: $AB=BA$. Assume that $A-B$ is a nilpotent... Close
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 2, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9998343586921692, "perplexity": 68.71226853259029}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2018-22/segments/1526794866772.91/warc/CC-MAIN-20180524190036-20180524210036-00637.warc.gz"}
https://ajayshahblog.blogspot.com/2007/06/tough-love.html
## Saturday, June 09, 2007 ### Tough love There is a certain tension in designing a bankruptcy code, where both extremes - of killing a firm too soon or too late - are bad, and the trick seems to be to find the right middle. In India, discussions about creditors rights are clouded by a bias in favour of being soft on an incumbent firm. There is a certain bleeding heart sentiment which comes into play, where the State tries to help a weak firm claw back to life, where it is felt that the death of firms is a bad thing. Tom Chang and Antoniette Schoar have an important paper The effect of judicial bias on Chapter 11 reorganisation where they obtain new insights on these questions. First, using US data they are able to watch judges repeatedly processing bankruptcy cases, and are able to identify some judges who are more lenient than others. This is, in itself, a fascinating case of the value of bringing economic reasoning to bear on legal thinking. Ideally, judges are not supposed to display such differences in behaviour, but bankruptcy proceedings constitute an opportunity to watch the same judge perform on many homogeneous cases and thus identify judge characteristics. They go on to look at what happens to the firms which benefited from dealing with a lenient judge at the bankruptcy proceeding. These firms don't bounce back, they don't do well. Please note: LaTeX mathematics works. This means that if you want to say $10 you have to say \$10.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.1635390669107437, "perplexity": 1963.7361222343204}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-51/segments/1512948592846.98/warc/CC-MAIN-20171217015850-20171217041850-00701.warc.gz"}
https://mathoverflow.net/questions/215295/calculating-homology-of-the-boundary-of-a-handlebody
Calculating Homology of the Boundary of a Handlebody Given a manifold $M$ with boundary $W = \partial M$, I know that having a handle decomposition of $M$ allows one to compute its homology, at least in nice cases, by - for example - using the Morse Homology of its critical points. Is it similarly easy to compute the homology of $W$, since the handle decomposition of $M$ provides a surgery description of $W$? If it helps, I'm interested in a particularly simple case: $M$ is a smooth $2n$-manifold which is described by simultaneously adding some number of $n$-handles to $D^{2n}$. However, I am primarily concerned with the integer homology, rather than over $\mathbb{Z}/2\mathbb{Z}$ or $\mathbb{Q}$. For $n>1$, manifold itself is determined by the linking numbers between the attaching spheres of your handles, and the framings of those handles. (For $n=1$, you have to take into account knotting and linking of the attaching circles.) The framings are in 1-1 correspondence with elements of $\pi_n(SO(n+1))$. The homology depends on the image of the framings under the map $\pi_n(SO(n+1))\to \pi_n(S^n)$. Geometrically, this means to take the linking number of your attaching sphere with a pushoff given by the first vector of the given framing. Then the homology is presented by the $(-1)^{n-1}$ symmetric matrix of linking numbers and images of these framings. The proof is an exercise using Poincar\'e duality and the long exact sequence of the pair $(W,M)$. This is a standard fact in the case $n=2$, in which case the map $\pi_1(SO(2))\to \pi_1(S^1)$ is a bijection, and I think you can find it in such sources as Gompf-Stipsicz. For the general high-dimensional case, look at Classification of (n-1)-Connected 2n-manifolds by C.T.C. Wall, Annals of Math Vol. 75, No. 1 (1962), pp. 163-189.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9658253192901611, "perplexity": 145.7146533625599}, "config": {"markdown_headings": false, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243989705.28/warc/CC-MAIN-20210512193253-20210512223253-00453.warc.gz"}
https://brilliant.org/problems/hot-integral-5/
# Hot Integral-5 Calculus Level 5 Let $$\displaystyle f(x)=\int\limits_0^1 \frac{\sqrt{1-xt^2}}{\sqrt{1-t^2}}dt$$ Then, the following integral can be expressed as : $$\displaystyle \int \frac{f(2015x^2)}{x^3}dx = -\frac{a}{bz^c}[\pi \left\{dz^f \;_4F_3(g,h,\frac{i}{k},\frac{j}{k};l,m,m;nz^o) -pz^q \log(r) + sz^t\log(-uz^v) + w\right\}]$$ + constant Evaluate $$a+b+c+d+f+g+h+i+j+k+l+m+n+o+p+q+r+s+t+u+v+w = ?$$ $$\textbf{Details and Assumptions}$$ 1)$$a,b,c,d,f,g,h,i,j,k,l,m,n,o,p,q,r,s,t,u,v,w$$ are all positive integers. 2) $$_4F_3(.,.,.,.;.,.,.;.)$$ is a $$\textbf{Hypergeometric function}$$. 3) Everything is in the $$\textbf{simplest form}$$ and $$r$$ is square free. $$\blacksquare$$ This is a part of Hot Integrals ×
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9170200824737549, "perplexity": 1464.825861508968}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-43/segments/1508187822145.14/warc/CC-MAIN-20171017163022-20171017183022-00682.warc.gz"}
http://www.ck12.org/book/CK-12-Foundation-and-Leadership-Public-Schools%252C-College-Access-Reader%253A-Geometry/r1/section/2.11/
<meta http-equiv="refresh" content="1; url=/nojavascript/"> Rhombus Properties | CK-12 Foundation # 2.11: Rhombus Properties Created by: CK-12 ## Learning Objectives • Identify and classify a rhombus. • Identify the relationship between diagonals in a rhombus. • Identify the relationship between diagonals and opposite angles in a rhombus. • Identify and explain biconditional statements. ## Perpendicular Diagonals in Rhombi Rhombi (plural of rhombus) are equilateral. • All four sides of a rhombus are congruent. • Also, a square is a special kind of rhombus and shares all of the properties of a rhombus. The diagonals of a rhombus not only bisect each other (because they are parallelograms), they do so at a right angle. In other words, the diagonals are perpendicular. This can be very helpful when you need to measure angles inside rhombi or squares. A rhombus has four _____________________________ sides. The diagonals of a rhombus are the same _____________________ and meet at a ________ angle, meaning they are perpendicular. Theorem for Rhombus Diagonals The diagonals of a rhombus are perpendicular bisectors of each other. ## Diagonals as Angle Bisectors Since a rhombus is a parallelogram, opposite angles are congruent. One property unique to rhombi is that in any rhombus, the diagonals will bisect the interior angles. Theorem for Rhombus Diagonals The diagonals of a rhombus bisect the interior angles. The diagonals of a rhombus are _____________________________ bisectors of each other. The diagonals of a rhombus also ___________________________ the interior angles. 1. Fill in the blank: A rhombus is a parallelogram with congruent ________________________. 2. Label the right angles in the picture below: 3. What is the measure of angle $ABC$ in the rhombus below? ## Biconditional Statements A biconditional statement is a conditional statement that also has a true converse. For example, a true biconditional statement is, “If a quadrilateral is a square then it has exactly four congruent sides and four congruent angles.” This statement is true, as is its converse: “If a quadrilateral has exactly four congruent sides and four congruent angles, then that quadrilateral is a square.” A biconditional statement is a true if-then statement whose _________________________ is also true. Remember... A conditional statement is an “if-then” statement. A converse is a statement in which the hypothesis and conclusion are reversed. Sometimes converses are true and sometimes they are not. When a conditional statement can be written as a biconditional, then we use the term “if and only if.” In the previous example, we could say: “A quadrilateral is a square if and only if it has four congruent sides and four congruent angles.” Example 1 Which of the following is a true biconditional statement? A. A polygon is a square if and only if it has four right angles. B. A polygon is a rhombus if and only if its diagonals are perpendicular bisectors. C. A polygon is a parallelogram if and only if its diagonals bisect the interior angles. D. A polygon is a rectangle if and only if its diagonals bisect each other. Examine each of the statements to see if it is true: A. A polygon is a square if and only if it has four right angles. • It is true that if a polygon is a square, it has four right angles. However, the converse statement is not necessarily true. A rectangle also has four right angles, and a rectangle is not necessarily a square. Providing an example that shows something is not true is called a counterexample. B. A polygon is a rhombus if and only if its diagonals are perpendicular bisectors. • The second statement seems correct. It is true that rhombi have diagonals that are perpendicular bisectors. The same is also true in converse—if a figure has perpendicular bisectors as diagonals, it is a rhombus. C. A polygon is a parallelogram if and only if its diagonals bisect the interior angles. • The third statement is not necessarily true. Not all parallelograms have diagonals that bisect the interior angles. This is true only of rhombi, not all parallelograms. D. A polygon is a rectangle if and only if its diagonals bisect each other. • This is not necessarily true. The diagonals in a rectangle do bisect each other, but parallelograms that are not rectangles also have bisecting diagonals. Choice D is not correct. So, after analyzing each statement carefully, only B is true. Choice B is the correct answer. 1. Write the following biconditional statement as an “if and only if” statement: The sun is the star at the center of our solar system. _______________________________________________________ if and only if _________________________________________________________________. 2. Is the following statement a true biconditional statement? If not, provide a counterexample. A polygon is a quadrilateral if and only if it has four sides. ## Graphic Organizer for Lesson 8 Logic Statements Type of Statement Description Example Conditional Statement • “________________” statement If a shape is a polygon, then it has straight sides. Converse • The hypothesis and conclusion are If a shape has straight sides, then it is a polygon. __________________________ • Not always true • Can be disproven with a ____________________ Biconditional Statement • Both the statement and its If a polygon has three sides, then it is a triangle. ____________________ are true. • Statement is true. • Good definitions are biconditional. • Converse (if it’s a triangle, then it’s a polygon with three sides) is also true. 8 , 9 , 10 ## Date Created: Feb 23, 2012 May 12, 2014 You can only attach files to None which belong to you If you would like to associate files with this None, please make a copy first.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 1, "texerror": 0, "math_score": 0.3059800863265991, "perplexity": 812.5614379925555}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-06/segments/1422115861027.55/warc/CC-MAIN-20150124161101-00100-ip-10-180-212-252.ec2.internal.warc.gz"}
http://link.springer.com/article/10.1007/s11537-012-1140-8
Japanese Journal of Mathematics , Volume 7, Issue 2, pp 235–294 Modular hyperbolas Original articles DOI: 10.1007/s11537-012-1140-8 Shparlinski, I.E. Jpn. J. Math. (2012) 7: 235. doi:10.1007/s11537-012-1140-8 Abstract We give a survey of a variety of recent results about the distribution and some geometric properties of points (x, y) on modular hyperbolas $${xy \equiv a\;(\mod m)}$$. We also outline a very diverse range of applications of such results, discuss multivariate generalisations and suggest a number of open problems of different levels of difficulty. Keywords modular hyperbola congruences exponential sums character sums Mathematics Subject Classification (2010) 11D79 11L07 11L40 11N69
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5895497798919678, "perplexity": 917.7220204246166}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917120187.95/warc/CC-MAIN-20170423031200-00554-ip-10-145-167-34.ec2.internal.warc.gz"}
http://www.emagtech.com/wiki/index.php?title=RF_Tutorial_Lesson_2:_Transient_Analysis_of_a_Simple_Transmission_Line_Circuit
# RF Tutorial Lesson 2: Transient Analysis of a Simple Transmission Line Circuit Tutorial Project: Transient Analysis of a Simple Transmission Line Circuit Objective: In this project, you will perform transient simulation and Fourier analysis of simple transmission line circuits. Concepts/Features: Transmission Line Generic T-Line Transient Test Sinusoidal Waveform Pulse Waveform Characteristic Impedance Reflection Coefficient Fourier Analysis Fundamental Frequency Harmonic Minimum Version Required: All versions ' Download Link: RF Lesson 2 ## What You Will Learn In this tutorial you will explore the transient response of transmission line circuits with various load configurations. You will also perform a Fourier analysis of non-sinusoidal signals. ## Putting the Circuit Together You will use the simple transmission line circuit of RF Tutorial Lesson 1 to start this project. However, instead of an AC voltage source, you will use a regular voltage source. The following is a list of parts needed for this part of the tutorial lesson: Part Name Part Type Part Value VS Voltage Source Waveform: TBD XTL1 Generic T-Line Defaults: Z0 = 50, eeff = 1, len = 10 RS Resistor 50 RL Resistor 50 SOURCE Voltage Probe Marker N/A IN Voltage Probe Marker N/A OUT Voltage Probe Marker N/A Set the length of the T-Line segment to 60mm. Set the waveform of the voltage source to "Sinusoidal" with a frequency of 2GHz and a peak amplitude of 1V. The free-space wavelength at this frequency is λ0 = 150mm. Therefore, the length of your T-Line segment is 0.4λ0. Place three voltage probe markers (keyboard shortcut: Alt+L) called "SOURCE", "IN" and "OUT" as shown in the figure. The basic transmission line circuit with three voltage probe markers. ## Transient Simulation of a Simple Transmission Line Circuit Run a Transient Test of your circuit with the specified parameters below. Note that your Plot Edit List must already contain v(SOURCE), v(IN) and v(OUT). The node voltages of all voltage probe markers are automatically added to the Plot Edit List. Start Time 0 3n 1p 1p v(source), v(in), v(out) Run the simulation and observe the time-domain voltages. The period of your sinusoidal waveform is T = 1/f0 = 1 / 2GHz = 500ps. As you would expect, the T-Line segment is matched to both the source and the load. There are no reflections off either the input or the output of the transmission line. You can see that the output voltage has a delay of 200ps with respect to the input voltage, which is consistent with the calculation given below: $\tau = \frac{L}{c} = \frac{60 mm}{3\times 10^8 m/s} = 200ps$ The graph of the source, input and output voltages for a sinusoidal waveform with f0 = 2GHz when RL = 50Ω. Changing the waveform in the property dialog of the voltage source. Next, you will try out a rectangular pulse waveform as your voltage source. Open the property dialog of VS and change the waveform type to "Pulse". Set the waveform parameters as specified below: Initial Voltage 0 1 0 1p 1p 50p 500p The duty cycle of the pulse train waveform is therefore 10%. Run a new transient test of your circuit with the same test parameters as before and compare the results to the previous case of a sinusoidal waveform. Here, too, there is a 200ps delay between the input and output voltages. Due to the perfect impedance match at both the input and output, v(in) and v(out) have equal amplitudes of 0.5V. Moreover, v(in) is simply is half-scaled replica of the source voltage. The graph of the source, input and output voltages for a pulse waveform with a period of T = 500ps when RL = 50Ω. ## Investigating the Effect of Load Mismatch Now change the load resistor to RL = 100Ω. Run the Transient Test with the same settings as before for both cases of sinusoidal and pulse waveforms: Start Time 0 3n 1p 1p v(source), v(in), v(out) The results of the two cases are shown in the figures below: The graph of the source, input and output voltages for a sinusoidal waveform with f0 = 2GHz when RL = 100Ω. The graph of the source, input and output voltages for a pulse waveform with a period of T = 500ps when RL = 100Ω. As you can see from these figures, the effect of the load mismatch in the sinusoidal signal case is the difference in the amplitudes of the input voltage and load voltage. This is due to the nonzero load reflection coefficient: $\Gamma_L = \frac{ Z_L - Z_0 }{ Z_L + Z_0 } = \frac{ 100 - 50 }{ 100 + 50 } = \frac{1}{3}$ In the case of the pulse waveform, the impedance mismatch introduces a reflected pulse that reaches node "IN" after a delay of 400ps, which is the round-trip time from the input point to the load and back. The amplitude of the smaller reflected pulse (166.7mv) is 1/3 the amplitude of the larger incident pulse (500mV) as you would expect from the value of the reflection coefficient. ## Analyzing a Quarter-Wave Impedance Transformer In RF Tutorial lesson 1, you designed a quarter-wave impedance transform to match an arbitrary resistive load to a 50Ω source. Set the length of the T-Line segment to L = λ0/4 = 37.5mm for an operating frequency of 2GHz. Also, set the characteristic impedance of the T-line to Z0 = √(100.50) = 70.71Ω. Run a new transient test with the same settings as before for both cases of sinusoidal and pulse waveforms: Start Time 0 3n 1p 1p v(source), v(in), v(out) The quarter-wave impedance transformer circuit designed for 2GHz operation. The results are shown and compared in the figures below: The graph of the source, input and output voltages for a sinusoidal waveform with f0 = 2GHz when the T-Line segment acts as a quarter-wave transformer. The graph of the source, input and output voltages for a pulse waveform with a period of T = 500ps when the T-Line segment acts as a quarter-wave transformer. The input impedance of the quarter-wave T-Line segment (at 2GHz) is given by: $Z_{in} = \frac{Z_{0c}^2}{Z_L} = \frac{70.71^2}{100} = 50\Omega$ As you can see from the above figure, in the case of the "pure harmonic" sinusoidal excitation, the source voltage is equally split between the source resistor RS and the input port of the T-Line. However, the situation is slightly different in the case of pulse waveform. The dispersive effects of the transmission line are in full display in this case. In other words, the matching condition is satisfied only at 2GHz and not at its harmonics present in the pulse waveform. Note that both load and source reflection coefficients are nonzero for this circuit: $\Gamma_L = \frac{ Z_L - Z_{0c} }{ Z_L + Z_{0c} } = \frac{ 100 - 70.71 }{ 100 + 70.71 } = 0.172$ $\Gamma_S = \frac{ Z_S - Z_{0c} }{ Z_S + Z_{0c} } = \frac{ 50 - 70.71 }{ 50 + 70.71 } = -0.172$ The Fourier Transform Settings dialog. To better understand this point, you can use RF.Spice's Fourier analysis feature, which is part of the transient test. You will run the "Transient Test" for the pulse waveform two more times with the "Fourier Analysis" enabled in the Transient Test Panel. Check the "Apply Fourier" checkbox and click the "Fourier Setup" button to open the Fourier Transform Settings dialog. Set the Fundamental Frequency to 2GHz and set the reference output node to "0" for the ground. Then, set the positive output node to "1" for the source voltage in the first run and then to "2" for the input voltage in the second run. At the end of the simulation, additional bar chart graphs are added to the Data Manager window. The spectral contents of the source and input voltages are shown in the figures below. You can clearly see that both signals have significant DC contents as well as sizable higher harmonics. The spectral contents of the source voltage v(source) with a pulse waveform. The spectral contents of the input voltage v(in) with a pulse waveform. ## Investigating the Effect of a Capacitive Load In the last part of this tutorial lesson, you add a 0.1pF shunt capacitor called "CL" to the load and see its effect in the case of pulse train signal. The modified circuit is shown in the opposite figure. The impedance of the shunt capacitor at the nth harmonic of the source's 2GHz fundamental frequency is given by: $Z_{Cap} = -\frac{j}{\omega C_L} = -\frac{j}{2\pi n (2\times 10^9)(0.1\times 10^{-12})} \approx -j(796/n) \ \Omega$ It can be seen that at the fundamental frequency, the capacitor has very negligible effect, but at higher harmonics, its impedance becomes comparable to the 100Ω resistive loads. The quarter-wave impedance transformer circuit with a parallel capacitor at the load. Run a transient test of your modified circuit with the same settings as before only for the case of pulse waveform: Start Time 0 3n 1p 1p v(source), v(in), v(out) Enable the "Fourier Analysis" for Node 2 with a fundamental frequency of 2GHz. The figures below show the simulation results. You can see from the top figure that the output voltage has been slightly deformed with a rounded rising edge. Also, the reflected pulse now has significant undershoot and overshoot at its rising and falling edges, respectively. Next, compare the spectral contents of v(in) for the two cases without and with the shunt capacitor. The contents at DC, fundamental frequency and first few harmonics are almost the same, but the results start to differ at higher harmonics as we justified earlier. The graph of the source, input and output voltages in the quarter-wave transformer circuit with a series short stub before the load. The graph of the source, input and output voltages in the quarter-wave transformer circuit with a shunt open stub before the load.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6567320227622986, "perplexity": 1395.0730895082818}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-26/segments/1498128323721.80/warc/CC-MAIN-20170628171342-20170628191342-00273.warc.gz"}
http://nrich.maths.org/public/leg.php?code=5039&cl=2&cldcmpid=10333
# Search by Topic #### Resources tagged with Interactivities similar to Dicey Perimeter, Dicey Area: Filter by: Content type: Stage: Challenge level: ### There are 223 results Broad Topics > Information and Communications Technology > Interactivities ### More Transformations on a Pegboard ##### Stage: 2 Challenge Level: Use the interactivity to find all the different right-angled triangles you can make by just moving one corner of the starting triangle. ### Junior Frogs ##### Stage: 1 and 2 Challenge Level: Have a go at this well-known challenge. Can you swap the frogs and toads in as few slides and jumps as possible? ### One to Fifteen ##### Stage: 2 Challenge Level: Can you put the numbers from 1 to 15 on the circles so that no consecutive numbers lie anywhere along a continuous straight line? ### Multiples Grid ##### Stage: 2 Challenge Level: What do the numbers shaded in blue on this hundred square have in common? What do you notice about the pink numbers? How about the shaded numbers in the other squares? ### Arrangements ##### Stage: 2 Challenge Level: Is it possible to place 2 counters on the 3 by 3 grid so that there is an even number of counters in every row and every column? How about if you have 3 counters or 4 counters or....? ### Combining Cuisenaire ##### Stage: 2 Challenge Level: Can you find all the different ways of lining up these Cuisenaire rods? ### A Square of Numbers ##### Stage: 2 Challenge Level: Can you put the numbers 1 to 8 into the circles so that the four calculations are correct? ### Difference ##### Stage: 2 Challenge Level: Place the numbers 1 to 10 in the circles so that each number is the difference between the two numbers just below it. ### Winning the Lottery ##### Stage: 2 Challenge Level: Try out the lottery that is played in a far-away land. What is the chance of winning? ### Four Triangles Puzzle ##### Stage: 1 and 2 Challenge Level: Cut four triangles from a square as shown in the picture. How many different shapes can you make by fitting the four triangles back together? ### Counters ##### Stage: 2 Challenge Level: Hover your mouse over the counters to see which ones will be removed. Click to remover them. The winner is the last one to remove a counter. How you can make sure you win? ### Tetrafit ##### Stage: 2 Challenge Level: A tetromino is made up of four squares joined edge to edge. Can this tetromino, together with 15 copies of itself, be used to cover an eight by eight chessboard? ### Code Breaker ##### Stage: 2 Challenge Level: This problem is based on a code using two different prime numbers less than 10. You'll need to multiply them together and shift the alphabet forwards by the result. Can you decipher the code? ### Factor Lines ##### Stage: 2 Challenge Level: Arrange the four number cards on the grid, according to the rules, to make a diagonal, vertical or horizontal line. ### Red Even ##### Stage: 2 Challenge Level: You have 4 red and 5 blue counters. How many ways can they be placed on a 3 by 3 grid so that all the rows columns and diagonals have an even number of red counters? ##### Stage: 1 and 2 Challenge Level: Place six toy ladybirds into the box so that there are two ladybirds in every column and every row. ### Teddy Town ##### Stage: 1, 2 and 3 Challenge Level: There are nine teddies in Teddy Town - three red, three blue and three yellow. There are also nine houses, three of each colour. Can you put them on the map of Teddy Town according to the rules? ### Seven Flipped ##### Stage: 2 Challenge Level: Investigate the smallest number of moves it takes to turn these mats upside-down if you can only turn exactly three at a time. ### Nine-pin Triangles ##### Stage: 2 Challenge Level: How many different triangles can you make on a circular pegboard that has nine pegs? ### Triangles All Around ##### Stage: 2 Challenge Level: Can you find all the different triangles on these peg boards, and find their angles? ### Fault-free Rectangles ##### Stage: 2 Challenge Level: Find out what a "fault-free" rectangle is and try to make some of your own. ##### Stage: 2 Challenge Level: If you have only four weights, where could you place them in order to balance this equaliser? ### Cycling Squares ##### Stage: 2 Challenge Level: Can you make a cycle of pairs that add to make a square number using all the numbers in the box below, once and once only? ### Square Tangram ##### Stage: 2 Challenge Level: This was a problem for our birthday website. Can you use four of these pieces to form a square? How about making a square with all five pieces? ### Countdown ##### Stage: 2 and 3 Challenge Level: Here is a chance to play a version of the classic Countdown Game. ##### Stage: 1 and 2 Challenge Level: Our 2008 Advent Calendar has a 'Making Maths' activity for every day in the run-up to Christmas. ### Magic Potting Sheds ##### Stage: 3 Challenge Level: Mr McGregor has a magic potting shed. Overnight, the number of plants in it doubles. He'd like to put the same number of plants in each of three gardens, planting one garden each day. Can he do it? ### Multiplication Square Jigsaw ##### Stage: 2 Challenge Level: Can you complete this jigsaw of the multiplication square? ### Number Differences ##### Stage: 2 Challenge Level: Place the numbers from 1 to 9 in the squares below so that the difference between joined squares is odd. How many different ways can you do this? ### Sorting Symmetries ##### Stage: 2 Challenge Level: Find out how we can describe the "symmetries" of this triangle and investigate some combinations of rotating and flipping it. ### A Dotty Problem ##### Stage: 2 Challenge Level: Starting with the number 180, take away 9 again and again, joining up the dots as you go. Watch out - don't join all the dots! ### One Million to Seven ##### Stage: 2 Challenge Level: Start by putting one million (1 000 000) into the display of your calculator. Can you reduce this to 7 using just the 7 key and add, subtract, multiply, divide and equals as many times as you like? ##### Stage: 2 Challenge Level: Three beads are threaded on a circular wire and are coloured either red or blue. Can you find all four different combinations? ### More Carroll Diagrams ##### Stage: 2 Challenge Level: How have the numbers been placed in this Carroll diagram? Which labels would you put on each row and column? ### Domino Numbers ##### Stage: 2 Challenge Level: Can you see why 2 by 2 could be 5? Can you predict what 2 by 10 will be? ### Cuisenaire Environment ##### Stage: 1 and 2 Challenge Level: An environment which simulates working with Cuisenaire rods. ### Play a Merry Tune ##### Stage: 2 Challenge Level: Explore the different tunes you can make with these five gourds. What are the similarities and differences between the two tunes you are given? ### Coded Hundred Square ##### Stage: 2 Challenge Level: This 100 square jigsaw is written in code. It starts with 1 and ends with 100. Can you build it up? ##### Stage: 2 Challenge Level: NRICH December 2006 advent calendar - a new tangram for each day in the run-up to Christmas. ### Which Symbol? ##### Stage: 2 Challenge Level: Choose a symbol to put into the number sentence. ### World of Tan 24 - Clocks ##### Stage: 2 Challenge Level: Can you fit the tangram pieces into the outlines of these clocks? ### Part the Piles ##### Stage: 2 Challenge Level: Try to stop your opponent from being able to split the piles of counters into unequal numbers. Can you find a strategy? ### Board Block for Two ##### Stage: 1 and 2 Challenge Level: Board Block game for two. Can you stop your partner from being able to make a shape on the board? ### Isosceles Triangles ##### Stage: 3 Challenge Level: Draw some isosceles triangles with an area of $9$cm$^2$ and a vertex at (20,20). If all the vertices must have whole number coordinates, how many is it possible to draw? ### Coordinate Tan ##### Stage: 2 Challenge Level: What are the coordinates of the coloured dots that mark out the tangram? Try changing the position of the origin. What happens to the coordinates now? ### Overlapping Circles ##### Stage: 2 Challenge Level: What shaped overlaps can you make with two circles which are the same size? What shapes are 'left over'? What shapes can you make when the circles are different sizes? ### Round Peg Board ##### Stage: 1 and 2 Challenge Level: A generic circular pegboard resource. ### Noughts and Crosses ##### Stage: 2 Challenge Level: A game for 2 people that everybody knows. You can play with a friend or online. If you play correctly you never lose! ### Twice as Big? ##### Stage: 2 Challenge Level: Investigate how the four L-shapes fit together to make an enlarged L-shape. You could explore this idea with other shapes too. ### World of Tan 22 - an Appealing Stroll ##### Stage: 2 Challenge Level: Can you fit the tangram pieces into the outline of the child walking home from school?
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.29585158824920654, "perplexity": 2948.603611913407}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-26/segments/1466783395166.84/warc/CC-MAIN-20160624154955-00022-ip-10-164-35-72.ec2.internal.warc.gz"}
https://wiki.seeedstudio.com/Grove-Electricity_Sensor/
edit # Grove - Electricity Sensor The Electricity sensor module is a member of Grove. It is based on the TA12-200 current transformer which can transform the large AC into small amplitude. You can use it to test large alternating current up to 5A. ## Features¶ • Grove compatible interface • Maximum 5A input • High accuracy • Small size Tip ## Application Ideas¶ • Alternating current measurement • Device condition monitoring ## Specification¶ ### Key Specification¶ Items Min PCB Size 2.0cm*4.0cm IO Structure SIG,NC,NC,GND RoHS YES ### Electronic Characteristics¶ Items Min Norm Max Unit Transformation ratio - 2000:1 - - Input Current 0 - 5 A Output Current 0 - 2.5 mA Sampling Resistance - 800 - Ω Sampling Voltage 0 - 2 V Working Frequency 20 - 20K HZ Nonlinear scale - - 0.2% - Phase Shift - - 5' - Operating Temperature -55 - 85 Dielectric strength - 6 - KVAC/1min ## Platforms Supported¶ Arduino Raspberry Pi Caution The platforms mentioned above as supported is/are an indication of the module's software or theoritical compatibility. We only provide software library or code examples for Arduino platform in most cases. It is not possible to provide software library / demo code for all possible MCU platforms. Hence, users have to write their own software library. ## Getting started¶ ### Play with Arduino¶ The following sketch demonstrates a simple application of measuring the amplitude of the alternating voltage. The SIG pin will output a alternating voltage based on the alternating current being measured. You can measure the value using ADC. #### Hardware¶ • Step 1. Prepare the below stuffs: Seeeduino V4.2 Base Shield Grove-Electricity_Sensor Get One Now Get One Now Get One Now • Step 2. Connect Grove-Electricity_Sensor to port A0 of Grove-Base Shield. • Step 3. Plug Grove - Base Shield into Seeeduino. • Step 4. Connect Seeeduino to PC via a USB cable. Note If we don't have Grove Base Shield, We also can directly connect this module to Seeeduino as below. Seeeduino Grove-Electricity_Sensor 5V Red GND Black Not Conencted White A0 Yellow #### Software¶ Step 1. Copy the code and flash it into the controller board and upload the code. /****************************************************************************/ // Function: Measure the amplitude current of the alternating current and // the effective current of the sinusoidal alternating current. // Hardware: Grove - Electricity Sensor // Date: Jan 19,2013 // by www.seeedstudio.com #define ELECTRICITY_SENSOR A0 // Analog input pin that sensor is attached to float amplitude_current; //amplitude current float effective_value; //effective current void setup() { Serial.begin(9600); pins_init(); } void loop() { int sensor_max; sensor_max = getMaxValue(); Serial.print("sensor_max = "); Serial.println(sensor_max); //the VCC on the Grove interface of the sensor is 5v amplitude_current=(float)sensor_max/1024*5/800*2000000; effective_value=amplitude_current/1.414;//minimum_current=1/1024*5/800*2000000/1.414=8.6(mA) //Only for sinusoidal alternating current Serial.println("The amplitude of the current is(in mA)"); Serial.println(amplitude_current,1);//Only one number after the decimal point Serial.println("The effective value of the current is(in mA)"); Serial.println(effective_value,1); } void pins_init() { pinMode(ELECTRICITY_SENSOR, INPUT); } /*Function: Sample for 1000ms and get the maximum value from the SIG pin*/ int getMaxValue() { int sensorValue; //value read from the sensor int sensorMax = 0; uint32_t start_time = millis(); while((millis()-start_time) < 1000)//sample for 1000ms { if (sensorValue > sensorMax) { /*record the maximum sensor value*/ sensorMax = sensorValue; } } return sensorMax; } Note The minimum effective current that can be sensed by the code can be calculated using the equation below. minimum_current=1/1024*5/800*2000000/1.414=8.6(mA). • Step 2. Open the serial monitor, The results is as follows: ### With Raspberry Pi¶ #### Hardware¶ • Step 1. Prepare the below stuffs: Raspberry pi GrovePi_Plus Grove-Electricity_Sensor Get One Now Get One Now Get One Now • Step 2. Plug the GrovePi_Plus into Raspberry. • Step 3. Connect Grove-Electricity_Sensor to A0 port of GrovePi_Plus. • Step 4. Connect the Raspberry to PC through USB cable. #### Software¶ • Step 1. Follow Setting Software to configure the development environment. • Step 2. Git clone the Github repository. cd ~ git clone https://github.com/DexterInd/GrovePi.git • Step 3. Excute below commands to use this sensor cd ~/GrovePi/Software/Python python grove_electricity_sensor.py Here is the code of example: #!/usr/bin/env python # # GrovePi Example for using the Grove Electricity Sensor (https://www.seeedstudio.com/wiki/Grove_-_Electricity_Sensor) # # The GrovePi connects the Raspberry Pi and Grove sensors. You can learn more about GrovePi here: http://www.dexterindustries.com/GrovePi # # ''' GrovePi for the Raspberry Pi: an open source platform for connecting Grove Sensors to the Raspberry Pi. Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. ''' import time import grovepi # Connect the Grove Electricity Sensor to analog port A0 # SIG,NC,NC,GND sensor = 0 grovepi.pinMode(sensor,"INPUT") # Vcc of the grove interface is normally 5v grove_vcc = 5 while True: try: # Get sensor value # Calculate amplitude current (mA) amplitude_current = (float)(sensor_value / 1024 * grove_vcc / 800 * 2000000) # Calculate effective value (mA) effective_value = amplitude_current / 1.414 # minimum_current = 1 / 1024 * grove_vcc / 800 * 2000000 / 1.414 = 8.6(mA) # Only for sinusoidal alternating current print("sensor_value", sensor_value) print("The amplitude of the current is", amplitude_current, "mA") print("The effective value of the current is", effective_value, "mA") time.sleep(1) except IOError: print ("Error") Here is the result. ## Tech Support¶ Please submit any technical issue into our forum.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.23213064670562744, "perplexity": 16259.461691794028}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656104354651.73/warc/CC-MAIN-20220704050055-20220704080055-00651.warc.gz"}
https://www.physicsforums.com/threads/car-driving-around-a-curve.175282/
# Car driving around a curve 1. Jun 27, 2007 ### canadiantiger7 1. The problem statement, all variables and given/known data a car drives around a curve with a radius 410 m at a speed of 38 m/s. the road is banked at 5.1 degrees, and the car weighs 1400 kg. what if the frictional force of the car? at what speed could you drive around this curve so that the force of friction is zero? I tried a few equations and got different answers, but none of them correct. any help is appreciated. 2. Relevant equations 3. The attempt at a solution 2. Jun 27, 2007 ### StatusX What exactly have you tried? You have to show some work before you'll get any help. 3. Jun 27, 2007 ### canadiantiger7 i got the acceleration (V2/R) and then plugged that in to get Fnet (m*a). but it did not work. that is what it said to do in the book though. 4. Jun 27, 2007 ### StatusX That is the net force on the car, but you need to determine where its coming from. Gravity always acts vertically with magnitude mg. The other two forces are the normal force of the road, which is perpendicular to the surface of the road, and the force of friction, which is parellel to the surface. These must be such that a) the vertical components add to zero, since there's no vertical acceleration and b) the horizontal components add to the net force, providing the centripetal force that keeps the car moving in a circle. 5. Mar 28, 2008 ### alexander38 Hi canadiantiger7, I’m not getting what you exactly want. Please provide more details so that I can help you. Thanks & Regards Find the location of police speed traps using your mobile phones and other types of devices, take a look at http://www.trapster.com 6. Mar 29, 2008 ### merryjman If you bank a curve steeply enough, the normal force will supply the needed centripetal acceleration rather than the frictional force. This is the principle at work at high speed race tracks - this way, the limited friction between the tires/road can be used strictly for accelerating and braking, and not turning as well. I would start with a free-body diagram if I were you, and then do what StatusX said above with regard to the horizontal and vertical forces. Note that the normal force does NOT point directly upward, as it would in other problems. Know someone interested in this topic? Share this thread via Reddit, Google+, Twitter, or Facebook Have something to add? Similar Discussions: Car driving around a curve
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8313265442848206, "perplexity": 692.8727369614636}, "config": {"markdown_headings": true, "markdown_code": false, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988718840.18/warc/CC-MAIN-20161020183838-00452-ip-10-171-6-4.ec2.internal.warc.gz"}
https://overbrace.com/bernardparent/viewtopic.php?f=95&t=1564&start=20
Fundamentals of Fluid Mechanics B Questions and Answers No, there are no other assumptions. These strain rates can be used for any fluid. Also, the same applies with the strain rates in cylindrical and spherical coordinates in the tables. The strain rates can be used in the general case. 03.24.20 Question by AME536A Student Could you also give us the final answers for Assignment 5 Question 3 (d) and (e)? Let me see if I can find them.. If I do, I'll post them soon. Question by AME536A Student I'm still having problems to solve Assignment#1 Question#2(b), the prove of Crocco's theorem. In class you gave us this hint: say $a=b-c$ $$da = d(b-c)$$ If $da=0$ and if s-s and uniform properties at some point upstream: $$\nabla a = 0$$ Then $$db-dc =0$$ and $$\nabla (b-c)=0$$ $$\nabla b = \nabla c$$ I don't understand how we can use this without ending up with a temperature gradient term in the equation. 03.25.20 Another hint. For two streamlines near each other and at a certain location, one has a temperature $T$, and the other the temperature $T+dT$. Expand terms and get rid of the terms that are necessarily much smaller than the others. Question by Student AME536B Do we need to memorize proofs for this midterm? The midterm is open book so you can consult your assignments, class notes, books if you want. However, if I were you, I would stick to consulting the tables: you won't waste time and score higher this way. Question by Student AME536B I was attempting to figure out the mean free path derivation question you gave us previously on how to get the $\sqrt2$ term. What I came up with using your tip of $$q_{rms}=c*q_{avg}$$ is, $$\overline{V_{rel}^{2}}=\overline{V_1^{2}}+\overline{V_2^{2}}$$ Is this process correct? I don't understand where your equation (2) is coming from. You need to clarify this. Question by Student AME536B On the tables, the equation given for the strain rate $S_{\theta r}$ in spherical coordinates has a term of $\frac{v_{\theta}}{r}$; is that a typo? Shouldn't the shear strain equation for spherical coordinates be given by: $$S_{r\theta} = \frac{1}{2} \left( \frac{1}{r}\frac{\partial v_r}{\partial \theta} + \frac{\partial v_{\theta}}{\partial r} \right)$$ 04.06.20 The equation in the tables is correct: there is no typo. Question by AME536A Student For question 1. b. on Assignment 6, do you want us to derive the expression you gave for the streamfunction? Or just start from the given equation? If it's given in the question don't derive it. Previous   1 ,  2   •  PDF 1✕1 2✕1 2✕2  •  New Question $\pi$
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 2, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 2, "x-ck12": 0, "texerror": 0, "math_score": 0.8197869658470154, "perplexity": 427.89344568757616}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585371803248.90/warc/CC-MAIN-20200407152449-20200407182949-00410.warc.gz"}
http://hackage.haskell.org/package/synthesizer-core-0.8.2.1/docs/Synthesizer-Plain-Cut.html
synthesizer-core-0.8.2.1: Audio signal processing coded in Haskell: Low level part Copyright (c) Henning Thielemann 2006 GPL [email protected] provisional requires multi-parameter type classes None Haskell2010 Synthesizer.Plain.Cut Contents Description Synopsis # dissection takeUntilPause :: C a => a -> Int -> T a -> T a Source # Take signal until it falls short of a certain amplitude for a given time. takeUntilInterval :: (a -> Bool) -> Int -> T a -> T a Source # Take values until the predicate p holds for n successive values. The list is truncated at the beginning of the interval of matching values. # glueing selectBool :: (T a, T a) -> T Bool -> T a Source # select :: Ix i => Array i (T a) -> T i -> T a Source # Arguments :: C v => T Int (T v) A list of pairs: (relative start time, signal part), The start time is relative to the start time of the previous event. -> T v The mixed signal. Given a list of signals with time stamps, mix them into one signal as they occur in time. Ideally for composing music. Cf. series
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.15050068497657776, "perplexity": 3091.5880205173166}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-30/segments/1563195526153.35/warc/CC-MAIN-20190719074137-20190719100137-00219.warc.gz"}
https://www.answers.com/Q/How_do_you_change_50_over_45_mixed_number_to_an_improper_fraction
Numbers # How do you change 50 over 45 mixed number to an improper fraction? ###### Wiki User 50/45 is an improper fraction. 1 and 1/9 is an equivalent mixed number in lowest form. 🙏 0 🤨 0 😮 0 😂 0 ###### Wiki User 50 over 45 or 50/45 IS an improper fraction! 🙏 0 🤨 0 😮 0 😂 0 ###### Wiki User It already is an improper fraction, the other form is 1 and 1/9 🙏 0 🤨 0 😮 0 😂 0 ## Related Questions An improper fraction is a fraction. It can't become a proper fraction. You do the numerator divided by the denominator, then simplify the fractions if you can. (this is how you change a improper fraction to a mixed number, you cannot change a proper fraction to a mixed number) Here is a video tutorial https://www.youtube.com/watch?v=HPetBqWKvZM on how to change an improper fraction to a mixed number. You change it into an improper fraction, then simplify it, then if you want to, change it back to a mixed number. Change the mixed number to an improper fraction and then divide. A mixed number can be converted to an improper fraction. I wouldconvert both to improper fraction,find a common multiple (CM),calculate equivalent fraction with CM as the denominator,add the numerators,change the improper fraction to a mixed fraction, if required,simplify the fractional part of the mixed fraction - if appropriate and required. I don't think you can but, you can change an improper fraction to a decimal. If you are using the same number, then the improper fraction and mixed number are equal. You change the mixed number to an improper fraction, multiply its denominator by the whole number. Then, if required, convert back to a mixed fraction. change it to an improper fraction and then divide it First you would want to change the mixed number to an improper fraction. Then you can subtract 595/24 is an improper fraction; to change it to mixed number = 24 19/24 A mixed number of the form AB/C, as an improper fraction, is equal to (AC + B)/CA mixed number of the form AB/C, as an improper fraction, is equal to (AC + B)/CA mixed number of the form AB/C, as an improper fraction, is equal to (AC + B)/CA mixed number of the form AB/C, as an improper fraction, is equal to (AC + B)/C 43 is an integer, not an improper fraction nor a mixed number. A mixed number can be changed into an improper fraction By changing the mixed number into an improper fraction. you have to change it into a improper fraction then divide the numerator into the domnator Yes, it can. In fact every mixed number has an equivalent improper fraction. mixed # into an improper fraction: multiply the whole number by the denominator(bottom number ont he fraction) then add that number to the numberator(the top number of the fraction. and that's it! improper fraction into a mised number: divide the numberator by the denominator. whatever WHOLE number you get is your number, now for the fraction. whatever you got as the remander when you divided but that on top of the fration. ex 47/5. 5 goes into 47 9 times with a remander of 2. that equals 9 and 2/5. Well you could change it into a mixed number or just leave it. ;)
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8691167831420898, "perplexity": 1149.8651848605923}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610704832583.88/warc/CC-MAIN-20210127183317-20210127213317-00142.warc.gz"}
http://motls.blogspot.com.au/2012/05/higgs-combo-vixra-java-applet.html
## Monday, May 21, 2012 ... ///// ### Higgs combo viXra java applet Cosmological update: Bovy and Tremaine of IAS looked at a recent Chilean claim – wildly hyped in the mainstream media (but only mentioned in one sentence on TRF, in an article on a different topic) – that there was no dark matter around the Solar System. When they corrected some profiles for the velocities, they found out that the dark matter density is nonzero and compatible with the usual estimates. Via Resonaances Phil Gibbs has written a cute and user-friendly java applet (download Java 7v4 instead of your dated Java 6v32 if you don't have the new one yet) that allows you to create thousands of charts relevant for the Higgs boson discovery: viXra combo applet (click) A screenshot of the applet is below. Once the page above shows up, try to change the "Plot Type" to Exclusion, Signal, Pvalue, Sigma and see how the bump near 125 GeV is immediately affected. If you want to spend more time, you may try to play with the decay channels and individual detectors that are included and other things. So far, in 2012, each major detector has recorded about 2.5 inverse femtobarns at 8 TeV. For the Higgs purposes, this is equivalent to nearly 3 inverse femtobarns at 7 TeV (and it is like 5/fb for the production of 1 TeV gluinos). So in combination with 5/fb from 2011, each detector has the equivalent of about 8/fb at 7 TeV. Click to zoom in. When you add all the channels at a single detector (CMS or ATLAS), the data accumulated so far are "more likely than not" to be enough for 5 sigma of local significance. For a global significance, one detector isn't quite enough but a combination of both ATLAS and CMS is surely enough by now to obtain 5-sigma global significance for the discovery of the 125 GeV Higgs boson, too. On July 7th, there will be a new announcement on the progress with the Higgs boson. We will have to see whether they will remain separated or not. My guess is that at this time, they will already have the combination from CMS and ATLAS and they will announce a shared discovery of the 125 GeV Higgs boson. (John Ramsden will strictly owe me \$500 at that moment, if true.) It's also plausible that they will actually collect much more than 2.5/fb in 2012 by then and evaluate the results quickly enough. If they collect much more than 5/fb for 2012 by the July conference, they could make the discovery announcement separately (or at least the "luckier ones"). Still, I find the inclusive "combo" solution more likely from a sociological viewpoint because it's kind of politically correct. Such an inclusive solution wouldn't occur for the first time. In the mid 1990s, the top quark was discovered once the data from CDF and D0 at the Tevatron were combined. When it comes to ATLAS and CMS, I don't think that one of them deserves to be credited with the Higgs discovery more than the other.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 1, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7633273601531982, "perplexity": 1317.246910541085}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-26/segments/1498128320395.62/warc/CC-MAIN-20170625032210-20170625052210-00187.warc.gz"}
https://www.springerprofessional.de/experiment-and-analysis-of-mechanical-properties-of-lightweight-/20034128
main-content ## Weitere Artikel dieser Ausgabe durch Wischen aufrufen Erschienen in: Open Access 01.12.2022 | Research # Experiment and Analysis of Mechanical Properties of Lightweight Concrete Prefabricated Building Structure Beams verfasst von: Yingguang Fang, Yafei Xu, Renguo Gu print DRUCKEN insite SUCHEN ## Abstract Recent years have witnessed that the prefabricated concrete structure is in the widespread use of building structures. This structure, however, still has some weaknesses, such as excessive weight of components, high requirements for construction equipment, difficult alignment of nodes, and poor installation accuracy. In order to handle the problems mentioned above, the prefabricated component made of lightweight concrete is adopted. At the same time, this prefabricated component is beneficial to reducing the load of the building structure itself and improving the safety and economy of the building structure. Nevertheless, it is rarely found that the researches and applications of lightweight concrete for stressed members are conducted. In this context, this paper replaces ordinary coarse aggregate with lightweight ceramsite or foam based on the C60 concrete mix ratio so as to obtain a mix ratio of C40 lightweight concrete that meets the engineering standards. Besides, ceramsite concrete beams and foamed concrete beams are fabricated. Moreover, through three-point bending tests, this paper further explores the mechanical properties of lightweight concrete beams and plain concrete beams during normal use conditions. As demonstrated in the results, the mechanical properties of the foamed concrete beam are similar to those of the plain concrete beam. Compared to plain concrete beams, the density of foamed concrete beams was lower by 23.4%; moreover, the ductility and toughness of foamed concrete were higher by 13% and 3%, respectively. However, in comparison with the plain concrete beam, the mechanical properties of the ceramsite concrete beam have some differences, with relatively large dispersion and obvious brittle failure characteristics. Moreover, in consideration of the nonlinear deformation characteristics of reinforced concrete beams, the theoretical calculation value of beam deflection was given in this paper based on the assumption of flat section and the principle of virtual work. The theoretically calculated deflection values of ordinary concrete beams and foamed concrete beams are in good agreement with the experimental values under normal use conditions, verifying the rationality and effectiveness of the calculation method. The research results of this paper can be taken as a reference for similar engineering designs. Hinweise ## Publisher's Note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. ## 1 Introduction The prefabricated concrete structure is widely used in building structures. Compared with the traditional cast-in-place structure, the prefabricated concrete structure has diverse advantages, such as short construction periods, high production efficiency, less material consumption, high quality of finished products, low carbon, and environmental protection (Liu et al., 2020; Shah et al., 2021). This structure, however, also has the disadvantage of weak structural integrity (Huang et al., 2021; Savoia et al., 2017), since it is difficult to guarantee the quality of the node construction of prefabricated components. To be specific, the narrow construction environment, high requirements for construction equipment due to excessive quality of components, the difficulty in the alignment of nodes, and the poor installation accuracy are the key factors affecting the quality of node connections (Chen et al., 2017; Nguyen & Hong, 2020). The adoption of lightweight concrete to reduce the weight of prefabricated components is helpful to solve the above problems. At the same time, this prefabricated component is beneficial to reducing the load of the building structure itself and improving the safety and economy of the building structure. Currently, there are a lot of researches on lightweight concrete, but many lightweight concretes are mainly adopted for functional components. Many scholars have obtained lightweight concrete with significantly improved heat insulation and sound absorption by replacing the aggregates of plain concrete. In recent years, with the deepening of researches, it is promising to utilize some lightweight concrete in structural stress members (Kozłowski & Kadela, 2018; Lee, Kang, et al., 2018; Yang et al., 2016). Many studies explored that the types and contents of lightweight aggregate additions have a great influence on their mechanical properties (Hamidian & Shafigh, 2021; Karamloo et al., 2020; Tian et al., 2020; Vakhshouri & Nejadi, 2018). Therefore, researchers have made a lot of efforts to find the lightweight aggregate concrete that is expected to be used in stressed components. Although the prefabricated beam functions as one of the main load-bearing components of the prefabricated concrete structure, the research on prefabricated beams made of lightweight concrete is not enough. Lee, Lim, et al. (2018) used lightweight foamed mortar with a 28-day compressive strength of 20 MPa to make reinforced concrete beams, and then conducted bending tests on the beams. Based on the research results, the ultimate load of the reinforced lightweight foam mortar beam was about 8–34% lower than that of plain reinforced concrete with the same steel configuration. However, Jones and McCarthy (2005) once pointed out that most engineers and designers were unlikely to pay much attention to the structural application of foamed concrete unless the strength of foamed concrete exceeds 25 Mpa. Therefore, Lim (2007) conducted bending tests on reinforced foam concrete beams made from 20 to 35 MPa, respectively, and found that both foam concrete beams and ordinary concrete beams showed bending failure models and similar ultimate loads. At the same time, further researches are called upon for beams made of foamed concrete with a compressive strength of 35 Mpa or more. In addition, ceramsite concrete is a type of lightweight concrete that is expected to be used for force members. In recent years, some scholars have studied the bearing capacity and crack width of ceramsite concrete beams. For example, Chen, Li, et al. (2020), Chen, Hui, et al. (2020)) explored the failure mode of shale ceramsite lightweight aggregate concrete beams and the width of diagonal cracks. Moreover, Liu et al (2021) made detailed analysis on the bearing capacity of H-shaped steel beams with circular holes on the webs wrapped in ceramsite concrete (SBWCC) and further proposed a short-term stiffness formula. According to the above researches, it can be found that in the past, the researches on lightweight concrete mainly focused on ultimate strength and ductility. However, flexural members, like beams, should not only have enough strength and ductility but also should meet the service limit state, such as crack width, vibration, and deflection. (Jahami et al., 2019; Wang & Tan, 2021). Nowadays, there are rare studies on the mechanical properties of foamed concrete beams and ceramsite beams with C40 strength grade concrete under normal use conditions. In particular, the mechanical differences between these two types of lightweight concrete beams and plain concrete beams, and the calculation method of the deflection of lightweight concrete beams are rarely explored in the previous studies. Based on the C60 concrete mix ratio, this study uses lightweight ceramsite or foam material to replace ordinary coarse aggregates so as to obtain a C40 light concrete mix ratio that meets the engineering standards. Besides, ceramsite concrete beams and foamed concrete beams are fabricated. Moreover, through three-point bending tests, this paper further explores the mechanical properties of lightweight concrete beams and plain concrete beams during normal use conditions. Then, the theoretical calculation method of the deflection of the foamed concrete beam is proposed in this paper based on the assumption of flat section and the principle of virtual work. The research results of this paper will help further understand the mechanical properties of lightweight concrete and promote the structural application of lightweight concrete. ## 2 Experimental Model ### 2.1 Test Materials and Mix Ratio This paper intends to explore the mechanical properties of plain concrete, ceramsite concrete, and foamed concrete. Considering that there are many factors that affect the strength of concrete beams, such as the amount of cement, the amount of foam, the amount of ceramsite, the curing conditions, and the water–cement ratio, the compressive strength and flexural strength of the specimens of different materials should meet the minimum engineering requirements in order to better enable the test beam to have sufficient strength to meet engineering requirements. This paper replaces the high-strength concrete with lightweight aggregates to obtain the ratio of C40 lightweight concrete beams, and then study the mechanical properties of different types of beams. To be specific, referring to the related literature (Chen, Hui, et al., 2020; Chen, Li, et al., 2020; Elrahman Abd, Chung, et al., 2019; Elrahman Abd, Chung, et al., 2019; Gong et al., 2018; Lee et al., 2019; Lotfy et al., 2015; Yu et al., 2013), the author replaced the high-strength C60 concrete coarse aggregate with light aggregates and made standard specimens of compressive and flexural strengths, whose strengths were tested accordingly. During the whole experiment, Yuexiu brand P·II52.5R cement was used to make lightweight concrete; S95 slag powder with activity index greater than 95% was used for mineral powder, and river sand with a particle size of 2.36 mm or less for fine aggregate The coarse aggregate adopts gravel piles with a bulk density of 1520 kg/m3 and a particle size of 15 mm or less. In addition, ceramsite with a density of 618 kg/m3, the cylinder strength of 1.8 Mpa, and the particle size of 8 mm–15 mm and a concentrated high-efficiency cement foaming agent were adopted. In this paper, the types of longitudinally stressed steel bars and stirrup steel bars in the specimens are all HRB500 and HPB300, respectively. The mechanical properties of these steel bars are shown in Table 1. Table 1 Mechanical properties of the rebars. Specimens fy (MPa) fu (MPa) Es (MPa) HRB500 500 630 2 × 105 HPB300 300 420 2 × 105 According to Chinese engineering requirements, the 28d compressive strength of the standard specimens is not less than 40 MPa, and the flexural strength 4.4 MPa. After various trials and tests, a lightweight ceramsite concrete with a 28d compressive strength characteristic value of 41 MPa and a flexural strength characteristic value of 6.62 MPa, and the foamed concrete with 28d compressive strength characteristic value of 41.4 MPa and the characteristic value of flexural strength of 12.97 MPa were prepared. The information of concrete mix ratio is shown in Table 2. The density of plain concrete, the density of foamed concrete, and the density of ceramsite concrete are 2480 kg/m3, 1900 kg/m3, and 2000 kg/m3 respectively. Compared with the density of plain concrete, the density of the two types of lightweight concrete has been reduced by 23.4% and 19.4%, respectively. Table 2 Concrete mix ratio. Category Plain concrete Per m3 Ceramsite concrete Per m3 Foamed concrete Per m3 Cement (kg) 434.0 (18.1%) 434.0 (26.1%) 557.3 (27.9%) Mineral powder (kg) 144.7 (6.0%) 144.7 (8.7%) 268.3 (13.4%) Sand (kg) 700.4 (29.2%) 700.4 (42.2%) 990.7 (49.5%) Water reducing agent (kg) 15.0 (0.6%) 15.0 (0.9%) 18.6 (0.9%) Water (kg) 138.8 (5.8%) 138.8 (8.4%) 165.1 (8.3%) Gravel (kg) 967.2 (40.3%) Ceramsite (kg) 228.5 (13.8%) Foam (L) 516 ### 2.2 Experimental Design Considering that T beams have high flexural and shear resistance (Gulec et al., 2021), the cross section of the test beam is designed as a T-section. A total of 5 simply supported beam members were designed and fabricated in the experiment, including 1 plain concrete beam––C, 2 ceramsite concrete beams––CC1 and CC2 (CC1 and CC2 are parallel samples), and 2 foamed concrete beams––FC1 and FC2 (FC1 and FC2 are parallel samples). Since the mid-span bending moment is the largest in the whole beam, the concrete cracks in the tension zone here are most serious and with the fastest speed in the whole beam. Since the strain gage attached to the middle span is prone to fracture, it is specially glued to the 1/4 beam length on the right in order to perform better. The thickness of the concrete protective layer is 30 mm. According to the Chinese standard, the minimum reinforcement ratio of tensile steel bars is 0.2%, and the reinforcement ratio of tensile steel bars should not exceed 2.5%. In this article, the reinforcement ratio of tension steel bars is 1.9%, and the reinforcement ratio of compression steel bars is 0.4%. The measuring points and reinforcement of the beam are displayed in Fig. 1. When the preloading is completed, the formal loading is carried out accordingly which is still controlled by load, with a step of 50 kN as the step length. After completing the loading of each level, the load is held for 10 min, and then the corresponding data of each measuring point under the test load of this level are recorded. After loading up to 200 kN, the loading step size is changed to 25 kN so as to better control the sudden occurrence of brittle failure and observe the deformation process of the component in detail. Apart from having enough strength and ductility, beams should also meet the service limit state (Wang & Tan, 2021). In this context, this test took the service limit into consideration. According to the relevant Chinese standards, the deflection of the flexural member during its service period shall meet a certain limit, which is 1/500 of the calculation span. Therefore, this paper selects the load termination value on the basis of the normal service limit state of the beam. 250 kN was taken as the load termination value by referring to the recommended deflection formula of the Chinese code and the pre-experiments, which means the test ends when the load reaches 250 kN or the ultimate strength of the specimen. Fig. 2 shows the loaded state of specimens. ## 3 Experimental Phenomena and Results ### 3.1 Experimental Phenomena The deformation characteristics of foamed concrete and plain concrete are similar throughout the load. When the load reaches about 100 kN, 1–3 vertical cracks in the area near the loading point can be found in the beam. As the load grows, the vertical cracks continue to increase and extend along the height direction and the neutral axis moves upwards. As the load continues, individual oblique cracks begin to appear at both ends of the beam, and the beam's deflection increases significantly. When continuing the loading, the number of cracks at both ends of the beam shows an upward trend. When the ceramsite concrete beam is loaded to 80 kN, the first vertical crack appears in the middle of the span, and the "click" sound of concrete cracks will be heard. With the continuous loading, it can be found that the number of cracks is more than that of plain concrete beams, and the cracks develop faster. When re-loaded, an oblique crack along the support to the loading point gradually develops. As shown in Fig. 3, when the loading reaches about 225 kN, the specimen suddenly breaks, demonstrating obvious characteristics of shear brittle failure. ### 3.2 Experimental Results Fig. 4 shows the load–deflection curve of the specimen. It can be found that at the initial stage of loading, the deflection of each concrete beam increases linearly with the load. With the increase of the load, the beam gradually cracks, the load deflection curve of plain concrete and foamed concrete gradually deviates from the straight line, and the stiffness of the beam decreases. The load deflection data of the two beams of ceramsite concrete are discrete, with the strength significantly lower than that of plain concrete. Since the density of ceramsite (about 600 kg/m3) is much smaller than that of cement paste (about 1500 kg/m3), ceramsite tends to float up during concrete solidification, resulting in uneven distribution of ceramsite in concrete. The uneven distribution in concrete, the discreteness of ceramsite strength, and the complexity and randomness of ceramsite interface bonding may lead to the obvious difference of the load deflection data of CC1 and CC2. ## 4 Calculation of Beam Deflection Under the load, the section bending moment of the beam member varies along the axis, and the average stiffness or curvature of the corresponding section changes in a complicated way, which is the main reason for accurately calculating the deformation of reinforced concrete members. The direct bilinear method, effective inertia method, and curvature integral method are mainly adopted for the calculation of beam deflection. For example, China's GB 50010-2002 "Concrete Structure Design" adopts the direct bilinear method to calculate the short-term stiffness of components that allow cracks. The American standard ACI 318-99 stipulates that the stiffness calculation after cracking adopts the effective moment of inertia method. In this paper, considering the nonlinear deformation of reinforced concrete beams, the virtual work principle is used to calculate the deflection, and the related calculations are completed with the help of commercial software matlab. ### 4.1 Basic Assumptions 1) The average strain distribution conforms to the plane section assumption, that is, the average strain of the section is linearly distributed along the height. 2) There is no bond slip between longitudinally tensioned steel bars and concrete. The stress–strain of the longitudinal steel bars adopts an ideal elastoplastic model, and the expression is $$\left\{ {\begin{array}{*{20}c} {\sigma_{s} = \varepsilon_{s} E_{{\text{s}}} ,\varepsilon_{s} \le \varepsilon_{{\text{y}}} \, } \\ {{\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} \sigma_{s} = f_{y} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} ,\varepsilon_{y} < \varepsilon_{s} \le \varepsilon_{su} } \\ \end{array} } \right.,$$ (1) where $$\sigma_{s}$$ is the steel bar stress, $$E_{{\text{s}}}$$ is the steel bar elastic modulus, $$\varepsilon_{{\text{y}}}$$ is the steel bar yield strain, and $$f_{y}$$ is the steel bar yield strength design value. 3) The constitutive models of plain concrete, ceramsite concrete, and foamed concrete are selected with reference to relevant specifications, without considering the tensile effect of concrete. The constitutive model of plain concrete adopts the formula recommended by relevant Chinese standards, and its expression is as follows: $$\left\{ {\begin{array}{*{20}c} {\sigma_{c} = f_{c} [1 - (1 - \frac{{\varepsilon_{c} }}{{\varepsilon_{0} }})^{n} ],\varepsilon_{c} \le \varepsilon_{{0}} \, } \\ {{\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} \sigma_{c} = f_{c} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} ,\varepsilon_{0} < \varepsilon_{{\text{c}}} \le \varepsilon_{cu} } \\ \end{array} } \right.,$$ (2) where $${\kern 1pt} {\kern 1pt} {\kern 1pt} \sigma_{c}$$ is the concrete stress, $$f_{c} {\kern 1pt}$$ is the design value of the concrete compressive strength, $$\varepsilon_{0}$$ is the concrete compressive strain when the concrete compressive stress reaches $$f_{c} {\kern 1pt}$$,$$\varepsilon_{cu}$$ is the ultimate compressive strain of the normal section concrete, and n is the coefficient. In this paper, for C40 concrete, $$\varepsilon_{0}$$, $$\varepsilon_{cu}$$, and n are 0.002, 0.0033, and 2.0 respectively. The constitutive model expressions of ceramsite concrete and foamed concrete are as follows: $$\left\{ {\sigma_{c} = \begin{array}{*{20}c} {f_{c} [1.5(\frac{{\varepsilon_{c} }}{{\varepsilon_{0} }}) - 0.5(\frac{{\varepsilon_{c} }}{{\varepsilon_{0} }})^{2} ],} & {\varepsilon_{c} \le \varepsilon_{0} } \\ {\sigma_{c} = f,} & {\varepsilon_{c} > \varepsilon_{0} } \\ \end{array} } \right.$$ (3) 4) The concrete deformation is considered to be continuous without considering cracks, satisfying the principle of virtual work. ### 4.2 Bending Analysis and Deflection Calculation of Normal Section According to the basic assumption of the flat section assumption, the section stress and the strain distribution are shown in Fig. 5 when the properly reinforced beam is in normal service. Assuming that the height of the compression zone of the section is xc, the strain at the distance y from the neutral axis of the section can be calculated by the assumption of the flat section as follows: $$\varepsilon (y{) = }\frac{y}{\rho }.$$ (4) In the formula,$$\rho$$ is the radius of curvature, and y is the coordinate with the neutral axis as the origin. According to the force balance condition of the cross section, the following two balance equations can be summarized: $$\sum {X = 0,\int_{0}^{{{\text{y}}_{c} }} {\sigma_{c} } } (y) \cdot bdy{ + }P_{{1}} { + }T_{1} + T_{2} { = }0,$$ (5) $$\sum {M = 0,M_{u} { + }\int_{0}^{{{\text{y}}_{c} }} {\sigma_{c} } } (y) \cdot b \cdot ydy + T_{1} y_{t1} + T_{2} 2y_{t2} + P_{1} y_{p1} { = }0,$$ (6) Taking the neutral axis as the origin of the ordinate, upward is positive, and downward is negative; yt1, yt2, and yp1, respectively, represent the ordinate of the resultant force of the first-row tensioned steel bars, that of the second-row tensioned steel bars, and that of compressed steel bars. From Eqs. (16), the relationship between curvature and bending moment can be obtained, and then the curve of deflection and load can be calculated based on the principle of virtual work in Eq. (7). With the increase of load, the neutral axis of the concrete will gradually move up, so the section effective moment of inertia Ie of the concrete will change accordingly, which can be determined by Eq. 8. When using the component, the deformation caused by the axial force and the shear force is negligible. Based on the mathematics software matlab for related programming, this paper calculates the theoretical calculation deflection of each beam by adopting the numerical integration method. In order to ensure sufficient accuracy and calculation speed, the integration step length is selected as 2 mm after multiple debugging. $$\Delta = \sum {\int {\frac{{\overline{M}M_{P} }}{EIM}} } {\text{ds ,}}$$ (7) $$I_{{\text{e}}} { = }\frac{{M_{P} \rho }}{E},$$ (8) where $$\Delta$$ is the deflection; $$\overline{M}$$ is the bending moment of the unit load on the virtual beam. Mp is the bending moment of section, and $$\rho (M)$$ is the corresponding curvature. E is the elastic modulus of reinforced concrete, and Ie is the effective moment of inertia of the section. ## 5 Results and Analysis The calculation results are shown in Fig. 6. It can be found that the theoretical calculation values of plain concrete and foamed concrete are in good agreement with the experimental values, the absolute error is mostly within 0.2 mm, and the relative error is mostly 10%–20% during the normal service period. The cracks of the two beams are smaller during the normal service period, which can be calculated based on the virtual work principle with little error. Due to the discreteness of the strength of the ceramsite itself and the uneven distribution of the ceramsite in the concrete, the deflection experimental data of the ceramsite concrete beam are discrete. The crack of ceramsite concrete beam develops rapidly with a large number under the load, and thus, there will be a large error in calculation when using the assumption of plane section and the principle of virtual work. In order to further verify whether the ceramsite concrete satisfies the plane section assumption, this test measured the concrete strain of the beam under different loads by using strain gages pasted on the side of the ceramsite concrete beam, and further calculated the average strain of the section, so as to obtain the average strain distribution curve of the section. The degree of compliance with the plain section assumption can be found from the curve. Fig. 7 demonstrates the strain distribution of the ceramsite concrete section, in which the strain point at 150 kN and with the section height of 47 mm fails to collect data at this point due to the quality of the strain gage. In addition, the dotted line connection is adopted for the data in this paper. Under lower loads (50 kN, 100 kN), the strain distribution of the cross section of the beam is basically straight at the elastic stage. As the load increases, the section strain begins to deviate from the straight line when the load reaches 150 kN. The section strain is far from the straight line. ### 5.2 Bending Stiffness In order to analyze the changes in the bending stiffness of the beam throughout the loading period, the following equation is used for the analysis of the section bending stiffness: $$B_{s} { = }\frac{{FL^{3} }}{48\Delta }.$$ (9) As shown in Fig. 8, the bending stiffness of foamed concrete and plain concrete gradually decreases as the load increases. The ceramsite concrete beam has comparatively large discreteness because of the uneven strength of ceramsite, uneven distribution of ceramsite in the beam, and the complexity and randomness of the ceramsite interface and the combination of colloidal materials. Thus, its bending stiffness fluctuates greatly as the load increases. Therefore, through the stiffness trend line of ceramsite concrete beams, the overall change trend of stiffness can be better observed, indicating that the overall trend of stiffness decreases as the load increases. The deflection of the ceramsite concrete beam decreases rapidly at 100 kN, and the stiffness also plummets. The possible reason is that the ceramsite concrete has some slight cracks during preloading. The bending rigidity of foamed concrete will decrease with the increase of the load throughout the service period. Table 3 illustrates the bending stiffness of various concrete beams when they reach the normal service limit. When the beam reaches the normal service limit, the bending stiffness of the foamed concrete beam changes by about 10%, and its density is about 23.4% lighter than plain concrete, proving that it is a very good lightweight concrete. The material properties and uneven distribution of ceramsite concrete have a greater impact on the test results, so further research is needed. Table 3 Limit bending stiffness. Plain concrete Foamed concrete Ceramsite concrete Deflection limit (mm) 3.16 3.16 3.16 The corresponding load of deflection limit(kN) 216.3 201.2 222.9 Bending stiffness (106 N·m2) 6.52 5.88 6.75 Initial bending stiffness (106 N·m2) 6.78 6.67 7.31 Rate of change 3.83% 11.84% 7.70% The initial bending stiffness represents the bending stiffness of 50kN. ### 5.3 Failure Analysis As shown in Table 4, in the experiment, the foamed concrete beam failed due to the deflection reaching the service limit, and the ceramsite concrete beam suddenly broke with the development of the shear crack. The ceramsite concrete suffers brittle shear failure at 225 kN, demonstrating that its bearing capacity is significantly lower than that of plain concrete. The strength of the coarse aggregate in plain concrete is higher than the strength of the interface between the cement base stone and the coarse aggregate (Xiao et al., 2013), so the failure of concrete generally starts from the interface. However, the strength of ceramsite is relatively lower than that of the cement base and the interface, so the failure of the ceramsite concrete beam is caused by the cracking of the coarse aggregate of the ceramsite concrete beam until it penetrates the entire oblique section. The phenomena are consistent with the opinion proposed by Zhang and Gjvorv (1991) that aggregate strength is the main factor affecting the strength of lightweight aggregate concrete. Table 4 Experimental results of the lightweight beam Foamed concrete beam Ceramsite concrete beam μr 1.13 0.94 Energy absorption ratio 1.03 0.92 Failure mode Service limit Shear crack ### 5.4 Displacement Ductility and Energy Absorption Ductility is often used to represent a structure's ability to resist inelastic behavior. Based on the previous literature, there are a variety of calculation models for calculating displacement ductility factors. Park mode (Gulec et al., 2021; Park., 1988) is a commonly used model for calculating ductility factor, which is the ratio of failure point displacement to yield point displacement (Khatib et al., 2020). By referring to the Park model and the service limit requirements of the beam, this paper defines the displacement factor (μ) as the ratio of the service displacement value (Δu) of the beam to the corresponding equivalent elastoplastic yield point displacement (Δy) (Eq. 10). In order to compare the ductility of the lightweight concrete beams and plain concrete beams, the relative ductility ratio ur is further defined as Eq. 11. $$\mu = \frac{{\Delta_{u} }}{{\Delta_{y} }},$$ (10) $$\mu_{r} = \frac{{\mu_{l} }}{{\mu_{p} }},$$ (11) where μl and μp, respectively, represent the ductility factor of lightweight concrete beams and that of the plain concrete beams. The average values of the relative ductility ratio of these two kinds of lightweight concrete beams are shown in Table 4. As shown in the table, compared with the plain concrete beam, the ductility of foamed concrete beams is 1.13 times of the plain concrete beam’s ductility, demonstrating that foamed concrete beam has higher ductility. However, the ductility of the ceramsite concrete beam is slightly lower than that of the plain concrete beam, which indicates that replacing coarse aggregate with ceramsite will reduce the quality of the beam but lead to a reduction in the ductility of the beam. The energy absorption capacity of the beam can be used to reflect the resistance ability to inelastic deformation. The energy absorption capacity of the beam can be obtained by calculating the area under the force–displacement curve (Gulec et al., 2021), which can be obtained by the sum of the area of two successive points (Eq. 12): $${\text{Energy absorption = 0}}{.5}\sum\limits_{i = 1}^{{n{ - }1}} {{(}d_{{i{ + 1}}} { - }d_{{\text{i}}} {)(}F_{{i{ + 1}}} { - }F_{i} {)}} ,$$ (12) where di represents the displacement; Fi represents the load (kN) at this displacement (mm); n represents the number of displacement points. The energy absorption ratio is defined as the ratio of the absorbed energy of lightweight concrete beams to that of plain concrete to compare the energy absorption of lightweight concrete beams and plain concrete beams. The average energy absorption of lightweight concrete beams is shown in Table 4. It can be found that the energy absorption capacity of the ceramsite concrete beam is the lowest in all beams, while the energy absorption capacity of the foam concrete beam is 3% higher than that of the plain concrete beam. ## 6 Conclusion The paper obtains the mix ratio of lightweight concrete by replacing the coarse aggregate of high-strength concrete with lightweight ceramsite or foam based on the C60 concrete mix ratio. After that, lightweight concrete beams are fabricated and three-point bending tests are carried out. Through the experiments and theoretical analysis, the following conclusions can be drawn from the results of this study. (1) The mechanical properties of C40 foam concrete beams are similar to those of plain concrete beams. Compared to plain concrete beams, the density of foamed concrete was lower by 23.4%; moreover, the ductility and toughness of foamed concrete were higher by 13 and 3%, respectively. (2) Considering the nonlinear deformation characteristics of reinforced concrete beams, the theoretical calculation method of beam deflection is proposed based on the flat section assumption and the principle of virtual work. Within the normal use deflection limits, the calculated results are in good agreement with the deflection of plain concrete beams and foam concrete beams, the absolute error is mostly within 0.2 mm, and the relative error is mostly 10–20% during the normal service period, verifying that the calculated value can be used as a design reference for the deflection of the foamed concrete beam during normal use. (3) The mechanical properties of C40 ceramsite concrete beam have comparatively large discreteness. This may be caused by the strength discreteness of the ceramsite, the uneven distribution of ceramsite in the concrete beam, and the complexity and randomness of the combination of ceramsite interface and colloid. Thus, the aforementioned points should be attached great importance to in subsequent research and design. ## Acknowledgements This work was financially supported by the National Key Scientific Instruments and Equipment Development Projects of China(Grant No. 41827807), the Open Research Fund of the State Key Laboratory of Geomechanics and Geotechnical Engineering, Institute of Rock and Soil Mechanics, Chinese Academy of Sciences (Grant No. Z018019), the State Key Laboratory of Subtropical Building Science in China(Grant No. 2017KB16), and Guangdong Provincial Key Laboratory of Modern Civil Engineering Technology in China (Grant No. 2021B1212040003). ## Declarations ### Competing interests The authors declare that they have no competing interests. Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://​creativecommons.​org/​licenses/​by/​4.​0/​. ## Publisher's Note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. print DRUCKEN Literatur Bawab, J., Khatib, J., Jahami, A., Elkordi, A., & Ghorbel, E. (2021). Structural Performance of Reinforced Concrete Beams Incorporating Cathode-Ray Tube (CRT) Glass Waste. Buildings, 11(2), 67. CrossRef Chen, M., Li, Z., Wu, J., et al. (2020). Shear behaviour and diagonal crack checking of shale ceramsite lightweight aggregate concrete beams with high-strength steel bars. Construction and Building Materials., 249, 118730. CrossRef Chen, Y., Hui, Q., Zhang, H., Zhu, Z., & Wang, C. (2020). Experiment and application of ceramsite concrete used to maintain roadway in coal mine. Measurement and Control., 53, 1832–1840. CrossRef Chen, Z., Liu, J., & Yu, Y. (2017). Experimental study on interior connections in modular steel buildings. Engineering Structures, 147, 625–638. CrossRef Elrahman Abd, M., Chung, S., & Stephan, D. (2019). Effect of different expanded aggregates on the properties of lightweight concrete. Magazine of Concrete Research., 71, 95–107. CrossRef Elrahman Abd, M., Madawy El, M., Chung, S., Sikora, P., & Stephan, D. (2019). Preparation and characterization of ultra-lightweight foamed concrete incorporating lightweight aggregates. Applied Sciences., 9, 1447. CrossRef Gong, J., Zeng, W., & Zhang, W. (2018). Influence of shrinkage-reducing agent and polypropylene fiber on shrinkage of ceramsite concrete. Construction and Building Materials., 159, 155–163. CrossRef Gulec, A., Kose, M. M., & Gogus, M. T. (2021). Experimental investigation of flexural performance of T-section prefabricated cage reinforced beams with self-compacting concrete. Structures. https://​doi.​org/​10.​1016/​j.​istruc.​2021.​05.​074 CrossRef Hamidian, M. R., & Shafigh, P. (2021). Post-peak behaviour of composite column using a ductile lightweight aggregate concrete. International Journal of Concrete Structures and Materials. https://​doi.​org/​10.​1186/​s40069-020-00453-6 CrossRef Huang, W., Hu, G., Miao, X., & Fan, Z. (2021). Seismic performance analysis of a novel demountable precast concrete beam-column connection with multi-slit devices. Journal of Building Engineering, 44, 102663. CrossRef Jahami, A., Khatib, J., Baalbaki, O., & Sonebi, M. (2019). Prediction of deflection in reinforced concrete beams containing plastic waste. SSRN J. https://​doi.​org/​10.​2139/​ssrn.​3510113 CrossRef Jones, M. R., & McCarthy, A. (2005). Preliminary views on the potential of foamed concrete as a structural material. Magazine of Concrete Research, 57(1), 21–31. CrossRef Karamloo, M., Afzali-Naniz, O., & Doostmohamadi, A. (2020). Impact of using different amounts of polyolefin macro fibers on fracture behavior, size effect, and mechanical properties of self-compacting lightweight concrete. Construction and Building Materials., 250, 118856. CrossRef Khatib, J. M., Jahami, A., Elkordi, A., Abdelgader, H., & Sonebi, M. (2020). Structural assessment of reinforced concrete beams incorporating waste plastic straws. Environments, 7(11), 96. CrossRef Kozłowski, M., & Kadela, M. (2018). Mechanical characterization of lightweight foamed concrete. Advances in Materials Science &amp; Engineering., 2018, 1–8. CrossRef Kyriakopoulos, P., Peltonen, S., Vayas, I., Spyrakos, C., & Leskela, M. V. (2021). Experimental and numerical investigation of the flexural behavior of shallow floor composite beams. Engineering Structures, 231, 111734. CrossRef Lee, J., Kang, S., Ha, Y., & Hong, S. (2018). Structural behavior of durable composite sandwich panels with high performance expanded polystyrene concrete. International Journal of Concrete Structures and Materials. https://​doi.​org/​10.​1186/​s40069-018-0255-6 CrossRef Lee, Y. L., Lim, J. H., Lim, S. K., & Tan, C. S. (2018). Flexural behaviour of reinforced lightweight foamed mortar beams and slabs. KSCE Journal of Civil Engineering, 22(8), 2880–2889. CrossRef Lee, K., Yang, K., Mun, J., & Tuan, N. (2019). Effect of sand content on the workability and mechanical properties of concrete using bottom ash and dredged soil-based artificial lightweight aggregates. International Journal of Concrete Structures and Materials. https://​doi.​org/​10.​1186/​s40069-018-0306-z CrossRef Lim, H. (2007). Structural response of LWC beams in flexure. National University of Singapore. Liu, X., Meng, K., Zhang, A., Zhu, T., & Yu, C. (2021). Bearing capacity of H-section beam wrapped with ceramsite concrete. Steel and Composite Structures, 40(5), 679–696. Liu, Y., Ma, H., Li, Z., & Wang, W. (2020). Seismic behaviour of full-scale prefabricated RC beam–CFST column joints connected by reinforcement coupling sleeves. Structures. https://​doi.​org/​10.​1016/​j.​istruc.​2020.​10.​066 CrossRef Lotfy, A., Hossain, K. M. A., & Lachemi, M. (2015). Lightweight self-consolidating concrete with expanded shale aggregates, modelling and optimization. International Journal of Concrete Structures and Materials., 9, 185–206. CrossRef Nguyen, D. H., & Hong, W. K. (2020). A novel erection technique of the L-shaped precast frames utilizing laminated metal plates. Journal of Asian Architecture and Building Engineering. https://​doi.​org/​10.​1080/​13467581.​2020.​1841644 CrossRef Park, R. (1988) Ductility evaluation from laboratory and analytical testing. Proceedings of the 9th world conference on earthquake engineering, Tokyo-Kyoto, Japan. Vol. 8. Porter, J. H., Cain, T. M., Fox, S. L., & Harvey, P. S. (2019). Influence of infill properties on flexural rigidity of 3D-printed structural members. Virtual and Physical Prototyping, 14(2), 148–159. CrossRef Savoia, M., Buratti, N., & Vincenzi, L. (2017). Damage and collapses in industrial precast buildings after the 2012 Emilia earthquake. Engineering Structures, 137, 162–180. CrossRef Shah, Y. I., Hu, Z., Yin, B. S., & Li, X. (2021). Flexural performance analysis of UHPC wet joint of prefabricated bridge deck. Arabian Journal for Science and Engineering. https://​doi.​org/​10.​1007/​s13369-021-05735-z CrossRef Tian, Y., Yan, X., Zhang, M., Yang, T., & Zhang, J. (2020). Effect of the characteristics of lightweight aggregates presaturated polymer emulsion on the mechanical and damping properties of concrete. Construction and Building Materials., 253, 119154. CrossRef Vakhshouri, B., & Nejadi, S. (2018). Size effect and age factor in mechanical properties of BST Light Weight Concrete. Construction and Building Materials, 177, 63–71. CrossRef Wang, S., & Tan, K. H. (2021). Flexural performance of reinforced carbon nanofibers enhanced lightweight cementitious composite (CNF-LCC) beams. Engineering Structures, 238, 112221. CrossRef Xiao, J., Li, W., Corr, D. J., & Shah, S. P. (2013). Effects of interfacial transition zones on the stress–strain behavior of modeled recycled aggregate concrete. Cement and Concrete Research, 52, 82–99. CrossRef Xie, J., Huang, L., Guo, Y., Li, Z., Fang, C., Li, L., & Wang, J. (2018). Experimental study on the compressive and flexural behaviour of recycled aggregate concrete modified with silica fume and fibres. Construction and Building Materials, 178, 612–623. CrossRef Yang, K. H., Lee, Y., & Joo, D. B. (2016). Flexural behavior of post-tensioned lightweight concrete continuous one-way slabs. International Journal of Concrete Structures and Materials, 10(4), 425–434. CrossRef Yu, X. Q., Lin, M., Geng, G. L., Li, Y. H., & Jia, L. (2013). Preparation test study on High strength lightweight concrete in engineering materials. Advanced Materials Research. https://​doi.​org/​10.​4028/​www.​scientific.​net/​AMR.​648.​55 CrossRef Zhang, M. H., & Gjvorv, O. E. (1991). Mechanical properties of high-strength lightweight concrete. Materials Journal, 88(3), 240–247. Titel Experiment and Analysis of Mechanical Properties of Lightweight Concrete Prefabricated Building Structure Beams verfasst von Yingguang Fang Yafei Xu Renguo Gu Publikationsdatum 01.12.2022 Verlag Springer Singapore Erschienen in International Journal of Concrete Structures and Materials / Ausgabe 1/2022 Print ISSN: 1976-0485 Elektronische ISSN: 2234-1315 DOI https://doi.org/10.1186/s40069-021-00496-3 Zur Ausgabe
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 2, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8113915324211121, "perplexity": 3124.300782405425}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-33/segments/1659882571911.5/warc/CC-MAIN-20220813081639-20220813111639-00317.warc.gz"}
https://brilliant.org/weekly-problems/2018-04-09/advanced/
# Problems of the Week Contribute a problem The parks service would like to construct a path that starts at point A and revolves around the mountain once before ending at point B. Which describes the uphill/downhill nature of the shortest possible path? Note: Point A is at the base of a mountain in the shape of a right circular cone. Point B is located on the segment between point A and the summit. Let $f,g : \mathbb{R} \longrightarrow \mathbb{R}$ such that $f(2x)=2f(x)$ and $g(x) = x+ f\big(g(x)\big).$ If $2g(x)$ is in the range of $g$ for all real $x,$ then is it always true that $g(2x)=2g(x)?$ Note: $f$ and $g$ are not necessarily continuous over $\mathbb{R}.$ The rate of a simple chemical reaction like $\ce{A + 2B->C}$ depends on the input concentrations in a simple way. Here, for example, the output species $\text{C}$ is made with a rate proportional to $\left[\text{A}\right]\cdot\left[\text{B}\right]^2$ and, in general, the rates of one step reactions depend on the input concentrations raised to the power of their stoichiometric coefficient. But this isn't always the case. Chain reactions occur when the output from one reaction $\mathcal{R}_i$ is the input to another $\mathcal{R}_j$ as in the following system of reactions: $\begin{array}{rrr} \text{A} &\longleftrightarrow& \text{2B} \\ \text{B} &\longleftrightarrow& \text{3C} \\ \text{C} &\longleftrightarrow& \text{4D} \\ \text{D} &\longrightarrow& \text{E}. \end{array}$ It's a lot less simple to find reaction order in a case like this, but as the concentration of $\text{A}$ rises, a simple behavior emerges: in the asymptotic limit where $\left[\text{A}\right]$ gets large, the rate of production of the output $\left[\text{E}\right]$ is approximately equal to $\text{(a constant)}\times\left[\text{A}\right]^\alpha.$ What is $\alpha?$ Assume that fresh $\text{A}$ is supplied continuously to the system so that its concentration is always $\left[\text{A}_0\right],$ and that the conversion of $\text{D}$ to $\text{E}$ is irreversible. For simplicity, assume all the magnitudes of all rate constants are equal to $1.$ $\begin{array}{lrr} &&A & B \\ -&&C& D \\ \hline \\ &&E & F \\ + &&G& H\\ \hline \\ &P&P& P \end{array}$ Each of the letters $A, B, C, D, E, F, G, H, P$ represents a distinct digit from 1 to 9. What is $A+B+C+D?$ Three points ${\color{#3D99F6}A}, {\color{#D61F06}B},$ and ${\color{#20A900}C}$ are moving counterclockwise around a unit circle with periods $\SI{1}{\second}, \SI{2}{\second},$ and $\SI{3}{\second},$ respectively. They started from the same position on the circle. What is the maximal area of triangle ${\color{#3D99F6}A}{\color{#D61F06}B}{\color{#20A900}C}?$ ×
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 38, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7911342978477478, "perplexity": 186.30199865026518}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243989819.92/warc/CC-MAIN-20210518094809-20210518124809-00169.warc.gz"}
https://electronics.stackexchange.com/questions/45456/transfer-function-and-bode-plot-from-poles-and-zeroes
# Transfer Function and Bode Plot from Poles and Zeroes I've got the following bode plot from a black box: I've calculated the zero to be 1055 and pole to be 67. Thus I'm using the transfer function $H(s) = (s-1055)/(s-67)$ but this gives a clearly wrong bode plot: WolframAlpha calculation I'm guessing my transfer function is wrong. Can anyone see how? The plot from Alpha looks pretty close to your "black box" measurement to me. The only difference is a pre-factor. Try $\frac{67}{1055}\frac{s-1055}{s-67}$: First, note that that the transfer function you give has the pole (and zero too) in the right hand plane, i.e., the system it describes is unstable. I suspect you meant: $H(s) = \dfrac{s + 1055}{s+67}$ However, even this is not in standard form. Putting this transfer function into standard form yields: $H(s) = \dfrac{s + 1055}{s + 67} = \dfrac{1055}{67} \dfrac{\frac{s}{1055} + 1}{\frac{s}{67} + 1}$ So, now you see where the undesired gain has come from. Knowing (only) the pole and zero, you should guess instead: $H(s) =\dfrac{\frac{s}{1055} + 1}{\frac{s}{67} + 1}$ Actually, you have half of a Bode plot from a Black Box. A Bode plot shows magnitude and phase. You cannot do the fit you're trying to do without employing phase information. More accurately, you can, but it might be difficult to interpret your results. If you had your phase, you'd have more confidence in your fit. • I was moreso worried in the gain of 24dB that my transfer function had. The phase information wasn't important and I shouldn't have included it. – Sticky Oct 31 '12 at 9:18
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9006001353263855, "perplexity": 782.9543736837717}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-35/segments/1566027315811.47/warc/CC-MAIN-20190821065413-20190821091413-00134.warc.gz"}
http://aas.org/archives/BAAS/v36n5/aas205/1244.htm
AAS 205th Meeting, 9-13 January 2005 Session 92 Ellipticals and Early Types Poster, Wednesday, January 12, 2005, 9:20am-6:30pm, Exhibit Hall ## [92.06] A Low-redshift Sample of E+A Galaxies K.M. Walker (University of Toledo), M.P. Bergmann (NOAO Gemini Science Center), J. Turner (Gemini Observatory) We present the results of a low-redshift survey for E+A galaxies. These galaxies are spectroscopically classified as having strong Balmer and metallic absorption with the absence of any emission. The absorption lines indicate an abundance of A-type stars as well as an old stellar population, while the absence of emission, especially {O~{\sc ii}}, denotes the lack of current star formation. Essential in determining the evolution of early-type galaxies, a low-redshift sample will allow easier morphology observations and further spectroscopic study. Fourteen low-redshift E+A galaxies were found using the Sloan Digital Sky Survey Data Release One following similar criteria as the H-delta strong survey of Goto (2003), except that only galaxies with a redshift lower than z=0.05 were selected. Two of these E+A galaxies were observed with the Cerro Tololo Inter-American Observatory 1.5-m telescope and confirmed to have no emission anywhere along the longslit, including outside the 3" region sampled by the SDSS fibers. This work was supported by the NSF through the CTIO REU program.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9137296676635742, "perplexity": 9455.452661837253}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-52/segments/1419447552343.99/warc/CC-MAIN-20141224185912-00093-ip-10-231-17-201.ec2.internal.warc.gz"}
http://tex.stackexchange.com/questions/7966/aligning-text-vertically-in-tikz
# Aligning text vertically in Tikz I need to know the correct way of handling vertical alignment in TIKZ. Let me elaborate an example: \documentclass{article} \usepackage{amsmath,tikz} \begin{document} \begin{tikzpicture} \node (A1) at (0,4) {($\overset{\surd}{1}$ \indent 2)}; \node (B1) at (0,3) {($\overset{\surd}{3}$ \indent 4)}; \node (A) at (2,4) {$\phi_1$} ; \node (B) at (2,3) {$\phi_2$} ; \draw [->, line width=1pt] (A) -- (A1) node [near start, above] { \footnotesize{R1} }; \node (A2) at (6,4) {$7^1$ \indent $8^1$ \indent $3^1$ \indent $4^1$\indent $\parallel$ 2}; \node (B2) at (6,3) {$1^3$ \indent $2^3$ \indent $5^3$ \indent $6^3$\indent $\parallel$ 4}; %fake caption \node at (8,4.5) {fake}; \end{tikzpicture} Which results in this: I just want to know is ther any way to force the lines into a bottom line here so that $\phi_1$ and the left would be aligned? Should I use phantom to do so or is there any thing better I am unaware of? The next point is about the word "fake" in the top right corner. I want it over the "2" but I have placed it as a node. I could use \overset{fake}{\overline{2}} to have a "fake" above the "2" but it pushes the line to the left which is not desired. And as you might have guessed I also need a line between "fake" and "2" should I use \overline{2}? - I assume you're using amsmath for the \overset. To get the alignment right, stick a [every node/.style = {anchor = base}] as an argument to {tikzpicture}, and to make the arrow from \phi work, you should change its target to (A-|A1.east) (this specifies the point which is at the intersection of a horizontal line from A and a vertical line through A1.east, which gives a nice flat arrow). For your second question, I suggest you abandon your current approach and use some TikZ libraries. In particular, you should use chains to get nodes in a line easily and with any spacing you like, and positioning to put the "fake" in the right place. Once you do that, you can use a \draw command to get the line you want. I've also used scopes to simplify the code. It allows me to write braces {...} instead of \begin{scope}...\end{scope}. (If you don't know: scopes in TikZ allow you to pass optional arguments which take effect locally.) Final code: \documentclass{article} \usepackage{amsmath,tikz} \usetikzlibrary{chains, scopes, positioning} \begin{document} \begin{tikzpicture} [every node/.style = {anchor = base}] \node (A1) at (0,4) {($\overset{\surd}{1}$ \quad 2)}; \node (B1) at (0,3) {($\overset{\surd}{3}$ \quad 4)}; \node (A) at (2,4) {$\phi_1$} ; \node (B) at (2,3) {$\phi_2$} ; \draw [->, line width=1pt] (A) -- (A-|A1.east) node [near start, above] { \footnotesize{R1} }; { [start chain, every node/.append style = on chain, node distance = 1em] \node at (4,4) {$7^1$}; \node {$8^1$}; \node {$3^1$}; \node {$4^1$}; \node {$\parallel$}; \node (two) {$2$}; } { [start chain, every node/.append style = on chain, node distance = 1em] \node at (4,3) {$1^3$}; \node {$2^3$}; \node {$5^3$}; \node {$6^3$}; \node {$\parallel$}; \node {$4$}; } \node (fake) [above = 1ex of two.north] {fake}; \draw (fake.south east) -- (fake.south west); \end{tikzpicture} \end{document} - Thank you for your answer. I update the code now to reflect a full tex document. The first problem is solved, but how about the fake keyword and line below it? Do you have any suggestion? –  Yasser Sobhdel Dec 30 '10 at 10:44 Included now, second paragraph plus updated example. –  Ryan Reich Dec 30 '10 at 11:19 @Ryan Reich: Can you please tell me how can I markup code block like that? –  user1996 Dec 30 '10 at 11:20 @an_ant: I used the scopes library. I've edited my answer to explain. –  Ryan Reich Dec 30 '10 at 11:39 @an_ant: Oh, I see. You put your code block inside a list; apparently, you need to indent it four additional spaces. You can use the preview to make sure it looks right before committing. –  Ryan Reich Dec 30 '10 at 12:12 This exact problem is discussed in pgfmanual (you can read it running texdoc pgf in bash, also) in "3 Tutorial: A Petri-Net for Hagen" (pg. 37) and in "5 Tutorial: Putting a Diagram in Chains" (pg. 56). Two placing methods discussed there are: • 3.8 Placing Nodes Using Relative Placement - works like this. You use \node (waiting) {}; \node (critical) [below of waiting] {}; \node (leave critical) [right of waiting] {}; to place the first node, and every following one in relative positioning to that 1st one, or any other node you name. • 5.3 Aligning the Nodes Using Matrices - is done this way \matrix[row sep=1mm,column sep=5mm] { % First row: & & & & \node [terminal] {+}; & \\ % Second row: \node [nonterminal] {unsigned integer}; & \node [terminal] {.}; & \node [terminal] {digit}; & \node [terminal] {E}; & & \node [nonterminal] {unsigned integer}; \\ % Third row: & & & & \node [terminal] {-}; & \\ }; You actually add nodes in matrix, like you would add values in tables - members of rows and columns are therefore aligned appropriately. In combination with row sep, column sep and leaving atoms in matrix empty you can modify spaces between particular items in matrix. - For this matter, I'd like placing nodes using coordinates and not using the matrix. About the relative placement, i don't think it would solve the base line problem. Actually if there are imbalanced nodes (in hight). I have the experience of using relative coordinates and it made me to use phantoms to level that up (may be my bad!). –  Yasser Sobhdel Dec 30 '10 at 11:16
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9381410479545593, "perplexity": 2803.4811006778555}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-35/segments/1408500813887.15/warc/CC-MAIN-20140820021333-00216-ip-10-180-136-8.ec2.internal.warc.gz"}
http://mathhelpforum.com/advanced-math-topics/3497-proof.html
# Math Help - Proof 1. ## Proof I have to prove the following: Prove euclidean parallel postulate iff parallel lines are equidistant from one another. My hint is this: Use the converse to Theorem 4.1 since it can be proved assuming Neutral geometery + euclidean parallel postulate. " Parallel lines are equidistan from one another" if the length of all perpendicular segments from a point on line "dropped" to the other line are congurent[equal in length]. 2. Originally Posted by Nichelle14 Prove euclidean parallel postulate iff parallel lines are equidistant from one another Please define for me in rigorous geometric terms what it means that two lines are equidistant. 3. the best I can do for that is say that they are the same measurement apart. I don't think I can give you an answer. Maybe someone else can. 4. Observe, the picture. As I understand, the equidistant postulate states that, AB is perpendicular to line AD DC is perpendicular to line BC Also, equidistance-meaning AB=DC Note, triangle ABD and triangle BDC They satisfy Hypotenuse-Leg. Thus, they are congruent. Thus, <ADB=<BDC Thus the parallel postulate is proved. Note, the congrunce of hypotenuse-leg is independent of parallel postulate, If it were not then you cannot use it. Attached Thumbnails
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9081746935844421, "perplexity": 1448.2193549891501}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-10/segments/1394678695499/warc/CC-MAIN-20140313024455-00095-ip-10-183-142-35.ec2.internal.warc.gz"}
http://mathhelpforum.com/advanced-statistics/204923-question-bias-estimators.html
# Math Help - Question on Bias and Estimators 1. ## Question on Bias and Estimators Hi all, this is my problem, hopefully the image is clear To be honest, I'm stuck at finding the bias. I understand I need to work out the expectation (though that seems difficult, so help would be welcome), but what does it mean by subtracting sigma, when sigma is unknown? Thanks 2. ## Re: Question on Bias and Estimators Hey Mick. Since you have been given a distribution for S^2 in terms of chi-square, you want to find the expectation of S by itself. Recall that the expectation of a function of a random variable E[g(X)] for a continuous distribution is Integral (-infinity,infinity) g(x)f(x)dx. Now X = S^2. Let g(X) = SQRT(X) and now given the expectation theorem, you can find out what E[SQRT(S^2)] = E[S] is. Also be aware that n and sigma^2 are constants and not random variables so E[S^2] = sigma^2/(n-1)*E[chi-square_n_distribution]. Also remember that the expectation of a known distribution always returns some number whether its an actual number or just an expression in terms of non-random constants. 3. ## Re: Question on Bias and Estimators Thanks for the reply. So I'm at the integral of g(x)f(x), you've defined g(X)=SQRT(X) which I understand. But where do you take f(x) from? I guess either the normal distribution function or the inside of the square root of S? 4. ## Re: Question on Bias and Estimators f(x) will be the chi-square distribution with (n-1) degrees of freedom and E[SQRT(S^2)] = sigma^2/(n-1) * E[g(X)] where X ~ chi-square(n-1). 5. ## Re: Question on Bias and Estimators Thanks, I'm making progress I think, albeit very slow! But there's some parts of it I don't understand. So I'm going to post a picture of my workings, and if you could tell me what's badly wrong, or just point me how to continue it, I'd be really grateful. It looks like it can be integrated by parts, but I'm really bad at integration, and I've never been good with gamma distribution either. Thanks 6. ## Re: Question on Bias and Estimators Once you get to this step, use the fact that the integral of the PDF over the whole domain is equal to 1. Now you haven't got a 1/(Gamma(k)*theta^k) in there so you need to balance the equation by factoring in this term, but that's all there is to do it to evaluate the integral. More specifically if r = 1/(Gamma(k)*theta^k) and I is the integral term (without the constant c) then r*I = 1 if theta = 2 and k = n/2 so this means I = 1/r and then you can go from there. Also recall that you are trying to calculate E[S^2] where E[S^2] = [sigma^2/(n-1)]*E[SQRT(X)] <- chi-square so don't forget to factor in the sigma^2/(n-1) in the end (and the goal is to show that E[S^2] = sigma^2 for unbiased-ness or something else for biased-ness). 7. ## Re: Question on Bias and Estimators I'm sorry, I'm just confused. Can you tell me what I should be aiming for, what this integral should equal because I think once I've got this the rest of the question should be straightforward. Thanks 8. ## Re: Question on Bias and Estimators You want to show that E[S^2] = sigma^2 for S^2 to be an unbiased estimator of sigma^2 and we know that E[S^2] = sigma^2/(n-1) * E[SQRT(X)] where X ~ Chi-Square with n-1 degrees of freedom. Now in your working out you get E[SQRT(X)] down to the form of a gamma distribution PDF multiplied by some constant, but since you have the PDF in the integral, you can balance it by calculating for the appropriate constant c*Integral = 1 since integrating a PDF over the whole domain will always give 1. So if you balance this you will get Integral = 1/c for some constant (Hint: look at the definition of the Gamma distribution to find out what the c is in your case). Then once you get the final result for E[SQRT(X)] above, substitute in and see if you get E[S^2] = sigma^2 or something else and make a conclusion based on that.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9655706286430359, "perplexity": 563.1371414076036}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-49/segments/1416400379555.49/warc/CC-MAIN-20141119123259-00146-ip-10-235-23-156.ec2.internal.warc.gz"}
https://cs.stackexchange.com/tags/complexity-classes/hot?filter=all
# Tag Info 361 I think the Wikipedia articles $\mathsf{P}$, $\mathsf{NP}$, and $\mathsf{P}$ vs. $\mathsf{NP}$ are quite good. Still here is what I would say: Part I, Part II [I will use remarks inside brackets to discuss some technical details which you can skip if you want.] Part I Decision Problems There are various kinds of computational problems. However in an ... 178 Part II Continued from Part I. The previous one exceeded the maximum number of letters allowed in an answer (30000) so I am breaking it in two. $\mathsf{NP}$-completeness: Universal $\mathsf{NP}$ Problems OK, so far we have discussed the class of efficiently solvable problems ($\mathsf{P}$) and the class of efficiently verifiable problems ($\mathsf{... 68 It is known that P$\subseteq$NP$\subset$R, where R is the set of recursive languages. Since R is countable and P is infinite (e.g. the languages$\{n\}$for$n \in \mathbb{N}$are in P), we get that P and NP are both countable. 40 Let's refresh the definitions. PSPACE is the class of problems that can be solved on a deterministic Turing machine with polynomial space bounds: that is, for each such problem, there is a machine that decides the problem using at most$p(n)$tape cells when its input has length$n$, for some polynomial$p$. EXP is the class of problems that can ... 37 No, there will be absolutely no implication, for several reasons: The P vs. NP problem is about classical computation rather than quantum computation. Even if quantum computers could solve NP-hard problems in polynomial time (which we don't expect them to be able to do), it could still be the case that classical computers cannot solve them in polynomial ... 27$k$-SUM can be solved more quickly as follows. For even$k$: Compute a sorted list$S$of all sums of$k/2$input elements. Check whether$S$contains both some number$x$and its negation$-x$. The algorithm runs in$O(n^{k/2}\log n)$time. For odd$k$: Compute the sorted list$S$of all sums of$(k-1)/2$input elements. For each input element$a$, ... 26 More than useful mentioned answers, I recommend you highly to watch "Beyond Computation: The P vs NP Problem" by Michael Sipser. I think this video should be archived as one of the leading teaching video in computer science.! Enjoy! 17 Just elaborating what @Lamine pointed out : When$k$is part of the input,$k$can be as large as$\frac{n}{2}$, in which case the number of potential clique sets are$\binom{n}{\frac{n}{2}}$which is at least$\left( \frac{n}{\frac{n}{2}} \right)^{\frac{n}{2}}$. Hence your naive algorithm would take time$2^{\frac{n}{2}}$which is clearly exponential in ... 17 No. It is another open problem and certainly related, but different. The complexity class co-$\mathsf{NP}$is the set of languages whose complements are in$\mathsf{NP}$; that is, the set of decision problems for which a "no" answer has a deterministic polynomial-time verifier. So for example, the question "Is this SAT formula unsatisfiable?" If the answer ... 17 It is maybe easier to consider the contrapositive, that is${\sf P}={\sf NP} \Rightarrow {\sf NP}={\sf coNP}$. So assume${\sf P}={\sf NP}$, then for every$L\in {\sf NP}$, we have$L\in {\sf P}$, and since the languages in${\sf P}$are closed under complement,$\bar L\in {\sf P}$and therefore$L\in {\sf coNP}$. for every$L\in {\sf coNP}$, we have$\... 17 No implications are known either way: classical simulation of quantum computers tells us nothing about how hard NP search problems are; fast solutions to NP search problems tell us nothing about how fast quantum computers can be simulated classically. The following scenarios are possible: $P=NP=BQP$ $P=NP\subsetneq BQP$ $P\subsetneq NP=BQP$ $P\subsetneq NP\... 16 If a problem is NP-Hard it means that there exists a class of instances of that problem whose are NP-Hard. It is perfectly possible for other specific classes of instances to be solvable in polynomial time. Consider for example the problem of finding a 3-coloration of a graph. It is a well-known NP-Hard problem. Now imagine that its instances are restricted ... 16 All the classes you mention are classes of languages, formally, even if P and NP are often discussed in different (more sloppy?) terms. Note that terminology revolving around decision problems is equivalent to formal languages; the decision is always whether a word is in the given language, i.e. the problem is to solve the word problem. What you need to do ... 16 Any problem in NP is in EXPTIME because you can either use exponential time to try all possible certificates or to enumerate all possible computation paths of a nondeterministic machine. More formally, there are two main definitions of NP. One is that a language$L$is in NP iff there is a relation$R$such that there is a polynomial$p$such ... 15 First of all, the question you are asking is open, since an affirmative answer shows that$\sf NP = coNP$. In fact it is one of the most prominent open problems in computer science. If$\sf P= NP$, then the class$\sf NP$is closed under complement since$\sf P$is. If on the other hand$\sf P \not = NP$then we cannot say whether$\sf NP = coNP$or not. ... 14 It depends on what definitions you use. Sipser [1] defines$\mathrm{SPACE}(f(n))$to be the class of languages decided by Turing machines using$O(f(n))$cells on their work tapes for inputs of length$n$. Papadimitriou [2], on the other hand defines it to be the class of languages decided by Turing machines using at most$f(n)$cells on the ... 13 Let me answer your questions in order: By definition, a problem has an FPTAS if there is an algorithm which on instances of length$n$gives an$1+\epsilon$-approximation and runs in time polynomial in$n$and$1/\epsilon$, that is$O((n/\epsilon)^C)$for some constant$C \geq 0$. A running time of$2^{1/\epsilon}$doesn't belong to$O((n/\epsilon)^C)$for ... 12 Vor's answer gives the standard definition. Let me try to explain the difference a bit more intuitively. Let$M$be a bounded error probabilistic polynomial-time algorithm for a language$L$that answers correctly with probability at least$p\geq\frac{1}{2}+\delta$. Let$x$be the input and$n$the size of the input. What distinguishes an arbitrary$\... 12 $d$-SUM requires time $n^{\Omega(d)}$ unless k-SAT can be solved in $2^{o(n)}$ time for any constant k. This was shown in a paper by Mihai Patrascu and Ryan Williams(1). In other words, assuming the exponential time hypothesis, your algorithm is optimal up to a constant factor in the exponent (a polynomial factor in $n$) (1) Mihai Patrascu and Ryan ... 12 No, special cases can be easier. Consider this IP, for example, given $a_i \geq 0$ for $i \in [1..n]$: $\qquad\displaystyle \min \sum_{i=1}^n x_ia_i$ s.t. $\quad\displaystyle\sum_{i=1}^n x_i \geq 1$ and $\ \displaystyle x_i \in \mathbb{N}$ for $i \in [1..n]$. It finds the minimum among $a_1, \dots, a_n$ (that for which, inevitably, $x_i=1$ in an optimal ... 12 The fact that P ≠ NP does not preclude the possibility that NP = co-NP, in which case NP ∩ co-NP = NP. So to further the discussion, let us assume that NP ≠ co-NP. In that case, Corollary 9 in Schöning's A uniform approach to obtain diagonal sets in complexity classes shows that there exists some language in NP – co-NP which is NP-intermediate. So NPI ... 11 $\oplus P^{\oplus P}$ denotes the class $\oplus P$ equipped with what's known as an oracle for $\oplus P$ — we say that it has been given the ability to determine whether or not a string $s$ is a member of a language $L$ contained in the class $\oplus P$ in a single operation. I see that another commenter (sdcwc) has linked to the proof of $\oplus P^{\oplus ... 11 Your problem is known as the$\text{UNIQUE-SAT}$problem which is$\mathsf{US}$-complete. The problem is in$\mathsf{D^p}$but not known to be$\mathsf{D^p}$-hard under deterministic polynomial time reductions, where the class$\mathsf{D^p} = \{ L_1 \cap \overline{L_2} \mid L_1,L_2 \in \mathsf{NP} \}$. It was shown by Papadimitriou and Yannakis [1] that ... 11 There are very many natural complete problems for$\Pi_2^p$, and there is a survey [1] on completeness for levels of the polynomial hierarchy, containing many such problems. The paper On the complexity of min-max optimization problems and their approximation [2] contains a nice overview of "min-max problems" with several proofs of completeness. The latter ... 11 Your prof was absolutely not rigorous (i.e. completely wrong), that's why the distinction between NP and co-NP doesn't make sense with his definition. Better definition: Def.: A decision problem (that is a problem where the result is either YES or NO and nothing else) is in NP if for every instance where the result is YES there exists a hint such that we ... 10 That looks correct to me. The difference between BPP and PP is that for BPP the probability has to be greater than$1/2$by a constant, whereas for PP it could be$1/2+ 1/2^n$. So for BPP problems you can do probability amplification with a small number of repetitions, whereas for general PP problems you can't. 10 Only in one direction. As$\mathsf{P}=\text{co-}\mathsf{P}$, if$\mathsf{NP}\neq\text{co-}\mathsf{NP}$then we would know that$\mathsf{P}\neq\mathsf{NP}$. However the reverse implication doesn't hold. If$\mathsf{P}\neq\mathsf{NP}$then it's possible that either$\mathsf{NP}\neq\text{co-}\mathsf{NP}$or$\mathsf{NP}=\text{co-}\mathsf{NP}$. 10 First of all, we don't know whether$NP=EXP$or not. So the initial answer is "it is an open question". However, we strongly believe (and there are supporting evidence) that$NP\neq EXP$. In fact, we believe that$NP\neq PSPACE$and that$PSPACE\neq EXP$(that is, there is a strict containment$NP\subsetneq PSPACE \subsetneq EXP$). Since you are looking ... 10 The concept you are looking for is called enumeration complexity, which is the study of the computational complexity of enumerating (listing) all the solutions to a problem (or the members of a language/set). Enumeration algorithms can be modeled as a two step process:a precomputation step and an enumeration phase with delay. Both of these steps have their ... 9 Third, since$\sf{L} \subseteq \sf{NC}^2$, is there an algorithm to convert any logspace algorithm into a parallel version? It can be shown (Arora and Barak textbook) given a$t(n)$-time TM$M$, that an oblivious TM$M'$(i.e. a TM whose head movement is independent of its input$x$) can construct a circuit$C_n$to compute$M(x)$where$|x| = n\$. The ... Only top voted, non community-wiki answers of a minimum length are eligible
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.92631995677948, "perplexity": 1201.5519873105527}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-39/segments/1568514572896.15/warc/CC-MAIN-20190924083200-20190924105200-00271.warc.gz"}
https://www.physicsforums.com/threads/lorentz-invariant-mass-of-electromagnetic-field.170928/
Lorentz invariant mass of electromagnetic field? 1. May 19, 2007 da_willem An photon has mass zero by virtue of its momentum canceling its energy in $$m^2c^4 = E^2-p^2c^2$$ But in electromagnetism a field configution only has momentum when both a magnetic field and an electric field are present, e.g. in an electromagnetic wave. Now when there is only an electric or magnetic field present, doesn't the field have an invariant rest mass E/c^2 with E the total energy stored in the field? Does it make any sense to think of it like that? (Problem is maybe that for e.g. a point charge this mass is infinite...so it can't be the correct picture gravitationally right?) 2. May 19, 2007 pmb_phy I disagree. On what do you base this on? I worked out an example which gives the opposite of your conclusion. See http://www.geocities.com/physics_world/sr/mass_mag_field.htm No. No. The mass of a point charge is finite even in the case of a point charge which has an infinite mass density. Best wishes Pete 3. May 20, 2007 Xezlec Wait, are you just saying no because he used the term "mass"? I mean, relativistically, the energy stored in an electric or magnetic field certainly behaves like a mass E/c^2. It has inertia and it gravitates. Right? 4. May 20, 2007 Xezlec Oh, wait, you're just saying he can't say for a photon that that energy is a rest mass. Of course if you are at rest relative to the photon it has no mass. That zero mass gets dilated to finite mass when the photons speed becomes c, because at c, the mass is dilated by a factor of infinity. Which really makes no rigorous sense to say at all. But it makes intuitive sense. I wonder if that was any help? 5. May 20, 2007 da_willem I will take a look at your website later, but for now I would like to say a few things. Of course I figured a field configuration with only an electric or magnetic field has zero momentum, because the Poynting vector vanishes! Now with zero momentum and a nonzero field energy density this would seem to imply a mass by the energy momentum relation. I know this would make no sense physically, but the equations do appear to indicate such a (sometimes infinite) mass, what's the deal here? 6. May 20, 2007 Xezlec Oh, I see what you're asking now: when you integrate the energy density of the E-field of a point charge, you get infinite energy. This makes no sense because it certainly doesn't behave as though it has infinite mass. Here's what a website I found says: 7. May 21, 2007 da_willem Thanks! But, is it wrong to associate a 'rest mass' to the energy of a field, e.g. in the light of its gravitational influence? If so, what's the reason, as the equations (naively) seem to indicate such a mass? 8. May 21, 2007 pmb_phy I disagree. As my derivation demonstrates you can have only a magnetic field in a frame S and still have a non-zero Poynting vector in a frame S' which is moving relative to S. The reason being is that in S' there will be a non-vanishing E field which, which crossed with the B field in S' will give a non-vanishing Poynting vector. Pete 9. May 21, 2007 da_willem Right, I know that having an E or a B field is observer dependent (you do have an E field in the S' frame!), which might not cause a problem for the Lorentz invariance of the quantity $$'m'= \frac{1}{c^2} \sqrt{\frac{1}{2}\epsilon \int E^2 dV+ \frac{1}{2\mu} \int B^2 dV - \frac{c^2}{\mu} \int |\vec{E} \times \vec{B}| dV}$$ So in the S frame the change (increase mainly due to the magnetic field) in field energy is probably cancelled by the arising of field momentum. If I find the time I will try to do the calculation using your example. 10. May 25, 2007 Xezlec By my reading of that reference, I think if you apply the right normalization, you can find out if a particular field actually has any energy or not, and if so how much. The electric field of an electron itself doesn't have energy, as the article says, no one really has figured out any "nice" explanation why. It's just an exception to the usual rule about energy densities of fields. So when you have true point charges lying around, you have to do that normalization thing to get the energy calculations right. The electric field due to a continuous distribution of charge does carry energy in the usual way. And in this case, as I understand it, it is correct to say it has a rest mass. Point charges are the only weird thing, where the rules stop applying nicely. Of course there are no continous charge distributions really, but if you want to approximate... meh.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9115129709243774, "perplexity": 399.8092747028155}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476990033880.51/warc/CC-MAIN-20161020190033-00024-ip-10-142-188-19.ec2.internal.warc.gz"}
http://mathhelpforum.com/pre-calculus/106293-growth-decay-question.html
# Thread: Growth/Decay Question 1. ## Growth/Decay Question This problem uses radioactive isotopes. I need to find the amount of grams after 4000 years. I'm given: Isotope:14-C Half-life (years): 5715 Initial Quantity: 15 g Amount After 4000 Years: ?? g Also unsure as to starting this one. 2. Originally Posted by BeSweeet This problem uses radioactive isotopes. I need to find the amount of grams after 4000 years. I'm given: [i]Isotope:[/]14-C Half-life (years): 5715 Initial Quantity: 15 g Amount After 4000 Years: ?? g Also unsure as to starting this one. Use the following two formulae: ${\lambda} = \frac{ln2}{t_{1/2}}$ $A(t) = A_0e^{-\lambda t}$ Where: • $\lambda$ = Decay Constant • $t_{1/2}$ = half life • $A(t)$ = Amount left at time t • $A_0$ = Amount at t=0 • $t$ = time. 3. Originally Posted by e^(i*pi) Use the following two formulae: ${\lambda} = \frac{ln2}{t_{1/2}}$ $A(t) = A_0e^{-\lambda t}$ Where: • $\lambda$ = Decay Constant • $t_{1/2}$ = half life • $A(t)$ = Amount left at time t • $A_0$ = Amount at t=0 • $t$ = time. Thanks for the quick reply. [tex]\lambda[/MATH - 15? $t_{1/2}$ - 5715 $A(t)$ - ? $A_0$ - 15 $t$ - ? Help? 4. Originally Posted by BeSweeet This problem uses radioactive isotopes. I need to find the amount of grams after 4000 years. I'm given: [i]Isotope:[/]14-C Half-life (years): 5715 Initial Quantity: 15 g Amount After 4000 Years: ?? g Also unsure as to starting this one. Let t denote the time measured in years, A_0 the initial amount and A(t) the amount after t years. Then A(t) is given by: $A(t)=A_0 \cdot e^{k\cdot t}$ where k is a constant which you have to calculate first. You already know: If the initial value is A_0 then $A(5715) = \frac12 A_0$. That means: $\frac12 A_0 = A_0 \cdot e^{k \cdot 5715}$ . Solve for k. I've got $k \approx -0.0001212856$ Now calculate A(4000) with $A_0 = 15\ g$ 5. Originally Posted by BeSweeet Thanks for the quick reply. [tex]\lambda[/MATH - 15? $t_{1/2}$ - 5715 $A(t)$ - ? $A_0$ - 15 $t$ - ? Help? $\lambda = \frac{ln2}{5715}$ (which gives the same value as k as indicated by earboth) A(t) is what you're trying to find. t = 4000 6. I am so freaking confused right now. This may be asking a lot, but could someone find the answer, and show the steps? The best way for me, personally, to figure out how to do something, is to analyze steps that get an answer.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 28, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8818871974945068, "perplexity": 3482.3770024577116}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-17/segments/1492917118743.41/warc/CC-MAIN-20170423031158-00272-ip-10-145-167-34.ec2.internal.warc.gz"}
https://cstheory.stackexchange.com/questions/42759/alternating-delivery-problem
# Alternating Delivery Problem What is known about the complexity of the following problem: Suppose we have a complete bipartite graph $$G(V,E)$$ with disjoint sets $$C$$ and $$T$$. The candidate vertices, and the target vertices respectively. We want to choose a subset $$S\subset C$$ of size $$2$$, a servicing set from the candidates. For a target vertex $$t\in T$$, there is a probability $$p_t$$ that it will issue an order at any turn. Every turn, only one vertex issues an order. We start at a vertex $$s_0\in S$$. We wait for the next order to come from a vertex $$t_0\in T$$, we go to service it. Then, we go to the nearest servicing vertex $$s_1$$. Next, we get an order from vertex $$t_1$$, we go from $$s_1$$ to service $$t_1$$, then go back to the nearest servicing vertex to $$s_2$$. We repeat this, going to the nearest vertex servicing vertex in $$S$$ once you have serviced $$t_i$$ in turn $$i$$. The question is how do we choose $$S$$ to minimize the average cost from a service node to a target node asymptotically. • How long does it take to service a vertex? – JRyan Apr 22 at 17:18
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 18, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6995669603347778, "perplexity": 226.36758903382355}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-43/segments/1570987779528.82/warc/CC-MAIN-20191021143945-20191021171445-00010.warc.gz"}
https://www.atmos-chem-phys.net/17/8553/2017/
Journal cover Journal topic Atmospheric Chemistry and Physics An interactive open-access journal of the European Geosciences Union Journal topic • IF 5.668 • IF 5-year 6.201 • CiteScore 6.13 • SNIP 1.633 • IPP 5.91 • SJR 2.938 • Scimago H index 174 • h5-index 87 # Abstracted/indexed Abstracted/indexed Atmos. Chem. Phys., 17, 8553-8575, 2017 https://doi.org/10.5194/acp-17-8553-2017 Atmos. Chem. Phys., 17, 8553-8575, 2017 https://doi.org/10.5194/acp-17-8553-2017 Research article 13 Jul 2017 Research article | 13 Jul 2017 # Exploring gravity wave characteristics in 3-D using a novel S-transform technique: AIRS/Aqua measurements over the Southern Andes and Drake Passage Corwin J. Wright1, Neil P. Hindley1,2, Lars Hoffmann3, M. Joan Alexander4, and Nicholas J. Mitchell1 Corwin J. Wright et al. • 1Centre for Space, Atmospheric and Oceanic Science, University of Bath, Bath, UK • 2Institute for Climate and Atmospheric Science, School of Earth and Environment, University of Leeds, Leeds, UK • 3Jülich Supercomputing Centre, Forschungszentrum Jülich, Jülich, Germany • 4Northwest Research Associates, Boulder, CO, USA Abstract. Gravity waves (GWs) transport momentum and energy in the atmosphere, exerting a profound influence on the global circulation. Accurately measuring them is thus vital both for understanding the atmosphere and for developing the next generation of weather forecasting and climate prediction models. However, it has proven very difficult to measure the full set of GW parameters from satellite measurements, which are the only suitable observations with global coverage. This is particularly critical at latitudes close to 60° S, where climate models significantly under-represent wave momentum fluxes. Here, we present a novel fully 3-D method for detecting and characterising GWs in the stratosphere. This method is based around a 3-D Stockwell transform, and can be applied retrospectively to existing observed data. This is the first scientific use of this spectral analysis technique. We apply our method to high-resolution 3-D atmospheric temperature data from AIRS/Aqua over the altitude range 20–60 km. Our method allows us to determine a wide range of parameters for each wave detected. These include amplitude, propagation direction, horizontal/vertical wavelength, height/direction-resolved momentum fluxes (MFs), and phase and group velocity vectors. The latter three have not previously been measured from an individual satellite instrument. We demonstrate this method over the region around the Southern Andes and Antarctic Peninsula, the largest known sources of GW MFs near the 60° S belt. Our analyses reveal the presence of strongly intermittent highly directionally focused GWs with very high momentum fluxes (∼ 80–100 mPa or more at 30 km altitude). These waves are closely associated with the mountains rather than the open ocean of the Drake Passage. Measured fluxes are directed orthogonal to both mountain ranges, consistent with an orographic source mechanism, and are largest in winter. Further, our measurements of wave group velocity vectors show clear observational evidence that these waves are strongly focused into the polar night wind jet, and thus may contribute significantly to the missing momentum at these latitudes. These results demonstrate the capabilities of our new method, which provides a powerful tool for delivering the observations required for the next generation of weather and climate models.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8038821220397949, "perplexity": 3975.1850521511105}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-26/segments/1560627999800.5/warc/CC-MAIN-20190625051950-20190625073950-00252.warc.gz"}
https://www.varsitytutors.com/algebra_ii-help/intermediate-single-variable-algebra/polynomials?page=13
Algebra II : Polynomials Example Questions Example Question #74 : Factoring Polynomials Factor: Explanation: To factor this, use trial and error to see what works. Since we have a  as the leading coefficient, it's helpful to remember that there's only one way to get  . Same with  as our third term--there's only one way to get  . Use these facts as you try to factor. Remember that signs matter. Therefore, your answer is: . Example Question #75 : Factoring Polynomials Factor the polynomial: Explanation: The best method for factoring this polynomial is by grouping: Example Question #76 : Factoring Polynomials Factor the following expression completely: Explanation: The expression is in quadratic form, so it can be factored as though it is quadratic. The resulting expression can be factored further, as there are two difference of squares quadratic expressions. Example Question #77 : Factoring Polynomials Factor the following polynomial: Explanation: In order to factor a polynomial, you have to first find the greatest common factor. In this case, each term has an x in it, so you can easily factor that out. Also, the greatest common factor for the three terms is 3. After you take both the 3 and the x out you are left with The next step is to figure out which two numbers multiply together to equal , but also add together to equal . The factors of 52 are . The only ones that add to equal are , and in order to equal a positive 52, and , both of the numbers need to be negative. Example Question #78 : Factoring Polynomials Factor . Explanation: We need two numbers that add to get , and multiply to get .  It looks like  and  would do the trick.  To double check our answer, we can use the quadratic formula. Knowing our solutions are  and , we can insert them into an equation: and set each equal to . so: Example Question #79 : Factoring Polynomials Factor . Explanation: Because of the leading  in the quadratic term of the function, it's hard to think of a clever combination of solutions, but we can just rely on the quadratic formula to help us out: Setting each equation equal to  yields: If we were to expand this function we would find that it's wrong by a factor of . To correct this, we multiply one of the terms by .  In this case we can multiply the  term by 2, which also serves to clean up the fraction.  This gives us an end result of: Example Question #80 : Factoring Polynomials Factor . Explanation: We need to find two numbers that multiply to get , but we don't have to worry about them adding at all because there's a quadratic term and a linear term. If we check, we can see that the quadratic and linear terms multiply to get , which is very convenient for us.  Because the quadratic term is negative, we pair that sign with the linear term when factored: Example Question #81 : Factoring Polynomials Factor . Explanation: We can start by factoring the numerator and the denominator.  Starting with the numerator, we need to find two numbers that multiply to get  and add to get .  It happens that 3 and 3 work, so: For the denominator, we need two numbers that multiply to get , and add to get .  Here,  and  work, so: If we needed to, we always could have used the quadratic formula to factor each of the above equations.  We can now put the factored equations back in to our original function: Then we can cancel the  terms. Example Question #82 : Factoring Polynomials Factor . Explanation: In the beginning, we can treat the polynomials in the numerator and denominator as separate functions. We can use the quadratic formula for each, if we can't come up with a clever guess at the solution: Returning the solutions back into the original function gives: We cancel the  terms and have: Factor .
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8428393602371216, "perplexity": 539.1474278322244}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-25/segments/1623488556133.92/warc/CC-MAIN-20210624141035-20210624171035-00453.warc.gz"}
https://itprospt.com/num/20173802/find-the-sum-of-the-first-terms-of-the-geometric-series
5 # Find the sum of the first terms of the geometric series | + +( the least value of n such that the sum exceeds 20.Hence, find(5 marks)... ## Question ###### Find the sum of the first terms of the geometric series | + +( the least value of n such that the sum exceeds 20.Hence, find(5 marks) Find the sum of the first terms of the geometric series | + +( the least value of n such that the sum exceeds 20. Hence, find (5 marks) #### Similar Solved Questions ##### Functionf() =x" Given f (x) = x2 _ 4x and g (x) =x+ 1 Find (f + g)(x), (f.9)(x) f(3) - 9(-2) , (f g)(x) f)(3) Let be the one-to-one function Find the inverse function: a) f(x) =23 b) f(x) = Vzx ~ 3 8. Perform the following transformations of f(x) = 2 b) Move the result of (a) units left. a) Shift f(x) units uP; Reflect the result of (b) with respect to the x-axis function f() =x" Given f (x) = x2 _ 4x and g (x) =x+ 1 Find (f + g)(x), (f.9)(x) f(3) - 9(-2) , (f g)(x) f)(3) Let be the one-to-one function Find the inverse function: a) f(x) =23 b) f(x) = Vzx ~ 3 8. Perform the following transformations of f(x) = 2 b) Move the result of (a) units left. a) Sh... ##### Poker Simulation) I- isoiten usefulto be abletosimulate thedrzwing 0i3 deck ofpokercards Write MATLAB function pokern,thar simulatesthe poker drawbyreturningsome randominteger berween and 52 every time thatitiscalled Write script,test_pokerm thacsimulates 5,0ootimes withdrawngan lotits histozram figure theend showthe generated dataoi the first 20 values in the poker drawing Thehistogr must have title x-label andy-label showthe function pokerm the script test_pokerm the histogram fizure,and disp Poker Simulation) I- isoiten usefulto be abletosimulate thedrzwing 0i3 deck ofpokercards Write MATLAB function pokern,thar simulatesthe poker drawbyreturningsome randominteger berween and 52 every time thatitiscalled Write script,test_pokerm thacsimulates 5,0ootimes withdrawngan lotits histozram fi... ##### Provida eccsptable nema for te elkenes heua belowDraw the structuro alkene with the name 3-cthyi-3-hepxeneProvide the produc each reacbon shown belb# HCIHzoHzSO4BH,2) HzOz. HO"The number of bondsacetyleneB) 2frcmWhich of tne following reagentsused prepare trans alkene fok 2-putyne?A) HzLindlar catalyst B) Oz Ihen ZnwaterCJNa / NH; ()none of tnese6) The enol form of each molecule shown below For each molecule draw tne corresponding keto formKelo fonm61dKeakn Provida eccsptable nema for te elkenes heua below Draw the structuro alkene with the name 3-cthyi-3-hepxene Provide the produc each reacbon shown belb# HCI Hzo HzSO4 BH, 2) HzOz. HO" The number of bonds acetylene B) 2 frcm Which of tne following reagents used prepare trans alkene fok 2-putyne? ... ##### Thc timc T to failurc of ccrtain devicc has Wcibull(2,1) distribution That i8, Fr(t) =1 - e Show that X _ T2 cxponential(1). In view of the above; what does L7 converge to, for an iid sequence "XITT Ti,T2, _ from thc Wcibull(2,1) distribution? (First rcwritc the ratio in terms quantities of the form % Ci_1 Xf .) Thc timc T to failurc of ccrtain devicc has Wcibull(2,1) distribution That i8, Fr(t) =1 - e Show that X _ T2 cxponential(1). In view of the above; what does L7 converge to, for an iid sequence "XITT Ti,T2, _ from thc Wcibull(2,1) distribution? (First rcwritc the ratio in terms quantities of t... ##### Prcltm 5.67cud nnud EculuEprtentnt Prcltm 5.67 cud nnud Eculu Eprte ntnt... ##### Recognize basic argument patterns LIETL hadvan yjoups (rom the condusion identily what rules Ill comycund statemerits &nd pj.an reduires tkat Yoj aildi Yol cobain Lndisiun requifes regular pradtio @rderta repetition inceasingly aPp ed. Develcairg U1z; skill cmpleting [ complex prools You arr OsKn problem seL. Eten try (h-ea d fferent versions you firid DfOo Pacn Daut one page ba easy, good idea Iry the other verions praol possible eac page mwuch Dradl cnpletingComnplete the following natural recognize basic argument patterns LIETL hadvan yjoups (rom the condusion identily what rules Ill comycund statemerits &nd pj.an reduires tkat Yoj aildi Yol cobain Lndisiun requifes regular pradtio @rderta repetition inceasingly aPp ed. Develcairg U1z; skill cmpleting [ complex prools You arr OsK... ##### Question 1 (1 point) A transposition leaves all elements except two fixedTrueFalseQuestion 2 (1 point) Every permutation can be written as product of transpositionsTrueFalseQuestion 3 (1 point) An r cycle Is even if ris even and Is odd il r Is oddTrueFalseQuestionpoint)Thc symmetric Rroup Sn on elemcnts has clementsTrue False Question 1 (1 point) A transposition leaves all elements except two fixed True False Question 2 (1 point) Every permutation can be written as product of transpositions True False Question 3 (1 point) An r cycle Is even if ris even and Is odd il r Is odd True False Question point) Thc symmetric Rroup... ##### F 367 _ 934(-1) -(-1)82 Ih; 27 f 367 _ 934 (-1) -(-1)8 2 Ih; 27... ##### IEIEZETSTIIIPatmtta 4ace7i?Lene teleaGcichtxr7 -Acatt eerernenFaeJa/deie Ur- Ju Catt-ot1Nhcwuis IEIEZETSTIIIPatmtta 4ace7i? Lene telea Gcichtxr7 - Acatt eerernenFae Ja/deie Ur- Ju Catt-ot1 Nhcwuis... ##### LARLINALG8 4.6,.03745DETAILSFind J basis for the nullspace of thelmatrix: (If there is no basis | enter MONElin any single cell )1107u2 LARLINALG8 4.6,.037 45 DETAILS Find J basis for the nullspace of thelmatrix: (If there is no basis | enter MONElin any single cell ) 110 7u2... ##### Question 6 Not complete Marked out of 5.00 Flag questionFind the inverse Laplace transform of the following functionF(s) 3-WtiSelect one:f(t) = e' cosh(2t) + ! 6sinhz516 sinh ? + 6{f() 72-51 cosh(5 v6t)cosh(2 v6t) + 3v6 sinh?,64) f() = 2'70751 3v6 sinh:64) f() cosh(V6t) +CheckPrevious pageNext page Question 6 Not complete Marked out of 5.00 Flag question Find the inverse Laplace transform of the following function F(s) 3-Wti Select one: f(t) = e' cosh(2t) + ! 6sinhz 516 sinh ? + 6{ f() 72-51 cosh(5 v6t) cosh(2 v6t) + 3v6 sinh?,64) f() = 2' 70751 3v6 sinh:64) f() cosh(V6t) + Check Pre... ##### In spite of the potential safety hazards, some people would like to have an Internet connection in their car. A preliminary survey of adult Americans has estimated this proportion to be somewhere around. 30 (USA Today. May 1,2009 ).a. Use the given preliminary estimate to determine the sample size required to estimate the proportion of adult Americans who would like an Internet connection in their car to within $.02$ with $95 %$ confidence.b. The formula for determining sample size given in this In spite of the potential safety hazards, some people would like to have an Internet connection in their car. A preliminary survey of adult Americans has estimated this proportion to be somewhere around. 30 (USA Today. May 1,2009 ). a. Use the given preliminary estimate to determine the sample size ... ##### Recall Cauchy's Integra Formula (I): Let G € € be open and f : G - € analytic. If Y is loop (closed rectifiable curve) in G such that n(Y,w) = 0 for all w €C - G, then for each a eG-{r}, f(z) n(Y,a) f(a) dz. 2ti Yz-a Use Cauchy's Integral Formula to derive this version of Cauchy's Theorem: Let G C € be open and g: G - € analytic: If Y is loop in G such that n(Y,w) = 0 for all w €C-G,then g = 0 Recall Cauchy's Integra Formula (I): Let G € € be open and f : G - € analytic. If Y is loop (closed rectifiable curve) in G such that n(Y,w) = 0 for all w €C - G, then for each a eG-{r}, f(z) n(Y,a) f(a) dz. 2ti Yz-a Use Cauchy's Integral Formula to derive this vers... ##### Convert 1.25 moles of FezOa to formula units of FezOg: In addition; indicate how many atoms of iron and how many this amount atoms of oxygen are of FezOg Show your work: Convert 1.25 moles of FezOa to formula units of FezOg: In addition; indicate how many atoms of iron and how many this amount atoms of oxygen are of FezOg Show your work:... ##### Rationalize the denominator of each expression. Assume that all variables are positive when they appear. $$\frac{\sqrt{2}}{\sqrt{7}+2}$$ Rationalize the denominator of each expression. Assume that all variables are positive when they appear. $$\frac{\sqrt{2}}{\sqrt{7}+2}$$... ##### (a) Determine the final temperature when 0.1 $\mathrm{kg}$ of wa-ter at $10^{\circ} \mathrm{C}$ is added to 0.3 $\mathrm{kg}$ of soup at $50^{\circ} \mathrm{C} .$ What assumptions did you make? (b) Estimate the entropy change of this water-soup system during the process. Does the second law of thermodynamics allow this process? (a) Determine the final temperature when 0.1 $\mathrm{kg}$ of wa-ter at $10^{\circ} \mathrm{C}$ is added to 0.3 $\mathrm{kg}$ of soup at $50^{\circ} \mathrm{C} .$ What assumptions did you make? (b) Estimate the entropy change of this water-soup system during the process. Does the second law of ther...
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8674963116645813, "perplexity": 15087.51170371598}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656103940327.51/warc/CC-MAIN-20220701095156-20220701125156-00174.warc.gz"}
http://mathoverflow.net/questions/110580/almost-universal-properties
# Almost universal properties Suppose that a small category $C$ has sets of objects and arrows that carry the structure of, say, Polish spaces, in some appropriately compatible way. For example $C$ might be an internal category in some category of Polish spaces (with continuous morphisms? Borel morphisms?). Then it might make sense to study categorical limits that only hold generically (according to some choosen ideal or $\sigma$-ideal). For example, even if $C$ does not have products, a priori one might still hope for an almost product $A \times_{\rm almost} B$ for objects $A$ and $B$ such that for, say, merely a co-meager set of triples $(C,f:C\rightarrow A,g:C\rightarrow B)$ there exists a unique map from $C$ to $A \times_{\rm almost} B$ so that the appropriate diagram commutes. For now I'm only asking for: trenchant examples and/or pointers to the literature where anything along these lines arises. - Doesn't this make the most sense in the category of Polish spaces with maps up to almost everywhere equivalence? –  Will Sawin Oct 24 '12 at 21:02 @Will I thought about that, but I phrased the question the way I did because I thought it would be more attractive to more people if $C$ still constituted an old-fashioned garden-variety category, albeit with some extra structure. Make $C$ internal to a non-concrete category like yours and it gets hard to talk about the source and target of specific arrows, etc. So my motivation, to repair or relax the condition of the existence of limits and colimits for specific diagrams needs reframing. If I understand you?? –  David Feldman Oct 24 '12 at 21:59
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8128339648246765, "perplexity": 380.5634141642939}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2015-22/segments/1432207928780.77/warc/CC-MAIN-20150521113208-00091-ip-10-180-206-219.ec2.internal.warc.gz"}
https://amp.en.depression.pp.ua/9558678/1/exponential-utility.html
# ⓘ Exponential utility. In economics and finance, exponential utility is a specific form of the utility function, used in some contexts because of its convenience .. ## ⓘ Exponential utility In economics and finance, exponential utility is a specific form of the utility function, used in some contexts because of its convenience when risk is present, in which case expected utility is maximized. Formally, exponential utility is given by: u c = { 1 − e − a c / a ≠ 0 c a = 0 {\displaystyle uc={\begin{cases}1-e^{-ac}/a&a\neq 0\\c&a=0\\\end{cases}}} c {\displaystyle c} is a variable that the economic decision-maker prefers more of, such as consumption, and a {\displaystyle a} is a constant that represents the degree of risk preference (a > 0 {\displaystyle a> 0} for risk aversion, a = 0 {\displaystyle a=0} for risk-neutrality, or a < 0 {\displaystyle a • aversion. Isoelastic function Constant elasticity of substitution Exponential utility Risk aversion Ljungqvist, Lars Sargent, Thomas J. 2000 Recursive • alternative utility functions such as: CES constant elasticity of substitution, or isoelastic utility Isoelastic utility Exponential utility Quasilinear • value of the utility function. Notable special cases of HARA utility functions include the quadratic utility function, the exponential utility function • measure which depends on the risk aversion of the user through the exponential utility function. It is a possible alternative to other risk measures as • with quadratic utility 2 two - periods, exponential utility and normally - distributed returns, 3 infinite - periods, quadratic utility and stochastic • in a graph in which each player has a utility function that depends only on him and his neighbors. As the utility function depends on fewer other players • risk aversion, a term in Economics referring to a property of the exponential utility function Cara Sucia disambiguation Caras disambiguation Carra • assertion of David P. Reed that the utility of large networks, particularly social networks, can scale exponentially with the size of the network The • this integral exists. Exponential discounting and hyperbolic discounting are the two most commonly used examples. Discounted utility Intertemporal choice • multiplying by a fixed constant. Logarithmic growth is the inverse of exponential growth and is very slow. A familiar example of logarithmic growth is • investor s remaining lifetime. Under certain assumptions including exponential utility and a single asset with returns following an ARMA 1, 1 process, a
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9888941049575806, "perplexity": 1638.6269583657354}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-49/segments/1637964363337.27/warc/CC-MAIN-20211207075308-20211207105308-00380.warc.gz"}
https://www.physicsforums.com/threads/fusion-simulation-software.865789/
# Fusion simulation software Tags: 1. Apr 6, 2016 ### l0st In a nutshell: is there any free software, which can take as input full description of a relatively large system (20-100 light atoms + photons), simulate its evolution over a period of time, and tell the probability of fusion over that time? If there's no such thing available, how close could one get to it with existing software? 2. Apr 6, 2016 ### phyzguy I don't think anybody simulates fusion at the level of atoms. It would take a huge number of atoms to simulate even a small fluid volume, I think most simulation is done using fluid dynamics codes where you treat the plasma as a fluid characterized by a composition, temperature, pressure, magnetic field, etc. You can then write the rate of fusion as a function of the fluid variables. There are several open source MHD codes that can simulate fusion plasmas. Most of these were developed for astrophysical applications. One example is Athena. This paper gives some details on simulations using fluid dynamics codes at NIF. Draft saved Draft deleted Similar Discussions: Fusion simulation software
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9400875568389893, "perplexity": 1174.5797540816745}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-43/segments/1508187825889.47/warc/CC-MAIN-20171023092524-20171023112524-00516.warc.gz"}
https://simple.wiktionary.org/wiki/subordinate
# subordinate Jump to navigation Jump to search Subordinate is part of the Academic Word List. It is important for students in college and university. ### Pronunciation Adjective and Noun • (UK) enPR: sə-bôʹdĭn-ət, IPA: /səˈbɔːdɪnət/ • (US) enPR: sə-bôrʹdĭn-ət, IPA: /səˈbɔrdɪnət/ • Audio (US) (file) Verb • (UK) enPR: sə-bôʹdĭn-āt, IPA: /səˈbɔːdɪneɪt/ • Audio (UK) (file) • (US) enPR: sə-bôrʹdĭn-āt, IPA: /səˈbɔrdɪneɪt/ ## Adjective Positive subordinate Comparative more subordinate Superlative most subordinate 1. If something is subordinate, it is less important than someone or something else. Keeping our natural world healthy remains subordinate to the health of the economy. In society a dominant group often ignores the wishes of subordinate groups. ## Noun 1. (countable) A subordinate is someone who is at a lower position in an organization. A good manager will talk with her subordinates before making most decisions. ## Verb Plain form subordinate Third-person singular subordinates Past tense subordinated Past participle subordinated Present participle subordinating 1. (transitive) If you subordinate ${\displaystyle x}$ to ${\displaystyle y}$, you give less importance to ${\displaystyle x}$ and more to ${\displaystyle y}$. When he fell in love, everything else was subordinated to it.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 4, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.1854621320962906, "perplexity": 28822.270848209926}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-29/segments/1593655897168.4/warc/CC-MAIN-20200714145953-20200714175953-00183.warc.gz"}
http://www.koreascience.or.kr/article/JAKO200833338747473.page
# ON GENERALIZED JORDAN LEFT DERIVATIONS IN RINGS • Ashraf, Mohammad (Department of Mathematics Aligarh Muslim University) ; • Ali, Shakir (Department of Mathematics Aligarh Muslim University) • Published : 2008.05.31 #### Abstract In this paper, we introduce the notion of generalized left derivation on a ring R and prow that every generalized Jordan left derivation on a 2-torsion free primp ring is a generalized left derivation on R. Some related results are also obtained. #### References 1. M. Ashraf, On left $({\theta},{\phi})$-derivations of prime rings, Arch. Math. (Brno) 41 (2005), no. 2, 157-166 2. M. Ashraf, A. Ali, and S. Ali, On Lie ideals and generalized $({\theta},{\phi})$-derivations in prime rings, Comm. Algebra 32 (2004), no. 8, 2977-2985 https://doi.org/10.1081/AGB-120039276 3. M. Ashraf and N. Rehman, On Jordan generalized derivations in rings, Math. J. Okayama Univ. 42 (2000), 7-9 4. M. Ashraf and N. Rehman, On Lie ideals and Jordan left derivations of prime rings, Arch. Math. (Brno) 36 (2000), no. 3, 201-206 5. M. Ashraf, N. Rehman, and S. Ali, On Jordan left derivations of Lie ideals in prime rings, Southeast Asian Bull. Math. 25 (2001), no. 3, 379-382 https://doi.org/10.1007/s100120100000 6. M. Ashraf, N. Rehman, and S. Ali, On Lie ideals and Jordan generalized derivations of prime rings, Indian J. Pure Appl. Math. 34 (2003), no. 2, 291-294 7. M. Bresar and J. Vukman, On left derivations and related mappings, Proc. Amer. Math. Soc. 110 (1990), no. 1, 7-16 8. W. Cortes and C. Haetinger, On Jordan generalized higher derivations in rings, Turkish J. Math. 29 (2005), no. 1, 1-10 9. Q. Deng, On Jordan left derivations, Math. J. Okayama Univ. 34 (1992), 145-147 10. I. N. Herstein, Jordan derivations of prime rings, Proc. Amer. Math. Soc. 8 (1957), 1104-1110 11. I. N. Herstein, Topics in Ring Theory, Univ. of Chicago Press, Chicago, 1969 12. B. Hvala, Generalized derivations in rings, Comm. Algebra 26 (1998), no. 4, 1147-1166 https://doi.org/10.1080/00927879808826190 13. W. Jing and S. Lu, Generalized Jordan derivations on prime rings and standard operator algebras, Taiwanese J. Math. 7 (2003), no. 4, 605-613 https://doi.org/10.11650/twjm/1500407580 14. K. W. Jun and B. D. Kim, A note on Jordan left derivations, Bull. Korean Math. Soc. 33 (1996), no. 2, 221-228 15. Y. S. Jung, Generalized Jordan triple higher derivations on prime rings, Indian J. Pure Appl. Math. 36 (2005), no. 9, 513-524 16. E. C. Posner, Derivations in prime rings, Proc. Amer. Math. Soc. 8 (1957), 1093-1100 17. J. Vukman, Jordan left derivations on semiprime rings, Math. J. Okayama Univ. 39 (1997), 1-6 18. S. M. A. Zaidi, M. Ashraf, and S. Ali, On Jordan ideals and left $({\theta},{\theta})$-derivations in prime rings, Int. J. Math. Math. Sci. 2004 (2004), no. 37-40, 1957-1964 https://doi.org/10.1155/S0161171204309075 19. B. Zalar, On centralizers of semiprime rings, Comment. Math. Univ. Carolin. 32 (1991), no. 4, 609-614 #### Cited by 1. On generalized left derivations in rings and Banach algebras vol.81, pp.3, 2011, https://doi.org/10.1007/s00010-011-0070-5 2. LEFT JORDAN DERIVATIONS ON BANACH ALGEBRAS AND RELATED MAPPINGS vol.47, pp.1, 2010, https://doi.org/10.4134/BKMS.2010.47.1.151 3. On Lie Ideals and Generalized Jordan Left Derivations of Prime Rings vol.65, pp.8, 2014, https://doi.org/10.1007/s11253-014-0855-5 4. Generalized Jordan left derivations on semiprime algebras vol.161, pp.1, 2010, https://doi.org/10.1007/s00605-009-0116-0 5. Generalized Jordan left derivations in rings with involution vol.45, pp.4, 2012, https://doi.org/10.1515/dema-2013-0420 6. Generalized left derivations acting as homomorphisms and anti-homomorphisms on Lie ideal of rings vol.22, pp.3, 2014, https://doi.org/10.1016/j.joems.2013.12.015 7. Left Derivations Characterized by Acting on Multilinear Polynomials vol.39, pp.6, 2011, https://doi.org/10.1080/00927872.2010.480960 8. Additive mappings satisfying algebraic conditions in rings vol.63, pp.2, 2014, https://doi.org/10.1007/s12215-014-0153-y 9. Some Theorems for Sigma Prime Rings with Differential Identities on Sigma Ideals vol.2013, 2013, https://doi.org/10.1155/2013/572690 10. CHARACTERIZATIONS OF REAL HYPERSURFACES OF TYPE A IN A COMPLEX SPACE FORM vol.47, pp.1, 2010, https://doi.org/10.4134/BKMS.2010.47.1.001 11. Additive mappings act as a generalized left $$(\alpha , \beta )$$(α,β)-derivation in rings pp.2198-2759, 2018, https://doi.org/10.1007/s40574-018-0165-1
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9005589485168457, "perplexity": 3526.2623328902023}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439738858.45/warc/CC-MAIN-20200811235207-20200812025207-00194.warc.gz"}
http://www.scoop.it/t/complexity-by-bernard-ryefield?page=8
Complexity & Systems 9.0K views | +5 today # Complexity & Systems Complex systems present problems both in mathematical modelling and philosophical foundations. The study of complex systems represents a new approach to science that investigates how relationships between parts give rise to the collective behaviors of a system and how the system interacts and forms relationships with its environment. The equations from which models of complex systems are developed generally derive from statistical physics, information theory and non-linear dynamics, and represent organized but unpredictable behaviors of natural systems that are considered fundamentally complex.  wikipedia (en) Scooped by Bernard Ryefield ## Evolution’s Contrarian Capacity for Creativity - Facts So Romantic - Nautilus One can imagine life evolving again and again, crashing on the rocks of time and circumstance, until finally it hit upon just the right mutation rate—one that eons later would produce organisms and species and ecosystems. No comment yet. Rescooped by Bernard Ryefield from Creativity - Problem Solving ## Studying Collective Human Decision Making and Creativity with Evolutionary Computation We report a summary of our interdisciplinary research project "Evolutionary Perspective on Collective Decision Making" that was conducted through close collaboration between computational, organizational and social scientists at Binghamton University. We redefined collective human decision making and creativity as evolution of ecologies of ideas, where populations of ideas evolve via continual applications of evolutionary operators such as reproduction, recombination, mutation, selection, and migration of ideas, each conducted by participating humans. Based on this evolutionary perspective, we generated hypotheses about collective human decision making using agent-based computer simulations. The hypotheses were then tested through several experiments with real human subjects. Throughout this project, we utilized evolutionary computation (EC) in non-traditional ways---(1) as a theoretical framework for reinterpreting the dynamics of idea generation and selection, (2) as a computational simulation model of collective human decision making processes, and (3) as a research tool for collecting high-resolution experimental data of actual collaborative design and decision making from human subjects. We believe our work demonstrates untapped potential of EC for interdisciplinary research involving human and social dynamics. No comment yet. Rescooped by Bernard Ryefield from Complex Systems and X-Events ## Are Big, Rich Cities Greener Than Poor Ones? When it comes to cities, being big and rich is better for the planet than being big and poor, according to a new study of carbon dioxide emissions from cities around the world. But is this correct? Via Claudia Mihai, Roger D. Jones, PhD No comment yet. Scooped by Bernard Ryefield ## Global Civil Unrest: Contagion, Self-Organization, and Prediction Civil unrest is a powerful form of collective human dynamics, which has led to major transitions of societies in modern history. The study of collective human dynamics, including collective aggression, has been the focus of much discussion in the context of modeling and identification of universal patterns of behavior. In contrast, the possibility that civil unrest activities, across countries and over long time periods, are governed by universal mechanisms has not been explored. Here, records of civil unrest of 170 countries during the period 1919–2008 are analyzed. It is demonstrated that the distributions of the number of unrest events per year are robustly reproduced by a nonlinear, spatially extended dynamical model, which reflects the spread of civil disorder between geographic regions connected through social and communication networks. The results also expose the similarity between global social instability and the dynamics of natural hazards and epidemics. No comment yet. Rescooped by Bernard Ryefield from Non-Equilibrium Social Science ## The Science of Inequality In 2011, the wrath of the 99% kindled Occupy movements around the world. The protests petered out, but in their wake an international conversation about inequality has arisen, with tens of thousands of speeches, articles, and blogs engaging everyone from President Barack Obama on down. Ideology and emotion drive much of the debate. But increasingly, the discussion is sustained by a tide of new data on the gulf between rich and poor. This special issue uses these fresh waves of data to explore the origins, impact, and future of inequality around the world. Archaeological and ethnographic data are revealing how inequality got its start in our ancestors (see pp. 822 and 824). New surveys of emerging economies offer more reliable estimates of people's incomes and how they change as countries develop (see p. 832). And in the past decade in developed capitalist nations, intensive effort and interdisciplinary collaborations have produced large data sets, including the compilation of a century of income data and two centuries of wealth data into the World Top Incomes Database (WTID) (see p. 826 and Piketty and Saez, p. 838). Science 23 May 2014: Vol. 344 no. 6186 pp. 818-821 DOI: 10.1126/science.344.6186.818 Via NESS No comment yet. Scooped by Bernard Ryefield ## Nigel Goldenfeld: We Need a Theory of Life I was intrigued when Carl Woese told me his collaboration with University of Illinois physicist Nigel Goldenfeld was the most productive one of his entire career, and was pleased to finally run into Goldenfeld last September at lunch in the courtyard... No comment yet. Scooped by Bernard Ryefield ## ▶ Large-Scale Structure in Networks - YouTube Mark Newman May 2, 2014 Annual Science Board Symposium and Meeting Complexity: Theory and Practice Eli Levine's curator insight, To know the structure is to know a HUMONGOUS part of the function and, thus, the ability to predict.  It seems to me to be a large fractal pattern of clusters, nodes and connections (but, that is just in my relatively uneducated eye). Never forget, though, that there are important qualitative aspects to networks (think of defacto qualities of the nodes, groups of nodes and the connections amongst them).  Very important for social and/or ecological/causal relation networks (essentially, a network that outlines and maps accurately the function of a system and all of the flows of information and material resources). Really cool stuff here. Scooped by Bernard Ryefield ## Dynamics of Complex Systems - Y. Bar-Yam Textbook for seminar/course on complex systems. View full text in PDF format The study of complex systems in a unified framework has become recognized in recent years as a new scientific discipline, the ultimate of interdisciplinary fields. Breaking down the barriers between physics, chemistry and biology and the so-called soft sciences of psychology, sociology, economics, and anthropology, this text explores the universal physical and mathematical principles that govern the emergence of complex systems from simple components. No comment yet. Scooped by Bernard Ryefield ## Agent Based Modeling of Complex Adaptive Systems - advanced : Course Home Building on Complex Adaptive Systems theory and basic Agent Based Modeling knowledge presented in SPM4530, the Advanced course will focus on the model development process. The students are expected to conceptualize, develop and verify a model during the course, individually or in a group. The modeling tasks will be, as much as possible, based on real life research problems, formulated by various research groups from within and outside the faculty. Study Goals The main goal of the course is to learn how to form a modeling question, perform a system decomposition, conceptualize and formalize the system elements, implement and verify the simulation and validate an Agent Based Model of a socio-technical system. No comment yet. Scooped by Bernard Ryefield ## Dynamics of Complex Systems | NECSI | Libros y ... Complejidady Economía's insight: Full Online Text (Dynamics of Complex Systems | NECSI | @scoopit http://t.co/sVePaP2sG2) No comment yet. Scooped by Bernard Ryefield ## Insights from the Wikipedia Contest (IEEE Contest for Data Mining 2011) The Wikimedia Foundation has recently observed that newly joining editors on Wikipedia are increasingly failing to integrate into the Wikipedia editors' community, i.e. the community is becoming increasingly harder to penetrate. To sustain healthy growth of the community, the Wikimedia Foundation aims to quantitatively understand the factors that determine the editing behavior, and explain why most new editors become inactive soon after joining. As a step towards this broader goal, the Wikimedia foundation sponsored the ICDM (IEEE International Conference for Data Mining) contest for the year 2011. The objective for the participants was to develop models to predict the number of edits that an editor will make in future five months based on the editing history of the editor. Here we describe the approach we followed for developing predictive models towards this goal, the results that we obtained and the modeling insights that we gained from this exercise. In addition, towards the broader goal of Wikimedia Foundation, we also summarize the factors that emerged during our model building exercise as powerful predictors of future editing activity. No comment yet. Scooped by Bernard Ryefield ## Complex contagion process in spreading of online innovation Diffusion of innovation can be interpreted as a social spreading phenomena governed by the impact of media and social interactions. Although these mechanisms have been identified by quantitative theories, their role and relative importance are not entirely understood, since empirical verification has so far been hindered by the lack of appropriate data. Here we analyse a dataset recording the spreading dynamics of the world's largest Voice over Internet Protocol service to empirically support the assumptions behind models of social contagion. We show that the probability of spontaneous service adoption is constant, the probability of adoption via social influence is linearly proportional to the fraction of adopting neighbours, and the probability of service termination is time-invariant and independent of the behaviour of peers. By implementing the detected diffusion mechanisms into a dynamical agent-based model, we are able to emulate the adoption dynamics of the service in several countries worldwide. This approach enables us to make medium-term predictions of service adoption and disclose dependencies between the dynamics of innovation spreading and the socioeconomic development of a country. No comment yet. Scooped by Bernard Ryefield ## Saving Human Lives: What Complexity Science and Information Systems can Contribute We discuss models and data of crowd disasters, crime, terrorism, war and disease spreading to show that conventional recipes, such as deterrence strategies, are often not effective and sufficient to contain them. Many common approaches do not provide a good picture of the actual system behavior, because they neglect feedback loops, instabilities and cascade effects. The complex and often counter-intuitive behavior of social systems and their macro-level collective dynamics can be better understood by means of complexity science. We highlight that a suitable system design and management can help to stop undesirable cascade effects and to enable favorable kinds of self-organization in the system. In such a way, complexity science can help to save human lives. No comment yet. Scooped by Bernard Ryefield ## Large-scale Fluctuations of Lyapunov Exponents in Diffusive Systems We present a general formalism for computing Lyapunov exponents and their fluctuations in spatially extended systems described by diffusive fluctuating hydrodynamics, thus extending the concepts of dynamical system theory to a broad range of non-equilibrium systems. Our analytical results compare favorably with simulations of a lattice model of heat conduction. We further show how the computation of Lyapunov exponents for the Symmetric Simple Exclusion Process relates to damage spreading and to a two-species pair annihilation process, for which our formalism yields new finite size results. No comment yet. Scooped by Bernard Ryefield ## Combining segregation and integration: Schelling model dynamics for heterogeneous population The Schelling model is a simple agent based model that demonstrates how individuals' relocation decisions generate residential segregation in cities. Agents belong to one of two groups and occupy cells of rectangular space. Agents react to the fraction of agents of their own group within the neighborhood around their cell. Agents stay put when this fraction is above a given tolerance threshold but seek a new location if the fraction is below the threshold. The model is well known for its tipping point behavior: an initial random (integrated) pattern remains integrated when the tolerance threshold is below 1/3 but becomes segregated when the tolerance threshold is above 1/3. In this paper, we demonstrate that the variety of the Schelling model steady patterns is richer than the segregation-integration dichotomy and contains patterns that consist of segregated patches for each of the two groups alongside patches where both groups are spatially integrated. We obtain such patterns by considering a general version of the model in which the mechanisms of agents' interactions remain the same but the tolerance threshold varies between agents of both groups. We show that the model produces patterns of mixed integration and segregation when the tolerance threshold of most agents is either below the tipping point or above 2/3. In these cases, the mixed patterns are relatively insensitive to the model's parameters. No comment yet. Rescooped by Bernard Ryefield from Global Brain ## Game Theory Makes New Predictions for Evolution | Simons Foundation An insight borrowed from computer science suggests that evolution values both fitness and diversity. Via Spaceweaver No comment yet. Rescooped by Bernard Ryefield from Papers ## The direction of evolution: The rise of cooperative organization Two great trends are evident in the evolution of life on Earth: towards increasing diversification and towards increasing integration. Diversification has spread living processes across the planet, progressively increasing the range of environments and free energy sources exploited by life. Integration has proceeded through a stepwise process in which living entities at one level are integrated into cooperative groups that become larger-scale entities at the next level, and so on, producing cooperative organizations of increasing scale (for example, cooperative groups of simple cells gave rise to the more complex eukaryote cells, groups of these gave rise to multi-cellular organisms, and cooperative groups of these organisms produced animal societies). The trend towards increasing integration has continued during human evolution with the progressive increase in the scale of human groups and societies. The trends towards increasing diversification and integration are both driven by selection. An understanding of the trajectory and causal drivers of the trends suggests that they are likely to culminate in the emergence of a global entity. This entity would emerge from the integration of the living processes, matter, energy and technology of the planet into a global cooperative organization. Such an integration of the results of previous diversifications would enable the global entity to exploit the widest possible range of resources across the varied circumstances of the planet. This paper demonstrates that it's case for directionality meets the tests and criticisms that have proven fatal to previous claims for directionality in evolution. The direction of evolution: The rise of cooperative organization John E. Stewart Biosystems Available online 1 June 2014 http://dx.doi.org/10.1016/j.biosystems.2014.05.006 Via Complexity Digest Eli Levine's curator insight, Cooperation is the best way to improve, sustain, maintain, and repair.  Competition is what drives everyone and everything towards something different, be it competition for resources or competition against the elements around us. I don't get what the point of competition amongst the species is for.  Part of cooperation, after all, is knowing what works, learning about what could work better or doesn't work, and then letting the negative or sub-optimal slip back beneath the waves of ignorance, such that the new ways can rise to prominence. Change is the only constant in this universe of universes. Yet cooperation, I think, yields the higher and stronger of the universal structures that are out there, even if it means that there are still losers and winners.  The only difference is the level of consent and consensus that's reached within the social, ecological, economical, and/or political landscape.  One way works towards what is best.  The other way simply yields what is best at competing, which is not the same as being the actual best solution to a given problem or condition. Luciano Lampi's curator insight, is this the end of stove pipes? Ra's curator insight, Have I been reading too much science fiction? Scooped by Bernard Ryefield ## NSF Research Helps UNESCO Preserve Subaks in Bali Immersed in the world of Balinese water temples and cooperative farms, Anthropologist J. Stephen Lansing’s NSF funded research helped win UNESCO World Heritage Site status for Bali’s subaks. No comment yet. Scooped by Bernard Ryefield ## Designing Complex Dynamics in Cellular Automata with Memory Since their inception at Macy conferences in later 1940s complex systems remain the most controversial topic of inter-disciplinary sciences. The term complex system' is the most vague and liberally used scientific term. Using elementary cellular automata (ECA), and exploiting the CA classification, we demonstrate elusiveness of complexity' by shifting space-time dynamics of the automata from simple to complex by enriching cells with {\it memory}. This way, we can transform any ECA class to another ECA class --- without changing skeleton of cell-state transition function --- and vice versa by just selecting a right kind of memory. A systematic analysis display that memory helps `discover' hidden information and behaviour on trivial --- uniform, periodic, and non-trivial --- chaotic, complex --- dynamical systems. No comment yet. Suggested by Samir ## ECCS 2014 Satellite Workshop on Robustness, Adaptability and Critical Transitions in Living Systems. Submit your Abstract at http://seis.bristol.ac.uk/~fs13378/eccs_2014_livingsys.html No comment yet. Scooped by Bernard Ryefield ## Networks as a Privileged Way to Develop Mesoscopic Level Approaches in Systems Biology The methodologies advocated in computational biology are in many cases proper system-level approaches. These methodologies are variously connected to the notion of “mesosystem” and thus on the focus on relational structures that are at the basis of biological regulation. Here, I describe how the formalization of biological systems by means of graph theory constitutes an extremely fruitful approach to biology. I suggest the epistemological relevance of the notion of graph resides in its multilevel character allowing for a natural “middle-out” causation making largely obsolete the traditional opposition between “top-down” and “bottom-up” styles of reasoning, so fulfilling the foundation dream of systems science of a direct link between systems analysis and the underlying physical reality. No comment yet. Scooped by Bernard Ryefield ## COLLECTIVE ADAPTATIVE SYSTEMS SYNTHESIS WITH NON-ZERO-SUM GAMES The objective of CASSTING is to develop a novel approach for analysing and designing collective adaptive systems in their totality, by setting up a game theoretic framework. Here components are viewed as players, their behaviour is captured by strategies, system runs are plays, and specifications are winning conditions. We will develop formalisms for modelling collective adaptive systems as games, and algorithms for synthesising optimal strategies (and components). No comment yet. Scooped by Bernard Ryefield ## 3rd International Conference on Complex Dynamical Systems and Their Applications: New Mathematical Concepts and Applications | The European Mathematical Society 3rd International Conference on Complex Dynamical Systems and Their Applications: New Mathematical Concepts and... http://t.co/Wd83nLP9Ik No comment yet. Scooped by Bernard Ryefield ## The influence of persuasion in opinion formation and polarization We present a model that explores the influence of persuasion in a population of agents with positive and negative opinion orientations. The opinion of each agent is represented by an integer number k that expresses its level of agreement on a given issue, from totally against k=-M to totally in favor k=M. Same-orientation agents persuade each other with probability p, becoming more extreme, while opposite-orientation agents become more moderate as they reach a compromise with probability q. The population initially evolves to (a) a polarized state for r=p/q>1, where opinions' distribution is peaked at the extreme values k=±M, or (b) a centralized state for r<1, with most opinions around k=±1. When r»1, polarization lasts for a time that diverges as rMlnN, where N is the population's size. Finally, an extremist consensus (k=M or -M) is reached in a time that scales as r-1 for r«1. No comment yet. Rescooped by Bernard Ryefield from Talks ## Nonlinear Dynamics and Chaos - Steven Strogatz, Cornell University - YouTube This course of 25 lectures, filmed at Cornell University in Spring 2014, is intended for newcomers to nonlinear dynamics and chaos. It closely follows Prof. Strogatz's book, "Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering." The mathematical treatment is friendly and informal, but still careful. Analytical methods, concrete examples, and geometric intuition are stressed. The theory is developed systematically, starting with first-order differential equations and their bifurcations, followed by phase plane analysis, limit cycles and their bifurcations, and culminating with the Lorenz equations, chaos, iterated maps, period doubling, renormalization, fractals, and strange attractors. A unique feature of the course is its emphasis on applications. These include airplane wing vibrations, biological rhythms, insect outbreaks, chemical oscillators, chaotic waterwheels, and even a technique for using chaos to send secret messages. In each case, the scientific background is explained at an elementary level and closely integrated with the mathematical theory. The theoretical work is enlivened by frequent use of computer graphics, simulations, and videotaped demonstrations of nonlinear phenomena. The essential prerequisite is single-variable calculus, including curve sketching, Taylor series, and separable differential equations. In a few places, multivariable calculus (partial derivatives, Jacobian matrix, divergence theorem) and linear algebra (eigenvalues and eigenvectors) are used. Fourier analysis is not assumed, and is developed where needed. Introductory physics is used throughout. Other scientific prerequisites would depend on the applications considered, but in all cases, a first course should be adequate preparation Nonlinear Dynamics and Chaos - Steven Strogatz, Cornell University
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4551045298576355, "perplexity": 2557.484533289482}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-44/segments/1476988725470.56/warc/CC-MAIN-20161020183845-00447-ip-10-171-6-4.ec2.internal.warc.gz"}
http://www.zazzle.com/states+calendars
Showing All Results 945 results Page 1 of 16 Related Searches: united, america, lantern Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo Got it! We won't show you this product again! Undo No matches for Showing All Results 945 results Page 1 of 16
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8189893364906311, "perplexity": 4434.912735365363}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-10/segments/1394678702437/warc/CC-MAIN-20140313024502-00076-ip-10-183-142-35.ec2.internal.warc.gz"}
https://chemistry.stackexchange.com/questions/117590/what-chemical-could-dissolve-low-density-polyethylene
# What chemical could dissolve low-density polyethylene? What chemical could dissolve plastic aside from acetone? Preferably, low-cost and environmentally friendly. Any suggestions would be helpful. I'm targeting for thin plastic (low-density polyethylene). Environmentally friendly for me is that it follows the EPA standards from the USA. I'm targeting for around 150 USD. I'm okay with solvents that make the plastic swell, but a chemical that could dissolve it is preferable. • 150 USD for... what? Per liter, per tonne, per day? Can you say anything at all about what you're trying to do? – jeffB Jul 3 at 1:50 • The safest solvent might be supercritical carbon dioxide, $$\ce{CO2}$$. Supercritical carbon dioxide is also used in "green" dry-cleaning. Use for dissolving polymers requires a pressure vessel for ~5 to 50 MPa, able to work at ~40 to 100°C. There are (paywalled) articles at ACS and NCBI, if you want to investigate further. Supercritical $$\ce{CO2}$$, while not very toxic, can cause asphyxiation above ~5%, and the container is, effectively, a bomb containing a warm or hot pressurized fluid.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 2, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.3059277832508087, "perplexity": 5883.97830950849}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-43/segments/1570986669057.0/warc/CC-MAIN-20191016163146-20191016190646-00379.warc.gz"}
https://au.mathworks.com/help/matlab/matlab_oop/scoping-classes-with-packages.html
Packages Create Namespaces Package Folders Packages are special folders that can contain class folders, function, and class definition files, and other packages. The names of classes and functions are scoped to the package folder. A package is a namespace within which names must be unique. Function and class names must be unique only within the package. Using a package provides a means to organize classes and functions. Packages also enable you to reuse the names of classes and functions in different packages. Note Packages are not supported for classes created before MATLAB® Version 7.6 (that is, classes that do not use classdef). Package folders always begin with the + character. For example, +mypack +mypack/pkfcn.m % a package function +mypack/@myClass % class folder in a package The parent of the top-level package folder must be on the MATLAB path. Listing the Contents of a Package List the contents of a package using the help command: help event Contents of event: EventData - event.EVENTDATA Base class for event data PropertyEvent - event.PROPERTYEVENT Event data for object property events listener - event.LISTENER Listener object proplistener - event.PROPLISTENER Listener object for property events You can also use the what command: what event Classes in directory Y:xxx\matlab\toolbox\matlab\lang\+event EventData PropertyEvent listener proplistener Internal Packages MathWorks® reserves the use of packages named internal for utility functions used by internal MATLAB code. Functions that belong to an internal package are intended for MathWorks use only. Using functions or classes that belong to an internal package is discouraged. These functions and classes are not guaranteed to work in a consistent manner from one release to the next. Any of these functions and classes might be removed from the MATLAB software in any subsequent release without notice and without documentation in the product release notes. Referencing Package Members Within Packages All references to packages, functions, and classes in the package must use the package name prefix, unless you import the package. (See Import Classes.) For example, call this package function: +mypack/pkfcn.m With this syntax: z = mypack.pkfcn(x,y); Definitions do not use the package prefix. For example, the function definition line of the pkfcn.m function would include only the function name: function z = pkfcn(x,y) Define a package class with only the class name: classdef myClass but call it with the package prefix: obj = mypack.myClass(arg1,arg2,...); Calling class methods does not require the package name because you have an object of the class. You can use dot or function notation: obj.myMethod(arg) myMethod(obj,arg) A static method requires the full class name, which includes the package name: mypack.myClass.stMethod(arg) Referencing Package Members from Outside the Package Functions, classes, and other packages contained in a package are scoped to that package. To reference any of the package members, prefix the package name to the member name, separated by a dot. For example, the following statement creates an instance of MyClass, which is contained in mypack package. obj = mypack.MyClass; Accessing Class Members — Various Scenarios This section shows you how to access various package members from outside a package. Suppose that you have a package mypack with the following contents: +mypack +mypack/myFcn.m +mypack/@MyFirstClass +mypack/@MyFirstClass/myFcn.m +mypack/@MyFirstClass/otherFcn.m +mypack/@MyFirstClass/MyFirstClass.m +mypack/@MySecondClass +mypack/@MySecondClass/MySecondClass.m +mypack/+mysubpack +mypack/+mysubpack/myFcn.m Invoke the myFcn function in mypack: mypack.myFcn(arg) Create an instance of each class in mypack: obj1 = mypack.MyFirstClass; obj2 = mypack.MySecondClass(arg); Invoke the myFcn function that is in the package mysubpack: mypack.mysubpack.myFcn(arg1,arg2); If mypack.MyFirstClass has a method called myFcn, call it like any method call on an object: obj = mypack.MyFirstClass; myFcn(obj,arg); If mypack.MyFirstClass has a property called MyProp, assign it using dot notation and the object: obj = mypack.MyFirstClass; obj.MyProp = x; Packages and the MATLAB Path You cannot add package folders to the MATLAB path, but you must add the package parent folder to the MATLAB path. Package members are not accessible if the package parent folder is not on the MATLAB path, even if the package folder is the current folder. Making the package folder the current folder is not sufficient to add the package parent folder to the path. Package members remain scoped to the package. Always refer to the package members using the package name. Alternatively, import the package into the function in which you call the package member, see Import Classes. Package folders do not shadow other package folders that are positioned later on the path, unlike classes, which do shadow other classes. If two or more packages have the same name, MATLAB treats them all as one package. If redundantly named packages in different path folders define the same function name, then MATLAB finds only one of these functions. Resolving Redundant Names Suppose a package and a class have the same name. For example: fldr_1/+foo fldr_2/@foo/foo.m A call to which foo returns the path to the executable class constructor: >> which foo fldr_2/@foo/foo.m A function and a package can have the same name. However, a package name by itself is not an identifier. Therefore, if a redundant name occurs alone, it identifies the function. Executing a package name alone returns an error. Package Functions vs. Static Methods In cases where a package and a class have the same name, a package function takes precedence over a static method. For example, path folder fldrA contains a package function and path folder fldrB contains a class static method: fldrA/+foo/bar.m % bar is a function in package foo fldrB/@foo/bar.m % bar is a static method of class foo A call to which foo.bar returns the path to the package function: which foo.bar fldrA\+foo\bar.m % package function In cases where the same path folder contains both package and class folders with the same name, the package function takes precedence over the static method. fldr/@foo/bar.m % bar is a static method of class foo fldr/+foo/bar.m % bar is a function in package foo A call to which foo.bar returns the path to the package function: which foo.bar fldr/+foo/bar.m If a path folder fldr contains a classdef file foo that defines a static method bar and the same folder contains a package +foo that contains a package function bar. fldr/foo.m % bar is a static method of class foo fldr/+foo/bar.m % bar is a function in package foo A call to which foo.bar returns the path to the package function: which foo.bar fldr/+foo/bar.m
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.5232376456260681, "perplexity": 3831.0282116128465}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320300616.11/warc/CC-MAIN-20220117182124-20220117212124-00681.warc.gz"}
https://www.rheumatologyadvisor.com/home/decision-support-in-medicine/imaging/bells-palsy/
### I. What imaging technique is first-line for this diagnosis? • MRI of the brain with contrast, 7th cranial nerve protocol ### II. Describe the advantages and disadvantages of this technique for diagnosis of Bell's palsy. • Can define and characterize the 7th cranial nerve from the brainstem throught the parotid in all three planes • Revealing of inflammation along the 7th cranial nerve • Time consuming • Requires significant patient cooperation to minimize motion artifact. ### III. What are the contraindications for the first-line imaging technique? Contraindicated in patients with non-MR-compatible hardware. ### IV. What alternative imaging techniques are available? No alternative imaging techniques exist. Not applicable Not applicable.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8408483266830444, "perplexity": 18966.414690720198}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-39/segments/1631780060877.21/warc/CC-MAIN-20210928153533-20210928183533-00315.warc.gz"}
http://mathhelpforum.com/statistics/182541-reliability-test.html
# Math Help - Reliability of a Test 1. ## Reliability of a Test 15%of the U.S. population has math fever. If an individual has math fever, a testaccurately identifies the condition 80% of the time. If an individual does nothave math fever, the test accurately identifies the individual as not havingmath fever 90% of the time. The following list describes theseoutcomes: MathFever 15% PosTest 80% NegTest 20% NoMath Fever 85% PosTest 10% NegTest 90% What percent of the time is the test inaccurate? What is the probability of a false positive? What is the probability that the individual does not have math fever given that thetest is negative? I know that the answers should be: 0.885, 0.115, 9623 but I'm not sure how to set up any of these 2. What does it mean to have a false positive? You know that a a test says you have the fever if you actually have the fever 80% of the time. It also says you don't have the fever when you don't actually have the fever 90% of the time. Consider now that you don't actually have the fever but it gives you a positive test anyway. You should have something like [Not Sick] * [Positive Test] + [Sick] * [Negative Test] = [False Positive] In other words, the given how many people are sick or not sick and the chances that the test gets it wrong, you should be able to figure out its false positive rate. I got the correct answer, and from the expression above, so should you. The key here is understanding why it is set up that way. Do you see why? 3. Ahh yes, that certainly makes much more sense. How would you approach the inaccurate question and what would be the difference? 4. Think about the definition. A false positive is precisely that: how likely is a test to indicate the false of what is actually (positively) happening? Inaccuracy is how off the mark a test is. This is about the positive rate itself. Using the above algorithm, just think how many of the population that are sick will test sick or how many of the population that aren't sick will test not sick? The answer falls right out. As for the last question, it is a conditional probability. What do you know about that so far? 5. Thanks so much Bryan! I finally got it! (I didn't realize the last one was conditional probability--thanks!)
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8753077983856201, "perplexity": 709.3953063031621}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-10/segments/1393999651529/warc/CC-MAIN-20140305060731-00057-ip-10-183-142-35.ec2.internal.warc.gz"}
https://worldwidescience.org/topicpages/l/levels+descriptive+note.html
#### Sample records for levels descriptive note 1. Immigrant fertility in Sweden, 2000-2011: A descriptive note Directory of Open Access Journals (Sweden) 2014-03-01 Full Text Available Background: Modern Scandinavian population registers provide excellent data sources that allow a user to quickly gain an impression of the level of fertility and its structure across subpopulations. This may also allow the analyst to check a feature of the much-cited disruption hypothesis, at least in part. Objective: The purpose of this note is to exploit this potential to give an overview of the structure of recent total fertility after immigration to Sweden from various groups of sending countries, separately for males and females. In the process we demonstrate to what extent the post-migration fertility compensation which is part of the fertility disruption hypothesis is fulfilled in our study population. Due to the nature of our data we have refrained from studying fertility before migration. Methods: Based on data from a combination of two Swedish administrative registers (the Historic Population Register and the Multi-Generation Register that cover both men and women in the entire population for the years 2000-2011, we compute and plot TFR-like age-cumulated fertility levels, specific for years since immigration, for six groups of sending countries, separately for men and women. Results: We find that the post-migration fertility compensation specified as part of the fertility disruption hypothesis is visibly confirmed in our Swedish study population for female European immigrants from non-EU countries and for female immigrants from non-European countries with a low or medium UN Human Development Index, but not so for other female immigrants, i.e. not for those who come from a Nordic country or from a non-Nordic EU country, and not for female immigrants from a non-European country with a high Human Development Index, including the United States. We find mild but less conclusive evidence for the same feature for males. Conclusions: This shows that at least as far as post-migration fertility compensation is concerned, the disruption 2. Notes on Japanese Rhizocephala, with description of two new species NARCIS (Netherlands) Boschma, H. 1935-01-01 The following notes are based upon material of parasites on two Crustaceans from Japan, viz., Pachygrapsus crassipes Randall and Petrolisthes japonicus de Haan 1). Each of these two species may be infested by two different species of Rhizocephala: on Pachygrapsus crassipes occur the parasites 3. A Note on the Field-Theoretical Description of Superfluids CERN Document Server Andrianopoli, L; Grassi, P A; Trigiante, M 2014-01-01 Recently, a Lagrangian description of superfluids attracted some interest from the fluid/gravity-correspondence viewpoint. In this respect, the work of Dubovksy et al. has proposed a new field theoretical description of fluids, which has several interesting aspects. On another side, we have provided in arXiv:1304.2206 a supersymmetric extension of the original works. In the analysis of the Lagrangian structures a new invariant appeared which, although related to known invariants, provides, in our opinion, a better parametrisation of the fluid dynamics in order to describe the fluid/superfluid phases. 4. Noted Science.gov (United States) Nunberg, Geoffrey 2013-01-01 Considering how much attention people lavish on the technologies of writing--scroll, codex, print, screen--it's striking how little they pay to the technologies for digesting and regurgitating it. One way or another, there's no sector of the modern world that is not saturated with note-taking--the bureaucracy, the liberal professions, the… 5. Taxonomic notes on some Polyglyptini: descriptions of new genus and new species (Homoptera, Membracidae, Smiliinae Directory of Open Access Journals (Sweden) Albino M. Sakakibara 1996-01-01 Full Text Available Taxonomic notes on some Polyglyptini; descriptions of new genus and new species (Homoptera, Membracidae, Smiliinae. The genera Hemiptycha Germar, Metheisa Fowler, Maturnaria Metcalf, Aphetea Fowler, Dioclophara Kirkaldy, and Phormophora Stål, are redescribed; Creonus, gen.n. (type species: Maturna lloydi Funkhouser, 1914, and Aphetea robustula, sp.n. (from Bolivia, are described. Some nomenclatural changes are introduced, as follow: - Hemiptycha Germar, 1833 = Polyrhyssa Stål, 1869, syn.n.: - Hemiptycha cultrata (Coquebert, 1801, comb.n., = Polyglyptodes flavocostatus Haviland, 1925, syn.n., = Polyrhyssa cultrata maculata Fonseca, 1942, syn.n. - Hemiptycha obtecta (Fabricius, 1803 = Hille herbicola Haviland, 1925, syn.n. - Maturnaria ephippigera (Fairmaire, 1846 = Publilia tumulata Buckton, 1903, syn.n., = Metheisa fowleri Funkhouser, 1927, syn.n. - Creonus lloydi (Funkhouser, 1914, comb.n. - Aphetea parvula (Fabricius, 1803, comb.n., = Aphetea affinis Haviland, 1925, syn.n. - Dioclophara Kirkaldy, 1904 = lncolea Goding, 1926, syn.n. - Dioclophara viridula (Fairmaire, 1846 = Maturna multilineata Fonseca, 1942, syn.n. - Dioclophara variegata (Goding, 1926, comb.n. = lncolea viridis Goding, 1926, syn.n. - Phormophora maura (Fabricius, 1803 = Darnis dorsata Fabricius, 1803, syn.n. 6. Self consitent description of nuclear level densitities International Nuclear Information System (INIS) Barranco, M.; Treiner, J. 1980-03-01 We present a self consistent calculation of the nuclear level density based on a generalization to finite temperature of a modified Thomas-Fermi method. A simple expression is derived for the so-called level density parameter a entering the expression of the density of states Xi(E)=Esup(-5/4)e 2 √aE and relating the excitation energy to the temperature E=a T 2 , in terms of nucleons equilibrium densities at T=0 only. One thus avoids the difficulty of adding external constraints to calculate isolated nuclei at finite temperature which are shown to be unstable against particle emission. The role of the nuclear surface is discussed. It is shown that the effective mass of the interaction plays a crucial role in determining the value of a and comparison with experiment confirms the value m*/m > = 1 near the Fermi level obtained through more microscopic analysis 7. Notes on Neotropical Eumeninae, with the description of a new species of Pachodynerus de Saussure (Hymenoptera, Vespidae Directory of Open Access Journals (Sweden) Marcel Gustavo Hermes 2013-06-01 Full Text Available Notes on Neotropical Eumeninae, with the description of a new species of Pachodynerus de Saussure (Hymenoptera, Vespidae. Taxonomic information on Neotropical Eumeninae is provided. A new species, Pachodynerus fessatus sp. nov. is described from southeastern São Paulo, Brazil. Additional material of Pachodynerus sericeus (Fox was examined, representing the first further specimens after the original description and including the previously unknown male. The examination of new material of the genus Stenonartonia adds some new distribution records and shows some previously unrecorded individual variation for some species. The males of Stenonartonia guaraya Garcete-Barrett and Stenonartonia rejectoides Garcete-Barrett are described for the first time. 8. Taxonomic notes of Hemixantha (Diptera: Richardiidae with description of a new species from the Amazon Region Directory of Open Access Journals (Sweden) Lisiane Dilli Wentd 2014-02-01 Full Text Available Richardiidae is represented by ca. 180 species in 33 genera. A new species of Hemixantha, now totaling 20 species, is described from the Amazon Region, state of Pará, Brazil, and from Teoponte, Bolivia. Hemixantha maculosa sp. nov., H. picta Hennig, 1937 and H. pulchripennis Hendel, 1911 are richardiids known to have an extensively banded wing pattern. Hemixantha maculosa sp. nov. represents the first record of Hemixantha from the state of Pará and this species can be distinguished from H. picta and H. pulchripennis by anepisternum and dorsal half of anepimeron bright yellow and vein R2+3 slightly sinuous. Besides the hyaline areas on the wing of H. maculosa are very distinct from these other species, as follows: (1 subtrapezoidal median area in cell cua1 extended anteriorly to vein M1, separated from hyaline area in cell dm; (2 two oblique areas from costal margin almost to or beyond vein R4+5, one proximal to and one distal to level of crossvein r-m; (3 a circular area on basal part of cell r4+5 and (4 a triangular area on basal region of cell m1. An identification key, diagnoses, description and redescriptions, and illustrations to the species of Hemixantha with an extensively banded wing pattern are provided. 9. Visual space perception at different levels of depth description Czech Academy of Sciences Publication Activity Database 2015-01-01 Roč. 77, č. 6 (2015), 2098–2107 ISSN 1943-3921 R&D Projects: GA ČR GA13-28709S Institutional support: RVO:68081740 Keywords : visual space perception * Depth scales * Level of description Subject RIV: AN - Psychology Impact factor: 1.782, year: 2015 10. A note on the translation of conceptual data models into description logics: disjointness and covering assumptions CSIR Research Space (South Africa) Casini, G 2012-10-01 Full Text Available possibilities for conceptual data modeling. It also raises the question of how existing conceptual models using ER, UML or ORM could be translated into Description Logics (DLs), a family of logics that have proved to be particularly appropriate for formalizing... 11. Multi-level barriers to LTBI treatment: a research note. Science.gov (United States) Hill, Linda; Blumberg, Elaine; Sipan, Carol; Schmitz, Katharine; West, Joshua; Kelley, Norma; Hovell, Melbourne 2010-08-01 This study describes the barriers to effective and timely LTBI treatment encountered in a research study on INH adherence in Latino adolescents. Participant study logs were reviewed, results of continuing medical education pretests for medical providers were examined, and participating medical facilities were contacted in order to construct a profile of multi-level barriers to LTBI treatment. A total of 285 TST positive Latino (96%) high school students were recruited into the trial. We encountered a lack of understanding of the gravity of tuberculosis infection among both the public and providers of health care. Parents and adolescents cited competing priorities, transportation problems and financial constraints as reasons for non-compliance. Improved education of the public and physicians is needed regarding the gravity of the disease and the value of treatment, as well as public and financial support for LTBI treatment by both the government and the medical community. 12. Description and biological notes of the first species of Xenos (Strepsiptera:Stylopidae) parasitic in Polistes carnifex F. (Hymenoptera:Vespidae) in Mexico DEFF Research Database (Denmark) Kathirithamby, J.; Hughes, David P. 2006-01-01 A description and biological notes on the first species of Xenos (X. hamiltoni) (Strepsiptera: Stylopidae) parasitic in Polistes carnifex F. from Mexico is given. A list of Strepsiptera and their hosts from Mexico is provided.... 13. Effect of skill level on recall of visually presented patterns of musical notes. Science.gov (United States) Kalakoski, Virpi 2007-04-01 Expertise effects in music were studied in a new task: the construction of mental representations from separate fragments. Groups of expert musicians and non-musicians were asked to recall note patterns presented visually note by note. Skill-level, musical well-formedness of the note patterns and presentation mode were varied. The musicians recalled note patterns better than the non-musicians, even though the presentation was visual and successive. Furthermore, only musicians' performance was affected by musical well-formedness of the note patterns when visual gestalt properties, verbal rehearsability, and familiarity of the stimuli were controlled. Musicians were also able to use letter names referring to notes as efficiently as visual notes, which indicates that the better recall of musicians cannot be explained by perceptual visual chunking. These results and the effect of skill level on the distribution of recall errors indicate that the ability to chunk incoming information into meaningful units does not require that complete familiar patterns are accessible to encoding processes, yet previous knowledge stored in long-term memory affects representation construction in working memory. The present method offers a new reliable tool, and its implications to the research on construction of representations and musical imagery are discussed. 14. Exercise in completing design information questionnaire for model research reactor: model description, notes, questionnaire International Nuclear Information System (INIS) Bellinger, J.; Ho, T. 1989-01-01 The document which defines the inspection measures which the IAEA can deploy at any given nuclear facility is known as the Facility Attachment. For the Agency to negotiate an effective Facility Attachment it must have available certain design information, including the facility's identity, capacity and location; the form, location and flow of nuclear material and the layout of important items of equipment; and a description of the features and procedures relating to nuclear material accountancy, containment and surveillance. In practice such information is solicited in a format, standardized for each facility type, known as the Design Information Questionnaire or the D.I.Q. The nuclear activities used as a model in this course are those of a fictitious country called Pacifica. These nuclear activities bear some resemblance to those at the Australian Atomic Energy Commission's Research Establishment at Lucas Heights. Specifically, Pacifica has a 10 MW heavy water cooled and moderated research reactor using enriched uranium fuel which is very similar to the HIFAR reactor. The reactor and the associated laboratories are described and the Design Information Questionnaire for them is completed. figs., tabs 15. Translation of a High-Level Temporal Model into Lower Level Models: Impact of Modelling at Different Description Levels DEFF Research Database (Denmark) Kraft, Peter; Sørensen, Jens Otto 2001-01-01 The paper attempts theoretically to clarify the interrelation between various levels of descriptions used in the modelling and the programming of information systems. We suggest an analysis where we characterise the description levels with respect to how precisely they may handle information abou...... and other textual models. We also consider the aptness of models that include procedural mechanisms such as active and object databases... 16. A thermodynamic description of quarks at the subquark level International Nuclear Information System (INIS) Fitzpatrick, G.L. 1985-01-01 A thermodynamic basis for the description of quarks at the subquark level is proposed. It is suggested that subquarks are ultrarelativistic objects confined to the quark radius R. Thus they experience accelerations of the order a≅c/sup 2//R. But this means that information excluding horizons (iota) comparable to quark radii R, namely iota≅c/sup 2//a≅R, are present. Such horizons force us to describe quarks, at the subquark level, via thermodynamics. This thermodynamic description must involve unconventional negative energy Rindler vacua, rather than the conventional zero energy Minkowski vacuum. In an average thermodynamic sense, these Rindler vacua cancel excess kinetic energy of the subquarks, thereby removing an objection to theories involving subquarks. In any such theory it is necessary to assign an Unruh temperature T, where kT≅(h/2πc)a≅(hc/2πR), to the subquark matter. The author argues that T must be the temperature of the early universe phase transition (probably first order) at which quarks condensed into hadrons. Thus quarks have a temperature T independent of hadron mass. He shows how quark properties may be derived in the foregoing thermodynamic context 17. Notes on the genus Abantis Hopffer, 1855 with description of two new species (Lepidoptera: Hesperiidae, Pyrginae Directory of Open Access Journals (Sweden) Claudio Belcastro 2016-06-01 Full Text Available The Afrotropical genus Abantis Hopffer, 1855 includes about 23 species so far, all quite rare. In Evans’ Catalogue of the African Hesperiidae only 14 species were listed but, apart from a few newly discovered ones, some of the subspecies indicated over there have subsequently been raised to species level, although they look very similar in facies with previously recognised taxa. This article shows the presence of an unrecognised species similar to A. lucretia Druce, 1909 in West Africa, namely A. fabiana sp. n., and another one similar to both A. contigua Evans, 1937 and A. elegantula Mabille, 1890 in Central Africa, namely A. torbeni sp. n. The rarity in collections of almost all forest-dwelling Abantis spp. is reported as connected to their habit of flying high in the forest canopy, but the authors have no conclusive evidence of it. Most Abantis are hill-toppers, mud lovers and are attracted by dead animals (fish and turtles, although some of them are also flower-visiting. 18. Description of the Seibersdorf incineration plant for low level waste International Nuclear Information System (INIS) Chalupa, G.; Petschnik, G. 1986-09-01 After a description of the design and the construction principles of the incinerator building, the furnace and its attached auxilary devices are explained. The incinerator is layed out for low level wastes. It has a vertical furnace, operates with discontinuous feeding for trashes with heat-values between 600 and 10000 kcal/kg waste. The maximum throughput amounts 40 kg/h. The purification of the off-gas is guaranteed by a multistage filter system: 2 stages with ceramic candles, cooling column and a HEPA-filter system. The control of the off-gas cleaning is carried out by a stack instrumentation, consisting of an aerosol-, gas-, Iodine- and Tritium-monitor; the building is surveilled by doserate- and aerosolmonitors. (Author) 19. Automated generation of partial Markov chain from high level descriptions International Nuclear Information System (INIS) Brameret, P.-A.; Rauzy, A.; Roussel, J.-M. 2015-01-01 We propose an algorithm to generate partial Markov chains from high level implicit descriptions, namely AltaRica models. This algorithm relies on two components. First, a variation on Dijkstra's algorithm to compute shortest paths in a graph. Second, the definition of a notion of distance to select which states must be kept and which can be safely discarded. The proposed method solves two problems at once. First, it avoids a manual construction of Markov chains, which is both tedious and error prone. Second, up the price of acceptable approximations, it makes it possible to push back dramatically the exponential blow-up of the size of the resulting chains. We report experimental results that show the efficiency of the proposed approach. - Highlights: • We generate Markov chains from a higher level safety modeling language (AltaRica). • We use a variation on Dijkstra's algorithm to generate partial Markov chains. • Hence we solve two problems: the first problem is the tedious manual construction of Markov chains. • The second problem is the blow-up of the size of the chains, at the cost of decent approximations. • The experimental results highlight the efficiency of the method 20. CHEMICAL REACTIONS ON ADSORBING SURFACE: KINETIC LEVEL OF DESCRIPTION Directory of Open Access Journals (Sweden) P.P.Kostrobii 2003-01-01 Full Text Available Based on the effective Hubbard model we suggest a statistical description of reaction-diffusion processes for bimolecular chemical reactions of gas particles adsorbed on the metallic surface. The system of transport equations for description of particles diffusion as well as reactions is obtained. We carry out the analysis of the contributions of all physical processes to the formation of diffusion coefficients and chemical reactions constants. 1. LLW (Low-Level Waste) Notes, Volume 13, Number 1, February 1998 International Nuclear Information System (INIS) 1998-02-01 LLW Notes is a newsletter distributed to Low-Level Radioactive Waste Forum Participants and other state and compact officials. The LLW Forum provides an opportunity for state and compact officials to share information with one another and to exchange views with officials of federal agencies and other interested parties. This issue focuses on the following topics: DOI approves Ward Valley permit application; Project evidentiary hearings begin in Texas; and Summary judgment motions in California breach of contract action 2. LLW (Low-Level Waste) Notes, Volume 13, Number 1, February 1998 Energy Technology Data Exchange (ETDEWEB) NONE 1998-02-01 LLW Notes is a newsletter distributed to Low-Level Radioactive Waste Forum Participants and other state and compact officials. The LLW Forum provides an opportunity for state and compact officials to share information with one another and to exchange views with officials of federal agencies and other interested parties. This issue focuses on the following topics: DOI approves Ward Valley permit application; Project evidentiary hearings begin in Texas; and Summary judgment motions in California breach of contract action. 3. A Note on Comparing the Power of Test Statistics at Low Significance Levels. Science.gov (United States) Morris, Nathan; Elston, Robert 2011-01-01 It is an obvious fact that the power of a test statistic is dependent upon the significance (alpha) level at which the test is performed. It is perhaps a less obvious fact that the relative performance of two statistics in terms of power is also a function of the alpha level. Through numerous personal discussions, we have noted that even some competent statisticians have the mistaken intuition that relative power comparisons at traditional levels such as α = 0.05 will be roughly similar to relative power comparisons at very low levels, such as the level α = 5 × 10 -8 , which is commonly used in genome-wide association studies. In this brief note, we demonstrate that this notion is in fact quite wrong, especially with respect to comparing tests with differing degrees of freedom. In fact, at very low alpha levels the cost of additional degrees of freedom is often comparatively low. Thus we recommend that statisticians exercise caution when interpreting the results of power comparison studies which use alpha levels that will not be used in practice. 4. A Unified Description of Structural Levels in Biomolecules. Science.gov (United States) Macarulla, Alberto; And Others 1990-01-01 A single, didactic criterion for the description of all biological macromolecules is proposed. This criterion is applicable to globular, fibrous or mixed proteins, as well as to nucleic acids and lipids or polysaccharides. Specific examples are given for all except lipids and polysaccarides. (KR) 5. Translation of a High-Level Temporal Model into Lower Level Models: Impact of Modelling at Different Description Levels DEFF Research Database (Denmark) Kraft, Peter; Sørensen, Jens Otto 2001-01-01 given types of properties, and examine how descriptions on higher levels translate into descriptions on lower levels. Our example looks at temporal properties where the information is concerned with the existence in time. In a high level temporal model with information kept in a three-dimensional space...... the existences in time can be mapped precisely and consistently securing a consistent handling of the temporal properties. We translate the high level temporal model into an entity-relationship model, with the information in a two-dimensional graph, and finally we look at the translations into relational...... and other textual models. We also consider the aptness of models that include procedural mechanisms such as active and object databases... 6. Telidon Videotex presentation level protocol: Augmented picture description instructions Science.gov (United States) Obrien, C. D.; Brown, H. G.; Smirle, J. C.; Lum, Y. F.; Kukulka, J. Z.; Kwan, A. 1982-02-01 The Telidon Videotex system is a method by which graphic and textual information and transactional services can be accessed from information sources by the general public. In order to transmit information to a Telidon terminal at a minimum bandwidth, and in a manner independent of the type of communications channel, a coding scheme was devised which permits the encoding of a picture into the geometric drawing elements which compose it. These picture description instructions are an alpha geometric coding model and are based on the primitives of POINT, LINE, ARC, RECTANGLE, POLYGON, and INCREMENT. Text is encoded as (ASCII) characters along with a supplementary table of accents and special characters. A mosaic shape table is included for compatibility. A detailed specification of the coding scheme and a description of the principles which make it independent of communications channel and display hardware are provided. 7. The first record of Litargus tetraspilotus (Coleoptera, Mycetophagidae in Brazil, with biological notes and complementary description of the species Directory of Open Access Journals (Sweden) Julianne Milléo 2011-06-01 Full Text Available Litargus tetraspilotus LeConte, 1856 was collected feeding on Oidium sp. (Fungi, Ascomycota, Erysiphaceae associated with fruit trees. This is the first time L. tetraspilotus is recorded in Brazil, totaling three species of Mycetophagidae for this country. This study aims to provide a complementary description of this species based on new characters and to present information on its life cycle under laboratory conditions and fluctuation in population in the field. During the period of inventories between July 2004 and August 2006, about every fifteen days, a total of 565 specimens of L. tetraspilotus were collected, with the highest abundance found on citrus plants, with values differing significantly between the two years. The population levels differed between the seasons; spring had the greatest abundance and autumn the least. There was a significant positive correlation of L. tetraspilotus abundance with rainfall and relative humidity. Mycetophagidae, as well as other mycophagous families of Brazilian coleopterans, are barely studied, warranting further future studies of their bioecology and systematics. 8. High-Level Waste System Process Interface Description International Nuclear Information System (INIS) D'Entremont, P.D. 1999-01-01 The High-Level Waste System is a set of six different processes interconnected by pipelines. These processes function as one large treatment plant that receives, stores, and treats high-level wastes from various generators at SRS and converts them into forms suitable for final disposal. The three major forms are borosilicate glass, which will be eventually disposed of in a Federal Repository, Saltstone to be buried on site, and treated water effluent that is released to the environment 9. Notes on the genus Conchapelopia Fittkau (Diptera: Chironomidae: Tanypodinae) from southern China, with description of a new species. Science.gov (United States) Niitsuma, Hiromi; Tang, Hongqu 2017-02-22 Two interesting species, Conchapelopia togamaculosa Sasa & Okazawa and a new species, Conchapelopia brachiata sp. n., were collected from southern China. The male, pupa and larva of the new species are described, and new distributions of the former species are noted. Although the male of the new species is very distinct from that of the former in the hypopygial median volsella, the pupa and larva stunningly resemble those of the former. 10. Descriptions of Improvisational Thinking by Artist-Level Jazz Musicians Science.gov (United States) Norgaard, Martin 2011-01-01 Thought processes of seven artist-level jazz musicians, each of whom recorded an improvised solo, were investigated. Immediately after completing their improvisations, participants listened to recordings of their playing and looked at the notation of their solos as they described in a directed interview the thinking processes that led to the… 11. High-level waste description, inventory and hazard International Nuclear Information System (INIS) Crandall, J.; Hennelly, E.J.; McElroy, J.L. 1983-01-01 High-level nuclear waste (HLW), including its origin, is described and the current differences in definitions discussed. Quantities of defense and commercial radioactive HLW, both volume and curie content, are given. Current waste handling, which is interimin nature, is described for the several sites. The HLW hazard is defined by the times during which various radionuclides are the dominant contributors. The hazard is also compared to that of the ore. Using ICRP-2, which is the legal reference in the US, the hazard of the waste reduces to a level equal to the ore in about 300 years. The disposal plans are summarized and it is shown that regulatory requirements will probably govern disposal operations in such a conservative manner that the risk (product of hazard times probability of release) may well be lower than for any other wastes in existence or perhaps lower than those for any other human endeavor 12. Note: A component-level frequency tunable isolator for vibration-sensitive chips using SMA beams Energy Technology Data Exchange (ETDEWEB) Zhang, Xiaoyong, E-mail: [email protected], E-mail: [email protected]; Yan, Xiaojun, E-mail: [email protected], E-mail: [email protected] [School of Energy and Power Engineering, Beihang University, Beijing 100191 (China); Collaborative Innovation Center of Advanced Aero-Engine, Beijing 100191 (China); National Key Laboratory of Science and Technology on Aero-Engine Aero-Thermodynamics, Beijing 100191 (China); Beijing Key Laboratory of Aero-Engine Structure and Strength, Beijing 100191 (China); Ding, Xin; Wu, Di; Qi, Junlei; Wang, Ruixin; Lu, Siwei [School of Energy and Power Engineering, Beihang University, Beijing 100191 (China) 2016-06-15 This note presents a component-level frequency tunable isolator for vibration-sensitive chips. The isolator employed 8 U-shaped shape memory alloy (SMA) beams to support an isolation island (used for mounting chips). Due to the temperature-induced Young’s modulus variation of SMA, the system stiffness of the isolator can be controlled through heating the SMA beams. In such a way, the natural frequency of the isolator can be tuned. A prototype was fabricated to evaluate the concept. The test results show that the natural frequency of the isolator can be tuned in the range of 64 Hz–97 Hz by applying different heating strategies. Moreover, resonant vibration can be suppressed significantly (the transmissibility decreases about 65% near the resonant frequency) using a real-time tuning method. 13. Note: A component-level frequency tunable isolator for vibration-sensitive chips using SMA beams International Nuclear Information System (INIS) Zhang, Xiaoyong; Yan, Xiaojun; Ding, Xin; Wu, Di; Qi, Junlei; Wang, Ruixin; Lu, Siwei 2016-01-01 This note presents a component-level frequency tunable isolator for vibration-sensitive chips. The isolator employed 8 U-shaped shape memory alloy (SMA) beams to support an isolation island (used for mounting chips). Due to the temperature-induced Young’s modulus variation of SMA, the system stiffness of the isolator can be controlled through heating the SMA beams. In such a way, the natural frequency of the isolator can be tuned. A prototype was fabricated to evaluate the concept. The test results show that the natural frequency of the isolator can be tuned in the range of 64 Hz–97 Hz by applying different heating strategies. Moreover, resonant vibration can be suppressed significantly (the transmissibility decreases about 65% near the resonant frequency) using a real-time tuning method. 14. Conceptual modular description of the high-level waste management system for system studies model development International Nuclear Information System (INIS) McKee, R.W.; Young, J.R.; Konzek, G.J. 1992-08-01 This document presents modular descriptions of possible alternative components of the federal high-level radioactive waste management system and the procedures for combining these modules to obtain descriptions for alternative configurations of that system. The 20 separate system component modules presented here can be combined to obtain a description of any of the 17 alternative system configurations (i.e., scenarios) that were evaluated in the MRS Systems Studies program (DOE 1989a). First-approximation descriptions of other yet-undefined system configurations could also be developed for system study purposes from this database. The descriptions include, in a modular format, both functional descriptions of the processes in the waste management system, plus physical descriptions of the equipment and facilities necessary for performance of those functions 15. Nurses’ Knowledge Levels of Chest Drain Management: A Descriptive Study Directory of Open Access Journals (Sweden) Merve Tarhan 2016-12-01 Full Text Available Objective: The physician is responsible for inserting one or more chest tubes into the pleural space or the mediastinal space and connecting them to an appropriate drainage system. When the general principles about care of patients with chest drains were implemented correctly and effectively by nurses, nurse will contribute to accelerate the healing process of patients. In this context, the aim of this study was to determine the nurses’ level of knowledge regarding the care of patients with chest drains. Methods: The study was conducted with 153 nurses who worked in a chest diseases and thoracic surgery hospital in July 2014. Questionnaire form of 35 questions prepared by investigators was used to collect data. For the analysis of results, frequency tests, independent sample t-test and oneway ANOVA test were used. Results: 69.3% of nurses stated that they had obtained information from colleguages. 35.3% considered their knowledge about chest drain management to be inadequate. 55.6% scored 13 points and above from knowledge questionnaire about chest drain management. There were statistically significant difference between knowledge level and educational background, clinic work type, working unit, years of professional experience and institutional experience, frequency of contact patients with chest drain and perception of knowledge level (p<0.05. Conclusion: Results of this study indicate that lack of evidence-based nursing care and insufficient training has resulted in uncertainty and knowledge deficit in important aspects of chest drain care. It can be concluded that nurses receive training needs and training protocols are about chest drain management. 16. Defense High Level Waste Disposal Container System Description Document International Nuclear Information System (INIS) Pettit, N. E. 2001-01-01 The Defense High Level Waste Disposal Container System supports the confinement and isolation of waste within the Engineered Barrier System of the Monitored Geologic Repository (MGR). Disposal containers are loaded and sealed in the surface waste handling facilities, transferred to the underground through the accesses using a rail mounted transporter, and emplaced in emplacement drifts. The defense high level waste (HLW) disposal container provides long-term confinement of the commercial HLW and defense HLW (including immobilized plutonium waste forms [IPWF]) placed within disposable canisters, and withstands the loading, transfer, emplacement, and retrieval loads and environments. US Department of Energy (DOE)-owned spent nuclear fuel (SNF) in disposable canisters may also be placed in a defense HLW disposal container along with commercial HLW waste forms, which is known as co-disposal. The Defense High Level Waste Disposal Container System provides containment of waste for a designated period of time, and limits radionuclide release. The disposal container/waste package maintains the waste in a designated configuration, withstands maximum handling and rockfall loads, limits the individual canister temperatures after emplacement, resists corrosion in the expected handling and repository environments, and provides containment of waste in the event of an accident. Defense HLW disposal containers for HLW disposal will hold up to five HLW canisters. Defense HLW disposal containers for co-disposal will hold up to five HLW canisters arranged in a ring and one DOE SNF canister inserted in the center and/or one or more DOE SNF canisters displacing a HLW canister in the ring. Defense HLW disposal containers also will hold two Multi-Canister Overpacks (MCOs) and two HLW canisters in one disposal container. The disposal container will include outer and inner cylinders, outer and inner cylinder lids, and may include a canister guide. An exterior label will provide a means by 17. Description of the male of Eburella pinima Martins and notes on the geographical distribution of Eburodacrys aenigma Galileo & Martins and Eburodacrys lanei Zajciw (Coleoptera, Cerambycidae Directory of Open Access Journals (Sweden) Juan Pablo Botero 2013-12-01 Full Text Available Description of the male of Eburella pinima Martins and notes on the geographical distribution of Eburodacrys aenigma Galileo & Martins and Eburodacrys lanei Zajciw (Coleoptera, Cerambycidae. The male of Eburella pinima Martins, 1997 is described and illustrated for the first time. Information on Eburodacrys aenigma Galileo & Martins, 2006, previously known only from the female holotype, which lacked locality label, is herein complemented. This species is recorded from Brazil and the male is depicted for the first time. The geographical distribution of Eburodacrys lanei Zajciw, 1958 is further restricted here as some previous records are confirmed to result from misidentifications of E. aenigma. 18. Taxonomic notes on Neoperla (Plecoptera: Perlidae) from Sichuan Province of China with the description of two new species. Science.gov (United States) Li, Weihai; Wang, Ying; Wang, Rongfeng 2017-01-17 Species of the perlid genus Neoperla from Sichuan Province, China are reviewed. Two new species are described, Neoperla caii Li & Wang, sp. nov. and N. emeishana Li & Wang, sp. nov. The new species are compared with related congeners. Available types of several known Neoperla species from Sichuan Province were studied and complementary descriptions or brief taxonomic comments are presented for N. bilineata Wu & Claassen, N. chui Wu & Claassen, N. microtumida Wu & Claassen, N. quadrata Wu & Claassen and N. truncata Wu. 19. Defense High Level Waste Disposal Container System Description Document International Nuclear Information System (INIS) 2000-01-01 The Defense High Level Waste Disposal Container System supports the confinement and isolation of waste within the Engineered Barrier System of the Monitored Geologic Repository (MGR). Disposal containers are loaded and sealed in the surface waste handling facilities, transferred to the underground through the accesses using a rail mounted transporter, and emplaced in emplacement drifts. The defense high level waste (HLW) disposal container provides long-term confinement of the commercial HLW and defense HLW (including immobilized plutonium waste forms (IPWF)) placed within disposable canisters, and withstands the loading, transfer, emplacement, and retrieval loads and environments. U.S. Department of Energy (DOE)-owned spent nuclear fuel (SNF) in disposable canisters may also be placed in a defense HLW disposal container along with commercial HLW waste forms, which is known as 'co-disposal'. The Defense High Level Waste Disposal Container System provides containment of waste for a designated period of time, and limits radionuclide release. The disposal container/waste package maintains the waste in a designated configuration, withstands maximum handling and rockfall loads, limits the individual canister temperatures after emplacement, resists corrosion in the expected handling and repository environments, and provides containment of waste in the event of an accident. Defense HLW disposal containers for HLW disposal will hold up to five HLW canisters. Defense HLW disposal containers for co-disposal will hold up to five HLW canisters arranged in a ring and one DOE SNF canister in the ring. Defense HLW disposal containers also will hold two Multi-Canister Overpacks (MCOs) and two HLW canisters in one disposal container. The disposal container will include outer and inner cylinders, outer and inner cylinder lids, and may include a canister guide. An exterior label will provide a means by which to identify the disposal container and its contents. Different materials 20. Description of the female of Navicordulia aemulatrix Pinto & Lamas and additional notes on the male (Odonata: Corduliidae). Science.gov (United States) Pinto, A P; Lamas, C J E 2011-01-01 The female of Navicordulia aemulatrix Pinto & Lamas is described and illustrated for the first time based on a single specimen from the same locality of the type series (state of Santa Catarina, [municipality of São Bento do Sul, 26°14'58"S, 49°22'59"W, railroad station] Rio Vermelho, 29.I.1952, in MZSP). In addition, further morphological notes for the male are provided based on three specimens collected at the type locality and at a new locality in the state of Santa Catarina (Timbó municipality). The pronotal process present in N. aemulatrix is re-evaluated and considered non-homologous to that found in Neocordulia setifera (Hagen in Selys) as previously suggested. 1. Project Notes Science.gov (United States) School Science Review, 1978 1978-01-01 Presents sixteen project notes developed by pupils of Chipping Norton School and Bristol Grammar School, in the United Kingdom. These Projects include eight biology A-level projects and eight Chemistry A-level projects. (HM) 2. Notes on some sertulariid hydroids (Cnidaria: Hydrozoa from the tropical western Pacific, with descriptions of nine new species Directory of Open Access Journals (Sweden) Horia R. Galea 2016-08-01 Full Text Available Forty-three species of sertulariid hydroids (Cnidaria: Hydrozoa: Sertulariidae, collected from the tropical western Pacific (Taiwan, Philippines, New Caledonia, French Polynesia, Vanuatu, Fiji, Tonga, Solomon Islands during various expeditions of the French Tropical Deep-Sea Benthos program, are discussed. Of these, nine are new to science: Gonaxia nova sp. nov., G. plumularioides sp. nov., Sertularella folliformis sp. nov., Se. plicata sp. nov., Se. pseudocatena sp. nov., Se. splendida sp. nov., Se. tronconica sp. nov., Se. tubulosa sp. nov., and Symplectoscyphus paucicatillus sp. nov. The subspecies Symplectoscyphus johnstoni (Gray, 1843 tropicus Vervoort, 1993 is raised to species but, in order to avoid the secondary homonymy with Sy. tropicus (Hartlaub, 1901, the replacement name, Sy. fasciculatus nom. nov., is introduced. The male and female gonothecae of Diphasia cristata Billard, 1920, the male gonothecae of Gonaxia elegans Vervoort, 1993, as well as the female gonothecae of Salacia macer Vervoort & Watson, 2003, are described for the first time. Additional notes on the morphology of several other species are provided. All taxa are illustrated, in most cases using figures drawn at the same scale, so as to highlight the differences between related species. 3. Technical Note: Updated durability/composition relationships for Hanford high-level waste glasses International Nuclear Information System (INIS) Piepel, G.F.; Hartley, S.A.; Redgate, P.E. 1996-03-01 This technical note presents empirical models developed in FYI 995 to predict durability as functions of glass composition. Models are presented for normalized releases of B, Li, Na, and Si from the 7-day Product Consistency Test (PCT) applied to quenched and canister centerline cooled (CCC) glasses as well as from the 28-day Materials Characterization Center-1 (MCC-1) test applied to quenched glasses. Models are presented for Composition Variation Study (CVS) data from low temperature melter (LTM) studies (Hrma, Piepel, et al. 1994) and high temperature melter (HTM) studies (Vienna et al. 1995). The data used for modeling in this technical note are listed in Appendix A 4. AutoBD: Automated Bi-Level Description for Scalable Fine-Grained Visual Categorization. Science.gov (United States) Yao, Hantao; Zhang, Shiliang; Yan, Chenggang; Zhang, Yongdong; Li, Jintao; Tian, Qi Compared with traditional image classification, fine-grained visual categorization is a more challenging task, because it targets to classify objects belonging to the same species, e.g. , classify hundreds of birds or cars. In the past several years, researchers have made many achievements on this topic. However, most of them are heavily dependent on the artificial annotations, e.g., bounding boxes, part annotations, and so on . The requirement of artificial annotations largely hinders the scalability and application. Motivated to release such dependence, this paper proposes a robust and discriminative visual description named Automated Bi-level Description (AutoBD). "Bi-level" denotes two complementary part-level and object-level visual descriptions, respectively. AutoBD is "automated," because it only requires the image-level labels of training images and does not need any annotations for testing images. Compared with the part annotations labeled by the human, the image-level labels can be easily acquired, which thus makes AutoBD suitable for large-scale visual categorization. Specifically, the part-level description is extracted by identifying the local region saliently representing the visual distinctiveness. The object-level description is extracted from object bounding boxes generated with a co-localization algorithm. Although only using the image-level labels, AutoBD outperforms the recent studies on two public benchmark, i.e. , classification accuracy achieves 81.6% on CUB-200-2011 and 88.9% on Car-196, respectively. On the large-scale Birdsnap data set, AutoBD achieves the accuracy of 68%, which is currently the best performance to the best of our knowledge.Compared with traditional image classification, fine-grained visual categorization is a more challenging task, because it targets to classify objects belonging to the same species, e.g. , classify hundreds of birds or cars. In the past several years, researchers have made many achievements on this topic 5. Thermodynamic approach to rheological modeling and simulations at the configuration space level of description NARCIS (Netherlands) Jongschaap, R.J.J.; Denneman, A.I.M.; Denneman, A.I.M.; Conrads, W. 1997-01-01 The so-called matrix model is a general thermodynamic framework for microrheological modeling. This model has already been proven to be applicable for a wide class of systems, in particular to models formulated at the configuration tensor level of description. For models formulated at the 6. Comparing the subjective task difficulty of human operators with task description levels International Nuclear Information System (INIS) Park, Jin Kyun; Jung, Won Dea; Yang, Joon Eon 2011-01-01 Without the loss of generality, it is reasonable to say that an operating procedure consists of many steps including detailed descriptions that provide necessary information in conducting the required tasks safely and effectively. In this regard, since it is widely perceived that procedures are effective for reducing the occurrence of human performance related problems, the use of procedures is very popular in large process control systems including nuclear power plants (NPPs), commercial airplanes and railway systems. However, the secure of an operational safety by using an operating procedure can be accomplished only if human operators are able to effectively obtain necessary information from it. In other words, it is hard to expect the reduction of human performance related problems, if task descriptions are so ambiguous or incomplete that human operators feel an undue difficulty in identifying 'what have to be done' and 'how to do it' from procedures. Unfortunately, it seems that a systematic method that can be used to distinguish the proper level of task descriptions is rare. For this reason, Park et al. developed a decision chart that could be helpful for characterizing the level of task descriptions. In this study, in order to ensure the appropriateness of the suggested decision chart, more detailed investigations were conducted with the support of human operators who are working as the operating personnel of NPPs 7. Systems Biology Graphical Notation: Process Description language Level 1 Version 1.3. Science.gov (United States) Moodie, Stuart; Le Novère, Nicolas; Demir, Emek; Mi, Huaiyu; Villéger, Alice 2015-09-04 The Systems Biological Graphical Notation (SBGN) is an international community effort for standardized graphical representations of biological pathways and networks. The goal of SBGN is to provide unambiguous pathway and network maps for readers with different scientific backgrounds as well as to support efficient and accurate exchange of biological knowledge between different research communities, industry, and other players in systems biology. Three SBGN languages, Process Description (PD), Entity Relationship (ER) and Activity Flow (AF), allow for the representation of different aspects of biological and biochemical systems at different levels of detail. The SBGN Process Description language represents biological entities and processes between these entities within a network. SBGN PD focuses on the mechanistic description and temporal dependencies of biological interactions and transformations. The nodes (elements) are split into entity nodes describing, e.g., metabolites, proteins, genes and complexes, and process nodes describing, e.g., reactions and associations. The edges (connections) provide descriptions of relationships (or influences) between the nodes, such as consumption, production, stimulation and inhibition. Among all three languages of SBGN, PD is the closest to metabolic and regulatory pathways in biological literature and textbooks, but its well-defined semantics offer a superior precision in expressing biological knowledge. 8. Cascade Boosting-Based Object Detection from High-Level Description to Hardware Implementation Directory of Open Access Journals (Sweden) K. Khattab 2009-01-01 Full Text Available Object detection forms the first step of a larger setup for a wide variety of computer vision applications. The focus of this paper is the implementation of a real-time embedded object detection system while relying on high-level description language such as SystemC. Boosting-based object detection algorithms are considered as the fastest accurate object detection algorithms today. However, the implementation of a real time solution for such algorithms is still a challenge. A new parallel implementation, which exploits the parallelism and the pipelining in these algorithms, is proposed. We show that using a SystemC description model paired with a mainstream automatic synthesis tool can lead to an efficient embedded implementation. We also display some of the tradeoffs and considerations, for this implementation to be effective. This implementation proves capable of achieving 42 fps for 320×240 images as well as bringing regularity in time consuming. 9. Using voice to create hospital progress notes: Description of a mobile application and supporting system integrated with a commercial electronic health record. Science.gov (United States) Payne, Thomas H; Alonso, W David; Markiel, J Andrew; Lybarger, Kevin; White, Andrew A 2018-01-01 We describe the development and design of a smartphone app-based system to create inpatient progress notes using voice, commercial automatic speech recognition software, with text processing to recognize spoken voice commands and format the note, and integration with a commercial EHR. This new system fits hospital rounding workflow and was used to support a randomized clinical trial testing whether use of voice to create notes improves timeliness of note availability, note quality, and physician satisfaction with the note creation process. The system was used to create 709 notes which were placed in the corresponding patient's EHR record. The median time from pressing the Send button to appearance of the formatted note in the Inbox was 8.8 min. It was generally very reliable, accepted by physician users, and secure. This approach provides an alternative to use of keyboard and templates to create progress notes and may appeal to physicians who prefer voice to typing. Copyright © 2017 Elsevier Inc. All rights reserved. 10. Technical NoteEarthquake dates and water level changes in wells in the Eskisehir region, Turkey Directory of Open Access Journals (Sweden) G. Yuce 2003-01-01 Full Text Available Although satisfactory results have yet to be obtained in earthquake prediction, one of the most common indicators of an anomalous precursor is a change in groundwater level in existing wells. Further wells should thus be drilled in unconfined aquifers since these are more susceptible to seismic waves. The Eskisehir region lies in the transition zone between the Aegean extensional domain and the compressible northern Anatolian block. Limnigraphs, installed in 19 exploration wells in the Eskisehir region, recorded pre-seismic, co-seismic and post-seismic level changes during the earthquakes of 17 August Izmit (Mw= 7.4 and 12 November Duzce (Mw= 7.2 1999 that occurred along the North Anatolian Fault Zone. The Izmit and Duzce earthquakes affected groundwater levels, especially in confined aquifers. The aquifer characteristics before and after the earthquakes were unchanged so the aquifer is elastic in its behaviour. Further detailed geo-mechanical investigation of the confined aquifer in the Eskisehir region may improve understanding of earthquake prediction. Keywords: earthquake prediction, Eskisehir, hydrological warning, monitoring groundwater levels 11. A Note on Aggregate Price-Level Elasticity and Supply-Side Shocks. Science.gov (United States) Findlay, David W. 1995-01-01 Investigates factors that influence the short-run and long-run effects of supply-side shocks on aggregate income and tax revenues. Concludes that the long-run relationship between tax revenues and the tax rate is completely independent of price-level elasticity. (CFR) 12. Correlation between radon level and confounders of cancer. A note on epidemiological inference at low doses International Nuclear Information System (INIS) Hajnal, M.A.; Toth, E.; Hamori, K.; Minda, M.; Koteles, Gy.J. 2007-01-01 Complete text of publication follows. Objective. The aim of this study was to examine and further clarify the extent of radon and progeny induced carcinogenesis, both separated from and combined with other confounders and health risk factors. This work was financed by National Development Agency, Hungary, with GVOP-3.1.1.-2004-05-0384/3.0. Methods. A case-control study was conducted in a Hungarian countryside region where the proportion of houses with yearly average radon level above 200 Bq.m -3 was estimated to be higher than 20% by our preceding regional surveys. Radon levels were measured with CR39 closed etched detectors for three seasons separately yielding yearly average by estimating the low summer level. The detectors were placed in the bedrooms, where people were expected to spend one third of a day. 520 patients with diagnosed cancers were included in these measurements, amongst which 77 developed lung or respiratory cancers. The control group consisted 6333 individuals, above 30 years of age. Lifestyle risk factors of cancers were collected by surveys including social status, pollution from indoor heating, smoking and alcohol history, nutrition, exercise and mental health index 5. Except smoking and alcohol habits, these cofactors were only available for the control group. Comparing disease occurrences the authors selected the multivariate generalised linear models. The case and control proportions along a given factor are binomially distributed, thus the logit link function was used. For radon both log and linear terms were probed for. Results. Many known health confounders of cancers correlated with radon levels, with an estimated total net increase of 50-150 Bq m -3 with increased risks. For lung cancers the model with the terms radon, age, gender and smoking was found to have the lowest Akaike Information Criterion (AIC). Heavy dependency on age, gender and smoking contribute largely to observed lung cancer incidence. However log linear relationship 13. A note on sea level variability at Clipperton Island from GEOSAT and in-situ observations Science.gov (United States) Maul, George A.; Hansen, Donald V.; Bravo, Nicolas J. During the 1986-1989 Exact Repeat Mission (ERM) of GEOSAT, in-situ observations of sea level at Clipperton Island (10°N/109°W) and satellite-tracked free-drifting drogued buoys in the eastern tropical Pacific Ocean are concurrently available. A map of the standard deviations of GEOSAT sea surface heights (2.9 years) shows a variance maximum along ˜12°N from Central America, past Clipperton to ˜160°W. Sea floor pressure gauge observations from a shallow (10m depth) site on Clipperton Island and an ERM crossover point in deep water nearby show a correlation of r = 0.76 with a residual of ±6.7 cm RMS. Approximately 17% of the difference (GEOSAT minus sea level) is characterized by a 4 cm amplitude 0° phase annual harmonic, which is probably caused by unaccounted-for tropospheric water vapor affecting the altimeter and/or ERM orbit error removal. Wintertime anticyclonic mesoscale eddies advecting past Clipperton Island each year have GEOSAT sea surface height and in-situ sea level signals of more than 30 cm, some of which are documented by the satellite-tracked drifters. Meridional profiles of the annual harmonic of zonal geostrophic current from GEOSAT and from the drifters both show synchronous maxima in the North Equatorial Countercurrent and the North Equatorial Current. Other Clipperton sea level maxima seen during late spring of each year may involve anticyclonic vortices formed along Central America the previous winter. 14. Site descriptions for preliminary radiological assessments of low-level waste repositories International Nuclear Information System (INIS) Burgess, W.E.; Jones, C.H.; Sumerling, T.J.; Ashton, J. 1988-07-01 The environmental contexts of four sites, previously under investigation by UK Nirex Ltd. as potential locations for low-level waste disposal facilities, are described. Information on geographical setting, climate, surface hydrology, land use, agriculture, fisheries, geology and hydrogeology is presented. The geological and hydrogeological data are interpreted with the support of deterministic modelling of groundwater conditions. The routes by which radionuclides may migrate from the site are identified and reduced to 1D statistical descriptions suitable for use in probabilistic risk assessments. Additional data required to improve the assessment of the performance of the site are identified. (author) 15. Technical note: Intercomparison of three AATSR Level 2 (L2 AOD products over China Directory of Open Access Journals (Sweden) Y. Che 2016-08-01 Full Text Available One of four main focus areas of the PEEX initiative is to establish and sustain long-term, continuous, and comprehensive ground-based, airborne, and seaborne observation infrastructure together with satellite data. The Advanced Along-Track Scanning Radiometer (AATSR aboard ENVISAT is used to observe the Earth in dual view. The AATSR data can be used to retrieve aerosol optical depth (AOD over both land and ocean, which is an important parameter in the characterization of aerosol properties. In recent years, aerosol retrieval algorithms have been developed both over land and ocean, taking advantage of the features of dual view, which can help eliminate the contribution of Earth's surface to top-of-atmosphere (TOA reflectance. The Aerosol_cci project, as a part of the Climate Change Initiative (CCI, provides users with three AOD retrieval algorithms for AATSR data, including the Swansea algorithm (SU, the ATSR-2ATSR dual-view aerosol retrieval algorithm (ADV, and the Oxford-RAL Retrieval of Aerosol and Cloud algorithm (ORAC. The validation team of the Aerosol-CCI project has validated AOD (both Level 2 and Level 3 products and AE (Ångström Exponent (Level 2 product only against the AERONET data in a round-robin evaluation using the validation tool of the AeroCOM (Aerosol Comparison between Observations and Models project. For the purpose of evaluating different performances of these three algorithms in calculating AODs over mainland China, we introduce ground-based data from CARSNET (China Aerosol Remote Sensing Network, which was designed for aerosol observations in China. Because China is vast in territory and has great differences in terms of land surfaces, the combination of the AERONET and CARSNET data can validate the L2 AOD products more comprehensively. The validation results show different performances of these products in 2007, 2008, and 2010. The SU algorithm performs very well over sites with different surface conditions in mainland 16. Preliminary notes on brain weight variation across labrid fish species with different levels of cooperative behaviour Institute of Scientific and Technical Information of China (English) Marta C.SOARES; Gon(c)alo I.ANDR(E); José R.PAULA 2015-01-01 Brain size and weight vary tremendously in the animal kingdom.It has been suggested that brain structural development must evolve balanced between the advantages of dealing with greater social challenges and the energetic costs of maintaining and developing larger brains.Here we ask if interspecific differences in cooperative behaviour (i.e.cleaning behaviour) are related to brain weight variations in four close-related species of Labrid fish:two are obligatory cleanerfish throughout their entire life (Labroides dimidiatus and L.bicolor),one facultative cleaner fish Labropsis australis and one last species that never engage in cleaning Labrichthys unilineatus.We first search for the link between the rate of species' cooperation and its relative brain weight,and finally,if the degree of social complexity and cooperation are reflected in the weight of its major brain substructures.Overall,no differences were found in relative brain weight (in relation to body weight) across species.Fine-scale differences were solely demonstrated for the facultative cleaner L.australis,at the brainstem level.Furthermore,data visual examination indicates that the average cerebellum and brainstem weights appear to be larger for L.dimidiatus.Because variation was solely found at specific brain areas (such as cerebellum and brainstem) and not for the whole brain weight values,it suggests that species social-ecological and cognitive demands may be directly contributing to a selective investment in relevant brain areas.This study provides first preliminary evidence that links potential differences in cognitive ability in cooperative behaviour to how these may mediate the evolution of brain structural development in non-mammal vertebrate groups [Current Zoology 61 (2):274-280,2015]. 17. A cautionary note on decadal sea level pressure predictions from GCMs Directory of Open Access Journals (Sweden) Stefan Liess 2018-03-01 Full Text Available A comparison of sea level pressure (SLP trends in a subset of seven Coupled Model Intercomparison Project (CMIP phase 5 general circulation models (GCM, namely decadal simulations with CCSM4, CanCM4, MPI-ESM-LR, FGOALS-g2, MIROC4h, MIROC5, and MRI-CGCM3, to their CMIP3 counterparts reveals an unrealistically strong forecast skill in CMIP3 models for trend predictions for 2001–2011 when using the 1979–2000 period to train the forecast. Boreal-winter SLP trends over five high-, mid-, and low-latitude zones were calculated over the 1979–2000 initialization period for each ensemble member and then ranked based on their performance relative to HadSLP2 observations. The same method is used to rank the ensemble members during the 2001–2011 period. In CMIP3, 17 out of 38 ensemble members retain their rank in the 2001–2011 hindcast period and 3 retain the neighboring rank. However, numbers are much lower in more recent CMIP5 decadal predictions over the similar 2001–2010 period when using the 1981–2000 period as initialization with the same number of ensembles. Different periods were used for CMIP3 and CMIP5 because although the 1979–2000 initialization is widely used for CMIP3, CMIP5 decadal predictions are only available for 30-year periods. The conclusion to consider the forecast skill in CMIP3 predictions during 2001–2011 as unrealistic is corroborated by comparisons to earlier periods from the 1960s to the 1980s in both CMIP3 and CMIP5 simulations. Thus, although the 2001–2011 CMIP3 predictions show statistically significant forecast skill, this skill should be treated as a spurious result that is unlikely to be reproduced by newer more accurate GCMs. 18. Assessment the Relationship Between Parents' Literacy Level with Children Growth in Mashhad: An Analytic Descriptive Study Directory of Open Access Journals (Sweden) Masumeh Saeidi 2013-12-01 Full Text Available Introduction: Present children are the investments of community in the future. Preparing children health which leads to the stability of community health, provided to accurate implementation of educational and health programs in the community and especially in mothers. So it is necessary to determine the relationship between parents' literacy  with growth rate in children. Materials and Methods: This cross-sectional descriptive-analytic study was conducted on 300 mothers referring to 10 selected Mashhad health-care centers for monitoring their 6-24-month year old infants. They completed questionnaire. Participants were selected by cluster and simple random sampling. Data were analyzed by descriptive- analytic statistics and using SPSS 16. Results: Presentfindings showed a significant relationship between literacy  level of parents with child growth status, breast feeding rate, junk food consumption, referring to health care center for growth monitoring, the age of initiating supplementary nutrition, the use of oil and butter in baby food and rate of attending in educational classes. So that higher literacy  level of parents was associated with using more formula, less junk food, oil and butter in baby's food and more referring times to health care center for monitoring child growth, desirable growth, and  also initiating supplementary food more at the assigned time (p 19. Assessment the Relationship between Parents' Literacy Level with Children Growth in Mashhad: An Analytic Descriptive Study Directory of Open Access Journals (Sweden) Bibi Leila Hoseini 2014-04-01 Full Text Available Introduction: Present children are the investments of community in the future. Preparing children health which leads to the stability of community health, provided to accurate implementation of educational and health programs in the community and especially in mothers. So it is necessary to determine the relationship between parents' literacy with growth rate in children.   Materials and Methods: This cross-sectional descriptive-analytic study was conducted on 300 mothers referring to 10 selected Mashhad health-care centers for monitoring their 6-24-month year old infants. They completed questionnaire. Participants were selected by cluster and simple random sampling. Data were analyzed by descriptive- analytic statistics and using SPSS 16.   Results: Presentfindings showed a significant relationship between literacy  level of parents with child growth status, breast feeding rate, junk food consumption, referring to health care center for growth monitoring, the age of initiating supplementary nutrition, the use of oil and butter in baby food and rate of attending in educational classes. So that higher literacy  level of parents was associated with using more formula, less junk food, oil and butter in baby's food and more referring times to health care center for monitoring child growth, desirable growth, and  also initiating supplementary food more at the assigned time (P 20. Simulation Experiment Description Markup Language (SED-ML) Level 1 Version 2. Science.gov (United States) Bergmann, Frank T; Cooper, Jonathan; Le Novère, Nicolas; Nickerson, David; Waltemath, Dagmar 2015-09-04 The number, size and complexity of computational models of biological systems are growing at an ever increasing pace. It is imperative to build on existing studies by reusing and adapting existing models and parts thereof. The description of the structure of models is not sufficient to enable the reproduction of simulation results. One also needs to describe the procedures the models are subjected to, as recommended by the Minimum Information About a Simulation Experiment (MIASE) guidelines. This document presents Level 1 Version 2 of the Simulation Experiment Description Markup Language (SED-ML), a computer-readable format for encoding simulation and analysis experiments to apply to computational models. SED-ML files are encoded in the Extensible Markup Language (XML) and can be used in conjunction with any XML-based model encoding format, such as CellML or SBML. A SED-ML file includes details of which models to use, how to modify them prior to executing a simulation, which simulation and analysis procedures to apply, which results to extract and how to present them. Level 1 Version 2 extends the format by allowing the encoding of repeated and chained procedures. 1. Amended diagnosis and redescription of Pristimantis marmoratus (Boulenger, 1900 (Amphibia: Craugastoridae, with a description of its advertisement call and notes on its breeding ecology and phylogenetic relationships Directory of Open Access Journals (Sweden) Philippe J.R. Kok 2018-01-01 Full Text Available The frog Pristimantis marmoratus was originally described as Hylodes marmoratus by George A. Boulenger in 1900 based on a single specimen reported to have been collected at the foot of Mount Roraima in Guyana in 1898. We herein discuss the exact location of the type locality of P. marmoratus and provide a redescription of the species based on new material from Kaieteur National Park and from the slopes of Maringma-tepui in Guyana. We also describe the previously unknown vocalization and breeding ecology of the species, and conducted an exploratory molecular analysis of the phylogenetic relationships within the genus Pristimantis represented by the members of the “unistrigatus species group” in the Guiana Shield. Pristimantis marmoratus is a small-sized species mainly distinguished from its known Guiana Shield congeners by the combination of F I < II, SVL ≤ 20.4 in males, presence of vocal slits in males, granular/pustulate dorsal skin with well-developed scapular ridges, basal webbing between fingers, fringes on fingers and toes, crossed iris, diffuse yellow or pale green wash on groin, and absence of flashy colour on axillary/pre-axillary region. The advertisement call consists of a single note repeated at a rate of ca 11 calls/min with a dominant frequency ranging from 2756 to 3101 Hz. Pristimantis marmoratus is primarily arboreal, exclusively active at dusk, and probably restricted to the pristine rainforests of the Pantepui uplands and highlands, east of the Gran Sabana between ca 600 and 1800 m above sea level. Preliminary molecular analyses recovered Pristimantis marmoratus as sister to an unnamed species from the Eastern Guiana Shield. On grounds of the newly established distributional extent we suggest maintaining the IUCN conservation status as Least Concern. 2. Simulation Experiment Description Markup Language (SED-ML Level 1 Version 3 (L1V3 Directory of Open Access Journals (Sweden) Bergmann Frank T. 2018-03-01 Full Text Available The creation of computational simulation experiments to inform modern biological research poses challenges to reproduce, annotate, archive, and share such experiments. Efforts such as SBML or CellML standardize the formal representation of computational models in various areas of biology. The Simulation Experiment Description Markup Language (SED-ML describes what procedures the models are subjected to, and the details of those procedures. These standards, together with further COMBINE standards, describe models sufficiently well for the reproduction of simulation studies among users and software tools. The Simulation Experiment Description Markup Language (SED-ML is an XML-based format that encodes, for a given simulation experiment, (i which models to use; (ii which modifications to apply to models before simulation; (iii which simulation procedures to run on each model; (iv how to post-process the data; and (v how these results should be plotted and reported. SED-ML Level 1 Version 1 (L1V1 implemented support for the encoding of basic time course simulations. SED-ML L1V2 added support for more complex types of simulations, specifically repeated tasks and chained simulation procedures. SED-ML L1V3 extends L1V2 by means to describe which datasets and subsets thereof to use within a simulation experiment. 3. Simulation Experiment Description Markup Language (SED-ML) Level 1 Version 3 (L1V3). Science.gov (United States) Bergmann, Frank T; Cooper, Jonathan; König, Matthias; Moraru, Ion; Nickerson, David; Le Novère, Nicolas; Olivier, Brett G; Sahle, Sven; Smith, Lucian; Waltemath, Dagmar 2018-03-19 The creation of computational simulation experiments to inform modern biological research poses challenges to reproduce, annotate, archive, and share such experiments. Efforts such as SBML or CellML standardize the formal representation of computational models in various areas of biology. The Simulation Experiment Description Markup Language (SED-ML) describes what procedures the models are subjected to, and the details of those procedures. These standards, together with further COMBINE standards, describe models sufficiently well for the reproduction of simulation studies among users and software tools. The Simulation Experiment Description Markup Language (SED-ML) is an XML-based format that encodes, for a given simulation experiment, (i) which models to use; (ii) which modifications to apply to models before simulation; (iii) which simulation procedures to run on each model; (iv) how to post-process the data; and (v) how these results should be plotted and reported. SED-ML Level 1 Version 1 (L1V1) implemented support for the encoding of basic time course simulations. SED-ML L1V2 added support for more complex types of simulations, specifically repeated tasks and chained simulation procedures. SED-ML L1V3 extends L1V2 by means to describe which datasets and subsets thereof to use within a simulation experiment. 4. Transport-level description of the 252Cf-source method using the Langevin technique International Nuclear Information System (INIS) Stolle, A.M.; Akcasu, A.Z. 1991-01-01 The fluctuations in the neutron number density and detector outputs in a nuclear reactor can be analyzed conveniently by using the Langevin equation approach. This approach can be implemented at any level of approximation to describe the time evolution of the neutron population, from the most complete transport-level description to the very basic point reactor analysis of neutron number density fluctuations. In this summary, the complete space- and velocity-dependent transport-level formulation of the Langevin equation approach is applied to the analysis of the 252 Cf-source-driven noise analysis (CSDNA) method, an experimental technique developed by J.T. Mihalczo at Oak Ridge National Laboratory, which makes use of noise analysis to determine the reactivity of subcritical media. From this analysis, a theoretical expression for the subcritical multiplication factor is obtained that can then be used to interpret the experimental data. Results at the transport level are in complete agreement with an independent derivation performed by Sutton and Doub, who used the probability density method to interpret the CSDNA experiment, but differed from other expressions that have appeared in the literature 5. Improving the description of collective effects within the combinatorial model of nuclear level densities International Nuclear Information System (INIS) Hilaire, S.; Girod, M.; Goriely, S. 2011-01-01 The combinatorial model of nuclear level densities has now reached a level of accuracy comparable to that of the best global analytical expressions without suffering from the limits imposed by the statistical hypothesis on which the latter expressions rely. In particular, it provides naturally, non Gaussian spin distribution as well as non equipartition of parities which are known to have a significant impact on cross section predictions at low energies. Our first global model developed in Ref. 1 suffered from deficiencies, in particular in the way the collective effects - both vibrational and rotational - were treated. We have recently improved this treatment using simultaneously the single particle levels and collective properties predicted by a newly derived Gogny interaction, therefore enabling a microscopic description of energy-dependent shell, pairing and deformation effects. In addition, for deformed nuclei, the transition to sphericity is coherently taken into account on the basis of a temperature-dependent Hartree-Fock calculation which provides at each temperature the structure properties needed to build the level densities. This new method is described and shown to give promising preliminary results with respect to available experimental data. (authors) 6. Allowable residual contamination levels for decommissioning. Part 1. A description of the method International Nuclear Information System (INIS) Napier, B.A.; Kennedy, W.E. Jr. 1984-10-01 This paper contains a description of the methods used in a study sponsored by UNC Nuclear Industries to determine Allowable Residual Contamination Levels (ARCL) for decommissioning facilities in the 100 Areas of the Hanford Site. The ARCL method is based on a scenario/exposure-pathway analysis and compliance with an annual dose limit for three specific modes of future use of the land and facilities. Thes modes of use are restricted, controlled, and unrestricted. The information on ARCL values for restricted and controlled use is intended to permit a full consideration of decommissioning alternatives. The analysis results in site-specific ARCL values that can be used for determining compliance with any annual dose limit selected. This flexibility permits proper consideration of field situations involving the radionuclide mixtures and physical conditions encountered. In addition, this method permits a full determination of as low as reasonably achievable (ALARA) conditions 7. Wetland Ecohydrology: stochastic description of water level fluctuations across the soil surface Science.gov (United States) Tamea, S.; Muneepeerakul, R.; Laio, F.; Ridolfi, L.; Rodriguez-Iturbe, I. 2009-12-01 Wetlands provide a suite of social and ecological critical functions such as being habitats of disease-carrying vectors, providing buffer zones against hurricanes, controlling sediment transport, filtering nutrients and contaminants, and a repository of great biological diversity. More recently, wetlands have also been recognized as crucial for carbon storage in the context of global climate change. Despite such importance, quantitative approaches to many aspects of wetlands are far from adequate. Therefore, improving our quantitative understanding of wetlands is necessary to our ability to maintain, manage, and restore these invaluable environments. In wetlands, hydrologic factors and ecosystem processes interplay and generate unique characteristics and a delicate balance between biotic and abiotic elements. The main hydrologic driver of wetland ecosystems is the position of the water level that, being above or below ground, determines the submergence or exposure of soil. When the water level is above the soil surface, soil saturation and lack of oxygen causes hypoxia, anaerobic functioning of microorganisms and anoxic stress in plants, that might lead to the death of non-adapted organisms. When the water level lies below the soil surface, the ecosystem becomes groundwater-dependent, and pedological and physiological aspects play their role in the soil water balance. We propose here a quantitative description of wetland ecohydrology, through a stochastic process-based water balance, driven by a marked compound Poisson noise representing rainfall events. The model includes processes such as rainfall infiltration, evapotranspiration, capillary rise, and the contribution of external water bodies, which are quantified in a simple yet realistic way. The semi-analytical steady-state probability distributions of water level spanning across the soil surface are validated with data from the Everglades (Florida, USA). The model and its results allow for a quantitative 8. Anxiety Level and Descriptive Features of Women Requesting Mammography at Early Diagnosis Center Directory of Open Access Journals (Sweden) 2009-08-01 9. Blunt traumatic injury during pregnancy: a descriptive analysis from a level 1 trauma center. Science.gov (United States) Al-Thani, Hassan; El-Menyar, Ayman; Sathian, Brijesh; Mekkodathil, Ahammed; Thomas, Sam; Mollazehi, Monira; Al-Sulaiti, Maryam; Abdelrahman, Husham 2018-03-27 The precise incidence of trauma in pregnancy is not well-known, but trauma is estimated to complicate nearly 1 in 12 pregnancies and it is the leading non-obstetrical cause of maternal death. A retrospective study of all pregnant women presented to national level 1 trauma center from July 2013 to June 2015 was conducted. Descriptive and inferential statistics applied for data analysis. Across the study period, a total of 95 pregnant women were presented to the trauma center. The average incidence rate of traumatic injuries was 250 per 1000 women of childbearing age presented to the Hamad Trauma Center. The mean age of patients was 30.4 ± SD 5.6 years, with age ranging from 20 to 42 years. The mean gestational age at the time of injury was 24.7 ± 8.7 weeks which ranged from 5 to 37 weeks. The majority (47.7%) was in the third trimester of the pregnancy. In addition, the large majority of injuries was due to MVCs (74.7%) followed by falls (15.8%). Trauma during pregnancy is not an uncommon event particularly in the traffic-related crashes. As it is a complex condition for trauma surgeons and obstetrician, an appropriate management protocol and multidisciplinary team are needed to improve the outcome and save lives of both the mother and fetus. 10. A Qualitative Descriptive Case Study of the Requirements of the IT Industry for Entry-Level IT Positions Science.gov (United States) Feuerherm, Todd Michael 2009-01-01 This qualitative descriptive case study explored the requirements of the IT industry for education, IT certification, and work experience for entry-level IT professionals. Research has shown a growing problem where IT graduates were not able to meet the requirements for entry-level IT jobs. IT enrollment has decreased considerably over the past… 11. Description of a new species of Apterotoxitiades Adlbauer, 2008 (Cerambycidae, Dorcasominae, Apatophyseini and the female of A. vivesi Adlbauer, 2008, with notes on the biology of the genus Directory of Open Access Journals (Sweden) 2015-02-01 Full Text Available Following the description of the Apatophyseini genus Apterotoxitiades Adlbauer, 2008 (Cerambycidae: Dorcasominae from South Africa, a new species has now been discovered in the eastern Drakensberg range of the country. The holotype female is here described as A. aspinosus Björnstad, sp. n. Also, a new small series collected at Hogsback, in the Amathole range, has allowed the description of the previously unknown female of the type species, A. vivesi Adlbauer, 2008. Both species are high altitude dwellers, occurring above 1300 m asl and their habitat consists mainly of mountain grassland interspersed with mistbelt forest pockets. All specimens were recorded in the austral winter to early spring, when these mountain ranges are occasionally covered in snow and night temperature plummet below 0 °C. They appear to be nocturnal and their complete lack of wings indicates a remarkable adaptation to cold conditions at high altitude. 12. Descriptions of two new species and one new subspecies from the Exocelina okbapensis-group, and notes on the E. aipo-group (Coleoptera, Dytiscidae, Copelatinae Directory of Open Access Journals (Sweden) Helena Shaverdo 2017-11-01 Full Text Available Two new species and one new subspecies of Exocelina Broun, 1886 from New Guinea are described: E. okbapensis Shaverdo & Balke, sp. n., E. okbapensis hajeki Shaverdo & Balke, ssp. n., and E. may Shaverdo & Balke, sp. n. These and two already described species are assigned to the E. okbapensis-group, which is morphologically (based on setation of the paramere and phylogenetically close to the E. aipo-group. On the latter, morphological and taxonomic notes are provided. An identification key to all known species of the groups is presented, and important diagnostic characters are illustrated. Data on the species distributions are mapped and show that the species occur only in the central mountain part of the island restricted by Wamena in the west and Sandaun Province in the east. 13. Revision of Bondariella Hustache & Bondar (Coleoptera: Curculionidae), with descriptions of the first species from the Amazon and notes on natural history. Science.gov (United States) Valente, Roberta De Melo; Cordeiro Júnior, Mariano Brandão 2015-09-15 Bondariella Hustache & Bondar, 1942 is revised. In addition to the redescription of the genus and its four previously known species, we report and describe the first species from the Amazonian region: Bondariella rudicula sp. nov. and Bondariella crenata sp. nov. We also provide a key to the species, illustrations and notes on natural history. All six species of Bondariella have been collected on palm flowers from Brazil, with the previously known species having only been recorded from species of Syagrus Mart. from the Atlantic Forest and Caatinga biomes. The two new species have only been recorded from species of Euterpe Mart. from the Amazon biome. Future work may prove if the associations with Syagrus and Euterpe are specific to the biomes. 14. Description of the last nymphal instar of Talipes appendiculatus (Hemiptera: Membracidae: Membracinae, and notes on the biology of the immatures and adults Directory of Open Access Journals (Sweden) Frederico Lencioni-Neto Full Text Available ABSTRACT The last instar nymph of Talipes appendiculatus (Fonseca is described and illustrated, and biological data are given on the immatures and adults. The 5th instar nymph has the body covered with tuberculate chalazae; head and thorax without scoli; abdomen with terga III-VIII with single dorsal scoli, all different in size, III and VIII small, V the largest about 2x longer than wide. Descriptions of the adults (male and female are also provided, with the addition of some new characters. The species (adults, immatures of various stages has been collected on branches of Cassia fistula L. (Leguminosae: Caesalpinioidea, and is frequently attended by bees and wasps. 15. Detailed description of an SSAC at the facility level for research laboratory facilities International Nuclear Information System (INIS) Jones, R.J. 1985-08-01 The purpose of this document is to provide a detailed description of a system for the accounting for and control of nuclear material in a research laboratory facility which can be used by a facility operator to establish his own system to comply with a national system for nuclear material accounting and control and to facilitate application of IAEA safeguards. The scope of this document is limited to descriptions of the following SSAC elements: (1) Nuclear Material Measurements; (2) Measurement Quality; (3) Records and Reports; (4) Physical Inventory Taking; (5) Material Balance Closing 16. Detailed description of an SSAC at the facility level for research reactors International Nuclear Information System (INIS) Jones, R.J. 1984-09-01 The purpose of this document is to provide a detailed description of a system for the accounting for and control of nuclear material in a research reactor facility which can be used by a facility operator to establish his own system to comply with a national system for nuclear material accounting and control and to facilitate application of IAEA safeguards. The scope of this document is limited to descriptions of the following SSAC elements: (1) Nuclear Material Measurements; (2) Measurement Quality; (3) Records and Reports; (4) Physical Inventory Taking; (5) Material Balance Closing 17. Detailed description of an SSAC at the facility level for mixed oxide fuel fabrication facilities International Nuclear Information System (INIS) Jones, R.J. 1985-09-01 The purpose of this document is to provide a detailed description of a system for the accounting for and control of nuclear material in a mixed oxide fuel fabrication facility which can be used by a facility operator to establish his own system to comply with a national system for nuclear material accounting and control and to facilitate application of IAEA safeguards. The scope of this document is limited to descriptions of the following SSAC elements: (1) Nuclear Material Measurements; (2) Measurement Quality; (3) Records and Reports; (4) Physical Inventory Taking; (5) Material Balance Closing 18. A description of two new species of the genus Erenna (Siphonophora: Physonectae: Erennidae), with notes on recently collected specimens of other Erenna species. Science.gov (United States) Pugh, P R; Haddock, S H D 2016-11-10 Two new Erenna species, E. insidiator sp. nov. and E. sirena sp. nov., are described from specimens collected in the vicinity of Monterey Bay, California, and also, for E. sirena at the southern end of the Gulf of California, Mexico. Further information on the three extant Erenna species is given, based on specimens collected in the same areas. These have enabled, for instance, the identification of three types of tentilla on the tentacles of E. cornuta Pugh, 2001, rather than the two noted on the single previously known specimen. The genus is remarkable for the presence of bioluminescent lures on the tentilla of all five species. In E. sirena sp. nov. the tentilla are also covered by a red-fluorescent layer, which was briefly described by Haddock et al. (2005), and further details are given herein. Another extraordinary feature of the colonies E. sirena sp. nov. is that the main part of the tentacle, with its tentilla, can be extended away from the siphosomal stem on a long peduncle. This phenomenon also appears to occur in E. laciniata Pugh, 2001, and has not been observed before for other physonect species. 19. Detailed description of a state system for accounting for and control of nuclear material at the state level International Nuclear Information System (INIS) Jones, R.J. 1985-02-01 The purpose of this document is to provide a detailed description of the technical elements of a system for the accounting for and control of nuclear material at the State Authority level which can be used by a state in the establishment of a national system for nuclear material accounting and control. It is expected that a state system designed along the lines described also will assist the IAEA in carrying out its safeguards responsibilities. The scope of this document is limited to descriptions of the technical elements of a state level system concerned with Laws and Regulations, the Information System, and the Establishment of Requirements for Nuclear Material Accounting and Control. The discussion shows the relationship of these technical elements at the state level to the principal elements of an SSAC at the facility levels 20. Generic packet descriptions : Verified Parsing and Pretty Printing of Low-Level Data NARCIS (Netherlands) van Geest, Marcell; Swierstra, Wouter 2017-01-01 Complex protocols describing the communication or storage of binary data are difficult to describe precisely. This paper presents a collection of data types for describing a binary data formats; the corresponding parser and pretty printer are generated automatically from a data description. By 1. Research Note: DEFF Research Database (Denmark) Behuria, Pritish; Buur, Lars; Gray, Hazel 2017-01-01 its core conceptual and methodological features. This Research Note starts by setting out our understanding of political settlements and provides an overview of existing political settlements literature on African countries. The note then explores how the key concept of ‘holding power’ has been... 2. Simulation Experiment Description Markup Language (SED-ML Level 1 Version 2 Directory of Open Access Journals (Sweden) Bergmann Frank T. 2015-06-01 Full Text Available The number, size and complexity of computational models of biological systems are growing at an ever increasing pace. It is imperative to build on existing studies by reusing and adapting existing models and parts thereof. The description of the structure of models is not sufficient to enable the reproduction of simulation results. One also needs to describe the procedures the models are subjected to, as recommended by the Minimum Information About a Simulation Experiment (MIASE guidelines. 3. Notes on black holes and three dimensional gravity International Nuclear Information System (INIS) 1999-01-01 In these notes we review some relevant results on 2+1 quantum gravity. These include the Chern-Simons formulation and its affine Kac-Moody symmetry, the asymptotic algebra of Brown and Henneaux, and the statistical mechanics description of 2+1 black holes. A brief introduction to the classical and semiclassical aspects of black holes is also included. The level of the notes is basic assuming only some knowledge on Statistical Mechanics, General Relativity and Yang-Mills theory 4. RESEARCH NOTE Navya RESEARCH NOTE. CDKN2A and MC1R ... Department of Pharmacy and Department of Nursing, School of Health Sciences, Frederick. University, Nicosia ..... Appears with highest frequency in African, Asian-Indian, and Papua. New Guinean ... 5. Editorial note Directory of Open Access Journals (Sweden) Tore Ahlbäck 2009-01-01 Full Text Available Editorial note of the Scripta Instituti Donneriani Aboensis, vol. 21, Postmodern Spirituality, based on papers read at the symposium on Postmodern Spirituality held at Åbo, Finland, on 11–13 June 2008. 6. NMR paves the way for atomic level descriptions of sparsely populated, transiently formed biomolecular conformers. Science.gov (United States) Sekhar, Ashok; Kay, Lewis E 2013-08-06 The importance of dynamics to biomolecular function is becoming increasingly clear. A description of the structure-function relationship must, therefore, include the role of motion, requiring a shift in paradigm from focus on a single static 3D picture to one where a given biomolecule is considered in terms of an ensemble of interconverting conformers, each with potentially diverse activities. In this Perspective, we describe how recent developments in solution NMR spectroscopy facilitate atomic resolution studies of sparsely populated, transiently formed biomolecular conformations that exchange with the native state. Examples of how this methodology is applied to protein folding and misfolding, ligand binding, and molecular recognition are provided as a means of illustrating both the power of the new techniques and the significant roles that conformationally excited protein states play in biology. 7. The Processing Speed of Scene Categorization at Multiple Levels of Description: The Superordinate Advantage Revisited. Science.gov (United States) Banno, Hayaki; Saiki, Jun 2015-03-01 Recent studies have sought to determine which levels of categories are processed first in visual scene categorization and have shown that the natural and man-made superordinate-level categories are understood faster than are basic-level categories. The current study examined the robustness of the superordinate-level advantage in a visual scene categorization task. A go/no-go categorization task was evaluated with response time distribution analysis using an ex-Gaussian template. A visual scene was categorized as either superordinate or basic level, and two basic-level categories forming a superordinate category were judged as either similar or dissimilar to each other. First, outdoor/ indoor groups and natural/man-made were used as superordinate categories to investigate whether the advantage could be generalized beyond the natural/man-made boundary. Second, a set of images forming a superordinate category was manipulated. We predicted that decreasing image set similarity within the superordinate-level category would work against the speed advantage. We found that basic-level categorization was faster than outdoor/indoor categorization when the outdoor category comprised dissimilar basic-level categories. Our results indicate that the superordinate-level advantage in visual scene categorization is labile across different categories and category structures. © 2015 SAGE Publications. 8. Writing a technical note. Science.gov (United States) Ng, K H; Peh, W C G 2010-02-01 A technical note is a short article giving a brief description of a specific development, technique or procedure, or it may describe a modification of an existing technique, procedure or device applicable to medicine. The technique, procedure or device described should have practical value and should contribute to clinical diagnosis or management. It could also present a software tool, or an experimental or computational method. Technical notes are variously referred to as technical innovations or technical developments. The main criteria for publication will be the novelty of concepts involved, the validity of the technique and its potential for clinical applications. 9. Notes on Piezoelectricity Energy Technology Data Exchange (ETDEWEB) Redondo, Antonio [Los Alamos National Lab. (LANL), Los Alamos, NM (United States) 2016-02-03 These notes provide a pedagogical discussion of the physics of piezoelectricity. The exposition starts with a brief analysis of the classical (continuum) theory of piezoelectric phenomena in solids. The main subject of the notes is, however, a quantum mechanical analysis. We first derive the Frohlich Hamiltonian as part of the description of the electron-phonon interaction. The results of this analysis are then employed to derive the equations of piezoelectricity. A couple of examples with the zinc blende and and wurtzite structures are presented at the end 10. Microscopic description of average level spacing in even-even nuclei International Nuclear Information System (INIS) Huong, Le Thi Quynh; Hung, Nguyen Quang; Phuc, Le Tan 2017-01-01 A microscopic theoretical approach to the average level spacing at the neutron binding energy in even-even nuclei is proposed. The approach is derived based on the Bardeen-Cooper-Schrieffer (BCS) theory at finite temperature and projection M of the total angular momentum J , which is often used to describe the superfluid properties of hot rotating nuclei. The exact relation of the J -dependent total level density to the M -dependent state densities, based on which the average level spacing is calculated, was employed. The numerical calculations carried out for several even-even nuclei have shown that in order to reproduce the experimental average level spacing, the M -dependent pairing gaps as well as the exact relation of the J -dependent total level density formula should be simultaneously used. (paper) 11. Accurate description of phase diagram of clathrate hydrates on molecular level Energy Technology Data Exchange (ETDEWEB) Belosludov, V.; Subbotin, O. [Niklaev Inst. of Inorganic Chemistry, Novosibirsk (Russian Federation). Siberian Branch of Russian Academy of Science; Belosludov, R.; Mizuseki, H.; Kawazoe, Y. [Tohoku Univ., Aoba-ku, Sendai (Japan). Inst. for Materials Research 2008-07-01 A number of experimental and theoretical studies of hydrogen hydrates have been conducted using different methods. In order to accurately estimate the thermodynamic properties of clathrate hydrates that multiply filling the cages, this paper presented a method based on the solid solution theory of van der Waals and Platteeuw with several modifications, including multiple occupancies, host relaxation, and the description of the quantum nature of hydrogen behavior in the cavities. The validity of the proposed approach was verified for argon, methane, and xenon hydrates. The results were in agreement with known experimental data. The model was then used to calculate the curves of monovariant three-phase equilibrium gas-hydrate-ice and the degree of filling of the large and small cavities for pure hydrogen and mixed hydrogen/propane hydrates in a wide range of pressure and at low temperatures. The paper presented the theory, including equations, monovariant equilibria, and computational details. It was concluded that the proposed model accounted for the influence of guest molecules on the host lattice and guest-guest interaction. The model could be used with other inclusion compounds with the same type of composition such as clathrate silicon, zeolites, and inclusion compounds of semiconductor elements. The calculated curves of monovariant equilibrium agree with the experiment. 33 refs., 1 tab., 9 figs. 12. Beyond the knowledge level : descriptions of rational behavior for sharing and reuse OpenAIRE Schmalhofer, Franz; Aitken, J. Stuart; Bourne, Lyle E. 1994-01-01 The currently dominant approach to the sharing and reuse of knowledge strives to develop ontologies, with clearly constrained interpretations. The idea of ontological commitments is based on the knowledge level perspective. Several shortcomings of the knowledge level have been identified (Clancey, 1991). Pursuing Clancey's argument, if KBS are to be situated in ever changing environments, their purposes and significance will change over time and they have to be redescribed accordingly. The be... 13. Code Description for Generation of Meteorological Height and Pressure Level and Layer Profiles Science.gov (United States) 2016-06-01 defined by user input height or pressure levels. It can process input profiles from sensing systems such as radiosonde, lidar, or wind profiling radar...routine may be required for different input types and formats. meteorological sounding interpolation , integrated mean layer values, US Army Research...or other radiosonde soundings. There are 2 main versions or “methods” that produce output in height- or pressure-based profiles of interpolated level 14. SAFETY NOTES CERN Document Server TIS Secretariat 2001-01-01 Please note that the revisions of safety notes no 3 (NS 3 Rev. 2) and no 24 (NS 24 REV.) entitled respectively 'FIRE PREVENTION FOR ENCLOSED SPACES IN LARGE HALLS' and 'REMOVING UNBURIED ELV AND LVA ELECTRIC CONDUITS' are available on the web at the following urls: http://edmsoraweb.cern.ch:8001/cedar/doc.download?document_id=322811&version=1&filename=version_francaise.pdf http://edmsoraweb.cern.ch:8001/cedar/doc.download?document_id=322861&version=2&filename=version_francaise.pdf Paper copies can also be obtained from the TIS Divisional Secretariat, email [email protected] 15. A spatial and seasonal description of return-levels for the Berlin-Brandenburg region (Germany) Science.gov (United States) Fischer, Madlen; Rust, Henning W.; Ulbrich, Uwe 2016-04-01 Extreme precipitation events have a strong impact on the environment, society and economy. Besides the direct effect, e.g. damage due to hail, extreme precipitation can cause flood events, mudslides and increased erosion, which in turn lead to serious damage. Typically, return levels derived from annual maxima of daily precipitation sums are used for the design of hydraulic structures or for risk assessment in insurance companies. Seasonally or monthly resolved return levels are rarely considered, although they provide additional information: the higher temporal resolution can be beneficial for risk management, e.g. for agriculture or tourism sector. In addition, annual return levels derived from monthly maxima offer lower uncertainties, since a larger data basis are used for estimation. Here, the generalized extreme value distribution (GEV) is used to calculate monthly resolved return levels for 323 stations in the region Berlin-Brandenburg (Germany). Instead of estimating the parameters of the GEV for each month separately, the seasonal variation is captured by harmonic functions. This natural approach is particularly suitable for an efficient characterization of the seasonal variation of extreme precipitation. In a first step, a statistical model is developed for each station separately to estimate the monthly return levels. Besides the seasonal smoothness, also smoothness in space is exploited here. We use functions of longitude, latitude and altitude to describe the spatial variation of GEV parameters in the second step. Thus, uncertainty is reduced at gauges with short time series and estimates for ungauged sites can be obtained in a meaningful way. 16. Process description and plant design for preparing ceramic high-level waste forms International Nuclear Information System (INIS) Grantham, L.F.; McKisson, R.L.; Guon, J.; Flintoff, J.F.; McKenzie, D.E. 1983-01-01 The ceramics process flow diagram has been simplified and upgraded to utilize only two major processing steps - fluid-bed calcination and hot isostatic press consolidating. Full-scale fluid-bed calcination has been used at INEL to calcine high-level waste for 18 y; and a second-generation calciner, a fully remotely operated and maintained calciner that meets ALARA guidelines, started calcining high-level waste in 1982. Full-scale hot isostatic consolidation has been used by DOE and commercial enterprises to consolidate radioactive components and to encapsulate spent fuel elements for several years. With further development aimed at process integration and parametric optimization, the operating knowledge of full-scale demonstration of the key process steps should be rapidly adaptable to scale-up of the ceramic process to full plant size. Process flowsheets used to prepare ceramic and glass waste forms from defense and commercial high-level liquid waste are described. Preliminary layouts of process flow diagrams in a high-level processing canyon were prepared and used to estimate the preliminary cost of the plant to fabricate both waste forms. The estimated costs for using both options were compared for total waste management costs of SRP high-level liquid waste. Using our design, for both the ceramic and glass plant, capital and operating costs are essentially the same for both defense and commercial wastes, but total waste management costs are calculated to be significantly less for defense wastes using the ceramic option. It is concluded from this and other studies that the ceramic form may offer important advantages over glass in leach resistance, waste loading, density, and process flexibility. Preliminary economic calculations indicate that ceramics must be considered a leading candidate for the form to immobilize high-level wastes 17. Effective description of general extensions of the Standard Model: the complete tree-level dictionary Science.gov (United States) de Blas, J.; Criado, J. C.; Pérez-Victoria, M.; Santiago, J. 2018-03-01 We compute all the tree-level contributions to the Wilson coefficients of the dimension-six Standard-Model effective theory in ultraviolet completions with general scalar, spinor and vector field content and arbitrary interactions. No assumption about the renormalizability of the high-energy theory is made. This provides a complete ultraviolet/infrared dictionary at the classical level, which can be used to study the low-energy implications of any model of interest, and also to look for explicit completions consistent with low-energy data. 18. Assessment the Relationship between Parents' Literacy Level with Children Growth in Mashhad: An Analytic Descriptive Study Directory of Open Access Journals (Sweden) M Saeidi 2014-04-01 Results: Presentfindings showed a significant relationship between literacy  level of parents with child growth status, breast feeding rate, junk food consumption, referring to health care center for growth monitoring, the age of initiating supplementary nutrition, the use of oil and butter in baby food and rate of attending in educational classes. So that higher literacy  level of parents was associated with using more formula, less junk food, oil and butter in baby's food and more referring times to health care center for monitoring child growth, desirable growth, and  also initiating supplementary food more at the assigned time (P 19. Low-Level Waste Forum notes and summary reports for 1994. Volume 9, Number 2, April 1994 International Nuclear Information System (INIS) 1994-04-01 This is a publication of the Low-Level Radioactive Waste Forum Participants. The topics of the publication include DOE policy, state concerns and activities, court hearings and decisions, federal agency activities, US NRC waste management function reorganization, low-level radioactive waste storage and compaction, and US NRC rulemaking and hearings 20. Low-Level Waste Forum notes and summary reports for 1994. Volume 9, Number 2, April 1994 Energy Technology Data Exchange (ETDEWEB) NONE 1994-04-01 This is a publication of the Low-Level Radioactive Waste Forum Participants. The topics of the publication include DOE policy, state concerns and activities, court hearings and decisions, federal agency activities, US NRC waste management function reorganization, low-level radioactive waste storage and compaction, and US NRC rulemaking and hearings. 1. SuperPixel based mid-level image description for image recognition NARCIS (Netherlands) Tasli, H.E.; Sicre, R.; Gevers, T. 2015-01-01 This study proposes a mid-level feature descriptor and aims to validate improvement on image classification and retrieval tasks. In this paper, we propose a method to explore the conventional feature extraction techniques in the image classification pipeline from a different perspective where 2. High-level intuitive features (HLIFs) for intuitive skin lesion description. Science.gov (United States) Amelard, Robert; Glaister, Jeffrey; Wong, Alexander; Clausi, David A 2015-03-01 A set of high-level intuitive features (HLIFs) is proposed to quantitatively describe melanoma in standard camera images. Melanoma is the deadliest form of skin cancer. With rising incidence rates and subjectivity in current clinical detection methods, there is a need for melanoma decision support systems. Feature extraction is a critical step in melanoma decision support systems. Existing feature sets for analyzing standard camera images are comprised of low-level features, which exist in high-dimensional feature spaces and limit the system's ability to convey intuitive diagnostic rationale. The proposed HLIFs were designed to model the ABCD criteria commonly used by dermatologists such that each HLIF represents a human-observable characteristic. As such, intuitive diagnostic rationale can be conveyed to the user. Experimental results show that concatenating the proposed HLIFs with a full low-level feature set increased classification accuracy, and that HLIFs were able to separate the data better than low-level features with statistical significance. An example of a graphical interface for providing intuitive rationale is given. 3. [The Italian deprivation index at census block level: definition, description and association with general mortality]. Science.gov (United States) Caranci, Nicola; Biggeri, Annibale; Grisotto, Laura; Pacelli, Barbara; Spadea, Teresa; Costa, Giuseppe 2010-01-01 the study is aimed at developing a nationwide deprivation index at municipality and census block level, based on the 2001 Census data, and meeting epidemiological needs. The study uses data drawn from the 2001 General Census of Population and Housing. From the 280 variables defined at census block level (352,605 census tracts with average number of inhabitants 169, standard deviation 225; and average area 0.6 km², sd 2.4 km²) five traits that operationally combine to represent the multidimensionality of the social and material deprivation concept have been selected; these are: low level of education, unemployment, non-home ownership, one parent family and overcrowding. The index is calculated by summing standardized indicators and it is also available as categorical by quintiles of population. The same procedure is applied to aggregate frequency data at municipality level. The correlation between mortality and deprivation has been evaluated using 2000-2004 general mortality. considering national data, a strong north-south gradient in deprivation was observed. The municipality deprivation index 2001 is highly correlated to the index likewise calculated on the basis of the previous 1991 Census (r=0.91). General mortality was positively correlated to the index (in particular in population up to 64 years and in larger size municipalities). the pattern described by the deprivation index was coherent with what is already known about geographic distribution of poverty and its impact on mortality. Such outcome bears out the index use for epidemiological purposes. 4. Descriptions of health states associated with increasing severity and frequency of hypoglycemia: a patient-level perspective Directory of Open Access Journals (Sweden) Harris SB 2013-09-01 Full Text Available Stewart B Harris,1 Kamlesh Khunti,2 Mona Landin-Olsson,3 Claus B Galbo-Jørgensen,4 Mette Bøgelund,4 Barrie Chubb,5 Jens Gundgaard,6 Marc Evans71Schulich School of Medicine and Dentistry, Western University, London, ON, Canada; 2Diabetes Research Unit, University of Leicester, Leicester, UK; 3Department of Medicine, Helsingborg Hospital, Lund University, Lund, Sweden; 4Incentive, Holte Stationsvej, Holte, Denmark; 5EU Health Economics and Outcomes Research, Novo Nordisk Ltd, Crawley, UK; 6Health Economics and HTA, Novo Nordisk A/S, Søborg, Denmark; 7Department of Diabetes, University Hospital Llandough, Cardiff, UKAims: We sought to develop descriptions of health states associated with daytime and nocturnal hypoglycemia in a structured fashion from the patient's perspective under different combinations of severity and frequency of hypoglycemic events.Methods: An expert meeting followed by two patient focus groups was used to develop comprehensive descriptions of acute consequences of severe and non-severe, daytime and nocturnal hypoglycemia. Patients with diabetes (type 1 = 85, type 2 = 162 from a survey panel then validated these descriptions and assessed how often they worried and took different actions to prevent hypoglycemia. Severity and frequency of hypoglycemia were compared with respect to how often people worried and took actions to prevent an event. The effect of hypoglycemia on 35 different life activities was quantitatively compared for patients who had and had not experienced a severe hypoglycemic event.Results: At least 95% of respondents agreed that the detailed patient-level descriptions of health states accurately reflected their experience of severe and non-severe, daytime and nocturnal hypoglycemia, thereby validating these descriptions. Respondents who had experienced a severe hypoglycemic event were generally more adversely affected in their worries and actions and life events than those who experienced only non 5. Technical Note African Journals Online (AJOL) In this report on four patients, we did not use any of these techniques. The existence and the site of the fistulas was clearly demonstrated using basic but important preoperative detailed assessment and two intraoperative findings. The preoperative referral note that indicated the site of technical difficulty during the previous ... CERN Multimedia 2006-01-01 Members of the personnel are invited to take note that only parcels corresponding to official orders or contracts will be handled at CERN. Individuals are not authorised to have private merchandise delivered to them at CERN and private deliveries will not be accepted by the Goods Reception services. Thank you for your understanding. (Version française la semaine prochaine.) 7. Editor's Note African Journals Online (AJOL) On another note: the editor and the editorial team acknowledge the financial support of the Carnegie Corporation of New York through the University of Ghana Building A New Generation of Academics in Africa (BANGA-Africa) Project. We also use this platform to express our gratitude for the support of various stakeholders, ... 8. Kinematic Description of Elite Vs. Low Level Players in Team-Handball Jump Throw Science.gov (United States) Wagner, Herbert; Buchecker, Michael; von Duvillard, Serge P.; Müller, Erich 2010-01-01 The jump throw is the most applied throwing technique in team- handball (Wagner et al., 2008); however, a comprehensive analysis of 3D-kinematics of the team-handball jump throw is lacking. Therefore, the purpose of our study was: 1) to measure differences in ball release speed in team- handball jump throw and anthropometric parameters between groups of different levels of performance and (2) to analyze upper body 3D-kinematics (flexion/extension and rotation) to determine significant differences between these groups. Three-dimensional kinematic data was analyzed via the Vicon MX 13 motion capturing system (Vicon Peak, Oxford, UK) from 26 male team-handball players of different performance levels (mean age: 21.2 ± 5.0 years). The participants were instructed to throw the ball (IHF Size 3) onto a target at 8 m distance, and to hit the center of a square of 1 × 1 m at about eye level (1.75 m), with maximum ball release speed. Significant differences between elite vs. low level players were found in the ball release speed (p handball players who were taller and of greater body weight have the ability to achieve a higher ball release speed in the jump throw, and that an increase in trunk flexion and rotation angular velocity improve the performance in team-handball jump throw that should result in an increase of ball release speed. Key points Team-handball players who were taller and of greater body weight have the ability to achieve a higher ball release speed. An increase in trunk flexion, trunk rotation and shoulder internal rotation angular velocity should result in an increase of ball release speed. Trunk movements are normally well observable for experienced coaches, easy correctable and therefore practical to improve the performance in team-handball jump throw of low level players during training without using complex measurement devices. PMID:24149381 9. Automation in the Teaching of Descriptive Geometry and CAD. High-Level CAD Templates Using Script Languages Science.gov (United States) Moreno, R.; Bazán, A. M. 2017-10-01 The main purpose of this work is to study improvements to the learning method of technical drawing and descriptive geometry through exercises with traditional techniques that are usually solved manually by applying automated processes assisted by high-level CAD templates (HLCts). Given that an exercise with traditional procedures can be solved, detailed step by step in technical drawing and descriptive geometry manuals, CAD applications allow us to do the same and generalize it later, incorporating references. Traditional teachings have become obsolete and current curricula have been relegated. However, they can be applied in certain automation processes. The use of geometric references (using variables in script languages) and their incorporation into HLCts allows the automation of drawing processes. Instead of repeatedly creating similar exercises or modifying data in the same exercises, users should be able to use HLCts to generate future modifications of these exercises. This paper introduces the automation process when generating exercises based on CAD script files, aided by parametric geometry calculation tools. The proposed method allows us to design new exercises without user intervention. The integration of CAD, mathematics, and descriptive geometry facilitates their joint learning. Automation in the generation of exercises not only saves time but also increases the quality of the statements and reduces the possibility of human error. 10. The choice of the optimum terms for semi-empirical description of s-wave neutron resonance level densities Energy Technology Data Exchange (ETDEWEB) Kaczmarczyk, Maria; Lason, Lech [Division of Nuclear Physics, University of Lodz, ul Pomorska 149/153, 90-236 Lodz (Poland) 2006-04-01 This paper presents a function describing the dependence of the neutron resonance level density {rho} on the neutron number N in the target nucleus. The function describes quite well, with an accuracy of one order, the experimental data for 284 nuclides. Moreover, it adequately describes the general tendency and shell model effects for magic nuclei and for nuclei close to magic ones. The achieved agreement between the values obtained from the proposed description and the experimental data {rho}{sub exp} can be improved if the {rho}{sub exp} values are normalized energetically and reduced to a narrow range of angular momentum J. 11. Multi-Level Simulated Fault Injection for Data Dependent Reliability Analysis of RTL Circuit Descriptions Directory of Open Access Journals (Sweden) NIMARA, S. 2016-02-01 Full Text Available This paper proposes data-dependent reliability evaluation methodology for digital systems described at Register Transfer Level (RTL. It uses a hybrid hierarchical approach, combining the accuracy provided by Gate Level (GL Simulated Fault Injection (SFI and the low simulation overhead required by RTL fault injection. The methodology comprises the following steps: the correct simulation of the RTL system, according to a set of input vectors, hierarchical decomposition of the system into basic RTL blocks, logic synthesis of basic RTL blocks, data-dependent SFI for the GL netlists, and RTL SFI. The proposed methodology has been validated in terms of accuracy on a medium sized circuit – the parallel comparator used in Check Node Unit (CNU of the Low-Density Parity-Check (LDPC decoders. The methodology has been applied for the reliability analysis of a 128-bit Advanced Encryption Standard (AES crypto-core, for which the GL simulation was prohibitive in terms of required computational resources. 12. Preparing medical students for obstetrics and gynecology milestone level one: a description of a pilot curriculum Directory of Open Access Journals (Sweden) Helen Morgan 2014-11-01 Full Text Available Background: The implementation of the Accreditation Council for Graduate Medical Education (ACGME Milestones in the field of obstetrics and gynecology has arrived with Milestones Level One defined as the level expected of an incoming first-year resident. Purpose: We designed, implemented, and evaluated a 4-week elective for fourth-year medical school students, which utilized a multimodal approach to teaching and assessing the Milestones Level One competencies. Methods: The 78-hour curriculum utilized traditional didactic lectures, flipped classroom active learning sessions, a simulated paging curriculum, simulation training, embalmed cadaver anatomical dissections, and fresh-frozen cadaver operative procedures. We performed an assessment of student knowledge and surgical skills before and after completion of the course. Students also received feedback on their assessment and management of eight simulated paging scenarios. Students completed course content satisfaction surveys at the completion of each of the 4 weeks. Results: Students demonstrated improvement in knowledge and surgical skills at the completion of the course. Paging confidence trended toward improvement at the completion of the course. Student satisfaction was high for all of the course content, and the active learning components of the curriculum (flipped classroom, simulation, and anatomy sessions had higher scores than the traditional didactics in all six categories of our student satisfaction survey. Conclusions: This pilot study demonstrates a practical approach for preparing fourth-year medical students for the expectations of Milestones Level One in obstetrics and gynecology. This curriculum can serve as a framework as medical schools and specific specialties work to meet the first steps of the ACGME's Next Accreditation System. 13. Description of a ceramic waste form and canister for Savannah River Plant high-level waste International Nuclear Information System (INIS) Butler, J.L.; Allender, J.S.; Gould, T.H. Jr. 1982-04-01 A canistered ceramic waste form for possible immobilization of Savannah River Plant (SRP) high-level radioactive wastes is described. Characteristics reported for the form include waste loading, chemical composition, heat content, isotope inventory, mechanical and thermal properties, and leach rates. A conceptual design of a potential production process for making this canistered form are also described. The ceramic form was selected in November 1981 as the primary alternative to the reference waste form, borosilicate glass, for making a final waste form decision for SRP waste by FY-1983. 11 tables 14. Low-Level Waste Forum notes and summary reports for 1994. Volume 9, Number 3, May-June 1994 Energy Technology Data Exchange (ETDEWEB) NONE 1994-06-01 This issue includes the following articles: Vermont ratifies Texas compact; Pennsylvania study on rates of decay for classes of low-level radioactive waste; South Carolina legislature adjourns without extending access to Barnwell for out-of-region generators; Southeast Compact Commission authorizes payments for facility development, also votes on petitions, access contracts; storage of low-level radioactive waste at Rancho Seco removed from consideration; plutonium estimates for Ward Valley, California; judgment issued in Ward Valley lawsuits; Central Midwest Commission questions courts jurisdiction over surcharge rebates litigation; Supreme Court decides commerce clause case involving solid waste; parties voluntarily dismiss Envirocare case; appellate court affirms dismissal of suit against Central Commission; LLW Forum mixed waste working group meets; US EPA Office of Radiation and Indoor Air rulemakings; EPA issues draft radiation site cleanup regulation; EPA extends mixed waste enforcement moratorium; and NRC denies petition to amend low-level radioactive waste classification regulations. 15. Low-Level Waste Forum notes and summary reports for 1994. Volume 9, Number 3, May-June 1994 International Nuclear Information System (INIS) 1994-06-01 This issue includes the following articles: Vermont ratifies Texas compact; Pennsylvania study on rates of decay for classes of low-level radioactive waste; South Carolina legislature adjourns without extending access to Barnwell for out-of-region generators; Southeast Compact Commission authorizes payments for facility development, also votes on petitions, access contracts; storage of low-level radioactive waste at Rancho Seco removed from consideration; plutonium estimates for Ward Valley, California; judgment issued in Ward Valley lawsuits; Central Midwest Commission questions court's jurisdiction over surcharge rebates litigation; Supreme Court decides commerce clause case involving solid waste; parties voluntarily dismiss Envirocare case; appellate court affirms dismissal of suit against Central Commission; LLW Forum mixed waste working group meets; US EPA Office of Radiation and Indoor Air rulemakings; EPA issues draft radiation site cleanup regulation; EPA extends mixed waste enforcement moratorium; and NRC denies petition to amend low-level radioactive waste classification regulations 16. Microscopic description of dissipative dynamics of a level-crossing transition Energy Technology Data Exchange (ETDEWEB) Scala, M.; Militello, B.; Messina, A. [Dipartimento di Fisica dell' Universita di Palermo, Via Archirafi 36, I-90123 Palermo (Italy); Vitanov, N. V. [Department of Physics, Sofia University, 5 James Bourchier Boulevard, BG-1164 Sofia (Bulgaria) 2011-08-15 We analyze the effect of a dissipative bosonic environment on the Landau-Zener-Stueckelberg-Majorana (LZSM) level crossing model by using a microscopic approach to derive the relevant master equation. For an environment at zero temperature and weak dissipation, our microscopic approach confirms the independence of the survival probability on the decay rate that has been predicted earlier by the simple phenomenological LZSM model. For strong decay the microscopic approach predicts a notable increase of the survival probability, which signals dynamical decoupling of the initial state. Unlike the phenomenological model, our approach makes it possible to study the dependence of the system dynamics on the temperature of the environment. In the limit of very high temperature we find that the dynamics is characterized by a very strong dynamical decoupling of the initial state--the temperature-induced quantum Zeno effect. 17. GELCOM: a geothermal levelized busbar cost model. Description and user's guide Energy Technology Data Exchange (ETDEWEB) Gupta, J.N.; Leigh, J.G. 1978-01-01 The model described is designed to project the costs of electricity generated from liquid dominated hydrothermal resources. It is presented in the form of an interactive computer program called GELCOM, written in Fortran IV, together with an analytic explanation of the methodology. GELCOM starts from resource data and estimates the capital and operating costs for the production field and electrical generating plant. The methodology implemented in GELCOM to determine the engineering cost, operating parameter estimates, and the levelized busbar cost is described. The output displayed by GELCOM for one prospect is explained. The input data files maintained in the computer system and the input data card formats are included. A sample session with GELCOM and the economic reports which may be generated from the GELCOM are illustrated. A completed listing of the computer program is given in an appendix. (MHR) 18. Developmental Math Students and College-Level Coursework. Data Notes. Volume 1, Number 8, November/December 2006 Science.gov (United States) Clery, Sue 2006-01-01 Using data from Achieving the Dream: Community Colleges Count, this issue examines the percentages and numbers of students who attempted and completed both developmental and college-level math. The analysis reveals that a large portion of students either do not attempt or do not complete developmental math coursework. Furthermore, the data show… 19. Application note : Energy Technology Data Exchange (ETDEWEB) Russo, Thomas V. 2013-08-01 The development of the XyceTM Parallel Electronic Simulator has focused entirely on the creation of a fast, scalable simulation tool, and has not included any schematic capture or data visualization tools. This application note will describe how to use the open source schematic capture tool gschem and its associated netlist creation tool gnetlist to create basic circuit designs for Xyce, and how to access advanced features of Xyce that are not directly supported by either gschem or gnetlist. 20. A descriptive study of noise in the neonatal intensive care unit: ambient levels and perceptions of contributing factors. Science.gov (United States) Darcy, Ashley E; Hancock, Lauren E; Ware, Emily J 2008-10-01 To examine the baseline acoustic environment in several mid-Atlantic region neonatal intensive care units (NICUs) and investigate the perceived factors contributing to noise levels in these NICUs. Quantitative data were collected from 3 urban, mid-Atlantic level IIIB and level IIIC NICUs. Qualitative data were collected via interview from 2 RNs employed in each of the study NICUs. This was an exploratory descriptive study utilizing a mixed-methods approach. A quantitative method was used for sound-level data collection, and a qualitative method was utilized during interviews with nurses to examine perceptions of factors contributing to noise. Ambient sound levels, measured in decibels, were taken at 5-minute intervals over a 2-hour period during both day and night shifts in a central location at each NICU. In addition, nurses were interviewed using a standardized interview questionnaire, and these interviews were then reviewed to determine themes regarding perceived factors contributing to sound levels. Hourly mean sound levels in each NICU ranged from 53.9 dB to 60.6 dB, with no statistically significant difference between noise levels recorded on day shift versus night shift, and no statistically significant difference among sites. Qualitative data showed that nurses' believed day shift to be louder than night shift and many perceived their own NICU as "pretty quiet." Key contributing factors to increased sound levels were stated as monitors or alarms, performing invasive procedures, presence of family, nurses or doctors giving report or rounds, and ringing phones. Noise levels were found to be above the American Academy of Pediatrics-recommended 45-dB level and often louder than the 50-dB level, which should not be exceeded more than 10% of the time. The recommended impulse maximum of 65 dB was also exceeded. Environmental Protection Agency recommendations for hospitals include sound levels no louder than 35 dB on night shift; this standard was also violated 1. A descriptive study of noise in the neonatal intensive care unit. Ambient levels and perceptions of contributing factors. Science.gov (United States) Darcy, Ashley E; Hancock, Lauren E; Ware, Emily J 2008-06-01 To examine the baseline acoustic environment in several mid-Atlantic region neonatal intensive care units (NICUs) and investigate the perceived factors contributing to noise levels in these NICUs. Quantitative data were collected from 3 urban, mid-Atlantic level IIIB and level IIIC NICUs. Qualitative data were collected via interview from 2 RNs employed in each of the study NICUs. This was an exploratory descriptive study utilizing a mixed-methods approach. A quantitative method was used for sound-level data collection, and a qualitative method was utilized during interviews with nurses to examine perceptions of factors contributing to noise. Ambient sound levels, measured in decibels, were taken at 5-minute intervals over a 2-hour period during both day and night shifts in a central location at each NICU. In addition, nurses were interviewed using a standardized interview questionnaire, and these interviews were then reviewed to determine themes regarding perceived factors contributing to sound levels. Hourly mean sound levels in each NICU ranged from 53.9 dB to 60.6 dB, with no statistically significant difference between noise levels recorded on day shift versus night shift, and no statistically significant difference among sites. Qualitative data showed that nurses' believed day shift to be louder than night shift and many perceived their own NICU as "pretty quiet." Key contributing factors to increased sound levels were stated as monitors or alarms, performing invasive procedures, presence of family, nurses or doctors giving report or rounds, and ringing phones. Noise levels were found to be above the American Academy of Pediatrics--recommended 45-dB level and often louder than the 50-dB level, which should not be exceeded more than 10% of the time. The recommended impulse maximum of 65 dB was also exceeded. Environmental Protection Agency recommendations for hospitals include sound levels no louder than 35 dB on night shift; this standard was also violated 2. Program of Hanford high-level waste retrieval task: a narrative description International Nuclear Information System (INIS) Wallskog, H.A. 1976-12-01 The objective of this task is to develop and demonstrate the equipment and methods for the retrieval of high-level radioactive wastes from underground storage tanks at Hanford. The approach will be to continue with engineering studies and the conceptual design in progress and follow on with the engineering design, construction, testing and demonstration of a Prototype Retrieval System. This system will consist of a large, mobile platform providing the support and control of an articulated arm used to remotely position waste recovery/removal tools. Other major components include the equipment needed to bring the material up to the platform for packaging and subsequent transport to a processing facility, and the television viewing and lighting subsystem. This prototype system will be functionally complete and will contain items such as a control center, tool change and maintenance/repair capability, etc. The program includes a complete non-radioactive demonstration of the system in a mock waste tank as well as a radioactive demonstration involving one or more waste tanks 3. Pressure ulcer prevention in nursing homes: nurse descriptions of individual and organization level factors. Science.gov (United States) Dellefield, Mary Ellen; Magnabosco, Jennifer L 2014-01-01 Sustaining pressure ulcer prevention (PUP) in nursing homes has been difficult to achieve. Implementation science researchers suggest that identification of individual staff and organizational factors influencing current practices is essential to the development of an effective and customized plan to implement practice changes in a specific setting. A mixed methods approach was used to describe nurses' perceptions of individual and organization-level factors influencing performance of PUP in two Veterans Health Administration (VHA) nursing homes prior to implementation of a national VHA initiative on Hospital Acquired Pressure Ulcers (HAPUs). Individual interviews of 16 nursing staff were conducted. Individual factors influencing practice were a personal sense of responsibility to Veterans and belief in the effectiveness and importance of preventive measures. Organizational factors were existence of cooperative practices between nursing assistants and licensed nurses in assessing risk; teamwork, communication, and a commitment to Veterans' well-being. Integration and reinforcement of such factors in the development and maintenance of customized plans of PUP initiatives is recommended. Published by Mosby, Inc. 4. Low-Level Waste Forum notes and summary reports for 1994. Volume 9, Number 4, July 1994 International Nuclear Information System (INIS) 1994-07-01 This issue includes the following articles: Federal Facility Compliance Act Task Force forms mixed waste workgroup; Illinois Department of Nuclear Safety considers construction of centralized storage facility; Midwest Commission agrees on capacity limit, advisory committee; EPA responds to California site developer's queries regarding application of air pollutant standards; county-level disqualification site screening of Pennsylvania complete; Texas Compact legislation introduced in US Senate; Generators ask court to rule in their favor on surcharge rebates lawsuit; Vermont authority and Battelle settle wetlands dispute; Eighth Circuit affirms decision in Nebraska community consent lawsuit; Nebraska court dismisses action filed by Boyd County local monitoring committee; NC authority, Chem-Nuclear, and Stowe exonerated; Senator Johnson introduces legislation to transfer Ward Valley site; Representative Dingell writes to Clinton regarding disposal of low-level radioactive waste; NAS committee on California site convenes; NRC to improve public petition process; NRC releases draft proposed rule on criteria for decontamination and closure of NRC-licensed facilities; and EPA names first environmental justice federal advisory council 5. Low-Level Waste Forum notes and summary reports for 1994. Volume 9, Number 4, July 1994 Energy Technology Data Exchange (ETDEWEB) NONE 1994-07-01 This issue includes the following articles: Federal Facility Compliance Act Task Force forms mixed waste workgroup; Illinois Department of Nuclear Safety considers construction of centralized storage facility; Midwest Commission agrees on capacity limit, advisory committee; EPA responds to California site developers queries regarding application of air pollutant standards; county-level disqualification site screening of Pennsylvania complete; Texas Compact legislation introduced in US Senate; Generators ask court to rule in their favor on surcharge rebates lawsuit; Vermont authority and Battelle settle wetlands dispute; Eighth Circuit affirms decision in Nebraska community consent lawsuit; Nebraska court dismisses action filed by Boyd County local monitoring committee; NC authority, Chem-Nuclear, and Stowe exonerated; Senator Johnson introduces legislation to transfer Ward Valley site; Representative Dingell writes to Clinton regarding disposal of low-level radioactive waste; NAS committee on California site convenes; NRC to improve public petition process; NRC releases draft proposed rule on criteria for decontamination and closure of NRC-licensed facilities; and EPA names first environmental justice federal advisory council. 6. Implementation of Malaria Dynamic Models in Municipality Level Early Warning Systems in Colombia. Part I: Description of Study Sites Science.gov (United States) Ruiz, Daniel; Cerón, Viviana; Molina, Adriana M.; Quiñónes, Martha L.; Jiménez, Mónica M.; Ahumada, Martha; Gutiérrez, Patricia; Osorio, Salua; Mantilla, Gilma; Connor, Stephen J.; Thomson, Madeleine C. 2014-01-01 As part of the Integrated National Adaptation Pilot project and the Integrated Surveillance and Control System, the Colombian National Institute of Health is working on the design and implementation of a Malaria Early Warning System framework, supported by seasonal climate forecasting capabilities, weather and environmental monitoring, and malaria statistical and dynamic models. In this report, we provide an overview of the local ecoepidemiologic settings where four malaria process-based mathematical models are currently being implemented at a municipal level. The description includes general characteristics, malaria situation (predominant type of infection, malaria-positive cases data, malaria incidence, and seasonality), entomologic conditions (primary and secondary vectors, mosquito densities, and feeding frequencies), climatic conditions (climatology and long-term trends), key drivers of epidemic outbreaks, and non-climatic factors (populations at risk, control campaigns, and socioeconomic conditions). Selected pilot sites exhibit different ecoepidemiologic settings that must be taken into account in the development of the integrated surveillance and control system. PMID:24891460 7. Hanford Waste Vitrification Plant Quality Assurance Program description for high-level waste form development and qualification International Nuclear Information System (INIS) 1993-08-01 The Hanford Waste Vitrification Plant Project has been established to convert the high-level radioactive waste associated with nuclear defense production at the Hanford Site into a waste form suitable for disposal in a deep geologic repository. The Hanford Waste Vitrification Plant will mix processed radioactive waste with borosilicate material, then heat the mixture to its melting point (vitrification) to forin a glass-like substance that traps the radionuclides in the glass matrix upon cooling. The Hanford Waste Vitrification Plant Quality Assurance Program has been established to support the mission of the Hanford Waste Vitrification Plant. This Quality Assurance Program Description has been written to document the Hanford Waste Vitrification Plant Quality Assurance Program 8. Editors' note Science.gov (United States) Denker, Carsten; Feller, Alex; Schmidt, Wolfgang; von der Lühe, Oskar 2012-11-01 This topical issue of Astronomische Nachrichten/Astronomical Notes is a collection of reference articles covering the GREGOR solar telescope, its science capabilities, its subsystems, and its dedicated suite of instruments for high-resolution observations of the Sun. Because ground-based telescopes have life spans of several decades, it is only natural that they continuously reinvent themselves. Literally, the GREGOR telescope builds on the foundations of the venerable Gregory-Coudé Telescope (GCT) at Observatorio del Teide, Tenerife, Spain. Acknowledging the fact that new discoveries in observational solar physics are driven by larger apertures to collect more photons and to scrutinize the Sun in finer detail, the GCT was decommissioned and the building was made available to the GREGOR project. 9. Editorial note DEFF Research Database (Denmark) Gani, Rafiqul; Hrymak, A.; Lee, J. 2009-01-01 industrial and educational applications. These articles will highlight theory, models, algorithms and applications with respect to value preservation and/or value creation or growth within the chemical product supply chain. To highlight and motivate research in the emerging challenges in PSE, we plan...... components of PSE—modeling, numerical analysis, optimization, systems and control theory, computer science, and, management science will be highlighted through the published articles (full-length papers, perspective papers, review papers, short notes and letters to the editor). They will cover...... and intelligent systems, integrated approaches to design, control and data analysis, systematic techniques for managing complexity, etc.), PSE emerging domains (product-process design, enterprise-wide optimization, energy and sustainability, biological engineering, pharmaceutical engineering, etc.) and novel... 10. Inventory of greenhouse gases at the municipality level. Description of calculation methods; Denmark; Drivhusgasopgoerelse paa kommuneniveau. Beskrivelse af beregningsmetoder Energy Technology Data Exchange (ETDEWEB) Nielsen, Ole-Kenneth; Winther, M.; Gyldenkaerne, S.; Lyck, E.; Thomsen, Marianne; Hoffmann, L.; Fauser, P. 2009-02-15 This report includes a description of methodologies, data and algorithms behind the inventories of greenhouse gases at the municipality level divided into sectors. The starting point for the sectors in this report is the sectors used for the official Danish emission inventories. A simplified generalization of the equations used in emission calculations is based on the assumption that emissions of a given activity is estimated using data descriptive for the size of the activity multiplied by an emission factor pr unit of activity. Emissions of CH{sub 4} and N{sub 2}O are converted to CO{sub 2} equivalents. In this project this generalization and these conversions are also the basis for all methodologies. The sectors included in this report are: the collective power and heating, individual heating, mobile sources, transportation and machinery, industrial processes, solvents, agriculture, land use and waste depositing and wastewater. The methods include calculations of the greenhouse gases that are most important for the sectors. The importance is estimated from the national emission inventory. This report covers methodologies for the greenhouse gases CO{sub 2}, CH{sub 4} and N{sub 2}O. Due to the mentioned importance criteria for some sectors not all greenhouse gases are included. As for the national inventories the calculation is built into several levels (Tiers) with increased requirements for municipalities regarding data. Tier 1 is mainly based on the Danish national greenhouse gas inventory data using appropriate distribution keys for a given activity into municipality level. Tier 2 is more detailed and includes emission factors used in the Danish national greenhouse gas inventories, for some sectors the emission factors are aggregated, while municipalities can enter their own activity data. At Tier 3, which is the most detailed level, there is - for some sectors - the opportunity to enter municipality specific emission factors and activity data. For other 11. Letters What causes an ice skater to accelerate? Note on the definitions of weight A-level physics is mathematical enough Correction to 'Confusion over the physics of circular motion' Science.gov (United States) 2011-11-01 What causes an ice skater to accelerate? Hugh Fricker Note on the definitions of weight Nenad Stojilovic A-level physics is mathematical enough Helen Hare Correction to 'Confusion over the physics of circular motion' 12. Built environment interventions aimed at improving physical activity levels in rural Ontario health units: a descriptive qualitative study. Science.gov (United States) Coghill, Cara-Lee; Valaitis, Ruta K; Eyles, John D 2015-05-03 Few studies to date have explored the relationship between the built environment and physical activity specifically in rural settings. The Ontario Public Health Standards policies mandate that health units in Ontario address the built environment; however, it is unclear how public health practitioners are integrating the built environment into public health interventions aimed at improving physical activity in chronic disease prevention programs. This descriptive qualitative study explored interventions that have or are being implemented which address the built environment specifically related to physical activity in rural Ontario health units, and the impact of these interventions. Data were collected through twelve in-depth semi-structured interviews with rural public health practitioners and managers representing 12 of 13 health units serving rural communities. Key themes were identified using qualitative content analysis. Themes that emerged regarding the types of interventions that health units are employing included: Engagement with policy work at a municipal level; building and working with community partners, committees and coalitions; gathering and providing evidence; developing and implementing programs; and social marketing and awareness raising. Evaluation of interventions to date has been limited. Public health interventions, and their evaluations, are complex. Health units who serve large rural populations in Ontario are engaging in numerous activities to address physical activity levels. There is a need to further evaluate the impact of these interventions on population health. 13. A quantitative description of state-level taxation of oil and gas production in the continental U.S International Nuclear Information System (INIS) Weber, Jeremy G.; Wang, Yongsheng; Chomas, Maxwell 2016-01-01 We provide a quantitative description of state-level taxation of oil and gas production in the continental U.S. for 2004–2013. Aggregate revenues from production taxes nearly doubled in real terms over the period, reaching $10.3 billion and accounting for 20% of tax receipts in the top ten revenue states. The average state had a tax rate of 3.6%; nationally, the average dollar of production was taxed at 4.2%. The oil-specific rate estimated for the study period is$2.4 per barrel or $5.5 per ton of carbon. Lastly, state-level tax rates are two-thirds higher in states excluding oil and gas wells from local property taxes, suggesting that the policies are substitutes for one another. - Highlights: •State tax revenue from oil and gas production nearly doubled from 2004 to 2013. •Nationally, the typical dollar of production is taxed at 4.2%. •The rate applied to the typical dollar of production did not increase over time. •On average oil is taxed at$2.4 per barrel or \$5.5 per ton of carbon. •State tax rates are two-thirds higher where oil and gas are not taxed as property. 14. Editorial Note Science.gov (United States) van der Meer, F.; Ommen Kloeke, E. 2015-07-01 With this editorial note we would like to update you on the performance of the International Journal of Applied Earth Observation and Geoinformation (JAG) and inform you about changes that have been made to the composition of the editorial team. Our Journal publishes original papers that apply earth observation data for the management of natural resources and the environment. Environmental issues include biodiversity, land degradation, industrial pollution and natural hazards such as earthquakes, floods and landslides. As such the scope is broad and ranges from conceptual and more fundamental work on earth observation and geospatial sciences to the more problem-solving type of work. When I took over the role of Editor-in-Chief in 2012, I together with the Publisher set myself the mission to position JAG in the top-3 of the remote sensing and GIS journals. To do so we strived at attracting high quality and high impact papers to the journal and to reduce the review turnover time to make JAG a more attractive medium for publications. What has been achieved? Have we reached our ambitions? We can say that: The submissions have increased over the years with over 23% for the last 12 months. Naturally not all may lead to more papers, but at least a portion of the additional submissions should lead to a growth in journal content and quality. 15. Note & Recensioni Directory of Open Access Journals (Sweden) 2014-12-01 Full Text Available VolumiVito Campanelli, Web Aesthetics. How Digital Media Affect Culture and Society; Id., Remix It Yourself. Analisi socio-estetica delle forme comunicative del Web [Emanuele Crescimanno] • Jean-François Bordron, Image et vérité. Essais sur les dimensions iconiques de la connaissance [Veronica Estay Stange] • Stéphane Dumas, Les peaux créatrices – Esthétique de la sécrétion [Marc-Vincent Howlett]NoteQuidam veritatis effectus. A proposito di A Plea for Balance in Philosophy. Essays in honour of Paolo Parrini [Fabrizio Desideri] • A Single Face to Capture the Whole World: Literary Shapes and Shadows. An Interview With Tolm Coibin [Fabrizia Abbate] •Convegno Le sensible a l’oeuvre: savoirs du corps entre esthetique et neurosciences, Parigi, 15 Maggio 2014 [Jessica Murano] 16. Science Notes. Science.gov (United States) Shaw, G. W.; And Others 1989-01-01 Provides a reading list for A- and S-level biology. Contains several experiments and demonstrations with topics on: the intestine, bullock corneal cells, valences, the science of tea, automated hydrolysis, electronics characteristics, bromine diffusion, enthalpy of vaporization determination, thermometers, pendulums, hovercraft, Bernoulli fluid… 17. Pass/fail patterns of candidates who failed COMLEX-USA level 2-PE because of misrepresentation of clinical findings on postencounter notes. Science.gov (United States) Langenau, Erik E; Sandella, Jeanne M 2011-07-01 In 2007, The National Board of Osteopathic Medical Examiners (NBOME) instituted a policy to address the accuracy and integrity of postencounter written documentation recorded during the Comprehensive Osteopathic Medical Licensing Examination Level 2-Performance Evaluation (COMLEX-USA Level 2-PE). This policy was instituted not only to protect the integrity of the examination, but also to highlight that overdocumentation of clinical findings not obtained during patient encounters may jeopardize patient safety. To investigate overall and domain pass/fail patterns of candidates who misrepresented clinical findings with regard to past and subsequent performance on COMLEX-USA Level 2-PE. Specifically, to investigate what percentage of candidates failed because of misrepresentation on first attempts and how they performed on subsequent administrations, as well as the previous performance patterns of candidates who failed because of misrepresentation on examination retakes. Historical records from NBOME's COMLEX-USA Level 2-PE database (testing cycles 2007-2008, 2008-2009, and 2009-2010) were used to analyze overall and domain pass/fail patterns of candidates who failed at least once because of misrepresentation of clinical findings. Of the 24 candidates who failed because of misrepresentation of postencounter (SOAP) notes, 20 candidates (83%) were first-time examinees. Four candidates (17%) were repeating the examination, 2 of whom were making a third attempt to pass. Among these 20 candidates who failed because of misrepresentation of clinical findings on their first attempt, 19 passed on their next attempt. At the time of study analysis, all but 2 candidates eventually passed the examination in subsequent attempts. Among candidates found to have misrepresented clinical findings on postencounter written documentation on COMLEX-USA Level 2-PE, no pattern existed between their past or subsequent performance with regard to overall or domain pass/fail results. The vast 18. Editor's Note Directory of Open Access Journals (Sweden) Daniel Shanahan 2016-07-01 ,500 visits a month from more than 50 countries while producing high-level original research, as well as insightful commentaries, from some of the field's most promising and established scholars. I'm also pleased to announce that Daniel Müllensiefen of Goldsmiths, University of London will be joining me as co-editor beginning in Volume 11, No.3. Additionally, we are lucky to be joined by Kelly Jakubowski of Durham University and Suzie Capps of Goldsmiths, who will be working as editorial assistants. 19. Editor's Note Directory of Open Access Journals (Sweden) Daniel Shanahan 2017-01-01 Full Text Available THIS issue marks the second of two special issues focusing on corpus approaches to musicological and music-theoretical research. What began as a simple call for papers turned into a fairly ambitious project, and I couldn't be more pleased with the result. The range and depth of the articles in these two issues demonstrates the level of scholarly work being conducted in this growing area of research. In this issue, Derek Strykowski's article on text-painting in Marenzio's madrigals provides a model for how one might search address the often difficult-to-test notion of text-painting, and Craig Sapp's commentary is a nice supplemental discussion of the topic, offering avenues of future research. Fernando Benadon's analysis of roughly 30,000 onset times in Afro-Cuban drumming focuses on the notion of "near onsets" and "onset clusters", and Peter Marten's commentary argues that, while this might be difficult to perceive, the methodology demonstrates an excellent model for future research. Steven Cannon's analysis of the form of 19th-century symphonies is enlightening, and provides a nice empirical backing to much of the extant literature on the repertoire, and Ben Duane's response provides a nice commentary on future directions and possible methodological improvements. Caitlyn Trevor and David Huron's article on the usage of open strings in nominally sad music deftly bridges emotion research and instrumental affordances, and Daniel Müllensiefen's commentary praises the methodology, laments the negative results, and issues a call for more negative results to be published. Perhaps the most unique aspect of Empirical Musicology Review is the "open peer review" format it pioneered a decade ago. This issue contains a number of commentaries —and responses to commentaries— that fit within that vision. Nat Condit-Schultz and Mitch Ohriner take two fascinating but different approaches to the notion of "flow" in hip-hop, and provide insightful 20. Detailed description of an SSAC at the facility level for on-load refueled power reactor facilities International Nuclear Information System (INIS) Jones, R.J. 1985-11-01 The purpose of this document is to provide a detailed description of a system for the accounting for and control of nuclear material in an on-load refueled power reactor facility which can be used by a facility operator to establish his own system to comply with a national system for nuclear material accounting and control and to facilitate application of IAEA safeguards. The scope of this document is limited to descriptions of the following SSAC elements: (1) Nuclear Material Measurements; (2) Measurement Quality; (3) Records and Reports; (4) Physical Inventory Taking; (5) Material Balance Closing 1. Editors note Directory of Open Access Journals (Sweden) Yngve Nordkvelle 2017-01-01 title “Networked e-Learning: The changing facilitator - learner relationship, a facilitators’ perspective; A Phenomenological Investigation”. The phenomenological case study deals with how the relationship between facilitator and student is changing. Networked e-Learning is the context and the research is undertaken at an Irish higher education institution.The author’s role as a highly experienced facilitator provides particular and specific insight into the guiding facilitator’s experiences during a time of institutional transition to Networked e-Learning.Gamification is a topic that has been declared as “up and coming” for a number of years. Marc Fabian Buck, of the Nord University, Norway, presents the article “Gamification of learning and teaching in schools – a critical stance”. He states that the aim of Gamification is to change learning for the better by making use of the motivating effects of (digital games and elements typical of games, like experience points, levelling, quests, rankings etc. His most contemporary example is of the “Summer of ‘16” and the apparent success of “Pokemon go”. He argues that gamified learning and teaching suspends the fundamental, subversive, and critical moments only schools can offer.The last article is provided by Ulf Olsson, of Stockholm University, Sweden: “Teachers’ Awareness of Guidelines for Quality Assurance when developing MOOCs”. His study focuses on higher education teachers’ awareness of quality issues in relation to Standards and Guidelines for Quality Assurance in the European Higher Education Area (ESG. Olsson conducted semi-structured interviews with 20 teachers at six Swedish HEIs while they developed open courses (MOOCs. The overall findings show that the teachers were not part of any transparent quality assurance system. Subsequently, he raises the question of the adequacy of a quality system for innovative activities. 2. Performance/Design Requirements and Detailed Technical Description for a Computer-Directed Training Subsystem for Integration into the Air Force Phase II Base Level System. Science.gov (United States) Butler, A. K.; And Others The performance/design requirements and a detailed technical description for a Computer-Directed Training Subsystem to be integrated into the Air Force Phase II Base Level System are described. The subsystem may be used for computer-assisted lesson construction and has presentation capability for on-the-job training for data automation, staff, and… 3. Description of a new species of skate of the genus Malacoraja Stehmann, 1970: the first species from the southwestern Atlantic Ocean, with notes on generic monophyly and composition (Chondrichthyes: Rajidae Directory of Open Access Journals (Sweden) Marcelo R. de Carvalho Full Text Available The first report of a western South Atlantic soft skate, genus Malacoraja Stehmann, 1970, is described as Malacoraja obscura, new species, from the southeastern Brazilian continental slope off the states of Espírito Santo and Rio de Janeiro, in depths ranging from 808-1105 m. The new species is known from five specimens and is distinguished from congeners by its unique dorsal coloration with small, faded white spots on disc and pelvic fins, by retaining in larger specimens an irregular row of thorns along dorsal midline of tail (extending from tail base to two-thirds of tail length in 680 mm total length female, and by presenting a ventral tail midline devoid of small denticles only at base (naked region not extending posterior to pelvic fin rear margin. Further diagnostic characters in combination include the lack of scapular thorns in larger specimens, elevated number of tooth rows (64/62 tooth rows in subadult male of 505 mm TL, and 76/74 in large female of 680 mm TL and vertebrae (27-28 Vtr, 68-75 Vprd, ventral disc and tail with a uniform dark brown coloration, paired postventral fenestrae on scapulocoracoid, enlarged posterior postventral fenestra, circular foramen magnum and paired internal carotid foramina on braincase floor. Adult males were unavailable for study, but an anatomical description of M. obscura, n. sp., is provided. Comparisons are made with all known material of M. kreffti, literature accounts of M. senta, and with abundant material of South African M. spinacidermis; M. obscura, n. sp., most closely resembles M. spinacidermis from the eastern South Atlantic in squamation, coloration and size. Malacoraja is monophyletic due to its unique squamation and rostral appendices, and apparently comprises two species-groups, one for M. obscura and M. spinacidermis, and the other for M. kreffti and M. senta, but clarification of species-level relationships must await more anatomical information, particularly of the latter two 4. Lecture notes on quantum statistics NARCIS (Netherlands) Gill, R.D. 2000-01-01 These notes are meant to form the material for an introductory course on quantum statistics at the graduate level aimed at mathematical statisticians and probabilists No background in physics quantum or otherwise is required They are still far from complete 5. Detailed description of an SSAC at the facility level for light water moderated (off-load refueled) power reactor facilities International Nuclear Information System (INIS) Jones, R.J. 1985-03-01 This report is intended to provide the technical details of an effective State Systems of Accounting for and Control of Nuclear Material (SSAC) which Member States may use, if they wish, to establish and maintain their SSACs. It is expected that systems designed along the lines described would be effective in meeting the objectives of both national and international systems for nuclear material accounting and control. This document accordingly provides a detailed description of a system for the accounting for and control of nuclear material in an off-load refueled light water moderated power reactor facility which can be used by a facility operator to establish his own system to comply with a national system for nuclear material accounting and control and to facilitate application of IAEA safeguards. The scope of this document is limited to descriptions of the following elements: (1) Nuclear Material Measurements; (2) Measurement Quality; (3) Records and Reports; (4) Physical Inventory Taking; (5) Material Balance Closing 6. Description of new species of Stenaelurillus Simon, 1886 from the Western Ghats of India with the redescription of Stenaelurillus lesserti Reimoser, 1934 and notes on mating plug in the genus (Arachnida, Araneae, Salticidae Directory of Open Access Journals (Sweden) Pothalil A. Sebastian 2015-03-01 Full Text Available A new species of the jumping spider genus Stenaelurillus Simon, 1886, S. albus sp. n., is described from the Western Ghats of India, one of the biodiversity hotspots of the world. Detailed morphological descriptions, diagnostic features and illustrations of copulatory organs of both sexes are given. Detailed redescription, diagnosis and illustration of S. lesserti Reimoser, 1934 are provided. The occurrence of a mating plug in the genus is reported. 7. Detailed description of an SSAC at the facility level for a low-enriched uranium conversion and fuel fabrication facility International Nuclear Information System (INIS) Jones, R.J. 1984-09-01 Some States have expressed a need for more detailed guidance with regard to the technical elements in the design and operation of SSACs for both the national and the international objectives. To meet this need the present document has been prepared, describing the technical elements of an SSAC in considerable detail. The purpose of this document is therefore, to provide a detailed description of a system for the accounting for and control of nuclear material in a model low enriched uranium conversion and fuel fabrication facility which can be used by a facility operator to establish his own system in a way which will provide the necessary information for compliance with a national system for nuclear material accounting and control and for the IAEA to carry out its safeguards responsibilities 8. Lecture Notes and Exercises for Course 21240 (Basic Analytical Chemistry) DEFF Research Database (Denmark) 1999-01-01 The publication contains notes dealing with difficult topics in analytical chemistry (cfr. Course Descriptions, DTU), relevant exercises as well as final examination problems from the last years.......The publication contains notes dealing with difficult topics in analytical chemistry (cfr. Course Descriptions, DTU), relevant exercises as well as final examination problems from the last years.... 9. Lecture Notes and Exercises for Course 21240 (Basic Analytical Chemistry) DEFF Research Database (Denmark) 1998-01-01 The publication contains notes dealing with difficult topics in analytical chemistry (cfr. Course Descriptions, DTU), relevant exercises as well as final examination problems from the last years.......The publication contains notes dealing with difficult topics in analytical chemistry (cfr. Course Descriptions, DTU), relevant exercises as well as final examination problems from the last years.... 10. Interim oceanographic description of the North-East Atlantic site for the disposal of low-level radioactive waste International Nuclear Information System (INIS) Gurbutt, P.A.; Dickson, R.R. 1983-01-01 Within the terms of the Surveillance Mechanism for Sea Dumping of Radioactive Wastes, NEA is requested to assess the suitability of dumping sites proposed by Participating countries and to keep under review those previously thought suitable. The aim of this volume is to provide a further interim description of the North-East Atlantic dumpsite itself. Quantities of known wastes dumpings are summarized. A review of the available data on sediment distribution in the area is presented. The flow field at the site is described so that an experiment to determine the influence on the deep flow of large scale topographic features at the dumpsite. The tide gauge results are briefly presented. The state of knowledge of the hydrographic and chemical conditions prior to 1977 is reviewed; recent results are added. The results of the radioactivity determination in the surface layer (3cm thick) of box cored sediments and the vertical radioactivity profiles are presented in tables. Some results on adsorption and geochemical partitioning of long-lived radionuclides on dumpsite sediments are briefly reviewed. Biological studies have been undertaken: concentration of radionuclides in biological materials, radiation effects on the dumpsite fauna. A dose-limit (critical group) calculation model is presented. Collective dose commitment and mass transfer are briefly discussed. The concentration of radionuclides in sediments and some organisms of the Bay of Biscay has been evaluated. Some isopycnal data for the eastern Atlantic, windstress and stratification are briefly mentioned 11. Antipsychotic Medication in Children and Adolescents : A Descriptive Review of the Effects on Prolactin Level and Associated Side Effects NARCIS (Netherlands) Roke, Yvette; van Harten, Peter N.; Boot, Annemieke M.; Buitelaar, Jan K. Objective: This review reports the incidence of hyperprolactinemia, its relationship with genotype, and prolactin-related side effects in children and adolescents treated with antipsychotics. Method: Data on prolactin levels were available for haloperidol, pimozide, risperidone, olanzapine, 12. Antipsychotic medication in children and adolescents: a descriptive review of the effects on prolactin level and associated side effects. NARCIS (Netherlands) Roke, Y.; Harten, P.N. van; Boot, A.M.; Buitelaar, J.K. 2009-01-01 OBJECTIVE: This review reports the incidence of hyperprolactinemia, its relationship with genotype, and prolactin-related side effects in children and adolescents treated with antipsychotics. METHOD: Data on prolactin levels were available for haloperidol, pimozide, risperidone, olanzapine, 13. Technical baseline description of high-level waste and low-activity waste feed mobilization and delivery International Nuclear Information System (INIS) Papp, I.G. 1997-01-01 This document is a compilation of information related to the high-level waste (HLW) and low-activity waste (LAW) feed staging, mobilization, and transfer/delivery issues. Information relevant to current Tank Waste Remediation System (TWRS) inventories and activities designed to feed the Phase I Privatization effort at the Hanford Site is included. Discussions on the higher level Phase II activities are offered for a perspective on the interfaces 14. A Species-Level Phylogeny of Extant Snakes with Description of a New Colubrid Subfamily and Genus. Science.gov (United States) Figueroa, Alex; McKelvy, Alexander D; Grismer, L Lee; Bell, Charles D; Lailvaux, Simon P 2016-01-01 With over 3,500 species encompassing a diverse range of morphologies and ecologies, snakes make up 36% of squamate diversity. Despite several attempts at estimating higher-level snake relationships and numerous assessments of generic- or species-level phylogenies, a large-scale species-level phylogeny solely focusing on snakes has not been completed. Here, we provide the largest-yet estimate of the snake tree of life using maximum likelihood on a supermatrix of 1745 taxa (1652 snake species + 7 outgroup taxa) and 9,523 base pairs from 10 loci (5 nuclear, 5 mitochondrial), including previously unsequenced genera (2) and species (61). Increased taxon sampling resulted in a phylogeny with a new higher-level topology and corroborate many lower-level relationships, strengthened by high nodal support values (> 85%) down to the species level (73.69% of nodes). Although the majority of families and subfamilies were strongly supported as monophyletic with > 88% support values, some families and numerous genera were paraphyletic, primarily due to limited taxon and loci sampling leading to a sparse supermatrix and minimal sequence overlap between some closely-related taxa. With all rogue taxa and incertae sedis species eliminated, higher-level relationships and support values remained relatively unchanged, except in five problematic clades. Our analyses resulted in new topologies at higher- and lower-levels; resolved several previous topological issues; established novel paraphyletic affiliations; designated a new subfamily, Ahaetuliinae, for the genera Ahaetulla, Chrysopelea, Dendrelaphis, and Dryophiops; and appointed Hemerophis (Coluber) zebrinus to a new genus, Mopanveldophis. Although we provide insight into some distinguished problematic nodes, at the deeper phylogenetic scale, resolution of these nodes may require sampling of more slowly-evolving nuclear genes. 15. Seasonal Variation in Objectively Assessed Physical Activity among Young Norwegian Talented Soccer Players: A Description of Daily Physical Activity Level Directory of Open Access Journals (Sweden) Stig A. Sæther, Nils P. Aspvik 2014-12-01 Full Text Available ‘Practise makes perfect’ is a well-known expression in most sports, including top-level soccer. However, a high training and match load increases the risk for injury, overtraining and burnout. With the use of accelerometers and a self-report questionnaire, the aim of this study was to describe talented players’ physical activity (PA level. Data were collected three times during the 2011 Norwegian Football season (March, June and October. The accelerometer output, counts·min–1 (counts per unit time registered, reports the daily PA-level for young talented soccer players. Results showed a stable PA-level across the season (March: 901.2 counts·min–1, June: 854.9 counts·min–1, October: 861.5 counts·min–1. Furthermore, comparison of five different training sessions across the season showed that the PA-level ranged from 2435.8 to 3745.4 counts·min–1. A one-way ANOVA showed no significant differences between the three measured weeks during the soccer season (p≤0.814. However, the training sessions in January had a significantly higher PA-level than those in June and October (p≤0.001. Based on these results, we discuss how potential implications of PA-level affect factors such as risk of injury, overtraining and burnout. We argue that player development must be seen as part of an overall picture in which club training and match load should be regarded as one of many variables influencing players’ PA-level. 16. A Species-Level Phylogeny of Extant Snakes with Description of a New Colubrid Subfamily and Genus Science.gov (United States) McKelvy, Alexander D.; Grismer, L. Lee; Bell, Charles D.; Lailvaux, Simon P. 2016-01-01 Background With over 3,500 species encompassing a diverse range of morphologies and ecologies, snakes make up 36% of squamate diversity. Despite several attempts at estimating higher-level snake relationships and numerous assessments of generic- or species-level phylogenies, a large-scale species-level phylogeny solely focusing on snakes has not been completed. Here, we provide the largest-yet estimate of the snake tree of life using maximum likelihood on a supermatrix of 1745 taxa (1652 snake species + 7 outgroup taxa) and 9,523 base pairs from 10 loci (5 nuclear, 5 mitochondrial), including previously unsequenced genera (2) and species (61). Results Increased taxon sampling resulted in a phylogeny with a new higher-level topology and corroborate many lower-level relationships, strengthened by high nodal support values (> 85%) down to the species level (73.69% of nodes). Although the majority of families and subfamilies were strongly supported as monophyletic with > 88% support values, some families and numerous genera were paraphyletic, primarily due to limited taxon and loci sampling leading to a sparse supermatrix and minimal sequence overlap between some closely-related taxa. With all rogue taxa and incertae sedis species eliminated, higher-level relationships and support values remained relatively unchanged, except in five problematic clades. Conclusion Our analyses resulted in new topologies at higher- and lower-levels; resolved several previous topological issues; established novel paraphyletic affiliations; designated a new subfamily, Ahaetuliinae, for the genera Ahaetulla, Chrysopelea, Dendrelaphis, and Dryophiops; and appointed Hemerophis (Coluber) zebrinus to a new genus, Mopanveldophis. Although we provide insight into some distinguished problematic nodes, at the deeper phylogenetic scale, resolution of these nodes may require sampling of more slowly-evolving nuclear genes. PMID:27603205 17. Seasonal Variation in Objectively Assessed Physical Activity among Young Norwegian Talented Soccer Players: A Description of Daily Physical Activity Level. Science.gov (United States) Sæther, Stig A; Aspvik, Nils P 2014-12-01 'Practise makes perfect' is a well-known expression in most sports, including top-level soccer. However, a high training and match load increases the risk for injury, overtraining and burnout. With the use of accelerometers and a self-report questionnaire, the aim of this study was to describe talented players' physical activity (PA) level. Data were collected three times during the 2011 Norwegian Football season (March, June and October). The accelerometer output, counts·min(-1) (counts per unit time registered), reports the daily PA-level for young talented soccer players. Results showed a stable PA-level across the season (March: 901.2 counts·min(-1), June: 854.9 counts·min(-1), October: 861.5 counts·min(-1)). Furthermore, comparison of five different training sessions across the season showed that the PA-level ranged from 2435.8 to 3745.4 counts·min(-1). A one-way ANOVA showed no significant differences between the three measured weeks during the soccer season (p≤0.814). However, the training sessions in January had a significantly higher PA-level than those in June and October (p≤0.001). Based on these results, we discuss how potential implications of PA-level affect factors such as risk of injury, overtraining and burnout. We argue that player development must be seen as part of an overall picture in which club training and match load should be regarded as one of many variables influencing players' PA-level. Key pointsIt is well established that to achieve a high performance level in sport, one must implement a high training and match load in childhood and youth.With the use of accelerometers and a self-reported questionnaire, the aim of this study was to describe talented players' total physical activity (PA) load.These results indicate that young talented soccer players must overcome large doses of PA on a weekly basis, exposing them to a high risk of injury, overtraining and burnout. 18. System-Level Shared Governance Structures and Processes in Healthcare Systems With Magnet®-Designated Hospitals: A Descriptive Study. Science.gov (United States) Underwood, Carlisa M; Hayne, Arlene N The purpose was to identify and describe structures and processes of best practices for system-level shared governance in healthcare systems. Currently, more than 64.6% of US community hospitals are part of a system. System chief nurse executives (SCNEs) are challenged to establish leadership structures and processes that effectively and efficiently disseminate best practices for patients and staff across complex organizations, geographically dispersed locations, and populations. Eleven US healthcare SCNEs from the American Nurses Credentialing Center's repository of Magnet®-designated facilities participated in a 35-multiquestion interview based on Kanter's Theory of Organizational Empowerment. Most SCNEs reported the presence of more than 50% of the empowerment structures and processes in system-level shared governance. Despite the difficulties and complexities of growing health systems, SCNEs have replicated empowerment characteristics of hospital shared governance structures and processes at the system level. 19. arXiv Effective description of general extensions of the Standard Model: the complete tree-level dictionary CERN Document Server de Blas, J.; Perez-Victoria, M.; Santiago, J. 2018-03-19 We compute all the tree-level contributions to the Wilson coefficients of the dimension-six Standard-Model effective theory in ultraviolet completions with general scalar, spinor and vector field content and arbitrary interactions. No assumption about the renormalizability of the high-energy theory is made. This provides a complete ultraviolet/infrared dictionary at the classical level, which can be used to study the low-energy implications of any model of interest, and also to look for explicit completions consistent with low-energy data. 20. Neutron-capture gamma-ray study of levels in 135Ba and description of nuclear levels in the interacting-boson-fermion model International Nuclear Information System (INIS) Chrien, R.E.; Koene, B.K.S.; Stelts, M.L.; Meyer, R.A.; Brant, S.; Paar, V.; Lopac, V. 1993-01-01 We have performed neutron-capture gamma-ray studies on natural and enriched targets of 134 Ba in order to investigate the nuclear levels of 135 Ba. The low-energy level spectra were compared with the calculations using the interacting-boson-fermion model (IBFM) and the cluster-vibration model. The level densities up to 5 MeV that are calculated within the IBFM are in accordance with the constant temperature Fermi gas model. From the spin distribution we have determined the corresponding spin cutoff parameter σ and compared it to the prediction from nuclear systematics 1. Note Taking and Recall Science.gov (United States) Fisher, Judith L.; Harris, Mary B. 1974-01-01 To study the effect of note taking and opportunity for review on subsequent recall, 88 college students were randomly assigned to five treatment groups utilizing different note taking and review combinations. No treatment effects were found, although quality of notes was positively correlated with free recall an multiple-choice measures.… 2. Levels of anxiety in parents in the 24 hr before and after their child's surgery: A descriptive study. Science.gov (United States) Pomicino, Laura; Maccacari, Elena; Buchini, Sara 2018-01-01 To (i) investigate pre- and postoperative anxiety levels in parents of surgical patients; (ii) identify factors that affect parental anxiety; and (iii) analyse assistance provided and overall parental satisfaction to assess whether and how this aspect can impact their anxiety level. Surgery as an event generates anxiety in children and their parents. Children who are anxious before surgery are likely to develop more postoperative psychological and physiological complications than those who are not. The role parents play in influencing emotional states of their children has been well demonstrated. However, specific national programmes aimed at helping parents develop new models for coping are relatively inexistent in Italy. Longitudinal study. One hundred and one parents of children undergoing surgery at a healthcare facility in Padua, Italy, completed the Italian version of the State-Trait Anxiety Inventory Form Y questionnaire. They also answered questions about their parents' socio-demographic situation, the amount and quality of preoperative information received, assistance provided and their overall satisfaction with this information. The preoperative level of anxiety in parents who were interviewed was higher than Italian normative data, especially in Pediatric Cardiac Surgery and Pediatric Urology departments. Mothers had a significantly higher level of anxiety than fathers. Communicating possible complications of surgical procedures increased anxiety, while providing information about pre- and postsurgery nutrition and pain management and providing local anaesthetic on children decreased parental anxiety. Parents expressed a sufficiently high level of satisfaction although they defined the hospital environment as uncomfortable. Aspects of care that can make hospitalisation less traumatic for parents are as follows: greater support, involving them in the treatment process, improving hospital department admission procedures and providing thorough preoperative 3. Can I Hear You? A Descriptive Longitudinal Study of Hearing Levels of Aeromedical Technicians and Flight Nurses Science.gov (United States) 1980-06-01 flightline) and in flight. The acuity with which they can perceive sound is important to the determination and the communi- cation of the patient’s... Preschool child 60dB 69dB School child 77dB 77dB Housewife 64dB 67dB Office Wroker 72dB 70dB Factory Worker 87dB 87dB ii eq( 2 4 ) - average sound level for...work or home environment, Jansen and Klensch have shown that the audiological re- sponses to pleasant music and to unpleasant noise were similar 4. Alternative methods for dispoal of low-level radioactive wastes. Task 1. Description of methods and assessment of criteria International Nuclear Information System (INIS) Bennett, R.D.; Miller, W.O.; Warriner, J.B.; Malone, P.G.; McAneny, C.C. 1984-04-01 The study reported herein contains the results of Task 1 of a four-task study entitled Criteria for Evaluating Engineered Facilities. The overall objective of this study is to ensure that the criteria needed to evaluate five alternative low-level radioactive waste (LLW) disposal methods are available to the Nuclear Regulatory Commission (NRC) and the Agreement States. The alternative methods considered are belowground vaults, aboveground vaults, earth mounded concrete bunkers, mined cavities, and augered holes. Each of these alternatives is either being used by other countries for low-level radioactive waste (LLW) disposal or is being considered by other countries or US agencies. In this report the performance requirements are listed, each alternative is described, the experience gained with its use is discussed, and the performance capabilities of each method are addressed. Next, the existing 10 CFR Part 61 Subpart D criteria with respect to paragraphs 61.50 through 61.53, pertaining to site suitability, design, operations and closure, and monitoring are assessed for applicability to evaluation of each alternative. Preliminary conclusions and recommendations are offered on each method's suitability as an LLW disposal alternative, the applicability of the criteria, and the need for supplemental or modified criteria 5. Hanford Waste Vitrification Plant Quality Assurance Program description for defense high-level waste form development and qualification International Nuclear Information System (INIS) Hand, R.L. 1992-01-01 This document describes the quality assurance (QA) program of the Hanford Waste Vitrification Plant (HWVP) Project. The purpose of the QA program is to control project activities in such a manner as to achieve the mission of the HWVP Project in a safe and reliable manner. A major aspect of the HWVP Project QA program is the control of activities that relate to high-level waste (HLW) form development and qualification. This document describes the program and planned actions the Westinghouse Hanford Company (Westinghouse Hanford) will implement to demonstrate and ensure that the HWVP Project meets the US Department of Energy (DOE) and ASME regulations. The actions for meeting the requirements of the Waste Acceptance Preliminary Specifications (WAPS) will be implemented under the HWVP product qualification program with the objective of ensuring that the HWVP and its processes comply with the WAPS established by the federal repository 6. Cd toxicity in Ash amended to plantations; the dilemma of different answers at different levels of description DEFF Research Database (Denmark) Christensen, Søren; Ekelund, Flemming; Kjøller, Rasmus 2018-01-01 . Other more abundant cations like Zn regulate Cd uptake. Mycorrhiza reduce Cd uptake in several plants. In barley Cd gradually decrease from root via stem to seeds. In conclusion for ash amounts up to 10 t/ha, we do not find metal toxicity of ash or evidence for bioaccumulation of metals in soil...... growth, equally because of the pH increase and the nutrients added. The pH increase stimulates decomposition activity. Soil carbon availability is not permanently increased with ash amendment, however, which may suggest that decomposition increase is coupled to primary production increase. Plantation....... Wood ash increase plantation soil Cd, and so did Cd in grass although at a very low level, Cd in lichens and Cd in the small biomass of earthworms. The toxicity of Cd is regulated by metal binding to soil being several orders of magnitude lower than in pure liquid. Ash is toxic towards microarthropods... 7. Detailed description of a new management system for solid, short-lived low and intermediate level radioactive waste at Ignalina NPP International Nuclear Information System (INIS) 2002-01-01 The objective is to modify and extend the existing system at the Ignalina Nuclear Power Plant (INPP) for handling of Very Low Level Waste (VLLW), short lived Low and Intermediate Level Waste (LLW-SL and ILW-SL). The ultimate aim is to reduce the risks and the influence on the personnel and the environment. According to the request from INPP, the modified system is based on the existence of an incineration plant. This system description describes treatment of non-combustible VLLW, LLW-SL and ILW-SL at a new waste handling facility (WHF) located in the future buildings 159/2 and 159/3 at the INPP. The new WHF is also handling Exempt Waste (EW), Reusable Material (RM) and Free Release Goods (FRG). The buildings 159/2 and 159/3 are future extensions of the existing building 159. (author) 8. OPG's long term management proposal for low and intermediate level radioactive waste: project description, operations International Nuclear Information System (INIS) Witzke, P. 2011-01-01 Although the Deep Geologic Repository (DGR) is approximately 8 years away from being placed into service, it is time to start planning for operations. Ontario Power Generation's (OPG) Nuclear Waste Management Division (NWMD) has a systematic approach to preparing for operation of any new facility that is readily applicable to the DGR. The DGR Operational Readiness Plan has been benchmarked at similar facilities in North America and Europe. The operating vision is a living model, and is constantly being reviewed and refined to align with the detailed design of the DGR as it proceeds through its phases of development. Combined with 40 years of operating surface storage facilities for the storage of Low and Intermediate Level Waste (LILW), the DGR operating vision will enable NWMD to provide meaningful input during COMS (Constructability, Operability, Maintainability, and Safety) review in the DGR project detailed-design phase in 2011/2012. A Work Breakdown Structure has been used to communicate the detail of the operating vision, and also to estimate the costs of Operational Readiness and Operations during the lifetime of the facility. (author) 9. Hanford Waste Vitrification Plant quality assurance program description for defense high-level waste form development and qualification International Nuclear Information System (INIS) Hand, R.L. 1990-12-01 The US Department of Energy-Office of Civilian Radioactive Waste Management has been designated the national high-level waste repository licensee and the recipient for the canistered waste forms. The Office of Waste Operations executes overall responsibility for producing the canistered waste form. The Hanford Waste Vitrification Plant Project, as part of the waste form producer organization, consists of a vertical relationship. Overall control is provided by the US Department of Energy-Environmental Restoration and Waste Management Headquarters; with the US Department of Energy-Office of Waste Operations; the US Department of Energy- Headquarters/Vitrification Project Branch; the US Department of Energy-Richland Operations Office/Vitrification Project Office; and the Westinghouse Hanford Company, operations and engineering contractor. This document has been prepared in response to direction from the US Department of Energy-Office of Civilian Radioactive Waste Management through the US Department of Energy-Richland Operations Office for a quality assurance program that meets the requirements of the US Department of Energy. This document provides guidance and direction for implementing a quality assurance program that applies to the Hanford Waste Vitrification Plant Project. The Hanford Waste Vitrification Plant Project management commits to implementing the quality assurance program activities; reviewing the program periodically, and revising it as necessary to keep it current and effective. 12 refs., 6 figs., 1 tab 10. Description of the immature stages of Sigara (Tropocorixa jensenhaarupi (Hemiptera: Heteroptera: Corixidae: Corixini, with ecological notes Descripción de los estadios larvales de Sigara (Tropocorixa jensenhaarupi (Heteroptera: Corixidae, con notas acerca de su ecología Directory of Open Access Journals (Sweden) María Cecilia Melo 2011-03-01 Full Text Available Sigara (Tropocorixa jensenhaarupi Jaczewski is the smallest species of the subgenus ranging from 4.2-4.7 mm, and it is characterized by the absence of a strigil, the small and narrow genital capsule with a short hypandrium in males, and the shape of the abdominal tergite VII in females. This species is endemic to the Patagonian subregion (Andean region in Argentina. A monthly sampling study was performed during a year in northern Mendoza, and additional material was collected in southern Mendoza, more precisely from Bañado Carilauquen in the Llancanelo Lake Reserve (Malargüe Department. Since little is known about the ecological requirements of S. (T. jensenhaarupi, herein we describe its habitat, the environmental conditions and its association with other macroinvertebrates. Also, we provide a morphological description of larval stages, and provide new records of this species.Sigara (Tropocorixa jensenhaarupi Jaczewski es la especie más pequeña del subgénero (4.2-4.7 mm, y se caracteriza por la ausencia de estrigilo, la cápsula genital masculina pequeña y angosta con el hipandro corto, y la forma característica del tergito abdominal VII en las hembras. Esta especie es endémica de la subregión Patagónica (región Andina en Argentina. Se realizaron muestreos mensuales durante un año en el norte de la provincia de Mendoza, material adicional fue recolectado en el sur de la provincia, más precisamente en el Bañado Carilauquen en la Reserva Laguna Llancanelo (Departamento de Malargüe. Poco se conoce acerca de los requerimientos ecológicos de S. (T. jensenhaarupi, por lo que en este trabajo describimos su hábitat, las condiciones ambientales y su asociación con otros macroinvertebrados. Además, describimos los estadios larvales y damos nuevos registros de esta especie. 11. Descriptive Research DEFF Research Database (Denmark) Wigram, Anthony Lewis 2003-01-01 Descriptive research is described by Lathom-Radocy and Radocy (1995) to include Survey research, ex post facto research, case studies and developmental studies. Descriptive research also includes a review of the literature in order to provide both quantitative and qualitative evidence of the effect...... starts will allow effect size calculations to be made in order to evaluate effect over time. Given the difficulties in undertaking controlled experimental studies in the creative arts therapies, descriptive research methods offer a way of quantifying effect through descriptive statistical analysis... 12. Level crossing analysis of chemically induced dynamic nuclear polarization: Towards a common description of liquid-state and solid-state cases Energy Technology Data Exchange (ETDEWEB) Sosnovsky, Denis V.; Ivanov, Konstantin L., E-mail: [email protected] [International Tomography Centre of SB RAS, Institutskaya 3a, 630090, Novosibirsk (Russian Federation); Novosibirsk State University, Pirogova 2, 630090, Novosibirsk (Russian Federation); Jeschke, Gunnar [Institut für Physikalische Chemie, ETH Zürich, Vladimir-Prelog-Weg 2, CH-8093 Zürich (Switzerland); Matysik, Jörg [Institut für Analytische Chemie, Universität Leipzig, Linnéstr. 3, D-04103 Leipzig (Germany); Vieth, Hans-Martin [International Tomography Centre of SB RAS, Institutskaya 3a, 630090, Novosibirsk (Russian Federation); Institut für Experimentalphysik, Freie Universität Berlin, Arnimallee 14, D-14195 Berlin (Germany) 2016-04-14 Chemically Induced Dynamic Nuclear Polarization (CIDNP) is an efficient method of creating non-equilibrium polarization of nuclear spins by using chemical reactions, which have radical pairs as intermediates. The CIDNP effect originates from (i) electron spin-selective recombination of radical pairs and (ii) the dependence of the inter-system crossing rate in radical pairs on the state of magnetic nuclei. The CIDNP effect can be investigated by using Nuclear Magnetic Resonance (NMR) methods. The gain from CIDNP is then two-fold: it allows one to obtain considerable amplification of NMR signals; in addition, it provides a very useful tool for investigating elusive radicals and radical pairs. While the mechanisms of the CIDNP effect in liquids are well established and understood, detailed analysis of solid-state CIDNP mechanisms still remains challenging; likewise a common theoretical frame for the description of CIDNP in both solids and liquids is missing. Difficulties in understanding the spin dynamics that lead to the CIDNP effect in the solid-state case are caused by the anisotropy of spin interactions, which increase the complexity of spin evolution. In this work, we propose to analyze CIDNP in terms of level crossing phenomena, namely, to attribute features in the CIDNP magnetic field dependence to Level Crossings (LCs) and Level Anti-Crossings (LACs) in a radical pair. This approach allows one to describe liquid-state CIDNP; the same holds for the solid-state case where anisotropic interactions play a significant role in CIDNP formation. In solids, features arise predominantly from LACs, since in most cases anisotropic couplings result in perturbations, which turn LCs into LACs. We have interpreted the CIDNP mechanisms in terms of the LC/LAC concept. This consideration allows one to find analytical expressions for a wide magnetic field range, where several different mechanisms are operative; furthermore, the LAC description gives a way to determine CIDNP sign 13. 1. On note taking. Science.gov (United States) Plaut, Alfred B J 2005-02-01 In this paper the author explores the theoretical and technical issues relating to taking notes of analytic sessions, using an introspective approach. The paper discusses the lack of a consistent approach to note taking amongst analysts and sets out to demonstrate that systematic note taking can be helpful to the analyst. The author describes his discovery that an initial phase where as much data was recorded as possible did not prove to be reliably helpful in clinical work and initially actively interfered with recall in subsequent sessions. The impact of the nature of the analytic session itself and the focus of the analyst's interest on recall is discussed. The author then describes how he modified his note taking technique to classify information from sessions into four categories which enabled the analyst to select which information to record in notes. The characteristics of memory and its constructive nature are discussed in relation to the problems that arise in making accurate notes of analytic sessions. 14. Descriptions and characterizations of water-level data and groundwater flow for the Brewster Boulevard and Castle Hayne Aquifer Systems and the Tarawa Terrace Aquifer Science.gov (United States) Faye, Robert E.; Jones, L. Elliott; Suárez-Soto, René J. 2013-01-01 This supplement of Chapter A (Supplement 3) summarizes results of analyses of groundwater-level data and describes corresponding elements of groundwater flow such as vertical hydraulic gradients useful for groundwater-flow model calibration. Field data as well as theoretical concepts indicate that potentiometric surfaces within the study area are shown to resemble to a large degree a subdued replica of surface topography. Consequently, precipitation that infiltrates to the water table flows laterally from highland to lowland areas and eventually discharges to streams such as Northeast and Wallace Creeks and New River. Vertically downward hydraulic gradients occur in highland areas resulting in the transfer of groundwater from shallow relatively unconfined aquifers to underlying confined or semi-confined aquifers. Conversely, in the vicinity of large streams such as Wallace and Frenchs Creeks, diffuse upward leakage occurs from underlying confined or semi-confined aquifers. Point water-level data indicating water-table altitudes, water-table altitudes estimated using a regression equation, and estimates of stream levels determined from a digital elevation model (DEM) and topographic maps were used to estimate a predevelopment water-table surface in the study area. Approximate flow lines along hydraulic gradients are shown on a predevelopment potentiometric surface map and extend from highland areas where potentiometric levels are greatest toward streams such as Wallace Creek and Northeast Creek. The distribution of potentiometric levels and corresponding groundwater-flow directions conform closely to related descriptions of the conceptual model. 15. Implications of observing and writing field notes through different lenses Directory of Open Access Journals (Sweden) Hellesø R 2015-04-01 Full Text Available Ragnhild Hellesø,1 Line Melby,1 Solveig Hauge21Department of Nursing Science, Institute of Health and Society, Faculty of Medicine, University of Oslo, Oslo, Norway; 2Faculty of Health and Social Studies, Telemark University College, Porsgrunn, NorwayBackground: From a philosophy of science perspective, the literature has posited that different research approaches influence field studies. Studies addressing interdisciplinary research have focused on the challenges of organizing and running interdisciplinary teams, cultural differences between and within disciplines, and constraints in conducting interdisciplinary research. Studies exploring and discussing the process and outcome of transferring observations to notes from an interdisciplinary point of view are not identified. The aim of this paper is to explore the characteristics of field notes created by researchers representing different disciplines and experiences.Methods: A case study using a modified dynamic observation method was employed. The analyses were initiated by a researcher who had not been involved in the data collection. The field notes were analyzed using three main steps.Results: The structures of both researchers' field notes were characterized by similarities in their descriptions, but the notes' foci and analytical levels differed.Conclusion: The findings contribute new insights concerning the execution of interdisciplinary observational studies. Our findings demonstrate that entering the field with different lenses produced richer and more varied data, providing a broader platform from which to discuss and interpret a study's findings. From a theoretical point of view, the findings enable a more nuanced discussion and a conceptual elaboration regarding how observational approaches should be pursued in future studies. On a practical level, the findings show that even if the researchers agree on what the overall focus in the observations should be, differences can occur in 16. Some notes about the nuclear data libraries International Nuclear Information System (INIS) Panini, G.C. 1984-01-01 The paper gives a short description of the main nuclear data collections. The features which are particular of each source are enhanced and compared. Notes about the Nuclear Data Processing are also outlined. The paper is intended as a preliminary approach for people interested in the Nuclear Data management 17. A note on notes: note taking and containment. Science.gov (United States) Levine, Howard B 2007-07-01 In extreme situations of massive projective identification, both the analyst and the patient may come to share a fantasy or belief that his or her own psychic reality will be annihilated if the psychic reality of the other is accepted or adopted (Britton 1998). In the example of' Dr. M and his patient, the paradoxical dilemma around note taking had highly specific transference meanings; it was not simply an instance of the generalized human response of distracted attention that Freud (1912) had spoken of, nor was it the destabilization of analytic functioning that I tried to describe in my work with Mr. L. Whether such meanings will always exist in these situations remains a matter to be determined by further clinical experience. In reopening a dialogue about note taking during sessions, I have attempted to move the discussion away from categorical injunctions about what analysis should or should not do, and instead to foster a more nuanced, dynamic, and pair-specific consideration of the analyst's functioning in the immediate context of the analytic relationship. There is, of course, a wide variety of listening styles among analysts, and each analyst's mental functioning may be affected differently by each patient whom the analyst sees. I have raised many questions in the hopes of stimulating an expanded discussion that will allow us to share our experiences and perhaps reach additional conclusions. Further consideration may lead us to decide whether note taking may have very different meanings for other analysts and analyst-patient pairs, and whether it may serve useful functions in addition to the one that I have described. 18. Description and assessment of regional sea-level trends and variability from altimetry and tide gauges at the northern Australian coast Science.gov (United States) Gharineiat, Zahra; Deng, Xiaoli 2018-05-01 This paper aims at providing a descriptive view of the low-frequency sea-level changes around the northern Australian coastline. Twenty years of sea-level observations from multi-mission satellite altimetry and tide gauges are used to characterize sea-level trends and inter-annual variability over the study region. The results show that the interannual sea-level fingerprint in the northern Australian coastline is closely related to El Niño Southern Oscillation (ENSO) and Madden-Julian Oscillation (MJO) events, with the greatest influence on the Gulf Carpentaria, Arafura Sea, and the Timor Sea. The basin average of 14 tide-gauge time series is in strong agreement with the basin average of the altimeter data, with a root mean square difference of 18 mm and a correlation coefficient of 0.95. The rate of the sea-level trend over the altimetry period (6.3 ± 1.4 mm/yr) estimated from tide gauges is slightly higher than that (6.1 ± 1.3 mm/yr) from altimetry in the time interval 1993-2013, which can vary with the length of the time interval. Here we provide new insights into examining the significance of sea-level trends by applying the non-parametric Mann-Kendall test. This test is applied to assess if the trends are significant (upward or downward). Apart from a positive rate of sea-level trends are not statistically significant in this region due to the effects of natural variability. The findings suggest that altimetric trends are not significant along the coasts and some parts of the Gulf Carpentaria (14°S-8°S), where geophysical corrections (e.g., ocean tides) cannot be estimated accurately and altimeter measurements are contaminated by reflections from the land. 19. A Revision of Lasionycta Aurivillius (Lepidoptera: Noctuidae for North America and notes on Eurasian species, with descriptions of 17 new species, 6 new subspecies, a new genus, and two new species of Tricholita Grote Directory of Open Access Journals (Sweden) Lars Crabo 2009-12-01 (Fabricius, comb. rev., Lasionycta dovrensis (Wocke, stat. rev. and L. fumida (Graeser, stat. rev. The genus Psammopolia Crabo & Lafontaine (type species: Polia wyatti Barnes & Benjamin is described, resulting in the following new combinations: Psammopolia arietis (Grote, comb. n., Psammopolia insolens (Grote, comb. n., Psammopolia ochracea (Smith, comb. n., Psammopolia sala (Troubridge & Mustelin, comb. n., and Psammopolia wyatti (Barnes & Benjamin, comb. n. Two new species of the related genus Tricholita Grote are described: T. ferrisi Crabo & Lafontaine from southwestern Arizona and T. knudsoni Crabo & Lafontaine from western Texas. Adults and genitalia of all North American Lasionycta and Psammopolia species and the new Tricholita are illustrated. Keys to species-groups and species are presented. DNA barcodes of 39 of the 43 species were sequenced and are presented as neighbor-joining phylograms. The barcodes support the taxonomy at genus and species-group level, but not consistently at the sub-group level. At the species-level, performance of DNA barcodes was variable; 17 of the 39 barcoded species exhibited haplotype variation discordant with morphology, phenotype and distribution. A high frequency (43.6 % of haplotypes were either shared among more than one species (representing eight species, or were closely similar and nested within haplotypic variation of other species (nine species. 20. On that Note... Science.gov (United States) Stein, Harry 1988-01-01 Provides suggestions for note-taking from books, lectures, visual presentations, and laboratory experiments to enhance student knowledge, memory, and length of attention span during instruction. Describes topical and structural outlines, visual mapping, charting, three-column note-taking, and concept mapping. Benefits and application of… 1. Making Notes, Making Meaning. Science.gov (United States) Burke, Jim 2002-01-01 Introduces notetaking tools used successfully with English-as-a-second-language students and low-achieving high school freshmen. Provides an overview of each tool and explains how students use them to take notes when reading textbooks and articles. Notes these tools and academic habits have helped students succeed in their mainstream academic… 2. Hanford Waste Vitrification Plant Quality Assurance Program description for high-level waste form development and qualification. Revision 3, Part 2 Energy Technology Data Exchange (ETDEWEB) 1993-08-01 The Hanford Waste Vitrification Plant Project has been established to convert the high-level radioactive waste associated with nuclear defense production at the Hanford Site into a waste form suitable for disposal in a deep geologic repository. The Hanford Waste Vitrification Plant will mix processed radioactive waste with borosilicate material, then heat the mixture to its melting point (vitrification) to forin a glass-like substance that traps the radionuclides in the glass matrix upon cooling. The Hanford Waste Vitrification Plant Quality Assurance Program has been established to support the mission of the Hanford Waste Vitrification Plant. This Quality Assurance Program Description has been written to document the Hanford Waste Vitrification Plant Quality Assurance Program. 3. User's guide for the implementation of level one of the proposed American National Standard Specifications for an information interchange data descriptive file on control data 6000/7000 series computers CERN Document Server Wiley, R A 1977-01-01 User's guide for the implementation of level one of the proposed American National Standard Specifications for an information interchange data descriptive file on control data 6000/7000 series computers 4. A note on Fukui’s note Directory of Open Access Journals (Sweden) Tsai Cheng-Yu Edwin 2015-12-01 Full Text Available This commentary relates Fukui’s (2015 note on weak vs. strong generation to two aspects of quantification in Chinese: quantifier scope and the syntactic licensing conditions of noninterrogative wh-expressions. It is shown that the phenomena under discussion echo Fukui’s (2015 view that only strong generation allows for a deeper understanding of natural language and that dependencies are to be distinguished structurally. 5. A study of actions in operative notes. Science.gov (United States) Wang, Yan; Pakhomov, Serguei; Burkart, Nora E; Ryan, James O; Melton, Genevieve B 2012-01-01 Operative notes contain rich information about techniques, instruments, and materials used in procedures. To assist development of effective information extraction (IE) techniques for operative notes, we investigated the sublanguage used to describe actions within the operative report 'procedure description' section. Deep parsing results of 362,310 operative notes with an expanded Stanford parser using the SPECIALIST Lexicon resulted in 200 verbs (92% coverage) including 147 action verbs. Nominal action predicates for each action verb were gathered from WordNet, SPECIALIST Lexicon, New Oxford American Dictionary and Stedman's Medical Dictionary. Coverage gaps were seen in existing lexical, domain, and semantic resources (Unified Medical Language System (UMLS) Metathesaurus, SPECIALIST Lexicon, WordNet and FrameNet). Our findings demonstrate the need to construct surgical domain-specific semantic resources for IE from operative notes. 6. Mysore study: A study of suicide notes. Science.gov (United States) Namratha, P; Kishor, M; Sathyanarayana Rao, T S; Raman, Rajesh 2015-01-01 Suicide is one of the leading causes of preventable deaths. Recent data suggest South India as one of the regions with highest suicide rates in the world. In 2013, 134,799 people committed suicide in India according to the statistics released by the National Crime Records Bureau. Suicide note is one of the most important sources to understand suicide, which may be beneficial in suicide prevention. Studies on suicidal notes from this part of the world are sparse. The aim was to study the themes in suicide notes that might be useful in prevention strategies. A descriptive study of all suicide notes of those individuals who committed suicide between 2010 and 2013 available with Police Department, Mysore district was obtained and analyzed. A total of 22 suicide note were available. A majority of suicide note was in age group of 16-40 years (86%) and most were men (59%). All suicide notes were handwritten, the majority (70%) in regional language Kannada. Length of notes varied from just few words to few pages. Contents of suicide notes included apology/shame/guilt (80%), love for those left behind (55%) and instruction regarding practical affairs (23%). Most have blamed none for the act (50%). 23% mentioned that they are committing suicide to prove their innocence. 32% mentioned a last wish. The majority of suicidal note contained "guilt" which is a strong indicator of possible depression in deceased. Creating awareness about suicide among public and ensuring access to professionals trained in suicide prevention is need of the hour in this part of the world. 7. Note: Possibilities of detecting the trace-level erosion products from an electric propulsion hollow cathode plasma source by the method of time-of-flight mass spectrometry Science.gov (United States) Ning, Zhong-Xi; Zhang, Hai-Guang; Zhu, Xi-Ming; Jiang, Bin-Hao; Zhou, Zhong-Yue; Yu, Da-Ren; An, Bing-Jian; Wang, Yan-Fei 2018-02-01 A hollow cathode produces electrons which neutralize ions from electric propulsion thrusters. After hundreds to thousands of hours of operation in space, the cathode materials can be significantly eroded due to ion bombardment. As a result, the electric propulsion system performance will be obviously changed or even fail. In this work, the erosion products from a LaB6 hollow cathode (widely used presently in electric propulsion systems) are studied by using a specific detection system, which consists of a molecular beam sampler and a time-of-flight mass spectrometer. This system measures trace-level-concentration (10-6-10-3) products. Boron (B), tantalum (Ta), and tungsten (W)—originating from the emitter, keeper, and orifice of the hollow cathode—are measured. It is found that the erosion rate is significantly influenced by the gas flow rate to the cathode. 8. MixedNotes DEFF Research Database (Denmark) Jokela, Tero; Lucero, Andrés 2014-01-01 Affinity Diagramming is a technique to organize and make sense of qualitative data. It is commonly used in Contextual Design and HCI research. However, preparing notes for and building an Affinity Diagram remains a laborious process, with a wide variety of different approaches and practices....... In this paper, we present MixedNotes, a novel technique to prepare physical paper notes for Affinity Diagramming, and a software tool to support this technique. The technique has been tested with large real-life Affinity Diagrams with overall positive results.... 9. [The values of different study designs on the levels of evidence: a descriptive analysis of the researches published in four general medical journals in 2009]. Science.gov (United States) Yang, Zu-yao; Zhang, Yuan; Wu, Shan-shan; Zhou, Yuan; DU, Yu-kun; Zhan, Si-yan 2010-12-01 To discuss the levels of evidence provided by different study designs. Websites of N Engl J Med, JAMA, Lancet, and BMJ were accessed to identify research articles (systematic review and meta-analysis included) published in 2009. A standardized data collection form was established using Epidata 3.1 software to extract the "title", "country of lead author", "clinical problem" (such as treatment, diagnosis, etc.) and "study design" of eligible studies. Descriptive statistics was conducted with SPSS 13.0. Over all, 844 studies were included, among which 35.7% were RCT, 9.4% systematic review and Meta-analysis, and 54.9% other types of studies. Regarding clinical problems, 34.2%, 19.7%, 13.7%, 6.0% and 5.1% of the included researches addressed the issues of treatment, etiology/risk factors, prevention, disease frequency and prognosis, respectively. The study designs that were most frequently adopted to explore these problems were RCT (70.6%), cohort study (44.6%), RCT (68.1%), cross-sectional study (56.9%), and cohort study (93.0%), respectively. High-level evidence does not come exclusively from RCT and systematic review, as each type of study may have its unique value in health related research. The clinical problem of interest, the previous work that has been done to approach the same issue, as well as other factors should be taken into account when deciding whether the selected study design is appropriate. 10. Lagos Notes and Records African Journals Online (AJOL) PROMOTING ACCESS TO AFRICAN RESEARCH ... Lagos Notes and Records is an annual, interdisciplinary journal of the humanities. ... Insuring the Nation: Europeans and the Emergence of Modern Insurance Business in Colonial Nigeria ... 11. Notes on Contributors African Journals Online (AJOL) NOTES ON CONTRIBUTORS. Samuel AMOAKO, Associate Researcher, South African Research Unit in Social Change. Contact Details: C/o Lucinda Bercony, Humanities Research Village (House No. 3). University of Johannesburg, P. O. Box 524. Bunting Road Campus, Auckland Park, 2006. 12. NCEP Internal Office Notes Data.gov (United States) National Oceanic and Atmospheric Administration, Department of Commerce — The National Centers for Environmental Prediction (NCEP) and its predecessors have produced internal publications, known as Office Notes, since the mid-1950's. In... 13. Writing a Condolence Note Science.gov (United States) ... through the eyes of others. For example, a mother whose son had died found out that her ... few suggestions on ending your condolence note: “Our love and support will always be here for you.” “ ... 14. A note on name individuation and identifying descriptions African Journals Online (AJOL) Kate H used to refer to either the famous philosopher or to the shipping magnate. ..... chain that produced the use of the name is a list containing the generic name and the ... arcane issues concerning currency individuation, object to my claim that I ... 15. Notes on the genus Digitaria, with descriptions of new species NARCIS (Netherlands) Henrard, J.Th. 1934-01-01 Some years ago I had the opportunity to study more extensively a very interesting group of grasses, belonging to what is now accepted as a distinct genus, the genus Digitaria, formerly belonging as a subgenus to the genus Panicum. As to living plants of this group I was familiar with two european 16. Analytical dynamics course notes CERN Document Server Lindenbaum, Samuel D 1994-01-01 This book comprises a set of lecture notes on rational mechanics, for part of the graduate physics curriculum, delivered by the late Prof. Shirley L. Quimby during his tenure at Columbia University, New York. The notes contain proofs of basic theorems, derivations of formulae and amplification of observations, as well as the presentation and solution of illustrative problems. Collateral readings from more than 50 source references are indicated at appropriate places in the text. 17. Notes on instrumentation and control CERN Document Server Roy, G J 2013-01-01 Notes on Instrumentation and Control presents topics on pressure (i.e., U-tube manometers and elastic type gauges), temperature (i.e. glass thermometer, bi-metallic strip thermometer, filled system thermometer, vapor pressure thermometer), level, and flow measuring devices. The book describes other miscellaneous instruments, signal transmitting devices, supply and control systems, and monitoring systems. The theory of automatic control and semi-conductor devices are also considered. Marine engineers will find the book useful. 18. Notes on modeling and simulation Energy Technology Data Exchange (ETDEWEB) Redondo, Antonio [Los Alamos National Lab. (LANL), Los Alamos, NM (United States) 2017-03-10 These notes present a high-level overview of how modeling and simulation are carried out by practitioners. The discussion is of a general nature; no specific techniques are examined but the activities associated with all modeling and simulation approaches are briefly addressed. There is also a discussion of validation and verification and, at the end, a section on why modeling and simulation are useful. 19. Note Taking and Note Sharing While Browsing Campaign Information DEFF Research Database (Denmark) Robertson, Scott P.; Vatrapu, Ravi; Abraham, George 2009-01-01 Participants were observed while searching and browsing the internet for campaign information in a mock-voting situation in three online note-taking conditions: No Notes, Private Notes, and Shared Notes. Note taking significantly influenced the manner in which participants browsed for information... 20. Descriptive statistics. Science.gov (United States) Nick, Todd G 2007-01-01 Statistics is defined by the Medical Subject Headings (MeSH) thesaurus as the science and art of collecting, summarizing, and analyzing data that are subject to random variation. The two broad categories of summarizing and analyzing data are referred to as descriptive and inferential statistics. This chapter considers the science and art of summarizing data where descriptive statistics and graphics are used to display data. In this chapter, we discuss the fundamentals of descriptive statistics, including describing qualitative and quantitative variables. For describing quantitative variables, measures of location and spread, for example the standard deviation, are presented along with graphical presentations. We also discuss distributions of statistics, for example the variance, as well as the use of transformations. The concepts in this chapter are useful for uncovering patterns within the data and for effectively presenting the results of a project. 1. The Impact of Novice Counselors' Note-Taking Behavior on Recall and Judgment Science.gov (United States) 2014-01-01 Purpose: To examine the effect of note-taking on novice counselors' recall and judgment of interview information in four situations: no notes, taking notes, taking notes and reviewing these notes, and reviewing notes taken by others. Method: The sample included 13 counselors-in-training recruited from a master's level training program in… 2. Interim oceanographic description of the North-East-Atlantic site for the disposal of low-level radioactive waste. Vol. 2 International Nuclear Information System (INIS) Dickson, R.R.; Gurbutt, P.A.; Kershaw, P.J. 1986-01-01 The NEA Co-ordinated Research and Environmental Surveillance Programme (CRESP) related to sea disposal of radioactive waste was started in 1981 following a recommendation of the Group of experts convened every five years by NEA to review the continued suitability of the dumping site for radioactive waste in the North-East Atlantic. The objective of CRESP is to increase the available scientific data base related to the oceanographic and biological characteristics of the dump site and elaborate a site specific model of the transfers of radionuclides to human populations. Volume one of the ''Interim Oceanographic Description of the North-East Atlantic Site for the Disposal of Low-Level Radioactive Waste'' was published in early 1983. It was an attempt to identify remaining gaps in current knowledge of conditions at the site and relate these conditions to the physical environment of the North-East Atlantic Ocean as a whole. The amount of data obtained by the CRESP Programme is now sufficient to justify publication of this second volume. Scientists present results of research which is of direct relevance to a better assessment of the impact from dumping radioactive waste in the North-East Atlantic, in particular an evaluation of the potential radiation doses to man. These two volumes represent part of the scientific contribution of the CRESP Programme to the 1985 Review of the Continued Suitability of the North-East Atlantic dump site 3. Physicians’ Progress Notes DEFF Research Database (Denmark) Bansler, Jørgen; Havn, Erling C.; Mønsted, Troels 2013-01-01 in patient care, they have not dealt specifically with the role, structure, and content of the progress notes. As a consequence, CSCW research has not yet taken fully into account the fact that progress notes are coordinative artifacts of a rather special kind, an open-ended chain of prose texts, written...... sequentially by cooperating physicians for their own use as well as for that of their colleagues. We argue that progress notes are the core of the medical record, in that they marshal and summarize the overwhelming amount of data that is available in the modern hospital environment, and that their narrative...... format is uniquely adequate for the pivotal epistemic aspect of cooperative clinical work: the narrative format enables physicians to not only record ‘facts’ but also—by filtering, interpreting, organizing, and qualifying information—to make sense and act concertedly under conditions of uncertainty... 4. Grouping Notes Through Nodes DEFF Research Database (Denmark) Dove, Graham; Abildgaard, Sille Julie Jøhnk; Biskjær, Michael Mose , both individually and when grouped, and their role in categorisation in semantic long-term memory. To do this, we adopt a multimodal analytical approach focusing on interaction between humans, and between humans and artefacts, alongside language. We discuss in detail examples of four different...... externalisation functions served by Post-ItTM notes, and show how these functions are present in complex overlapping combinations rather than being discrete. We then show how the temporal development of Post-ItTM note interactions supports categorisation qualities of semantic long-term memory.... 5. Grouping Notes Through Nodes DEFF Research Database (Denmark) Dove, Graham; Abildgaard, Sille Julie; Biskjær, Michael Mose 2017-01-01 , both individually and when grouped, and their role in categorisation in semantic long-term memory. To do this, we adopt a multimodal analytical approach focusing on interaction between humans, and between humans and artefacts, alongside language. We discuss in detail examples of four different...... externalisation functions served by Post-ItTM notes, and show how these functions are present in complex overlapping combinations rather than being discrete. We then show how the temporal development of Post-ItTM note interactions supports categorisation qualities of semantic long-term memory.... 6. Note by Note: a New Revolution in Cooking OpenAIRE Burke, Roisin; Danaher, Pauline 2016-01-01 Note by note cooking is an application of Molecular Gastronomy. It was first proposed by French Physical Chemist and Molecular Gastronomy Co-founder, Hervé This. Note by Note dishes are being created as part of Ph.D. research in the Dublin Institute of Technology, Cathal Brugha Street. 7. Note taking, review, memory, and comprehension. Science.gov (United States) Bohay, Mark; Blakely, Daniel P; Tamplin, Andrea K; Radvansky, Gabriel A 2011-01-01 In previous work assessing memory at various levels of representation, namely the surface form, textbase, and situation model levels, participants read texts but were otherwise not actively engaged with the texts. The current study tested the influence of active engagement with the material via note taking, along with the opportunity to review such notes, and the modality of presentation (text vs. spoken). The influence of these manipulations was assessed both immediately and 1 week later. In Experiment 1 participants read a text, whereas in Experiment 2 participants watched a video recording of the material being read as a lecture. For each experiment the opportunity to take notes was manipulated within participants, and the opportunity to review these notes before the test was manipulated between participants. Note taking improved performance at the situation model level in both experiments, although there was also some suggestion of benefit for the surface form. Thus, active engagement with material, such as note taking, appears to have the greatest benefit at the deeper levels of understanding. 8. Note Taking for Geography Students. Science.gov (United States) Kneale, Pauline E. 1998-01-01 Addresses geography students' questions about why, when, and how to take notes. Outlines a step-by-step process for taking notes from written sources and from class lectures. Discusses what types of notes are appropriate for various types of sources. Suggests some ideas for making notes useful for individual learning styles. (DSK) 9. OpenLabNotes DEFF Research Database (Denmark) List, Markus; Franz, Michael; Tan, Qihua 2015-01-01 be advantageous if an ELN was Integrated with a laboratory information management system to allow for a comprehensive documentation of experimental work including the location of samples that were used in a particular experiment. Here, we present OpenLabNotes, which adds state-of-the-art ELN capabilities to Open......LabFramework, a powerful and flexible laboratory information management system. In contrast to comparable solutions, it allows to protect the intellectual property of its users by offering data protection with digital signatures. OpenLabNotes effectively Closes the gap between research documentation and sample management......, thus making Open-Lab Framework more attractive for laboratories that seek to increase productivity through electronic data management.... 10. A note on Marx OpenAIRE Olesen, Finn; Jensen, Frank 2001-01-01 Throughout all his life Karl Marx wrote angrily about capitalism. By use of a dialectic approach he was convinced that the working class had to unite and make a social revolution and thereby free them selves from exploitation. Marx himself was in many ways a dialectic person as we try to show in the note. So in some sense he became one with his scientific methodology. 11. Descriptive data analysis. Science.gov (United States) Thompson, Cheryl Bagley 2009-01-01 This 13th article of the Basics of Research series is first in a short series on statistical analysis. These articles will discuss creating your statistical analysis plan, levels of measurement, descriptive statistics, probability theory, inferential statistics, and general considerations for interpretation of the results of a statistical analysis. 12. Notes on Laser Acceleration International Nuclear Information System (INIS) Tajima, T. 2008-01-01 This note intends to motivate our effort toward the advent of new methods of particle acceleration, utilizing the fast rising laser technology. By illustrating the underlying principles in an intuitive manner and thus less jargon-clad fashion, we seek a direction in which we shall be able to properly control and harness the promise of laser acceleration. First we review the idea behind the laser wakefield. We then go on to examine ion acceleration by laser. We examine the sheath acceleration in particular and look for the future direction that allows orderly acceleration of ions in high energies 13. MISCELLANEOUS BOTANICAL NOTES 2 Directory of Open Access Journals (Sweden) A.J.G.H KOSTERMANS 2014-01-01 Full Text Available 1.   Durio  cupreus Ridley is considered to  represent a  distinct  species.2.   Durio wyatt-smithii Kosterm. is reported from Borneo.3.   Machilus nervosa Merr. represents Meliosma bontoeensis Merr.4.   Beilschmiedia brassii Allen represents Vavaea brassii (Allen Kosterm.5.   The author of the generic name Heritiera is Aiton.6.   Heritiera macrophylla (non Wall. Merr. is conspecific with H. ungus-tata Pierre.7.   Some specimens from N. Celebes, attributed formerly to H. sylvatica Merr., belong to H. arafurensis Kosterm.8.   Additional note on Heritiera littoralis Ait. and H. macrophylla Wall, ex Kurz.9.   Heritiera   montana   Kosterm.,   nov.   spec,   from   New   Guinea   and H. khidii Kosterm., nov. spec, from Northern Siam.10.   Additional note on Heritiera, novoguineensis Kosterm. and H. pereo-riacea Kosterm. and an undescribed species.11.   Heritiera acuminata Wall, ex Kurz represents a distinct species.12.   Heritiera  solomonensis  Kosterm.,  nov.  spec,  from the  Solomon  Isl.13.   A note on Firmiana bracteata A. DC.14.   Firmiana fulgens (Wall, ex King  Corner is based on a mixtum com-positum and has been the source of constant confusion. For the element, which occurs in Malaysia a new name is coined: F. malayana Kosterm. It does not occur in Tenasserim.15.   A revised bibliography of Firmiana colorata R. Br., F. pallens Stearn and F. malayana Kosterm. is presented.16.   Additional note on Firmiana hainanensis Kosterm.17.   Firmiana kerrii (Craib Kosterm., comb, nov., based on Sterculia kerrii Craib.18.   Additional specimens of Firmiana papuana Mildbr.19.   Cryptocarya hintonii Allen is referred to Primus as Primus hintonii (Allen  Kosterm.20.   Beilschmiedia wallichiana (G. Don   Kosterm., based on Sideroxylon wallichianum, G. Don, is described. Formerly it was relegated to Litsea by Kurz.21.   New species in Lauraceae: Beilschmiedia aborensis Kosterm., B 14. Notes on functional analysis CERN Document Server Bhatia, Rajendra 2009-01-01 These notes are a record of a one semester course on Functional Analysis given by the author to second year Master of Statistics students at the Indian Statistical Institute, New Delhi. Students taking this course have a strong background in real analysis, linear algebra, measure theory and probability, and the course proceeds rapidly from the definition of a normed linear space to the spectral theorem for bounded selfadjoint operators in a Hilbert space. The book is organised as twenty six lectures, each corresponding to a ninety minute class session. This may be helpful to teachers planning a course on this topic. Well prepared students can read it on their own. 15. Figures of Merit Software: Description, User's Guide, Installation Notes, Versions Description, and License Agreement Science.gov (United States) hoelzer, H. D.; Fourroux, K. A.; Rickman, D. L.; Schrader, C. M. 2011-01-01 Figures of Merit (FoMs) and the FoM software provide a method for quantitatively evaluating the quality of a regolith simulant by comparing the simulant to a reference material. FoMs may be used for comparing a simulant to actual regolith material, specification by stating the value a simulant s FoMs must attain to be suitable for a given application and comparing simulants from different vendors or production runs. FoMs may even be used to compare different simulants to each other. A single FoM is conceptually an algorithm that computes a single number for quantifying the similarity or difference of a single characteristic of a simulant material and a reference material and provides a clear measure of how well a simulant and reference material match or compare. FoMs have been constructed to lie between zero and 1, with zero indicating a poor or no match and 1 indicating a perfect match. FoMs are defined for modal composition, particle size distribution, particle shape distribution, (aspect ratio and angularity), and density. This TM covers the mathematics, use, installation, and licensing for the existing FoM code in detail. 16. Towards a worldsheet description of N=8 supergravity International Nuclear Information System (INIS) Lipstein, Arthur; Schomerus, Volker 2015-10-01 In this note we address the worldsheet description of 4-dimensional N=8 supergravity using ambitwistors. After gauging an appropriate current algebra, we argue that the only physical vertex operators correspond to the N=8 supermultiplet. It has previously been shown that worldsheet correlators give rise to supergravity tree level scattering amplitudes. We extend this work by proposing a definition for genus-one amplitudes that passes several consistency checks such as exhibiting modular invariance and reproducing the expected infrared behavior of 1-loop supergravity amplitudes. 17. Note e Recensioni Directory of Open Access Journals (Sweden) a cura di Mariagrazia Portera 2013-12-01 Full Text Available Volumi Winfried Menninghaus, La promessa della bellezza, [Fabrizio Desideri, p. 272] • David Rothenberg, Survival of the Beautiful. Art, Science and Evolution [Danae Crocchiola, p. 274] • Lev Manovich, Software Takes Command [Angela Maiello, p. 277]. Note Method in Aesthetics: Philosophy, Evolution and the Cognitive Sciences [Aaron Meskin, Matthew Kieran, Gregory Currie, p. 280] • L’Abitare possibile. Estetica, Architettura e New Media, Ravello, Auditorium Oscar Niemeyer, 28-30 maggio 2013 [Sara Matetich, p. 282] • Copenhagen Summer School in Phenomenology and Philosophy of Mind, University of Copenhagen, 12-16 Agosto 2013 [Raoul Frauenfelder, p. 289] • Ciò che è vivo e ciò che è morto nell’estetica di Archibald Ali- son. Nota a margine del convegno: Neoestetica ed emozione. Archibald Alison e l’estetica con- temporanea, Palermo, 4-5 ottobre 2013 [Giuseppe Pucci, p. 294 18. Notes on quantum groups International Nuclear Information System (INIS) Pressley, A.; Chari, V.; Tata Inst. of Fundamental Research, Bombay 1990-01-01 The authors presents an introduction to quantum groups defined as a deformation of the universal enveloping algebra of a Lie algebra. After the description of Hopf algebras with some examples the approach of Drinfel'd and Jimbo is described, where the quantization of a Lie algebra represents a Hopf algebra, defined over the algebra of formal power series in an indetermined h. The authors show that this approach arises from a r-matrix, which satisfies the classical Yang-Baxter equation. As example quantum sl 2 is considered. Furthermore the approaches of Manin and Woroniwicz and the R-matrix approach are described. (HSI) 19. System Description: DEFF Research Database (Denmark) 2009-01-01 Delphin is a functional programming language [Adam Poswolsky and Carsten Schürmann. Practical programming with higher-order encodings and dependent types. In European Symposium on Programming (ESOP), 2008] utilizing dependent higher-order datatypes. Delphin's two-level type-system cleanly separates... 20. LLW notes. Volume 11, No.8 International Nuclear Information System (INIS) 1996-12-01 'LLW Notes' is distributed by Afton Associates, Inc. to Low-Level Radioactive Waste Forum Participants and other state, and compact officials identified by those Participants to receive 'LLW Notes'. The Low-Level Radioactive Waste Forum (LLW Forum) is an association of state and compact representatives, appointed by governors and compact commissions, established to facilitate state and compact implementation of the Low-Level Radioactive Waste Policy Act of 1980 and the Low-Level Radioactive Waste Policy Amendments Act of 1985 and to promote the objectives of low-level radioactive waste regional compacts. The LLW Forum provides an opportunity for state and compact officials to share information with one another and to exchange views with officials of federal agencies and other interested parties 1. LLW notes. Vol. 11, No. 1 International Nuclear Information System (INIS) 1996-02-01 'LLW Notes' is distributed by Afton Associates, Inc. to Low-Level Radioactive Waste Forum Participants and other state and compact officials identified by those Participants to receive 'LLW Notes'. The Low-Level Radioactive Waste Forum (LLW Forum) is an association of state and compact representatives, appointed by governors and compact commissions, established to facilitate state and compact implementation of the Low-Level Radioactive Waste Policy Act of 1980 and the Low-Level Radioactive Waste Policy Amendments Act of 1985 and to promote the objectives of low-level radioactive waste regional compacts. The LLW Forum provides an opportunity for state and compact officials to share information with one another and to exchange views with officials of federal agencies and other interested parties 2. LLW notes. Vol. 11, No. 1 Energy Technology Data Exchange (ETDEWEB) NONE 1996-02-01 LLW Notes is distributed by Afton Associates, Inc. to Low-Level Radioactive Waste Forum Participants and other state and compact officials identified by those Participants to receive LLW Notes. The Low-Level Radioactive Waste Forum (LLW Forum) is an association of state and compact representatives, appointed by governors and compact commissions, established to facilitate state and compact implementation of the Low-Level Radioactive Waste Policy Act of 1980 and the Low-Level Radioactive Waste Policy Amendments Act of 1985 and to promote the objectives of low-level radioactive waste regional compacts. The LLW Forum provides an opportunity for state and compact officials to share information with one another and to exchange views with officials of federal agencies and other interested parties. 3. LLW notes, Vol. 11, No. 2 Energy Technology Data Exchange (ETDEWEB) NONE 1996-03-01 LLW Notes is distributed by Afton Associates, Inc. to Low-Level Radioactive Waste Forum Participants and other state, and compact officials identified by those Participants to receive LLW Notes. The Low-Level Radioactive Waste Forum (LLW Forum) is an association of state and compact representatives, appointed by governors and compact commissions, established to facilitate state and compact implementation of the Low-Level Radioactive Waste Policy Act of 1980 and the Low-Level Radioactive Waste Policy amendments Act of 1985 and to promote the objectives of low-level radioactive waste regional compacts. The LLW Forum provides an opportunity for state and compact officials to share information with one another and to exchange views with officials of federal agencies and other interested parties. 4. LLW notes, Vol. 11, No. 2 International Nuclear Information System (INIS) 1996-03-01 'LLW Notes' is distributed by Afton Associates, Inc. to Low-Level Radioactive Waste Forum Participants and other state, and compact officials identified by those Participants to receive LLW Notes. The Low-Level Radioactive Waste Forum (LLW Forum) is an association of state and compact representatives, appointed by governors and compact commissions, established to facilitate state and compact implementation of the Low-Level Radioactive Waste Policy Act of 1980 and the Low-Level Radioactive Waste Policy amendments Act of 1985 and to promote the objectives of low-level radioactive waste regional compacts. The LLW Forum provides an opportunity for state and compact officials to share information with one another and to exchange views with officials of federal agencies and other interested parties 5. Revisiting convergence: A research note. Science.gov (United States) Clark, Rob 2015-09-01 A number of recent studies show that income inequality is declining between countries. In this research note, I question the significance of this trend by examining the role of initial conditions in producing convergence. An important (but neglected) property of inequality dynamics is the tendency for extreme distributions to become more moderate. When income disparities are large, the subsequent trend is biased toward convergence. Conversely, when initial conditions approach parity, divergence becomes the more likely long-term outcome. I apply this principle to trends in GDP PC across 127 countries during the 1980-2010 period. Using counterfactual analysis, I manipulate the initial level of inequality in GDP PC while holding constant each country's observed growth rate during the sample period. I find that the growth dynamics of GDP PC produce either convergence or divergence based simply on the initial distribution of income. The point of transition occurs at a moderate level of inequality, whether using population weights (Gini=.365) or not (Gini=.377). I conclude that the recent convergence observed in GDP PC is primarily a function of large income gaps between countries and would not have materialized at more moderate levels of initial inequality. By contrast, an examination of the pre-1950 period reveals divergent growth patterns that are not sensitive to initial conditions. Copyright © 2015 Elsevier Inc. All rights reserved. 6. PHENIX WBS notes International Nuclear Information System (INIS) 1994-02-01 The Work Breakdown Structure (WBS) Book begins with this Overview section, which contains the high-level summary cost estimate, the cost profile, and the global construction schedule. The summary cost estimate shows the total US cost and the cost in terms of PHENIX construction funds for building the PHENIX detector. All costs in the WBS book are shown in FY 1993 dollars. Also shown are the institutional and foreign contributions, the level of pre-operations funding, and the cost of deferred items. Pie charts are presented at PHENIX WBS level 1 and 2 that show this information. The PHENIX construction funds are shown broken down to PHENIX WBS level 3 items per fiscal year, and the resulting profile is compared to the RHIC target profile. An accumulated difference of the two profiles is also shown. The PHENIX global construction schedule is presented at the end of the Overview section. Following the Overview are sections for each subsystem. Each subsystem section begins with a summary cost estimate, cost profile, and critical path. The total level 3 cost is broken down into fixed costs (M ampersand S), engineering costs (EDIA) and labor costs. Costs are further broken down in terms of PHENIX construction funds, institutional and foreign contributions, pre-operations funding, and deferred items. Also shown is the contingency at level 3 and the level 4 breakdown of the total cost. The cost profile in fiscal years is shown at level 3. The subsystem summaries are followed by the full cost estimate and schedule sheets for that subsystem. These detailed sheets are typically carried down to level 7 or 8. The cost estimate Total, M ampersand S, EDIA, and Labor breakdowns, as well as contingency, for each WBS entry 7. Notes in Colombian Herpetology, II Notes in Colombian Herpetology, II Directory of Open Access Journals (Sweden) Dunn Emmett Reid 1944-03-01 Full Text Available The Lizard Genus Echinosaura (Teiidae in Colombia / Notes on the habits of the Tadpole-Carrying Frog Hyloxalus granuliventris / A New Marsupian Frog (Gastrotheca from Colombia The Lizard Genus Echinosaura (Teiidae in Colombia / Notes on the habits of the Tadpole-Carrying Frog Hyloxalus granuliventris / A New Marsupian Frog (Gastrotheca from Colombia. 8. Notes on Three Newly Naturalized Plants in Taiwan Directory of Open Access Journals (Sweden) Shih-Huei Chen 2005-03-01 Full Text Available Chloris divaricata R. Br. var. cynodontoides (Bal. Lazarides, Boerhavia coccinea Mill., and Hyptis pectinata (L. Poit. are recently found naturalized in Taiwan. The present study gives the taxonomic description and line drawings of the three species. In addition, their distribution and notes on ecology and taxonomy are provided. 9. Effect of fat level on the perception of five flavor chemicals in ice cream with or without fat mimetics by using a descriptive test. Science.gov (United States) Liou, B K; Grün, I U 2007-10-01 Fat mimetics are commonly used in the manufacture of low-fat and fat-free ice creams. However, the use of fat mimetics affects flavor and texture characteristics of ice cream, which results in decreased overall acceptability by consumers. The initial objective of this study was to investigate the release behavior of 5 strawberry flavor compounds in ice creams with Simplesse((R)), Litesse((R)), and Litesse((R))/Simplesse((R)) mixes using descriptive analysis. Fat mimetics and flavor formulation significantly influenced the perception of Furaneoltrade mark (cooked sugar flavor), alpha-ionone (violet flavor), and gamma-undecalactone (peach flavor), but there was no interaction between ice cream type and flavor formulation for the 3 flavors. Furaneol and ethyl-3-methyl-3-phenylglycidate (candy flavor) were perceived more strongly in full-fat ice cream, while cis-3-hexen-1-ol (grassy flavor), alpha-ionone, and gamma-undecalactone were perceived more strongly in low-fat ice cream. Ice creams with Simplesse and full-fat ice cream had similar sensory characteristics, while ice creams with Litesse were similar to low-fat ice creams in flavor characteristics, and ice creams with Litesse/Simplesse mixes were closer in flavor profile to low-fat ice cream but had similar texture properties to those of full-fat ice cream. Simplesse was found to be a better fat mimetic for duplicating the flavor profiles and mouthfeel of full-fat ice cream. 10. Description of the U.S. Geological Survey's water-quality sampling and water-level monitoring program at the Hallam Nuclear Facility, August through September 1997 International Nuclear Information System (INIS) 1997-01-01 A water-quality and water-level program between the US Department of Energy (USDOE) and the US Geological Survey (USGS) was re-established in August 1997 to (1) collect one set of water-quality samples from 17 of the 19 USDOE monitor wells, and (2) make five water-level measurements during a 2-month period from the 19 USDOE monitor wells at the Hallam Nuclear Facility, Hallam, Nebraska. Data from these wells are presented 11. Improving operation notes to meet British Orthopaedic Association guidelines. Science.gov (United States) Morgan, David; Fisher, Noel; Ahmad, Aman; Alam, Fazle 2009-04-01 Operation notes are an important part of medical records for clinical, academic and medicolegal reasons. This study audited the quality of operative note keeping for total knee replacements against the standards set by the British Orthopaedic Association (BOA). A prospective review of all patients undergoing total knee replacement at a district general hospital over 8 months. Data recorded were compared with those required by the BOA good-practice guidelines. Change in practice was implemented and the audit cycle completed. Data were statistically analysed. A total of 129 operation notes were reviewed. There was a significant improvement in the mean number of data points recorded from 9.6 to 13.1. The least well recorded data were diagnosis, description of findings, alignment and postoperative flexion range. All had a significant improvement except description of findings. The operating surgeon writing the note improved from 56% to 67%. Detailed postoperative instructions also improved in quality. Surgeon education and the use of a checklist produce better quality total knee replacement operation notes in line with BOA guidelines. Further improvements may be made by making the data points part of the operation note itself. 12. Note on plagiarism especially of plagiarism (including selfplagiarism), as something directly affecting the integrity of the sci- entific process. Such behaviour is unacceptable and deserves exposure and an appropriate level of penalty. In case it is inadvertently published, a paper containing plagiarized material steals credit from the original. 13. A note on axial symmetries International Nuclear Information System (INIS) Beetle, Christopher; Wilder, Shawn 2015-01-01 This note describes how to characterize and normalize an axial Killing field on a general Riemannian geometry or four-dimensional Lorentzian geometry. No global assumptions are necessary, such as that the orbits of the Killing field all have period 2π. Rather, any Killing field that vanishes at at least one point necessarily has the expected global properties. (note) 14. The Anatomy of a Note. Science.gov (United States) Moore, Herb 1986-01-01 Suggests that students can learn the physics of a musical note by learning how to synthesize sounds on a computer. Discusses ADSR (attack, decay, sustain, and release of a note) and includes a program (with listing) which students can use to examine ADSR on a Commodore 64 microcomputer. (JN) 15. Factors Associated with Maternal Serum Levels of Perfluoroalkyl Substances and Organochlorines: A Descriptive Study of Parous Women in Norway and Sweden Science.gov (United States) Lauritzen, Hilde B.; Larose, Tricia L.; Øien, Torbjørn; Odland, Jon Ø.; van de Bor, Margot; Jacobsen, Geir W.; Sandanger, Torkjel M. 2016-01-01 Introduction Perfluoroalkyl substances (PFASs) and organochlorines (OCs) are ubiquitous and persistent in the environment and proposed endocrine disrupting chemicals (EDCs). They can be transferred across the placenta during pregnancy, and studies suggest that the prenatal period may be particularly sensitive for influences on fetal growth and development. Several studies have investigated socio-demographic and pregnancy related factors associated with maternal serum PFAS and OC levels, but few studies have been conducted in time periods with increasing emissions of PFASs and recent emissions of OCs. Methods Serum from 424 pregnant women participating in the NICHD Scandinavian Successive Small-for-gestational Age (SGA) births study was collected in 1986–1988, and analyses of two PFASs and six OCs were conducted. Associations between EDCs and geographic, time dependent, socio-demographic and pregnancy related variables were evaluated by using multivariable linear regression models. Results Previous breastfeeding duration, time since last breastfeeding period, sampling date and country of residence were important factors associated with serum levels of PFOS and PFOA. Smoking status and pre-pregnancy BMI were negatively associated with PFOS, and maternal height was borderline negatively associated with PFOS and PFOA. Glomerular filtration rate (GFR) was negatively associated with PFOS in a sub-sample. Maternal serum levels of OCs were positively associated with maternal age, and negatively associated with previous breastfeeding duration and sampling date. Smoking had a consistently negative association with PCB 118 in a dose-dependent manner. Education level, pre-pregnancy BMI and alcohol consumption varied in importance according to the compound under study. Conclusions Several maternal factors, including potentially modifiable factors, markers of pregnancy physiology and factors also related to perinatal outcomes were associated with EDC levels. Results from this 16. Factors Associated with Maternal Serum Levels of Perfluoroalkyl Substances and Organochlorines: A Descriptive Study of Parous Women in Norway and Sweden. Directory of Open Access Journals (Sweden) Hilde B Lauritzen Full Text Available Perfluoroalkyl substances (PFASs and organochlorines (OCs are ubiquitous and persistent in the environment and proposed endocrine disrupting chemicals (EDCs. They can be transferred across the placenta during pregnancy, and studies suggest that the prenatal period may be particularly sensitive for influences on fetal growth and development. Several studies have investigated socio-demographic and pregnancy related factors associated with maternal serum PFAS and OC levels, but few studies have been conducted in time periods with increasing emissions of PFASs and recent emissions of OCs.Serum from 424 pregnant women participating in the NICHD Scandinavian Successive Small-for-gestational Age (SGA births study was collected in 1986-1988, and analyses of two PFASs and six OCs were conducted. Associations between EDCs and geographic, time dependent, socio-demographic and pregnancy related variables were evaluated by using multivariable linear regression models.Previous breastfeeding duration, time since last breastfeeding period, sampling date and country of residence were important factors associated with serum levels of PFOS and PFOA. Smoking status and pre-pregnancy BMI were negatively associated with PFOS, and maternal height was borderline negatively associated with PFOS and PFOA. Glomerular filtration rate (GFR was negatively associated with PFOS in a sub-sample. Maternal serum levels of OCs were positively associated with maternal age, and negatively associated with previous breastfeeding duration and sampling date. Smoking had a consistently negative association with PCB 118 in a dose-dependent manner. Education level, pre-pregnancy BMI and alcohol consumption varied in importance according to the compound under study.Several maternal factors, including potentially modifiable factors, markers of pregnancy physiology and factors also related to perinatal outcomes were associated with EDC levels. Results from this study are relevant to 17. A Description of Methodologies Used in Estimation of A-Weighted Sound Levels for FAA Advisory Circular AC-36-3B. Science.gov (United States) 1982-01-01 second) Dia propeller diameter (expressed in inches) T°F air temperature in degrees Farenheit T°C air temperature in degrees Celsius T:dBA total dBA...eMpiriC31 function to the absolute noise level ordinate. The term 240 log ( MH is the most sensitive and important part of the equation. The constant (240...standard day, zero wind, dry, zero gradient runway, at a sea level airport. 2. All aircraft operate at maximum takeoff gross weight. 3. All aircraft climb 18. Description of the U.S. Geological Survey's water-quality sampling and water-level monitoring program at the Hallam Nuclear Facility, June through September 1996 International Nuclear Information System (INIS) 1997-01-01 A water-quality and water-level program of the US Department of Energy (USDOE) in cooperation with the US Geological Survey (USGS) was re-established in June 1996 to develop six new USDOE observation wells, collect one set of water-quality samples from 17 of the 19 USDOE observation wells, and take monthly water-level measurements for a 3-month period in all 19 USDOE observation wells at the Hallam Nuclear Facility, Hallam, Nebraska. Thirteen of the observation wells were installed by HWS Consulting Group, Inc., in June 1993 and the remaining six were installed by Applied Research Associates in August 1995 19. Relativistic description of atomic nuclei International Nuclear Information System (INIS) Krutov, V.A. 1985-01-01 Papers on the relativistic description of nuclei are reviewed. The Brown and Rho ''small'' bag'' model is accepted for hardrons. Meson exchange potentials of the nucleon-nucleon interaction have been considered. Then the transition from a system of two interacting nucleons has been performed to the relativistic nucleus description as a multinucleon system on the basis of OBEP (one-boson exchange potential). The proboem of OPEP (one-pion-exchange potential) inclusion to a relativistic scheme is discussed. Simplicity of calculations and attractiveness of the Walecka model for specific computations and calculations was noted. The relativistic model of nucleons interacting through ''effective'' scalar and vector boson fields was used in the Walacka model for describing neutronaand nuclear mater matters 20. Comparing probabilistic and descriptive analyses of time–dose–toxicity relationship for determining no-observed-adverse-effect level in drug development International Nuclear Information System (INIS) Glatard, Anaïs; Berges, Aliénor; Sahota, Tarjinder; Ambery, Claire; Osborne, Jan; Smith, Randall; Hénin, Emilie; Chen, Chao 2015-01-01 The no-observed-adverse-effect level (NOAEL) of a drug defined from animal studies is important for inferring a maximal safe dose in human. However, several issues are associated with its concept, determination and application. It is confined to the actual doses used in the study; becomes lower with increasing sample size or dose levels; and reflects the risk level seen in the experiment rather than what may be relevant for human. We explored a pharmacometric approach in an attempt to address these issues. We first used simulation to examine the behaviour of the NOAEL values as determined by current common practice; and then fitted the probability of toxicity as a function of treatment duration and dose to data collected from all applicable toxicology studies of a test compound. Our investigation was in the context of an irreversible toxicity that is detected at the end of the study. Simulations illustrated NOAEL's dependency on experimental factors such as dose and sample size, as well as the underlying uncertainty. Modelling the probability as a continuous function of treatment duration and dose simultaneously to data from multiple studies allowed the estimation of the dose, along with its confidence interval, for a maximal risk level that might be deemed as acceptable for human. The model-based data integration also reconciled between-study inconsistency and explicitly provided maximised estimation confidence. Such alternative NOAEL determination method should be explored for its more efficient data use, more quantifiable insight to toxic doses, and the potential for more relevant animal-to-human translation. - Highlights: • Simulations revealed issues with NOAEL concept, determination and application. • Probabilistic modelling was used to address these issues. • The model integrated time-dose-toxicity data from multiple studies. • The approach uses data efficiently and may allow more meaningful human translation. 1. Comparing probabilistic and descriptive analyses of time–dose–toxicity relationship for determining no-observed-adverse-effect level in drug development Energy Technology Data Exchange (ETDEWEB) Glatard, Anaïs; Berges, Aliénor; Sahota, Tarjinder; Ambery, Claire [Clinical Pharmacology Modelling and Simulation, GlaxoSmithKline, 1 Iron Bridge Road, Uxbridge, UB11 1BT London (United Kingdom); Osborne, Jan [Non-Clinical Safety Projects, GlaxoSmithKline, Ware (United Kingdom); Smith, Randall [Computational Toxicology, GlaxoSmithKline, Upper Merion (United States); Hénin, Emilie [UMR 5558 Laboratoire Biométrie et de Biologie Evolutive, Equipe EMET (Evaluation et Modélisation des Effets Thérapeutiques), Université Claude Bernard Lyon1, Service de Pharmacologie Clinique et Essais Thérapeutiques, Hospices Civils de Lyon, Lyon (France); Chen, Chao, E-mail: [email protected] [Clinical Pharmacology Modelling and Simulation, GlaxoSmithKline, 1 Iron Bridge Road, Uxbridge, UB11 1BT London (United Kingdom) 2015-10-15 The no-observed-adverse-effect level (NOAEL) of a drug defined from animal studies is important for inferring a maximal safe dose in human. However, several issues are associated with its concept, determination and application. It is confined to the actual doses used in the study; becomes lower with increasing sample size or dose levels; and reflects the risk level seen in the experiment rather than what may be relevant for human. We explored a pharmacometric approach in an attempt to address these issues. We first used simulation to examine the behaviour of the NOAEL values as determined by current common practice; and then fitted the probability of toxicity as a function of treatment duration and dose to data collected from all applicable toxicology studies of a test compound. Our investigation was in the context of an irreversible toxicity that is detected at the end of the study. Simulations illustrated NOAEL's dependency on experimental factors such as dose and sample size, as well as the underlying uncertainty. Modelling the probability as a continuous function of treatment duration and dose simultaneously to data from multiple studies allowed the estimation of the dose, along with its confidence interval, for a maximal risk level that might be deemed as acceptable for human. The model-based data integration also reconciled between-study inconsistency and explicitly provided maximised estimation confidence. Such alternative NOAEL determination method should be explored for its more efficient data use, more quantifiable insight to toxic doses, and the potential for more relevant animal-to-human translation. - Highlights: • Simulations revealed issues with NOAEL concept, determination and application. • Probabilistic modelling was used to address these issues. • The model integrated time-dose-toxicity data from multiple studies. • The approach uses data efficiently and may allow more meaningful human translation. 2. Calculation notes that support accident scenario and consequence determination of a waste tank criticality International Nuclear Information System (INIS) Marusich, R.M. Westinghouse Hanford 1996-01-01 The purpose of this calculation note is to provide the basis for criticality consequences for the Tank Farm Safety Analysis Report (FSAR). Criticality scenario is developed and details and description of the analysis methods are provided 3. Feasibility of disposal of high-level radioactive waste into the seabed. Volume 5: Dispersal of radionuclides in the oceans: Models, data sets and regional descriptions International Nuclear Information System (INIS) Marietta, M.G.; Simmons, W.F. 1988-01-01 One of the options suggested for disposal of high-level radioactive waste resulting from the generation of nuclear power is burial beneath the deep ocean floor in geologically stable sediment formations which have no economic value. The 8-volume series provides an assessment of the technical feasibility and radiological safety of this disposal concept based on the results obtained by ten years of co-operation and information exchange among the Member countries participating in the NEA Seabed Working Group. This report summarizes the development of a realistic and credible methodology to describe the oceanic dispersion of radionuclides for risk assessment calculations 4. Note on bouncing backgrounds Science.gov (United States) de Haro, Jaume; Pan, Supriya 2018-05-01 The theory of inflation is one of the fundamental and revolutionary developments of modern cosmology that became able to explain many issues of the early Universe in the context of the standard cosmological model (SCM). However, the initial singularity of the Universe, where physics is indefinite, is still obscure in the combined SCM +inflation scenario. An alternative to SCM +inflation without the initial singularity is thus always welcome, and bouncing cosmology is an attempt at that. The current work is thus motivated to investigate the bouncing solutions in modified gravity theories when the background universe is described by the spatially flat Friedmann-Lemaître-Robertson-Walker (FLRW) geometry. We show that the simplest way to obtain the bouncing cosmologies in such spacetime is to consider some kind of Lagrangian whose gravitational sector depends only on the square of the Hubble parameter of the FLRW universe. For these modified Lagrangians, the corresponding Friedmann equation, a constraint in the dynamics of the Universe, depicts a curve in the phase space (H ,ρ ), where H is the Hubble parameter and ρ is the energy density of the Universe. As a consequence, a bouncing cosmology is obtained when this curve is closed and crosses the axis H =0 at least twice, and whose simplest particular example is the ellipse depicting the well-known holonomy corrected Friedmann equation in loop quantum cosmology (LQC). Sometimes, a crucial point in such theories is the appearance of the Ostrogradski instability at the perturbative level; however, fortunately enough, in the present work, as long as the linear level of perturbations is concerned, this instability does not appear, although it may appear at the higher order of perturbations. 5. Laparotomy operative note template constructed through a modified Delphi method. Science.gov (United States) Moore, Lolonya; Churley-Strom, Ruth; Singal, Bonita; O'Leary, Sharon 2009-05-01 An operative note is indispensable to physician documentation and decision-making; however, there are no accepted standards for operative note content. Our aim was to use a modified Delphi consensus-building method to construct a uniform operative note template for laparotomy. Using Joint Commission on Accreditation of Healthcare Organizations requirements, literature review, and feedback from 15 medical malpractice defense attorneys, we compiled a draft operative note template of 31 elements. We surveyed 37 Association of Professor of Gynecology and Obstetrics/Solvay scholars asking for their input on inclusion of each item as essential content of the operative note. Two iterations of the survey were required to reach a predetermined 75% level of consensus. Nine elements were eliminated from the template: 6 original and 3 expert-suggested elements. We provide an operative note template that was compiled through a Delphi process. 6. Could the description on polynuclear superhalogens by DFT be comparable with high-level ab initio results? A comparison between DFT and CCSD(T) International Nuclear Information System (INIS) Sun, Yin-Yin; Li, Jin-Feng; Li, Miao-Miao; Zhou, Fu-Qiang; Li, Jian-Li; Yin, Bing 2016-01-01 A systematic density functional theory study including 17 exchange-correlation functionals was performed on different types of superhalogens with high level coupled-cluster single double including perturbative triple excitations (CCSD(T)) results as the reference. The superhalogens selected here cover the ranges from mononuclear to polynuclear structures and from structures with halogen-atom ligands to those with non-halogen ligands, e.g., [MgX 3 ] − , [Mg 2 X 5 ] − , and [Mg 3 X 7 ] − (X = F, Cl, CN). It is clearly indicated that three double-hybrid functionals B2T-PLYP, B2GP-PLYP, B2K-PLYP as well as the range-separated hybrid functional ωB97X are capable of providing results which approach the accuracy at the CCSD(T) level. The basis set effect is usually moderate and, in most cases, it is enough to utilize the basis set of triple-ξ quality, e.g., Def2-TZVP. In addition, the results of the HF and MP2 method are also acceptable here, especially for polynuclear superhalogens where CCSD(T) is probably unpractical International Nuclear Information System (INIS) Welch, R.W. 1981-06-01 The policy of the Ontario government is to ensure that consumers have access to a full range of competitively-priced energy forms. Nuclear power is more desirable than coal, both environmentally and from the point of view of cost, and the nuclear industry is also important to Ontario, providing jobs and contributing to the province's industrial development. The government of Ontario is exploring the possibility of exporting surplus electrical power when a profit can be made. Ontario Hydro has been exporting electricity for about 60 years, and sales to the United States have increased significantly since 1973, allowing rates to Ontario consumers to be lowered. Exports at present are from fossil-fired plants, increasing nuclear capacity frees the more expensive coal-fuelled capacity for export purposes. In past generating plants have not been built exclusively for the export market, and only short-term power surplusses have been exported. Over the next ten years exports are likely to remain about 10 billion kilowatt-hours per year. It would be environmentally advantageous for Ontario to build nuclear plants to generate electricity for the American market, rather than have to import acid rain from US coal-fired power plants. Much depends on public opinion and government policies at the federal level in both countries 8. Olkiluoto site description 2006 International Nuclear Information System (INIS) Andersson, J.; Ahokas, H.; Hudson, J.A. 2007-03-01 subjective criteria are used for identifying faults; Regarding the rock mechanics description, presented in Chapter 5, additional data have been assessed and the overall level of rock mechanics understanding has increased compared with previous model versions; A new hydrogeological model has been developed, which is presented in Chapter 6, which places more emphasis on the hydraulic data. Pressure anomalies are now stronger drivers for including hydraulic zones in the model and the number of explicit uncertain cases is greater; The hydrogeochemical description, presented in Chapter 7, is consistent with the previous model and there is an increased level of internal consistency in the hydrogeochemical understanding. Information on gases and understanding of their origin has increased significantly. Based on these descriptions, the report also makes a second set of predictions in Chapter 9 concerning the expected geology and rock mechanics properties to be found during the excavation of the ONKALO and also predicts the hydrogeological and hydrogeochemical impacts of these excavations. Two types of predictions are made: type A predictions, that use only the latest version of the overall Site Model; and type B predictions, that also use all the data from the tunnel, which are derived from activities, such as tunnel mapping and Pilot holes. The Site Descriptive Modelling involves uncertainties and it is necessary to assess the confidence in such modelling. This has been assessed through special protocols in a technical auditing exercise, which is presented in Chapter 10. These protocols investigate whether all data have been considered and understood; where the uncertainties lie and what the potential is for alternative interpretations; whether there is sufficient consistency between disciplines and consistency with the past evolution of the site; as well as comparisons with previous model versions. Chapter 11 concludes that, overall, the uncertainty and confidence assessment 9. Note on the ABC Conjecture OpenAIRE Carella, N. A. 2006-01-01 This note imparts heuristic arguments and theorectical evidences that contradict the abc conjecture over the rational numbers. In addition, the rudimentary datails for transforming this problem into the doimain of equidistribution theory are provided. 10. Alcune Note di Analisi Matematica OpenAIRE Vasselli, Ezio 2011-01-01 Lectures notes (in italian) of some arguments of classical analysis, with exercises. A particular emphasis to functional analysis and elementary operator algebra theory is given, by means of exercises and examples. 11. SAFETY INSTRUCTION AND SAFETY NOTE CERN Multimedia TIS Secretariat 2002-01-01 Please note that the SAFETY INSTRUCTION N0 49 (IS 49) and the SAFETY NOTE N0 28 (NS 28) entitled respectively 'AVOIDING CHEMICAL POLLUTION OF WATER' and 'CERN EXHIBITIONS - FIRE PRECAUTIONS' are available on the web at the following urls: http://edms.cern.ch/document/335814 and http://edms.cern.ch/document/335861 Paper copies can also be obtained from the TIS Divisional Secretariat, email: [email protected] 12. Editor’s Note Directory of Open Access Journals (Sweden) 2009-04-01 Full Text Available In rehabilitation, we don't often think about globalization or how what we do may have impact worldwide. Many of us don't have the time, inclination, yet maybe we need to think beyond our borders to pick up a global concept. Apparently the era of globalization is having negative implications for the provision of effective psychiatric rehabilitation programs. The question is whether the impact of globalization can be eliminated or even "neutralized" for the benefit of the patient and the family. The answer is positive. There are solutions like the empowerment of the international movement of families and users of services, with parallel efforts for self-actualization and self-determination of any suffering person. At the community level, the atmosphere could be changed by the systematic implementation of various mental health interventions directed at the modification of the local community's beliefs and attitudes towards the interactions of the mentally ill. We hope the authors discuss these issues in our journal. Regarding Iranian Rehabilitation Journal, beginning with this issue, Volume 7, Number 9, two issues will be published each year, one for spring and one for autumn. In this issue, we begin by exploring in more detail the concept of rehabilitation, then use a historical perspective to examine the changing nature and support for rehabilitation as a correctional goal over time. Our attention next turns to the current treatment programs that are found within the correctional system. Perhaps the most important consideration is whether rehabilitation "works" to reduce the likelihood that offenders will "recidivate". Accordingly, we also review the latest research on the effectiveness of treatment interventions. We conclude this entry with comments on the future of rehabilitation as a correctional goal. 13. Political Investment in Economic Protection: A Note OpenAIRE Arsenio M. Balisacan 1987-01-01 This note presents some aspects of the neoclassical political economy of economic protection. It develops a conceptual framework which focuses on the costs and benefits of investment in political in?uence to certain conditions in the society and how the benefit-cost structure may affect the level of economic protection and the amount of what Bhagwati called "directly unproductive profit-seeking" activities. 14. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 6. Characterization and description of areas. Sjaelland International Nuclear Information System (INIS) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-01-01 The low and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities, strong sorption capacity for many radionuclides and self-sealing properties. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks. Nevertheless, the Tertiary clays were mapped as well. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas potentially useful for a waste disposal. The 20 areas have to be reduced to 1-3 most potential locations where detailed field investigations of the geological, hydrogeological - hydrochemical and geotechnical conditions will be performed. The present report describes the areas 5 and 6 on Zealand. (LN) 15. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 7. Characterization and description of areas. Langeland, Taesinge and Fyn International Nuclear Information System (INIS) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-01-01 The low and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities, high sorption capacity for many radionuclides and self-sealing properties. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks. Nevertheless, the Tertiary clays were mapped as well. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas where a waste disposal potentially can be located. The 20 areas have to be reduced to 1-3 most potential locations where detailed field investigations of the geological, hydrogeological - hydrochemical and geotechnical conditions will be performed. The present report describes the areas 7,8,9,10, and 11 on the islands Langeland, Taasinge and Funen. (LN) 16. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 8. Characterization and description of areas. Oestjylland International Nuclear Information System (INIS) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-01-01 The low and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities, high sorption capacity for many radionuclides and self-sealing properties. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks. Nevertheless, the Tertiary clays were mapped as well. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas where a waste disposal potentially can be located. The 20 areas have to be reduced to 1-3 most potential locations where detailed field investigations of the geological, hydrogeological - hydrochemical and geotechnical conditions will be performed. The present report describes the areas 12,13,14 and 15 in Eastern Jutland. (LN) 17. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 11. Description of areas. Danish and English summary International Nuclear Information System (INIS) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-01-01 The low - and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by choosing deposits with low water flow and high sorption potential of the sediments or rocks. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks but the Tertiary clays were also mapped. The salt diapirs, salt pillows and salt deposits and deep basement rocks are not included in the present study. These rocks and deposits are situated too deep for the present study and salt deposits seem to be unstable for a disposal (e.g. German salt mines). The regional geologic survey based on existing data was concluded by selecting 22 areas in Denmark. There remains now to reduce the number of potential areas to 1-3 where detailed field studies will be performed in order to select the final location. (LN) 18. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 4. Characterization and description of areas. Bornholm International Nuclear Information System (INIS) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-01-01 The low - and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities and high sorption potentials of the sediments or rocks. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier been focused on deep seated salt deposits and basement rocks, but the Tertiary clays were also mapped. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas where a waste disposal potentially can be located. The 20 areas have to be reduced to 2-3 more precise locations, where detailed field investigations of the geological, hydrogeological-hydrochemical and technical conditions will be performed. The present report describes areas 1 and 2 on Bornholm, East Denmark. (LN) 19. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 9. Characterization and description of areas. Limfjorden International Nuclear Information System (INIS) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-01-01 The low and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities, strong sorption capacity for many radionuclides and self-sealing properties. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks. Nevertheless, the Tertiary clays were mapped as well. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas potentially useful for a waste disposal. The 20 areas have to be reduced to 1-3 most potential locations where detailed field investigations of the geological, hydrogeological, hydrochemical and geotechnical conditions will be performed. The present report describes the areas 16,17,18,19,20 and 21 around Limfjorden. (LN) 20. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 5. Characterization and description of areas. Falster and Lolland International Nuclear Information System (INIS) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-01-01 The low and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities, strong sorption capacity for many radionuclides and self-sealing properties. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks. Nevertheless, the Tertiary clays were mapped as well. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas potentially useful for a waste disposal. The 20 areas have to be reduced to 1-3 most potential locations where detailed field investigations of the geological, hydrogeological, hydrochemical and geotechnical conditions will be performed. The present report describes areas 3 and 4 on Falster and Lolland. (LN) 1. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 4. Characterization and description of areas. Bornholm Energy Technology Data Exchange (ETDEWEB) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-07-01 The low - and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities and high sorption potentials of the sediments or rocks. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier been focused on deep seated salt deposits and basement rocks, but the Tertiary clays were also mapped. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas where a waste disposal potentially can be located. The 20 areas have to be reduced to 2-3 more precise locations, where detailed field investigations of the geological, hydrogeological-hydrochemical and technical conditions will be performed. The present report describes areas 1 and 2 on Bornholm, East Denmark. (LN) 2. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 10. Characterization and description of areas. Nordjylland Energy Technology Data Exchange (ETDEWEB) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-07-01 The low and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities, strong sorption capacity for many radionuclides and self-sealing properties. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks. Nevertheless, the Tertiary clays were mapped as well. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas potentially useful for a waste disposal. The 20 areas have to be reduced to 1-3 most potential locations where detailed field investigations of the geological, hydrogeological, hydrochemical and geotechnical conditions will be performed. The present report describes the area 22 in Northern Jutland. (LN) 3. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 8. Characterization and description of areas. OEstjylland Energy Technology Data Exchange (ETDEWEB) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-07-01 The low and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities, high sorption capacity for many radionuclides and self-sealing properties. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks. Nevertheless, the Tertiary clays were mapped as well. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas where a waste disposal potentially can be located. The 20 areas have to be reduced to 1-3 most potential locations where detailed field investigations of the geological, hydrogeological - hydrochemical and geotechnical conditions will be performed. The present report describes the areas 12,13,14 and 15 in Eastern Jutland. (LN) 4. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 7. Characterization and description of areas. Langeland, Taasinge and Fyn Energy Technology Data Exchange (ETDEWEB) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-07-01 The low and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities, high sorption capacity for many radionuclides and self-sealing properties. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks. Nevertheless, the Tertiary clays were mapped as well. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas where a waste disposal potentially can be located. The 20 areas have to be reduced to 1-3 most potential locations where detailed field investigations of the geological, hydrogeological - hydrochemical and geotechnical conditions will be performed. The present report describes the areas 7,8,9,10, and 11 on the islands Langeland, Taasinge and Funen. (LN) 5. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 9. Characterization and description of areas. Limfjorden Energy Technology Data Exchange (ETDEWEB) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-07-01 The low and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities, strong sorption capacity for many radionuclides and self-sealing properties. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks. Nevertheless, the Tertiary clays were mapped as well. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas potentially useful for a waste disposal. The 20 areas have to be reduced to 1-3 most potential locations where detailed field investigations of the geological, hydrogeological, hydrochemical and geotechnical conditions will be performed. The present report describes the areas 16,17,18,19,20 and 21 around Limfjorden. (LN) 6. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 5. Characterization and description of areas. Falster and Lolland Energy Technology Data Exchange (ETDEWEB) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-07-01 The low and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities, strong sorption capacity for many radionuclides and self-sealing properties. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks. Nevertheless, the Tertiary clays were mapped as well. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas potentially useful for a waste disposal. The 20 areas have to be reduced to 1-3 most potential locations where detailed field investigations of the geological, hydrogeological, hydrochemical and geotechnical conditions will be performed. The present report describes areas 3 and 4 on Falster and Lolland. (LN) 7. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 6. Characterization and description of areas. Sjaelland Energy Technology Data Exchange (ETDEWEB) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-07-01 The low and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by low water flow possibilities, strong sorption capacity for many radionuclides and self-sealing properties. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks. Nevertheless, the Tertiary clays were mapped as well. The salt diapirs and the salt deposits are not included in the present study. The task is to find approximately 20 areas potentially useful for a waste disposal. The 20 areas have to be reduced to 1-3 most potential locations where detailed field investigations of the geological, hydrogeological - hydrochemical and geotechnical conditions will be performed. The present report describes the areas 5 and 6 on Zealand. (LN) 8. FleCSPH notes Energy Technology Data Exchange (ETDEWEB) Lim, Hyun [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Loiseau, Julien [Los Alamos National Lab. (LANL), Los Alamos, NM (United States) 2017-07-17 good level of performance. Then the driver represent the general execution of the resolution without worrying of the data locality and communications. As FleCSI is an On-Development code the structure may change in the future and we keep track of these changes in FleCSPH. 9. Note DEFF Research Database (Denmark) Kullman, Mikael; Campillo, Javier; Dahlquist, Erik 2016-01-01 Globally, more than 50% of all people are living in cities today. Enhancing sustainability and efficiency of urban energy systems is thus of high priority for global sustainable development. The European research project PLEEC (Planning for Energy Efficient Cities) focuses on technological... 10. Technical note on drainage systems DEFF Research Database (Denmark) Bentzen, Thomas Ruby This technical note will present simple but widely used methods for the design of drainage systems. The note will primarily deal with surface water (rainwater) which on a satisfactorily way should be transport into the drainage system. Traditional two types of sewer systems exist: A combined system......, where rainwater and sewage is transported in the same pipe, and a separate system where the two types of water are transported in individual pipe. This note will only focus on the separate rain/stormwater system, however, if domestic sewage should be included in the dimensioning procedure, it......’s not major different than described below - just remember to include this contribution for combined systems where the surface water (rain) and sewage are carried in the same pipes in the system and change some of the parameters for failure allowance (this will be elaborated further later on). The technical... 11. Gaz de France. Operation note International Nuclear Information System (INIS) 2003-01-01 This note was published for the public at the occasion of the admission to Euronext's Eurolist of the existing shares that make the capital of Gaz de France company, the French gas utility. The note gives some informations about Gaz de France activity, and about its strategy of development in the European gas market. Then it describes the offer relative to the opening of Gaz de France capital. Some selected financial data and some precision about the risk factors and the management of the company complete the document. (J.S.) 12. Guidelines for Description NARCIS (Netherlands) Links, P.; Horsman, Peter; Kühnel, Karsten; Priddy, M.; Reijnhoudt, Linda; Merenmies, Mark 2013-01-01 The Guidelines follow the conceptual metadata model (deliverable 17.2). They include guidelines for description of collection-holding institutions, document collections, organisations, personalities, events, camps and ghettos. As much as possible the guidelines comply with the descriptive standards 13. Analysing playing using the note-time playing path. Science.gov (United States) de Graaff, Deborah L E; Schubert, Emery 2011-03-01 This article introduces a new method of data analysis that represents the playing of written music as a graph. The method, inspired by Miklaszewski, charts low-level note timings from a sound recording of a single-line instrument using high-precision audio-to-MIDI conversion software. Note onset times of pitch sequences are then plotted against the score-predicted timings to produce a Note-Time Playing Path (NTPP). The score-predicted onset time of each sequentially performed note (horizontal axis) unfolds in performed time down the page (vertical axis). NTPPs provide a visualisation that shows (1) tempo variations, (2) repetitive practice behaviours, (3) segmenting of material, (4) precise note time positions, and (5) time spent on playing or not playing. The NTPP can provide significant new insights into behaviour and cognition of music performance and may also be used to complement established traditional approaches such as think-alouds, interviews, and video coding. 14. Churchill on Stalin: A note. NARCIS (Netherlands) Ellman, M.J. 2006-01-01 The purpose of this note is to draw attention to two limitations of Churchill's war memoirs as a source of accurate information about Stalin, his views ans actions. they concern, first Stalin's alleged remarks in 1942 about collectivisation, and second Stalin's allleged response to Churchill's 15. Lecture notes on ideal magnetohydrodynamics International Nuclear Information System (INIS) Goedbloed, J.P. 1983-03-01 Notes, prepared for a course of lectures held at the Instituto de Fisica, Universidade Estadual de Campinas, Brazil (June-August 1978). An extensive theoretical treatment of the behaviour of hot plasmas caught in equations and mathematical models is presented in 12 chapters 16. Developing INDCs: a guidance note DEFF Research Database (Denmark) Bakkegaard, Riyong Kim; Bee, Skylar; Naswa, Prakriti needs and low capability, would need means of implementation (MoI) for adaptation and to take ambitious mitigation actions. Developing countries would include MoI needs in the context of mitigation and adaptation. The note explains briefly how countries can identify their unconditional contributions.... 17. A note on hypoplastic yielding OpenAIRE 2010-01-01 This note discusses briefly the definition of yield surface in hypoplasticity in connection with the physical notion of yielding. The relation of yielding with the vanishing of the material time derivative of the stress tensor and the vanishing of the corotational stress rate is investigated. 18. NOAA Weather Radio - EAS Description Science.gov (United States) Non-Zero All Hazards Logo Emergency Alert Description Event Codes Fact Sheet FAQ Organization Search Search For Go NWS All NOAA Emergency Alert System (EAS) List of EAS Event Codes NWS EAS fact sheet What Management Agency (FEMA) and the NWS, implements the EAS at the federal level. The EAS is the nation's public 19. Optimization for Guitar Fingering on Single Notes Science.gov (United States) Itoh, Masaru; Hayashida, Takumi This paper presents an optimization method for guitar fingering. The fingering is to determine a unique combination of string, fret and finger corresponding to the note. The method aims to generate the best fingering pattern for guitar robots rather than beginners. Furthermore, it can be applied to any musical score on single notes. A fingering action can be decomposed into three motions, that is, a motion of press string, release string and move fretting hand. The cost for moving the hand is estimated on the basis of Manhattan distance which is the sum of distances along fret and string directions. The objective is to minimize the total fingering costs, subject to fret, string and finger constraints. As a sequence of notes on the score forms a line on time series, the optimization for guitar fingering can be resolved into a multistage decision problem. Dynamic programming is exceedingly effective to solve such a problem. A level concept is introduced into rendering states so as to make multiple DP solutions lead a unique one among the DP backward processes. For example, if two fingerings have the same value of cost at different states on a stage, then the low position would be taken precedence over the high position, and the index finger would be over the middle finger. 20. LLW Notes, vol. 9, no. 1. February/March 1994 International Nuclear Information System (INIS) 1994-03-01 LLW Notes is published ten times each year and is distributed to Low- Level Radioactive Waste Forum Participants and other state and compact officials identified by those Participants to receive LLW Notes. The Low-Level Radioactive Waste Forum is an association of representatives of states and compacts established to facilitate state and compact commission implementation of the Low-Level Radioactive Waste Policy Amendments Act of 1985 and to promote the objectives of low-level radioactive waste regional compacts. The Forum provides an opportunity for states and compacts to share information with one another and to exchange views with officials of federal agencies 1. Studies on Neotropical Phasmatodea XVI: Revision of Haplopodini Günther, 1953 (rev. stat.), with notes on the subfamily Cladomorphinae Bradley & Galil, 1977 and the descriptions of a new tribe, four new genera and nine new species (Phasmatodea: "Anareolatae": Phasmatidae: Cladomorphinae). Science.gov (United States) Hennemann, Frank H; Conle, Oskar V; Perez-Gelabert, Daniel E 2016-06-27 eggs. Haplopodini Günther, 1953 is re-established (rev. stat.) and comprises almost exclusively Caribbean genera previously placed in Hesperophasmatini by Bradley & Galil (1977) or Cranidiini by Zompro, (2004). Aploploides Rehn & Hebard, 1938, Diapherodes Gray, 1835, Haplopus Burmeister, 1838 and Paracranidium Brock, 1998 were misplaced in Cranidiini and are transferred to Haplopodini. On the basis of numerous morphological characters of the insects and eggs Hesperophasmatini is removed from Pseudophasmatidae: Xerosomatinae and re-transferred to its previous position in the subfamily Cladomorphinae sensu lato. A detailed newdiagnosis of Hesperophasmatini is presented, but is only provisional since the true diversity is as yet only fractionally known. The lack of a gula distinguishes Hesperophasmatini from all other tribes. The genus Laciphorus Redtenbacher, 1908 is removed from Hesperophasmatini and transferred to Diapheromeridae: Diapheromerinae: Diapheromerini. The new tribe Pterinoxylini n. trib. is established to contain only the type-genus Pterinoxylus Audinet-Serville, 1838. It is closely related and perhaps the sister taxon of Hesperophasmatini, with which it shares the presence of rough sensory areas on the probasisternum and profurcasternum. It differs from Hesperophasmatini and Haplopodini by the presence of a tympanal region (= stridulatory organ) in the alae of females and the alveolar eggs, which possess peripheral opercular and polar structures. Haplopodini is likely to be the sister group of Pterinoxylini n. trib. + Hesperophasmatini.        The tribe Haplopodini rev. stat. is revised at the species level and comprises eight almost exclusively Caribbean genera, four of which are newly described. All eight genera now contained in Haplopodini are described in detail, differentiated from their closest relatives and their relationships and systematic position within Haplopodini are discussed. Keys and maps showing their distributions are presented 2. A note on simultaneous Monte Carlo tests DEFF Research Database (Denmark) Hahn, Ute In this short note, Monte Carlo tests of goodness of fit for data of the form X(t), t ∈ I are considered, that reject the null hypothesis if X(t) leaves an acceptance region bounded by an upper and lower curve for some t in I. A construction of the acceptance region is proposed that complies to a...... to a given target level of rejection, and yields exact p-values. The construction is based on pointwise quantiles, estimated from simulated realizations of X(t) under the null hypothesis.... 3. YUCCA MOUNTAIN SITE DESCRIPTION International Nuclear Information System (INIS) Simmons, A.M. 2004-01-01 The ''Yucca Mountain Site Description'' summarizes, in a single document, the current state of knowledge and understanding of the natural system at Yucca Mountain. It describes the geology; geochemistry; past, present, and projected future climate; regional hydrologic system; and flow and transport within the unsaturated and saturated zones at the site. In addition, it discusses factors affecting radionuclide transport, the effect of thermal loading on the natural system, and tectonic hazards. The ''Yucca Mountain Site Description'' is broad in nature. It summarizes investigations carried out as part of the Yucca Mountain Project since 1988, but it also includes work done at the site in earlier years, as well as studies performed by others. The document has been prepared under the Office of Civilian Radioactive Waste Management quality assurance program for the Yucca Mountain Project. Yucca Mountain is located in Nye County in southern Nevada. The site lies in the north-central part of the Basin and Range physiographic province, within the northernmost subprovince commonly referred to as the Great Basin. The basin and range physiography reflects the extensional tectonic regime that has affected the region during the middle and late Cenozoic Era. Yucca Mountain was initially selected for characterization, in part, because of its thick unsaturated zone, its arid to semiarid climate, and the existence of a rock type that would support excavation of stable openings. In 1987, the United States Congress directed that Yucca Mountain be the only site characterized to evaluate its suitability for development of a geologic repository for high-level radioactive waste and spent nuclear fuel 4. YUCCA MOUNTAIN SITE DESCRIPTION Energy Technology Data Exchange (ETDEWEB) A.M. Simmons 2004-04-16 The ''Yucca Mountain Site Description'' summarizes, in a single document, the current state of knowledge and understanding of the natural system at Yucca Mountain. It describes the geology; geochemistry; past, present, and projected future climate; regional hydrologic system; and flow and transport within the unsaturated and saturated zones at the site. In addition, it discusses factors affecting radionuclide transport, the effect of thermal loading on the natural system, and tectonic hazards. The ''Yucca Mountain Site Description'' is broad in nature. It summarizes investigations carried out as part of the Yucca Mountain Project since 1988, but it also includes work done at the site in earlier years, as well as studies performed by others. The document has been prepared under the Office of Civilian Radioactive Waste Management quality assurance program for the Yucca Mountain Project. Yucca Mountain is located in Nye County in southern Nevada. The site lies in the north-central part of the Basin and Range physiographic province, within the northernmost subprovince commonly referred to as the Great Basin. The basin and range physiography reflects the extensional tectonic regime that has affected the region during the middle and late Cenozoic Era. Yucca Mountain was initially selected for characterization, in part, because of its thick unsaturated zone, its arid to semiarid climate, and the existence of a rock type that would support excavation of stable openings. In 1987, the United States Congress directed that Yucca Mountain be the only site characterized to evaluate its suitability for development of a geologic repository for high-level radioactive waste and spent nuclear fuel. 5. What Limits the Encoding Effect of Note-Taking? A Meta-Analytic Examination Science.gov (United States) Kobayashi, K. 2005-01-01 Previous meta-analyses indicate that the overall encoding effect of note-taking is positive but modest. This meta-analysis of 57 note-taking versus no note-taking comparison studies explored what limits the encoding effect by examining the moderating influence of seven variables: intervention, schooling level, presentation mode and length, test… 6. Fotonovela. Technical Note No. 13. Science.gov (United States) Barriga, Patricio; Villacis, Rodrigo This report contains a description of a program that sought to develop the fotonovela as an instrument for increasing literacy and consciousness awareness in Ecuador. Following a general introduction to the work of the Nonformal Education Project, the creator of this program, the report presents information about the antecedents and… 7. A Guide to Field Notes for Qualitative Research: Context and Conversation. Science.gov (United States) Phillippi, Julia; Lauderdale, Jana 2018-02-01 Field notes are widely recommended in qualitative research as a means of documenting needed contextual information. With growing use of data sharing, secondary analysis, and metasynthesis, field notes ensure rich context persists beyond the original research team. However, while widely regarded as essential, there is not a guide to field note collection within the literature to guide researchers. Using the qualitative literature and previous research experience, we provide a concise guide to collection, incorporation, and dissemination of field notes. We provide a description of field note content for contextualization of an entire study as well as individual interviews and focus groups. In addition, we provide two "sketch note" guides, one for study context and one for individual interviews or focus groups for use in the field. Our guides are congruent with many qualitative and mixed methodologies and ensure contextual information is collected, stored, and disseminated as an essential component of ethical, rigorous qualitative research. 8. Evalutive Descriptions of Art Directory of Open Access Journals (Sweden) Nataša Lah 2016-01-01 Full Text Available Taking into account the fact that, throughout history, certain artworks have been considered as “worth of watching” (according to the Greek etymon ἀξιοϑέατος / aksioteatos, preservation, or theorizing, while others were not, one is led to investigate the various types of evaluative descriptions. Those artworks that are more valuable than others, or simply valuable in themselves on the basis of rather specific features, have always represented the paradigmatic model for the evaluator, thus revealing the identitary nature of value as different from one epoch to another. Our aim has been to discern, with regard to this starting point, the way in which the process of evaluating artworks fits the general matrix of the universal theory of value, with its clearly distinguished levels of evaluation, beginning with value descriptions, continuing through the features of evaluation or abstract qualities of values extracted from these descriptions, and ending with value norms or systems of accepted generalizations in evaluation. Value standpoints in such an evaluation matrix represent dispositions or preferences in procedures, which reflect the norms or signifying concepts of the time. Corresponding procedures, or applications of the hierarchicized signification of artworks, are manifested in all known forms of artwork assessment: attribution, institutionalization, and setting of priorities in terms of exhibition, conservation, acquisition, restoration, and so on. Research in the history of European art-historical ideas has corroborated the hypothesis that, prior to the late 18th century, clear normative patterns were applied when it came to the evaluation of artworks. However, with the emergence of early Romanticism, this could no longer be done in the traditional way. Before the period in question, visual art was created (regardless of some stylistic discrepancies between individual authors and classified according to well-defined thematic 9. Note Taking Activity and its assessment in a Blended Learning Environment OpenAIRE Nakayama, Minoru; Mutsuura, Kouichi; Yamamoto, Hiroh 2017-01-01 "Note-taking" is a popular skill for all types of learning activities. In recent years, the online educational environment has began spreading rapidly at institutes of higher educational, obviating the need for printed materials or written notes. This means that students’ ability to take notes may decline and this may influence the success of their learning.  In order to examine this phenomenon, students' notes were surveyed during a blended learning course in a bachelor level program at... 10. Descrição do girino de Lysapsus laevis (Parker, com notas sobre o ambiente, hábitos e desenvolvimento (Anura, Hylidae, Pseudinae Description of the tadpole of Lysapsus laevis (Parker, with notes on the habitat, habits, and development (Anura, Hylidae, Pseudinae Directory of Open Access Journals (Sweden) Ulisses Caramaschi 2004-09-01 11. Revising Lecture Notes: How Revision, Pauses, and Partners Affect Note Taking and Achievement Science.gov (United States) Luo, Linlin; Kiewra, Kenneth A.; Samuelson, Lydia 2016-01-01 Note taking has been categorized as a two-stage process: the recording of notes and the review of notes. We contend that note taking might best involve a three-stage process where the missing stage is revision. This study investigated the benefits of revising lecture notes and addressed two questions: First, is revision more effective than… 12. Descriptive set theory CERN Document Server Moschovakis, YN 1987-01-01 Now available in paperback, this monograph is a self-contained exposition of the main results and methods of descriptive set theory. It develops all the necessary background material from logic and recursion theory, and treats both classical descriptive set theory and the effective theory developed by logicians. 13. Description logics of context CSIR Research Space (South Africa) Klarman, S 2013-05-01 Full Text Available We introduce Description Logics of Context (DLCs) - an extension of Description Logics (DLs) for context-based reasoning. Our approach descends from J. McCarthy's tradition of treating contexts as formal objects over which one can quantify... 14. Physics 3204. Course Description. Science.gov (United States) Newfoundland and Labrador Dept. of Education. A description of the physics 3204 course in Newfoundland and Labrador is provided. The description includes: (1) statement of purpose, including general objectives of science education; (2) a list of six course objectives; (3) course content for units on sound, light, optical instruments, electrostatics, current electricity, Michael Faraday and… 15. Lecture notes for criticality safety International Nuclear Information System (INIS) Fullwood, R. 1992-03-01 These lecture notes for criticality safety are prepared for the training of Department of Energy supervisory, project management, and administrative staff. Technical training and basic mathematics are assumed. The notes are designed for a two-day course, taught by two lecturers. Video tapes may be used at the options of the instructors. The notes provide all the materials that are necessary but outside reading will assist in the fullest understanding. The course begins with a nuclear physics overview. The reader is led from the macroscopic world into the microscopic world of atoms and the elementary particles that constitute atoms. The particles, their masses and sizes and properties associated with radioactive decay and fission are introduced along with Einstein's mass-energy equivalence. Radioactive decay, nuclear reactions, radiation penetration, shielding and health-effects are discussed to understand protection in case of a criticality accident. Fission, the fission products, particles and energy released are presented to appreciate the dangers of criticality. Nuclear cross sections are introduced to understand the effectiveness of slow neutrons to produce fission. Chain reactors are presented as an economy; effective use of the neutrons from fission leads to more fission resulting in a power reactor or a criticality excursion. The six-factor formula is presented for managing the neutron budget. This leads to concepts of material and geometric buckling which are used in simple calculations to assure safety from criticality. Experimental measurements and computer code calculations of criticality are discussed. To emphasize the reality, historical criticality accidents are presented in a table with major ones discussed to provide lessons-learned. Finally, standards, NRC guides and regulations, and DOE orders relating to criticality protection are presented 16. Notes on elementary particle physics CERN Document Server 1972-01-01 Notes of Elementary Particle Physics is a seven-chapter text that conveys the ideas on the state of elementary particle physics. This book emerged from an introductory course of 30 lectures on the subject given to first-year graduate students at the University of Liverpool. The opening chapter deals with pertinent terminologies in elementary particle physics. The succeeding three chapters cover the concepts of transition amplitudes, probabilities, relativistic wave equations and fields, and the interaction amplitude. The discussion then shifts to tests of electromagnetic interactions, particul 17. Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 11. Description of areas. Danish and English summary; Low- and intermediate level radioactive waste from Risoe, Denmark. Location studies for potential disposal areas. Report no. 11. Omraadebeskrivelser - Description of areas. Dansk og engelsk resume Energy Technology Data Exchange (ETDEWEB) Gravesen, P.; Nilsson, B.; Schack Pedersen, S.A.; Binderup, M. 2011-07-01 The low - and intermediate level radioactive waste from Risoe: the nuclear reactor buildings, different types of material from the research periods and waste from hospitals and research institutes have to be stored in a final disposal in Denmark for at least 300 years. The task is to locate and recognize sediments or rocks with low permeability which can isolate the radioactive waste from the surrounding deposits, the groundwater resources, the recipients and from human activities. The sediments or rocks shall also act as a protection if the waste disposal leaks radioactive material to the surroundings. This goal can be reached by choosing deposits with low water flow and high sorption potential of the sediments or rocks. The investigation of geological deposits as potential waste disposals for high radioactive waste from nuclear power plants has earlier focused on deep seated salt deposits and basement rocks but the Tertiary clays were also mapped. The salt diapirs, salt pillows and salt deposits and deep basement rocks are not included in the present study. These rocks and deposits are situated too deep for the present study and salt deposits seem to be unstable for a disposal (e.g. German salt mines). The regional geologic survey based on existing data was concluded by selecting 22 areas in Denmark. There remains now to reduce the number of potential areas to 1-3 where detailed field studies will be performed in order to select the final location. (LN) 18. Debriefing Note Secondary Education Support Programme DEFF Research Database (Denmark) Webster, Neil; Vagnby, Bo Hellisen; Thomsen, Thomas J. Debriefing note regarding joint programming for the Secondary Education Support Programme (2003- 2007). The note specifies preparation of SIP Physical Guidelines; Training needs assessment for Physical School Status and Rapid Technical Assessments; SIP/DEP preparation; Selection criteria... 19. Note-Making in Social Studies. Science.gov (United States) Fowler, Robert W. 1985-01-01 Note-making is one excellent method for helping students retain important points made by the teacher. Techniques that elementary and secondary social studies teacher can use to teach note-making skills are described. (RM) 20. Taking notes as an interactive process OpenAIRE Hornig, Wolfgang 1984-01-01 Taking notes as an interactive process : how to improve students´ notes / Hornig W. ; Nowak, J. - In: Nowak, Johann: Textverstehen und Textrekonstruktion in Vorlesungen. - Augsburg : HDZ, 1984. - S. 227-253. - (Augsburger Studien zur Hochschuldidaktik ; 12) 1. The new bank note distribution system OpenAIRE Gerrit Bilkes 1997-01-01 In this article, the author outlines the recent changes made to the way Canada's bank notes are distributed. The new system allows financial institutions to exchange notes directly with one another at designated points across the country, rather than through Bank of Canada agencies, as was previously the case. The institutions communicate with the Bank of Canada through a computerized inventory-management system. Two Bank of Canada operations centres monitor note quality and supply new notes ... Science.gov (United States) Vodicka, Elisabeth; Mejilla, Roanne; Leveille, Suzanne G; Ralston, James D; Darer, Jonathan D; Delbanco, Tom; Walker, Jan; Elmore, Joann G 2013-09-26 3. What Predicts Skill in Lecture Note Taking? Science.gov (United States) Peverly, Stephen T.; Ramaswamy, Vivek; Brown, Cindy; Sumowski, James; Alidoost, Moona; Garner, Joanna 2007-01-01 Despite the importance of good lecture notes to test performance, very little is known about the cognitive processes that underlie effective lecture note taking. The primary purpose of the 2 studies reported (a pilot study and Study 1) was to investigate 3 processes hypothesized to be significantly related to quality of notes: transcription… 4. 1995 Baseline solid waste management system description International Nuclear Information System (INIS) Anderson, G.S.; Konynenbelt, H.S. 1995-09-01 This provides a detailed solid waste system description that documents the treatment, storage, and disposal (TSD) strategy for managing Hanford's solid low-level waste, low-level mixed waste, transuranic and transuranic mixed waste, and greater-than-Class III waste. This system description is intended for use by managers of the solid waste program, facility and system planners, as well as system modelers. The system description identifies the TSD facilities that constitute the solid waste system and defines these facilities' interfaces, schedules, and capacities. It also provides the strategy for treating each of the waste streams generated or received by the Hanford Site from generation or receipt through final destination 5. Note on off-shell relations in nonlinear sigma model International Nuclear Information System (INIS) Chen, Gang; Du, Yi-Jian; Li, Shuyi; Liu, Hanqing 2015-01-01 In this note, we investigate relations between tree-level off-shell currents in nonlinear sigma model. Under Cayley parametrization, all odd-point currents vanish. We propose and prove a generalized U(1) identity for even-point currents. The off-shell U(1) identity given in http://dx.doi.org/10.1007/JHEP01(2014)061 is a special case of the generalized identity studied in this note. The on-shell limit of this identity is equivalent with the on-shell KK relation. Thus this relation provides the full off-shell correspondence of tree-level KK relation in nonlinear sigma model. 6. Lecture notes on diophantine analysis CERN Document Server Zannier, Umberto 2014-01-01 These lecture notes originate from a course delivered at the Scuola Normale in Pisa in 2006. Generally speaking, the prerequisites do not go beyond basic mathematical material and are accessible to many undergraduates. The contents mainly concern diophantine problems on affine curves, in practice describing the integer solutions of equations in two variables. This case historically suggested some major ideas for more general problems. Starting with linear and quadratic equations, the important connections with Diophantine Approximation are presented and Thue's celebrated results are proved in full detail. In later chapters more modern issues on heights of algebraic points are dealt with, and applied to a sharp quantitative treatment of the unit equation. The book also contains several Supplements, hinted exercises and an Appendix on recent work on heights. 7. Lecture notes on wavelet transforms CERN Document Server Debnath, Lokenath 2017-01-01 This book provides a systematic exposition of the basic ideas and results of wavelet analysis suitable for mathematicians, scientists, and engineers alike. The primary goal of this text is to show how different types of wavelets can be constructed, illustrate why they are such powerful tools in mathematical analysis, and demonstrate their use in applications. It also develops the required analytical knowledge and skills on the part of the reader, rather than focus on the importance of more abstract formulation with full mathematical rigor.  These notes differs from many textbooks with similar titles in that a major emphasis is placed on the thorough development of the underlying theory before introducing applications and modern topics such as fractional Fourier transforms, windowed canonical transforms, fractional wavelet transforms, fast wavelet transforms, spline wavelets, Daubechies wavelets, harmonic wavelets and non-uniform wavelets. The selection, arrangement, and presentation of the material in these ... 8. Hardware description languages Science.gov (United States) Tucker, Jerry H. 1994-01-01 Hardware description languages are special purpose programming languages. They are primarily used to specify the behavior of digital systems and are rapidly replacing traditional digital system design techniques. This is because they allow the designer to concentrate on how the system should operate rather than on implementation details. Hardware description languages allow a digital system to be described with a wide range of abstraction, and they support top down design techniques. A key feature of any hardware description language environment is its ability to simulate the modeled system. The two most important hardware description languages are Verilog and VHDL. Verilog has been the dominant language for the design of application specific integrated circuits (ASIC's). However, VHDL is rapidly gaining in popularity. 9. Some Notes and Critiques of Selected Lexicographic Texts in Bikol Directory of Open Access Journals (Sweden) Louward Allen Zubiri 2014-12-01 Full Text Available This study revolves around notes and critiques of Bikol lexicography. Based on a selection of six dictionaries, the structure and diachronic development of Bikol lexicography were analyzed. Half of the selected dictionaries were authored in full or part by Malcolm Mintz, a renowned Bikol expert. These enabled a linear analysis of changes in lexicography that have occurred within the span of four decades. The dictionaries studied include the earliest and most influential Bikol dictionary of Lisboa (1865. The study presents a preliminary description of lexicographic work done in Bikol and traces the shift in the paradigm of dictionary making from the Spanish era to the present. 10. Resource description and access 2013 revision CERN Document Server 2013-01-01 This e-book contains the 2013 Revision of RDA: Resource Description and Access, and includes the July 2013 Update. This e-book offers links within the RDA text and the capability of running rudimentary searches of RDA, but please note that this e-book does not have the full range of content or functionality provided by the subscription product RDA Toolkit. Included: A full accumulation of RDA- the revision contains a full set of all current RDA instructions. It replaces the previous version of RDA Print as opposed to being an update packet to that version. RDA has gone through many changes sin 11. Quality of outpatient clinical notes: a stakeholder definition derived through qualitative research Directory of Open Access Journals (Sweden) Hanson Janice L 2012-11-01 content. Conclusions Perspectives of these four stakeholder groups provide a comprehensive description of quality in outpatient clinical documentation. The resulting description of characteristics and content necessary for quality notes provides a research-based foundation for assessing the quality of clinical documentation in outpatient health care settings. 12. Quality of outpatient clinical notes: a stakeholder definition derived through qualitative research. Science.gov (United States) Hanson, Janice L; Stephens, Mark B; Pangaro, Louis N; Gimbel, Ronald W 2012-11-19 groups provide a comprehensive description of quality in outpatient clinical documentation. The resulting description of characteristics and content necessary for quality notes provides a research-based foundation for assessing the quality of clinical documentation in outpatient health care settings. 13. Notes on neutron flux measurement International Nuclear Information System (INIS) Alcala Ruiz, F. 1984-01-01 The main purpose of this work is to get an useful guide to carry out topical neutron flux measurements. Although the foil activation technique is used in the majority of the cases, other techniques, such as those based on fission chambers and self-powered neutron detectors, are also shown. Special interest is given to the description and application of corrections on the measurement of relative and absolute induced activities by several types of detectors (scintillators, G-M and gas proportional counters). The thermal arid epithermal neutron fluxes, as determined in this work, are conventional or effective (West cots fluxes), which are extensively used by the reactor experimentalists; however, we also give some expressions where they are related to the integrated neutron fluxes, which are used in neutron calculations. (Author) 16 refs 14. Profiling Chilean Suicide Note-Writers through Content Analysis Directory of Open Access Journals (Sweden) Francisco Ceballos-Espinoza 2016-09-01 Full Text Available Suicides account for 2000 deaths in Chile each year. With a suicide rate of 11.3, it is classified as a country with high suicide risk. Aims: to identify personality and cognitive characteristics of the group of Chilean suicides that left suicide notes, through a content analysis. Methods: descriptive field study with an ex post facto design. All suicides registered between 2010 and 2012 by the Investigations Police of Chile were analyzed, obtaining 203 suicide notes from 96 cases. The Darbonne categories for content analysis were used with the inter-judge method. Results: The mean age of the suicides was 44.2 (SD = 18.53. Most of the notes were addressed to family members (51.7%. The most expressed reasons were marital- or interpersonal-related (24.6%; another 23.6% expressed a lack of purpose or hopelessness (including depression, wish to die, low self-esteem. The most frequent content expressed were instructions (about money, children, and funeral. All of the notes showed logical thinking and were written with coherence and clarity. Notably 42% of the notes were marked by affections of fondness, love or dependence of others. Regarding attitudes, the most common were of escape or farewell (42.4%, followed by fatalism, hopelessness, frustration or tiredness (40%. 24 statistically significant differences were found throughout the categories of analysis, according to cohorts of age, marital status and sex. Conclusions: the findings contribute to the profiling of Chilean suicides and to the implementation of suicide prevention programs 15. Notes on the quantum tetrahedron CERN Document Server Coquereaux, Robert 2002-01-01 This is a set of notes describing several aspects of the space of paths on ADE Dynkin diagrams, with a particular attention paid to the graph E6. Many results originally due to A. Ocneanu are here described in a very elementary way (manipulation of square or rectangular matrices). We define the concept of essential matrices for a graph and describe their module properties with respect to right and left actions of fusion algebras. In the case of the graph E6, essential matrices build up a right module with respect to its fusion algebra but a left module with respect to the fusion algebra of A11. We present two original results: 1) We show how to recover the Ocneanu graph of quantum symmetries of the Dynkin diagram E6 from the natural multiplication defined in the tensor square of its fusion algebra (the tensor product should be taken over a particular subalgebra); this is the Cayley graph for the two generators of the twelve dimensional algebra (E6 \\otimes_A3 E6); here A3 and E6 refer to the commutative fusion... 16. Notes on the Spatial Turn Directory of Open Access Journals (Sweden) Stipe Grgas 2012-06-01 Full Text Available On the basis of ever-mounting evidence, amongst which is the “zone” problematic of the Zadar conference that occassioned these notes, it can be concluded that the spatial turn has insinuated itself as an all-pervading heuristic tool throughout the humanities and the social sciences. The extent to which space and spatiality have usurped the central stage in the various branches of reasearch can be gauged by admonishments that what we are witnessing is a new fundamentalism that has simply inverted the terms of the dualism of time and space (May and Thrift 2001: “Introduction”. According to Michael Dear the sway of space is manifested in multifold ways: in the ubiquity of spatial analysis in social theories and practices; in the explosion of publications devoted to the exploration of the interface of the social and the spatial; in the reintegration of human geography into various domains of knowledge; in the focus given to difference and the consequent diversification of theoretical and empirical practices; in a theoretically informed exploration of the relation between geographical knowledge and social action; and, finally, in the unprecedented proliferation of research agendas and publications pertaining to these isuuses (Dear 2001: 24. Two recent collections of papers are indicative of the ubiquity of spatial issues in scholarly work. 17. LLW Notes, volume 9, No. 7. November and December 1994 International Nuclear Information System (INIS) 1994-12-01 LLW Notes is distributed to Low-Level Radioactive Waste Forum Participants and other state and compact officials identified by those Participants to receive LLW Notes. The Low-Level Radioactive Waste Forum is an association of state and compact representatives appointed by governors and compact commissions, established to facilitate state and compact commission implementation of the Low- Level Radioactive Waste Policy Act of 1980 and the Low-Level Radioactive Waste Policy Amendments Act of 1985 and to promote the objectives of low-level radioactive waste regional compacts. The Forum provides an opportunity for states and compacts to share information with one another and to exchange views with officials of federal agencies and other interested parties 18. LLW Notes, vol.9, no. 5. August/September 1994 International Nuclear Information System (INIS) 1994-09-01 LLW Notes is distributed to Low-Level Radioactive Waste Forum Participants and other state and compact officials identified by those Participants to receive LLW Notes. The Low-Level Radioactive Waste Forum is an association of state and compact representatives appointed by governors and compact commissions, established to facilitate state and compact commission implementation of the Low- Level Radioactive Waste Policy Act of 1980 and the Low-Level Radioactive Waste Policy Amendments Act of 1985 and to promote the objectives of low-level radioactive waste regional compacts. The Forum provides an opportunity for states and compacts to share information with one another and to exchange views with officials of federal agencies and other interested parties 19. LLW Notes, vol.9, no. 5. August/September 1994 Energy Technology Data Exchange (ETDEWEB) NONE 1994-09-01 LLW Notes is distributed to Low-Level Radioactive Waste Forum Participants and other state and compact officials identified by those Participants to receive LLW Notes. The Low-Level Radioactive Waste Forum is an association of state and compact representatives appointed by governors and compact commissions, established to facilitate state and compact commission implementation of the Low- Level Radioactive Waste Policy Act of 1980 and the Low-Level Radioactive Waste Policy Amendments Act of 1985 and to promote the objectives of low-level radioactive waste regional compacts. The Forum provides an opportunity for states and compacts to share information with one another and to exchange views with officials of federal agencies and other interested parties. 20. LLW Notes, Volume 9, Number 6. October 1994 International Nuclear Information System (INIS) 1994-10-01 LLW Notes is distributed to Low-Level Radioactive Waste Forum Participants and other state and compact officials identified by those Participants to receive LLW Notes. The Low-Level Radioactive Waste Forum is an association of state and compact representatives appointed by governors and compact commissions, established to facilitate state and compact commission implementation of the Low- Level Radioactive Waste Policy Act of 1980 and the Low-Level Radioactive Waste Policy Amendments Act of 1985 and to promote the objectives of low-level radioactive waste regional compacts. The Forum provides an opportunity for states and compacts to share information with one another and to exchange views with officials of federal agencies and other interested parties 1. LLW Notes, volume 9, No. 7. November and December 1994 Energy Technology Data Exchange (ETDEWEB) NONE 1994-12-01 LLW Notes is distributed to Low-Level Radioactive Waste Forum Participants and other state and compact officials identified by those Participants to receive LLW Notes. The Low-Level Radioactive Waste Forum is an association of state and compact representatives appointed by governors and compact commissions, established to facilitate state and compact commission implementation of the Low- Level Radioactive Waste Policy Act of 1980 and the Low-Level Radioactive Waste Policy Amendments Act of 1985 and to promote the objectives of low-level radioactive waste regional compacts. The Forum provides an opportunity for states and compacts to share information with one another and to exchange views with officials of federal agencies and other interested parties. 2. LLW Notes, Volume 9, Number 6. October 1994 Energy Technology Data Exchange (ETDEWEB) NONE 1994-10-01 LLW Notes is distributed to Low-Level Radioactive Waste Forum Participants and other state and compact officials identified by those Participants to receive LLW Notes. The Low-Level Radioactive Waste Forum is an association of state and compact representatives appointed by governors and compact commissions, established to facilitate state and compact commission implementation of the Low- Level Radioactive Waste Policy Act of 1980 and the Low-Level Radioactive Waste Policy Amendments Act of 1985 and to promote the objectives of low-level radioactive waste regional compacts. The Forum provides an opportunity for states and compacts to share information with one another and to exchange views with officials of federal agencies and other interested parties. 3. Open Oncology Notes: A Qualitative Study of Oncology Patients' Experiences Reading Their Cancer Care Notes. Science.gov (United States) Kayastha, Neha; Pollak, Kathryn I; LeBlanc, Thomas W 2018-04-01 4. The Effect of Speed Alterations on Tempo Note Selection. Science.gov (United States) 1986-01-01 Investigated the tempo note preferences of 100 randomly selected college-level musicians using familiar orchestral music as stimuli. Subjects heard selections at increased, decreased, and unaltered tempi. Results showed musicians were not accurate in estimating original tempo and showed consistent preference for faster than actual tempo.… 5. CRAC2 model description International Nuclear Information System (INIS) Ritchie, L.T.; Alpert, D.J.; Burke, R.P.; Johnson, J.D.; Ostmeyer, R.M.; Aldrich, D.C.; Blond, R.M. 1984-03-01 The CRAC2 computer code is a revised version of CRAC (Calculation of Reactor Accident Consequences) which was developed for the Reactor Safety Study. This document provides an overview of the CRAC2 code and a description of each of the models used. Significant improvements incorporated into CRAC2 include an improved weather sequence sampling technique, a new evacuation model, and new output capabilities. In addition, refinements have been made to the atmospheric transport and deposition model. Details of the modeling differences between CRAC2 and CRAC are emphasized in the model descriptions 6. Olkiluoto site description 2011 International Nuclear Information System (INIS) 2012-12-01 This fourth version of the Olkiluoto Site Report, produced by the OMTF (Olkiluoto Modelling Task Force), updates the Olkiluoto Site Report 2008 with the data and knowledge obtained up to December 2010. A descriptive model of the site (the Site Descriptive Model, SDM), i.e. a model describing the geological and hydrogeological structure of the site, properties of the bedrock and the groundwater and its flow, and the associated interacting processes and mechanisms. The SDM is divided into six parts: surface system, geology, rock mechanics, hydrogeology, hydrogeochemistry and transport properties 7. Olkiluoto site description 2011 Energy Technology Data Exchange (ETDEWEB) NONE 2012-12-15 This fourth version of the Olkiluoto Site Report, produced by the OMTF (Olkiluoto Modelling Task Force), updates the Olkiluoto Site Report 2008 with the data and knowledge obtained up to December 2010. A descriptive model of the site (the Site Descriptive Model, SDM), i.e. a model describing the geological and hydrogeological structure of the site, properties of the bedrock and the groundwater and its flow, and the associated interacting processes and mechanisms. The SDM is divided into six parts: surface system, geology, rock mechanics, hydrogeology, hydrogeochemistry and transport properties. 8. Station Program Note Pull Automation Science.gov (United States) 2016-01-01 9. Biomarkers of genotoxicity of urban air pollution. Overview and descriptive data from a molecular epidemiology study on populations exposed to moderate-to-low levels of polycyclic aromatic hydrocarbons: the AULIS project DEFF Research Database (Denmark) Kyrtopoulos, S.A.; Georgiadis, P.; Autrup, H. 2001-01-01 in the context of the European Union Environment and Climate Programme, known by the short name of AULIS project, is discussed and descriptive statistics of the main findings of the project are presented. These findings indicate that for cohorts suffering moderate-to-low exposures to airborne particulate... 10. The transportation operations system: A description International Nuclear Information System (INIS) Best, R.E.; Danese, F.L.; Dixon, L.D.; Peterson, R.W.; Pope, R.B. 1990-01-01 This paper presents a description of the system for transporting radioactive waste that may be deployed to accomplish the assigned system mission, which includes accepting spent nuclear fuel (SNF) and high-level radioactive waste (HLW) from waste generator sites and transporting them to the FWMS destination facilities. The system description presented here contains, in part, irradiated fuel and waste casks, ancillary equipments, truck, rail, and barge transporters, cask and vehicle traffic management organizations, maintenance facilities, and other operations elements. The description is for a fully implemented system, which is not expected to be achieved, however, until several years after initial operations. 6 figs 11. Employers' views on the fit note. Science.gov (United States) Kotze, E 2014-12-01 12. Clinical Note Creation, Binning, and Artificial Intelligence OpenAIRE Deliberato, Rodrigo Oct?vio; Celi, Leo Anthony; Stone, David J 2017-01-01 The creation of medical notes in software applications poses an intrinsic problem in workflow as the technology inherently intervenes in the processes of collecting and assembling information, as well as the production of a data-driven note that meets both individual and healthcare system requirements. In addition, the note writing applications in currently available electronic health records (EHRs) do not function to support decision making to any substantial degree. We suggest that artifici... 13. Addressing the Language Description Deficit Directory of Open Access Journals (Sweden) M. Ali Bolgiin 2010-01-01 Full Text Available Well-described language features are key to successful teaching and learning, especially for achieving advanced levels of proficiency. Other measures, such as simply increasing the number of reading and listening passages in a language program alone are not enough to bring the student to a higher level in a given skill. In fact, even being present in the target culture does not suffice. Angelelli and Degueldre (2002 argue that at advanced levels, even spending time in a country where the language is spoken is not necessarily sufficient for learners: "They do not need just exposure; they need answers to questions and explanations that they can rarely get by simply being immersed in a language/ culture." Less commonly taught languages (LCTLs lack descriptions that have such answers and explanations (cf. Fotos, 2002. It is argued in this paper that corpuslinguistic analyses help to provide actual usage-based, rather than intuition-based, descriptions and explanations of language features. Such approach is illustrated through English and Turkish examples. 14. Osiris reactor descriptive report International Nuclear Information System (INIS) 1976-03-01 OSIRIS is a swimming pool reactor of 70 MW thermal power. Its main purpose is the irradiation of reactor materials in high neutron flux. A description is given of the air conditioning, ventilation, and radioactive gas removal system. (R.L.) 15. Plot Description (PD) Science.gov (United States) Robert E. Keane 2006-01-01 The Plot Description (PD) form is used to describe general characteristics of the FIREMON macroplot to provide ecological context for data analyses. The PD data characterize the topographical setting, geographic reference point, general plant composition and cover, ground cover, fuels, and soils information. This method provides the general ecological data that can be... 16. Generalizing: The descriptive struggle Directory of Open Access Journals (Sweden) Barney G. Glaser, Ph.D.; Hon Ph.D. 2006-11-01 Full Text Available The literature is not kind to the use of descriptive generalizations. Authors struggle and struggle to find and rationalize a way to use them and then fail in spite of trying a myriad of work-arounds. And then we have Lincoln and Guba’s famous statement: “The only generalization is: there is no generalization” in referring to qualitative research. (op cit, p. 110 They are referring to routine QDA yielding extensive descriptions, but which tacitly include conceptual generalizations without any real thought of knowledge about them. In this chapter I wish to explore this struggle for the purpose of explaining that the various contra arguments to using descriptive generalizations DO NOT apply to the ease of using conceptual generalizations yielded in SGT and especially FGT. I will not argue for the use of descriptive generalization. I agree with Lincoln and Guba with respect to QDA, “the only generalization is: there is no generalization.” It is up to the QDA methodologists, of whom there are many; to continue the struggle and I wish them well. 17. Fractal description of fractures International Nuclear Information System (INIS) Lung, C.W. 1991-06-01 Recent studies on the fractal description of fractures are reviewed. Some problems on this subject are discussed. It seems hopeful to use the fractal dimension as a parameter for quantitative fractography and to apply fractal structures to the development of high toughness materials. (author). 28 refs, 7 figs 18. Tore Supra: technical description International Nuclear Information System (INIS) 1985-08-01 This report is aimed, after a brief recall of physics and technologic perspectives of Tore Supra, at giving a detailed description of the basic machine; details of each component are defined. Volume 1 is specifically concerned with the general aspects of Tore Supra and the toroidal field system [fr 19. BNL ALARA Center: ALARA Notes, No. 9 International Nuclear Information System (INIS) Khan, T.A.; Xie, J.W.; Beckman, M.C. 1994-02-01 This issue of the Brookhaven National Laboratory's Alara Notes includes the agenda for the Third International Workshop on ALARA and specific instructions on the use of the on-line fax-on-demand service provided by BNL. Other topics included in this issue are: (1) A discussion of low-level discharges from Canadian nuclear plants, (2) Safety issues at French nuclear plants, (3) Acoustic emission as a means of leak detection, (4) Replacement of steam generators at Doel-3, Beaznau, and North Anna-1, (5) Remote handling equipment at Bruce, (6) EPRI's low level waste program, (7) Radiation protection during concrete repairs at Savannah River, (8) Reactor vessel stud removal/repair at Comanche Peak-1, (9) Rework of reactor coolant pump motors, (10) Restoration of service water at North Anna-1 and -2, (11) Steam generator tubing problems at Mihama-1, (12) Full system decontamination at Indian Point-2, (13) Chemical decontamination at Browns Ferry-2, and (14) Inspection methodolody in France and Japan 20. The silica road: Field trip notes Directory of Open Access Journals (Sweden) María Soto 2016-09-01 This paper present some keynotes distributed to the symposium attendants. It consists on: 1 a brief geological framework of the Montsant Massif, located at the southern margin of the Catalan Central Depression. It includes the definition of their Palaeozoic to Cenozoic depositional sequence and the localization of some points of interests with panoramic views and source areas where Tertiary chert nodules were available. They represent a significant focus for the raw materials procurement of several Paleolithic occupations since Lower Pleistocene. 2 A general presentation of the Middle Palaeolithic site of the Abric Romaní site, including a brief reference of the three research phases initiated at the beginning of 20th Century, the description of its 50 meters stratigraphic sequence, where 27 archaeological levels have been identified dating since 110 to 39 ka BP, and some of the main traits of the archaeological assemblages recovered. 1. ELECTRICAL POWER SYSTEM DESCRIPTION DOCUMENT Energy Technology Data Exchange (ETDEWEB) M. Maniyar 2004-06-22 The purpose of this revision of the System Description Document (SDD) is to establish requirements that drive the design of the electrical power system and their bases to allow the design effort to proceed to License Application. This SDD is a living document that will be revised at strategic points as the design matures over time. This SDD identifies the requirements and describes the system design as they exist at this time, with emphasis on those attributes of the design provided to meet the requirements. This SDD has been developed to be an engineering tool for design control. Accordingly, the primary audience are design engineers. This type of SDD leads and follows the design process. It leads the design process with regard to the flow down of upper tier requirements onto the system. Knowledge of these requirements is essential to performing the design process. This SDD follows the design with regard to the description of the system. The description provided in the SDD is a reflection of the results of the design process to date. Functional and operational requirements applicable to this system are obtained from ''Project Functional and Operational Requirements'' (F&OR) (Siddoway, 2003). Other requirements to support the design process have been taken from higher level requirements documents such as ''Project Design Criteria Document'' (PDC) (Doraswamy 2004), the fire hazards analyses, and the preclosure safety analysis. The above mentioned low-level documents address ''Project Requirements Document'' (PRD) (Canori and Leitner 2003) requirements. This SDD includes several appendices with supporting information. Appendix B lists key system charts, diagrams, drawings, and lists; and Appendix C is a list of system procedures. 2. ELECTRICAL SUPPORT SYSTEM DESCRIPTION DOCUMENT Energy Technology Data Exchange (ETDEWEB) S. Roy 2004-06-24 The purpose of this revision of the System Design Description (SDD) is to establish requirements that drive the design of the electrical support system and their bases to allow the design effort to proceed to License Application. This SDD is a living document that will be revised at strategic points as the design matures over time. This SDD identifies the requirements and describes the system design as they exist at this time, with emphasis on those attributes of the design provided to meet the requirements. This SDD has been developed to be an engineering tool for design control. Accordingly, the primary audience/users are design engineers. This type of SDD both ''leads'' and ''trails'' the design process. It leads the design process with regard to the flow down of upper tier requirements onto the system. Knowledge of these requirements is essential in performing the design process. The SDD trails the design with regard to the description of the system. The description provided in the SDD is a reflection of the results of the design process to date. Functional and operational requirements applicable to electrical support systems are obtained from the ''Project Functional and Operational Requirements'' (F&OR) (Siddoway 2003). Other requirements to support the design process have been taken from higher-level requirements documents such as the ''Project Design Criteria Document'' (PDC) (Doraswamy 2004), and fire hazards analyses. The above-mentioned low-level documents address ''Project Requirements Document'' (PRD) (Canon and Leitner 2003) requirements. This SDD contains several appendices that include supporting information. Appendix B lists key system charts, diagrams, drawings, and lists, and Appendix C includes a list of system procedures. 3. ELECTRICAL SUPPORT SYSTEM DESCRIPTION DOCUMENT International Nuclear Information System (INIS) Roy, S. 2004-01-01 The purpose of this revision of the System Design Description (SDD) is to establish requirements that drive the design of the electrical support system and their bases to allow the design effort to proceed to License Application. This SDD is a living document that will be revised at strategic points as the design matures over time. This SDD identifies the requirements and describes the system design as they exist at this time, with emphasis on those attributes of the design provided to meet the requirements. This SDD has been developed to be an engineering tool for design control. Accordingly, the primary audience/users are design engineers. This type of SDD both ''leads'' and ''trails'' the design process. It leads the design process with regard to the flow down of upper tier requirements onto the system. Knowledge of these requirements is essential in performing the design process. The SDD trails the design with regard to the description of the system. The description provided in the SDD is a reflection of the results of the design process to date. Functional and operational requirements applicable to electrical support systems are obtained from the ''Project Functional and Operational Requirements'' (F andOR) (Siddoway 2003). Other requirements to support the design process have been taken from higher-level requirements documents such as the ''Project Design Criteria Document'' (PDC) (Doraswamy 2004), and fire hazards analyses. The above-mentioned low-level documents address ''Project Requirements Document'' (PRD) (Canon and Leitner 2003) requirements. This SDD contains several appendices that include supporting information. Appendix B lists key system charts, diagrams, drawings, and lists, and Appendix C includes a list of system procedures 4. A Note on the Stability of the Job Descriptive Index (J. D. I.). Science.gov (United States) 1978-05-01 AND A D D R E S S 10. P R O G R A M ELEMENT , PROJE CT , TASK A R E A & WORK UNIT NUMBER S Departme nt of Psychology 6 l l53 N University of...Center , University of Mary land , College Park 19. KEY WOR DS (Con t inue on re v e rse aids II necalaary ~ td ident ify by block num ber) Attitudes 5. Notes on Wyeomyia Mosquitoes of Suriname, with a Description of Wyeomyia surinamensis sp.n. NARCIS (Netherlands) Bruijning, C.F.A. 1959-01-01 Most of the material recorded in this paper was collected by the author during his stay in Suriname from 1949 to 1955. By courtesy of Mrs. J. BONNE-WEPSTER, the author was enabled to study the Wyeomyia specimens which were collected by BONNE and BONNE-WEPSTER in Suriname and are at present in the 6. Some preliminary notes on Surinam Sting Rays, including the description of a new species NARCIS (Netherlands) Boeseman, M. 1948-01-01 In Dr. D. C. Geijskes' collection of Surinam fishes, mentioned in my previous paper on this subject (Boeseman, 1948), I found six specimens belonging to the so-called "sting rays" (Dasyatidae), representing three different species, and all very interesting in some respects. I. Dasyatis schmardae 7. Note on the hydrodynamic description of thin nematic films: Strong anchoring model KAUST Repository Lin, Te-Sheng; Cummings, Linda J.; Archer, Andrew J.; Kondic, Lou; Thiele, Uwe 2013-01-01 We discuss the long-wave hydrodynamic model for a thin film of nematic liquid crystal in the limit of strong anchoring at the free surface and at the substrate. We rigorously clarify how the elastic energy enters the evolution equation for the film thickness in order to provide a solid basis for further investigation: several conflicting models exist in the literature that predict qualitatively different behaviour. We consolidate the various approaches and show that the long-wave model derived through an asymptotic expansion of the full nemato-hydrodynamic equations with consistent boundary conditions agrees with the model one obtains by employing a thermodynamically motivated gradient dynamics formulation based on an underlying free energy functional. As a result, we find that in the case of strong anchoring the elastic distortion energy is always stabilising. To support the discussion in the main part of the paper, an appendix gives the full derivation of the evolution equation for the film thickness via asymptotic expansion. © 2013 AIP Publishing LLC. 8. Taxonomic notes and description of the male of Xenochlora nigrofemorata (Smith, 1879) (Hymenoptera: Apidae: Halictinae). Science.gov (United States) Santos, Leandro M; Melo, Gabriel A R 2013-01-01 The present work describes for the first time the male of the bee genus Xenochlora Engel, Brooks & Yanega, 1997. The male of X. nigrofemorata (Smith, 1879) is described and illustrated. Additionally, Megalopta opacicollis Friese, 1926 is placed as a junior synonym of X. nigrofemorata (Smith, 1879). 9. Notes on the tribe Mimagathidini Enderlein, with the description of a new genus (Hymenoptera: Braconidae: Orgilinae) NARCIS (Netherlands) Braet, Y.; Achterberg, van C.; Chen, X. 2000-01-01 Eleonoria gen. nov. (type species: E. mesembria spec. nov. from Malaysia; five additional species from Japan, China (including Taiwan), Philippines, and Indonesia (Halmahera) are included), Orgilonia pasohensis spec. nov. from Malaysia, Orgilus reclinatus spec. nov. from Congo, O. alboannulatus 10. Technical Note: Description and assessment of a nudged version of the new dynamics Unified Model Directory of Open Access Journals (Sweden) O. Morgenstern 2008-03-01 Full Text Available We present a "nudged" version of the Met Office general circulation model, the Unified Model. We constrain this global climate model using ERA-40 re-analysis data with the aim of reproducing the observed "weather" over a year from September 1999. Quantitative assessments are made of its performance, focusing on dynamical aspects of nudging and demonstrating that the "weather" is well simulated. 11. Taxonomic notes on Chironomidae (Diptera from Okinawa Island, Japan, with the description of three new species Directory of Open Access Journals (Sweden) Masaru Yamamoto 2015-07-01 Full Text Available Three new species: Ionthosmittia cuneipenne sp. nov., Orthocladius (Euorthocladius okinawanus sp. nov. and Parakiefferiella semiovata sp. nov., are described from Okinawa Island, Ryukyus, Japan. In addition to these species, twelve species are newly recorded from this island. Diagnostic characters of Tokyobrillia tamamegaseta (Kobayashi et Sasa are emended. Pentapedilum yakuabeum Sasa et Suzuki syn. nov. and Polypedilum yakucedeum Sasa et Suizuki syn. nov. are junior synonyms of Ainuyusurika tuberculatum (Tokunaga. Einfeldia kanazawai (Yamamoto is transferred to the genus Chironomus Meigen and its systematic position is discussed. 12. Description of male Tylorida sataraensis Kulkarni, 2014 (Araneae, Tetragnathidae with notes on habits and conservation status Directory of Open Access Journals (Sweden) Siddharth Kulkarni 2015-02-01 Full Text Available The male sex of Tylorida sataraensis Kulkarni, 2014 is described based on specimens from the type locality. The distinguishing characters from its closest species Tylorida ventralis (Thorell, 1877 are detailed. An interesting behaviour of going underwater by T. sataraensis, on disturbance is recorded and tested for significance. The surveys have shown sighting of this species only to the perennial streams of the rocky outcrops in Satara region. The potential threats to this species and the possible conservation status based on known distribution are discussed. 13. Notes on Pauropoda (Myriapoda from U.S.A. with descriptions of two new species Directory of Open Access Journals (Sweden) Ulf Scheller 2011-07-01 Full Text Available Two new species of Pauropoda are described from USA, Kionopauropus alyeskaensis sp. n. (Pauropodidae, and Eurypauropus arcuatus sp. n. (Eurypauropodidae. The genus Kionopauropus is reported from the Western Hemisphere for the first time. 14. Revisionary notes on Bentonia van Achterberg, 1992 (Hymenoptera: Braconidae: Orgilinae) with description of two new species NARCIS (Netherlands) Braet, Y.; Tignon, M. 1998-01-01 Two new species of the genus Bentonia van Achterberg, 1992 (Braconidae: Orgilinae) (B. inca from Peru and B. xochiquetzalis from Mexico) are described and partly illustrated. A third undescribed species was found for which some characters are listed. The distribution of B. scutellaris van 15. Coping with Stress. Research Notes. Science.gov (United States) Jordan, Debra J. 1995-01-01 Research related to the impact of exercise on stress indicates that a regular aerobic exercise program is important to control the negative effects of stress. It was also reported that those who are physically fit have higher levels of self-esteem. Implications for camp staff involve starting a regular exercise program to offset job-related… 16. Notes on Loricata (Mollusca) 11-14 NARCIS (Netherlands) Kaas, P. 1985-01-01 Four new species of chitons (Mollusca, Polyplacophora) are described, viz Leptochiton (L.) dispersus and L. (L.) permodestus from Transkei, L. (L.) meiringae from the eastern Cape Province, S Africa (note 11) and Ischnochiton (Stenosemus) vanbellei from the Mediterranean Sea (note 14). New records 17. Suicide Notes in Hong Kong in 2000 Science.gov (United States) Wong, Paul W. C.; Yeung, April W. M.; Chan, Wincy S. C.; Yip, Paul S. F.; Tang, Arthur K. H. 2009-01-01 Suicide notes have been regarded as one of the most informative data sources to understand the reasons why people commit suicide. However, there is a paucity of suicide note studies, leaving researchers with an assumption that this phenomenon remains static over time. This study examines this assumption by comparing the characteristics of… 18. Stalin and Marxism: a research note NARCIS (Netherlands) van Ree, E. 1997-01-01 This article concerns the research done by the author in Stalin‘s private library. The notes made in the works of Marx, Engels and Lenin suggest that until the end of his life Stalin felt himself in general agreement with these "classics." The choice of books and the notes support the thesis that, 19. Children's note taking as a mnemonic tool. Science.gov (United States) Eskritt, Michelle; McLeod, Kellie 2008-09-01 When given the opportunity to take notes in memory tasks, children sometimes make notes that are not useful. The current study examined the role that task constraints might play in the production of nonmnemonic notes. In Experiment 1, children played one easy and one difficult memory game twice, once with the opportunity to make notes and once without that opportunity. More children produced functional notations for the easier task than for the more difficult task, and their notations were beneficial to memory performance. Experiment 2 found that the majority of children who at first made nonmnemonic notations were able to produce functional notations with minimal training, and there was no significant difference in notation quality or memory performance between spontaneous and trained note takers. Experiment 3 revealed that the majority of children could transfer their training to a novel task. The results suggest that children's production of nonmnemonic notes may be due in part to a lack of knowledge regarding what task information is important to represent or how to represent it in their notes rather than to an inability to make functional notes in general. 20. Observing professionals taking notes on screen NARCIS (Netherlands) Melenhorst, M.S. 2005-01-01 In this study 38 participants wrote a piece of advice based on reading and annotating information from an extensive Web site. Half of the participants took notes in a separate window, the other half used an advanced annotation tool. In text annotations were far more used than separate notes. The 1. RETRIEVAL EQUIPMENT DESCRIPTIONS International Nuclear Information System (INIS) J. Steinhoff 1997-01-01 The objective and the scope of this document are to list and briefly describe the major mobile equipment necessary for waste package (WP) retrieval from the proposed subsurface nuclear waste repository at Yucca Mountain. Primary performance characteristics and some specialized design features of the equipment are explained and summarized in the individual subsections of this document. There are no quality assurance requirements or QA controls in this document. Retrieval under normal conditions is accomplished with the same fleet of equipment as is used for emplacement. Descriptions of equipment used for retrieval under normal conditions is found in Emplacement Equipment Descriptions, DI: BCAF00000-01717-5705-00002 (a document in progress). Equipment used for retrieval under abnormal conditions is addressed in this document and consists of the following: (1) Inclined Plane Hauler; (2) Bottom Lift Transporter; (3) Load Haul Dump (LHD) Loader; (4) Heavy Duty Forklift for Emplacement Drifts; (5) Covered Shuttle Car; (6) Multipurpose Vehicle; and (7) Scaler 2. Multidimensional nonlinear descriptive analysis CERN Document Server Nishisato, Shizuhiko 2006-01-01 Quantification of categorical, or non-numerical, data is a problem that scientists face across a wide range of disciplines. Exploring data analysis in various areas of research, such as the social sciences and biology, Multidimensional Nonlinear Descriptive Analysis presents methods for analyzing categorical data that are not necessarily sampled randomly from a normal population and often involve nonlinear relations. This reference not only provides an overview of multidimensional nonlinear descriptive analysis (MUNDA) of discrete data, it also offers new results in a variety of fields. The first part of the book covers conceptual and technical preliminaries needed to understand the data analysis in subsequent chapters. The next two parts contain applications of MUNDA to diverse data types, with each chapter devoted to one type of categorical data, a brief historical comment, and basic skills peculiar to the data types. The final part examines several problems and then concludes with suggestions for futu... 3. TMACS system description International Nuclear Information System (INIS) Scaief, C.C. 1995-01-01 This document provides a description of the Tank Monitor and Control System (TMACS). It is intended as an introduction for those persons unfamiliar with the system as well as a reference document for the users, maintenance personnel, and system designers. In addition to describing the system, the document outlines the associated drawing documentation, provides maintenance and spare parts information, and discusses other TMACS documents that provide additional detail 4. Input description for BIOPATH International Nuclear Information System (INIS) Marklund, J.E.; Bergstroem, U.; Edlund, O. 1980-01-01 The computer program BIOPATH describes the flow of radioactivity within a given ecosystem after a postulated release of radioactive material and the resulting dose for specified population groups. The present report accounts for the input data necessary to run BIOPATH. The report also contains descriptions of possible control cards and an input example as well as a short summary of the basic theory.(author) 5. Confidence assessment. Site-descriptive modelling SDM-Site Laxemar International Nuclear Information System (INIS) 2009-06-01 independent data from different disciplines. While some aspects have lower confidence this lack of confidence is handled by providing wider uncertainty ranges, bounding estimates and/or alternative models to repository engineering and long term safety assessment. It is judged that most, of the low confidence aspects have little impact on repository engineering design or for long-term safety. It may also be noted that the feedback requirements from SR-Can to the site modelling are now met in the completed site investigations, subject to levels of uncertainty that are viewed as acceptable. Only a few data points and a few types of data have been omitted from the modelling, mainly because they are judged less relevant and reliable than the data considered. Inclusion of data from outside the Laxemar subarea might have enhanced confidence in the regional model, but only at the locations of the data and these changes in confidence would have been of little significance in relation to implications for the local model area and would not, therefore, have been of any real significance to design or safety assessment. These omissions are judged to have little or no negative impact on confidence in the Laxemar subarea model. In fact, identification of unreliable data and their elimination should have a positive effect on confidence. Poor precision in the measured data is judged to have a limited impact on uncertainties in the site descriptive model, with the exceptions of interpretation and combination of borehole and outcrop fracture data and general uncertainties in sorption data 6. Confidence assessment. Site-descriptive modelling SDM-Site Laxemar Energy Technology Data Exchange (ETDEWEB) 2008-12-15 independent data from different disciplines. While some aspects have lower confidence this lack of confidence is handled by providing wider uncertainty ranges, bounding estimates and/or alternative models to repository engineering and long term safety assessment. It is judged that most, of the low confidence aspects have little impact on repository engineering design or for long-term safety. It may also be noted that the feedback requirements from SR-Can to the site modelling are now met in the completed site investigations, subject to levels of uncertainty that are viewed as acceptable. Only a few data points and a few types of data have been omitted from the modelling, mainly because they are judged less relevant and reliable than the data considered. Inclusion of data from outside the Laxemar subarea might have enhanced confidence in the regional model, but only at the locations of the data and these changes in confidence would have been of little significance in relation to implications for the local model area and would not, therefore, have been of any real significance to design or safety assessment. These omissions are judged to have little or no negative impact on confidence in the Laxemar subarea model. In fact, identification of unreliable data and their elimination should have a positive effect on confidence. Poor precision in the measured data is judged to have a limited impact on uncertainties in the site descriptive model, with the exceptions of interpretation and combination of borehole and outcrop fracture data and general uncertainties in sorption data 7. XML Diagnostics Description Standard International Nuclear Information System (INIS) Neto, A.; Fernandes, H.; Varandas, C.; Lister, J.; Yonekawa, I. 2006-01-01 A standard for the self-description of fusion plasma diagnostics will be presented, based on the Extensible Markup Language (XML). The motivation is to maintain and organise the information on all the components of a laboratory experiment, from the hardware to the access security, to save time and money when problems arises. Since there is no existing standard to organise this kind of information, every Association stores and organises each experiment in different ways. This can lead to severe problems when the organisation schema is poorly documented or written in national languages. The exchange of scientists, researchers and engineers between laboratories is a common practice nowadays. Sometimes they have to install new diagnostics or to update existing ones and frequently they lose a great deal of time trying to understand the currently installed system. The most common problems are: no documentation available; the person who understands it has left; documentation written in the national language. Standardisation is the key to solving all the problems mentioned. From the commercial information on the diagnostic (component supplier; component price) to the hardware description (component specifications; drawings) to the operation of the equipment (finite state machines) through change control (who changed what and when) and internationalisation (information at least in the native language and in English), a common XML schema will be proposed. This paper will also discuss an extension of these ideas to the self-description of ITER plant systems, since the problems will be identical. (author) 8. Management control system description Energy Technology Data Exchange (ETDEWEB) Bence, P. J. 1990-10-01 This Management Control System (MCS) description describes the processes used to manage the cost and schedule of work performed by Westinghouse Hanford Company (Westinghouse Hanford) for the US Department of Energy, Richland Operations Office (DOE-RL), Richland, Washington. Westinghouse Hanford will maintain and use formal cost and schedule management control systems, as presented in this document, in performing work for the DOE-RL. This MCS description is a controlled document and will be modified or updated as required. This document must be approved by the DOE-RL; thereafter, any significant change will require DOE-RL concurrence. Westinghouse Hanford is the DOE-RL operations and engineering contractor at the Hanford Site. Activities associated with this contract (DE-AC06-87RL10930) include operating existing plant facilities, managing defined projects and programs, and planning future enhancements. This document is designed to comply with Section I-13 of the contract by providing a description of Westinghouse Hanford's cost and schedule control systems used in managing the above activities. 5 refs., 22 figs., 1 tab. 9. Suicide note themes and suicide prevention. Science.gov (United States) Foster, Tom 2003-01-01 The aim was to determine if suicide note themes might inform suicide prevention strategies. The themes of 42 suicide notes from the Northern Ireland Suicide Study (major psychological autopsy study) were examined. The commonest themes were "apology/shame" (74%), "love for those left behind" (60%), "life too much to bear" (48%), "instructions regarding practical affairs post-mortem" (36%), "hopelessness/nothing to live for" (21%) and "advice for those left behind" (21%). Notes of suicides with major unipolar depression were more likely than notes of suicides without major unipolar depression to contain the themes "instructions regarding practical affairs post-mortem" (67% versus 19%, p = 0.005) and "hopelessness/nothing to live for" (40% versus 11%, p = 0.049). Notes of suicides with a previous history of deliberate self-harm were less likely than notes of suicides without a history of deliberate self-harm to contain the theme "apology/shame" (58% versus 87%, p = 0.04). Notes of elderly suicides were more likely than non-elderly notes to contain the theme "burden to others" (40% versus 3%, p = 0.03). The fact that three quarters of suicide notes contained the theme "apology/shame" suggests that the deceased may have welcomed alternative solutions for their predicaments. Scrutiny of suicide note themes in the light of previous research findings suggests that cognitive therapy techniques, especially problem solving, may have an important role to play in suicide prevention and that potential major unipolar depressive (possibly less impulsive) suicides, in particular, may provide fertile ground for therapeutic intervention (physical and psychological). Ideally all primary care doctors and mental health professionals working with (potentially) suicidal people should be familiar with basic cognitive therapy techniques, especially problem solving skills training. 10. Positive Effects of Restricting Student Note-Taking in a Capstone Psychology Course: Reducing the Demands of Divided Attention in the Classroom Science.gov (United States) Long, Gerald M. 2014-01-01 Two versions of a senior-level capstone course with differing note-taking strategies were compared. In one semester, a traditional student note-taking format was used; in another semester, student note-taking was rendered unnecessary by providing students with complete instructor notes. Student performance in the course as well as student opinions… 11. Note Taking on Trial: A Legal Application of Note-Taking Research Science.gov (United States) Kiewra, Kenneth A. 2016-01-01 This article is about note taking, but it is not an exhaustive review of note-taking literature. Instead, it portrays the application of note-taking research to an unusual and important area of practice--the law. I was hired to serve as an expert witness on note taking in a legal case that hinged, in part, on the completeness and accuracy of… 12. NOTE: Do acupuncture points exist? Science.gov (United States) Yan, Xiaohui; Zhang, Xinyi; Liu, Chenglin; Dang, Ruishan; Huang, Yuying; He, Wei; Ding, Guanghong 2009-05-01 We used synchrotron x-ray fluorescence analysis to probe the distribution of four chemical elements in and around acupuncture points, two located in the forearm and two in the lower leg. Three of the four acupuncture points showed significantly elevated concentrations of elements Ca, Fe, Cu and Zn in relation to levels in the surrounding tissue, with similar elevation ratios for Cu and Fe. The mapped distribution of these elements implies that each acupuncture point seems to be elliptical with the long axis along the meridian. 13. Nuclear asset management. Slide notes International Nuclear Information System (INIS) Puglia, W.; Bailey, H.; Kubinova, J. 2004-01-01 Nuclear asset management is defined as the process for making resource allocation and risk management decisions at all levels of nuclear generation business to maximize value/profitability for all stakeholders while maintaining plant safety. In the presentation, the NAM concept is explained, financial benefits achieved in US industry over the past 12 years are outlined, Data Systems and Solutions (DS and S) is presented as a joint venture between Rolls-Royce and SAIC, and NAM benefits in nuclear industry from DS and S client experience are demonstrated. (P.A.) 14. Outside home. Notes on reflexivity Directory of Open Access Journals (Sweden) Mara Clemente 2017-01-01 The paper proffers the idea in which a “reflexive process” on subjectivity can involve and/or hopefully involve the entire experience of the researcher, going beyond the borders of a single research. In the process, unexpected elements of subjectivity can come into play; in other cases the meaning attributed to them can change in time or can have a role different from what had been expected. Some elements, objects of epistemological analyses, as imposed by a reflexive approach, can become objects of attention also on the phenomenological level. 15. Lecture Notes on Mixing Ventilation DEFF Research Database (Denmark) Nielsen, Peter Vilhelm The main task of the ventilation system or the air-conditioning system is to supply· and remove air and airborne materials and to supply or remove heat from a room. The necessary level of fresh air will be supplied to· a room by a ventilation system, and heat from equipment or solar radiation can...... be removed by an air-conditioning system. An industrial ventilation system may both take care of the occupants' comfort and the industrial processes in the area.... 16. Bank note recognition for the vision impaired. Science.gov (United States) Hinwood, A; Preston, P; Suaning, G J; Lovell, N H 2006-06-01 Blind Australians find great difficulty in recognising bank notes. Each note has the same feel, with no Braille markings, irregular edges or other tangible features. In Australia, there is only one device available that can assist blind people recognise their notes. Internationally, there are devices available; however they are expensive, complex and have not been developed to cater for Australian currency. This paper discusses a new device, the MoneyTalker that takes advantage of the largely different colours and patterns on each Australian bank note and recognises the notes electronically, using the reflection and transmission properties of light. Different coloured lights are transmitted through the inserted note and the corresponding sensors detect distinct ranges of values depending on the colour of the note. Various classification algorithms were studied and the final algorithm was chosen based on accuracy and speed of recognition. The MoneyTalker has shown an accuracy of more than 99%. A blind subject has tested the device and believes that it is usable, compact and affordable. Based on the devices that are available currently in Australia, the MoneyTalker is an effective alternative in terms of accuracy and usability. 17. Open notes: doctors and patients signing on. Science.gov (United States) Delbanco, Tom; Walker, Jan; Darer, Jonathan D; Elmore, Joann G; Feldman, Henry J; Leveille, Suzanne G; Ralston, James D; Ross, Stephen E; Vodicka, Elisabeth; Weber, Valerie D 2010-07-20 Few patients read their doctors' notes, despite having the legal right to do so. As information technology makes medical records more accessible and society calls for greater transparency, patients' interest in reading their doctors' notes may increase. Inviting patients to review these notes could improve understanding of their health, foster productive communication, stimulate shared decision making, and ultimately lead to better outcomes. Yet, easy access to doctors' notes could have negative consequences, such as confusing or worrying patients and complicating rather than improving patient-doctor communication. To gain evidence about the feasibility, benefits, and harms of providing patients ready access to electronic doctors' notes, a team of physicians and nurses have embarked on a demonstration and evaluation of a project called OpenNotes. The authors describe the intervention and share what they learned from conversations with doctors and patients during the planning stages. The team anticipates that "open notes" will spread and suggests that over time, if drafted collaboratively and signed by both doctors and patients, they might evolve to become contracts for care. 18. EndNote 7.0 Directory of Open Access Journals (Sweden) Eapen Bell 2006-01-01 Full Text Available EndNote is a useful software for online literature search and efficient bibliography management. It helps to format the bibliography according to the citation style of each journal. EndNote stores references in a library file, which can be shared with others. It can connect to online resources like PubMed and retrieve search results as per the search criteria. It can also effortlessly integrate with popular word processors like MS Word. The Indian Journal of Dermatology, Venereology and Leprology website has a provision to import references to EndNote. 19. Description logic rules CERN Document Server Krötzsch, M 2010-01-01 Ontological modelling today is applied in many areas of science and technology,including the Semantic Web. The W3C standard OWL defines one of the most important ontology languages based on the semantics of description logics. An alternative is to use rule languages in knowledge modelling, as proposed in the W3C's RIF standard. So far, it has often been unclear how to combine both technologies without sacrificing essential computational properties. This book explains this problem and presents new solutions that have recently been proposed. Extensive introductory chapters provide the necessary 20. MCO Monitoring activity description International Nuclear Information System (INIS) SEXTON, R.A. 1998-01-01 Spent Nuclear Fuel remaining from Hanford's N-Reactor operations in the 1970s has been stored under water in the K-Reactor Basins. This fuel will be repackaged, dried and stored in a new facility in the 200E Area. The safety basis for this process of retrieval, drying, and interim storage of the spent fuel has been established. The monitoring of MCOS in dry storage is a currently identified issue in the SNF Project. This plan outlines the key elements of the proposed monitoring activity. Other fuel stored in the K-Reactor Basins, including SPR fuel, will have other monitoring considerations and is not addressed by this activity description 1. Analysis of medication adherence-related notes from a service-oriented community pharmacy. Science.gov (United States) Witry, Matthew; Parry, Rachel; McDonough, Randal; Deninger, Michael 2017-07-15 2. Usage Notes in the Oxford American Dictionary. Science.gov (United States) Berner, R. Thomas 1981-01-01 Compares the "Oxford American Dictionary" with the "American Heritage Dictionary." Examines the dictionaries' differences in philosophies of language, introductory essays, and usage notes. Concludes that the "Oxford American Dictionary" is too conservative, paternalistic, and dogmatic for the 1980s. (DMM) 3. Interpreters' notes. On the choice of language DEFF Research Database (Denmark) Dam, Helle Vrønning 2004-01-01 This paper reports on a small-scale empirical study on note-taking in consecutive interpreting. As data, the study draws on the notes produced by four subjects while interpreting one Spanish source text consecutively into Danish, on the one hand, and one Danish source text into Spanish...... to particular scrutiny here. However, somewhat surprisingly, the results of the analyses indicate that the choice of language in note-taking is governed mainly by the status of the language in the interpreters' language combination, i.e. whether it is an A- or a B-language, and much less by its status......, on the other. The aim of the study is to explore what governs conference interpreters' choice of language for their notes. The categories traditionally used to discuss, describe and explain this choice are those of 'source language' and 'target language', and these categories are therefore subject... 4. A NOTE ON THE POCHHAMMER FREQUENCY EQUATION African Journals Online (AJOL) cistvr A note on the Pochhammer frequency equation. ),,,,(. ),,,,(;/. 2 zwura. ZWURA. tT ρω. µ. = ω= , where ωis the angular frequency of the wave, which is considered to be imposed in this problem. We also introduce a material parameter α defined by. )2. /(. µ+λµ=α , which is related to Poisson's ratio ν by n- n-. =a. 22. 21 . We note ... 5. Electronic health records improve clinical note quality. Science.gov (United States) Burke, Harry B; Sessums, Laura L; Hoang, Albert; Becher, Dorothy A; Fontelo, Paul; Liu, Fang; Stephens, Mark; Pangaro, Louis N; O'Malley, Patrick G; Baxi, Nancy S; Bunt, Christopher W; Capaldi, Vincent F; Chen, Julie M; Cooper, Barbara A; Djuric, David A; Hodge, Joshua A; Kane, Shawn; Magee, Charles; Makary, Zizette R; Mallory, Renee M; Miller, Thomas; Saperstein, Adam; Servey, Jessica; Gimbel, Ronald W 2015-01-01 The clinical note documents the clinician's information collection, problem assessment, clinical management, and its used for administrative purposes. Electronic health records (EHRs) are being implemented in clinical practices throughout the USA yet it is not known whether they improve the quality of clinical notes. The goal in this study was to determine if EHRs improve the quality of outpatient clinical notes. A five and a half year longitudinal retrospective multicenter quantitative study comparing the quality of handwritten and electronic outpatient clinical visit notes for 100 patients with type 2 diabetes at three time points: 6 months prior to the introduction of the EHR (before-EHR), 6 months after the introduction of the EHR (after-EHR), and 5 years after the introduction of the EHR (5-year-EHR). QNOTE, a validated quantitative instrument, was used to assess the quality of outpatient clinical notes. Its scores can range from a low of 0 to a high of 100. Sixteen primary care physicians with active practices used QNOTE to determine the quality of the 300 patient notes. The before-EHR, after-EHR, and 5-year-EHR grand mean scores (SD) were 52.0 (18.4), 61.2 (16.3), and 80.4 (8.9), respectively, and the change in scores for before-EHR to after-EHR and before-EHR to 5-year-EHR were 18% (pquality scores significantly improved over the 5-year time interval. The EHR significantly improved the overall quality of the outpatient clinical note and the quality of all its elements, including the core and non-core elements. To our knowledge, this is the first study to demonstrate that the EHR significantly improves the quality of clinical notes. © The Author 2014. Published by Oxford University Press on behalf of the American Medical Informatics Association. 6. Remedial action programs annual meeting: Meeting notes International Nuclear Information System (INIS) 1987-01-01 The US Department of Energy Grand Junction Projects Office was pleased to host the 1987 Remedial Action programs Annual Meeting and herein presents notes from that meeting as prepared (on relatively short notice) by participants. These notes are a summary of the information derived from the workshops, case studies, and ad hoc committee reports rather than formal proceedings. The order of the materials in this report follows the actual sequence of presentations during the annual meeting 7. Lecture Notes in Statistics. 3rd Semester DEFF Research Database (Denmark) The lecture note is prepared to meet the requirements for the 3rd semester course in statistics at the Aarhus School of Business. It focuses on multiple regression models, analysis of variance, and log-linear models.......The lecture note is prepared to meet the requirements for the 3rd semester course in statistics at the Aarhus School of Business. It focuses on multiple regression models, analysis of variance, and log-linear models.... 8. Note-taking Strategies and Academic Achievement OpenAIRE Haghverdi, Hamid; Biria, Reza; Karimi, Lotfollah 2014-01-01 The purpose of the present article is two fold. That is, revealing the significance, underlying theory and findings concerning note-taking in the literature related and exploring, through a survey study, the Iranian professors‘ and students‘ attitudes towards the effect of teaching note-taking strategies on the students‘ academic achievement. To this end, many previous studies were reviewed and we knew that many scholars, conducting empirical studies, surveys and interviews, have highlighted ... 9. Notes on Computational Methodology and Tools of Thermoelectric Energy Systems DEFF Research Database (Denmark) Chen, Min; Bach, Inger Palsgaard; Rosendahl, Lasse 2007-01-01 The SPICE model allows the concurrent simulation of thermoelectric devices and application electric sub-models. It is an important step to implement the thermoelectric modeling at the system level. In this paper, temperature dependent material properties in the SPICE model, temperature and heat...... flow obtained by the code ANSYS Multiphysics and SPICE (Simulation Program with Integrated Circuit Emphasis), as well as some notes on the 3-D extension of the SPICE model are introduced.... 10. MITHRAS REDISCOVERED II. FURTHER NOTES ON CIMRM 1938 AND 1986 Directory of Open Access Journals (Sweden) Szabo Csaba 2015-06-01 Full Text Available Recently, an important Mithraic relief was rediscovered and republished by the author of these lines in collaboration with George Bounegru and Victor Sava. The relief, known in the literature as CIMRM 1938 was for a long time considered a „disappeared” monument, the only laconic description being that of Marteen J. Vermaseren from his monumental corpus. Due to the recently rediscovered photographs of the relief and the detailed analysis of the correspondence between Béla Cserni and Franz Cumont, the CIMRM 1938 is now became available for further research. In this article, I will add some further historiographic and iconographic notes on one of the biggest Mithraic reliefs found in Dacia, solving also another mysterious piece in Vermaseren’s catalogue, the CIMRM 1986. The article is also the first publication of Béla Cserni’s photograph about the relief. 11. THULE: A detailed description International Nuclear Information System (INIS) Terry, M.J. 1964-07-01 This report describes the THULE scheme of lattice physics calculation which has been developed in FORTRAN for the IBM 7090. This scheme predicts the neutron flux over energy and space, for many groups and regions, together with reactivity and reaction rate edits for both a single lattice cell and a reactor core. This report describes in detail the input requirements for the THULE programme which forms the main part of the scheme. Brief descriptions of the 7090 programmes TED 6 and NOAH are included as appendices. TED 6 will produce the THULE edits from a WDSN output tape and NOAH is a version of the METHUSELAH programme which contains many of the THULE edits and will also produce input cards for THULE. (author) 12. Metaphorical descriptions of wrongdoers Directory of Open Access Journals (Sweden) Dryll Ewa Marta 2017-10-01 Full Text Available What is a metaphoric picture of an evil person made of? In a study devoted to the development of the ability to use metaphorical descriptions of humans, the semantic fields of four target metaphors - Human-Swamp, Human-Snake, Human-Knife, and Human-Nettle - were established and compared. Subjects (365 young adults were asked to decipher the metaphors’ meanings. The results were obtained mainly by qualitative analysis, with frequency analysis of clusters containing synonymous meanings. The results indicate that when creating imaginary characteristics of evil people, young adults seem to be more concerned about the possibility of suffering verbal harassment (most commonly: vulgarity, mockery, gossip, jeering than the threat of actual physical assault. The results may prove useful for developmental comparisons. 13. Polemic and Descriptive Negations DEFF Research Database (Denmark) Horslund, Camilla Søballe 2011-01-01 to semantics and pragmatics, negations can be used in three different ways, which gives rise to a typology of three different types of negations: 1) the descriptive negation, 2) the polemic negation, and 3) the meta-linguistic negation (Nølke 1999, 4). This typology illuminates the fact that the negation...... common in certain social context or genres, while polemic negations are more likely to come up in other genres and social settings. Previous studies have shown a relation between articulatory prominence and register, which may further inform the analysis. Hence, the paper investigates how articulatory...... prominence and register may either work in concert or oppose each other with respect to the cues they provide for the interpretation.... 14. Descriptive sensory evaluations DEFF Research Database (Denmark) Dehlholm, Christian A recent trend in descriptive sensory evaluation methodology has been the application of rapid evaluation techniques. The ease in use makes the techniques extremely easy to implement by industry and university environments. Thus, one might not consider validity in the choice of method. The overall...... aim of this thesis is to compare and evaluate selected rapid evaluation techniques for sensory profiling. Method variations have been suggested for evaluations in product development and quality control, and method insight is provided. The thesis includes three original studies, designed...... as a consequence of the current practices and needs faced in the industry. Study I compared applicability and validity of rapid methods across several panels of trained assessors. Two rapid approaches were introduced for the evaluation of foods. The first method, ‘Free Multiple Sorting’, allows subjects to perform... 15. A System Description of the Cocaine Trade Science.gov (United States) 1994-01-01 72 C.17. Drug Market Hirarchy Tables (Cells A112 to N155) ............ 74 C.18. Purity Levels (Cells A156 to E71...This report also provides detailed information on how to use the model. The spreadsheets are available for either IBM (DOS) or Apple -based machines upon...red square (IBM) or arrow ( Apple ) in the upper right- hand corner have a note "behind" the cell explaining something about the data in the cell, or if 16. Note on Louis de Broglie's duality principle International Nuclear Information System (INIS) Noguez, G. 1988-01-01 For a material corpuscle, the principle of relativity ought to have a more complex expression due to the wave-particle dualism. As a matter of fact, three frame changes are necessary to link descriptions of two objects with uniform relative velocity, as a particle, as a wave, as both wave and particle. Galilean, dual, and Lorentz transformations correspond to these three descriptions [fr 17. GeneNotes – A novel information management software for biologists Directory of Open Access Journals (Sweden) Wong Wing H 2005-02-01 Full Text Available Abstract Background Collecting and managing information is a challenging task in a genome-wide profiling research project. Most databases and online computational tools require a direct human involvement. Information and computational results are presented in various multimedia formats (e.g., text, image, PDF, word files, etc., many of which cannot be automatically processed by computers in biologically meaningful ways. In addition, the quality of computational results is far from perfect and requires nontrivial manual examination. The timely selection, integration and interpretation of heterogeneous biological information still heavily rely on the sensibility of biologists. Biologists often feel overwhelmed by the huge amount of and the great diversity of distributed heterogeneous biological information. Description We developed an information management application called GeneNotes. GeneNotes is the first application that allows users to collect and manage multimedia biological information about genes/ESTs. GeneNotes provides an integrated environment for users to surf the Internet, collect notes for genes/ESTs, and retrieve notes. GeneNotes is supported by a server that integrates gene annotations from many major databases (e.g., HGNC, MGI, etc.. GeneNotes uses the integrated gene annotations to (a identify genes given various types of gene IDs (e.g., RefSeq ID, GenBank ID, etc., and (b provide quick views of genes. GeneNotes is free for academic usage. The program and the tutorials are available at: http://bayes.fas.harvard.edu/genenotes/. Conclusions GeneNotes provides a novel human-computer interface to assist researchers to collect and manage biological information. It also provides a platform for studying how users behave when they manipulate biological information. The results of such study can lead to innovation of more intelligent human-computer interfaces that greatly shorten the cycle of biology research. 18. Ophthalmology patients' interest in online access to clinic notes at three US clinics. Science.gov (United States) Lee, Bryan S; Oster, Natalia V; Chen, Galen Y; Ding, Leona L; Walker, Janice D; Elmore, Joann G 2017-07-01 19. Quality Assurance Requirements and Description International Nuclear Information System (INIS) Ram Murthy 2002-01-01 The Quality Assurance Requirements and Description (QARD) is the principal Quality Assurance (QA) document for the Civilian Radioactive Waste Management Program (Program). It establishes the minimum requirements for the QA program [INTRODUCTION :1p2s (NOT A REQUIREMENT)]. The QARD contains regulatory requirements and program commitments necessary for the development of an effective QA program [INTRODUCTION :1p3s (NOT A REQUIREMENT)]. Implementing documents must be based on, and be consistent with the QARD. The QARD applies to the following: (1) Acceptance of spent nuclear fuel and high-level waste. (2) Transport of spent nuclear fuel and high-level waste. (3) Storage of spent nuclear fuel through receipt of storage cask certification or a facility operating license. (4) Monitored Geologic Repository, including the site characterization activities [Exploratory Studies Facility (ESF) and surface based testing], through receipt of an operating license. (5) High-level waste form development through qualification, production, and acceptance. (6) Characterization of DOE spent nuclear fuel, and conditioning through acceptance of DOE spent nuclear fuel. Section 2.0, Quality Assurance Program, defines in greater detail criteria for determining work subject to the QARD 20. Description of surface systems. Preliminary site description Simpevarp sub area - Version 1.2 Energy Technology Data Exchange (ETDEWEB) Lindborg, Tobias [ed. 2005-03-01 Swedish Nuclear Fuel and Waste Management Co is currently conducting site characterisation in the Simpevarp area. The area is divided into two subareas, the Simpevarp and the Laxemar subarea. The two subareas are surrounded by a common regional model area, the Simpevarp area. This report describes both the regional area and the subareas. This report is an interim version (model version 1.2) of the description of the surface systems at the Simpevarp area, and should be seen as a background report to the site description of the Simpevarp area, version 1.2, SKB-R--05-08. The basis for this description is quality-assured field data available in the SKB SICADA and GIS databases, together with generic data from the literature. The Surface system, here defined as everything above the bedrock, comprises a number of separate disciplines (e.g. hydrology, geology, topography, oceanography and ecology). Each discipline has developed descriptions and models for a number of properties that together represent the site description. The current methodology for developing the surface system description and the integration to ecosystem models is documented in a methodology strategy report SKB-R--03-06. The procedures and guidelines given in that report were followed in this report. Compared with version 1.1 of the surface system description SKB-R--04-25, this report presents considerable additional features, especially in the ecosystem description (Chapter 4) and in the description of the surface hydrology (Section 3.4). A first attempt has also been made to connect the flow of matter (carbon) between the different ecosystems into an overall ecosystem model at a landscape level. A summarised version of this report is also presented in SKB-R--05-08 together with geological-, hydrogeological-, transport properties-, thermal properties-, rock mechanics- and hydrogeochemical descriptions. 1. Description of surface systems. Preliminary site description Simpevarp sub area - Version 1.2 International Nuclear Information System (INIS) Lindborg, Tobias 2005-03-01 Swedish Nuclear Fuel and Waste Management Co is currently conducting site characterisation in the Simpevarp area. The area is divided into two subareas, the Simpevarp and the Laxemar subarea. The two subareas are surrounded by a common regional model area, the Simpevarp area. This report describes both the regional area and the subareas. This report is an interim version (model version 1.2) of the description of the surface systems at the Simpevarp area, and should be seen as a background report to the site description of the Simpevarp area, version 1.2, SKB-R--05-08. The basis for this description is quality-assured field data available in the SKB SICADA and GIS databases, together with generic data from the literature. The Surface system, here defined as everything above the bedrock, comprises a number of separate disciplines (e.g. hydrology, geology, topography, oceanography and ecology). Each discipline has developed descriptions and models for a number of properties that together represent the site description. The current methodology for developing the surface system description and the integration to ecosystem models is documented in a methodology strategy report SKB-R--03-06. The procedures and guidelines given in that report were followed in this report. Compared with version 1.1 of the surface system description SKB-R--04-25, this report presents considerable additional features, especially in the ecosystem description (Chapter 4) and in the description of the surface hydrology (Section 3.4). A first attempt has also been made to connect the flow of matter (carbon) between the different ecosystems into an overall ecosystem model at a landscape level. A summarised version of this report is also presented in SKB-R--05-08 together with geological-, hydrogeological-, transport properties-, thermal properties-, rock mechanics- and hydrogeochemical descriptions 2. An Information Foraging Analysis of Note Taking and Note Sharing While Browsing Campaign Information DEFF Research Database (Denmark) Vatrapu, Ravi; Robertson, Scott 2010-01-01 In this paper, we present an experimental study of political information foraging in the context of e-voting. Participants were observed while searching and browsing the internet for campaign information in a mock-voting situation in three online note-taking conditions: No Notes, Private Notes...... with lack of scent, low value perception, and value depletion of information. Implications for the voter centered design of e-voting portals are discussed.... 3. What happens when patients can see their doctors' note? - the Open Notes movement OpenAIRE Mende, Susan 2017-01-01 Introduction:  The Open Notes movement represents a culture change, enabling patients’ access to their providers’ notes, thereby increasing transparency and patient engagement.Policy context, objective and highlights:  OpenNotes involves allowing patients on-line or hard copy access to their providers’ notes. The one-year initial pilot began in 2010 with twenty thousand patients and one hundred primary care physicians at three medical centers in the United States.  The pilot’s evaluation foun... 4. Detecting clinically relevant new information in clinical notes across specialties and settings. Science.gov (United States) Zhang, Rui; Pakhomov, Serguei V S; Arsoniadis, Elliot G; Lee, Janet T; Wang, Yan; Melton, Genevieve B 2017-07-05 Automated methods for identifying clinically relevant new versus redundant information in electronic health record (EHR) clinical notes is useful for clinicians and researchers involved in patient care and clinical research, respectively. We evaluated methods to automatically identify clinically relevant new information in clinical notes, and compared the quantity of redundant information across specialties and clinical settings. Statistical language models augmented with semantic similarity measures were evaluated as a means to detect and quantify clinically relevant new and redundant information over longitudinal clinical notes for a given patient. A corpus of 591 progress notes over 40 inpatient admissions was annotated for new information longitudinally by physicians to generate a reference standard. Note redundancy between various specialties was evaluated on 71,021 outpatient notes and 64,695 inpatient notes from 500 solid organ transplant patients (April 2015 through August 2015). Our best method achieved at best performance of 0.87 recall, 0.62 precision, and 0.72 F-measure. Addition of semantic similarity metrics compared to baseline improved recall but otherwise resulted in similar performance. While outpatient and inpatient notes had relatively similar levels of high redundancy (61% and 68%, respectively), redundancy differed by author specialty with mean redundancy of 75%, 66%, 57%, and 55% observed in pediatric, internal medicine, psychiatry and surgical notes, respectively. Automated techniques with statistical language models for detecting redundant versus clinically relevant new information in clinical notes do not improve with the addition of semantic similarity measures. While levels of redundancy seem relatively similar in the inpatient and ambulatory settings in the Fairview Health Services, clinical note redundancy appears to vary significantly with different medical specialties. 5. BLT-EC (Breach, Leach Transport, and Equilibrium Chemistry), a finite-element model for assessing the release of radionuclides from low-level waste disposal units: Background, theory, and model description International Nuclear Information System (INIS) MacKinnon, R.J.; Sullivan, T.M.; Simonson, S.A.; Suen, C.J. 1995-08-01 Performance assessment models typically account for the processes of sorption and dissolution-precipitation by using an empirical distribution coefficient, commonly referred to as K d that combines the effects of all chemical reactions between solid and aqueous phases. In recent years, however, there has been an increasing awareness that performance assessments based solely on empirically based K d models may be incomplete, particularly for applications involving radionuclides having sorption and solubility properties that are sensitive to variations in the in-situ chemical environment. To accommodate variations in the in-situ chemical environment, and to assess its impact on radionuclide mobility, it is necessary to model radionuclide release, transport, and chemical processes in a coupled fashion. This modeling has been done and incorporated into the two-dimensional, finite-element, computer code BLT-EC (Breach, Leach, Transport, Equilibrium Chemistry). BLT-EC is capable of predicting container degradation, waste-form leaching, and advective-dispersive, multispecies, solute transport. BLT-EC accounts for retardation directly by modeling the chemical processes of complexation, sorption, dissolution-precipitation, ion-exchange, and oxidation-reduction reactions. In this report we: (1) present a detailed description of the various physical and chemical processes that control the release and migration of radionuclides from shallow land LLW disposal facilities; (2) formulate the mathematical models that represent these processes; (3) outline how these models are incorporated and implemented in BLT-EC; and (4) demonstrate the application of BLT-EC on a set of example problems 6. Polyphonic Piano Transcription with a Note-Based Music Language Model Directory of Open Access Journals (Sweden) Qi Wang 2018-03-01 Full Text Available This paper proposes a note-based music language model (MLM for improving note-level polyphonic piano transcription. The MLM is based on the recurrent structure, which could model the temporal correlations between notes in music sequences. To combine the outputs of the note-based MLM and acoustic model directly, an integrated architecture is adopted in this paper. We also propose an inference algorithm, in which the note-based MLM is used to predict notes at the blank onsets in the thresholding transcription results. The experimental results show that the proposed inference algorithm improves the performance of note-level transcription. We also observe that the combination of the restricted Boltzmann machine (RBM and recurrent structure outperforms a single recurrent neural network (RNN or long short-term memory network (LSTM in modeling the high-dimensional note sequences. Among all the MLMs, LSTM-RBM helps the system yield the best results on all evaluation metrics regardless of the performance of acoustic models. 7. Clinical Note Creation, Binning, and Artificial Intelligence. Science.gov (United States) Deliberato, Rodrigo Octávio; Celi, Leo Anthony; Stone, David J 2017-08-03 The creation of medical notes in software applications poses an intrinsic problem in workflow as the technology inherently intervenes in the processes of collecting and assembling information, as well as the production of a data-driven note that meets both individual and healthcare system requirements. In addition, the note writing applications in currently available electronic health records (EHRs) do not function to support decision making to any substantial degree. We suggest that artificial intelligence (AI) could be utilized to facilitate the workflows of the data collection and assembly processes, as well as to support the development of personalized, yet data-driven assessments and plans. ©Rodrigo Octávio Deliberato, Leo Anthony Celi, David J Stone. Originally published in JMIR Medical Informatics (http://medinform.jmir.org), 03.08.2017. 8. Description of textures by a structural analysis. Science.gov (United States) Tomita, F; Shirai, Y; Tsuji, S 1982-02-01 A structural analysis system for describing natural textures is introduced. The analyzer automatically extracts the texture elements in an input image, measures their properties, classifies them into some distinctive classes (one ground'' class and some figure'' classes), and computes the distributions of the gray level, the shape, and the placement of the texture elements in each class. These descriptions are used for classification of texture images. An analysis-by-synthesis method for evaluating texture analyzers is also presented. We propose a synthesizer which generates a texture image based on the descriptions. By comparing the reconstructed image with the original one, we can see what information is preserved and what is lost in the descriptions. Science.gov (United States) Nazi, Kim M; Turvey, Carolyn L; Klein, Dawn M; Hogan, Timothy P; Woods, Susan S 2015-03-01 10. Xyce parallel electronic simulator release notes. Energy Technology Data Exchange (ETDEWEB) Keiter, Eric R; Hoekstra, Robert John; Mei, Ting; Russo, Thomas V.; Schiek, Richard Louis; Thornquist, Heidi K.; Rankin, Eric Lamont; Coffey, Todd S; Pawlowski, Roger P; Santarelli, Keith R. 2010-05-01 The Xyce Parallel Electronic Simulator has been written to support, in a rigorous manner, the simulation needs of the Sandia National Laboratories electrical designers. Specific requirements include, among others, the ability to solve extremely large circuit problems by supporting large-scale parallel computing platforms, improved numerical performance and object-oriented code design and implementation. The Xyce release notes describe: Hardware and software requirements New features and enhancements Any defects fixed since the last release Current known defects and defect workarounds For up-to-date information not available at the time these notes were produced, please visit the Xyce web page at http://www.cs.sandia.gov/xyce. 11. Application Note: Power Grid Modeling With Xyce. Energy Technology Data Exchange (ETDEWEB) Sholander, Peter E. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States) 2015-06-01 This application note describes how to model steady-state power flows and transient events in electric power grids with the SPICE-compatible Xyce TM Parallel Electronic Simulator developed at Sandia National Labs. This application notes provides a brief tutorial on the basic devices (branches, bus shunts, transformers and generators) found in power grids. The focus is on the features supported and assumptions made by the Xyce models for power grid elements. It then provides a detailed explanation, including working Xyce netlists, for simulating some simple power grid examples such as the IEEE 14-bus test case. 12. Do Simulation-Based Skill Exercises and Post-Encounter Notes Add Additional Value to a Standardized Patient-Based Clinical Skills Examination? Directory of Open Access Journals (Sweden) Michael D. Prislin 2011-01-01 Full Text Available Background. Standardized patient (SP clinical assessments have limited utility in assessing higher-level clinical competencies. This study explores the value of including simulation exercises and postencounter notes in an SP clinical skills examination. Methods. Two exercises involving cardiac auscultation and ophthalmic funduscopy simulations along with written post encounter notes were added to an SP-based performance examination. Descriptive analyses of students' performance and correlations with SP-based performance measures were obtained. Results. Students' abilities to detect abnormalities on physical exam were highly variable. There were no correlations between SP-based and simulation-derived measures of physical examination competency. Limited correlations were found between students' abilities to perform and document physical examinations and their formulation of appropriate differential diagnoses. Conclusions. Clinical simulation exercises add depth to SP-based assessments of performance. Evaluating the content of post encounter notes offers some insight into students' integrative abilities, and this appears to be improved by the addition of simulation-based post encounter skill exercises. However, further refinement of this methodology is needed. 13. Qualitative and quantitative descriptions of glenohumeral motion. Science.gov (United States) Hill, A M; Bull, A M J; Wallace, A L; Johnson, G R 2008-02-01 Joint modelling plays an important role in qualitative and quantitative descriptions of both normal and abnormal joints, as well as predicting outcomes of alterations to joints in orthopaedic practice and research. Contemporary efforts in modelling have focussed upon the major articulations of the lower limb. Well-constrained arthrokinematics can form the basis of manageable kinetic and dynamic mathematical predictions. In order to contain computation of shoulder complex modelling, glenohumeral joint representations in both limited and complete shoulder girdle models have undergone a generic simplification. As such, glenohumeral joint models are often based upon kinematic descriptions of inadequate degrees of freedom (DOF) for clinical purposes and applications. Qualitative descriptions of glenohumeral motion range from the parody of a hinge joint to the complex realism of a spatial joint. In developing a model, a clear idea of intention is required in order to achieve a required application. Clinical applicability of a model requires both descriptive and predictive output potentials, and as such, a high level of validation is required. Without sufficient appreciation of the clinical intention of the arthrokinematic foundation to a model, error is all too easily introduced. Mathematical description of joint motion serves to quantify all relevant clinical parameters. Commonly, both the Euler angle and helical (screw) axis methods have been applied to the glenohumeral joint, although concordance between these methods and classical anatomical appreciation of joint motion is limited, resulting in miscommunication between clinician and engineer. Compounding these inconsistencies in motion quantification is gimbal lock and sequence dependency. 14. 15 CFR Notes Applicable to State... - Notes applicable to State of Understanding related to Medical Equipment: Science.gov (United States) 2010-01-01 ... 15 Commerce and Foreign Trade 2 2010-01-01 2010-01-01 false Notes applicable to State of Understanding related to Medical Equipment: applicable Notes applicable to State of Understanding related to Medical Equipment: Commerce and Foreign Trade Regulations Relating to Commerce and Foreign Trade (Continued) BUREAU OF INDUSTRY AND SECURITY,... 15. Olkiluoto biosphere description 2006 International Nuclear Information System (INIS) Haapanen, R.; Aro, L.; Ilvesniemi, H.; Kareinen, T.; Kirkkala, T.; Mykrae, S.; Turkki, H.; Lahdenperae, A.-M.; Ikonen, A.T.K. 2007-02-01 This report summarises the current knowledge of the biosphere of Olkiluoto, and it is the first Biosphere Description Report. The elements considered were climate, topography, land use, overburden, terrestrial vegetation and fauna and sea flora, fauna and water. The principal aim was to present a synthesis of the present state (now to 2020) and the main features of past evolution of the biosphere at the site using currently available data. The lack of site specific parameters and their importance was discussed. Conceptual ecosystem models are presented for land and sea. Currently available data made it possible to calculate the biomass of the terrestrial vegetation and further convert it to carbon. In the case of terrestrial animals, preliminary figures are given for moose alone due to lack of sitespecific data. For the same reason, the sea ecosystem model was not quantified within this work. The ecosystems on Olkiluoto do not deviate from the surrounding areas. Since mires are few on Olkiluoto, forests are the most important land ecosystem. However, coastal areas are the transition zones between land and sea, and also potential sites for deep groundwater discharge. The major interest concerning aquatic ecosystems was laid on four future lakes potentially developing from the sea due to the land up-lift. Current sea sediments near Olkiluoto are future land areas, and thus very important. Spatially, the forest ecosystems of Olkiluoto are now most comprehensively covered, while the temporal coverage is highest in sea ecosystems. Lack of data is greatest in terrestrial fauna and sea sediments. During this work, the system boundaries were crossed and the use of data over disciplines was started. The data were mostly in agreement, but some discrepancies were detected. To solve these, and to supplement the existing data, some recommendations were given. (orig.) 16. Description of nuclear properties International Nuclear Information System (INIS) Faessler, A. 1991-01-01 The lectures want to give a survey about new developments in the description of nuclei. In a first chapter we try to derive nuclear properties from the basis theory of quantum chromodynamics. This is not rigorously possible. There are still many cracks in the bridge between QCD and nuclear structure. The basic ingredient for nuclear structure calculations is the nucleon-nucleon interaction. We shall discuss the nucleon-nucleon interaction in a quark model. In a further chapter we discuss the way to come from a bare nucleon-nucleon interaction in free space to an effective nucleon-nucleon interaction in a limited model space for nuclear structure calculations. Such nuclear structure calculations can be done as shell model calculations. But they are due to the large number of configurations limited to light nuclei. We discuss possibilities (MONSTER and VAMPIR) to enlarge the model space for medium heavy and heavy nuclei. As the example of the low lying isovector 1 + states we discuss collective models (Bohr - Mottelson Model, interacting Boson Model) with proton and neutron degrees of freedom. The same states can also be described microscopically with the Quasi-Particle Random Phase Approximation (QRPA). We discuss the removal of spurious states in RPA. We also discuss the calculation of form factors and compare with inelastic electron scattering data. Finally we apply QRPA to the double-beta decay. Grand unified models predict, that the neutrino is identical with his antiparticle, that it has a finite mass and a weak right-handed interaction. If these properties are found the standard model of the strong and the electro-weak interaction can not be correct. Presently we can only derive from lower limits of the half lives of neutrinoless double-beta decays upper limits of the neutrino mass and of the right-handedness of the weak interaction and lower limits of the mass of the right handed heavy vector boson, if a specific grand unified model is given. (author) 17. NOTES: issues and technical details with introduction of NOTES into a small general surgery residency program. Science.gov (United States) Kavic, Michael S; Mirza, Brian; Horne, Walter; Moskowitz, Jesse B 2008-01-01 Natural orifice translumenal endoscopic surgery (NOTES) is a development of recent origin. In 2004, Kalloo et al first described NOTES investigation in an animal model. Since then, several investigators have pursued NOTES study in animal survival and nonsurvival models. Our objectives for this project included studying NOTES intervention in a laboratory environment using large animal (swine) models and learning to do so in a safe, controlled manner. Ultimately, we intend to introduce NOTES methodology into our surgical residency training program. The expertise of an experienced laparoscopic surgeon, fellowship-trained laparoendoscopic surgeon, and veterinarian along with a senior surgical resident was utilized to bring the input of several disciplines to this study. The Institutional Animal Care and Use Committee (IACUC) of Northeastern Ohio Universities College of Medicine and Pharmacy (NEOUCOM/COP) approved this study. A series of 5 laboratory sessions using mixed breed farm swine varying in weight from 37 kg to 43.1 kg was planned for the initial phase of NOTES introduction into our residency program. Animals were not kept alive in this investigation. All animals were anesthetized using a standard swine protocol and euthanized following guidelines issued by the American Veterinary Medical Association Panel on Euthanasia. Equipment included a Fujinon EVE endoscope 0.8 cm in diameter with a suction/irrigation channel and one working channel. Initially, a US Endoscopy gastric overtube, 19.5 mm OD and 50 cm in length, was used to facilitate passage of the endoscope. However, this device was found to have insufficient length. Subsequently, commercially available 5/8" diameter clear plastic tubing, 70 cm to 80 cm in length, was adapted for use as an overtube. Standard endoscopic instruments included Boston Scientific biopsy forceps, needle-knife, papillotome, endoscopic clip applier, and Valley Lab electrosurgical unit. A Karl Storz laparoscope and tower were used for 18. EPICS release 3.11.6 specific documentation -- Release notes for EPICS 3.11.6 International Nuclear Information System (INIS) 1994-01-01 These notes cover the following: (1) directions for switching to production APS release R3.11.6; (2) unbundling of channel access clients; (3) access security; (4) channel access additions; synchronous time support; and (5) description of major differences between R3.11.3 and R3.11.6 Also included is a list of new and/or updated documentation for the program 19. Grouping Notes Through NodesThe Functions of Post-It™ Notes in Design Team Cognition DEFF Research Database (Denmark) Dove, Graham; Abildgaard, Sille Julie; Biskjaer, Michael Mose The Post-It™ note is a frequently used, and yet seldom studied, design material. We investigate the functions Post-It™ notes serve when providing cognitive support for creative design team practice. Our investigation considers the ways in which Post-It™ notes function as design externalisations......, both individually and when grouped, and their role in categorisation in semantic long-term memory. To do this, we adopt a multimodal analytical approach focusing on interaction between humans, and between humans and artefacts, alongside language. We discuss in detail examples of four different...... externalisation functions served by Post-It™ notes, and show how these functions are present in complex overlapping combinations rather than being discrete. We then show how the temporal development of Post-It™ note interactions supports categorisation qualities of semantic long-term memory.... 20. Employing Picture Description to Assess the Students' Descriptive Paragraph Writing Directory of Open Access Journals (Sweden) Ida Ayu Mega Cahyani 2018-03-01 Full Text Available Writing is considered as an important skill in learning process which is needed to be mastered by the students. However, in teaching learning process at schools or universities, the assessment of writing skill is not becoming the focus of learning process and the assessment is administered inappropriately. In this present study, the researcher undertook the study which dealt with assessing descriptive paragraph writing ability of the students through picture description by employing an ex post facto as the research design. The present study was intended to answer the research problem dealing with the extent of the students’ achievement of descriptive paragraph writing ability which is assessed through picture description. The samples under the study were 40 students determined by means of random sampling technique with lottery system. The data were collected through administering picture description as the research instrument. The obtained data were analyzed by using norm-reference measure of five standard values. The results of the data analysis showed that there were 67.50% samples of the study were successful in writing descriptive paragraph, while there were 32.50% samples were unsuccessful in writing descriptive paragraph which was assessed by administering picture description test 1. A note on totally normal spaces International Nuclear Information System (INIS) Zougdani, H.K. 1990-10-01 In this note we give the necessary and sufficient condition for a topological space X such that the product space X x Y is totally normal for any (non discrete) metric space Y, and we show that a totally normal p-space need not be a perfectly normal in general, which makes Theorem 2 doubtful. (author). 6 refs 2. SHORT COMMUNICATIONS, NOTES AND REPORTS Record of ... African Journals Online (AJOL) Campbell Murn 50. SHORT COMMUNICATIONS, NOTES AND. REPORTS. Record of Cinereous Vulture Aegypius monachus from Mukundara Hills Tiger Reserve, Rajasthan,. India. Sailaja Nayak1*, Sunny Shah2 and Jimmy Borah2 .... Prey status and abundance in Mukundara Hills Tiger. Reserve. Technical Report, WWF-India. Rao, J.P. ... 3. Botswana - Accrual Accounting Policy Note and Guide OpenAIRE World Bank 2011-01-01 This paper comprises two volumes: volume one, this concise policy and guidance note that deals with the request as outlined, and a volume two which provides more detailed technical guidance on the implementation of International Public Sector accounting Standards (IPSAS) accrual based standards. Recommended reforms directly relevant to this paper include: 1) fiscal reforms on both the reve... 4. Additional notes on the Rubiaceae of Surinam NARCIS (Netherlands) Bremekamp, C.E.B. 1936-01-01 Since the appearance of my „Notes on the Rubiaceae of Surinam” (in Rec. d. Trav. bot. néerl. XXXI, 1934, 248; also in Meded. Bot. Mus. Herb. Utrecht no. 11, 1934) a number of species and varieties new to the flora of that country have come to light. The majority have been collected by Mr. Rombouts 5. CIFSRF Concept Notes Call 2013 Instructions International Development Research Centre (IDRC) Digital Library (Canada) IDRC CRDI Cc: (the other principal investigators and third-party organisations). Subject: CultiAF 2013 Call for Concept Notes Submitted by (names of all principal ..... clerical, accounting, or secretarial help, general office expenses, office rental and utility ... Please provide the details of financial contributions that will be made to the ... 6. Modern systematics, a further botanical note NARCIS (Netherlands) Holttum, R.E. 1948-01-01 As noted in Dr L.G.M. Baas Becking’s Postscript to Mr van Bemmel’s article in Chronica Naturae Vol. 104, part 4, the new systematics has not been entirely neglected by botanists. I would like to put a further botanical vieuwpoint on this subject. Firstly, I suggest that there is no sharp distinction 7. NOTE receives the prestigious ALICE Industrial Award CERN Multimedia 2006-01-01 "NOTE Lund has been given the ALICE Industrial Award due to good co-operation, great capacity for innovation and high quality of work, as a PCB manufacturer in the CERN project ALICE. Only a small number of awards have so far been conferred to a select number of companies." 8. Note from TS/FM Group CERN Document Server 2007-01-01 Renovation work at Restaurant no. 1 Please note that starting from 5 March and for one week, the men's toilets in Bldg. 500 (underground) will be closed because of the renovation work at Restaurant No. 1. The women's toilets in the same place will exceptionally be open to men and women during the same period. We apologize for the inconvenience and we are at your disposal for further information. Shuttle requests Please note that starting from 1 March 2007, shuttle requests: for official visits or bidders' conferences on the CERN site; to/from the airport or the centre of Geneva; for long distances, must be made via [email protected] or by calling 77777. The radio taxi will still be reachable on 76969. Please note that starting from 1st of March, requests for the transport of PCs from your office to the PC store (Bldg. 513) and vice-versa, must be made the 'transport request' form on EDH under the field 'removals'). Mail distribution Please note that starting from 1 March 2007, mail distrib... 9. [Natural Orifice Transluminal Endoscopic Surgery (NOTES)]. Science.gov (United States) Kim, Yong Sik; Kim, Chul Young; Chun, Hoon Jai 2008-03-01 Recently, the field of gastrointestinal endoscopy is developing rapidly. Once limited to the gastroinstestinal lumen, the endoscopic technology is now breaking the barriers and extending its boundary to peritoneal and pleural space. In 2004, Dr. Kalloo, a gastroenterologist, observed intraperitoneal organs of a pig using a conventional endoscope through the stomach wall. Since then, new endoscopic technique of intraperitoneal intervention with transluminal approach named the Natural Orifice Transluminal Endoscopic Surgery or NOTES has been introduced. NOTES reaches the target organ by inserting the endoscope through a natural orifice (e.g. mouth, anus, vagina, urethra) and entering the peritoneal lumen by means of making an incision on the luminal wall. After a series of successful experiences in animal studies, NOTES are now being tried on human subjects. There are still many obstacles to overcome, but bright future for this new technology is expected because of its proposed advantages of less pain, lower complication rate, short recovery time, and scarless access. In this review, we plan to learn about NOTES. 10. Note from the Goods Reception services CERN Multimedia FI Department 2008-01-01 Members of the personnel are invited to take note that only parcels corresponding to official orders or contracts will be handled at CERN. Individuals are not authorised to have private merchandise delivered to them at CERN and private deliveries will not be accepted by the Goods Reception services. Goods Reception Services 11. Interpreters' notes. On the choice of language DEFF Research Database (Denmark) Dam, Helle Vrønning 2004-01-01 This paper reports on a small-scale empirical study on note-taking in consecutive interpreting. As data, the study draws on the notes produced by four subjects while interpreting one Spanish source text consecutively into Danish, on the one hand, and one Danish source text into Spanish, on the ot...... in the interpreting task, i.e. whether it functions as the source or the target language. Drawing on the concept of processing capacity and the Effort Model of consecutive, a tentative explanation of these findings is suggested......., on the other. The aim of the study is to explore what governs conference interpreters' choice of language for their notes. The categories traditionally used to discuss, describe and explain this choice are those of 'source language' and 'target language', and these categories are therefore subject...... to particular scrutiny here. However, somewhat surprisingly, the results of the analyses indicate that the choice of language in note-taking is governed mainly by the status of the language in the interpreters' language combination, i.e. whether it is an A- or a B-language, and much less by its status... 12. European Science Notes. Volume 41, Number 9. Science.gov (United States) 1987-09-01 management system and of the following: Secretaria de la Revista the inference engine. de Psicologia Social , Facultad de Psico- Application of Knowledge-Based...Notes NATO Advanced Research Workshop on Social and Environmental Psychology .................... William D. Crano 519 Applied Artificial Intelligence...An International Journal ................................. William D. Crano 520 New Spanish Journal of Social Psychology ................. William D 13. Lecture notes for Advanced Time Series Analysis DEFF Research Database (Denmark) 1997-01-01 A first version of this notes was used at the lectures in Grenoble, and they are now extended and improved (together with Jan Holst), and used in Ph.D. courses on Advanced Time Series Analysis at IMM and at the Department of Mathematical Statistics, University of Lund, 1994, 1997, ... 14. Note Taking in Multi-Media Settings Science.gov (United States) Black, Kelly; Yao, Guangming 2014-01-01 We provide a preliminary exploration into the use of note taking when combined with video examples. Student volunteers were divided into three groups and asked to perform two problems. The first problem was explored in a classroom setting and the other problem was a novel problem. The students were asked to complete the two questions. Furthermore,… 15. Biomarkers of genotoxicity of urban air pollution. Overview and descriptive data from a molecular epidemiology study on populations exposed to moderate-to-low levels of polycyclic aromatic hydrocarbons: the AULIS project DEFF Research Database (Denmark) Kyrtopoulos, S.A.; Georgiadis, P.; Autrup, H. 2001-01-01 studies may serve as complementary tools providing a better understanding of the relative contribution of ambient atmospheric pollution to the overall genotoxic burden suffered by city dwellers. However, past efforts to apply biomarkers to studies of low levels exposure to urban air pollution have given......-based population studies to aid the assessment of the genotoxic and carcinogenic effects of urban air pollution is reviewed by reference to the achievements and limitations of earlier reported studies. The design and methodology adopted in a recently completed large-scale population study, carried out......Epidemiologic studies indicate that prolonged exposure to high pollution levels is associated with increased risk of cancer, especially lung cancer. However, under conditions of moderate or low air pollution, epidemiologic evidence does not permit reliable conclusions. Biomarker-based population... 16. The impact of note taking style and note availability at retrieval on mock jurors' recall and recognition of trial information. Science.gov (United States) Thorley, Craig; Baxter, Rebecca E; Lorek, Joanna 2016-01-01 Jurors forget critical trial information and what they do recall can be inaccurate. Jurors' recall of trial information can be enhanced by permitting them to take notes during a trial onto blank sheets of paper (henceforth called freestyle note taking). A recent innovation is the trial-ordered-notebook (TON) for jurors, which is a notebook containing headings outlining the trial proceedings and which has space beneath each heading for notes. In a direct comparison, TON note takers recalled more trial information than freestyle note takers. This study investigated whether or not note taking improves recall as a result of enhanced encoding or as a result of note access at retrieval. To assess this, mock jurors watched and freely recalled a trial video with one-fifth taking no notes, two-fifths taking freestyle notes and two-fifths using TONs. During retrieval, half of the freestyle and TON note takers could access their notes. Note taking enhanced recall, with the freestyle note takers and TON note takers without note access performing equally as well. Note taking therefore enhances encoding. Recall was greatest for the TON note takers with note access, suggesting a retrieval enhancement unique to this condition. The theoretical and applied implications of these findings are discussed. 17. The Danish contribution to the European DEMOCOPHES project: A description of cadmium, cotinine and mercury levels in Danish mother-child pairs and the perspectives of supplementary sampling and measurements Energy Technology Data Exchange (ETDEWEB) Mørck, Thit A. [Department of Public Health, University of Copenhagen, Copenhagen (Denmark); Nielsen, Flemming [Department of Public Health, University of Southern Denmark, Odense (Denmark); Nielsen, Jeanette K.S.; Jensen, Janne F.; Hansen, Pernille W.; Hansen, Anne K.; Christoffersen, Lea N. [Department of Public Health, University of Copenhagen, Copenhagen (Denmark); Siersma, Volkert D. [The Research Unit for General Practice and Section of General Practice, Department of Public Health, University of Copenhagen, Copenhagen (Denmark); Larsen, Ida H.; Hohlmann, Linette K. [Department of Public Health, University of Copenhagen, Copenhagen (Denmark); Skaanild, Mette T. [Danish Environmental Protection Agency (Denmark); Frederiksen, Hanne [Department of Growth and Reproduction, University Hospital, Copenhagen (Denmark); Biot, Pierre [Federal Public Service Health, Food Chain Safety and Environment, Brussels (Belgium); Casteleyn, Ludwine [University of Leuven, Leuven (Belgium); Kolossa-Gehring, Marike; Schwedler, Gerda [Federal Environment Agency (UBA), Berlin (Germany); Castaño, Argelia [Instituto de Salud Carlos III (ISCIII), Majadahonda, Madrid (Spain); Angerer, Jürgen; Koch, Holger M. [Institute for Prevention and Occupational Medicine of the German social Accident Insurance, Institute of the Ruhr-Universität Bochum (IPA), Bochum (Germany); Esteban, Marta [Instituto de Salud Carlos III (ISCIII), Majadahonda, Madrid (Spain); and others 2015-08-15 Human biomonitoring (HBM) is an important tool, increasingly used for measuring true levels of the body burdens of environmental chemicals in the general population. In Europe, a harmonized HBM program was needed to open the possibility to compare levels across borders. To explore the prospect of a harmonized European HBM project, DEMOCOPHES (DEMOnstration of a study to COordinate and Perform Human biomonitoring on a European Scale) was completed in 17 European countries. The basic measurements performed in all implemented countries of DEMOCOPHES included cadmium, cotinine and phthalate metabolites in urine and mercury in hair. In the Danish participants, significant correlations between mothers and children for mercury in hair and cotinine in urine were found. Mercury in hair was further significantly associated with intake of fish and area of residence. Cadmium was positively associated with BMI in mothers and an association between cadmium and cotinine was also found. As expected high cotinine levels were found in smoking mothers. For both mercury and cadmium significantly higher concentrations were found in the mothers compared to their children. In Denmark, the DEMOCOPHES project was co-financed by the Danish ministries of health, environment and food safety. The co-financing ministries agreed to finance a number of supplementary measurements of substances of current toxicological, public and regulatory interest. This also included blood sampling from the participants. The collected urine and blood samples were analyzed for a range of other persistent and non-persistent environmental chemicals as well as two biomarkers of effect. The variety of supplementary measurements gives the researchers further information on the exposure status of the participants and creates a basis for valuable knowledge on the pattern of exposure to various chemicals. - Highlights: • Levels of cadmium, mercury and cotinine in the Danish subpopulation are comparable to levels in the 18. Feedback produces divergence from prospect theory in descriptive choice. Science.gov (United States) Jessup, Ryan K; Bishara, Anthony J; Busemeyer, Jerome R 2008-10-01 A recent study demonstrated that individuals making experience-based choices underweight small probabilities, in contrast to the overweighting observed in a typical descriptive paradigm. We tested whether trial-by-trial feedback in a repeated descriptive paradigm would engender choices more correspondent with experiential or descriptive paradigms. The results of a repeated gambling task indicated that individuals receiving feedback underweighted small probabilities, relative to their no-feedback counterparts. These results implicate feedback as a critical component during the decision-making process, even in the presence of fully specified descriptive information. A model comparison at the individual-subject level suggested that feedback drove individuals' decision weights toward objective probability weighting. 19. Dual descriptions of supersymmetry breaking International Nuclear Information System (INIS) Intrilligator, K.; Thomas, S. 1996-08-01 Dynamical supersymmetry breaking is considered in models which admit descriptions in terms of electric, confined, or magnetic degrees of freedom in various limits. In this way, a variety of seemingly different theories which break supersymmetry are actually interrelated by confinement or duality. Specific examples are given in which there are two dual descriptions of the supersymmetry breaking ground state 20. The Danish contribution to the European DEMOCOPHES project: A description of cadmium, cotinine and mercury levels in Danish mother-child pairs and the perspectives of supplementary sampling and measurements. Science.gov (United States) Mørck, Thit A; Nielsen, Flemming; Nielsen, Jeanette K S; Jensen, Janne F; Hansen, Pernille W; Hansen, Anne K; Christoffersen, Lea N; Siersma, Volkert D; Larsen, Ida H; Hohlmann, Linette K; Skaanild, Mette T; Frederiksen, Hanne; Biot, Pierre; Casteleyn, Ludwine; Kolossa-Gehring, Marike; Schwedler, Gerda; Castaño, Argelia; Angerer, Jürgen; Koch, Holger M; Esteban, Marta; Schoeters, Greet; Den Hond, Elly; Exley, Karen; Sepai, Ovnair; Bloemen, Louis; Joas, Reinhard; Joas, Anke; Fiddicke, Ulrike; Lopez, Ana; Cañas, Ana; Aerts, Dominique; Knudsen, Lisbeth E 2015-08-01 Human biomonitoring (HBM) is an important tool, increasingly used for measuring true levels of the body burdens of environmental chemicals in the general population. In Europe, a harmonized HBM program was needed to open the possibility to compare levels across borders. To explore the prospect of a harmonized European HBM project, DEMOCOPHES (DEMOnstration of a study to COordinate and Perform Human biomonitoring on a European Scale) was completed in 17 European countries. The basic measurements performed in all implemented countries of DEMOCOPHES included cadmium, cotinine and phthalate metabolites in urine and mercury in hair. In the Danish participants, significant correlations between mothers and children for mercury in hair and cotinine in urine were found. Mercury in hair was further significantly associated with intake of fish and area of residence. Cadmium was positively associated with BMI in mothers and an association between cadmium and cotinine was also found. As expected high cotinine levels were found in smoking mothers. For both mercury and cadmium significantly higher concentrations were found in the mothers compared to their children. In Denmark, the DEMOCOPHES project was co-financed by the Danish ministries of health, environment and food safety. The co-financing ministries agreed to finance a number of supplementary measurements of substances of current toxicological, public and regulatory interest. This also included blood sampling from the participants. The collected urine and blood samples were analyzed for a range of other persistent and non-persistent environmental chemicals as well as two biomarkers of effect. The variety of supplementary measurements gives the researchers further information on the exposure status of the participants and creates a basis for valuable knowledge on the pattern of exposure to various chemicals. Copyright © 2014 Elsevier Inc. All rights reserved. 1. Total Vaginal NOTES Hysterectomy: A New Approach to Hysterectomy. Science.gov (United States) Baekelandt, Jan 2015-01-01 2. Towards reproducible descriptions of neuronal network models. Directory of Open Access Journals (Sweden) Eilen Nordlie 2009-08-01 Full Text Available Progress in science depends on the effective exchange of ideas among scientists. New ideas can be assessed and criticized in a meaningful manner only if they are formulated precisely. This applies to simulation studies as well as to experiments and theories. But after more than 50 years of neuronal network simulations, we still lack a clear and common understanding of the role of computational models in neuroscience as well as established practices for describing network models in publications. This hinders the critical evaluation of network models as well as their re-use. We analyze here 14 research papers proposing neuronal network models of different complexity and find widely varying approaches to model descriptions, with regard to both the means of description and the ordering and placement of material. We further observe great variation in the graphical representation of networks and the notation used in equations. Based on our observations, we propose a good model description practice, composed of guidelines for the organization of publications, a checklist for model descriptions, templates for tables presenting model structure, and guidelines for diagrams of networks. The main purpose of this good practice is to trigger a debate about the communication of neuronal network models in a manner comprehensible to humans, as opposed to machine-readable model description languages. We believe that the good model description practice proposed here, together with a number of other recent initiatives on data-, model-, and software-sharing, may lead to a deeper and more fruitful exchange of ideas among computational neuroscientists in years to come. We further hope that work on standardized ways of describing--and thinking about--complex neuronal networks will lead the scientific community to a clearer understanding of high-level concepts in network dynamics, and will thus lead to deeper insights into the function of the brain. 3. Notes on Two Newly Naturalized Plants in Taiwan: Evolvulus nummularius (L. L. (Convolvulaceae and Acalypha aristata Kunth (Euphorbiaceae Directory of Open Access Journals (Sweden) Shih-Huei Chen 2009-09-01 Full Text Available Evolvulus nummularius (L. L. and Acalypha aristata Kunth, originally native to tropical America, were recently found naturalized in disturbed sites of Taiwan. The present study gives the taxonomic description and line drawings of the two species. In addition, their distributions and notes on ecology are provided. 4. Calculation notes that support accident scenario and consequence of the in-tank fuel fire/deflageration International Nuclear Information System (INIS) Crowe, R.D. Westinghouse Hanford 1996-01-01 The purpose of this calculation note is to provide the basis for In-Tank Fuel fire/Deflageration consequence for the Tank Farm Safety Analysis Report (FSAR). Tank Fuel Fire/Deflageration scenario is developed and details and description of the analysis methods are provided 5. Calculation notes that support accident scenario and consequence of the in-tank fuel fire/deflagration Energy Technology Data Exchange (ETDEWEB) Crowe, R.D. 1996-09-27 The purpose of this calculation note is to provide the basis for In-Tank Fuel Fire/Deflageration consequence for the Tank Farm Safety Analysis Report (FSAR). Tank Fuel Fire/Deflageration scenario is developed and details and description of the analysis methods are provided. 6. Content of system design descriptions International Nuclear Information System (INIS) 1998-10-01 A System Design Description (SDD) describes the requirements and features of a system. This standard provides guidance on the expected technical content of SDDs. The need for such a standard was recognized during efforts to develop SDDs for safety systems at DOE Hazard Category 2 nonreactor nuclear facilities. Existing guidance related to the corresponding documents in other industries is generally not suitable to meet the needs of DOE nuclear facilities. Across the DOE complex, different contractors have guidance documents, but they vary widely from site to site. While such guidance documents are valuable, no single guidance document has all the attributes that DOE considers important, including a reasonable degree of consistency or standardization. This standard is a consolidation of the best of the existing guidance. This standard has been developed with a technical content and level of detail intended to be most applicable to safety systems at DOE Hazard Category 2 nonreactor nuclear facilities. Notwithstanding that primary intent, this standard is recommended for other systems at such facilities, especially those that are important to achieving the programmatic mission of the facility. In addition, application of this standard should be considered for systems at other facilities, including non-nuclear facilities, on the basis that SDDs may be beneficial and cost-effective 7. Using Guided Notes to Enhance Instruction for All Students Science.gov (United States) 2011-01-01 Taking notes from lectures or reading material can be challenging, especially for those who have learning disabilities. An alternative to traditional note-taking is a method called "guided notes," which has been found to improve the accuracy of students' notes, increase the frequency of student responses, and improve students' quiz and test… 8. DigiMemo: Facilitating the Note Taking Process Science.gov (United States) Kurt, Serhat 2009-01-01 Everyone takes notes daily for various reasons. Note taking is very popular in school settings and generally recognized as an effective learning strategy. Further, note taking is a complex process because it requires understanding, selection of information and writing. Some new technological tools may facilitate the note taking process. Among such… 9. Waste Management Systems Requirements and Descriptions (SRD) International Nuclear Information System (INIS) Conner, C.W. 1986-01-01 The Department of Energy (DOE), Office of Civilian Radioactive Waste Management (OCRWM) is responsible for the development of a system for the management of high-level radioactive waste and spent fuel in accordance with the Nuclear Waste Policy Act of 1982. The Waste Management system requirements and description document is the program-level technical baseline document. The requirements include the functions that must be performed in order to achieve the system mission and performance criteria for those functions. This document covers only the functional requirements of the system; it does not cover programmatic or procedural requirements pertaining to the processes of designing, siting and licensing. The requirements are largely based on the Nuclear Waste Policy Act of 1982, Environmental Protection Agency standards, Nuclear Regulatory Commission regulations, and DOE orders and guidance. However, nothing in this document should be construed as to relieve the DOE or its contractors from their responsibilities to comply with applicable statutes, regulations, and standards. This document also provides a brief description of the system being developed to meet the requirements. In addition to the described ''authorized system,'' a system description is provided for an ''improved-performance system'' which would include a monitored retrievable storage (MRS) facility. In the event that an MRS facility is approved by Congress, the improved-performance system will become the reference system. Neither system description includes Federal Interim Storage (FIS) capabilities. Should the need for FIS be identified, it will be included as an additional system element. The descriptions are focused on the interfaces between the system elements, rather than on the detail of the system elements themselves 10. Justifications on social ineaqualities. Notes on the Salvadoran case Directory of Open Access Journals (Sweden) Irene Lungo Rodríguez 2016-05-01 Full Text Available This note presents the theoretical and methodological guidelines of my phd dissertation, which is under preparation. My concern is about cultural aspects that underlie social inequalities in contemporary El Salvador. Currently, this country is characterized by the primacy of neoliberal acumulation model, the rise of an unprecedented democracy, processes of reconcentration of capital and the persistence of high levels of social exclusion. Specifically, I study the justification of inequalities from the case of the “elite” middle classes in El Salvador. This research is inscribed whitin the sociological debates on the cultural aspects after reproduction of stratification and inequality. 11. Writing and Music: Album Liner Notes Directory of Open Access Journals (Sweden) Dean Leonard Biron 2011-09-01 Full Text Available A deceptive aspect of the ‘writing about music is like dancing about architecture’ cliché is the function of the preposition ‘about’. Literature and music, dance and architecture, painting and film – all are discrete aesthetic forms that nonetheless simultaneously feed off and provide nourishment for each other as part of art’s perpetual drive toward diversity and innovation. Nowhere is the aptness of the association between writing and music more obvious than in the phenomenon of album liner notes. Rather than merely an attempt at describing or translating musical experience, liner notes contribute to the dialogue between composer and listener and are a significant part of the culture of contemporary music. 12. Flatland an edition with notes and commentary CERN Document Server Abbott, Edwin A; Banchoff, Thomas F 2010-01-01 Flatland, Edwin Abbott Abbott's story of a two-dimensional universe, as told by one of its inhabitants who is introduced to the mysteries of three-dimensional space, has enjoyed an enduring popularity from the time of its publication in 1884. This fully annotated edition enables the modern-day reader to understand and appreciate the many "dimensions" of this classic satire. Mathematical notes and illustrations enhance the usefulness of Flatland as an elementary introduction to higher-dimensional geometry. Historical notes show connections to late-Victorian England and to classical Greece. Citations from Abbott's other writings as well as the works of Plato and Aristotle serve to interpret the text. Commentary on language and literary style includes numerous definitions of obscure words. An appendix gives a comprehensive account of the life and work of Flatland's remarkable author. 13. Teaching Furlow palatoplasty: the sticky note method. Science.gov (United States) Liu, Mona Mengyue; Kim, JeeHong; Jabbour, Noel 2014-11-01 The double-opposing Z-plasty (Furlow palatoplasty) procedure is a well-established method for palate repair in children. We propose a simple and easily accessible sticky note model to demonstrate the lengthening in palatal anatomy afforded by this technically challenging procedure. Our model involves creating a lengthened three-dimensional representation of the Z-plasty through making specified incisions and rearrangements of the palatal layers. The sticky note model was made a total of 20 times and length of the palate model pre and post Z-plasty was measured. The average length of the palate pre-procedure was 72 mm. The average length of the palate post procedure was 78.9 mm, showing an increase of 6.9 mm (9.6%). Our model provides an accurate and valuable educational tool that will aid in the visualization and understanding of the Furlow palatoplasty procedure. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved. 14. Molecular gastronomy is a scientific discipline, and note by note cuisine is the next culinary trend Directory of Open Access Journals (Sweden) This Hervé 2013-01-01 Full Text Available Abstract For the past two decades, there has been much confusion about molecular gastronomy. This confusion has arisen because people ignore that the word gastronomy does not mean cuisine, it means knowledge about food. Similar to ‘molecular biology’, molecular gastronomy is a scientific discipline that looks for the mechanisms of phenomena occurring during dish preparation and consumption. As with any other scientific discipline, it can have many applications. One of the first was ‘molecular cuisine’ but since 1994, ‘note by note cuisine’ has also been promoted. The latter involves preparing dishes using pure compounds, or more practically mixture of compounds obtained by fractioning plant or animal tissues, instead of using these tissues themselves. Note by note cuisine raises issues in various fields: science, technology, nutrition, physiology, toxicology and politics. 15. Notes on HTR applications in methanol production International Nuclear Information System (INIS) Santoso, B.; Barnert, H. 1997-01-01 Notes on the study of HTR applications are presented. The study in particular should be directed toward the most feasible applications of HTR for process heat generation. A prospective study is the conversion of CO 2 gas from Natuna to methanol or formic acid. Further studies needs to be deepened under the auspices of IAEA and countries that have similar interest. (author). 3 refs, 3 figs 16. Mission climate note n.7, November 2005 International Nuclear Information System (INIS) Dominicis, A. de 2005-01-01 The year 2005 was the year of the implementing of the CO 2 quotas market and the implementing of the kyoto protocol after the ratification by Russia. This context confirmed the presence of a 'carbon finance' based on new financial products: the CO 2 actives or carbon credits which materialize the emissions reductions and exist on the new markets. This note presents the main points of these markets: the purchase funds implementing, the investments and the purchase mechanisms management. (A.L.B.) 17. Note: How Does Product Proliferation Affect Responsiveness? OpenAIRE Diwakar Gupta; Mandyam M. Srinivasan 1998-01-01 In this note we consider some strategies that a manufacturing firm may use to deal with an increase in the variety of products it offers. We indicate how alternate strategies for dealing with product proliferation impact the firm's responsiveness, measured in terms of average production lead time and average work-in-process inventory. Focusing on the make-to-order environment and using queueing models, we derive conditions under which an increase in product variety can improve both individual... 18. European Scientific Notes. Volume 36, Number 2, Science.gov (United States) 1982-02-28 noted that the alleged Symptom-limited exercise testing is gen- immunity of marathon runners to coronary erally accepted as an integral part of any...marathon runners out of coronary patients. risk, and the motivation pf both physician and Wenger emphasized that there is no fixed dose patient...never 45 ESN 36-2 (1982) RESCUE KITE The accuracy of addition of columns of 2-digit numbers was significantly worse at In emergencies at sea, detection 19. A note on Ramsey numbers for fans NARCIS (Netherlands) Zhang, Yanbo; Broersma, Haitze J.; Chen, Yaojun For two given graphs G1 and G2, the Ramsey number R(G1,G2) is the smallest integer N such that, for any graph G of order N, either G contains G1 as a subgraph or the complement of G contains G2 as a subgraph. A fan Fl is l triangles sharing exactly one vertex. In this note, it is shown that R(Fn, 20. AREST model description International Nuclear Information System (INIS) Engel, D.W.; McGrail, B.P. 1993-11-01 The Office of Civilian Radioactive Waste Management and the Power Reactor and Nuclear Fuel Development Corporation of Japan (PNC) have supported the development of the Analytical Repository Source-Term (AREST) at Pacific Northwest Laboratory. AREST is a computer model developed to evaluate radionuclide release from an underground geologic repository. The AREST code can be used to calculate/estimate the amount and rate of each radionuclide that is released from the engineered barrier system (EBS) of the repository. The EBS is the man-made or disrupted area of the repository. AREST was designed as a system-level models to simulate the behavior of the total repository by combining process-level models for the release from an individual waste package or container. AREST contains primarily analytical models for calculating the release/transport of radionuclides to the lost rock that surrounds each waste package. Analytical models were used because of the small computational overhead that allows all the input parameters to be derived from a statistical distribution. Recently, a one-dimensional numerical model was also incorporated into AREST, to allow for more detailed modeling of the transport process with arbitrary length decay chains. The next step in modeling the EBS, is to develop a model that couples the probabilistic capabilities of AREST with a more detailed process model. This model will need to look at the reactive coupling of the processes that are involved with the release process. Such coupling would include: (1) the dissolution of the waste form, (2) the geochemical modeling of the groundwater, (3) the corrosion of the container overpacking, and (4) the backfill material, just to name a few. Several of these coupled processes are already incorporated in the current version of AREST 1. Notes on two species of Processa (Decapoda: Processidae from the Mexican Pacific Directory of Open Access Journals (Sweden) Manuel Ayón-Parente Full Text Available Material belonging to the genus Processa, held in the Regional Marine Invertebrates Collection in Mazatlán, Mexico, is revised including specimens of the widely distributed P. peruviana, and the scarcely collected P. hawaiensis. For comparative purposes with the specimens from Hawaii and other localities, a detailed description of a male of P. hawaiensis collected in continental Mexico is provided, including illustrations of all appendages. Small differences are noted with previous description and partial redescriptions of this species, including proportion between propodus and dactylus of the fourth pereopod, and between merus and carpus of the right cheliped. In addition, the shape and setation of the first pair of pleopods in the Mexican material differs from the description of P. hawaiensis based on African material. 2. Energy and greenhouse effect. Twelve short notes International Nuclear Information System (INIS) Prevot, Henri 2013-12-01 The author proposes twelve brief notes aimed at discussing the reduction of fossil energy consumption in order to reduce CO 2 emissions and to improve the French energy supply security, without any useless expense. These notes address the reason for energy savings, the cost and price of a CO 2 ton, the issue of thermal regulation for buildings (it's not in compliance with the law, and results in higher expenses and increased CO 2 emissions), the introduction of a carbon tax to incite investments for energy saving, the status and health of the CO 2 European market, the support of actions aimed at reducing fossil energy consumption, the fact that bio-heat is ten times more efficient than bio-fuel and that therefore car holders should finance bio-heat, the development of hybrid uses of energy to avoid the difficulty of energy storage, the reduction of CO 2 emissions at low cost (by consuming as much renewable energy as nuclear energy but without wind or photovoltaic energy), the cost of less CO 2 , less fossil energy and less nuclear, and the interest of France to act on its own to reduce CO 2 emissions. The author proposes a brief synthesis of these notes and some proposals regarding thermal regulation for buildings, taxes, the European CO 2 market, the forest biomass, electricity production, and the European and word dimensions of these issues 3. Core Benchmarks Descriptions International Nuclear Information System (INIS) Pavlovichev, A.M. 2001-01-01 Actual regulations while designing of new fuel cycles for nuclear power installations comprise a calculational justification to be performed by certified computer codes. It guarantees that obtained calculational results will be within the limits of declared uncertainties that are indicated in a certificate issued by Gosatomnadzor of Russian Federation (GAN) and concerning a corresponding computer code. A formal justification of declared uncertainties is the comparison of calculational results obtained by a commercial code with the results of experiments or of calculational tests that are calculated with an uncertainty defined by certified precision codes of MCU type or of other one. The actual level of international cooperation provides an enlarging of the bank of experimental and calculational benchmarks acceptable for a certification of commercial codes that are being used for a design of fuel loadings with MOX fuel. In particular, the work is practically finished on the forming of calculational benchmarks list for a certification of code TVS-M as applied to MOX fuel assembly calculations. The results on these activities are presented 4. Notes on Soho and a Reminiscence. Science.gov (United States) Melton, Hollis 1992-01-01 Discusses George Maciunas' pivotal contributions to the renaissance of SoHo, the New York City community south of Houston Street. Recounts the establishment of Fluxus cooperatives, the history of the FilmMakers' Cinematheque, Maciunas' long struggle with the Attorney General's office, and closes with a description of the February 1978 erotic Flux… 5. Notes on Caribbean Phymatidae (Hemiptera: Heteroptera) NARCIS (Netherlands) Kormilev, N.A.; Doesburg, van P.H. 1991-01-01 New descriptive and distributional data are presented for Ambush bugs from the Dominican Republic: Phymatocoris iviei gen. nov., spec. nov. and Lophoscutus hispaniolensis spec. nov. New records are given for Phymata interjecta Dudich, 1922, Lophoscutus confusus Kormilev, 1989, and L. ypsilon 6. 1993 baseline solid waste management system description International Nuclear Information System (INIS) Armacost, L.L.; Fowler, R.A.; Konynenbelt, H.S. 1994-02-01 Pacific Northwest Laboratory has prepared this report under the direction of Westinghouse Hanford Company. The report provides an integrated description of the system planned for managing Hanford's solid low-level waste, low-level mixed waste, transuranic waste, and transuranic mixed waste. The primary purpose of this document is to illustrate a collective view of the key functions planned at the Hanford Site to handle existing waste inventories, as well as solid wastes that will be generated in the future. By viewing this system as a whole rather than as individual projects, key facility interactions and requirements are identified and a better understanding of the overall system may be gained. The system is described so as to form a basis for modeling the system at various levels of detail. Model results provide insight into issues such as facility capacity requirements, alternative system operating strategies, and impacts of system changes (ie., startup dates). This description of the planned Hanford solid waste processing system: defines a baseline system configuration; identifies the entering waste streams to be managed within the system; identifies basic system functions and waste flows; and highlights system constraints. This system description will evolve and be revised as issues are resolved, planning decisions are made, additional data are collected, and assumptions are tested and changed. Out of necessity, this document will also be revised and updated so that a documented system description, which reflects current system planning, is always available for use by engineers and managers. It does not provide any results generated from the many alternatives that will be modeled in the course of analyzing solid waste disposal options; such results will be provided in separate documents 7. Health Topic XML File Description Science.gov (United States) ... this page: https://medlineplus.gov/xmldescription.html Health Topic XML File Description: MedlinePlus To use the sharing ... information categories assigned. Example of a Full Health Topic Record A record for a MedlinePlus health topic ... 8. Understanding Digital Note-Taking Practice for Visualization. Science.gov (United States) Willett, Wesley; Goffin, Pascal; Isenberg, Petra 2015-05-13 We present results and design implications from a study of digital note-taking practice to examine how visualization can support revisitation, reflection, and collaboration around notes. As digital notebooks become common forms of external memory, keeping track of volumes of content is increasingly difficult. Information visualization tools can help give note-takers an overview of their content and allow them to explore diverse sets of notes, find and organize related content, and compare their notes with their collaborators. To ground the design of such tools, we conducted a detailed mixed-methods study of digital note-taking practice. We identify a variety of different editing, organization, and sharing methods used by digital note-takers, many of which result in notes becoming "lost in the pile''. These findings form the basis for our design considerations that examine how visualization can support the revisitation, organization, and sharing of digital notes. 9. A note on BPS vortex bound states Directory of Open Access Journals (Sweden) A. Alonso-Izquierdo 2016-02-01 Full Text Available In this note we investigate bound states, where scalar and vector bosons are trapped by BPS vortices in the Abelian Higgs model with a critical ratio of the couplings. A class of internal modes of fluctuation around cylindrically symmetric BPS vortices is characterized mathematically, analyzing the spectrum of the second-order fluctuation operator when the Higgs and vector boson masses are equal. A few of these bound states with low values of quantized magnetic flux are described fully, and their main properties are discussed. 10. A note on BPS vortex bound states Energy Technology Data Exchange (ETDEWEB) Alonso-Izquierdo, A., E-mail: [email protected] [Departamento de Matematica Aplicada, Universidad de Salamanca (Spain); Garcia Fuertes, W., E-mail: [email protected] [Departamento de Fisica, Universidad de Oviedo (Spain); Mateos Guilarte, J., E-mail: [email protected] [Departamento de Fisica Fundamental, Universidad de Salamanca (Spain) 2016-02-10 In this note we investigate bound states, where scalar and vector bosons are trapped by BPS vortices in the Abelian Higgs model with a critical ratio of the couplings. A class of internal modes of fluctuation around cylindrically symmetric BPS vortices is characterized mathematically, analyzing the spectrum of the second-order fluctuation operator when the Higgs and vector boson masses are equal. A few of these bound states with low values of quantized magnetic flux are described fully, and their main properties are discussed. 11. Notes on human error analysis and prediction International Nuclear Information System (INIS) Rasmussen, J. 1978-11-01 The notes comprise an introductory discussion of the role of human error analysis and prediction in industrial risk analysis. Following this introduction, different classes of human errors and role in industrial systems are mentioned. Problems related to the prediction of human behaviour in reliability and safety analysis are formulated and ''criteria for analyzability'' which must be met by industrial systems so that a systematic analysis can be performed are suggested. The appendices contain illustrative case stories and a review of human error reports for the task of equipment calibration and testing as found in the US Licensee Event Reports. (author) 12. A note on the microeconomics of migration. Science.gov (United States) Stahl, K 1983-11-01 "The purpose of this note is to demonstrate in a simple model that an individual's migration from a small town to a large city may be rationalized purely by a consumption motive, rather than the motive of obtaining a higher income. More specifically, it is shown that in a large city an individual may derive a higher utility from spending a given amount of income than in a small town." A formal model is first developed that includes the principal forces at work and is then illustrated using a graphic example. The theoretical and empirical issues raised are considered in the concluding section. excerpt 13. A note on oblique water entry KAUST Repository Moore, M. R. 2012-10-02 A minor error in Howison et al. (J. Eng. Math. 48:321-337, 2004) obscured the fact that the points at which the free surface turns over in the solution of the Wagner model for the oblique impact of a two-dimensional body are directly related to the turnover points in the equivalent normal impact problem. This note corrects some of the earlier results given in Howison et al. (J. Eng. Math. 48:321-337, 2004) and discusses the implications for the applicability of the Wagner model. © 2012 Springer Science+Business Media B.V. 14. European Science Notes. Volume 40, Number 1. Science.gov (United States) 1986-01-01 Mass Spectrometry mers and copolymers of polyacrylate salt series edited by Professor J.F.J. Todd latex) rather than an inorganic or or- (University...changes in the popu- cy with two potassium dihydrogen phos- lation of a vibrational manifold were phate (KDP) crystals. Following a fil- determined by...AD-A162 235 EUROPEAN SCIENCE NOTES VOLUME 48 NUMBER I(U) OFFICE OF i/1 NAVAL RESEARCH LONDON (ENGLAND) L E SHAFFER JAN 86 UNCLASSIFIED F/G 5/2 DEFF Research Database (Denmark) Landex, Alex; Andersen, Jonas Lohmann Elkjær; Salling, Kim Bang 2006-01-01 these effects must be described qualitatively. This note describes the socio-economic evaluation based on market prices and not factor prices which has been the tradition in Denmark till now. This is due to the recommendation from the Ministry of Transport to start using calculations based on market prices...... alternative. In socio-economic evaluations it is intended to describe the effects in economic terms whenever possible (”+” is used when it is positive for the society, and ”–” when it is negative for the society). However, not all the effects for the society can be described in economic terms, and instead... 16. A note on Jauch-Piron states Science.gov (United States) Pták, Pavel 1986-04-01 A state s on a logic L is called Jauch-Piron if the equality s( a)= s( b)=1 for a,b∈ L implies the existence of such an element c∈ L that c⩽ a, c⩽ b and s( c)=1. If L is a lattice then the present definition coincides with Jauch-Piron states as defined in [7], [9] and [13]. A logic is called Jauch-Piron if its every state is Jauch-Piron. In this note we investigate the size of the class of Jauch-Piron logics. 17. Notes on Timed Concurrent Constraint Programming DEFF Research Database (Denmark) Nielsen, Mogens; Valencia, Frank D. 2004-01-01 and program reactive systems. This note provides a comprehensive introduction to the background for and central notions from the theory of tccp. Furthermore, it surveys recent results on a particular tccp calculus, ntcc, and it provides a classification of the expressive power of various tccp languages.......A constraint is a piece of (partial) information on the values of the variables of a system. Concurrent constraint programming (ccp) is a model of concurrency in which agents (also called processes) interact by telling and asking information (constraints) to and from a shared store (a constraint... 18. Note from the radioprotection group's shipping service CERN Multimedia 2006-01-01 Le service SHIPPING du groupe de radioprotection souhaite vous rappeler qu'avant toute expédition de matériel susceptible d'être radioactif, une demande de transport doit être établie par EDH en cochant la case appropriée (danger radioactif). Merci de bien vouloir prendre note des informations figurant dans le site Web: http://cern.ch/service-rp-shipping Toute demande non conforme ne sera pas prise en compte. Radioactive Shipping Service http://cern.ch/service-rp-shippingTél: 73171Fax: 69200 19. European Science Notes. Volume 41, Number 7. Science.gov (United States) 1987-07-01 college preparatory that the social constructionist view, as group. (Chances are only slightly better he describes it, offers anything more for the...new method for the recognition and positioning of two-dimensionl objects. The method uses segmented descriptions of the object contours to generate...and recursively evaluate a number of selected hypotheses. This article dis- cusses the method in detail. International Symposium on Simulation of 20. European Scientific Notes. Volume 34, Number 4, Science.gov (United States) 1980-04-30 used to heat another body of water and most of the petroleum- exporting coun- to, say, 80°C and then condensed to form tries. And energy and water are...stocks, export subsidies, predator-prey relationship) and/or 194 ESN 34-4 (1980) Stages of Xoonoic Development and Structure of the Individual Farm Unit...such were not useful in agriculture because as cotton, are more tolerant, while they supplied only descriptive infor- others, including avocados (a 1. The Husting dilemma: A methodological note Science.gov (United States) Nichols, J.D.; Hepp, G.R.; Pollock, K.H.; Hines, J.E. 1987-01-01 Recently, Gill (1985) discussed the interpretation of capture history data resulting from his own studies on the red-spotted newt, Notophthalmus viridescens , and work by Husting (1965) on spotted salamanders, Ambystoma maculatum. Gill (1985) noted that gaps in capture histories (years in which individuals were not captured, preceded and followed by years in which they were) could result from either of two very different possibilities: (1) failure of the animal to return to the fenced pond to breed (the alternative Husting (1965) favored), or (2) return of the animal to the breeding pond, but failure of the investigator to capture it and detect its presence. The authors agree entirely with Gill (1985) that capture history data such as his or those of Husting (1965) should be analyzed using models that recognize the possibility of 'census error,' and that it is important to try to distinguish between such 'error' and skipped breeding efforts. The purpose of this note is to point out the relationship between Gill's (1985:347) null model and certain capture-recapture models, and to use capture-recapture models and tests to analyze the original data of Husting (1965). 2. Transport description of damped nuclear reactions International Nuclear Information System (INIS) Randrup, J. 1984-01-01 This lecture series is concerned with the transport description of damped nuclear reactions. Part 1 is an elementary introduction to the general transport theory of nuclear dynamics. It can be read without any special knowledge of the field, although basic quantum mechanics is required for the formal derivation of the general expressions for the transport coefficients. The results can also be used in a wider context than the present one. Part 2 gives the student an up-to-date orientation about recent progress in the understanding of the angular-momentum variables in damped reactions. The emphasis is here on the qualitative understanding of the physics rather than the, at times somewhat tedious, formal derivations. More detailed presentations are due to be published soon. By necessity entire topics have been omitted. For example, no discussion is given of the calculation of the form factors, and the several instructive applications of the theory to transport of mass and change are not covered at all. For these topics they refer to the literature. It is hoped that the present notes provide a sufficient basis to make the literature on the subject accessible to the student 3. Current knowledge on bioacoustics of the subfamily Lophyohylinae (Hylidae, Anura and description of Ocellated treefrog Itapotihyla langsdorffii vocalizations Directory of Open Access Journals (Sweden) Lucas Rodriguez Forti 2018-05-01 4. Situation awareness and documentation of changes that affect patient outcomes in progress notes. Science.gov (United States) Tower, Marion; Chaboyer, Wendy 2014-05-01 To report on registered nurses' situation awareness as a precursor to decision-making when recording changes in patients' conditions. Progress notes are important to communicate patients' progress and detail changes in patients' conditions. However, documentation is often poorly completed. There is little work that examines nurses' decision-making during documentation. This study focused on describing situation awareness as a precursor to decision-making during documentation. This study used Endsley's (Situation Awareness Analysis and Measurement, 2000, Lawrence Erlbaum Associates, NJ) work on situation awareness to guide and conceptualise information. The study was situated in a naturalistic paradigm to provide an interpretation of nurses' decision-making. Think-aloud research methods and semi-structured interviews were employed to illuminate decision-making processes. Audio recordings and interview texts were individually examined for evidence of cues, informed by Endsley's (Situation Awareness Analysis and Measurement, 2000, Lawrence Erlbaum Associates, NJ) descriptions of situation awareness. As patients' conditions changed, nurses used complex mental models and pattern-matching of information, drawing on all 3 levels of situation awareness during documentation. Level 1 situation awareness provided context, level 2 situation awareness signified a change in condition and its significance for the patient, and level 3 situation awareness was evident when nurses thought aloud about what this information indicated. Three themes associated with changes in patients' conditions emerged: deterioration in condition, not responding to prescribed treatments as expected and issues related to professional practice that impacted on patients' conditions. Nurses used a complex mental model for decision-making, drawing on 3 levels of situation awareness. Hamm's cognitive continuum theory, when related to situation awareness, is a useful decision-making theory to provide a 5. The Effectiveness of Note-Taking on Reading Comprehension of Iranian EFL Learners Directory of Open Access Journals (Sweden) Farnoush Bahrami 2017-10-01 Full Text Available By taking notes students could save time for reading all textbooks for their exams or for their representations. Taking notes increases attention of students to read or heard materials, and this increases their comprehension. Thus, the present study is important because note-taking could help them to remember what they learnt, absolutely important information. The method used in this research was survey. The 40 Persian EFL learners were selected from a language institute in Karaj to participate in the present study. These learners were divided into two groups; one of them is experimental group (N=20 and the other one is control group (N=20. Pretest and post test were two instruments that were used to carry out this study, a pretest about skill of note-taking of passages of the lessons was used for both experimental and control group. This test consisted of 4 passages. The same test was administrated again as the post test for both groups by the end of the course to see the different conclusion between taking note of experimental group and control group. Reliability between 4 texts is in oscillation from 0.6 to 0.81 (from 0.6 upwards. Therefore this reliability was an acceptable one. To analyze data descriptive statistics (that was contained percentage, frequency and mean score and also inferential statistics (that was contained ANOVA, Pearson correlation, independent sample t-test, multivariate’s test, regression were carried out by using SPSS16 soft ware. The findings confirmed that note taking is effective in reading comprehension. 6. Gaz de France. Operation note; Gaz de France. Note d'operation Energy Technology Data Exchange (ETDEWEB) NONE 2003-07-01 This note was published for the public at the occasion of the admission to Euronext's Eurolist of the existing shares that make the capital of Gaz de France company, the French gas utility. The note gives some informations about Gaz de France activity, and about its strategy of development in the European gas market. Then it describes the offer relative to the opening of Gaz de France capital. Some selected financial data and some precision about the risk factors and the management of the company complete the document. (J.S.) 7. 29 CFR Appendix to Subpart B of... - Reporter's Notes Science.gov (United States) 2010-07-01 ... Part 18—Reporter's Notes Reporter's Introductory Note The Rules of Evidence for the United States... contained in its files, such as the revenue statistics contained in the reports submitted to it by a... 8. Notes on the IMACON 500 streak camera system International Nuclear Information System (INIS) Clendenin, J.E. 1985-01-01 The notes provided are intended to supplement the instruction manual for the IMACON 500 streak camera system. The notes cover the streak analyzer, instructions for timing the streak camera, and calibration 9. CANISTER HANDLING FACILITY DESCRIPTION DOCUMENT Energy Technology Data Exchange (ETDEWEB) J.F. Beesley 2005-04-21 The purpose of this facility description document (FDD) is to establish requirements and associated bases that drive the design of the Canister Handling Facility (CHF), which will allow the design effort to proceed to license application. This FDD will be revised at strategic points as the design matures. This FDD identifies the requirements and describes the facility design, as it currently exists, with emphasis on attributes of the design provided to meet the requirements. This FDD is an engineering tool for design control; accordingly, the primary audience and users are design engineers. This FDD is part of an iterative design process. It leads the design process with regard to the flowdown of upper tier requirements onto the facility. Knowledge of these requirements is essential in performing the design process. The FDD follows the design with regard to the description of the facility. The description provided in this FDD reflects the current results of the design process. 10. CANISTER HANDLING FACILITY DESCRIPTION DOCUMENT International Nuclear Information System (INIS) Beesley. J.F. 2005-01-01 The purpose of this facility description document (FDD) is to establish requirements and associated bases that drive the design of the Canister Handling Facility (CHF), which will allow the design effort to proceed to license application. This FDD will be revised at strategic points as the design matures. This FDD identifies the requirements and describes the facility design, as it currently exists, with emphasis on attributes of the design provided to meet the requirements. This FDD is an engineering tool for design control; accordingly, the primary audience and users are design engineers. This FDD is part of an iterative design process. It leads the design process with regard to the flowdown of upper tier requirements onto the facility. Knowledge of these requirements is essential in performing the design process. The FDD follows the design with regard to the description of the facility. The description provided in this FDD reflects the current results of the design process 11. Fuel Handling Facility Description Document International Nuclear Information System (INIS) M.A. LaFountain 2005-01-01 The purpose of the facility description document (FDD) is to establish the requirements and their bases that drive the design of the Fuel Handling Facility (FHF) to allow the design effort to proceed to license application. This FDD is a living document that will be revised at strategic points as the design matures. It identifies the requirements and describes the facility design as it currently exists, with emphasis on design attributes provided to meet the requirements. This FDD was developed as an engineering tool for design control. Accordingly, the primary audience and users are design engineers. It leads the design process with regard to the flow down of upper tier requirements onto the facility. Knowledge of these requirements is essential to performing the design process. It trails the design with regard to the description of the facility. This description is a reflection of the results of the design process to date 12. How to write cases and teaching notes in marketing education? Directory of Open Access Journals (Sweden) Damnjanović Vesna 2017-01-01 Full Text Available This paper presents guidelines for developing cases and teaching notes for marketing classes in higher education. Cases are often used in problem-based learning, where students are placed in a problem or decision-making situation and are encouraged to take the role of problem solvers. The recommendations for case writing are based on analysis of high-level examples of marketing cases, published in the most prominent global publishing databases, along with more than 15 years of experience of the authors. This paper explains the process of writing cases depending on the source of data and tackles all necessary issues that should be addressed. Writing teaching notes is focused on providing the supportive material for a case that helps a lecturer understand the reason for writing the case, the questions that might arise from it and the professional or theoretical points that should be raised in discussion. This paper should contribute to interactive learning and provide lecturers with detailed recommendations for writing cases in the marketing field. 13. A Pedagogic Cycle for EFL Note-Taking Science.gov (United States) Siegel, Joseph 2016-01-01 More attention has been given to note-taking in EFL classrooms in recent years, probably due to growing numbers of EFL students who attend lectures given in English and the importance of note-taking during language proficiency tests. While textbooks and materials for note-taking practice are readily available, classroom techniques for teaching EFL… 14. Insights regarding the Usefulness of Partial Notes in Mathematics Courses Science.gov (United States) Cardetti, Fabiana; Khamsemanan, Nirattaya; Orgnero, M. Carolina 2010-01-01 Note-taking is a widespread practice used by college students to record information from lectures. Unfortunately, even successful students' notes are incomplete and, therefore, may lack the potential to positively impact their academic performance. Research suggests that instructors can help students improve their note-taking skills by using… 15. SNF/HLW Transfer System Description Document International Nuclear Information System (INIS) W. Holt 2005-01-01 The purpose of this system description document (SDD) is to establish requirements that drive the design of the spent nuclear fuel (SNF)/high-level radioactive waste (HLW) transfer system and associated bases, which will allow the design effort to proceed to license application. This SDD will be revised at strategic points as the design matures. This SDD identifies the requirements and describes the system design, as it currently exists, with emphasis on attributes of the design provided to meet the requirements. This SDD is an engineering tool for design control. Accordingly, the primary audience and users are design engineers. This SDD is part of an iterative design process. It leads the design process with regard to the flowdown of upper tier requirements onto the system. Knowledge of these requirements is essential in performing the design process. The SDD follows the design with regard to the description of the system. The description provided in this SDD reflects the current results of the design process 16. Grouping notes through nodes: The functions of Post-It notes in design team cognition DEFF Research Database (Denmark) Dove, Graham; Abildgaard, Sille Julie Jøhnk; Biskjaer, Michael Mose 2018-01-01 We investigate the way Post-It notes support creative design team practice, focusing on how they function as cognitive externalisations that, through grouping activities, support categorisation qualities associated with semantic long-term memory. We use a multimodal approach, drawing... 17. Lecture Notes on Topics in Accelerator Physics Energy Technology Data Exchange (ETDEWEB) Chao, Alex W. 2002-11-15 These are lecture notes that cover a selection of topics, some of them under current research, in accelerator physics. I try to derive the results from first principles, although the students are assumed to have an introductory knowledge of the basics. The topics covered are: (1) Panofsky-Wenzel and Planar Wake Theorems; (2) Echo Effect; (3) Crystalline Beam; (4) Fast Ion Instability; (5) Lawson-Woodward Theorem and Laser Acceleration in Free Space; (6) Spin Dynamics and Siberian Snakes; (7) Symplectic Approximation of Maps; (8) Truncated Power Series Algebra; and (9) Lie Algebra Technique for nonlinear Dynamics. The purpose of these lectures is not to elaborate, but to prepare the students so that they can do their own research. Each topic can be read independently of the others. 18. Lecture Notes on Topics in Accelerator Physics International Nuclear Information System (INIS) Chao, Alex W. 2002-01-01 These are lecture notes that cover a selection of topics, some of them under current research, in accelerator physics. I try to derive the results from first principles, although the students are assumed to have an introductory knowledge of the basics. The topics covered are: (1) Panofsky-Wenzel and Planar Wake Theorems; (2) Echo Effect; (3) Crystalline Beam; (4) Fast Ion Instability; (5) Lawson-Woodward Theorem and Laser Acceleration in Free Space; (6) Spin Dynamics and Siberian Snakes; (7) Symplectic Approximation of Maps; (8) Truncated Power Series Algebra; and (9) Lie Algebra Technique for nonlinear Dynamics. The purpose of these lectures is not to elaborate, but to prepare the students so that they can do their own research. Each topic can be read independently of the others 19. A note on migration with borrowing constraints. Science.gov (United States) Ghatak, S; Levine, P 1994-12-01 "This note examines an important conflict between the theory and evidence on migration in LDCs. While the Harris-Todaro class of models explain the phenomenon of migration mainly by expected income differential between the economically advanced and the backward regions, the actual evidence in some cases suggests that migration could actually rise following a rise in income in backward areas. We resolve this puzzle by analysing migration in the context of the existence of imperfect credit markets in LDCs. We show that under certain plausible conditions, the rate of migration from the rural to the urban areas may actually rise when rural wages rise, as they ease the constraints on borrowing by potential migrants." excerpt 20. Note on the extended noncommutativity of coordinates International Nuclear Information System (INIS) Boulahoual, Amina; Sedra, My. Brahim 2001-04-01 We present in this short note an idea about a possible extension of the standard noncommutative algebra to the formal differential operators framework. In this sense, we develop an analysis and derive an extended noncommutative algebra given by [x a , x b ] * =i(θ+χ) ab where θ ab , is the standard noncommutative parameter and χ ab (x)≡χ ab μ (x)δ μ =1/2(x a θ μ b -x b θ a )δ μ is an antisymmetric non-constant vector-field shown to play the role of the extended deformation parameter. This idea was motivated by the importance of noncommutative geometry framework in the current subject of D-brane and matrix theory physics. (author)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.4980227053165436, "perplexity": 5536.106286024223}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-45/segments/1603107881369.4/warc/CC-MAIN-20201023102435-20201023132435-00671.warc.gz"}
https://terrytao.wordpress.com/2015/10/07/sweeping-a-matrix-rotates-its-graph/?replytocom=460285
I recently learned about a curious operation on square matrices known as sweeping, which is used in numerical linear algebra (particularly in applications to statistics), as a useful and more robust variant of the usual Gaussian elimination operations seen in undergraduate linear algebra courses. Given an ${n \times n}$ matrix ${A := (a_{ij})_{1 \leq i,j \leq n}}$ (with, say, complex entries) and an index ${1 \leq k \leq n}$, with the entry ${a_{kk}}$ non-zero, the sweep ${\hbox{Sweep}_k[A] = (\hat a_{ij})_{1 \leq i,j \leq n}}$ of ${A}$ at ${k}$ is the matrix given by the formulae $\displaystyle \hat a_{ij} := a_{ij} - \frac{a_{ik} a_{kj}}{a_{kk}}$ $\displaystyle \hat a_{ik} := \frac{a_{ik}}{a_{kk}}$ $\displaystyle \hat a_{kj} := \frac{a_{kj}}{a_{kk}}$ $\displaystyle \hat a_{kk} := \frac{-1}{a_{kk}}$ for all ${i,j \in \{1,\dots,n\} \backslash \{k\}}$. Thus for instance if ${k=1}$, and ${A}$ is written in block form as $\displaystyle A = \begin{pmatrix} a_{11} & X \\ Y & B \end{pmatrix} \ \ \ \ \ (1)$ for some ${1 \times n-1}$ row vector ${X}$, ${n-1 \times 1}$ column vector ${Y}$, and ${n-1 \times n-1}$ minor ${B}$, one has $\displaystyle \hbox{Sweep}_1[A] = \begin{pmatrix} -1/a_{11} & X / a_{11} \\ Y/a_{11} & B - a_{11}^{-1} YX \end{pmatrix}. \ \ \ \ \ (2)$ The inverse sweep operation ${\hbox{Sweep}_k^{-1}[A] = (\check a_{ij})_{1 \leq i,j \leq n}}$ is given by a nearly identical set of formulae: $\displaystyle \check a_{ij} := a_{ij} - \frac{a_{ik} a_{kj}}{a_{kk}}$ $\displaystyle \check a_{ik} := -\frac{a_{ik}}{a_{kk}}$ $\displaystyle \check a_{kj} := -\frac{a_{kj}}{a_{kk}}$ $\displaystyle \check a_{kk} := \frac{-1}{a_{kk}}$ for all ${i,j \in \{1,\dots,n\} \backslash \{k\}}$. One can check that these operations invert each other. Actually, each sweep turns out to have order ${4}$, so that ${\hbox{Sweep}_k^{-1} = \hbox{Sweep}_k^3}$: an inverse sweep performs the same operation as three forward sweeps. Sweeps also preserve the space of symmetric matrices (allowing one to cut down computational run time in that case by a factor of two), and behave well with respect to principal minors; a sweep of a principal minor is a principal minor of a sweep, after adjusting indices appropriately. Remarkably, the sweep operators all commute with each other: ${\hbox{Sweep}_k \hbox{Sweep}_l = \hbox{Sweep}_l \hbox{Sweep}_k}$. If ${1 \leq k \leq n}$ and we perform the first ${k}$ sweeps (in any order) to a matrix $\displaystyle A = \begin{pmatrix} A_{11} & A_{12} \\ A_{21} & A_{22} \end{pmatrix}$ with ${A_{11}}$ a ${k \times k}$ minor, ${A_{12}}$ a ${k \times n-k}$ matrix, ${A_{12}}$ a ${n-k \times k}$ matrix, and ${A_{22}}$ a ${n-k \times n-k}$ matrix, one obtains the new matrix $\displaystyle \hbox{Sweep}_1 \dots \hbox{Sweep}_k[A] = \begin{pmatrix} -A_{11}^{-1} & A_{11}^{-1} A_{12} \\ A_{21} A_{11}^{-1} & A_{22} - A_{21} A_{11}^{-1} A_{12} \end{pmatrix}.$ Note the appearance of the Schur complement in the bottom right block. Thus, for instance, one can essentially invert a matrix ${A}$ by performing all ${n}$ sweeps: $\displaystyle \hbox{Sweep}_1 \dots \hbox{Sweep}_n[A] = -A^{-1}.$ If a matrix has the form $\displaystyle A = \begin{pmatrix} B & X \\ Y & a \end{pmatrix}$ for a ${n-1 \times n-1}$ minor ${B}$, ${n-1 \times 1}$ column vector ${X}$, ${1 \times n-1}$ row vector ${Y}$, and scalar ${a}$, then performing the first ${n-1}$ sweeps gives $\displaystyle \hbox{Sweep}_1 \dots \hbox{Sweep}_{n-1}[A] = \begin{pmatrix} -B^{-1} & B^{-1} X \\ Y B^{-1} & a - Y B^{-1} X \end{pmatrix}$ and all the components of this matrix are usable for various numerical linear algebra applications in statistics (e.g. in least squares regression). Given that sweeps behave well with inverses, it is perhaps not surprising that sweeps also behave well under determinants: the determinant of ${A}$ can be factored as the product of the entry ${a_{kk}}$ and the determinant of the ${n-1 \times n-1}$ matrix formed from ${\hbox{Sweep}_k[A]}$ by removing the ${k^{th}}$ row and column. As a consequence, one can compute the determinant of ${A}$ fairly efficiently (so long as the sweep operations don’t come close to dividing by zero) by sweeping the matrix for ${k=1,\dots,n}$ in turn, and multiplying together the ${kk^{th}}$ entry of the matrix just before the ${k^{th}}$ sweep for ${k=1,\dots,n}$ to obtain the determinant. It turns out that there is a simple geometric explanation for these seemingly magical properties of the sweep operation. Any ${n \times n}$ matrix ${A}$ creates a graph ${\hbox{Graph}[A] := \{ (X, AX): X \in {\bf R}^n \}}$ (where we think of ${{\bf R}^n}$ as the space of column vectors). This graph is an ${n}$-dimensional subspace of ${{\bf R}^n \times {\bf R}^n}$. Conversely, most subspaces of ${{\bf R}^n \times {\bf R}^n}$ arises as graphs; there are some that fail the vertical line test, but these are a positive codimension set of counterexamples. We use ${e_1,\dots,e_n,f_1,\dots,f_n}$ to denote the standard basis of ${{\bf R}^n \times {\bf R}^n}$, with ${e_1,\dots,e_n}$ the standard basis for the first factor of ${{\bf R}^n}$ and ${f_1,\dots,f_n}$ the standard basis for the second factor. The operation of sweeping the ${k^{th}}$ entry then corresponds to a ninety degree rotation ${\hbox{Rot}_k: {\bf R}^n \times {\bf R}^n \rightarrow {\bf R}^n \times {\bf R}^n}$ in the ${e_k,f_k}$ plane, that sends ${f_k}$ to ${e_k}$ (and ${e_k}$ to ${-f_k}$), keeping all other basis vectors fixed: thus we have $\displaystyle \hbox{Graph}[ \hbox{Sweep}_k[A] ] = \hbox{Rot}_k \hbox{Graph}[A]$ for generic ${n \times n}$ ${A}$ (more precisely, those ${A}$ with non-vanishing entry ${a_{kk}}$). For instance, if ${k=1}$ and ${A}$ is of the form (1), then ${\hbox{Graph}[A]}$ is the set of tuples ${(r,R,s,S) \in {\bf R} \times {\bf R}^{n-1} \times {\bf R} \times {\bf R}^{n-1}}$ obeying the equations $\displaystyle a_{11} r + X R = s$ $\displaystyle Y r + B R = S.$ The image of ${(r,R,s,S)}$ under ${\hbox{Rot}_1}$ is ${(s, R, -r, S)}$. Since we can write the above system of equations (for ${a_{11} \neq 0}$) as $\displaystyle \frac{-1}{a_{11}} s + \frac{X}{a_{11}} R = -r$ $\displaystyle \frac{Y}{a_{11}} s + (B - a_{11}^{-1} YX) R = S$ we see from (2) that ${\hbox{Rot}_1 \hbox{Graph}[A]}$ is the graph of ${\hbox{Sweep}_1[A]}$. Thus the sweep operation is a multidimensional generalisation of the high school geometry fact that the line ${y = mx}$ in the plane becomes ${y = \frac{-1}{m} x}$ after applying a ninety degree rotation. It is then an instructive exercise to use this geometric interpretation of the sweep operator to recover all the remarkable properties about these operations listed above. It is also useful to compare the geometric interpretation of sweeping as rotation of the graph to that of Gaussian elimination, which instead shears and reflects the graph by various elementary transformations (this is what is going on geometrically when one performs Gaussian elimination on an augmented matrix). Rotations are less distorting than shears, so one can see geometrically why sweeping can produce fewer numerical artefacts than Gaussian elimination.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 358, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9449103474617004, "perplexity": 182.03684065955974}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 20, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-51/segments/1575540521378.25/warc/CC-MAIN-20191209173528-20191209201528-00229.warc.gz"}
https://www.math.rutgers.edu/academics/graduate-program/course-descriptions/1018-640-507-functional-analysis-i
# Course Descriptions ## 16:640:507 - Functional Analysis I Denis Kriventsov ### Text: Haim Brezis, "Functional Analysis, Sobolev Spaces, and Partial Differential Equations ### Prerequisites: 16:640: 501 and 16:640:502 ### Description: Sobolev spaces and the variational formulation of boundary value problems in one dimension. The Hahn-Banach theorems. Conjugate convex functions. The uniform boundedness principle and the closed graph theorem. Characterization of surjective operators. Weak topologies. Reflexive spaces. Separable spaces. Uniform convexity. L^{p} spaces. Hilbert spaces. Compact operators. Spectral decomposition of self-adjoint compact operators. ## Contacts Departmental Chair Michael Saks
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8606194257736206, "perplexity": 4130.486689282585}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-47/segments/1573496665985.40/warc/CC-MAIN-20191113035916-20191113063916-00249.warc.gz"}
http://tdunning.blogspot.ru/2011/05/
## Sunday, May 29, 2011 ### Online algorithms for boxcar-ish moving averages One problem with exponentially weighted moving averages is that the weight that older samples decays sharply even for very recent samples. The impulse response of such an average shows this clearly. In the graph to the right, the red line is the impulse response of the exponential weighted average is shown by the red line. The impulse response of a different kind of moving average derived by John Canny in the early 80's is shown by the black line. The Canny filter puts higher weight on the events in the recent past which makes it preferable when you don't want to forget things right away, but do want to forget them before long and also want an on-line algorithm. The cool thing about the Canny filter is that it is only twice as much work as a simple exponential moving average. The idea is that if you take the difference between two exponentially weighted averages with different time constants, you can engineer things to give you an impulse response of the sort that you would like. The formula for such a difference looks like this \begin{eqnarray*} w(x) &=& k_1 e^{-x/\alpha} - k_2 e^{-x/\beta} \end{eqnarray*} Here $k_1$ and $k_2$ scale the two component filters in magnitude and $\alpha$ and $\beta$ are the time constants for the filters. It is nice to set the magnitude of the filter at delay $0$ to be exactly 1. We can use this to get a value for $k_2$ in terms of $k_1$ \begin{eqnarray*} w(0) &=& k_1 - k_2 = 1 \\ k_2 &=& k_1 -1 \end{eqnarray*} Similarly, we can constrain the slope of the impulse response to be $0$ at delay $0$. This gives us $\beta$ in terms $\alpha$ \begin{eqnarray*} w'(0) &=& {k_1 \over \alpha} - {k_2 \over \beta} = 0\\ {k_1 \over \alpha} &=& {k_1-1 \over \beta} \\ \beta &=& \alpha \frac{ k_1-1} { k_1} \end{eqnarray*} The final result is this impulse response \begin{eqnarray*} w(x) = k \exp \left(-{x \over \alpha}\right)-(k-1) \exp\left(-{k x\over \alpha (k-1)}\right) \end{eqnarray*} We can do a bit better if we note that as $k \to \infty$, the shape of the impulse quickly converges to a constant shape with mean of $w(x) \to \frac{3a}{2}$ and a total volume of $2a$.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9682269096374512, "perplexity": 513.828526916887}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 5, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2017-22/segments/1495463608004.38/warc/CC-MAIN-20170525063740-20170525083740-00219.warc.gz"}
https://origamiusa.org/catalog/products/paper-form
Paper & Form Your price: $45.00 Author Zsebe SKU: B15-118 Regular price:$45.00 Member price: \$40.50 13 original 3-dimensional models, each from a single, uncut square. Models include Dove, Lion, Horse, Colorado Beetle, Ladybird, Bison, Warthog, Wild Boar Piglet, Wild Boar, Fly-Agaric (mushroom), Brown Bear, Polar Bear, and Bear on Rock (from a single sheet). The author recommends paper that is good for shaping or wet-folding, such as tissue foil, thicker papers, dampened cardstock, and Elephant Hide. A short paragraph at the beginning of each design describes the inspiration, folding hints, and paper recommendations. There is also a crease pattern and drawing of the model dimensions as related to paper size. Designs are expressive, make efficient use of the paper, and are 3-D, so they stand up and make for excellent displays. Quality hardcover book includes color photos and diagrams with text descriptions on each step. In English and Hungarian. 92 pp. HC. (I-HI)
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.19257226586341858, "perplexity": 13215.3574685376}, "config": {"markdown_headings": false, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439735885.72/warc/CC-MAIN-20200804220455-20200805010455-00568.warc.gz"}
https://link.springer.com/article/10.1007/s10658-019-01869-x
## Introduction Ramularia leaf spot (RLS) is an important disease of barley crops in many temperate countries (Havis et al. 2015; McGrann and Havis 2017). The disease is caused by the fungus Ramularia collo-cygni, which is transmitted through infected seed and by air-borne spores (Havis et al. 2014). R. collo-cygni has a long endophytic phase (Kaczmarek et al. 2017) with necrotrophic disease symptoms usually only visible once the crop has entered the reproductive phase (Schützendübel et al. 2008). Despite R. collo-cygni possessing the genetic capability to synthesise a range of secondary metabolites that could potentially affect disease severity (Dussart et al. 2018), expression of RLS appears to be linked with adverse environmental conditions (Makepeace et al. 2008; McGrann and Brown 2018; Peraldi et al. 2014). The vertical transmission by seeds and relationship between environmental conditions and symptom expression has led to the suggestion that R. collo-cygni may be an endophyte that only causes disease under specific circumstances (McGrann and Havis 2017). Resistance to RLS appears to be a quantitative trait although genetic studies in barley have indicated that mutant MILDEW LOCUS O (mlo) alleles increase susceptibility to RLS (McGrann et al. 2014), although this phenotype is not observed in all environments (Makepeace et al. 2007). Mutant mlo alleles are agriculturally important as they confer broad-spectrum resistance to all known field isolates of powdery mildew caused by Blumeria graminis f. sp. hordei in spring barley (Piffanelli et al. 2004). This resistance has been used to great effect for more than 40 years. The mutant allele is currently present in more than 70% of elite European varieties (Dreiseitl 2012; Jørgensen 1992) with most commercial spring barley with mlo-mediated mildew resistance carrying the mlo11 allele although some older varieties contain mlo9 (Jørgensen 1992). The MLO gene encodes a seven-transmembrane domain protein (Büschges et al. 1997). Wild type MLO alleles are susceptibility factors for biotrophic powdery mildew fungi, which co-evolved before the divergence of monocot and dicot plants (Acevedo-Garcia et al. 2014; Appiano et al. 2015). However, introducing mlo in commercial spring barley varieties came with a number of undesirable pleiotropic effects. Plants containing mlo alleles developed spontaneous foliar necrosis spotting, tended to yield less (Kjaer et al. 1990; Wolter et al. 1993; Thomas et al. 1998) and were more susceptible to diseases with necrotrophic stages (Jarosch et al. 1999; Kumar et al. 2001; Jansen et al. 2005; McGrann et al. 2014). Plant breeding has managed to alleviate the necrotic spotting and yield penalties associated with mlo (Bjornstad and Aastveit 1990; Kjaer et al. 1990) but there are still concerns over increased disease susceptibility associated with this mutant allele. Mutant analyses have indicated that mlo-mediated susceptibility to the pathogens with necrotrophic growth stages is regulated by other genes. Two REQUIRED FOR MLO RESISTANCE (ROR1 and ROR2) genes are essential for the full expression of mlo-mediated resistance to powdery mildew (Freialdenhoven et al. 1996). Enhanced RLS symptom formation in mlo plants is reduced in ror1 and ror2 mutants although levels of R. collo-cygni DNA in colonised leaves are still elevated (McGrann et al. 2014). Mutant ror1 plants show decreased sensitivity to toxins produced by Bipolaris sorokiniana compared to mlo plants with the wild type ROR1 allele (Kumar et al. 2001). However, loss of ROR1 function has no effect on barley susceptibility to Magnaporthe oryzae (Jarosch et al. 1999) suggesting ROR genes may operate through mechanisms that differentially affect mlo-mediated susceptibility to specific fungal species. NEC1 encodes a cyclic nucleotide gated channel 4 protein, which when defective, results in a spontaneous lesion mimic phenotype in barley (Rostoks et al. 2006). Mutant nec1 alleles reduce lesion formation by Fusarium culmorum and R. collo-cygni in barley but these nec1-mediated phenotypes are differentially affected by mlo mutations. The presence of nec1 mutation in an mlo5 background was shown to have no effect on the development of RLS symptoms or the accumulation of R. collo-cygni DNA. However, the presence of mlo5 in a nec1 background compromises the reduction in F. culmorum lesion size usually seen in nec1 mutants, although disease levels did not reach those seen in wild type MLO or NEC1 plants or in mlo5 mutants (McGrann et al. 2015a). A screen for restoring resistance to M. oryzae in mlo-barley identified two loss of function mutations, ENHANCED MAGNAPORTHE RESISTANCE (EMR1 and EMR2), which showed partially restored resistance to M. oryzae with no compromise on mildew resistance (Jansen et al. 2007; Jansen and Schaffrath 2009). The gene responsible for the emr1 or emr2 phenotype have not yet been characterised. Genetic mapping of the emr1 locus indicated that the emr1 mutant contains at least one additional mutant gene, 3-KETOACYL-CoA SYNTHASE (HvKCS6), located on chromosome 4H that results in emr1 plants having reduced lower levels of cuticular waxes (Weidenbach et al. 2014). These mutagenesis studies have shown that the negative effect of mlo on certain pathogens with necrotrophic colonisation stages can be mitigated. Here, the effect of the emr1 mutation on RLS was examined to assess the potential of this locus to mitigate mlo-mediated susceptibility to this disease. ## Materials and methods ### Plant material The emr1mlo5 mutant was derived from NaN3 mutagenesis of IngridBCmlo5 mutant (mlo5) and shown to restore resistance to the blast fungus, Magnaporthe oryzae, in the highly susceptible mlo5 background (Jansen et al. 2007). Experiments included the emr1mlo5 double mutant, the single mutant parental line mlo5 and the wild type donor cultivar of both mutants cv. Ingrid. Seeds were sown in to Levington M3 compost (ICL, Ipswich, UK) in rows of ten seeds per plastic tray (210 × 156 × 47 mm). Barley seedlings were cultivated in a MD1400 modular climate chamber (Snijders Labs, Tilberg, The Netherlands) at 18 °C with a 16:8 h light: dark cycle under 200 μmol m−2 s−1 artificial light at 80% relative humidity. ### Ramularia collo-cygni isolates and inoculation method The R. collo-cygni isolate DK05 Rcc001 (McGrann et al. 2016) collected from Denmark in 2005 was used in this study. Fungal cultures were stored on potato dextrose agar (PDA) plates supplemented with streptomycin (100 μg mL−1) at 15 °C. Liquid cultures were prepared by adding two 5 mm2 plugs from 14 to 28 day old PDA storage plates to 200 mL potato dextrose broth (PDB) supplemented with streptomycin (100 μg mL−1). PDB flasks with R. collo-cygni were incubated for 14 days at 15 °C in the dark on an orbital shaker set at 150 rpm. Barley seedlings were inoculated with R. collo-cygni as previously described (McGrann et al. 2014; Peraldi et al. 2014). Inoculum was prepared by fragmenting R. collo-cygni hyphal growth in the liquid cultures in a food processor (Kenwood Electronics, London, UK). A single drop of Tween20 (Sigma, Dorset, UK) was added to each batch of 50 mL inoculum that was used to inoculate approximately 240 barley plants at growth stage (GS) 12 (Zadoks et al. 1974). Plants were incubated in the dark for 48 h post inoculation and RLS lesions were scored as the proportion of the prophyll leaf covered with disease symptoms from 8 to 21 days post inoculation (dpi). Green leaf area (GLA) retention was assessed as the amount of prophyll leaf area not senescent at the same time point that RLS was scored in the emr1 mutant seedling inoculation experiments. A minimum of three independent inoculation experiments were assessed for each seedling inoculation experiment. ### Field trials Barley cv. Ingrid, the near isogenic line mlo5 and the emr1mlo5 mutant were sown in tussock plots at a trial site in Lanark, Scotland in 2015 and 2016. The Lanark site was selected as it typically exhibits high levels of R. collo-cygni (Havis et al. 2014). To maintain green leaf area retention and control other foliar diseases during the growing season the fungicides Comet® 200 (0.4 L ha−1 Pyraclostrobin [200 g L−1] BASF, Cheshire, UK) plus Vegas ® (0.25 L ha−1 Cyflufenamid [50 g L−1] Certis, Cambridgeshire, UK) were applied as a mixture to each plot at both growth stage (GS) 25–20 and GS45. RLS symptoms were scored 2–3 times after the crop had flowered (GS60 onwards) as the proportion of leaf F-1 covered with disease lesions. ### Quantification of in planta Ramularia collo-cygni DNA levels Genomic DNA (gDNA) was extracted from five leaves of cv. Ingrid, mlo5 and the emr1mlo5 mutant 21 dpi with R. collo-cygni isolate DK05 using the DNeasy Plant Mini Kit (Qiagen, Hildem, Germany) following the manufacturer’s protocol. R. collo-cygni DNA levels were quantified using a dilution series of R. collo-cygni DNA (50 ng to 0.1 pg) with qPCR following the method of Taylor et al. (2010). Data was collected from two independent inoculation experiments. ### Dark-induced senescence assays Differences in relative chlorophyll content and senescence between emr1mlo5, mlo5 and cv. Ingrid were tested using a SPAD 502 Plus Chlorophyll meter (Konica Minolta, Warrington, UK) in a dark-induced senescence assay as previously described (McGrann et al. 2015b). Relative chlorophyll content was taken from three points across the blade of GS12 prophyll leaves after excision from the plant (day 0). Leaves were placed on damp tissue paper in clear plastic boxes covered with aluminium foil and stored at room temperature. Additional measurements were taken every two days up to and including day eight of the experiment. Three independent experiments were examined with 6–8 leaves of each line tested in each experiment. ### Sensitivity to reactive oxygen species-induced cell death The sensitivity of emr1mlo5, mlo5 and cv. Ingrid to cell death induced by reactive oxygen species (ROS) was tested using detached prophyll leaves from barley plants at GS11–12 as described previously (McGrann et al. 2015b). Cell death was induced by the hydrogen peroxide donor alloxan (200 mM), the mitochondrial superoxide donor menadione (100 mM) and the chloroplastic superoxide donor methyl viologen (25 μM). Lesions were measured 96 h after treatment with each ROS donor using ImageJ software (Abràmoff et al. 2004). Three independent experiments were performed with eight leaves of each line assessed for each ROS donor in each replicate experiment. ### Data analysis RLS scores over each seeding inoculation time course experiment were used to calculate the area disease under progress curve (AUDPC; Shaner and Finney 1977). AUDPC data was converted into the percentage of the maximum AUDPC (%maxAUDPC) across the entire seedling inoculation time course experiment. Prior to analysis %maxAUDPC data was LOGIT+ transformed (McGrann et al. 2014) and then analysed using a general linear model (GLM). Differences in seedling levels of RLS %maxAUDPC and R. collo-cygni gDNA in seedling prophyll leaves 21 dpi of the emr1mlo5 mutant compared to mlo5 and Ingrid were tested using GLM with line and experiment as factors. Differences in GLA retention in emr1mlo5, mlo5 and cv. Ingrid was examined using linear mixed modelling of repeated measures. Differences in GLA retention were evaluated using the uniform correlation/split plot in time covariance matrix with dpi, line, experiment and the interactions between these factors set as fixed factors. The dpi by individual plant interaction term was set as the random factor in the model. Significant differences between mlo5 and emr1mlo5 or cv. Ingrid at specific dpi were assessed using a t-test. GLM was also used to assess the variation in RLS AUDPC in the field trials with line and year as factors following a Poisson transformation of the AUDPC data. Variation in ROS-induced cell death between emr1mlo5, mlo5 and cv. Ingrid was assessed using a separate GLM for each ROS donor. Each GLM assessed the contribution of experiment, line and the interactions between these terms on the observed phenotypes. Differences in dark-induced senescence were tested using linear mixed modelling with repeated measurements as described above for GLA. ## Results ### Development of Ramularia leaf spot on the spring barley emr1mlo5 mutant In seedling assays typical RLS symptoms were observed on leaves of barley cv. Ingrid, the mlo5 single mutant and the emr1mlo5 double mutant (Fig. 1a) 21 dpi with lesions first visible 8–10 dpi. The mlo5 mutant had increased RLS development compared to the wild type Ingrid (P < 0.001) as previously reported (McGrann et al. 2014, 2015a), whilst the emr1mlo5 mutant developed significantly less RLS than mlo5 (P < 0.001) but significantly more disease than cv. Ingrid (P = 0.007; Fig. 1b). By 8 dpi both mlo5 and emr1mlo5 had significantly less GLA than cv. Ingrid (P < 0.01) whereas emr1mlo5 plants had significantly lower GLA than mlo5 (P < 0.05) at this time point (Fig.1c). From 10 dpi onwards there were no significant differences in GLA between mlo5 and emr1mlo5 whereas both mutants had significantly lower levels of GLA compared to cv. Ingrid at all further time points (P < 0.001; Fig. 1c). Despite the differences in RLS development between emr1mlo5 and mlo5 there was no significant difference in the accumulation of R. collo-cygni gDNA 21 dpi between the mutants (P = 0.98) whereas cv. Ingrid leaves had significantly less fungal DNA than mlo5 (P = 0.022) and emr1mlo5 (P = 0.039) leaves (Fig. 1d). In field grown plants no significant differences in RLS development were recorded between the two years of the field trials (P = 0.37) nor were any differences observed between the lines (P = 0.74). However, even though the differences in RLS development between field grown cv. Ingrid, mlo5 and emr1mlo5 were not statistically significant there was a trend for mlo5 plants to exhibit more RLS symptoms than the other two lines (Fig. 1e). ### Effect of emr1 mutation on relative chlorophyll content and dark-induced senescence Processes linked with foliar senescence and chlorophyll breakdown have been linked with development of RLS (Schützendübel et al. 2008; McGrann et al. 2015b; McGrann and Brown 2018). Therefore, the effects of the emr1 mutation on senescence were examined using a dark-induced senescence assay. Prophyll leaves of the emr1mlo5 mutant at GS12 were a paler shade of green when compared to mlo5 or cv. Ingrid (Fig. 2a). This may be related to the altered wax status of emr1mlo5 plants (Weidenbach et al. 2014) or due to changes in chlorophyll content of the leaves associated with the emr1 mutation. Relative chlorophyll content of the emr1mlo5 mutant was significantly lower (P < 0.001) at day 0 than either mlo5 or cv. Ingrid. The relative chlorophyll content of emr1mlo5 remained significantly lower than mlo5 throughout the dark-induced senescence time course (day 2–6) but the SPAD readings for the two mutants were not significantly different by day 8 (Fig. 2b). Relative chlorophyll content of cv. Ingrid leaves was significantly higher than both emr1mlo5 (P < 0.001) and mlo5 (P < 0.05) across all time points. ### Sensitivity of emr1mlo5 mutant to reactive oxygen species-induced cell death Changes in ROS homeostasis associated with plant development and spontaneous cell death phenotypes have been linked with expression of RLS symptoms (McGrann and Brown 2018). No significant differences in the size of the alloxan-induced lesion (P = 0.49) were observed between emr1mlo5, mlo5 and cv. Ingrid (Fig. 3a+d). No significant difference in the size of the lesions formed by the mitochondrial superoxide donor menadione were recorded between the lines (Fig. 3b+e; P = 0.07) but the emr1mlo5 mutant did produce a significantly smaller lesion compared to mlo5 and cv. Ingrid (P < 0.001) when treated with the chloroplastic superoxide donor methyl viologen (Fig. 3c+f). ## Discussion Ramularia leaf spot has established itself as a serious threat to modern barley production across temperate climates (McGrann and Havis 2017; Havis et al. 2015; Walters et al. 2008). Yield losses typically range from 5 to 10% but can be as high as 70% without considering potential losses to grain quality which can reduce the value of the crop further (Havis et al. 2015). Resistance against RLS is most probably a quantitatively controlled genetic trait and the strong genotype by environment effects observed for the expression of disease symptoms between field trials, has led to problems in breeding effective resistance against this disease (Havis et al. 2015). Durable powdery mildew resistance conferred by the mlo locus has resulted in mildew not being the threat it once was (Makepeace et al. 2007) and the importance of this resistance source is clear by the presence of the mlo allele in approximately 70% of European spring barley varieties (Dreiseitl 2012; Jørgensen 1992). Even though mlo alleles can increase the threat of emerging diseases such as RLS (McGrann et al. 2014) the overriding importance of this mutation to spring barley production means that it is essential to develop varieties that are able to mitigate the negative effect of mlo on RLS (Brown and Rant 2013). Limiting the effect of mlo on the necrotrophic phase of fungal diseases is an important breeding target to protect the use of mlo in commercial spring barley. Recent studies by Aghnoum et al. (2019) suggest that the genetic background of barley varieties can mitigate mlo-mediated susceptibility to blast disease. Through mutation analysis of an experimental mlo barley line, Jansen et al. (2007) demonstrated the presence of genetic loci that can moderate the effects of mlo on necrotrophic diseases. The mutant alleles, emr1 and emr2, both reduce mlo-mediated enhanced susceptibility to M. oryzae but have no effect on powdery mildew resistance (Jansen et al. 2007; Jansen and Schaffrath 2009). This suggests that there are genetic loci that have contrasting effects on the different pathways regulated by mlo, as previously suggested (McGrann et al. 2015a). Mutant emr1mlo5 plants showed reduced RLS development, but this reduction in disease was not accompanied by a reduction in fungal biomass. Similar findings were observed in barley ror mutants where both ror1 and ror2 reduce RLS expression in leaves but do not lower levels of R. collo-cygni biomass (McGrann et al. 2014). Not all genes that affect the expression of RLS interact with mlo. Mutant alleles of the cyclic nucleotide gated channel NEC1 lower RLS symptoms in wild type MLO plants but do not affect disease levels in a nec1 mlo double mutant (McGrann et al. 2015a) indicating that multiple pathways affect the expression of RLS symptoms and that mlo may only affect some of these pathways. The emr1 mutant contains at least one other mutant locus, iwa1, which results in defective wax biosynthesis (Weidenbach et al. 2014). Whether or not altered surface wax composition affects RLS is currently unknown but emr1mlo5 mutants had reduced relative chlorophyll content and were more sensitive to chloroplastic ROS. Defects in ROR genes and NEC1 result in mis-regulated ROS and affect the transition of R. collo-cygni from asymptomatic growth to necrotrophic development. Changes in foliar chlorophyll content associated with the onset of senescence have been linked to RLS symptom expression (Schützendübel et al. 2008; McGrann and Brown 2018). Host responses associated with foliar senescence processes are regulated during R. collo-cygni infection suggesting that this pathogen triggers premature senescence in the plant (Sjökvist et al. 2019). Transgenic barley plants that have delayed leaf senescence due to over-expression of a Stress-induced NAC1 transcription factor show increased resistance against this disease (McGrann et al. 2015b). It is possible that defects in relative chlorophyll content, senescence and ROS balance in emr1mlo5 plants affect the transition of R. collo-cygni entering its necrotrophic phase. Determining the function of the gene at the EMR1 locus responsible for the effects on RLS symptom development may provide further insights into how plant genetics trigger this fungus to change growth habit and cause disease symptoms. Whilst there is promise in modern commercial spring barley varieties for good RLS resistance in plants with mlo mildew resistance, a better understanding of the mechanisms and genetics controlling these phenotypes is required. Abiotic stress is known to affect expression of RLS (Peraldi et al. 2014; Makepeace et al. 2008; McGrann and Brown 2018) and Makepeace et al. (2007) suggested that the mlo effects on diseases such as RLS are influenced by environmental conditions. This finding is supported by the emr1mlo5 field trial data, where statistically significant differences between mlo, wild type MLO (cv. Ingrid) and emr1mlo5 were not recorded, despite the trend in RLS levels observed being similar to results obtained in the controlled environment tests. Assessing RLS development in varieties with good RLS resistance in mlo backgrounds in multiple environments is essential to confirm that the effect of genes moderating mlo-mediated susceptibility to RLS is not environmentally sensitive. This could be particularly important considering genetic loci such as EMR1, which are able to moderate the effect of mlo on RLS symptom expression, but do not appear to affect the accumulation of fungal biomass (McGrann et al. 2014, 2015a). As such R. collo-cygni DNA levels may remain high in spring barley varieties with mlo which have low levels of RLS. This could have negative consequences on grain quality and could result in epidemics when environmental conditions are favourable for the expression of RLS symptoms. The emr1 mutation has potential to mitigate the effect of mlo on promoting RLS symptom production in spring barley but further experimental evidence is needed to evaluate how the environment influences the effects of genetic loci that interaction with the mlo-response against this disease.
{"extraction_info": {"found_math": false, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 0, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8304859399795532, "perplexity": 10591.222443201179}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.3, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499911.86/warc/CC-MAIN-20230201045500-20230201075500-00299.warc.gz"}
http://24solarhome.com/tag/efficiency/
# Solar Panels from Used solar battery Researchers at Massachusetts Institute of Technology (MIT) have demonstrated a procedure to convert used lead acid batteries from automobiles into solar panels with the help of a solar power company. A single battery can be used to produce solar panels for as many as 30 homes. It must be noted that with the advancement in solar battery technology it is expected that 200 million lead acid batteries would be retired soon from USA alone. This development shows us a way forward to reusing a huge resource of lead that would be otherwise go to dumping sites. A material that is making this possible is organo lead halide perovskite. A layer of perovskite only 1/2 a micrometer thick is enough to produce a solar panel and does not require a very high manufacturing process like for other silicon based solar panels. One might think that this is another experimental material that achieves an efficiency in single digits. But this is hardly the case. In just a few years of research perovskite based solar cells have achieved efficiency of more than 19%. It is expected by the end of 2014 the efficiency would cross the psychological barrier of 20%. 1 Battery Provides Solar Panels for 30 Homes Perovskite (source: Wikipedia)  is a calcium titanium oxide mineral species composed of calcium titanate, with the chemical formula CaTiO3. The mineral was discovered in the Ural Mountains of Russia by Gustav Rose in 1839 and is named after Russian mineralogist Lev Perovski (1792–1856). It lends its name to the class of compounds which have the same type of crystal structure as CaTiO3 known as the perovskite structure. The perovskite crystal structure was first described by Victor Goldschmidt in 1926, in his work on tolerance factors. The crystal structure was later published in 1945 from X-ray diffraction data on barium titanate by Helen Dick Megaw. Efficiency of a solar panel is the ratio of the electrical energy produced to the incident solar radiation e.g. a 20% efficient solar panel of 1 m2 area would produce 200 W when the incident solar radiation reaches a level of 1000 W/m2. # First Solar Sets New CdTe Solar Cell Efficiency Record First Solar Inc. an American manufacturer of thin film Photovoltaic (PV) modules announced that it has achieved a record efficiency of 21% for its CdTe Solar Cells. The previous best achieved by the company was 20.4% in Feb of 2014. The record has been accepted by US Department of Energy's National Renewable Energy Laboratory (NREL) and included in "Best Research Cell Efficiency" reference chart. First solar aims to achieve an efficiency of 22% by 2015. Thin Film Solar Panels Note: Efficiency of a Solar Cell is the ratio of the Electrical Energy produced to the incident Solar Energy e.g. if the incident Solar Radiation is 1000 W/m2 the Electrical Energy produced by a 21% efficient Solar Panel of 1 m2 area is 210 W (neglecting the various losses that might be encountered). # Cave Houses Around the World It is well known that cave houses provide a noise free and weather proof environment (cool during the summer and warm during the winter). These structures also shield one from man-made Electromagnetic radiation that is present everywhere. Furthermore, since most of the materials used in construction are local, the environmental impact is minimal. Shown below are some cave houses from around the world. Some are simple dwellings with the most basic necessities, while other have running water, electricity and wireless access. Cave Hotel Cave House Cave House Cave House in Baluchistan Cave House in Utah # Earthquake Resistant and Energy Efficient Homes - PAKSBAB After the devastating earthquake of 2005 which destroyed nearly half a million rural homes in Pakistan, there was an urgent need to build earthquake resistant homes. Thus came into being PAKSBAB which is the short form of Pakistan Straw Bale and Appropriate Buildings. So far PAKSBAB, from its limited resources, has build 27 homes in northern parts of Pakistan. These homes have been built using indigenous resources and by training the local people in construction of straw bale houses. A typical straw bale house has an area of 576 square feet and costs $3000 on average. This turns out to be$5.2/square feet which is less than half the cost of brick and mortar houses. A typical home comprises of two rooms, a verandah and a kitchen and requires around 1200 hours of labour. So 6 people working for 8 hours daily can construct a straw bale house in 25 days! The main advantages of straw bale houses over brick and mortar houses are highlighted below. 1. Energy efficiency, since straw is a good insulator 2. Non toxic products are used (light straw, clay and wood) 3. Cheap materials are used resulting in a cost that is half that of a regular house 4. Resistant to earthquakes Tightly packed walls and a gravel weighted foundation creates better weather-proofing Twice as energy efficient as a conventional house, straw bale makes for environmentally friendly earthquake-proof homes Clay-plaster reinforced, a fabricated straw bale house costs half the expenses of modern building for every square foot Straw bales required for the construction of these energy efficient and earthquake resilient homes are built using manually operated farm jacks and locally manufactured compression moulds. Furthermore the local industry is being encouraged to supply straw bales and other materials required for these projects. Additional appropriate building methods that PAKSBAB is promoting include passive solar, rainwater catchment, solar lamps, high-efficiency cooking and heating, and the use of natural building materials such as light straw clay, wattle and daub, and cob. # Quaid-e-Azam Solar Park - ROI The government of Pakistan has recently launched the Quaid-e-Azam Solar Park in the Cholistan desert near Bahawalpur. The project aims to produce 100 MW of electrical energy by end of 2014 and 1000 MW by end of 2016. This is a small step in the right direction. Countries like India, China and Germany are much ahead in the game with installed solar projects of 2600 MW, 20000 MW and 36000 MW respectively. Let us take a closer look at the price that we will have to pay for the energy produced. The cost of the 100 MW project is around $131 million, that is the price per Watt is$1.31. That seems to be quite good, lets look closely. We know that 400,000 panels are to be installed in the first phase to produce 100 MW of electrical energy. This means that each Solar Panel would produce 250 W and the cost of each panel would be \$327.5 or Rs.32750. Assuming that there is peak solar energy available for six hours daily, each solar panel would produce 1.5 kWhr of energy each day or 547.5 kWhr of energy per year. This amounts to 13687.5 kWhr of energy over a 25 year period (assuming that the performance of the Solar Panels does not degrade over the 25 year period). Now assuming that each unit of energy (kWhr) is sold at Rs.15 the total energy produced by the Solar Panel over its life period amounts to Rs.205312.5 i.e the revenue earned from selling electricity is 6.27 times the investment (205312.5/32750=6.27). Solar Park Bahawalpur In other words the investment is recovered in 4 years and you have free electricity for the remaining 21 years. Please note that the above calculations do not include the operational costs, if any. Also, the above analysis assumes that the performance of the Solar Panels does not degrade over its life time. Final Comment: The location of the proposed project does not seem to be optimum as Bahawalpur is receiving 2000 kWhr per squared meter per year as opposed to vast expanses of Balochistan that receive 2200 kWhr per squared meter per year. # Do Numbers Make Sense Press Release "The Fraunhofer Institute for Solar Energy Systems ISE, Soitec, CEA-Leti and the Helmholtz Center Berlin have jointly announced that they have achieved a new world record for the conversion of sunlight into electricity using a new solar cell structure with four solar subcells. Surpassing competition after only over three years of research, and entering the roadmap at world class level, a new record efficiency of 44.7% was measured at a concentration of 297 suns. This indicates that 44.7% of the solar spectrum's energy, from ultraviolet through to the infrared, is converted into electrical energy. This is a major step towards reducing further the costs of solar electricity and continues to pave the way to the 50% efficiency roadmap". "These solar cells are used in concentrator photovoltaics (CPV), a technology which achieves more than twice the efficiency of conventional PV power plants in sun-rich locations. The terrestrial use of so-called III-V multi-junction solar cells, which originally came from space technology, has prevailed to realize highest efficiencies for the conversion of sunlight to electricity. In this multi-junction solar cell, several cells made out of different III-V semiconductor materials are stacked on top of each other. The single subcells absorb different wavelength ranges of the solar spectrum". ## A Closer Look at the Numebrs Let us now try to validate the numbers given above. Power is the product of voltage and current. Pideal=Voc Isc=(4.165)(0.1921)=0.8001Watts Pmax=Pideal*FF=(0.8001)(0.8650)=0.6921Watts This is the power produced by 5.20mm2 of solar cell. 1mm2 of solar cell would produce 0.1331Watts. 1m2 of solar cell would produce 133.1kW of solar energy. This is the power generated due to 297.3 suns. A single sun would produce 133.1kW/297.3=447.67Watts of power. This gives us an efficiency of 447.67/1000=0.4477=44.77%. # Solar Cell Temperature and Efficiency It is a common misconception that higher the temperature higher is the output of the solar cell. This is not true as the efficiency of a solar cell decreases with increase in temperature and lower efficiency results in lower output power. So in fact a bright sunny day, with sun rays perpendicular to the solar panel and a cool weather is the ideal combination for higher performance of a solar panel. Let us now look at this in a bit more detail. There are two basic reasons for decrease in efficiency due to increase in temperature.  One is the decrease in the band gap energy (Eg) and the other is the decrease in open circuit voltage (Voc) with increase in temperature. The relationship between band gap energy and and temperature is quite straightforward and is given as. $E_{g} = E_{g}(0)+\genfrac{}{}{1}{0}{\alpha T^2}{T+\beta}$ One might argue that the decrease in band gap would allow for more carriers to be transferred to the conduction band and yield a higher output power. However this is not true as the output power is the product of current and voltage and a lower voltage would reduce the power. In fact there is an ideal range of band gap which produces the maximum energy. Going too high or too low would not yield the optimum results. Next we turn out attention to the open circuit voltage Voc. The relationship between temperature and open circuit voltage is not that straightforward. At first it might seem the open circuit voltage increases with increase in temperature as shown in the expression below. $V_{oc} = \genfrac{}{}{1}{0}{k_{B} T}{e} ln \biggl[1+\genfrac{}{}{1}{0}{J_{L}}{J_{s}}\biggr]$ But in reality this is not the case. Increase in temperature results in increase in intrinsic carrier concentration n which in turn results in higher reverse bias saturation current Js. There is a squared relationship here so increase in intrinsic carrier concentration would cause a very large increase in reverse bias saturation current. And as is evident from the above formula this causes a decrease in open circuit voltage. This is also shown in the figure below. So to conclude a bright sunny morning in winter might not be the worst time to produce some solar energy (provided you have got your solar panel tilt right 🙂 ). # Fill Factor and Efficiency The Efficiency of a solar cell is an important metric that determines how much of the incident solar energy is converted to useful electrical energy e.g. a 1m2 solar panel with 15% Efficiency would convert a radiant energy of 1000W/m2 into 150W of useful electrical energy. The Efficiency of a solar cell is sometimes defined in terms of the Fill Factor (FF) which is defined as. $FF = \genfrac{}{}{1}{0}{J_{max} V_{max}}{J_{sc} V_{oc}}$ Simply put its the ratio of area defined by (Vmax, Imax) to the area defined by (Voc, Isc) on the IV curve. And the Efficiency in terms of the Fill Factor is defined as. $\eta = \genfrac{}{}{1}{0}{J_{sc} V_{oc} FF}{P_{s}}$ The expression for Efficiency can be simplified by substituting FF in the above equation. $\eta = \genfrac{}{}{1}{0}{J_{sc} V_{oc}}{P_{s}} \genfrac{}{}{1}{0}{J_{max} V_{max}}{J_{sc} V_{oc}}$ or $\eta = \genfrac{}{}{1}{0}{J_{max} V_{max}}{P_{s}}$ Let us now look at some practical values for Efficiency and Fill Factor. $\eta = \genfrac{}{}{1}{0}{J_{sc} V_{oc} FF}{P_{s}}$ $\eta = \genfrac{}{}{1}{0}{(400) (0.70) (0.84)} {1000}$ $\eta = 0.2352$ This is the Efficiency ignoring certain practical issues of solar cells. Thus the typical Efficiency of mono-crystalline solar cells would be somewhat lower (15%-20%). Note: 1. Vmax, Imax is the Voltage and Current respectively at the Maximum Power Point on the IV curve. Remember that Power is just the product of Voltage and Current. 2. From basic circuit theory, the power delivered from or to a device is optimized where the derivative (graphically, the slope) dI/dV of the I-V curve is equal and opposite the I/V ratio (where dP/dV=0). This is known as the Maximum Power Point (MPP) and corresponds to the "knee" of the curve. 3. A solar charge controller is used to charge the batteries from the solar panel operating at its Maximum Power Point. 4. The more rapid the drop in Current as the Voltage approaches the Open Circuit Voltage the closer will be the Fill Factor to the ideal value of 100%.
{"extraction_info": {"found_math": true, "script_math_tex": 9, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 9, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.6209750771522522, "perplexity": 1224.8084722897147}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-21/segments/1620243989874.84/warc/CC-MAIN-20210518222121-20210519012121-00002.warc.gz"}
http://clay6.com/qa/43663/a-thin-semi-circular-ring-of-radius-r-has-a-positive-charge-q-distributed-u
Browse Questions # A thin semi-circular ring of radius r has a positive charge q distributed uniformly over it. The net field E at the centre O is $(C) \large\frac{-q}{2 \pi ^2 \in_0 r^2} \hat i$
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8087001442909241, "perplexity": 603.2412296285187}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2016-50/segments/1480698542002.53/warc/CC-MAIN-20161202170902-00067-ip-10-31-129-80.ec2.internal.warc.gz"}
https://www.cheenta.com/least-possible-value-problem-amc-10a-2019-quesstion19/
What is the NO-SHORTCUT approach for learning great Mathematics? # How to Pursue Mathematics after High School? For Students who are passionate for Mathematics and want to pursue it for higher studies in India and abroad. Contents [hide] Try this beautiful problem from Algebra based on Least Possible Value. ## Least Possible Value - AMC-10A, 2019- Problem 19 What is the least possible value of $((x+1)(x+2)(x+3)(x+4)+2019)$ where (x) is a real number? • $(2024)$ • $(2018)$ • $(2020)$ ### Key Concepts Algebra least value Answer: $(2018)$ AMC-10A (2019) Problem 19 Pre College Mathematics ## Try with Hints To find out the least positive value of $(x+1)(x+2)(x+3)(x+4)+2019$, at first we have to expand the expression .$((x+1)(x+2)(x+3)(x+4)+2019)$ $\Rightarrow (x+1)(x+4)(x+2)(x+3)+2019=(x^2+5x+4)(x^2+5x+6)+2019)$ Let us take $((x^2+5x+5=m))$ then the above expression becomes $((m-1)(m+1)+2019)$ $\Rightarrow m^2-1+2019$ $\Rightarrow m^2+2018$ Can you now finish the problem .......... Clearly in $(m^2+2018).......(m^2)$ is positive ( squares of any number is non-negative) and least value is 0 can you finish the problem........ Therefore minimum value of $m^2+2108$ is $2018$ since $m^2 \geq 0$ for all m belongs to real . ## What to do to shape your Career in Mathematics after 12th? From the video below, let's learn from Dr. Ashani Dasgupta (a Ph.D. in Mathematics from the University of Milwaukee-Wisconsin and Founder-Faculty of Cheenta) how you can shape your career in Mathematics and pursue it after 12th in India and Abroad. These are some of the key questions that we are discussing here: • What are some of the best colleges for Mathematics that you can aim to apply for after high school? • How can you strategically opt for less known colleges and prepare yourself for the best universities in India or Abroad for your Masters or Ph.D. Programs? • What are the best universities for MS, MMath, and Ph.D. Programs in India? • What topics in Mathematics are really needed to crack some great Masters or Ph.D. level entrances? • How can you pursue a Ph.D. in Mathematics outside India? • What are the 5 ways Cheenta can help you to pursue Higher Mathematics in India and abroad? ## Want to Explore Advanced Mathematics at Cheenta? Cheenta has taken an initiative of helping College and High School Passout Students with its "Open Seminars" and "Open for all Math Camps". These events are extremely useful for students who are really passionate for Mathematic and want to pursue their career in it. To Explore and Experience Advanced Mathematics at Cheenta This site uses Akismet to reduce spam. Learn how your comment data is processed.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.287310391664505, "perplexity": 1750.4833126934143}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046152085.13/warc/CC-MAIN-20210805224801-20210806014801-00611.warc.gz"}
https://socratic.org/questions/how-do-you-find-the-asymptotes-for-1-x-2x-2-5x-3
Precalculus Topics # How do you find the asymptotes for (1-x)/(2x^2-5x-3)? Feb 25, 2016 Two vertical asymptotes are $2 x + 1 = 0$ and $x - 3 = 0$. #### Explanation: To find all the asymptotes for function y=(1-x)/(2x^2-5x−3), let us first start with vertical asymptotes, which are given by putting denominator equal to zero or (2x^2-5x−3)=0. Let us find factors of (2x^2-5x−3) by splitting middle term in $- 6 s$ and $x$ i.e. (2x^2-6x+x−3=2x(x-3)+1(x-3)=(2x+1)(x-3). As the factors of denominators are $\left(2 x + 1\right)$ and $\left(x - 3\right)$, two vertical asymptotes are $2 x + 1 = 0$ and $x - 3 = 0$. As the highest degree of numerator is less than that of denominator, there is no horizontal or slanting asymptote. ##### Impact of this question 97 views around the world
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 12, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.8396652340888977, "perplexity": 829.9901368022817}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547583835626.56/warc/CC-MAIN-20190122095409-20190122121409-00566.warc.gz"}
http://polymathprojects.org/
# The polymath blog ## January 20, 2014 ### Two polymath (of a sort) proposed projects Filed under: discussion,polymath proposals — Gil Kalai @ 5:20 pm Tags: , , This post is meant to propose and discuss a polymath project and a sort of polymath project. # I. A polymath proposal: Convex hulls of real algebraic varieties. One of the interesting questions regarding the polymath endeavor was: Can polymath be used to develop a theory/new area? My idea is to have a project devoted to develop a theory of “convex hulls of real algebraic varieties”. The case where the varieties are simply a finite set of points is a well-developed area of mathematics – the theory of convex polytopes, but the general case was not studied much. I suppose that for such a project the first discussions will be devoted to raise questions/research directions. (And mention some works already done.) In general (but perhaps more so for an open-ended project), I would like to see also polymath projects which are on longer time scale than existing ones but perhaps less intensive, and that people can “get in” or “spin-off” at will in various times. # II. A polymath-of-a-sort proposal: Statements about the Riemann Hypothesis The Riemann hypothesis is arguably the most famous open question in mathematics. My view is that it is premature to try to attack the RH by a polymath project (but I am not an expert and, in any case, a project of this kind is better conducted with some specific program in mind). I propose something different. In a sort of polymath spirit the project I propose invite participants, especially professional mathematicians who thought about the RH over the years,  to share their thoughts about RH. Ideally each comment will be 1) One or a few paragraphs long 2) Well-thought, focused and rather polished A few comments by the same contributors are also welcome. To make it clear, the thread I propose is not going to be a research thread and also not a place for further discussions beyond some clarifying questions. Rather it is going to be a platform for interested mathematician to make statements and expressed polished thoughts about RH. (Also, if adopted, maybe we will need a special name for such a thing.) ____________________ This thread is not launching any of the two suggested projects, but rather a place to discuss further these proposals. For the second project,  it will be better still if the person who runs it will be an expert in the area, and certainly not an ignorant. For the first project, maybe there are better ideas for areas/theories appropriate for polymathing. ## November 4, 2013 ### Polymath9: P=NP? (The Discretized Borel Determinacy Approach) Filed under: polymath proposals — Gil Kalai @ 2:07 pm Tags: , Tim Gowers Proposed and launched a new polymath proposal aimed at a certain approach he has for proving that $NP \ne P$. ## September 20, 2013 ### Polymath8 – A Success ! Filed under: news — Gil Kalai @ 5:58 pm Tags: The main objectives of the polymath8 project, initiated by Terry Tao  back in June, were “to understand the recent breakthrough paper of Yitang Zhang establishing an infinite number of prime gaps bounded by a fixed constant ${H}$, and then to lower that value of ${H}$ as much as possible.” Polymath8 was a remarkable success! Within two months the best value of H that was 70,000,000 in Zhang’s proof was reduced to 5,414. Moreover, the polymath setting looked advantageous for this project, compared to traditional ways of doing mathematics. (I have written a post with some more details and thoughts about it, looked from a distance.) ## August 9, 2013 ### Polymath7 research thread 5: the hot spots conjecture Filed under: hot spots,research — Terence Tao @ 7:22 pm Tags: This post is the new research thread for the Polymath7 project to solve the hot spots conjecture for acute-angled triangles, superseding the previous thread; this project had experienced a period of low activity for many months, but has recently picked up again, due both to renewed discussion of the numerical approach to the problem, and also some theoretical advances due to Miyamoto and Siudeja. On the numerical side, we have decided to focus first on the problem of obtaining validated upper and lower bounds for the second Neumann eigenvalue ${\mu_2}$ of a triangle ${\Omega=ABC}$. Good upper bounds are relatively easy to obtain, simply by computing the Rayleigh quotient of numerically obtained approximate eigenfunctions, but lower bounds are trickier. This paper of Liu and Oshii has some promising approaches. After we get good bounds on the eigenvalue, the next step is to get good control on the eigenfunction; some approaches are summarised in this note of Lior Silberman, mainly based on gluing together exact solutions to the eigenfunction equation in various sectors or disks. Some recent papers of Kwasnicki-Kulczycki, Melenk-Babuska, and Driscoll employ similar methods and may be worth studying further. However, in view of the theoretical advances, the precise control on the eigenfunction that we need may be different from what we had previously been contemplating. These two papers of Miyamoto introduced a promising new method to theoretically control the behaviour of the second Neumann eigenfunction ${u_2}$, by taking linear combinations of that eigenfunction with other, more explicit, solutions to the eigenfunction equation ${\Delta u = - \mu_2 u}$, restricting that combination to nodal domains, and then computing the Dirichlet energy on each domain. Among other things, these methods can be used to exclude critical points occurring anywhere in the interior or on the edges of the triangle except for those points that are close to one of the vertices; and in this recent preprint of Siudeja, two further partial results on the hot spots conjecture are obtained by a variant of the method: • The hot spots conjecture is established unconditionally for any acute-angled triangle which has one angle less than or equal to ${\pi/6}$ (actually a slightly larger region than this is obtained). In particular, the case of very narrow triangles have been resolved (the dark green region in the area below). • The hot spots conjecture is also established for any acute-angled triangle with the property that the second eigenfunction ${u_2}$ has no critical points on two of the three edges (excluding vertices). So if we can develop more techniques to rule out critical points occuring on edges (i.e. to keep eigenfunctions monotone on the edges on which they change sign), we may be able to establish the hot spots conjecture for a further range of triangles. In particular, some hybrid of the Miyamoto method and the numerical techniques we are beginning to discuss may be a promising approach to fully resolve the conjecture. (For instance, the Miyamoto method relies on upper bounds on ${\mu_2}$, and these can be obtained numerically.) The arguments of Miyamoto also allow one to rule out critical points occuring for most of the interior points of a given triangle; it is only the points that are very close to one of the three vertices which we cannot yet rule out by Miyamoto’s methods. (But perhaps they can be ruled out by the numerical methods we are also developing, thus giving a hybrid solution to the conjecture.) Below the fold I’ll describe some of the theoretical tools used in the above arguments. ## June 4, 2013 ### Polymath proposal: bounded gaps between primes Filed under: planning,polymath proposals — Terence Tao @ 4:31 am Two weeks ago, Yitang Zhang announced his result establishing that bounded gaps between primes occur infinitely often, with the explicit upper bound of 70,000,000 given for this gap.  Since then there has been a flurry of activity in reducing this bound, with the current record being 4,802,222 (but likely to improve at least by a little bit in the near future). It seems that this naturally suggests a Polymath project with two interrelated goals: 1. Further improving the numerical upper bound on gaps between primes; and 2. Understanding and clarifying Zhang’s argument (and other related literature, e.g. the work of Bombieri, Fouvry, Friedlander, and Iwaniec on variants of the Elliott-Halberstam conjecture). Part 1 of this project splits off into somewhat independent sub-projects: 1. Finding narrow prime admissible tuples of a given cardinality (or, dually, finding large prime admissible tuples in a given interval).  This part of the project would be relatively elementary in nature, relying on combinatorics, elementary number theory, computer search, and perhaps some clever algorithm design.  (Scott Morrison has already been hosting a de facto project of this form at this page, and is happy to continue doing so). 2. Solving a calculus of variations problem associated with the Goldston-Yildirim-Pintz argument (discussed at this blog post, or in this older survey of Soundararajan) [in particular, this could lead to an improvement of a certain key parameter $k_0$, currently at 341,640, even without any improvement in the parameter $\varpi$ mentioned in part 3. below.] 3. Delving through the “hard” part of Zhang’s paper in order to improve the value of a certain key parameter $\varpi$ (which Zhang sets at 1/1168, but is likely to be enlargeable). Part 2 of this project could be run as an online reading seminar, similar to the online reading seminar of the Furstenberg-Katznelson paper that was part of the Polymath1 project.  It would likely focus on the second half of Zhang’s paper and would fit well with part 1.3.  I could run this on my blog, and this existing blog post of mine could be used for part 1.2. As with other polymath projects, it is conceivable that enough results are obtained to justify publishing one or more articles (which, traditionally, we would publish under the D.H.J. Polymath pseudonym).  But it is perhaps premature to discuss this possibility at this early stage of the process. Anyway, I would be interested to gauge the level of interest and likely participation in these projects, together with any suggestions for improving the proposal or other feedback. ## March 2, 2013 ### Polymath proposal (Tim Gowers): Randomized Parallel Sorting Algorithm Filed under: polymath proposals — Gil Kalai @ 4:41 pm From Holroyd’s sorting networks picture gallery A celebrated theorem of Ajtai, Komlos and Szemeredi describes a sorting network for  $n$ numbers of depth $O(log N)$. rounds where in each runs $n/2$. Tim Gowers proposes to find collectively a randomized sorting with the same properties. ## February 14, 2013 ### Next Polymath Project(s): What, When, Where? Filed under: polymath proposals — Gil Kalai @ 3:26 pm Let us have a little discussion about it. We may also discuss both general and specific open research mathematical projects which are of different flavor/rules. Proposals for polymath projects appeared on this blog,  in this post on Gowers’s blog, and in several other places. ## September 10, 2012 ### Polymath7 research threads 4: the Hot Spots Conjecture Filed under: hot spots,research — Terence Tao @ 7:28 pm It’s time for another rollover of the  Polymath7 “Hot Spots” conjecture, as the previous research thread has again become full. Activity has now focused on a numerical strategy to solve the hot spots conjecture for all acute angle triangles ABC.  In broad terms, the strategy (also outlined in this document) is as follows.   (I’ll focus here on the problem of estimating the eigenfunction; one also needs to simultaneously obtain control on the eigenvalue, but this seems to be to be a somewhat more tractable problem.) 1. First, observe that as the conjecture is scale invariant, the only relevant parameters for the triangle ABC are the angles $\alpha,\beta,\gamma$, which of course lie between 0 and $\pi/2$ and add up to $\pi$.  We can also order $\alpha \leq \beta \leq \gamma$, giving a parameter space which is a triangle between the values $(\alpha,\beta,\gamma) = (0,\pi/2,\pi/2), (\pi/4,\pi/4,\pi/2), (\pi/3,\pi/3,\pi/3)$. 2. The triangles that are too close to the degenerate isosceles triangle ${}(0,\pi/2,\pi/2)$ or the equilateral triangle ${}(\pi/3,\pi/3,\pi/3)$ need to be handled by analytic arguments.  (Preliminary versions of these arguments can be found here and Section 6 of these notes  respectively, but the constants need to be made explicit (and as strong as possible)). 3. For the remaining parameter space, we will use a sufficiently fine discrete mesh of angles $(\alpha,\beta,\gamma)$; the optimal spacing of this mesh is yet to be determined. 4. For each triplet of angles in this mesh,  we partition the triangle ABC (possibly after rescaling it to a reference triangle $\hat \Omega$, such as the unit right-angled triangle) into smaller subtriangles, and approximate the second eigenfunction $w$ (or the rescaled triangle $\hat w$) by the eigenfunction $w_h$ (or $\hat w_h$) for a finite element restriction of the eigenvalue problem, in which the function is continuous and piecewise polynomial of low degree (probably linear or quadratic) in each subtriangle; see Section 2.2 of these notes.    With respect to a suitable basis, $w_h$ can be represented by a finite vector $u_h$. 5. Using numerical linear algebra methods (such as Lanczos iteration) with interval arithmetic, obtain an approximation $\tilde u$ to $u_h$, with rigorous bounds on the error between the two.  This gives an approximation to $w_h$ or $\hat w_h$ with rigorous error bounds (initially of L^2 type, but presumably upgradable). 6. After (somehow) obtaining a rigorous error bound between $w$ and $w_h$ (or $\hat w$ and $\hat w_h$), conclude that $w$ stays far from its extremum when one is sufficiently far away from the vertices A,B,C of the triangle. 7. Using $L^\infty$ stability theory of eigenfunctions (see Section 5 of these notes), conclude that $w$ stays far from its extremum even when $(\alpha,\beta,\gamma)$ is not at a mesh point.  Thus, the hot spots conjecture is not violated away from the vertices.  (This argument should also handle the vertex that is neither the maximum nor minimum value for the eigenfunction, leaving only the neighbourhoods of the two extremising vertices to deal with.) 8. Finally, use an analytic argument (perhaps based on these calculations) to show that the hot spots conjecture is also not violated near an extremising vertex. This all looks like it should work in principle, but it is a substantial amount of effort; there is probably still some scope to try to simplify the scheme before we really push for implementing it. ## July 13, 2012 ### Minipolymath4 project, second research thread Filed under: polymath proposals — Terence Tao @ 7:49 pm It’s been almost 24 hours since the mini-polymath4 project was opened in order to collaboratively solve Q3 from the 2012 IMO.  In that time, the first part of the question was solved, but the second part remains open.  In other words, it remains to show that for any sufficiently large $k$ and any $N$, there is some $n \geq (1.99)^k$ such that the second player B in the liar’s guessing game cannot force a victory no matter how many questions he asks. As the previous research thread is getting quite lengthy (and is mostly full of attacks on the first part of the problem, which is now solved), I am rolling over to a fresh thread (as is the standard practice with the polymath projects).  Now would be a good time to summarise some of the observations from the previous thread which are still relevant here.  For instance, here are some relevant statements made in previous comments: 1. Without loss of generality we may take N=n+1; if B can’t win this case, then he certainly can’t win for any larger value of N (since A could simply restrict his number x to a number up to n+1), and if B can win in this case (i.e. he can eliminate one of the n+1 possibilities) he can also perform elimination for any larger value of N by partitioning the possible answers into n+1 disjoint sets and running the N=n+1 strategy, and then one can iterate one’s way down to N=n+1. 2. In order to show that B cannot force a win, one needs to show that for any possible sequence of questions B asks in the N=n+1 case, it is possible to construct a set of responses by A in which none of the n+1 possibilities of x are eliminated, thus each x belongs to at least one of each block of k+1 consecutive sets that A’s answers would indicate. 3. It may be useful to look at small examples, e.g. can one show that B cannot win when k=1, n=1? Or when k=2, n=3? It seems that some of the readers of this blog have already obtained a solution to this problem from other sources, or from working separately on the problem, so I would ask that they refrain from giving spoilers for this question until at least one solution has been arrived at collaboratively. Also, participants are encouraged to edit the wiki as appropriate with new developments and ideas, and participate in the discussion thread for any meta-discussion about the polymath project. ## July 12, 2012 ### Minipolymath4 project: IMO 2012 Q3 Filed under: research — Terence Tao @ 10:00 pm Tags: This post marks the official opening of the mini-polymath4 project to solve a problem from the 2012 IMO.  This time, I have selected Q3, which has an interesting game-theoretic flavour to it. Problem 3.   The liar’s guessing game is a game played between two players $A$ and $B$.  The rules of the game depend on two positive integers $k$ and $n$ which are known to both players. At the start of the game, $A$ chooses two integers $x$ and $N$ with $1 \leq x \leq N$.  Player $A$ keeps $x$ secret, and truthfully tells $N$ to player $B$.  Player $B$ now tries to obtain information about $x$ by asking player A questions as follows.  Each question consists of $B$ specifying an arbitrary set $S$ of positive integers (possibly one specified in a previous question), and asking $A$ whether $x$ belongs to $S$.  Player $B$ may ask as many such questions as he wishes.  After each question, player $A$ must immediately answer it with yes or no, but is allowed to lie as many times as she wishes; the only restriction is that, among any $k+1$ consecutive answers, at least one answer must be truthful. After $B$ has asked as many questions as he wants, he must specify a set $X$ of at most $n$ positive integers.  If $x$ belongs to $X$, then $B$ wins; otherwise, he loses.  Prove that: 1. If $n \geq 2^k$, then $B$ can guarantee a win. 2. For all sufficiently large $k$, there exists an integer $n \geq 1.99^k$ such that $B$ cannot guarantee a win. The comments to this post shall serve as the research thread for the project, in which participants are encouraged to post their thoughts and comments on the problem, even if (or especially if) they are only partially conclusive.  Participants are also encouraged to visit the discussion thread for this project, and also to visit and work on the wiki page to organise the progress made so far. This project will follow the general polymath rules.  In particular: 1. All are welcome. Everyone (regardless of mathematical level) is welcome to participate.  Even very simple or “obvious” comments, or comments that help clarify a previous observation, can be valuable. 2. No spoilers! It is inevitable that solutions to this problem will become available on the internet very shortly.  If you are intending to participate in this project, I ask that you refrain from looking up these solutions, and that those of you have already seen a solution to the problem refrain from giving out spoilers, until at least one solution has already been obtained organically from the project. 3. Not a race. This is not intended to be a race between individuals; the purpose of the polymath experiment is to solve problems collaboratively rather than individually, by proceeding via a multitude of small observations and steps shared between all participants.   If you find yourself tempted to work out the entire problem by yourself in isolation, I would request that you refrain from revealing any solutions you obtain in this manner until after the main project has reached at least one solution on its own. 4. Update the wiki. Once the number of comments here becomes too large to easily digest at once, participants are encouraged to work on the wiki page to summarise the progress made so far, to help others get up to speed on the status of the project. 5. Metacomments go in the discussion thread. Any non-research discussions regarding the project (e.g. organisational suggestions, or commentary on the current progress) should be made at the discussion thread. 6. Be polite and constructive, and make your comments as easy to understand as possible. Bear in mind that the mathematical level and background of participants may vary widely. Have fun! Next Page » The Rubric Theme. Blog at WordPress.com.
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 0, "mathjax_asciimath": 0, "img_math": 0, "codecogs_latex": 0, "wp_latex": 76, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 0, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.7439523935317993, "perplexity": 812.3845060018973}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": false}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2014-35/segments/1409535921872.11/warc/CC-MAIN-20140909040043-00342-ip-10-180-136-8.ec2.internal.warc.gz"}
https://robotics.stackexchange.com/questions/7331/position-control-for-linear-model-of-quadrotor-problem-with-tracking-task
# position control for linear model of quadrotor (problem with tracking task) Lately, if you notice I have posted some questions regarding position tracking for nonlinear model. I couldn't do it. I've switched to linear model, hope I can do it. For regulation problem, the position control seems working but once I switch to tracking, the system starts oscillating. I don't know why. I have stated what I've done below hope someone guides me to the correct path. The linear model of the quadrotor is provided here which is \begin{align} \ddot{x} &= g \theta \ \ \ \ \ \ \ \ \ \ (1)\\ \ddot{y} &= - g \phi \ \ \ \ \ \ \ \ \ \ (2)\\ \ddot{z} &= \frac{U_{1}}{m} - g \\ \ddot{\phi} &= \frac{L}{J_{x}} U_{2} \\ \ddot{\theta} &= \frac{L}{J_{y}} U_{2} \\ \ddot{\psi} &= \frac{1}{J_{z}} U_{2} \\ \end{align} In this paper, the position control based on PD is provided. In the aforementioned paper, from (1) and (2) the desired angles $\phi^{d}$ and $\theta^{d}$ are obtained, therefore, \begin{align} \theta^{d} &= \frac{\ddot{x}^{d}}{g} \\ \phi^{d} &= - \frac{\ddot{y}^{d}}{g} \end{align} where \begin{align} \ddot{x}^{d} &= Kp(x^{d} - x) + Kd( \dot{x}^{d} - \dot{x} ) \\ \ddot{y}^{d} &= Kp(y^{d} - y) + Kd( \dot{y}^{d} - \dot{y} ) \\ U_{1} &= Kp(z^{d} - z) + Kd( \dot{z}^{d} - \dot{z} ) \\ U_{2} &= Kp(\phi^{d} - \phi) + Kd( \dot{\phi}^{d} - \dot{\phi} ) \\ U_{3} &= Kp(\theta^{d} - \theta) + Kd( \dot{\theta}^{d} - \dot{\theta} ) \\ U_{4} &= Kp(\psi^{d} - \psi) + Kd( \dot{\psi}^{d} - \dot{\psi} ) \\ \end{align} with regulation problem where $x^{d} = 2.5 m, \ y^{d} = 3.5 m$ and $z^{d} = 4.5 m$, the results are Now if I change the problem to the tracking one, the results are messed up. In the last paper, they state A saturation function is needed to ensure that the reference roll and pitch angles are within specified limits Unfortunately, the max value for $\phi$ and $\theta$ are not stated in the paper but since they use Euler angles, I believe $\phi$ in this range $(-\frac{\pi}{2},\frac{\pi}{2})$ and $\theta$ in this range $[-\pi, \pi]$ I'm using Euler method as an ODE solver because the step size is fixed. For the derivative, Euler method is used. This is my code %######################( PD Controller & Atittude )%%%%%%%%%%%%%%%%%%%% clear all; clc; dt = 0.001; t = 0; % initial values of the system x = 0; dx = 0; y = 0; dy = 0; z = 0; dz = 0; Phi = 0; dPhi = 0; Theta = 0; dTheta = 0; Psi = pi/3; dPsi = 0; %System Parameters: m = 0.75; % mass (Kg) L = 0.25; % arm length (m) Jx = 0.019688; % inertia seen at the rotation axis. (Kg.m^2) Jy = 0.019688; % inertia seen at the rotation axis. (Kg.m^2) Jz = 0.039380; % inertia seen at the rotation axis. (Kg.m^2) g = 9.81; % acceleration due to gravity m/s^2 errorSumX = 0; errorSumY = 0; errorSumZ = 0; errorSumPhi = 0; errorSumTheta = 0; % Set desired position for tracking task DesiredX = pose(:,1); DesiredY = pose(:,2); DesiredZ = pose(:,3); % Set desired position for regulation task % DesiredX(:,1) = 2.5; % DesiredY(:,1) = 5; % DesiredZ(:,1) = 7.2; dDesiredX = 0; dDesiredY = 0; dDesiredZ = 0; DesiredXpre = 0; DesiredYpre = 0; DesiredZpre = 0; dDesiredPhi = 0; dDesiredTheta = 0; DesiredPhipre = 0; DesiredThetapre = 0; for i = 1:6000 % torque input %&&&&&&&&&&&&( Ux )&&&&&&&&&&&&&&&&&& Kpx = 90; Kdx = 25; Kix = 0.0001; errorSumX = errorSumX + ( DesiredX(i) - x ); % Euler Method Derivative dDesiredX = ( DesiredX(i) - DesiredXpre ) / dt; DesiredXpre = DesiredX(i); Ux = Kpx*( DesiredX(i) - x ) + Kdx*( dDesiredX - dx ) + Kix*errorSumX; %&&&&&&&&&&&&( Uy )&&&&&&&&&&&&&&&&&& Kpy = 90; Kdy = 25; Kiy = 0.0001; errorSumY = errorSumY + ( DesiredY(i) - y ); % Euler Method Derivative dDesiredY = ( DesiredY(i) - DesiredYpre ) / dt; DesiredYpre = DesiredY(i); Uy = Kpy*( DesiredY(i) - y ) + Kdy*( dDesiredY - dy ) + Kiy*errorSumY; %&&&&&&&&&&&&( U1 )&&&&&&&&&&&&&&&&&& Kpz = 90; Kdz = 25; Kiz = 0; errorSumZ = errorSumZ + ( DesiredZ(i) - z ); dDesiredZ = ( DesiredZ(i) - DesiredZpre ) / dt; DesiredZpre = DesiredZ(i); U1 = Kpz*( DesiredZ(i) - z ) + Kdz*( dDesiredZ - dz ) + Kiz*errorSumZ; %####################################################################### %####################################################################### %####################################################################### % Desired Phi and Theta %disp('before') DesiredPhi = -Uy/g; DesiredTheta = Ux/g; %&&&&&&&&&&&&( U2 )&&&&&&&&&&&&&&&&&& KpP = 20; KdP = 5; KiP = 0.001; errorSumPhi = errorSumPhi + ( DesiredPhi - Phi ); % Euler Method Derivative dDesiredPhi = ( DesiredPhi - DesiredPhipre ) / dt; DesiredPhipre = DesiredPhi; U2 = KpP*( DesiredPhi - Phi ) + KdP*( dDesiredPhi - dPhi ) + KiP*errorSumPhi; %-------------------------------------- %&&&&&&&&&&&&( U3 )&&&&&&&&&&&&&&&&&& KpT = 90; KdT = 10; KiT = 0.001; errorSumTheta = errorSumTheta + ( DesiredTheta - Theta ); % Euler Method Derivative dDesiredTheta = ( DesiredTheta - DesiredThetapre ) / dt; DesiredThetapre = DesiredTheta; U3 = KpT*( DesiredTheta - Theta ) + KdP*( dDesiredTheta - dTheta ) + KiT*errorSumTheta; %-------------------------------------- %&&&&&&&&&&&&( U4 )&&&&&&&&&&&&&&&&&& KpS = 90; KdS = 10; KiS = 0; DesiredPsi = 0; dDesiredPsi = 0; U4 = KpS*( DesiredPsi - Psi ) + KdS*( dDesiredPsi - dPsi ); %###################( ODE Equations of Quadrotor )################### ddx = g * Theta; dx = dx + ddx*dt; x = x + dx*dt; %======================================================================= ddy = -g * Phi; dy = dy + ddy*dt; y = y + dy*dt; %======================================================================= ddz = (U1/m) - g; dz = dz + ddz*dt; z = z + dz*dt; %======================================================================= ddPhi = ( L/Jx )*U2; dPhi = dPhi + ddPhi*dt; Phi = Phi + dPhi*dt; %======================================================================= ddTheta = ( L/Jy )*U3; dTheta = dTheta + ddTheta*dt; Theta = Theta + dTheta*dt; %======================================================================= ddPsi = (1/Jz)*U4; dPsi = dPsi + ddPsi*dt; Psi = Psi + dPsi*dt; %======================================================================= %store the erro ErrorX(i) = ( x - DesiredX(i) ); ErrorY(i) = ( y - DesiredY(i) ); ErrorZ(i) = ( z - DesiredZ(i) ); ErrorPsi(i) = ( Psi - 0 ); X(i) = x; Y(i) = y; Z(i) = z; T(i) = t; t = t + dt; end Figure1 = figure(1); set(Figure1,'defaulttextinterpreter','latex'); subplot(2,2,1) plot(T, ErrorX, 'LineWidth', 2) title('Error in $x$-axis Position (m)') xlabel('time (sec)') ylabel('$x_{d}(t) - x(t)$', 'LineWidth', 2) subplot(2,2,2) plot(T, ErrorY, 'LineWidth', 2) title('Error in $y$-axis Position (m)') xlabel('time (sec)') ylabel('$y_{d}(t) - y(t)$', 'LineWidth', 2) subplot(2,2,3) plot(T, ErrorZ, 'LineWidth', 2) title('Error in $z$-axis Position (m)') xlabel('time (sec)') ylabel('$z_{d} - z(t)$', 'LineWidth', 2) subplot(2,2,4) plot(T, ErrorPsi, 'LineWidth', 2) title('Error in $\psi$ (m)') xlabel('time (sec)') ylabel('$\psi_{d} - \psi(t)$','FontSize',12); grid on Figure2 = figure(2); set(Figure2,'units','normalized','outerposition',[0 0 1 1]); figure(2) plot3(X,Y,Z, 'b') grid on hold on plot3(DesiredX, DesiredY, DesiredZ, 'r') pos = get(Figure2,'Position'); set(Figure2,'PaperPositionMode','Auto','PaperUnits','Inches','PaperSize',[pos(3),pos(4)]); print(Figure2,'output2','-dpdf','-r0'); For the trajectory code clear all; clc; fileID = fopen('xyTrajectory.txt','w'); angle = -pi; z = 0; t = 0; for i = 1:6000 if ( z < 2 ) z = z + 0.1; x = 0; y = 0; end if ( z >= 2 ) angle = angle + 0.1; angle = wrapToPi(angle); z = 2; end X(i) = x; Y(i) = y; Z(i) = z; fprintf(fileID,'%f \t %f \t %f\n',x, y, z); end fclose(fileID); plot3(X,Y,Z) grid on • $\phi, \theta$ are inside the intervall -90,90 deg. All the linearization's story is about the fact that the system doesn't go to much far from the linearization's point $\phi = 0 \theta = 0$. About the tracking problem: please try again with the following gains: P = 8, I = 0, D = 35 . I had a lot of problem due to the integrating term. And be sure that the timing is right. 'For example the point to be reached must be not be too far from the last point. Small steps. – Dave May 23 '15 at 10:33 • And if I were you I would still implement a non linear controller. I was really happy with the following one: math.ucsd.edu/~mleok/pdf/LeLeMc2010_quadrotor.pdf You need just a trajectory with position, velocity, acceleration. Snap and jerk are not necessary. That was the first controller I implemented, that was working "out of the box". I was using C++. If you use Matlab, then you have it really fast – Dave May 23 '15 at 10:39 • @Dave, I will try your suggestions. Is there a way to share your code in C++? My ultimate goal is to reimplement the system in C++. Also, which ODE solver did you use? Regarding the nonlinear controller, I've tried my best but always arcsin yields undefined number based on the approach in this link researchgate.net/publication/… – CroCo May 23 '15 at 15:15 • Cont., I have seen one of your posts that you are implementing backstepping controller based on the approach in the aforementioned paper, if you can help me to carry out the experiment, I will be more than happy. Thanks. – CroCo May 23 '15 at 15:18 • About timing, could you please elaborate a bit since everything in my code in one for loop. – CroCo May 23 '15 at 15:20 for i = 1:6000 if ( z < 2 ) z = z + 0.1; x = 0; y = 0; end if ( z >= 2 ) angle = angle + 0.1; angle = wrapToPi(angle); z = 2; end X(i) = x; Y(i) = y; Z(i) = z; fprintf(fileID,'%f \t %f \t %f\n',x, y, z); end there is a jump when the altitude reachs 2. The x coordinates goes in only one single step from the value 0 to the value stored in the variable radius. Please arrange your code and let the generation of a much more soft trajectory in your code, in which you have more interpolated points between the position 0 and the radius. As a general rule, you must not have too much distance between two contiguous points. In the case of a quadrotor the fact that those points too far away are, leads to a very huge pitch and/or roll angle (the quadrotor tries to do its best to reach the next point in a very small time, so it accelerates suddenly by setting a very steep angle). This causes the asin function to not work properly, since the arguments are not anymore in the wanted function's domain ($]-1,1[$) An example could be: for i = 1:6000 if ( z < 2 ) z = z + 0.1; x = 0; y = 0; end if ( z >= 2 ) angle = angle + 0.1; angle = wrapToPi(angle); x = ( i / 6000 ) * radius * cos(angle); y = ( i / 6000 ) * radius * sin(angle); z = 2; end X(i) = x; Y(i) = y; Z(i) = z; fprintf(fileID,'%f \t %f \t %f\n',x, y, z); end regards • this is useful, because it shows the error in the algorithm. it would be beneficial to also define some constraints for a "good, followable trajectory". – Gürkan Çetin Jul 23 '15 at 19:35
{"extraction_info": {"found_math": true, "script_math_tex": 0, "script_math_asciimath": 0, "math_annotations": 0, "math_alttext": 0, "mathml": 0, "mathjax_tag": 0, "mathjax_inline_tex": 1, "mathjax_display_tex": 1, "mathjax_asciimath": 1, "img_math": 0, "codecogs_latex": 0, "wp_latex": 0, "mimetex.cgi": 0, "/images/math/codecogs": 0, "mathtex.cgi": 0, "katex": 0, "math-container": 0, "wp-katex-eq": 0, "align": 3, "equation": 0, "x-ck12": 0, "texerror": 0, "math_score": 0.9678505659103394, "perplexity": 4598.823753662116}, "config": {"markdown_headings": true, "markdown_code": true, "boilerplate_config": {"ratio_threshold": 0.18, "absolute_threshold": 10, "end_threshold": 15, "enable": true}, "remove_buttons": true, "remove_image_figures": true, "remove_link_clusters": true, "table_config": {"min_rows": 2, "min_cols": 3, "format": "plain"}, "remove_chinese": true, "remove_edit_buttons": true, "extract_latex": true}, "warc_path": "s3://commoncrawl/crawl-data/CC-MAIN-2019-51/segments/1575540484477.5/warc/CC-MAIN-20191206023204-20191206051204-00411.warc.gz"}